You are currently on the new version of our website. Access the old version .
HealthcareHealthcare
  • Article
  • Open Access

24 June 2022

Detection of Skin Cancer Based on Skin Lesion Images Using Deep Learning

,
,
,
and
1
Department of Computer Engineering and Network, College of Computer and Information Sciences, Jouf University, Sakaka 72341, Al Jouf, Saudi Arabia
2
Electrical Engineering Department, Faculty of Engineering at Shoubra, Benha University, Cairo 4272077, Egypt
3
Faculty of Computer Science and Information Technology, Universiti Malaysia Sarawak, Kota Samarahan 94300, Malaysia
4
Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakaka 72341, Al Jouf, Saudi Arabia

Abstract

An increasing number of genetic and metabolic anomalies have been determined to lead to cancer, generally fatal. Cancerous cells may spread to any body part, where they can be life-threatening. Skin cancer is one of the most common types of cancer, and its frequency is increasing worldwide. The main subtypes of skin cancer are squamous and basal cell carcinomas, and melanoma, which is clinically aggressive and responsible for most deaths. Therefore, skin cancer screening is necessary. One of the best methods to accurately and swiftly identify skin cancer is using deep learning (DL). In this research, the deep learning method convolution neural network (CNN) was used to detect the two primary types of tumors, malignant and benign, using the ISIC2018 dataset. This dataset comprises 3533 skin lesions, including benign, malignant, nonmelanocytic, and melanocytic tumors. Using ESRGAN, the photos were first retouched and improved. The photos were augmented, normalized, and resized during the preprocessing step. Skin lesion photos could be classified using a CNN method based on an aggregate of results obtained after many repetitions. Then, multiple transfer learning models, such as Resnet50, InceptionV3, and Inception Resnet, were used for fine-tuning. In addition to experimenting with several models (the designed CNN, Resnet50, InceptionV3, and Inception Resnet), this study’s innovation and contribution are the use of ESRGAN as a preprocessing step. Our designed model showed results comparable to the pretrained model. Simulations using the ISIC 2018 skin lesion dataset showed that the suggested strategy was successful. An 83.2% accuracy rate was achieved by the CNN, in comparison to the Resnet50 (83.7%), InceptionV3 (85.8%), and Inception Resnet (84%) models.

1. Introduction

The uncontrollable development of tissues in a specific body area is known as cancer [1]. One of the most quickly spreading diseases in the world looks to be skin cancer. Skin cancer is a disease in which abnormal skin cells develop out of control [2]. In order to determine potential cancer therapies, early detection and accurate diagnosis are essential. Melanoma, the deadliest form of skin cancer, is responsible for most skin cancer-related deaths in developed countries. The major skin cancer types comprise basal cell carcinoma [3], squamous cell carcinoma [4], Merkel cell cancer [5], dermatofibroma [6], vascular lesion [7], and benign keratosis [8].
In order to diagnose abnormalities in various regions of the body, such as skin cancer [9], breast cancer [10], brain tumors [11], lung cancer [12], and stomach cancer [13], diagnostic imaging assessment plays an important part. According to the GLOBOCAN survey, there will be 19.2 million new cancer diagnoses and 9.9 million cancer deaths in 2020. Lung cancer is the leading cause of death (18.2%), followed by colorectal cancer (9.5%), liver cancer (8.4%), stomach cancer (7.8%), breast cancer (6.9%), esophageal cancer (5.5%), and pancreatic cancer (4.7%). The GLOBOCAN survey also points out more than half of cancer deaths occur in Asia, along with about 20% of cancer deaths occurring in Europe. Furthermore, the areas most affected by skin cancer around the globe are shown in Figure 1, with North America accounting for about half of the total.
Figure 1. Skin cancer cases globally (22 March 2022) [1].
To ensure better prognosis and death rates, early skin cancer identification is crucial, yet solid tumor detection typically relies mostly on screening mammography with inadequate sensitivity, which is then validated by clinical specimens. Cancer screening and treatment reaction evaluations are usually not appropriate uses for this approach [2,3]. An increasing number of healthcare providers are using artificial intelligence (AI) for medical diagnostics to improve and accelerate the diagnosis decision-making procedure [4]. However, despite some current evidence of improvement in this domain, the accurate assessment and adequate reporting of predicted flaws have been entirely or partly ignored by currently available AI research for clinical diagnosis.
Computer-aided design (CAD) can quickly, reliably, and consistently diagnose various disorders. CAD also provides the option for advanced tumor disease detection and protection that is both precise and cost-effective. Human organ disorders are typically assessed using a variety of imaging technologies, including magnetic resonance imaging (MRI) [5], positron emission tomography (PET) [6], and X-rays [7]. Computed tomography (CT) [8,9], dermatoscopy image analysis, clinical screening, and other approaches were initially used to visually diagnose skin lesions. Dermatologists with little expertise have shown reduced accuracy in skin lesion diagnostics [10,11,12]. The methods for physicians to evaluate and analyze lesion images are time-consuming, complex, subjective, and error-prone. This is mainly because the images of skin lesions are so complicated. Unambiguous identification of lesion pixels is essential to performing image analysis, for evaluation and awareness of skin lesions. Using machine learning approaches in computer vision has led to a significant advance in computer-aided diagnostic and prediction systems for skin cancer detection [13]. Image preprocessing and classification of lesion images are some of the main processes used to outline the entire cancer detection and diagnosis, as described in Figure 2 [14].
Figure 2. Process of cancer detection.
The exponential growth in processing power has led to tremendous advancements in computer vision technologies, particularly in the development of deep learning models such as CNN. The earliest possible detection of skin cancer is now required. Skin cancer is the second most common cancer (after breast cancer) in women between the ages of 30 and 35, and the most common cancer in women between the ages of 25 and 29, according to Dr. Lee [15], who serves several young patients with skin cancer. Early identification of skin cancer using deep learning outperformed human specialists in many computer vision challenges [15,16], resulting in reduced death rates. It is possible to get outstanding and cutting-edge processing and classification accuracy by including efficient formulations into deep learning techniques [17,18,19].
In order to correctly diagnose early cancer signs from lesion images, this study proposes a crossbred DL model for cancer classification and prediction. Preprocessing and classification are key components of the system under consideration. During the preprocessing phase, the entire intensity of the image is improved to decrease the inconsistencies among photos. The image is additionally scaled and standardized to fit the training model’s scale during this procedure. Many different metrics were used to evaluate the suggested model in the comparison studies. These metrics included precision and recall metrics, the F1-score, and the area under the curve (AUC). The publicly available, large-scale ISIC 2018 dataset comprises a massive number of lesion images with diagnosed cancer. Pretrained networks such as Resnet50, InceptionV3, and Inception Resnet were employed for comparison. A training process with varying configurations of training strategies (e.g., validation patience and data augmentation) was employed to boost the recommended technique’s universal efficiency and prevent overfitting.
The remainder of this paper is broken down as follows: Section 2 summarizes existing investigations, Section 3 describes the methods used to build the cancer dataset and the recommended system’s design requirements, Section 4 offers the findings of the study, and Section 5 finishes with the conclusion and suggestions for further studies.

3. Proposed System

A CNN model using images from the image data store is presented schematically to generate discriminative and relevant attribute interpretations for the cancer detection technique, as shown in Algorithm 1. To begin, a basic explanation of the used dataset is provided. Moreover, the details of the implementation of proposed model, including preprocessing techniques and the basic architecture, are presented.
Algorithm 1: Processes in the mechanism suggested
Let Ƀ = lesion image, aug = augmentation, ppr = preprocessing, ig = image, ig€ = image enhancement algorithm (ESRGAN), rt = rotation, sc = scaling, rl = reflection, and sh = shifting method
Input: {Lesion image Ƀ}
Output: {confusion matrix, accuracy, precision, ROC, F1, AUC, recall}
Step 1: Browse(Ƀ)
Step 2: Implement (ppr (ig))
2.1. Operate (ig€)_
2.2. aug(ig) w.r.t. rt, sc, rl, sh
  2.2.1. perform rt
  2.2.2. perform sc
  2.2.3. perform rl
  2.2.4. perform sh
2.3. Resize (ig)/224*24*3
2.4 Normalize pixelvalue (ig)/interval [0,1]
Step 2: Split (dataset)/training, testing, and validating
Step 3: Train CNN model
Step 4: Train pretrained models (Resnet, Inception, Inception Resnet)
  4.1 Fine-tune model parameters (freeze layers, learning rate, epochs, batch size)
Step 5: Compute VPM (confusion matrix, accuracy, precision, ROC, F1, AUC, recall)
Step 6: Evaluation (existing work)

3.1. ISIC 2018 Image Dataset

Data are at the core of DL, representing what these learning techniques run on. Cancer is a unique disease, and there have already been many datasets published. We used lesion images from publicly accessible image databases of identified affected individuals. The ISIC 2018 dataset was utilized for training the proposed approach, which contained 10,015 training and 1512 test images for a total of 11,527 images [30]. ISIC 2018 provided the ground-truth data only for the training set, consisting of seven classes, melanoma, melanocytic nevus, basal cell carcinoma, squamous cell carcinoma, vascular lesions, dermatofibroma, and benign keratosis, as shown in Figure 3.
Figure 3. Classes of ISIC2018 dataset.
We applied the proposed CNN model to the ISIC 2018 skin lesion classification challenge test set; our data store consisted of 3533 lesion scans where 1760 of them are benign and 1773 are malignant, and we tested the proposed system using a total of 960 images consisting of 360 benign and 300 malignant cases. The lesion images were acquired from an openly accessible data repository ISIC 2018 [31]. For evaluation, the authors obtained radiological scans from many legitimate databases of cancer incidences; images from this source are used in most cancer diagnostics. The database, which is updated regularly, offers a free library of cancer cases and lesion images. The Kaggle list “Lesion Images” was used to collect lesion images; 3533 images from these sources are included in the ISIC2018 collection [41]. Figure 4 shows various lesion image examples from the ISIC2018 dataset, demonstrating the collection’s diversity of patient situations. It was decided to build ISIC2018 because the library is openly available and openly available to the academic communities and the public society.
Figure 4. Lesion images from ISIC2018 dataset.

3.2. Image Preprocessing

This process involved data augmentation, image improvement using (ESRGAN), image resizing, and normalization.

3.2.1. ESRGAN

Approaches such as super-resolution generative adversarial network enhanced SRGAN [42] can help improve the detection of skin lesions. The enhanced edition of the super-resolution GAN (Ledig et al.) [43] uses a resilient-in-residual block instead of a basic residual network or a simple convolution trunk when it comes to microscopic-level gradients. Additionally, the model does not have a batch normalization layer for smoothing down the image. Accordingly, the sharp edges of the image artefacts can be better approximated in the images produced by ESRGAN. When determining if an image is real or false, ESRGAN employs a relativistic discriminator https://arxiv.org/pdf/1807.00734.pdf (accesed on 10 April 2022). This method yields more accurate results. Perceptual differences between the actual and false images are combined with the relativistic average loss and pixelwise absolute difference between the real and fake images as the loss function during adversarial training. A two-phase training scheme is used to sharpen the generator’s skills. This reduces the pixelwise L1 distance between the input and target high-resolution image to avoid local minima when beginning with complete randomization in the first phase of the algorithm.
In the second stage, the goal is to refine and improve the reconstructed images of the smallest artefacts. The final trained model is interpolated between the L1 loss and the adversarially trained models for a photorealistic reconstruction.
A discriminator network was trained to distinguish between super-resolved images and actual photo images. By rearranging the lightness elements in the source image’s histogram, an evolutionary contrast enhancement algorithm was used to strengthen the lesion picture’s minutiae, textures, and poor contrast. As a result, this method enhanced the appearance of borders and arcs in each section of the picture, as shown in Figure 5, while simultaneously increasing the image’s contrast level.
Figure 5. Images after the enhancement process.

3.2.2. Augmentation

For each image in the dataset, upgraded images with associated masks including rotation, reflection, shifting, brightness, and resizing were produced. Detection and assessment are restricted by the poor quality of raw lesion images generated by electronic detectors. There were a total of 1440 benign and 1197 malignant training images. After conducting augmentation, there were a total of 1760 benign and 1773 malignant images. The imbalanced distribution of classes was addressed by performing oversampling on the malignant images.
To avoid biased prediction consequences, the ISIC2018 dataset was split into three mutually distinct sets (training, validation, and evaluation sets) to address the overfitting issue caused by the short number of training photographs. The output of the image augmentation process after applying different augmentation parameters is shown in Figure 6.
Figure 6. Output of the proposed image augmentation process.

3.2.3. Data Preparation

Image acquisition factors can vary due to the fact that certain photos in the dataset have low pixel dimensions, and all images should be resized. As a result, the image’s luminance and size can change dramatically. Each acquisition tool has its own unique set of criteria; hence, the lesion image dataset is likely to contain a variety of images. In order to verify that the data were consistent and free of noise, the pixel strength of all images was standardized within the interval [−1, 1]. Normalization computed using Equation (1) ensured that the model was less susceptible to minor weight changes, facilitating its improvement. Below, Inorm, MinI, and MaxI represent image, normalize, minimum, and maximum, respectively.
I n o r m = ( I M i n I ) ( 2 M a x I M i n I ) 1

3.3. Proposed CNN for ISIC2018 Detection

Due to the enormous number of hyperparameters and structures that need to be accounted for, DL models face significant difficulty (e.g., learning rate, number of frozen layers, batch size, and number of epochs). Several hyperparameter values were tested to see how they affected the efficiency of the suggested systems. The proposed CNN model consisted of three layers, as shown in Figure 7. As depicted in Figure 7, the skin cancer detection system employed a transfer DL strategy to learn discriminative and informative feature representations from preprocessed images in the image dataset.
Figure 7. An illustration of the skin cancer detection technique.
The presented system’s core architecture was built on three learning models: Resnet50, Inception, and Inception Resnet50.

3.3.1. Resnet50

Resnet50is a 50-layer residual network [44]. Several difficulties emerged when scholars tried to apply the adage “the deeper the better” to deep learning methods. In comparison to networks having 20–30 layers, the deep network with 52 layers produced subpar outcomes, disproving the theory that “the deeper the network, the higher the network’s efficiency”. Resnet-50, a residual learning feature of the CNN model, was developed by experts. The residual unit is compensated for by using a conventional layer with a skip connection. Tying a layer’s incoming signal to a certain layer’s output using a skip connection is possible. The residual units allowed the training of a 152-layer model that was used to win the 2015 LSVRC2015 challenge. There is less of a learning curve because of its novel residual structure. A top five false-positive rate of <3.6% can be achieved using this machine.

3.3.2. Inception V3

An essential feature of the Inception module is its capacity to perform multiresolution processing [45]. To capture characteristics in standard CNN models, kernels with distinct receptive areas are utilized in certain layers. In an inception model, on the other hand, many kernels with differing receptive fields are employed in tandem to retrieve features of various sizes. The Inception module’s outcome is created by stacking the parallel features that were extracted one on top of the other. The subsequent convolutional layer of the CNN uses the rich attribute maps produced by the Inception module’s merged result. Because of this, the Inception module’s effectiveness in medical imaging, specifically on lesion pictures, is exceptional [46].

3.3.3. Inception Resnet

The Resnet50 and Inception frameworks were combined into one model to classify hyperspectral images. More than one million photos from the ImageNet collection were used to train the Inception ResnetV2 convolutional neural network. In total, there are 164 layers in this network, and it is capable of classifying photos into 1000 different object categories. Consequently, the network has amassed a diverse set of feature descriptions. The network accepts a 299-by-299-pixel picture as an input and gives a set of classifiers.

4. Experimental Results

Experiments were conducted on the ISIC2018 dataset to illustrate the effectiveness of the suggested DL systems and to compare their findings to those of the current state of the art.

4.1. Parameter Setting and Experimental Evaluation Index

Simulations on the ISIC2018 dataset were carried out to illustrate the performance of the suggested DL systems and to compare their results to the current state of the art. On a linux desktop with a GPU RTX3060 and 8GB of RAM, the TensorFlow Keras program for the present scheme was tested. Training and testing sets were separated using a ratio of 80 to 20%, as shown in Figure 8. The training set contained 1760 benign and 1773 malignant images, while the testing set comprised 360 benign and 300 malignant images.
Figure 8. Distribution of dataset.
The suggested training set comprised an 80% randomized array of lesion images. All testing was conducted using this set. Then, 10% of the data were used for verification throughout the learning phase. The weight combinations with the greatest accuracy values were retained. On the ISIC2018 dataset, the Adam optimizer was used to pretrain the suggested architecture, which employs a learning rate technique that slows down learning when it becomes static for an extended period (i.e., validation patience). Furthermore, we implemented a batch rebalancing technique to improve the prevalence of infection forms during the batching process. The hyperparameters and their values used by the Adam optimizer for training are presented in Table 2.
Table 2. Hyperparameters of Adam optimizer.

4.2. Performance Assessment

This part of the study includes an in-depth explanation of the evaluation metrics utilized and their outcomes. Classifier accuracy (Acc) is the primarily used statistic for evaluating classification effectiveness. It is described as the number of instances (images) categorized accurately divided by the number of examples (images) in the dataset under analysis, as expressed in Equation (2). There are two used metrics generally used for evaluating the effectiveness of image categorization systems: precision (Pr) and recall (Rc). Precision is a measure of how many classified photos are correctly labeled compared to the total number of images, as expressed in Equation (3). Recall is the percentage of successfully categorized images in the database compared to the number of associated images, as expressed in Equation (4). The F-score is the harmonic mean of precision and recall; a greater value is an indicator of the system’s ability to forecast the future. The effectiveness of systems cannot be judged just on the basis of precision or recall. Equation (5) is the mathematical representation of the F-score (Fs).
A c c = T p + T n T p + T n + F p   + F n ,
P r = T p T p + F p   ,
R c = T p T p + F n   ,
F s = 2 × ( P r × R c P r + R c ) ,
where T p indicates a true positive, T n indicates a true negative, F p   indicates a false positive, and F n indicates a false negative.

4.3. Performance of Different DCNN Models

Different DCNNs (CNN, Resnet50, Inception, and Inception Resnet) were implemented for training and testing tasks on the ISIC 2018 skin lesion classification challenge dataset. Using an 80–20 split between training and testing, the results are presented of various assessments on the ISIC2018 dataset for the suggested systems. This division was chosen to minimize the impact on execution time. CNN, Resnet50, Inception, and Inception Resnet models were trained for 50 epochs employing 10% of the training set as a validation set, with a batch size ranging from 2 to 32, and learning rates varying from 1 × 104 to 1 × 106. Moreover, fine-tuning was performed regarding Resnet50, Inception, and Inception Resnet by freezing different numbers of layers to achieve the best accuracy. In order to train the models using similar parameters (runs 1–3, Table 3, Table 4, Table 5 and Table 6), we used several runs (three runs for the similar parameters) to construct the model ensemble. The accuracy fluctuated from run to run since the weights were generated at random for each run; only the best run outcome was saved. Table 3, Table 4, Table 5 and Table 6 show the accuracy results for the proposed CNN.
Table 3. Average accuracy of CNN model using ISIC dataset (optimizer = Adam, learning rate = 1 × 10−6).
Table 4. Average accuracy of CNN model using ISIC dataset (optimizer = Adam, learning rate = 1 × 10−5).
Table 5. Average accuracy of CNN model using ISIC dataset (optimizer = Adam, learning rate = 1 × 10−4).
Table 6. Best accuracy after fine-tuning using several transfer learning models.
Figure 9 and Figure 10 show the confusion matrices of the benign and malignant cancer findings using CNN, Resnet50, InceptionV3, and Inception Resnet. A total of 37 benign infected pictures were misinterpreted as malignant, while 74 malignant images were inaccurately classified as benign when using the CNN, in contrast to 62 benign images misclassified as malignant and 34 malignant images misinterpreted as benign when using InceptionV3.
Figure 9. Best confusion matrix of CNN.
Figure 10. Best confusion matrix of InceptionV3.
Diagnostic effectiveness was assessed using the AUC receiver operating characteristic curve (ROC), depicting the model’s categorization effectiveness as a function of two parameters: true positives and false positives. The AUC is calculated as the area under the ROC curve covered by small trapezoidal segments. As shown in Figure 11, we performed ROC analyses using a CNN model with an area of 0.83. The best-case ROC outcome for the suggested model after fine-tuning using InceptionV3 is shown in Figure 12.
Figure 11. ROC curve for CNN model.
Figure 12. ROC curve for InceptionV3.
It is now clear that the proposed strategy can be used in real-world settings to help radiologists diagnose cancer infection more correctly by utilizing lesion images, while simultaneously lowering their burden.

4.4. Comparison with Other Methods

A comparison of the suggested method’s efficacy to that of existing methods was performed to better demonstrate its viability. Table 7 shows that our strategy was superior to other networks in terms of performance. In the proposed approach, the Inception model had an overall accuracy rate of 85.7%, outperforming the existing models.
Table 7. Comparison with other methods.

4.5. Discussion

According to our findings, none of the other approaches could match our level of precision. We attribute this to (i) ESRGAN’s overall resolution improvement, (ii) the fine-tuning to learn particular dataset aspects, and (iii) our use of numerous architectures, each with a different capacity to generalize and adapt to various data. The lack of unique medical image features meant that the transfer learning architectures could not achieve a higher level of classification accuracy. Despite being better at classifying natural pictures, Resnet50’s classification accuracy was lower than that of InceptionV3 when used on medical images. These findings suggest that shallower networks, such as that in InceptionV3, have more generalizable properties that may be used for a larger variety of imagery. On the other hand, deeper networks such as Resnet50 and Inception Resnet learn abstract characteristics that may be applied to any domain. Because the features of InceptionV3 are less semantically suited to natural pictures, they are more generalizable and adaptable when applied to medical images (compared to Resnet50 and Inception Resnet). Furthermore, fine-tuning the networks improved the accuracy of the four models. Compared to Resnet50 and Inception Resnet, InceptionV3’s accuracy increased the most. According to the results of this study, deep networks are more likely to acquire relevant features when fine-tuned on a smaller dataset than shallow networks. The confusion matrices and numerical data shown in Figure 9 and Figure 10 indicate that the suggested procedures were sufficient.

5. Conclusions and Future Work

By analyzing images of lesions on the skin, we developed a technique for quickly and accurately diagnosing both benign and malignant forms of cancer. The suggested system uses image enhancement approaches to boost the luminance of the lesion image and reduce noise. Resnet50, InceptionV3, and Resnet Inception were all trained on the upper edge of the preprocessed lesion medical images to prevent overfitting, as well as improve the overall competencies of the suggested DL methods. A lesion image dataset called the ISIC2018 dataset was used to test the proposed system’s performance. In the proposed approach, the Inception model had an overall accuracy rate of 85.7%, which is comparable to that of experienced dermatologists. In addition to experimenting with several models (designed CNN, Resnet50, InceptionV3, and Inception Resnet), this study’s innovation and contribution are the use of ESRGAN as a preprocessing step. Our designed model showed results comparable to the pretrained model. According to the comparative research, the proposed system outperformed current models. To establish the effectiveness of the suggested method, there is a need to conduct tests on a large, complex dataset that includes many cancer cases. It is possible that, in the future, we will employ Densenet, VGG, or AlexNet to analyze the cancer dataset.

Author Contributions

Conceptualization, W.G., M.H. and N.U.S.; methodology, W.G. and N.U.S.; software, W.G.; validation, G.A.-W., M.H. and N.U.S.; formal analysis, W.G., M.H. and N.U.S.; resources, G.A.-W. and N.Z.J.; data curation, W.G., M.H. and N.U.S.; writing—original draft preparation, W.G., M.H. and N.U.S.; writing—review and editing, W.G., M.H., N.U.S., G.A.-W. and N.Z.J.; visualization, G.A.-W.; supervision, M.H. and N.Z.J.; project administration, M.H. and N.Z.J.; funding acquisition, G.A.-W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data set is available at https://challenge.isic-archive.com/data/, 20 April 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization. Global Health Observatory; World Health Organization: Geneva, Switzerland, 2022. [Google Scholar]
  2. Han, H.S.; Choi, K.Y. Advances in nanomaterial-mediated photothermal cancer therapies: Toward clinical applications. Biomedicines 2021, 9, 305. [Google Scholar] [CrossRef]
  3. Fuzzell, L.N.; Perkins, R.B.; Christy, S.M.; Lake, P.W.; Vadaparampil, S.T. Cervical cancer screening in the United States: Challenges and potential solutions for underscreened groups. Prev. Med. 2021, 144, 106400. [Google Scholar] [CrossRef] [PubMed]
  4. Ting, D.S.; Liu, Y.; Burlina, P.; Xu, X.; Bressler, N.M.; Wong, T.Y. AI for medical imaging goes deep. Nat. Med. 2018, 24, 539–540. [Google Scholar] [CrossRef] [PubMed]
  5. Wolf, M.; de Boer, A.; Sharma, K.; Boor, P.; Leiner, T.; Sunder-Plassmann, G.; Moser, E.; Caroli, A.; Jerome, N.P. Magnetic resonance imaging T1-and T2-mapping to assess renal structure and function: A systematic review and statement paper. Nephrol. Dial. Transplant. 2018, 33 (Suppl. S2), ii41–ii50. [Google Scholar] [CrossRef] [Green Version]
  6. Hooker, J.M.; Carson, R.E. Human positron emission tomography neuroimaging. Annu. Rev. Biomed. Eng. 2019, 21, 551–581. [Google Scholar] [CrossRef] [PubMed]
  7. Jaiswal, A.K.; Tiwari, P.; Kumar, S.; Gupta, D.; Khanna, A.; Rodrigues, J.J. Identifying pneumonia in chest X-rays: A deep learning approach. Measurement 2019, 145, 511–518. [Google Scholar] [CrossRef]
  8. Morawitz, J.; Dietzel, F.; Ullrich, T.; Hoffmann, O.; Mohrmann, S.; Breuckmann, K.; Herrmann, K.; Buchbender, C.; Antoch, G.; Umultu, L.; et al. Comparison of nodal staging between CT, MRI, and [18F]-FDG PET/MRI in patients with newly diagnosed breast cancer. Eur. J. Nucl. Med. Mol. Imaging 2022, 49, 992–1001. [Google Scholar] [CrossRef] [PubMed]
  9. Jinzaki, M.; Yamada, Y.; Nagura, T.; Nakahara, T.; Yokoyama, Y.; Narita, K.; Ogihara, N.; Yamada, M. Development of upright computed tomography with area detector for whole-body scans: Phantom study, efficacy on workflow, effect of gravity on human body, and potential clinical impact. Investig. Radiol. 2020, 55, 73. [Google Scholar] [CrossRef]
  10. Celebi, M.E.; Codella, N.; Halpern, A. Dermoscopy image analysis: Overview and future directions. IEEE J. Biomed. Health Inform. 2019, 23, 474–478. [Google Scholar] [CrossRef]
  11. Barata, C.; Celebi, M.E.; Marques, J.S. A survey of feature extraction in dermoscopy image analysis of skin cancer. IEEE J. Biomed. Health Inform. 2018, 23, 1096–1109. [Google Scholar] [CrossRef]
  12. Adeyinka, A.A.; Viriri, S. Skin lesion images segmentation: A survey of the state-of-the-art. In Proceedings of the International Conference on Mining Intelligence and Knowledge Exploration, Cluj-Napoca, Romania, 22 December 2018; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  13. Al-Masni, M.A.; Al-Antari, M.A.; Choi, M.-T.; Han, S.-M.; Kim, T.-S. Skin lesion segmentation in dermoscopy images via deep full resolution convolutional networks. Comput. Methods Programs Biomed. 2018, 162, 221–231. [Google Scholar] [CrossRef] [PubMed]
  14. Ünver, H.M.; Ayan, E. Skin lesion segmentation in dermoscopic images with combination of YOLO and grabcut algorithm. Diagnostics 2019, 9, 72. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Hu, Z.; Tang, J.; Wang, Z.; Zhang, K.; Zhang, L.; Sun, Q. Deep learning for image-based cancer detection and Diagnosis—A survey. Pattern Recognit. 2018, 83, 134–149. [Google Scholar] [CrossRef]
  16. Fujisawa, Y.; Otomo, Y.; Ogata, Y.; Nakamura, Y.; Okiyama, N.; Ohara, K.; Fujimoto, M.; Fujita, R.; Ishitsuka, Y.; Watanabe, R. Deep-learning-based, computer-aided classifier developed with a small dataset of clinical images surpasses board-certified dermatologists in skin tumour diagnosis. Br. J. Dermatol. 2019, 180, 373–381. [Google Scholar] [CrossRef]
  17. Adegun, A.; Viriri, S. Deep learning techniques for skin lesion analysis and melanoma cancer detection: A survey of state-of-the-art. Artif. Intell. Rev. 2021, 54, 811–841. [Google Scholar] [CrossRef]
  18. Iqbal, S.; Siddiqui, G.F.; Rehman, A.; Hussain, L.; Saba, T.; Tariq, U.; Abbasi, A.A. Prostate cancer detection using deep learning and traditional techniques. IEEE Access 2021, 9, 27085–27100. [Google Scholar] [CrossRef]
  19. Dildar, M.; Akram, S.; Mahmood, A.R.; Mahnashi, M.H.; Alsaiari, S.A.; Irfan, M.; Khan, H.U.; Saeed, A.H.M.; Ramzan, M.; Alraddadi, M.O. Skin cancer detection: A review using deep learning techniques. Int. J. Environ. Res. Public Health 2021, 18, 5479. [Google Scholar] [CrossRef]
  20. Vaishnavi, K.; Ramadas, M.A.; Chanalya, N.; Manoj, A.; Nair, J.J. Deep learning approaches for detection of COVID-19 using chest X-ray images. In Proceedings of the 2021 Fourth International Conference on Electrical, Computer and Communication Technologies (ICECCT), Piscataway, NJ, USA, 15–17 September 2021. [Google Scholar]
  21. Duc, N.T.; Lee, Y.-M.; Park, J.H.; Lee, B. An ensemble deep learning for automatic prediction of papillary thyroid carcinoma using fine needle aspiration cytology. Expert Syst. Appl. 2022, 188, 115927. [Google Scholar] [CrossRef]
  22. Kassem, M.A.; Hosny, K.M.; Damaševičius, R.; Eltoukhy, M.M. Machine learning and deep learning methods for skin lesion classification and Diagnosis: A systematic review. Diagnostics 2021, 11, 1390. [Google Scholar] [CrossRef]
  23. Abayomi-Alli, O.O.; Damasevicius, R.; Misra, S.; Maskeliunas, R.; Abayomi-Alli, A. Malignant skin melanoma detection using image augmentation by oversamplingin nonlinear lower-dimensional embedding manifold. Turk. J. Electr. Eng. Comput. Sci. 2021, 29, 2600–2614. [Google Scholar] [CrossRef]
  24. Kadry, S.; Taniar, D.; Damaševičius, R.; Rajinikanth, V.; Lawal, I.A. Extraction of abnormal skin lesion from dermoscopy image using VGG-SegNet. In Proceedings of the 2021 Seventh International Conference on Bio Signals, Images, and Instrumentation (ICBSII), Chennai, India, 25 March 2021. [Google Scholar]
  25. Hekler, A.; Utikal, J.S.; Enk, A.H.; Hauschild, A.; Weichenthal, M.; Maron, R.C.; Berking, C.; Haferkamp, S.; Klode, J.; Schadendorf, D.; et al. Superior skin cancer classification by the combination of human and artificial intelligence. Eur. J. Cancer 2019, 120, 114–121. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Reis, H.C.; Turk, V.; Khoshelham, K.; Kaya, S. InSiNet: A deep convolutional approach to skin cancer detection and segmentation. Med. Biol. Eng. Comput. 2022, 7, 1–20. [Google Scholar] [CrossRef] [PubMed]
  27. Nawaz, M.; Mehmood, Z.; Nazir, T.; Naqvi, R.A.; Rehman, A.; Iqbal, M.; Saba, T. Skin cancer detection from dermoscopic images using deep learning and fuzzy k-means clustering. Microsc. Res. Tech. 2022, 85, 339–351. [Google Scholar] [CrossRef] [PubMed]
  28. Humayun, M.; Sujatha, R.; Almuayqil, S.N.; Jhanjhi, N.Z. A Transfer Learning Approach with a Convolutional Neural Network for the Classification of Lung Carcinoma. Healthcare 2022, 10, 1058. [Google Scholar] [CrossRef]
  29. Adegun, A.A.; Viriri, S. Deep learning-based system for automatic melanoma detection. IEEE Access 2019, 8, 7160–7172. [Google Scholar] [CrossRef]
  30. Codella, N.; Rotemberg, V.; Tschandl, P.; Celebi, M.E.; Dusza, S.; Gutman, D.; Helba, B.; Kalloo, A.; Liopyris, K.; Marchetti, M.; et al. Skin lesion analysis toward melanoma detection 2018, a challenge hosted by the international skin imaging collaboration (isic). arXiv 2019, arXiv:1902.03368. [Google Scholar]
  31. Kinyanjui, N.M.; Odonga, T.; Cintas, C.; Codella, N.C.; Panda, R.; Sattigeri, P.; Varshney, K.R. Fairness of classifiers across skin tones in dermatology. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4 October 2020; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  32. Ech-Cherif, A.; Misbhauddin, M.; Ech-Cherif, M. Deep neural network based mobile dermoscopy application for triaging skin cancer detection. In Proceedings of the 2019 2nd International Conference on Computer Applications & Information Security (ICCAIS), Riyadh, Saudi Arabia, 1 May 2019. [Google Scholar]
  33. Le, D.N.; Le, H.X.; Ngo, L.T.; Ngo, H.T. Transfer learning with class-weighted and focal loss function for automatic skin cancer classification. arXiv 2020, arXiv:2009.05977. [Google Scholar]
  34. Dorj, U.-O.; Lee, K.-K.; Choi, J.-Y.; Lee, M. The skin cancer classification using deep convolutional neural network. Multimed. Tools Appl. 2018, 77, 9909–9924. [Google Scholar] [CrossRef]
  35. Rahman, M.M.; Nasir, M.K.; Nur, A.; Khan, S.I.; Band, S.; Dehzangi, I.; Beheshti, A.; Rokny, H.A. Hybrid Feature Fusion and Machine Learning Approaches for Melanoma Skin Cancer Detection; UNSW: Sydney, Australia, 2022. [Google Scholar]
  36. Murugan, A.; Nair, S.A.H.; Preethi, A.A.P.; Kumar, K.S. Diagnosis of skin cancer using machine learning techniques. Microprocess. Microsyst. 2021, 81, 103727. [Google Scholar] [CrossRef]
  37. Ali, M.S.; Miah, M.S.; Haque, J.; Rahman, M.M.; Islam, M.K. An enhanced technique of skin cancer classification using deep convolutional neural network with transfer learning models. Mach. Learn. Appl. 2021, 5, 100036. [Google Scholar] [CrossRef]
  38. Guan, Q.; Wang, Y.; Ping, B.; Li, D.; Du, J.; Qin, Y.; Lu, H.; Wan, X.; Xiang, J. Deep convolutional neural network VGG-16 model for differential diagnosing of papillary thyroid carcinomas in cytological images: A pilot study. J. Cancer 2019, 10, 4876. [Google Scholar] [CrossRef] [PubMed]
  39. Rajput, G.; Agrawal, S.; Raut, G.; Vishvakarma, S.K. An accurate and noninvasive skin cancer screening based on imaging technique. Int. J. Imaging Syst. Technol. 2022, 32, 354–368. [Google Scholar] [CrossRef]
  40. Kanani, P.; Padole, M. Deep learning to detect skin cancer using google colab. Int. J. Eng. Adv. Technol. Regul. Issue 2019, 8, 2176–2183. [Google Scholar] [CrossRef]
  41. MNOWAK061. Skin Lesion Dataset. ISIC2018 Kaggle Repository. 2021. Available online: https://www.kaggle.com/datasets/mnowak061/isic2018-and-ph2-384x384-jpg (accessed on 10 April 2022).
  42. Takano, N.; Alaghband, G. Srgan: Training dataset matters. arXiv 2019, arXiv:1903.09922. [Google Scholar]
  43. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21 July 2017. [Google Scholar]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 30 June 2016. [Google Scholar]
  45. Wang, C.; Chen, D.; Hao, L.; Liu, X.; Zeng, Y.; Chen, J.; Zhang, G. Pulmonary image classification based on inception-v3 transfer learning model. IEEE Access 2019, 7, 146533–146541. [Google Scholar] [CrossRef]
  46. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-ResNet and the impact of residual connections on learning. arXiv 2018, arXiv:1602.07261. [Google Scholar]
  47. Foahom Gouabou, A.C.; Damoiseaux, J.-L.; Monnier, J.; Iguernaissi, R.; Moudafi, A.; Merad, D. Ensemble Method of Convolutional Neural Networks with Directed Acyclic Graph Using Dermoscopic Images: Melanoma Detection Application. Sensors 2021, 21, 3999. [Google Scholar] [CrossRef]
  48. Lopez, A.R.; Giro-i-Nieto, X.; Burdick, J.; Marques, O. Skin lesion classification from dermoscopic images using deep learning techniques. In Proceedings of the 2017 13th IASTED International Conference on Biomedical Engineering (BioMed), Innsbruck, Austria, 20 February 2017. [Google Scholar]
  49. Harangi, B. Skin lesion classification with ensembles of deep convolutional neural networks. J. Biomed. Inform. 2018, 86, 25–32. [Google Scholar] [CrossRef]
  50. Kim, C.-I.; Hwang, S.-M.; Park, E.-B.; Won, C.-H.; Lee, J.-H. Computer-Aided Diagnosis Algorithm for Classification of Malignant Melanoma Using Deep Neural Networks. Sensors 2021, 21, 5551. [Google Scholar] [CrossRef]
  51. Alnowami, M. Very Deep Convolutional Networks for Skin Lesion Classification. J. King Abdulaziz Univ. Eng. Sci. 2019, 30, 43–54. [Google Scholar]
  52. Ameri, A. A deep learning approach to skin cancer detection in dermoscopy images. J. Biomed. Phys. Eng. 2020, 10, 801. [Google Scholar] [CrossRef] [PubMed]
  53. Sae-Lim, W.; Wettayaprasit, W.; Aiyarak, P. Convolutional neural networks using MobileNet for skin lesion classification. In Proceedings of the 2019 16th International Joint Conference on Computer Science and Software Engineering (JCSSE), Chonburi, Thailand, 10 July 2019. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.