Next Article in Journal
Neo-Adjuvant Therapy for Metastatic Melanoma
Previous Article in Journal
Enhancing Image-Guided Radiation Therapy for Pancreatic Cancer: Utilizing Aligned Peak Response Beamforming in Flexible Array Transducers
Previous Article in Special Issue
Prognostic Factors and Nomogram for Choroid Plexus Tumors: A Population-Based Retrospective Surveillance, Epidemiology, and End Results Database Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Skin Cancer Recognition Using Unified Deep Convolutional Neural Networks

by
Nasser A. AlSadhan
1,*,
Shatha Ali Alamri
2,
Mohamed Maher Ben Ismail
1 and
Ouiem Bchir
1
1
Computer Science Department, College of Computer and Information Sciences, King Saud University, Riyadh 12372, Saudi Arabia
2
Faculty of Computers and Information Technology, University of Tabuk, Tabuk 47512, Saudi Arabia
*
Author to whom correspondence should be addressed.
Cancers 2024, 16(7), 1246; https://doi.org/10.3390/cancers16071246
Submission received: 31 January 2024 / Revised: 13 March 2024 / Accepted: 19 March 2024 / Published: 22 March 2024
(This article belongs to the Special Issue Rethinking Cancer Epidemiology, Aging and Machine Learning)

Abstract

:

Simple Summary

According to the World Fund for Research on Cancer, skin cancer is one of the most common cancers. The early diagnosis of skin cancer lesions plays an essential role in the patient’s recovery. Nevertheless, recognizing skin cancer and differentiating it from benign skin lesions is a challenging task for dermatologists due to the visual similarities of benign nevi, seborrheic keratoses, and malignant melanomas. In this context, image-based skin lesion recognition systems have appeared as a solution to recognize these lesions and therefore reduce the number of biopsy procedures. This research investigated the performance of the latest versions of the You Only Look Once (YOLO) deep learning models. Unlike classification-based solutions, the proposed YOLO-based approach locates the skin lesions and categorizes them into the predefined classes. The experiments were conducted using 2750 images from the publicly accessible International Skin Imaging Collaboration (ISIC) dataset.

Abstract

The incidence of skin cancer is rising globally, posing a significant public health threat. An early and accurate diagnosis is crucial for patient prognoses. However, discriminating between malignant melanoma and benign lesions, such as nevi and keratoses, remains a challenging task due to their visual similarities. Image-based recognition systems offer a promising solution to aid dermatologists and potentially reduce unnecessary biopsies. This research investigated the performance of four unified convolutional neural networks, namely, YOLOv3, YOLOv4, YOLOv5, and YOLOv7, in classifying skin lesions. Each model was trained on a benchmark dataset, and the obtained performances were compared based on lesion localization, classification accuracy, and inference time. In particular, YOLOv7 achieved superior performance with an Intersection over Union (IoU) of 86.3%, a mean Average Precision (mAP) of 75.4%, an F1-measure of 80%, and an inference time of 0.32 s per image. These findings demonstrated the potential of YOLOv7 as a valuable tool for aiding dermatologists in early skin cancer diagnosis and potentially reducing unnecessary biopsies.

1. Introduction

Skin cancer represents a global public health crisis with an ever-increasing incidence. It stands as the 19th most common cancer globally, with a disconcerting upward trend [1]. Skin cancer weighs heavily on white-skinned individuals of both sexes, particularly in the United States, where the number of diagnosed cases has skyrocketed in recent decades [2,3]. While preventative measures like UV exposure reduction and screening programs have been implemented, the global prevalence of skin cancer continues its relentless ascent [4]. In fact, an early diagnosis of skin cancer is not merely a matter of statistics. It holds the key to extending life expectancy and improving treatment outcomes. However, the timely and accurate identification of skin cancer remains a challenging, mainly due to the diverse complexity and rapid spread of certain cancer types [5]. In addition, the limitations of traditional diagnostic methods, such as dermoscopy, has further exacerbated this issue. In particular, despite the valuable assistance offered by dermoscopy, its subjectivity, time-consuming procedures, and diagnosis subjectivity have limited its widespread use and promoted the investigation of alternative approaches [6]. On the other hand, coupled spectroscopic and imaging techniques, such as real-time Raman spectroscopy, have shown promise in conjunction with conventional approaches [7]. However, their ability to accurately discriminate between malignant and benign lesions, particularly those with overlapping characteristics, such as malignant melanoma and seborrheic keratosis, remains below expectation [8]. Additionally, the intra-class variations in terms of size, color, and form make the identification problem even more acute [9].
The transformative power of Machine Learning (ML), particularly deep learning, has emerged as a beacon of hope for diverse challenges in the medical field. In particular, ML techniques have revolutionized skin cancer identification through rapid, cost-effective, and accurate diagnoses that couple image classification techniques with deep learning models [10].
One should note that the motivation for this research stems from the profound impact an early diagnosis can have on patients. In fact, early detection offers a critical window of opportunity to (i) improve the prognosis, (ii) significantly enhance the likelihood of successful treatment, and (iii) ultimately save human lives [11]. Recent advancements in YOLO models have represented a potentially transformative alternative for skin cancer detection [12]. In particular, these models are ideally suited for real-time applications due to their computational efficiency, which makes them a promising avenue for rapid and accessible diagnoses.
This research investigated the broader landscape of Machine Learning (ML) and deep learning (DL) in medical applications, particularly in Computer-Aided Diagnosis (CAD) systems. In particular, the proposed YOLO model-based system aligns with the paradigm shift toward deep learning, and here, we demonstrate its potential for skin cancer recognition. Specifically, this study focused on:
  • ▪ Analyzing the state-of-the-art deep learning techniques employed for skin cancer lesion recognition;
  • ▪ Conducting a comprehensive evaluation of the latest YOLO models, namely, YOLOv3, YOLOv4, YOLOv5, and YOLOv7, in terms of their performance and computational efficiency;
  • ▪ Designing and implementing a YOLO-based approach for accurate skin cancer lesion recognition, focusing on the identification of “Malignant Melanoma”, “Benign Nevus”, and “Seborrheic Keratosis” lesions;
  • ▪ Assessing the performance of the proposed system in comparison to the existing methods, using established metrics such as accuracy, sensitivity, specificity, and computational time.
Accordingly, this article addresses the limitations of the current skin cancer recognition methods through the proposed YOLO-based approach. In fact, no existing study has empirically compared YOLOv3, YOLOv4, YOLOv5, and YOLOv7 in classifying skin cancer cases. Moreover, this article represents the first research to investigate YOLOv5 and YOLOv7 for skin lesion recognition. Furthermore, data augmentation was considered for improving the generalization of the designed YOLO model-based approach.
The rest of this manuscript is organized as follows: Section 2 surveys the relevant literature, while Section 3 depicts the proposed skin cancer recognition approach. The experiments and discussion of the obtained results are reported in Section 4. Finally, Section 5 concludes the article.

2. Literature Review

In this section, we survey the state-of-the-art deep learning approaches used in the localization, classification, and recognition of skin lesions. First, we will focus on the deep learning methods adopted for skin lesion localization [13]. These methods, often employed as preprocessing steps, aim to identify potentially cancerous regions within an image. In addition, the various architectures and strategies used for accurate localization are outlined in order to highlight their strengths and limitations. Secondly, we will delve into deep learning techniques for classifying skin lesions. In particular, diverse classification models and their ability to distinguish between benign and malignant lesions based on extracted features are explored. Moreover, the performance of different network architectures and training strategies are investigated. Finally, we will address the development of deep learning models for the simultaneous localization and recognition of skin lesions. In fact, these powerful “one-shot” approaches eliminate the need for separate localization and classification steps, potentially saving time and resources. Specifically, we will examine the design principles and performance of such models, offering a perspective on their future in clinical settings.
Deep learning-based approaches rely on object detection models, particularly YOLOv3 [14,15,16,17] for skin lesion localization. Similarly, faster R-CNN was used in [18] to address the skin lesion detection problem. One should mention that most methods target melanoma localization [15,18], while others focus on four common skin cancer types [16,17]. Moreover, the performance measures vary, with some evaluating segmentation post-localization [15,18] and others directly assessing localization accuracy [17]. Notably, these approaches leverage the localization capabilities of deep learning recognition models. Table 1 summarizes the state-of-the-art studies using deep learning-based approaches for localization purposes.
Other studies primarily tackle the binary classification problem (melanoma vs. benign) using CNNs [20], pre-trained models [21], and ensemble methods [22]. Multi-class classification (e.g., melanoma, common nevus, atypical nevus) was also addressed in [23], [24]. Specifically, AlexNet and DenseNet201 were explored to overcome the skin lesion detection challenges. Recently, YOLO-based approaches have been adapted for combined localization and classification tasks [25,26]. Table 2 summarizes the state-of-the-art studies that exploited deep learning-based approaches for classification purposes.
Accordingly, deep learning has gained traction for skin lesion recognition. Namely, YOLO [14,33,34,35] and Faster R-CNN [19] models were exploited to address these object detection and classification tasks. One should note that YOLOv2 [33] outperformed other YOLO variants in binary classification tasks (e.g., melanoma vs. benign) [11,36]. On the other hand, multi-class recognition (e.g., multiple cancer types) has been explored using YOLOv3 [37]. Moreover, Faster R-CNN has also been applied for melanoma and actinic keratosis recognition [38,39]. Table 3 summarizes the YOLO deep learning-based approaches that have been employed to recognize skin lesions.

3. Methodology

An early and accurate diagnosis is crucial for curing skin cancer pathologies. Conventional manual approaches have proven to be costly and time-consuming. This promoted the need for automated intelligent systems. In particular, image processing and deep learning have been introduced as promising alternatives for automated lesion recognition, potentially saving time, costs, and lives.
This research provides a comprehensive comparison between YOLOv3, YOLOv4, YOLOv5, and YOLOv7 models in the context of skin cancer detection. Specifically, the performances of these models in classifying cancers cases, namely melanomas, nevi, and seborrheic keratosis, were investigated.

3.1. YOLOv3-Based Model

The proposed YOLOv3 model leverages a pre-trained 53-layer Darknet53 [14] backbone for feature extraction, followed by an additional 53 layers dedicated to recognition, resulting in a total of 106 convolutional layers. This architecture incorporates a Feature-Pyramid Network (FPN) [42] as its neck, which utilizes bottom–up and top–down pathways to extract multi-scale feature maps. The final predictions are made by the YOLO layer, located in the head. A key strength of the proposed YOLOv3 is its ability to perform detections at three different scales, addressing the historical limitation of small object detection in previous YOLO versions. Figure 1 depicts the YOLOv3 architecture that was designed and implemented in this research.
In fact, the considered YOLOv3 loss function combines both binary cross-entropy loss and logistic regression for classification and object confidence prediction, respectively. In addition, class scores are predicted using logistic regression. Note that the classes exceeding a pre-defined threshold are only assigned to a bounding box.

3.2. YOLOv4-Based Model

The proposed YOLOv4 architecture was built upon YOLOv3. Specifically, a Cross Stage Partial Network (CSPNet) [43] was coupled with the Darknet model to form a novel CSPDarknet53 backbone for feature extraction. The resulting DenseNet-inspired [44] convolution architecture addresses the gradient vanishing problem, optimizes the backpropagation, and eliminates processing bottlenecks. This yields improved learning and generalization capabilities. Specifically, the proposed YOLOv4 architecture, shown in Figure 2, involves three sequential blocks, the (i) backbone: CSPDarknet53 extracts features from the input image; (ii) neck: Spatial Pyramid Pooling (SPP) [45] and Path Aggregation Network (PANet) [46] modules expand the receptive field and refine features from the backbone; and (iii) head: basic YOLO layers that generate the final recognition results.
In addition, YOLOv4 employs the Bag of Freebies (BoF) [47] and Bag of Specials (BoS) [35] techniques for optimization. BoF enhances the detector accuracy without increasing inference costs. It employs a Complete IoU (CIoU) loss, a drop block regularization, and diverse augmentation techniques to improve the model generalization. On the other hand, BoS involves plugins and post-processing modules, such as Mish activation, DIoU-NMS [48], and modified PANet, to enhance the object detection accuracy and maintain an acceptable inference cost.

3.3. YOLOv5-Based Model

The proposed YOLOv5 architecture distinguishes itself from YOLOv3 and YOLOv4 through key architectural choices. Specifically, YOLOv5 exhibits a Focus structure, which reduces the model size and addresses gradient redundancy. This yields a faster inference and improved accuracy. Unlike YOLOv3 and YOLOv4 which enclose three separate YOLO layers, YOLOv5 includes one single YOLO layer in the head. This simplification reduces model complexity while maintaining effective multi-scale prediction through three distinct feature maps. One should mention that both YOLOv4 and YOLOv5 utilize a Path Aggregation Network (PANet) [46] in the neck block. Figure 3 illustrates the main blocks of the considered YOLOv5 architecture. It can be seen that the input image is conveyed through the CSPDarknet53 backbone for feature extraction, followed by Spatial Pyramid Pooling (SPP) [45] to generate features at different scales. The extracted features are then fed into PANet for refinement and aggregation. Finally, the single YOLO layer in the head produces the final recognition results.

3.4. YOLOv7-Based Model

The proposed YOLOv7 architecture inherits a groundbreaking advancement in real-time object detection [49]. Like its predecessors, YOLOv7 adheres to the traditional backbone–neck–head architecture. To ensure optimal inference speed, YOLOv7 leverages Extended Efficiency Layer Aggregation Networks (E-ELANs) [50] as its final layer aggregation method. An E-ELAN, an enhanced version of the ELAN computational block, was meticulously designed considering the memory requirements, input/output channel ratios, and gradient propagation distances.
In fact, the considered YOLOv7 introduces a novel multi-head structure, incorporating a lead head for final predictions and an auxiliary head that assists in training in intermediate layers. Additionally, it employs a compound scaling strategy that scales the network depth and width while concatenating layers to balance speed and accuracy. In contrast to a scaled YOLOv4 architecture, YOLOv7 backbones are trained exclusively on the MS COCO dataset, opting against ImageNet pre-trained backbones. A noteworthy innovation in YOLOv7 is re-parameterization, a new “bag of freebies” technique that enhances model performance without increasing training costs. When it is applied after training, it further improves the inference results. Notably, YOLOv7 surpasses previous object detectors in both speed and accuracy, solidifying its position at the forefront of the field [49].

3.5. Proposed Skin Lesion Recognition Approach

This section outlines the proposed approach to compare the performances of YOLOv3, YOLOv4, YOLOv5, and YOLOv7 in recognizing “Malignant Melanoma” (MM), “Benign Nevus” (BN), and “Seborrheic Keratosis” (SK) skin lesions. Specifically, the four considered models are trained as depicted in Figure 4. Each YOLO model is fed with images representing skin lesions. The labels of these images are also conveyed to the recognition systems.
These labels consist of the type of the lesion, Malignant Melanoma (MM), Seborrheic Keratosis (SK), or Benign Nevus (BN), in addition to their bounding boxes’ details. Namely, the upper left corner coordinates, the width, and the height of the box are provided. In addition to training YOLO models, the network hyperparameters are tuned using the pre-defined validation set. The methodology depicted in Figure 5 was adopted to determine the best-performing YOLO version. More specifically, the performances of the considered YOLO models were assessed using the test set in terms of object classification, object localization, and computation time for recognition.
Then, two data augmentation strategies were adopted. The first one was applied to all training instances without allocating a specific proportion for each class. In other words, the data were increased without taking the data balance into account. In contrast, the second data augmentation uses a certain percentage from each class to accomplish a balance between the classes. The latter was employed to design the proposed system, as illustrated in Figure 6.
The subsequent sections detail the experimental setup, results, and comparisons, providing insights into the proposed model’s performance.

4. Experiments

The conducted experiments relied on the International Skin Imaging Collaboration (ISIC) 2017 dataset [13] from the International Skin Imaging Collaboration (ISIC) archive’s “Skin Lesion Analysis Towards Melanoma Detection” challenge. Comprising 2750 dermoscopy images, the dataset exhibits high intra-class variability in texture, color, and size. The images are labeled as “Malignant Melanoma” (MM), “Seborrheic Keratosis” (SK), or “Benign Nevus” (BN) and are distributed across a training set (2000 images), validation set (150 images), and test set (600 images). Their distribution with respect to each class is reported in Table 4. The images were labeled by dermatology experts and their sized range from 540 × 576 pixels to 4499 × 6748 pixels. The details of the considered dataset are depicted in Table 4.
The dataset images were annotated by expert dermatologists. Moreover, as shown in Figure 7, the ground truth required for the training of YOLO models was provided as the details of the bounding boxes surrounding the lesion of interest, along with the corresponding class labels.

4.1. Performance Measures

Four key metrics were used to evaluate the recognition performance and speed of YOLOv3, YOLOv4, YOLOv5, and YOLOv7 in detecting skin lesions:
  • Intersection over Union (IoU): Quantifies the spatial overlap between predicted and ground truth bounding boxes (0–1 scale; 1 = perfect overlap), which is shown in Equation (1).
  • Average Precision (AP): Integrates IoU, precision, and recall across various confidence thresholds, summarizing the localization and classification accuracy for each class, which is shown in Equation (2).
  • Mean Average Precision (mAP): Averages AP across all classes, providing a single overall performance indicator, which is shown in Equation (3).
  • F1-measure: Harmonic mean of precision and recall, offering a balanced view of model performance for each class, which is shown in Equation (4).
These standard metrics strike a balance between accuracy, detail, and conciseness, allowing for an effective comparison of model performances in skin lesion recognition. We considered both spatial accuracy (IoU) and classification accuracy (AP, mAP, F1-measure), while accounting for the impact of confidence thresholds (AP). Additionally, we measured the running time to evaluate the feasibility of real-world application.
IoU = Area   of   Intersection   Area   of   Union
AP = 0 1 p r d r  
mAP   = 1 N i = 1 N AP c
F 1 Measure = 2 Recall Precision Recall + Precision

4.2. Results

To comprehensively evaluate the performance of YOLOv3, YOLOv4, YOLOv5, and YOLOv7 in skin lesion recognition, we conducted four distinct experiments. Each experiment employed a specific model with a defined configuration, ensuring a fair and systematic comparison. Table 5 shows the detailed configuration adopted for each considered model. Specifically, the settings of the Learning rate (Lr), the Momentum (M), and the number of batches (B) are reported.
Figure 8 summarizes the performance of each configuration and for each model on the dataset across the chosen metrics. Additionally, the running time per image is reported to assess the feasibility of real-world application.
To further validate the findings from the initial experiment, we evaluated the performance of each model with its best configuration on the held-out test and validation sets. This assessment provides a more rigorous evaluation of generalizability and robustness using unseen data. The achieved results are detailed in Table 6. The best result for each metric is shown in bold.
To further illustrate the capabilities of the models for skin lesion recognition, we showcase the detection results of YOLOv7 on different lesion types. In particular, Figure 9 shows sample detections of benign nevi (BN), malignant melanomas (MM), and seborrheic keratosis (SK). It can be seen that YOLOv7 accurately located and identified the lesions of interest. Moreover, the generated bounding boxes precisely surround the lesions. This showcases YOLOv7’s strong spatial localization capabilities.

4.3. Discussion

Table 6 demonstrates YOLOv7’s significant edge over YOLOv3 and YOLOv4 in recognizing the three considered lesion types, namely BN, MM, and SK. This finding is consolidated by superior mAP, IoU, and F1-measure values compared to those of both YOLOv3 and YOLOv4. While YOLOv5 exhibited slightly better performance for benign nevus detection and a marginally higher IoU, YOLOv7 surpassed it in terms of inference time, F1-measures, and mAP. Notably, YOLOv7 boasted a significantly faster processing time per image (0.31 s vs. 0.51 s for YOLOv5), making it more suitable for real-time applications. Additionally, YOLOv7 demonstrated consistently higher F1-measures and mAP across all lesion categories, indicating superior overall accuracy and balanced performance.
These advancements can be attributed to YOLOv7’s architectural improvements: a novel model scaling component that optimizes factors like layer count, channel count, feature pyramid stages, and input image resolution for optimal performance, and a multi-head structure that further guides the training process, contributing to the model’s superior accuracy and efficiency.
The experimental results reveal that while YOLOv5 and YOLOv7 showcased comparable detection performances, YOLOv7 reigns supreme in terms of inference time, F1-measures, and mAP. Consequently, YOLOv7 served as the chosen model for the proposed approach due to its combined advantages in accuracy, speed, and robustness.
However, it is important to acknowledge that YOLOv7 exhibited a lower AP for malignant melanoma compared to benign nevi and seborrheic keratosis. This can be attributed to two factors: the visual similarities that malignant melanoma shares with the other categories, potentially leading to misclassification, and the data imbalance within the dataset, where the number of malignant melanoma instances was smaller than the number of instances of the other two categories.
To address this challenge, data augmentation was employed on the training set. This technique artificially expands the malignant melanoma data, providing the model with diverse training examples and potentially mitigating classification errors for this category.
To further enhance YOLOv7’s performance, we employed data augmentation to artificially expand the training set through a variety of transformations including rotation, flipping, cropping, scaling, translation, shearing, and color adjustments. Specifically, we implemented two strategies: first, applying augmentation to the entire training set, generating a pool of 20,000 images and second, recognizing the class imbalance, we balanced the data by keeping all augmented malignant melanoma (MM) images while adding only 20% and 70% more augmented benign nevus (BN) and seborrheic keratosis (SK) images, respectively, resulting in a roughly equal class representation, which is shown in Table 7.
Table 8 shows that data augmentation consistently boosted YOLOv7’s performance, with the exception of processing time, with balanced augmentation yielding the most pronounced impact. Notably, MM’s Average Precision (AP) soared from 64.9% to 68.4% (compared to 66.8% with unbalanced augmentation), showcasing the effectiveness of targeted data augmentation in addressing class imbalances.
Table 9 reports the performance achieved using the YOLOv7 model with those obtained using relevant state-of-the-art approaches. Namely, we compared YOLOv7 with a Convolutional Neural Network (CNN) with 69 layers introduced in [51]. Furthermore, the results obtained using the MobileNet-V2, Xception, and InceptionResNet-V2 models proposed in [52] were also considered in this comparison study.
As can be seen in Table 9, YOLOv7 consistently outperformed the state-of-the-art models according to the evaluation metrics. Notably, YOLOv7 yielded the highest AP, F1-score, IoU, and mAP values. One can also note that InceptionResNet-V2 model outperformed Xception model. Moreover, the CNN-based approach in [51] was overtaken by all considered models.

5. Conclusions and Future Work

The escalating prevalence of skin cancer demands urgent attention, as it poses a global public health threat. Early detection plays a crucial role in successful treatment, highlighting the need for accurate and efficient diagnostic methods. While traditional physician assessments remain the standard, their inherent limitations like subjectivity and time constraints necessitate alternative approaches. This project tackled this challenge by proposing an image-based system powered by deep learning to automatically recognize and differentiate cancerous skin lesions from benign ones. We focused on three prevalent lesions—malignant melanoma, benign nevi, and seborrheic keratosis.
To lay a solid foundation, we explored the background of these lesions, outlining their distinctive visual characteristics. We then delved into supervised learning and deep learning paradigms, specifically examining Convolutional Neural Networks (CNNs), ResNets, Darknets, and YOLO models. Furthermore, we investigated recent deep learning-based approaches for skin lesion recognition, focusing on those involving localization, classification, and recognition. The existing works using YOLOv3 and YOLOv4 demonstrated promising results, but a gap existed in comparing all four YOLO models—YOLOv3, YOLOv4, YOLOv5, and YOLOv7—for this specific task. Moreover, YOLOv5 and YOLOv7 remained unexplored in the domain of skin lesion recognition.
This paper bridged this gap by comprehensively evaluating and comparing the performance of all four YOLO models in automatically recognizing the three targeted lesions from the ISIC 2017 dataset. The results revealed YOLOv7 as the most effective model, further improving its performance by utilizing a balanced augmented training set. This highlights the value of targeted data augmentation in addressing class imbalances.
Looking ahead, exciting future research opportunities lie in exploring semantic segmentation deep learning approaches like DeepLabv3 for more nuanced lesion recognition along with the newer ISIC 2019 [53] and ISIC 2020 [54] datasets. Additionally, investigating the potential of Generative Adversarial Networks (GANs) for data generation presents a promising avenue for expanding and enriching training datasets, potentially unlocking further advancements in skin cancer diagnosis through the power of deep learning.

Author Contributions

Conceptualization, N.A.A., M.M.B.I., S.A.A. and O.B.; methodology, N.A.A., M.M.B.I., S.A.A. and O.B.; validation, N.A.A., M.M.B.I., S.A.A. and O.B.; formal analysis, N.A.A., M.M.B.I., S.A.A. and O.B.; investigation, N.A.A., M.M.B.I., S.A.A. and O.B.; resources, N.A.A., M.M.B.I., S.A.A. and O.B.; data curation, N.A.A., M.M.B.I., S.A.A. and O.B.; writing—original draft preparation, N.A.A.; writing—review and editing, N.A.A. and M.M.B.I.; visualization, N.A.A., M.M.B.I., S.A.A. and O.B.; supervision, N.A.A., M.M.B.I. and O.B.; project administration, N.A.A., M.M.B.I. and O.B.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

A publicly available dataset was analyzed in this research. This data can be found at https://challenge.isic-archive.com/data/#2017 (5 February 2023).

Acknowledgments

The authors would like to acknowledge the Researchers Supporting Project (number RSPD2024R846) from King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Skin Cancer|World Cancer Research Fund International. WCRF International. 2021. Available online: https://www.wcrf.org/dietandcancer/skin-cancer/ (accessed on 9 February 2023).
  2. Melanoma Skin Cancer Statistics. American Cancer Society. 2022. Available online: https://www.cancer.org/cancer/melanoma-skin-cancer/about/key-statistics.html (accessed on 9 February 2023).
  3. Rundle, C.W.; Militello, M.; Barber, C.; Presley, C.L.; Rietcheck, H.R.; Dellavalle, R.P. Epidemiologic Burden of Skin Cancer in the US and Worldwide. Curr. Dermatol. Rep. 2020, 9, 309–322. [Google Scholar] [CrossRef]
  4. Ferrante di Ruffano, L.; Takwoingi, Y.; Dinnes, J.; Chuchu, N.; Bayliss, S.E.; Davenport, C.; Matin, R.N.; Godfrey, K.; O’Sullivan, C.; Gulati, A.; et al. Computer-assisted diagnosis techniques (dermoscopy and spectroscopy-based) for diagnosing skin cancer in adults. Cochrane Database Syst. Rev. 2018, 2018, CD013186. [Google Scholar] [CrossRef] [PubMed]
  5. Fargnoli, M.C.; Kostaki, D.; Piccioni, A.; Micantonio, T.; Peris, K. Dermoscopy in the diagnosis and management of non-melanoma skin cancers. Eur. J. Dermatol. 2012, 22, 456–463. [Google Scholar] [CrossRef] [PubMed]
  6. Sigurdsson, S.; Philipsen, P.; Hansen, L.; Larsen, J.; Gniadecka, M.; Wulf, H. Detection of Skin Cancer by Classification of Raman Spectra. IEEE Trans. Biomed. Eng. 2004, 51, 1784–1793. [Google Scholar] [CrossRef]
  7. Lützow-Holm, C.; Gjersvik, P.; Helsing, P. Melanom, føflekk eller talgvorte? Tidsskr. Nor. Legeforening 2013, 133, 1167–1168. [Google Scholar] [CrossRef]
  8. Berseth, M. ISIC 2017—Skin Lesion Analysis towards Melanoma Detection. arXiv 2017, arXiv:1703.00523. [Google Scholar]
  9. Agrahari, P.; Agrawal, A.; Subhashini, N. Skin Cancer Detection Using Deep Learning. In Futuristic Communication and Network Technologies; Springer: Singapore, 2022; pp. 179–190. [Google Scholar]
  10. Arora, G.; Kumar, A.; Abdurohman, M. Classifiers for the Detection of Skin Cancer. In Smart Computing and Informatics; Springer: Singapore, 2017; pp. 351–360. [Google Scholar]
  11. Hasya, H.; Nuha, H.; Abdurohman, M. Real Time-Based Skin Cancer Detection System Using Convolutional Neural Network and YOLO. In Proceedings of the 4th International Conference of Computer and Informatics Engineering (IC2IE), Depok, Indonesia, 14–15 September 2021. [Google Scholar]
  12. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  13. Zafar, M.; Sharif, M.I.; Sharif, M.I.; Kadry, S.; Bukhari, S.A.C.; Rauf, H.T. Skin Lesion Analysis and Cancer Detection Based on Machine/Deep Learning Techniques: A Comprehensive Survey. Life 2023, 13, 146. [Google Scholar] [CrossRef]
  14. Banerjee, S.; Singh, S.; Das, A.; Bag, R. Diagnoses of Melanoma Lesion Using YOLOv3. In Computational Advancement in Communication, Circuits and Systems; Springer: Singapore, 2022; pp. 291–302. [Google Scholar]
  15. Ünver, H.; Ayan, E. Skin Lesion Segmentation in Dermoscopic Images with Combination of YOLO and GrabCut Algorithm. Diagnostics 2019, 9, 72. [Google Scholar] [CrossRef]
  16. Bagheria, F.; Tarokh, M.; Ziaratbanb, M. Semantic Segmentation of Lesions from Dermoscopic Images using Yolo-DeepLab Networks. Int. J. Eng. 2021, 34, 458–469. [Google Scholar]
  17. Saini, S.; Gupta, D.; Tiwari, A. Detector-Segmentor Network for Skin Lesion Localization and Segmentation. In Computer Vision, Pattern Recognition, Image Processing, and Graphics; Springer: Singapore, 2020; pp. 589–599. [Google Scholar]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  19. Hasan, M.; Barman, S.D.; Islam, S.; Reza, A.W. Skin cancer detection using convolutional neural network. In Proceedings of the 5th International Conference on Computing and Artificial Intelligence, Bali, Indonesia, 19–22 April 2019. [Google Scholar]
  20. Kalouche, S.; Ng, A.; Duchi, J. Vision-Based Classification of Skin Cancer Using Deep Learning. 2016. Available online: https://www.semanticscholar.org/paper/Vision-Based-Classification-of-Skin-Cancer-using-Kalouche/b57ba909756462d812dc20fca157b3972bc1f533 (accessed on 9 February 2023).
  21. Demir, A.; Yilmaz, F.; Köse, O. Early detection of skin cancer using deep learning architectures: Resnet-101 and inception-v3. In Proceedings of the Medical Technologies Congress (TIPTEKNO), Izmir, Turkey, 3–5 October 2019. [Google Scholar]
  22. Hosny, K.; Kassem, M.; Foaud, M. Skin cancer classification using deep learning and transfer learning. In Proceedings of the 9th Cairo International Biomedical Engineering Conference (CIBEC), Cairo, Egypt, 20–22 December 2018; pp. 90–93. [Google Scholar]
  23. Thurnhofer-Hemsi, K.; Domínguez, E. A Convolutional Neural Network Framework for Accurate Skin Cancer Detection. Neural Process. Lett. 2020, 53, 3073–3093. [Google Scholar] [CrossRef]
  24. Nersisson, R.; Iyer, T.; Raj, A.J.; Rajangam, V. A Dermoscopic Skin Lesion Classification Technique Using YOLO-CNN and Traditional Feature Model. Arab. J. Sci. Eng. 2021, 46, 9797–9808. [Google Scholar] [CrossRef]
  25. Banerjee, S.; Singh, S.K.; Chakraborty, A.; Das, A.; Bag, R. Melanoma Diagnosis Using Deep Learning and Fuzzy Logic. Diagnostics 2020, 10, 577. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, S.; Hamian, M. Skin Cancer Detection Based on Extreme Learning Machine and a Developed Version of Thermal Exchange Optimization. Comput. Intell. Neurosci. 2021, 2021, 9528664. [Google Scholar] [CrossRef]
  27. Florkowski, M. Classification of Partial Discharge Images Using Deep Convolutional Neural Networks. Energies 2020, 13, 5496. [Google Scholar] [CrossRef]
  28. Hinton, G. Deep Belief Networks. Scholarpedia. 2022. Available online: http://scholarpedia.org/article/Deep_belief_networks (accessed on 26 February 2023).
  29. Simonyan, R.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  30. Krizhevsky, A.; Sutskever, I.; Hinton, G. Imagenet classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems (NIPS’12), Lake Tahoe, NV, USA, 3–6 December 2012; Volume 1, pp. 1097–1105. [Google Scholar]
  31. Huang, G.; Liu, Z.; Maaten, L. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  32. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  33. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  34. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  35. Nie, Y.; Sommella, P.; O’Nils, M.; Liguori, C.; Lundgren, J. Automatic detection of melanoma with yolo deep convolutional neural networks. In Proceedings of the E-Health and Bioengineering Conference (EHB), Iasi, Romania, 21–23 November 2019; pp. 1–4. [Google Scholar]
  36. Roy, S.; Haque, A.; Neubert, J. Automatic Diagnosis of Melanoma from Dermoscopic Image Using Real-Time Object Detection. In Proceedings of the 52nd Annual Conference on Information Sciences and Systems (CISS), Princeton, NJ, USA, 21–23 March 2018; pp. 1–5. [Google Scholar]
  37. Nawaz, M.; Masood, M.; Javed, A.; Iqbal, J.; Nazir, T.; Mehmood, A.; Ashraf, R. Melanoma localization and classification through faster region-based convolutional neural network and SVM. Multimed. Tools Appl. 2021, 80, 28953–28974. [Google Scholar] [CrossRef]
  38. Hartanto, C.A.; Wibowo, A. Development of Mobile Skin Cancer Detection using Faster R-CNN and MobileNet v2 Model. In Proceedings of the 7th International Conference on Information Technology, Computer, and Electrical Engineering (ICITACEE), Semarang, Indonesia, 24–25 September 2020; pp. 58–63. [Google Scholar]
  39. Veneman, R. Real-Time Skin Cancer Detection Using Neural Networks on an Embedded Device; University of Twente: Enschede, The Netherlands, 2021. [Google Scholar]
  40. Zhang, J.; Huang, Y.; Zhang, X.; Xue, Y.; Bi, X.; Chen, Z. Improved YOLO V3 Network for Basal Cell Carcinomas and Bowen’s Disease Detection. Preprint. 2021. Available online: https://www.researchgate.net/publication/352206363_Improved_YOLO_V3_Network_for_Basal_Cell_Carcinomas_and_Bowen%27s_Disease_Detection (accessed on 9 February 2024).
  41. Lin, T.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–25 July 2017; pp. 2117–2125. [Google Scholar]
  42. Wang, C.-Y.; Liao, H.-Y.M.; Wu, Y.-H.; Chen, P.-Y.; Hsieh, J.-W.; Yeh, I.-H. CSPNet: A New Backbone that can Enhance Learning Capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1571–1580. [Google Scholar]
  43. Iandola, F.; Moskewicz, M.; Karayev, S.; Girshick, R.; Darrell, T.; Keutzer, K. DenseNet: Implementing efficient convnet descriptor pyramids. arXiv 2014, arXiv:1404.1869. [Google Scholar]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
  45. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  46. Zhang, Z.; He, T.; Zhang, H.; Zhang, Z.; Xie, J.; Li, K. Bag of freebies for training object detection neural networks. arXiv 2019, arXiv:1902.04103. [Google Scholar]
  47. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. arXiv 2019, arXiv:1911.08287. [Google Scholar] [CrossRef]
  48. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
  49. Rochet, F.; Elahi, T. Towards Flexible Anonymous Networks. arXiv 2022, arXiv:2203.03764. [Google Scholar]
  50. Codella, N.C.F.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin lesion analysis toward melanoma detection: A challenge at the 2017 International symposium on biomedical imaging (ISBI), hosted by the international skin imaging collaboration (ISIC). In Proceedings of the IEEE 15th International Symposium on Biomedical Imaging (ISBI), Washington, DC, USA, 4–7 April 2018; pp. 168–172. [Google Scholar]
  51. Rastegar, H.; Giveki, D. Designing a new deep convolutional neural network for skin lesion recognition. Multimed. Tools Appl. 2023, 82, 18907–18923. [Google Scholar] [CrossRef]
  52. Popescu, D.; El-khatib, M.; Ichim, L. Skin Lesion Classification Using Collective Intelligence of Multiple Neural Networks. Sensors 2022, 22, 4399. [Google Scholar] [CrossRef]
  53. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 180161. [Google Scholar] [CrossRef]
  54. Rotemberg, V.; Kurtansky, N.; Betz-Stablein, B.; Caffery, L.; Chousakos, E.; Codella, N.; Combalia, M.; Dusza, S.; Guitera, P.; Gutman, D.; et al. A patient-centric dataset of images and metadata for identifying melanomas using clinical context. Sci. Data 2021, 8, 34. [Google Scholar] [CrossRef]
Figure 1. The proposed YOLOv3 architecture.
Figure 1. The proposed YOLOv3 architecture.
Cancers 16 01246 g001
Figure 2. The proposed YOLOv4 architecture.
Figure 2. The proposed YOLOv4 architecture.
Cancers 16 01246 g002
Figure 3. The proposed YOLOv5 architecture.
Figure 3. The proposed YOLOv5 architecture.
Cancers 16 01246 g003
Figure 4. YOLO models’ training framework.
Figure 4. YOLO models’ training framework.
Cancers 16 01246 g004
Figure 5. Methodology to design YOLO-based recognition system for skin lesions.
Figure 5. Methodology to design YOLO-based recognition system for skin lesions.
Cancers 16 01246 g005
Figure 6. Proposed system architecture.
Figure 6. Proposed system architecture.
Cancers 16 01246 g006
Figure 7. Sample (a) dataset image and (b) the corresponding bounding box.
Figure 7. Sample (a) dataset image and (b) the corresponding bounding box.
Cancers 16 01246 g007
Figure 8. Performance results obtained using (a) YOLOv3, (b) YOLOv4, (c) YOLOv5, and (d) YOLOv7.
Figure 8. Performance results obtained using (a) YOLOv3, (b) YOLOv4, (c) YOLOv5, and (d) YOLOv7.
Cancers 16 01246 g008
Figure 9. Sample detections of (ah): benign nevi (BN), (ip): malignant melanoma (MM), and (qx): seborrheic keratosis (SK) lesions using YOLOv7.
Figure 9. Sample detections of (ah): benign nevi (BN), (ip): malignant melanoma (MM), and (qx): seborrheic keratosis (SK) lesions using YOLOv7.
Cancers 16 01246 g009
Table 1. Summary of deep learning-based approaches adopted for skin lesion localization.
Table 1. Summary of deep learning-based approaches adopted for skin lesion localization.
PerformanceSkin CancerModel Ref.
SpecificitySensitivityMelanomaYOLOv3 [14] [15]
97.0597.33
97.597.5
97.0297.97
Intersection Over Union (IOU)AccuracyBenign,
Melanoma,
Seborrheic Keratosis,
Atypical Nevi,
YOLOv3 [14] [16]
9094.4
8696
Mean Box IOUMean Average Precision (mAP)Benign,
Melanoma,
Seborrheic Keratosis,
Atypical Nevi,
YOLOv3 [14] [17]
79.0391.85
JaccardDiceSpecificitySensitivityAccuracyMelanomaFaster R-CNN [19][18]
80.991.597.396.895.9
89.195.298.897.597.9
80.994.798.197.297.1
Table 2. Summary of deep learning-based approaches for classifying skin lesions.
Table 2. Summary of deep learning-based approaches for classifying skin lesions.
PerformanceSkin CancerModelRef.
NPVPPVSpecificitySensitivityAccuracyNormal
Melanoma
Deep Believe Network [27][28]
94.1286.7689.791.1892.65
F1-ScorePrecisionRecallAccuracyBenign
Malignant
CNN [29][20]
83.2583.258489.5
AccuracyBenign
Malignant
VGG-16 [30][21]
78
PrecisionSpecificitySensitivityAccuracyMelanoma,
Nevi,
Atypical nevi
AlexNET [31][23]
97.7398.9398.3398.61
Avg. F-measureAvg. PrecisionAvg. RecallAccuracyArchitectureActinic Keratosis,
Basal Cell Carcinoma,
Benign Keratosis,
Dermatofibroma,
Nevi,
Melanoma,
Vascular
DenseNet2 [32][24]
91.2692.0390.594.52Plain DenseNet2
85.0585.384.891.73Two-level DenseNet2
AUCPrecisionRecallAccuracyBenign
Malignant
YOLOv2 [33][25]
0.95858894
AUCPrecisionSpecificitySensitivityAccuracyMelanoma
Non-Melanoma
YOLOv3 [14][26]
0.9997.599.3797.599
0.9997.4499.3897.4499
0.9994.6498.1394.2297.11
Table 3. Summary of YOLO deep learning-based approaches for recognizing skin lesions.
Table 3. Summary of YOLO deep learning-based approaches for recognizing skin lesions.
PerformanceSkin CancerModelRef.
mAPModelBenign
Malignant
YOLOv1 [34]
YOLOv2 [33]
YOLOv3 [14]
[36]
37YOLOv1
83YOLOv2
77YOLOv3
AUCSpecificitySensitivityAccuracyMelanoma
Non-Melanoma
YOLOv2 [33][11]
9185.986.3586
mAPF1-ScorePrecisionRecallAccuracyBenign
Malignant
YOLOv4 [35][40]
89.3485818994.04
PrecisionRecallAccuracySkin CancerBasal Cell Carcinomas,
Bowen’s Disease
YOLOv3 [14][41]
91.332.891.3BCC
90.930.390.9Bowen’s Disease
Table 4. Details of ISIC 2017 dataset.
Table 4. Details of ISIC 2017 dataset.
Class TypeTraining SetValidation SetTesting Set
Malignant Melanoma (MM)37430117
Seborrheic Keratosis (SK)2544290
Benign Nevus (BN)137278393
Table 5. Models’ hyper-parameter settings.
Table 5. Models’ hyper-parameter settings.
YOLOv3YOLOv4YOLOv5YOLOv7
Configuration\HyperparameterLrMBLrMBLrMBLrMB
Configuration 10.0010.937640.0010.900640.0010.949320.0010.93750
Configuration 20.00010.949640.00010.960640.0050.937160.00010.94950
Configuration 30.0010.950160.0010.949320.010.937320.0010.95032
Configuration 40.10.990320.010.990320.010.900160.010.94916
Configuration 50.010.949320.010.937640.0010.949640.010.99032
Configuration 60.0050.900640.0050.900640.010.950640.0050.90060
Table 6. Performance achieved using test and validation sets.
Table 6. Performance achieved using test and validation sets.
Validation SetTest Set
AP(BN)AP(MM)AP(SK)mAPIoUF1Time to Process (s)AP(BN)AP(MM)AP(SK)mAPIoUF1Time to Process (s)
YOLOv3 81.956.369.769.353.969.20.3679.542.760.560.950.865.00.45
YOLOv482.666.777.975.761.676.80.4781.552.665.466.560.872.00.50
YOLOv583.976.386.782.589.879.90.4981.661.974.972.887.474.20.51
YOLOv784.176.884.381.788.582.10.2480.164.981.375.486.377.90.31
Table 7. Class distribution with/without data augmentation.
Table 7. Class distribution with/without data augmentation.
ClassNo Data AugmentationData Augmentation
BN13722740
MM2542540
SK3742610
Table 8. Effect of data augmentation on YOLOv7 performance results.
Table 8. Effect of data augmentation on YOLOv7 performance results.
Test ResultsClassAPF1-ScoreIoUmAPProcessing Time (s)
No data augmentationBN80.177.986.375.40.44
MM64.9
SK81.3
Data augmentation—imbalancedBN81.478.387.576.50.59
MM66.8
SK81.4
Data augmentation—balancedBN81.979.688.778.00.54
MM68.4
SK83.8
Table 9. Comparison of YOLOv7 performance with relevant state-of-the-art approaches.
Table 9. Comparison of YOLOv7 performance with relevant state-of-the-art approaches.
ApproachClassAPF1-ScoreIoUmAP
CNN from [51]BN74.373.592.772.2
MM61.7
SK80.5
MobileNet-V2 from [52]BN78.574.983.174.7
MM67.4
SK78.1
Xception from [52]BN77.076.685.975.1
MM67.3
SK81.0
InceptionResNet-V2 from [52]BN76.877.986.575.4
MM67.4
SK82.0
Proposed YOLOv7 with balanced data augmentationBN81.979.688.778.0
MM68.4
SK83.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

AlSadhan, N.A.; Alamri, S.A.; Ben Ismail, M.M.; Bchir, O. Skin Cancer Recognition Using Unified Deep Convolutional Neural Networks. Cancers 2024, 16, 1246. https://doi.org/10.3390/cancers16071246

AMA Style

AlSadhan NA, Alamri SA, Ben Ismail MM, Bchir O. Skin Cancer Recognition Using Unified Deep Convolutional Neural Networks. Cancers. 2024; 16(7):1246. https://doi.org/10.3390/cancers16071246

Chicago/Turabian Style

AlSadhan, Nasser A., Shatha Ali Alamri, Mohamed Maher Ben Ismail, and Ouiem Bchir. 2024. "Skin Cancer Recognition Using Unified Deep Convolutional Neural Networks" Cancers 16, no. 7: 1246. https://doi.org/10.3390/cancers16071246

APA Style

AlSadhan, N. A., Alamri, S. A., Ben Ismail, M. M., & Bchir, O. (2024). Skin Cancer Recognition Using Unified Deep Convolutional Neural Networks. Cancers, 16(7), 1246. https://doi.org/10.3390/cancers16071246

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop