Abstract
Melanoma is widely recognized as one of the most lethal forms of skin cancer, with its incidence showing an upward trend in recent years. Nonetheless, the timely detection of this malignancy substantially enhances the likelihood of patients’ long-term survival. Several computer-based methods have recently been proposed, in the pursuit of diagnosing skin lesions at their early stages. Despite achieving some level of success, there still remains a margin of error that the machine learning community considers to be an unresolved research challenge. The primary objective of this study was to maximize the input feature information by combining multiple deep models in the first phase, and then to avoid noisy and redundant information by downsampling the feature set, using a novel evolutionary feature selection technique, in the second phase. By maintaining the integrity of the original feature space, the proposed idea generated highly discriminant feature information. Recent deep models, including Darknet53, DenseNet201, InceptionV3, and InceptionResNetV2, were employed in our study, for the purpose of feature extraction. Additionally, transfer learning was leveraged, to enhance the performance of our approach. In the subsequent phase, the extracted feature information from the chosen pre-existing models was combined, with the aim of preserving maximum information, prior to undergoing the process of feature selection, using a novel entropy-controlled gray wolf optimization (ECGWO) algorithm. The integration of fusion and selection techniques was employed, initially to incorporate the feature vector with a high level of information and, subsequently, to eliminate redundant and irrelevant feature information. The effectiveness of our concept is supported by an assessment conducted on three benchmark dermoscopic datasets: PH2, ISIC-MSK, and ISIC-UDA. In order to validate the proposed methodology, a comprehensive evaluation was conducted, including a rigorous comparison to established techniques in the field.
1. Introduction
Cancer is caused by the uncontrolled multiplication of abnormal cells. Human cells frequently possess the capacity to replicate and divide, and abnormally replicated cells can spread through the lymphatic and vascular systems, wreaking havoc on a healthy body [1]. The five main forms of cancer recognized by Stanford Health Care (SHC) are carcinoma, sarcoma, lymphoma, leukemia, and myeloma. Most cases of the malignant melanoma variety belong to the class of carcinoma [2,3]. One of the most lethal and prevalent cancers in the world is skin cancer [4]. Sunlight has been linked to skin cancer in recent studies, because radiation is the main source of these rays; however, some artificial light also causes DNA damage to skin cells. Skin cancer can also be caused by genetic abnormalities or diseases that run in families [5].
Skin cancer in any of its forms affects an estimated 9500 people every day in the United States alone, as reported by the American Cancer Society (ACS) [6]. In the year 2022, a total of 99,780 incidents of melanoma were identified, with 57,180 cases affecting males and 42,600 cases affecting females [7]. It is anticipated that almost 5080 men and 2570 women will lose their lives to this disease this year. Incidences of malignant melanoma have been steadily climbing over the course of the past few decades, with rates varying according to the ages of the people affected [8]. The percentage of skin cancer caused by sun exposure in different age categories and the number of cases diagnosed on each continent are presented in Figure 1.
Figure 1.
Epidemiological data on skin cancer: (a) WHO projections for skin cancer in 2022; (b) average accumulated sun exposure vs. age groups.
With a cautious prediction of 17,756 new cases in 2022, the Australian Institute of Health and Welfare (AIHW) predicts that both melanoma and non-melanoma cases will rise to the third-most-often diagnosed cancer type in Australia. Diagnosis rates are as follows: 58.5% male and 41.5% female [9,10]. Clinical examinations often involve a practitioner or dermatologist observing a suspect’s skin in a series of phases. The most common method is the ABCDE rule, in which the appearance of the lesion (symmetry, border, color, and diameter) and evolution of the lesion are observed [11]. Assessment of the skin’s appearance is heavily influenced by the observer’s eyesight, which varies from person to person. Such observational screening for skin lesions has significant limitations and cannot provide an accurate diagnosis. Despite the best efforts of dermatologists, a recent study found that only 80% of cases were correctly diagnosed [12].
Machine learning methods have been widely implemented in several domains, including activity recognition [13], experimental systems [14], embedded systems [13], and public health care [2], for nearly two decades. The employment of these cutting-edge methods has simplified the entire process of disease detection and diagnosis [15]. Computer-aided diagnostic (CAD) systems [16] have the potential to replace conventional surgical assessment methods based on auto-generated feature analysis using machine learning approaches [17]. Therefore, scientists are certain that machine learning techniques will eventually replace conventional approaches to evaluating surgical procedures [18,19,20]. Computer vision techniques potentially assist medical practitioners in efficiently diagnosing skin cancer within a reduced time frame. This study primarily focused on the use of feature fusion and selection methods together. The research presented in this study makes two primary contributions: firstly, the introduction of a bio-inspired feature selection strategy aimed at addressing the challenges posed by the "curse of dimensionality" and over-fitting; secondly, to enhance the efficacy of the extracted features, a fusion mechanism is employed that leverages the complementary strengths of four pretrained models.
The subsequent sections of the article are structured as follows: Section 2 of the paper encompasses the literature review, which is subsequently followed by the problem statement and the contributions made in Section 3. Section 4 comprises two distinct subsections, the first of which elucidates the datasets and models employed, while the latter expounds upon the proposed framework. The final section of the paper elucidates the simulation and analysis, providing a comprehensive account of the obtained results. The final section of the article, Section 6, serves as a conclusion and provides an overview of potential future research directions.
2. Literature Review
In this section, we provide a concise literature assessment of work done on skin lesion classification using CNNs. In a few cases, the classification frameworks were applied directly to the provided image samples, while in others, the images were initially pre-processed before being subjected to the main processing steps. We begin with a discussion of non-traditional or deep-model approaches to image classification. Several researchers have approached this issue by considering it as a binary classification problem, where the images are categorized into two primary classes: malignant and benign. Several image samples are presented to the readers as references in Figure 2. However, a small number of researchers even employed seven classes.
Figure 2.
Selected skin lesion samples showing the benign class (left) and the malignant class (right).
The research of [21] employed deep learning models for the automatic categorization of multi-class skin lesions. Their presented algorithm was based on the deep convolutional neural network (DCNN), which contains several stacked layers and variable filter sizes. The authors claimed to have attained 90.1% precision, 93.23% sensitivity, and 91.1% specificity on the ISIC-17 dataset. The proposed algorithm demonstrated superior performance in comparison to various alternative methods—particularly in the context of low-resolution images. Similarly, the authors in [22] proposed a DCNN framework, to categorize skin lesions images into seven different classes that were subsequently consolidated into two overarching classes: healthy and cancerous. One limitation of this study was the occasional inability to directly consolidate classes into a smaller number. An evaluation of different CNN architectures was undertaken in the work by [23], in which the authors utilized different configurations of 12 CNN models, and set seven different classifiers. The DenseNet201 combined with the KNN classifier resulted in the best F1-score, accuracy, recall, and precision values.
The methodology proposed by Bi et al. [24] employed a hyper-connected convolutional neural network (HcCNN), to classify skin lesion images. The proposed approach involved the implementation of a deep hierarchical convolutional neural network (HcCNN) that incorporated a multi-scale attention block. This integration enabled the model to effectively capture and utilize the visual characteristics present in both dermoscopy and clinical skin cancer image datasets. The method proposed in this study demonstrated a slightly reduced level of performance, in terms of accuracy (82.70%), sensitivity (68.18%), specificity (84.62%), and precision (75.98%). Similarly, the work of [25] addressed the classification of high-resolution images and class variation present in real datasets. They proposed a framework known as patch-based attention architecture (pretrained CNNs). The outlined algorithm provided a global context in between low- and high-resolution regions. The mean values of the achieved sensitivity, specificity, and F1-score were 73.3%, 96.3%, and 85.3%, which were quite low, as some of the methods achieved better results compared to the existing techniques. In [26], the authors outlined a method of accelerating the performance of classifying the skin lesions, by using generated adversarial networks (GANs) based on data augmentation technology. On the dataset ISIC-2018, the obtained accuracy, specificity, sensitivity, and average precision were, respectively, 95.25%, 96.61%, 83.21%, and 83.11%. Despite the authors’ claims that their acquired parameters were better than the CNN model, improvements are still needed to their multi-class accuracy, specificity, and sensitivity. The presented algorithm was effective only for skin lesion regions with high resolution and better diversity.
The proposed framework of [27] combined a skin lesion boundary segmentation (conducted using a full-resolution convolutional network) stage and a multiple skin cancer lesions classification stage. Then, a CNN, such as ResNet-50, Inception-v3, DenseNet-201, and Inception-ResNet-v2, was employed. The maximum achieved values of specificity, sensitivity, accuracy, true-negative rate (F1-score), and area under the curve were 80.62%, 75.67%, 75.75%, and 81.57%, respectively, on the ISIC 2017 dataset. In their study, Behara et al. [28] presented a model for categorizing skin lesions, which was founded on the utilization of deep convolutional generative adversarial networks (DCGAN). The methodology employed in this study yielded real-time images that were suitable for training purposes. Furthermore, these generated images were further improved by the application of different image processing techniques. The classification task was executed by the final layer of the discriminator, which predicted the desired class. The claimed performance metrics on the ISIC2017 dataset included accuracy of 99.83%, and precision and recall rates of 99%. While the generated images exhibited a certain degree of resemblance to genuine images, they were found to be deficient in terms of richness and diversity. The authors additionally provided a restricted level of control over the hyperparameters. The proposed method in [29] classified the cancer lesion by using ensembles of CNN models known as multi-resolution EfficientNets with metadata. Lesion classification was conducted using EfficientNets, SENet, and ResNet WSI. The achieved values of the area under the curve were in the range 77.5–96% and those of sensitivity were in the range 0.283–71%, obtained on the ISIC-2019 dataset. In [30], the authors proposed a cascade knowledge diffusion network (CKDNet) that transferred and accumulated the information gathered from various sub-tasks, to increase the efficiency of segmenting and classifying cancer images. They reported better performance without ensemble approaches or external datasets and every time neural networks needed to be trained, which took a lot of time: hence, in some applications, this could be a drawback. On the contrary, the authors in [31] proposed an approach for multi-label ensemble multi-class classification of skin cancer images. The efficiency of this method was only compared to that of the specialist’s advice.
The integration of conventional and contemporary frameworks is a subject of interest among researchers in the field. The work of [32] used a wavelet-based CNN model. The method decomposed the input image into seven different directional sub-bands. The sub-band images were fed to eight pretrained CNNs, as an input, to generate eight probabilistic classifiers. The efficiency of the proposed method was evaluated for seborrheic keratosis and melanoma classification. The authors concluded that the model I-GR0235 outperformed other models, in terms of performance. But the achieved values of accuracy (83%), the receiver operating characteristic curve (91%), and sensitivity (13%) were not convincing. Similarly, the authors in [33] presented a multi-level, multi-class algorithm implemented by available machine learning tools and advanced deep learning methods based on the divide and conquer rule. They achieved specificity, sensitivity, precision, and accuracy of 98.45%, 87.21%, 98.25%, and 92.82%, respectively, for the testing phase.
Researchers are still developing conventional methods for classifying cutaneous lesions. The work proposed in [34] was based on sparse representation for classification of lesion images. The developed algorithm produced discriminating sparse codes representing the features in a high-dimensional feature set. The reported values of sensitivity, accuracy, and specificity were 96.61%, 94.83%, and 93.31%, respectively, on the ISIC 2016 dataset. Similarly, the approach in [35] utilized a network called the self-supervised topology clustering network (STCN), to transform an invariant network, using a self-supervised modularity clustering algorithm based on the principles of topology analysis. The efficacy of the proposed STCN was compromised, due to its inability to effectively filter negative sample images, resulting in a decrease in classification performance. Additionally, the hand-crafted features included in the STCN also suffered from this limitation. There are some other applications in the medical imaging domain that have adopted hybrid techniques [36,37,38,39,40].
The literature review is concisely presented in Table 1. The given indices are PRC (precision), SEN (sensitivity), ACC (accuracy), SPC (specificity), AUC (area under the curve), F1-S (F1-score), Ppv (+ve predictive value), and Hm (Harmonic Mean).
Table 1.
A comparative analysis of performance, techniques, and datasets in the literature, for various techniques and their evaluations.
3. Problem Statement and Contributions
Over the past few years, computer-aided detection (CAD) systems have become increasingly important in the detection and assessment of skin lesions. Nevertheless, the classification process is hindered by various limitations at both the image level—including low-contrast lesion regions, skin flakes, the presence of hair, and air bubbles—and at the feature level, such as redundant or missing feature information. Consequently, achieving accurate classification becomes challenging. The presence of these undesirable characteristics has a direct or indirect impact on the segmentation and classification processes, leading to a decline in the overall performance of the system. Hence, it is imperative to tackle these issues at various stages, to establish a resilient framework for detection and classification. This study primarily examined the impact of feature-level information on the ultimate classification outcome. Following the feature extraction phase, conventional feature selection techniques frequently encounter challenges related to increased computational cost and diminished accuracy. Hence, in order to address the aforementioned issue, hybrid metaheuristic algorithms were introduced, to enhance performance. Two main contributions can be drawn from the findings of this study:
- Introduction of a bio-inspired feature selection strategy called the entropy-controlled gray wolf optimization algorithm, which is designed to resolve the challenges posed by the “curse of dimensionality” and over-fitting. This technique emphasizes identifying the most discriminant features, to mitigate these issues.
- Adoption of a fusion method, to combine the strengths of four pretrained models, so as to improve the efficiency of the extracted features.
Given a database of dermoscopic images, we had to attribute a label to each and every image, classifying them as either benign or malignant. We let an image be a dermoscopic image for a given database . The set of images were . For a given image, the number of channels were fixed to be three, and the number of classes C were provided by the user. Therefore, for each image, the extracted features, , were later subjected to the classifier for the label assignment, , against each image. The cascaded system, which consisted of a series of steps, including feature fusion and selection, was ultimately represented as
where denoted the features extracted after employing the transfer learning, depicted the fused feature set from fully connected layers of different architectures, and was the representation of the selected feature set as the output of a hierarchical structural design.
4. Material and Methods
4.1. Convolutional Neural Networks (CNNs)
CNNs are the most spectacular versions of deep feedforward neural networks used for feature detecting and classifying [16,45]. Each neuron in a CNN is linked to a group of other neurons in the higher layer, using a feedforward technique. Convolution, pooling, and fully linked layers make up the three main sub-blocks of a CNN’s fundamental architecture, as depicted in Figure 3.
Figure 3.
Basic Architecture of a CNN.
- Convolution layer: In the CNN architecture, this is the most basic and crucial element. The primary goal of it is to identify and extract local feature sets from an input image, . Let the image database be divided into training () and testing databases , where {. The training samples are represented as , where n denotes the training image database size. For each given input image, the resulting output image is , where signifies the class number. The convolutional layer consists of a kernel filter that goes through each pixel of the input image as . The local feature set is obtained, based on the following equation:where denotes the output feature map for that particular layer, where are the trainable parameters for the layer, and where is the activation function.
4.1.1. Pretrained CNN Models
In this study, we utilized four State-of-the-Art pretrained models for feature extraction, including DarkNet53, InceptionV3, InceptionResNetV2, and DenseNet201. There are various proposed sets of CNN architectures for computer vision applications. This decision was made based on their performance, number of parameters, and Top-1 accuracy.
- Inception-V3: The two essential components of Inception-V3 are feature extraction and classification. It is trained using the ImageNet database. Using inception units, an Inception-framework V3 can increase a network’s depth and width while also reducing its computing load.
- Inception-ResNet-V2: As with the development of Inception-V3, Inception-ResNet-V2 is likewise trained using the ImageNet database. It combines the ResNet module and inception. The other connections enable bypass in the model, which strengthens the network. The computational prowess of the inception units and the optimization leverage provided by the residual connections are combined in Inception-ResNet-V2.
- DenseNet-201: The ImageNet database is also used to train DenseNet-201. It is built on an advanced connectivity scheme that continuously integrates all of the output properties in a feedforward manner. Furthermore, it strengthens feature propagation, decreases the number of input and functional parameters, and mitigates the problem of vanishing gradient.
4.1.2. Datasets
In this study, we carried out our simulations on the three publicly available benchmark datasets:
- consists of 200 RGB images, divided into 160 benign and 40 melanoma image samples. The database is maintained by the Hospital Pedro Hispano, Matsinhos, through clinical observation using a dermoscope. The real physician’s response is also provided, i.e., normal, melanoma, or typical nevus.
- ISIC-MSK: the other database incorporated here is the International Skin Imaging Collaboration (ISIC). It includes 225 RGB dermoscopic image samples obtained from different well-reputed international cancer institutes, captured by various modalities.
- ISIC-UDA: is another publicly accessible dataset for the characterization and study of skin cancer (total images: 2750; training images: 2200; testing samples: 550). It contains three cancer types: melanoma, keratosis, and benign; but, since keratosis is a fairly common benign skin indication, the images can be divided into two classes: malignant and benign.
For evaluation purposes, dermatologists manually labeled all the datasets. Table 2 displays the distribution of images within the previously mentioned datasets.
Table 2.
Selected skin lesion image datasets and their respective ratio of training to testing.
4.2. Proposed Framework
In this study, a conventional hierarchical approach was employed, encompassing feature extraction, and concluding with the final classification. The proposed framework employed transfer learning, to extract deep features from pretrained models. Subsequently, the extracted features were combined in a predetermined order, and these combinations were then subjected to the proposed feature selection method. The feature vectors obtained at the end of the process were subsequently employed for classification purposes. Figure 4 demonstrates the detailed flow of the proposed framework, from the image acquisition to the final classification.
Figure 4.
Detailed illustration of proposed skin lesion classification framework.
4.2.1. Transfer Learning
Convolutional algorithms operate under the assumption that the feature sets of both the training and testing datasets are nearly identical, allowing for straightforward estimation. Although numerous pretrained models have undergone extensive training on general image datasets, they may not be optimal for specialized applications. Transfer learning (TL) is a viable approach, as it effectively classifies images using a limited number of training instances, even in scenarios where acquiring real-world data poses challenges. The optimal performance of transfer learning is achieved when the input and output source databases exhibit a significant degree of dispersion, in terms of their sizes, thereby ensuring a diverse source domain.
Consider a source domain, , where
; with particular learning assignments, , and target domain having the learning assignment as , where .
Let us consider, for a given dataset, the number of image samples are and the model is trained over a large dataset , where ≪, and their labels are and . The primary objective of transfer learning (TL) is to enhance the learning effectiveness of the target function , by leveraging the information derived from both the source dataset and the target dataset .
4.2.2. Feature Fusion
The availability of highly discriminant information is a crucial factor in enhancing classification accuracy. The presence of redundancy and irrelevant information not only diminishes the accuracy of classification but also imposes a greater computational load. Furthermore, the likelihood of attaining a high level of classification accuracy through the sole utilization of a standard feature extraction approach is quite low. Hence, a methodology for feature fusion has been selected, which not only generates a comprehensive information vector but also leads to an increase in redundancy [46]. In order to address this issue, the utilization of feature fusion in conjunction with feature selection algorithms is employed. In this study, we integrated the extracted set of features obtained from the chosen pretrained models, following the implementation of transfer learning. It was supposed that for a given set of features extracted from the selected model after applying transfer learning, . The dimensions for the extracted features were given as , extracted from the fully connected and average pooling layers of all the selected models. The fusion process involved a sequential concatenation of feature vectors, where each new vector was embedded into the existing one. The resultant feature vectors were generated from the combination of all the extracted feature vectors. We let , , , and . The concatenated form followed the property given: , where and . For the rest of the combinations, the property still held: , , , , and .
4.2.3. Entropy-Controlled Gray Wolf Optimization
We employed entropy-controlled gray wolf optimization (GWO) [47], to achieve the desired result. In this section, we offer a brief but concise background on the method.
GWO is a metaheuristic optimization technique that imitates the hunting strategy and social organizational behavior of gray wolves. Like other metaheuristic algorithms, GWO possesses a distinct array of merits and demerits when compared to alternative optimization techniques. GWO exhibits several potential advantages in comparison to alternative evolutionary strategies, encompassing simplicity, efficient exploration and exploitation capabilities, reduced parameter requirements, and enhanced convergence speed. It may not exhibit superior performance compared to other optimization strategies, in all scenarios, but the outcomes achieved by this approach for the given application are remarkable. This framework counts on three primary steps: skirting the prey, encircling the prey, and finally attacking and hunting the prey. In GWO, the population is categorized into alpha wolf, which is the leader of the gang, beta wolf, the second leader, and delta wolf, which is the third leader. The beta wolf assists the alpha leader in making the decisions, and the delta wolf dominates the pack of wolves . The hunting process is originally guided by three leaders, whereas the wolves only follow the leaders. The first step, i.e., the hunting step of the pack, is given as:
where is the new position of the wolf, is current position, and represents the coefficient vector. The variable depends on the current location of the prey and is defined as
here, is a random vector in the range . Other coefficients can be further explored in the cited article [48]. If we assume that , , and are the three optimum solutions, the new position of the other wolves is modified using the following set of rules:
here, is the leader count—selected to be three. The position vectors are calculated by Equation (6):
The parameters , , and are the positions of , , and at t iteration. Other sets of parameters, including and , are calculated using the reference article [48], such as
where , , and are calculated as in [48]. GWO, in general, is utilized, to solve the continuous optimization problem. It optimizes by considering a set of random solutions; for each solution there is a vector that keeps the parameters’ values of the problem. The first step is to estimate the objective function value for each solution. For the current solution, the entropy-based fitness value is calculated on the basis of the total amount of information in an entire probability distribution. The population vector subjected to the entropy calculation offers a maximum information range. The fitness is therefore calculated using Shannon entropy:
where is the selected vector. Hence, each solution has one variable, to keep its objective value. There are vectors and parameters other than the aforementioned. These vectors and parameters store the objective function and location values of , , and wolves. These values are updated before updating the position of the wolves. The GWO algorithm keeps updating the solutions, using Equations (5)–(7).
As mentioned above, we utilized GWO to solve the continuous optimization problem, but in the case of feature selection, we extended the work of [48] and embedded the concept of the entropy fitness function. A detailed flow of the proposed entropy-controlled gray wolf optimization algorithm is given in Figure 5.
Figure 5.
Detailed flow diagram of proposed entropy-controlled gray wolf optimization.
5. Results and Analysis
The simulations awere carried out on three publicly available datasets, as shown in Table 2. Three families of contemporary classifiers—including support vector machines (SVM), k-nearest neighbors (KNN), and ensemble (ES)—were used for classification. The proposed framework was evaluated utilizing two configurations. In the initial configuration, the classification results were obtained without feature selection. In the second simulation setup, the proposed feature selection step was incorporated, to obtain the classification results. In order to make a fair comparison, we also evaluated the proposed framework alongside other classifiers. The training/testing ratio of 70:30 was selected, and hold-out cross-validation was chosen as the cross-validation technique. Table 3 provides all the necessary base parameters for the chosen classifiers. The selected parameters were selected based on the default values for all the Matlab sessions. In this study, we endeavored to employ a diverse range of classifiers, encompassing SVM, KNN, and ensemble methods. This selection was predicated on the previous empirical evidence of consistently achieving superior outcomes in comparison to alternative sets of classifiers for this specific application.
Table 3.
Selected classifiers and their functional parameters.
In the results section, we will discuss the impact of the feature vectors produced by applying transfer learning to four pretrained models. The flow was designed to take into account the feature combination vectors, their initial sizes, and the reduction percentage obtained after implementing the proposed feature selection algorithm.
The findings presented in Table 4 demonstrate that the greatest reduction percentage was observed when all the extracted feature vectors (FV1–FV2–FV3–FV4) were combined. This suggests a high likelihood of redundant information. Despite the extent of the reduction achieved, the classification accuracy remained satisfactory. The average reduction percentage for the last feature combination, which included all feature vectors, was at its maximum value of 91.33%. By contrast, the average reduction percentages for the remaining feature combinations were 74.33%, 82.33%, 90.66%, and 88.66%, respectively. Additionally, an alternative manifestation of the impact can be observed in Figure 6, which illustrates that the greatest level of reduction was attained on the ISIC-MSK dataset. Based on the obtained results, we strongly believe that our proposed algorithm exhibits superior performance in handling large feature vectors, primarily due to its notable capability of effectively detecting and eliminating redundant information.
Table 4.
The chosen set of feature vectors and their respective dimensions, along with the percentage of reduction achieved.
Figure 6.
Comparison of reduction percentage for each selected dataset.
Based on the findings presented in Table 4, it can be inferred that the final combination exhibited the highest reduction rate. Building upon this observation, we proceeded to generate the testing accuracies, as well as other relevant parameters, such as sensitivity, specificity, false negative rate (FNR), false positive rate (FPR), and F1-score, in Table 5. We took accuracy as the primary measure, and we compared the performance specifically on this measure in addition to other measures. The FNR and the FPR of all the classifiers with greater accuracy and sensitivity were at their lowest, which clearly indicates the superior performance of these classifiers, including Fine KNN, Q-SVM, and ES-KNN. We focused primarily on accuracy in our comparisons, and we used this and other metrics to evaluate performance. Classifiers like Fine KNN, Q-SVM, and ES-KNN that had a high level of accuracy and sensitivity also had low levels of FNR and FPR, respectively, demonstrating their superior performance.
Table 5.
Performance comparison of various classifiers over selected datasets.
To provide a better insight, a fair comparison of the feature fusion approach with and without applying feature selection is also provided in Table 6. Three classifiers were employed, due to their superior accuracy and computational efficiency. The results demonstrate a noticeable enhancement in performance, following the implementation of the feature selection technique. When comparing the classification accuracy obtained using Fine KNN with and without feature selection, it is important to evaluate the impact of feature selection on the accuracy of the classification model. In the case of PH2, the maximum achieved accuracy was 98.89%, while the accuracy without feature selection was 85.22%. A discernible disparity of approximately 13% could be observed. Similar patterns were observed in other datasets. When considering the ISIC-MSK dataset, the accuracy rate was observed to be 99.01%. However, when utilizing Fine KNN with the same dataset, the accuracy rate decreased to 81.23%, although the classification accuracy achieved with other classifiers was 83.73%. Regarding ISIC-UDA, ES-KNN demonstrated an accuracy of 99.09%. Conversely, the maximum accuracy attained for the original fused feature vector was 89.74%. Additionally, in order to ensure a comprehensive evaluation, the classification accuracy of several established algorithms is also presented in Table 7. It is evident that our proposed method surpasses these existing algorithms, by exhibiting enhanced classification accuracy. Based on the statistical data, we hold the firm belief that our proposed feature selection techniques have the potential to yield exceptional results in various other applications.
Table 6.
Overall accuracy comparison of simple fusion approach with the proposed framework.
Table 7.
Performance comparison of existing algorithms.
6. Conclusions
Melanoma is widely acknowledged to be a highly fatal variant of skin cancer, with its occurrence demonstrating an increasing pattern in recent times [54]. Also in recent times, a number of computer-based methodologies have been put forth, with the aim of early detection and diagnosis of skin lesions. Despite having attained a certain degree of accomplishment, there persists a margin of error that is regarded as an unresolved research challenge within the machine learning community. The present study introduces an innovative framework for the categorization of skin lesions. The framework integrates deep features, in order to produce a feature vector that is highly discriminative, while simultaneously preserving the integrity of the original feature space. Our study utilized a selection of contemporary deep models—namely, Darknet53, DenseNet201, InceptionV3, and InceptionResNetV2—to perform feature extraction. Furthermore, the utilization of transfer learning was employed, to augment the efficacy of our methodology and, subsequently, feature selection was employed, to identify the most discriminant information. The approach demonstrated satisfactory performance in the majority of cases. However, it is important to note that the feature selection method may not be effective for feature vectors exhibiting maximum correlation. Furthermore, the inclusion of a pre-processing step has the potential to enhance accuracy even further.
There is room for improvement in a number of areas that could be investigated in further studies. Contrast enhancement, vision transformers (ViT), and feature selection are a few examples. Improved segmentation and classification accuracy could be achieved with the use of contrast enhancement techniques, by providing more refined pictures to the CNN/ViT models. Additionally, a dedicated CNN/ViT model may improve the system’s accuracy. As the feature selection mechanism is crucial in discarding superfluous data, other evolutionary methods and hybrid evolutionary methods could be proposed.
Author Contributions
Conceptualization, T.A. and A.A. (Anas Alsuhaibani); Methodology, T.A. and A.A. (Anas Alsuhaibani); Software, R.J., M.R. and A.A. (Adeel Akram); Validation, T.A. and N.A.A.; Formal analysis, M.R.; Investigation, A.A. (Adeel Akram); Writing—original draft, T.A.; Writing—review & editing, R.J.; Funding acquisition, N.A.A. All authors have read and agreed to the published version of the manuscript.
Funding
The funding of this work was provided by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R410), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
PH2 dataset: https://www.kaggle.com/datasets/synked/ph2-modified (accessed on 4 August 2023). ISIC-MSK and ISIC-UDA: https://challenge.isic-archive.com/data/ (accessed on 4 August 2023).
Acknowledgments
The authors are thankful to Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R410), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Amazon, E. Instances [WWW Document]. 2015. Available online: http://aws.amazon.com/ec2/instance-types (accessed on 6 July 2015).
- BF2SkNet: Best deep learning features fusion-assisted framework for multiclass skin lesion classification. Neural Comput. Appl. 2022, 1–17. [CrossRef]
- Akram, T.; Lodhi, H.M.J.; Naqvi, S.R.; Naeem, S.; Alhaisoni, M.; Ali, M.; Haider, S.A.; Qadri, N.N. A multilevel features selection framework for skin lesion classification. Hum.-Centric Comput. Inf. Sci. 2020, 10, 1–26. [Google Scholar] [CrossRef]
- The global burden of melanoma: Results from the Global Burden of Disease Study 2015. Br. J. Dermatol. 2017, 177, 134–140. [CrossRef] [PubMed]
- Skin Cancer Foundation. The Skin Cancer Foundation. Skin Cancer Information. [WWW Document]. 2018. Available online: https://www.skincancer.org/skin-cancer-information/ (accessed on 17 February 2020).
- Siegel, R.L.; Miller, K.D.; Fuchs, H.E.; Jemal, A. Cancer statistics. CA Cancer J. Clin. 2022, 72, 7–33. [Google Scholar] [CrossRef]
- Cokkinides, V.; Albano, J.; Samuels, A.; Ward, M.; Thum, J. American cancer society: Cancer facts and figures. Atlanta Am. Cancer Soc. 2022. Available online: https://www.cancer.org/research/cancer-facts-statistics/all-cancer-facts-figures/cancer-facts-figures-2022.html (accessed on 4 August 2023).
- American Cancer Society. Key Statistics for Melanoma Skin Cancer. [WWW Document]. 2022. Available online: https://www.cancer.org/cancer/melanoma-skin-cancer/about/key-statistics.html/ (accessed on 17 October 2022).
- Australia Government Cancer Australia. Melanoma of the Skin. [WWW Document]. 2022. Available online: https://www.canceraustralia.gov.au/cancer-types/melanoma/statistics/ (accessed on 18 October 2022).
- Australian Institute of Health and Welfare. Skin Cancer. [WWW Document]. 2022. Available online: https://www.aihw.gov.au/ (accessed on 18 October 2022).
- Rigel, D.S.; Friedman, R.J.; Kopf, A.W.; Polsky, D. ABCDE—An evolving concept in the early detection of melanoma. Arch. Dermatol. 2005, 141, 1032–1034. [Google Scholar] [CrossRef]
- Marks, R. Epidemiology of melanoma. Clin. Exp. Dermatol. 2000, 25, 459–463. [Google Scholar] [CrossRef]
- Naqvi, S.R.; Akram, T.; Haider, S.A.; Kamran, M. Artificial neural networks based dynamic priority arbitration for asynchronous flow control. Neural Comput. Appl. 2018, 29, 627–637. [Google Scholar] [CrossRef]
- Haider, S.A.; Naqvi, S.R.; Akram, T.; Kamran, M.; Qadri, N.N. Modeling electrical properties for various geometries of antidots on a superconducting film. Appl. Nanosci. 2017, 7, 933–945. [Google Scholar] [CrossRef]
- Tschandl, P.; Rosendahl, C.; Akay, B.N.; Argenziano, G.; Blum, A.; Braun, R.P.; Cabo, H.; Gourhant, J.-Y.; Kreusch, J.; Lallas, A.; et al. Expert-level diagnosis of nonpigmented skin cancer by combined convolutional neural networks. JAMA Dermatol. 2019, 155, 58–65. [Google Scholar] [CrossRef]
- Attique Khan, M.; Sharif, M.; Akram, T.; Kadry, S.; Hsu, C.H. A two-stream deep neural network-based intelligent system for complex skin cancer types classification. Int. J. Intell. Syst. 2022, 37, 10621–10649. [Google Scholar] [CrossRef]
- Khan, M.A.; Akram, T.; Sharif, M.; Saba, T.; Javed, K.; Lali, I.U.; Tanik, U.J.; Rehman, A. Construction of saliency map and hybrid set of features for efficient segmentation and classification of skin lesion. Microsc. Res. Tech. 2019, 82, 741–763. [Google Scholar] [CrossRef] [PubMed]
- Afza, F.; Khan, M.A.; Sharif, M.; Rehman, A. Microscopic skin laceration segmentation and classification: A framework of statistical normal distribution and optimal feature selection. Microsc. Res. Tech 2019, 82, 1471–1488. [Google Scholar] [CrossRef]
- Khan, M.A.; Javed, M.Y.; Sharif, M.; Saba, T.; Rehman, A. Multi-model deep neu- ral network based features extraction and optimal selection approach for skin lesion classification. In Proceedings of the 2019 International Conference on Computer and Information Sciences (ICCIS), Aljouf, Saudi Arabia, 3–4 April 2019; pp. 1–7. [Google Scholar]
- Zhang, N.; Cai, Y.X.; Wang, Y.Y.; Tian, Y.T.; Wang, X.L.; Badami, B. Skin cancer diagnosis based on optimized convolutional neural network. Artif. Intell. Med. 2020, 102, 101756. [Google Scholar] [CrossRef] [PubMed]
- Iqbal, I.; Younus, M.; Walayat, K.; Kakar, M.U.; Ma, J. Automated multi-class classification of skin lesions through deep convolutional neural network with dermoscopic images. Comput. Med Imaging Graph. 2021, 88, 101843. [Google Scholar] [CrossRef]
- Harangi, B.; Baran, A.; Hajdu, A. Assisted deep learning framework for multi-class skin lesion classification considering a binary classification support. Biomed. Signal Process. Control 2020, 62, 102041. [Google Scholar] [CrossRef]
- Rodrigues, D.d.A.; Ivo, R.F.; Satapathy, S.C.; Wang, S.; Hemanth, J.; Reboucas Filho, P.P. A new approach for classification skin lesion based on transfer learning, deep learning, and IoT system. Pattern Recognit. Lett. 2020, 136, 8–15. [Google Scholar] [CrossRef]
- Bi, L.; Feng, D.D.; Fulham, M.; Kim, J. Multi-label classification of multi-modality skin lesion via hyper-connected convolutional neural network. Pattern Recognit. 2020, 107, 107502. [Google Scholar] [CrossRef]
- Gessert, N.; Sentker, T.; Madesta, F.; Schmitz, R.; Kniep, H.; Baltruschat, I.; Werner, R.; Schlaefer, A. Skin lesion classification using CNNs with patch-based attention and diagnosis-guided loss weighting. IEEE Trans. Biomed. Eng. 2019, 67, 495–503. [Google Scholar] [CrossRef]
- Qin, Z.; Liu, Z.; Zhu, P.; Xue, Y. A GAN-based image synthesis method for skin lesion classification. Comput. Methods Programs Biomed. 2020, 195, 105568. [Google Scholar] [CrossRef]
- Al-Masni, M.A.; Kim, D.H.; Kim, T.S. Multiple skin lesions diagnostics via integrated deep convolutional networks for segmentation and classification. Comput. Methods Programs Biomed. 2020, 190, 105351. [Google Scholar] [CrossRef] [PubMed]
- Behara, K.; Bhero, E.; Agee, J.T. Skin Lesion Synthesis and Classification Using an Improved DCGAN Classifier. Diagnostics 2023, 13, 2635. [Google Scholar] [CrossRef]
- Gessert, N.; Nielsen, M.; Shaikh, M.; Werner, R.; Schlaefer, A. Skin lesion classification using ensembles of multi-resolution EfficientNets with meta data. MethodsX 2020, 7, 100864. [Google Scholar] [CrossRef]
- Jin, Q.; Cui, H.; Sun, C.; Meng, Z.; Su, R. Cascade knowledge diffusion network for skin lesion diagnosis and segmentation. Appl. Soft Comput. 2021, 99, 106881. [Google Scholar] [CrossRef]
- Chatterjee, S.; Dey, D.; Munshi, S.; Gorai, S. Extraction of features from cross correlation in space and frequency domains for classification of skin lesions. Biomed. Signal Process. Control 2019, 53, 101581. [Google Scholar] [CrossRef]
- Serte, S.; Demirel, H. Gabor wavelet-based deep learning for skin lesion classification. Comput. Biol. Med. 2019, 113, 103423. [Google Scholar] [CrossRef] [PubMed]
- Hameed, N.; Shabut, A.M.; Ghosh, M.K.; Hossain, M.A. Multi-class multi-level classification algorithm for skin lesions classification using machine learning techniques. Expert Syst. Appl. 2020, 141, 112961. [Google Scholar] [CrossRef]
- Moradi, N.; Mahdavi-Amiri, N. Kernel sparse representation based model for skin lesions segmentation and classification. Comput. Methods Programs Biomed. 2019, 182, 105038. [Google Scholar] [CrossRef] [PubMed]
- Wang, D.; Pang, N.; Wang, Y.; Zhao, H. Unlabeled skin lesion classification by self-supervised topology clustering network. Biomed. Signal Process. Control 2021, 66, 102428. [Google Scholar] [CrossRef]
- Kıymaç, E.; Kaya, Y. A novel automated CNN arrhythmia classifier with memory-enhanced artificial hummingbird algorithm. Expert Syst. Appl. 2023, 213, 119162. [Google Scholar] [CrossRef]
- Shokouhifar, A.; Shokouhifar, M.; Sabbaghian, M.; Soltanian-Zadeh, H. Swarm intelligence empowered three-stage ensemble deep learning for arm volume measurement in patients with lymphedema. Biomed. Signal Process. Control 2023, 85, 105027. [Google Scholar] [CrossRef]
- Aryai, P.; Khademzadeh, A.; Jafarali Jassbi, S.; Hosseinzadeh, M.; Hashemzadeh, O.; Shokouhifar, M. Real-time health monitoring in WBANs using hybrid Metaheuristic-Driven Machine Learning Routing Protocol (MDML-RP). AEU-Int. J. Electron. Commun. 2023, 168, 154723. [Google Scholar] [CrossRef]
- Veeraiah, N.; Alotaibi, Y.; Subahi, A.F. MayGAN: Mayfly Optimization with Generative Adversarial Network-Based Deep Learning Method to Classify Leukemia Form Blood Smear Images. Comput. Syst. Sci. Eng. 2023, 46, 2039–2058. [Google Scholar] [CrossRef]
- Ismail, W.N. Snake-Efficient Feature Selection-Based Framework for Precise Early Detection of Chronic Kidney Disease. Diagnostics 2023, 13, 2501. [Google Scholar] [CrossRef]
- Rebouças Filho, P.P.; Peixoto, S.A.; da Nóbrega, R.V.M.; Hemanth, D.J.; Medeiros, A.G.; Sangaiah, A.K.; de Albuquerque, V.H.C. Automatic histologically-closer classification of skin lesions. Comput. Med Imaging Graph. 2018, 68, 40–54. [Google Scholar] [CrossRef]
- Mahbod, A.; Schaefer, G.; Wang, C.; Dorffner, G.; Ecker, R.; Ellinger, I. Transfer learning using a multi-scale and multi-network ensemble for skin lesion classification. Comput. Methods Programs Biomed. 2020, 193, 105475. [Google Scholar] [CrossRef]
- Chatterjee, S.; Dey, D.; Munshi, S. Integration of morphological preprocessing and fractal based feature extraction with recursive feature elimination for skin lesion types classification. Comput. Methods Programs Biomed. 2019, 178, 201–218. [Google Scholar] [CrossRef]
- Liu, L.; Mou, L.; Zhu, X.X.; Mandal, M. Automatic skin lesion classification based on mid-level feature learning. Comput. Med Imaging Graph. 2020, 84, 101765. [Google Scholar] [CrossRef]
- Kamga, G.A.F.; Bitjoka, L.; Akram, T.; Mbom, A.M.; Naqvi, S.R.; Bouroubi, Y. Advancements in satellite image classification: Methodologies, techniques, approaches and applications. Int. J. Remote Sens. 2021, 42, 7662–7722. [Google Scholar] [CrossRef]
- Kamga, F.; Akram, T.; Laurent, B.; Naqvi, S.R.; Alex, M.M.; Muhammad, N. A deep heterogeneous feature fusion approach for automatic land-use classification. Inf. Sci. 2018, 467, 199–218. [Google Scholar]
- Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
- Faris, H.; Aljarah, I.; Al-Betar, M.A.; Mirjalili, S. Grey wolf optimizer: A review of recent variants and applications. Neural Comput. Appl. 2018, 30, 413–435. [Google Scholar] [CrossRef]
- Khan, M.A.; Sharif, M.; Akram, T.; Damaševičius, R.; Maskeliūnas, R. Skin Lesion Segmentation and Multiclass Classification Using Deep Learning Features and Improved Moth Flame Optimization. Diagnostics 2021, 11, 811. [Google Scholar] [CrossRef] [PubMed]
- El-Khatib, H.; Popescu, D.; Ichim, L. Deep learning–based methods for automatic diagnosis of skin lesions. Sensors 2020, 20, 1753. [Google Scholar] [CrossRef] [PubMed]
- Alizadeh, S.M.; Mahloojifar, A. Automatic skin cancer detection in dermoscopy images by combining convolutional neural networks and texture features. Int. J. Imaging Syst. Technol. 2021, 31, 695–707. [Google Scholar] [CrossRef]
- Chatterjee, S.; Dey, D.; Munshi, S.; Gorai, S. Dermatological expert system implementing the ABCD rule of dermoscopy for skin disease identification. Expert Syst. Appl. 2021, 167, 114204. [Google Scholar] [CrossRef]
- Khan, M.A.; Zhang, Y.D.; Sharif, M.; Akram, T. Pixels to classes: Intelligent learning framework for multiclass skin lesion localization and classification. Comput. Electr. Eng. 2021, 90, 106956. [Google Scholar] [CrossRef]
- Oliveira, R.B.; Papa, J.P.; Pereira, A.S.; Tavares, J.M.R. Computational methods for pigmented skin lesion classification in images: Review and future trends. Neural Comput. Appl. 2018, 29, 613–636. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).





