Next Article in Journal
Optimized 18F-FDG PET-CT Method to Improve Accuracy of Diagnosis of Metastatic Cancer
Previous Article in Journal
Ultrasound Assessment of Larynx and Trachea in the Neonatal Period, Examination Standard with Predictive Values—Study Protocol
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Skin Cancer Detection Model Using Transfer Learning and Dynamic-Opposite Hunger Games Search

by
Abdelghani Dahou
1,
Ahmad O. Aseeri
2,*,
Alhassan Mabrouk
3,
Rehab Ali Ibrahim
4,
Mohammed Azmi Al-Betar
5 and
Mohamed Abd Elaziz
4,5,6,7,*
1
Mathematics and Computer Science Department, University of Ahmed DRAIA, Adrar 01000, Algeria
2
Department of Computer Science, College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
3
Mathematics and Computer Science Department, Faculty of Science, Beni-Suef University, Beni-Suef 65214, Egypt
4
Department of Mathematics, Faculty of Science, Zagazig University, Zagazig 44519, Egypt
5
Artificial Intelligence Research Center (AIRC), College of Engineering and Information Technology, Ajman University, Ajman P.O. Box 346, United Arab Emirates
6
Faculty of Computer Science & Engineering, Galala University, Suez 43511, Egypt
7
Department of Electrical and Computer Engineering, Lebanese American University, Byblos 10999, Lebanon
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(9), 1579; https://doi.org/10.3390/diagnostics13091579
Submission received: 14 March 2023 / Revised: 21 April 2023 / Accepted: 25 April 2023 / Published: 28 April 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Recently, pre-trained deep learning (DL) models have been employed to tackle and enhance the performance on many tasks such as skin cancer detection instead of training models from scratch. However, the existing systems are unable to attain substantial levels of accuracy. Therefore, we propose, in this paper, a robust skin cancer detection framework for to improve the accuracy by extracting and learning relevant image representations using a MobileNetV3 architecture. Thereafter, the extracted features are used as input to a modified Hunger Games Search (HGS) based on Particle Swarm Optimization (PSO) and Dynamic-Opposite Learning (DOLHGS). This modification is used as a novel feature selection to alloacte the most relevant feature to maximize the model’s performance. For evaluation of the efficiency of the developed DOLHGS, the ISIC-2016 dataset and the PH2 dataset were employed, including two and three categories, respectively. The proposed model has accuracy 88.19% on the ISIC-2016 dataset and 96.43% on PH2. Based on the experimental results, the proposed approach showed more accurate and efficient performance in skin cancer detection than other well-known and popular algorithms in terms of classification accuracy and optimized features.

1. Introduction

Among the common spreading cancers worldwide, skin cancer can threaten human lives and causes serious danger. Skin cancer can affect the cells of the skin in any region of the body, especially sun-exposed areas. Based on the grown abnormal skin cell, skin cancer can be categorized into three types including the common type which is basal cell carcinoma, squamous cell carcinoma, and melanoma which is the less common type and the more dangerous compared to the other two types [1,2]. A significant challenge to facing skin cancer and preventing its negative impact on the skin is early detection which can be challenging, especially in its early stages. In addition, most people avoid periodic checks due to the lack of medical resources, far clinics, lack of specialists, or expensive diagnoses and treatments which can change the case of skin cancer to a more severe case and help the spread. The necessity of early detection, monitoring, and taking precautions seriously can decrease the dangerous complications rate and lower the physical effects [2,3].
Skin lesions can be examined by specialists or with the help of a diagnostic tool such as Dermoscopy which generates a dermoscopy image. Relying on specialists’ opinions only to examine the skin lesions can not be reliable in most cases where the need for dermoscopy images is highly important. However, dermoscopy images can suffer from various limitations, which make the interpretation difficult, including the existence of a highly trained expert, images can be complex, and the quality of the images can differ based on the device used to capture the image, affecting the appearance of the lesion. In addition, the captured area of the body in the dermoscopy image can affect the quality of the image in terms of the location, skin type, color, lighting, magnification, and skin thickness [4,5]. Thus, an automatic melanoma detection tool or algorithm based on dermoscopy images is valuable in improving skin lesions’ diagnosis and management rather than relying only on clinical expertise [3].
In computer vision, various studies have incorporated different features extracted from dermoscopy images to improve the detection accuracy of different skin cancer types including handcrafted features [3,6] and automatically learned features [7,8]. The commonly used algorithms to automatically learn and extract features uses Convolutional Neural Network (CNN) which can achieve remarkable performance on the detection task [7,8]. In addition, deep learning (DL) networks can require a large amount of data during the training phase. Thus, using transfer learning to fine-tune a pre-trained network on similar tasks can minimize the training complexity, fast convergence, and lower the training time [9]. Besides, pre-trained DL models can be employed as feature extractors without further training if they are already trained on similar tasks or related domains [10]. The learned features by the DL model can hold some noise which affects the final classification accuracy due to the presence of nonrelevant features of the high dimensionality of the learned features. Thus, optimization techniques such as metaheuristic algorithms can offer a great solution in the case of performing feature selection and only selecting the most relevant features to boost detection accuracy [11,12].
The proposed framework incorporates deep learning and optimization algorithms. At first, a deep learning model is proposed to process the inputted skin cancer images and learn to extract relevant and meaningful representations automatically without human intervention. At this stage, a pre-trained version of MobileNetV3 is used and fine-tuned to extract the image embeddings. Thus, rather than relying on raw images, we extract new input image representations that serve as the input of the feature selection algorithm. Second, a novel feature selection algorithm is proposed to filter each image embeddings and select only the most relevant attributes to improve the overall framework skin lesion recognition performance and reduce the representation dimensionality to fit on edge devices. This FS method depends on improving the behavior of a new metaheuristic technique named Hunger Games Search (HGS), called Dynamic-Opposite Learning (DOL). The aim of the developed FS, named DOLHGS, is to find the relevant features from the extracted ones using MobileNetV3. This is achieved through integrating the operators of HGS and DOL. The metaheuristic approaches have generally established their performance as FS techniques among the traditional wrapper or filter FS methods. These traditional FS methods such as exhaustive search, random search, and greedy search suffer from premature convergence, and high computational. So, the MH are used since they are efficient and effective techniques. To validate the developed framework, two real-world datasets which are PH2 and ISIC-2016 were used to evaluate and analyze the performance and report extensive experimental results.
The following summarizes the significant contributions of this study:
  • A pre-trained deep learning is used to learn and extract new representations for skin cancer images.
  • A novel FS algorithm is proposed to reduce the dimensionality of extracted features and improve the overall performance by determining the relevant features.
  • Two real-world datasets are used to validate and compare the proposed method to well-known methods.
  • A more general framework is suggested to integrate the proposed method into the system.
The remainder of the paper is laid out as follows: Section 2 includes a review of prior relevant works. Section 3 presents the background of the convolutional neural network as feature extraction and the Hunger Games Search (HGS). Section 4 contains the details of the suggested framework. The outcomes of the experiments are analyzed and discussed in Section 5. Finally, Section 6 gives our concluding remarks and suggests possible future developments.

2. Related Works

Medical classification is a crucial area of study due to its ability to assist in medical diagnosis. Recently, researchers have used deep learning and FS optimization to enhance classification performance on the Internet of Medical Things. This section discusses the classification of medical images using deep learning and FS optimization algorithms.

2.1. Deep Learning-Based Medical Images

Deep Learning techniques have recently demonstrated excellent performance in natural language processing [13,14,15] and image processing [16,17]. Convolutional Neural Networks (CNN) has become a popular deep learning model, where the use of CNNs in object recognition has recently shown encouraging results, and they have become a key study field in medical image analysis categorization [11,18]. Moreover, machine learning algorithms are applied in the majority of skin cancer detection investigations, however deep learning algorithms are used in a small number of skin lesions categorization investigations [19]. Otherwise, deep learning techniques need big quantities of well-labeled training examples [12,20].
As a result, to address the problem, transfer learning has been implemented. Due to its ability to successfully address the flaws of reinforcement learning and supervised learning, transfer learning is becoming increasingly popular [21]. According to the traditional view, in the test, machine learning is pre-trained and then refined for use with specific data. Transfer learning can be used to fine-tune already pre-trained models on new related tasks which shows to be efficient in improving many DL tasks rather than training the DL model from scratch [22]. It is theoretically possible to create effective objectives relevant way only a limited subset of training samples and by transferring information learned from other domains and activities. A deep learning algorithm on input from another set of medical centers (perhaps from different regions) may also result in a demographic incompatibility between the training and testing due to differences in patient features, as well as differences in imaging methods. However, this could result in very weak results [23]. Learning the algorithm with relevant data from the health center where the algorithm is planned to be utilized (target data) is, therefore, an essential undertaking that faces the issue of dealing with extremely small labeled datasets [23]. Several annotated source domains with considerable inconsistencies profit from a unique feature-based technique [24]. It’s also capable of multi-class categorization and routinely produces positive results.
To classify skin cancer images, researchers recently included a pretrained CNN and transfer learning rather than building a CNN from scratch with random initialization parameters [25]. This pre-training significantly decreased CNN’s training time, which led to an accuracy of 84.8% across five categories. In this situation, transfer learning enables models to be used on various and related tasks since they are learned on a single task or huge dataset. In [26], they used a deep learning-based method to find melanoma early. They applied a proposed VGGNet structure and a transfer learning approach to the skin lesion classification algorithm. The suggested method’s sensitivity on the ISIC Archive dataset was 78.66%. In [27], the performance of the classification algorithm for identifying skin lesions was assessed using an augmented and non-augmented dataset.
Small data sets are one problem that prevents the successful detection of skin cancer. Therefore, for the purpose of effective detection of melanoma skin cancer, Abayomi et al. [19] proposed a new method to augment the data. In order to facilitate the automatic identification of melanoma, Kadry et al. [16] suggested a scheme that applied a CNN-based method. To extract the skin melanoma from the dermoscopy image, they used the VGG-SegNet model. After that, the segmented skin melanoma and the ground truth are compared, and the key performance indices are calculated. In Ref. [17], they presented the DenseNet-based UNET model for efficient melanoma segmentation. To determine representative features, they added the DenseNet model to the UNET encoder unit. However, they noted that deep learning techniques might be advantageous without sufficient data. Recent years have seen a significant increase in the use of CNNs in medical image processing because of their strong feature representation capabilities. Tang et al. [28] proposed a multi-stage model based on an extremely deep residual network for fully automated skin cancer identification in medical images. In contrast to earlier approaches, they developed a classification method in the lightweight melanoma prediction model to improve feature selection, reduce computational, and limit the number of hyperparameters.
Lately, it has become clear that the Internet of Medical Things (IoMT) innovation is perfect for building intelligent systems that can diagnose illnesses precisely, just like experts do [29]. In Ref. [6], vital medical systems have benefited from the development of IoMT technology. It is now available to doctors in a variety of settings, improving their ability to diagnose patients without affecting subjective factors. In contrast, the problem of unbalanced data between uncommon and common illnesses was still there for any approach. This problem consequently led to bad performance. However, in the medical field, the classifier should be confident in its precision with a high percentage when identifying the cancer type. A previous study found that timely melanoma recognition is essential for giving patients the right care. We are therefore working to enhance melanoma medical diagnostics.

2.2. Medical Images Classification Using FS Optimizers

FS optimizers have been used to successfully resolve a wide range of challenging optimization problems in the real world. They can effectively move through the solution space because they can use a list of candidate solutions instead of just one. Meta-heuristic optimization techniques, therefore, outperform other optimization techniques. Task scheduling in the IoMT has benefited from the development of numerous meta-heuristic techniques [30], such as particle swarm optimization (PSO) [31], Multi-Verse Optimizer (MVO) [32], and BAT Algorithm [33]. A few experience local minima and convergence rates, notably when a sizable solution space is involved [34]. This restriction frequently leads to ineffective task scheduling techniques, negatively affecting the system’s performance. Consequently, there is an important need for an optimal global solution to the scheduling algorithm problem. So, this study aims to identify the best solutions that enhance the rate of convergence, as demonstrated in the following sections.

3. Background

This section provides the basic information on Efficient neural networks, Hunger Games Search, and Dynamic-Opposite Learning.

3.1. Efficient Neural Networks

Recently, researchers have proposed various convolutional neural network designs and architectures to enhance efficiency based on time and space. Depthwise convolutions such as NASNet [35], MobileNets [36,37], MnasNet [38], ShuffleNets [39], and EfficientNet [40] can fullfil the aforementioned requirements. Thus, the networks are increasingly used in different applications [41,42,43]. The MobileNetV3 [37] is a computationally efficient and optimized architecture for the classification of images and the deployment on embedded systems or IoT devices. In addition, the traditional convolutional layers is replaced with depthwise and pointwise convolutional layers in the MobileNetV3 offers the power of learning more complex representations, thus, achieving remarkable performance on the image classification task. In this section, we will briefly introduce the recently proposed MobileNetV3 as we will use it in our framework.
Recently, Howard et al. [37] introduced MobileNetV3 which came to enhance the previous versions of MobileNet (V1 and V2) with the ability to learn the optimal kernel size using network architecture search (NAS) technique (NetAdapt algorithm). The MobileNetV3 is implemented using the following components: the depthwise separable convolutional layer and the global average pooling layer. The depthwise convolutional layer comprises the depthwise convolutional kernel, the batch normalization layer and the ReLU activation function. In addition, the MobileNetV3 combines different modules from previous versions of MobileNet in the Inverted Residual Block (MBConv Block). These modules can include the Squeeze-And-Excite block [38] and a modified nonlinearity called hard swish introduced in [44,45].
In MobileNetV3, the depthwise separable convolutional layer can be seen as the core building block of the network as shown in Figure 1. The aim of using the depthwise separable convolutional is to replace the traditional convolution layer with a factorized version, reducing the model size. The block of the depthwise separable convolutional layer is composed of two layers which are: (1) depthwise convolution which applies a single convolutional filter to each input channel, (2) and a 1 × 1 convolution (pointwise convolution) to compute the linear combinations of the input channels and generate new feature maps.
Meanwhile, inspired by the intuition behind the bottleneck blocks where several blocks can contain an input followed by a sequence of bottlenecks [46]. The MobileNet architecture implements an inverted residual connection design in the bottleneck block to improve the model performance and lower memory usage. The inverted residual block shown in Figure 1 has been implemented in MobileNetV2 with a residual structure to build a robust feature descriptor and learn more complex nonlinear relationships with feature expansion. The inverted residual block is built using: (1) 1 × 1 expansion convolution to enhance the model performance with fewer calculations, (2) depthwise separable convolution, (3) and a 1 × 1 projection layer with residual connection.

3.2. Hunger Games Search

Yutao and Huiling [47] proposed the Hunger Games Search (HGS) optimizer which simulates the behaviors of animals and hunger. The mathematical modeling of HGS starts with a set of N solutions (X), then get the values of the fitness value ( F i t i ) of X. Then the phase of modernization is performed by the following equations:
X = X t × 1 + r a n d , r 1 < l W 1 × X b R × W 2 × X b X t , r 1 > l , r 2 < E W 1 × X b + R × W 2 × X b X t , r 1 > l , r 2 > E
where r 1 , r 2 and r a n d are random numbers. R is a random value from [ a , a ] and it is defined as:
R = 2 × r a n d × s s , s = 2 × 1 t T
The E is defined as:
E = sech F i t i F i t b
Also, F i t b stands for the finest value. Whereas W 1 and W 2 are the hunger weights defined as:
W 1 = H i × N S H × r 4 , r 3 < l 1 , r 3 > l
W 2 = 2 1 e ( | H i S H | ) × r 5
where r 3 , r 4 and r 5 are random numbers, and S H stands for the hungers feeling summation that defined as:
S H = i H i
As well as, the variable H i is defined as:
H i = 0 , F i t i = F i t b H i + H n , o t h e r w i s e
where F i t b is the finest value and F i t i is the fitness of X i . The new hunger H n is formulated as:
H n = L H × 1 + r , T H < L H T H , o t h e r w i s e
T H = 2 F i t i F i t b F i t w F i t b × r 6 × U B L B
In addition, the objective function has a worse value given by F i t w , also r 6 [ 0 , 1 ] is a random value which can tell that the hunger has positive or negative actions according to several factors (Algorithm 1).
Algorithm 1 Steps of HGS
1:
Initial the number of iterations T, size of population N.
2:
Build the population X.
3:
while  t T  do
4:
    Compute the fitness value for X i .
5:
    Allocate the value of X b ,   F i t W , and F i t b .
6:
    Enhance H i using Equation (7)
7:
    Update W 1 and W 2 based on Equation (4) and Equation (5), respectively.
8:
    for  do i = 1 : N
9:
         Enhance R using Equation (2)
10:
        Enhance E using Equation (3)
11:
        Enhance X i using Equation (1)
12:
     t = t + 1
13:
Return X b .

3.3. Particle Swarm Optimization

Kennedy and Eberhart [48] created Particle Swarm Optimization (PSO), which replicates the evolution of the understanding of a social activity [49]. First, random particles are created and their positions ( x i ) and speeds ( v i ) in a given dimension j-th are determined. Particle locations [49,50] are updated using Equations (10) and (11).
x i j ( t + 1 ) = x i j ( t ) + v i j ( t + 1 )
v i j ( t + 1 ) = w v i j ( t ) + c 1 r 1 ( x i j p ( t ) x i j ( t ) ) + c 2 r 2 ( x j g ( t ) x i j ( t ) )
where x i j refers to the location of particle i in dimension j, v i j is the i-th velocities in the j-th dimension, t is the current state, and w is a weight vector used to accelerate population merging. The acceleration coefficients c 1 and c 2 are constants. x i j p ( t ) denotes the particle i’s best prior position at dimension j, while x j g ( t ) indicates the optimum global location in dimension j. The random values on r 1 and r 2 are ∈ [ 0 , 1 ] .
This process is repeated until the stopping requirements (for example, a set number of iterations) are met. The PSO are described in Algorithm 2.
Algorithm 2 Algorithm of PSO
1:
Input N the number of solutions and total number of generations.
2:
Set the population to its initial state.
3:
repeat
4:
    for  i = 1 to N do
5:
         Compute the fitness value for X.
6:
         if Fitness value of updated X i < its P B e s t  then
7:
           Using updated X i as current solution.
8:
         Find the best solution with the best fitness ( G B e s t ) overall X.
9:
         Using Equation (11) to update the velocity.
10:
       Equation (10) used to update X i
11:
until The termination requirement has been reached.

3.4. Dynamic-Opposite Learning

This section gives the basic steps of the Dynamic-Opposite Learning (DOL) optimization strategy. First, we will discuss the traditional Opposition-based learning (OBL) strategy [51] that is used to improve different optimization approaches [52,53]. OBL strategy is applied to construct a new opposition solution to the current one. This strategy seeks to allocate the best solution that improves the convergence rate.
For the real number X [ U , L ] , its opposite value X o is computed as:
X o = U + L X
Opposite point [52]: Consider X = [ X 1 , X 2 , …, X D i m ] be a solution in a space of D i m -dimensional, and X j [ U j , L j ]. So, the opposite point X o of X is given as:
X j o = U B j + L j X j , w h e r e j = 1 , , D i m .
Besides, the best of the two points X o and X is selected based on the fitness value and the other one is ignored.
Similar to X o , the Dynamic opposite value X D O of the X is defined as:
X D o = w × r 8 ( r 9 × X o X ) + X , w > 0
where w is the weighting factor. The r 8 and r 9 refer to random numbers.
So, the Dynamic opposite point X j D O of point X = [ X 1 , X 2 , …, X D i m ] is defined as:
X j D o = X j + w × r a n d ( r a n d × X j o X j ) , w > 0
Therefore, DOL optimization starts by generating the initial solutions X = ( X 1 , , X D i m and compute its dynamic opposite solution X D o using Equation (15). Then according to the fitness value, the best of them (i.e., X D o and X) is determined and the other one is removed.

4. Proposed Model

The structure of the developed model is discussed in this section. First of all, the skin cancer images are collected. In case the aim is to train the developed model, there are three steps; the first step is to apply the DL model to extract the feature. Followed by the second step that aims to allocate the relevant features based on the modified HGS, named DOLHGS since it depends on DOL. The third step is to train the classifier. However, if the aim is to predict the case of the collected image, then the trained model is used directly. The details of the developed model are given in the following sections.

4.1. Deep Learning for Feature Extraction

This section describes the discriminative learning model used in the experiments to perform feature extraction. Convolutional neural networks and their variants have flexible architectures known for their success in applications such as image recognition [9,54]. Compared to previous studies, we introduced the usage of swarm optimization on top of features extracted from a pre-trained DL model to improve recognition accuracy, extracting the most relevant features, and reducing the number of features.
In our experiments, a pre-trained MobileNetV3 [37] on the ImageNet dataset is used as the backbone for the feature extraction phase in our framework. Based on resource capacity, there are two variants of MobileNetV3 which are MobileNetV3-Large and MobileNetV3-Small. This study used MobileNetV3-Large and adapted it to skin cancer detection tasks via fine-tuning the pre-trained model on different skin cancer datasets (ISIC-2016 and PH2). MobileNetV3 combines several building blocks, including depth-wise separable convolutions, linear bottleneck, and inverted residual structure. The depth-wise separable convolutions have been introduced in MobileNetV1 [36] to replace the traditional convolution layers, lower the model’s size, and make it easier to execute on mobile devices.
Transfer learning is used as the main mechanism in the feature extraction phase with the following steps:
(1)
Replacing the two last output layers in MobileNetV3 with dense connected blocks including two 1 × 1 convolutions for feature extraction and classification, respectively;
(2)
Fine-tuning the modified MobileNetV3 on the skin cancer dataset;
(3)
Extracting the corresponding feature vector of each image from the convolution layer added to the MobileNetV3 model; where the extracted features for each image are flattened into a vector of size 128.
(4)
Later, the extracted features for each image are fed to the feature selection part in our framework.
The 1 × 1 convolution can be seen as a multilayer perceptron (MLP) that can perform operations such as dimensionality reduction and applying non-linearity after convolutions. The added 1 × 1 convolution block receives input channels of size 960 from the last layer feature extractor in MobileNetV3 and outputs 128 channels (used for feature extraction) with a corresponding kernel of size 1. The last 1 × 1 convolution block acts as a fully-connected layer for image classification.
Table 1 shows the architecture of MobileNetV3 used as a feature extractor backbone for skin cancer detection. The image embedding (features) is collected for each image in the dataset and fed to the feature selection phase. The adapted MobileNetV3 was fine-tuned for 50 epochs before performing feature extraction with a batch of size 16. RMSprop, a form of stochastic gradient descent, trains the network with the learning rate set to 0.0001. The ensure model generalization and overcome the overfitting, data augmentation is performed on the training set using several image transformations including: resizing images into 224 × 224 , random crop, color jitter, random horizontal flip, and random vertical flip.

4.2. Steps of DOLHGS Feature Selection Algorithm

Within this section, we introduces the steps of the improved HGS, as the FS method, based on DOL. The general steps of the proposed FS approach, named DOLHGS, are given in Figure 2. The first step in the developed DOLHGS is to generate the set of N agents X which represents the solutions to the FS problem. This process is performed using the following equation:
X i = r a n d × ( U L ) + L , i = 1 , 2 , , N , j = 1 , 2 , , D i m
In Equation (16), U and L are the boundaries of search domain. D i m refers to the dimension of the given data (i.e., the number of features). Then we obtain the Boolean version of X i and this is achieved using Equation (17).
B X i j = 1 i f X i j > 0.5 0 o t h e r w i s e
The next process is to compute the fitness value of X i as defined in Equation (18).
F i t i = λ × γ i + ( 1 λ ) × | B X i | D i m ,
In Equation (18), ( | B X i | D i m ) is the ratio of selected features. γ i is the classification error (we used K N N at K = 5 ). λ stands for a parameter that applied to balance between ( | B X i | D i m ) and γ i .
Since the initial population has the largest effect on the convergence of the agents towards the optimal solution, the DOL is applied to the initial population X using Equation (15). Then computing the fitness value for each X D O , select the best N agents from X X D O according to the fitness value. Thereafter, the best solution X b is determined which has the smallest fitness value F i t b .
Thereafter, the value of X i is updated through using either the operators of HGS or PSO. This is achieved according to the probability P r i of each X i . In the case of P r i > 0.5 , the operators of HGS are used as defined in Equations (1)–(7); otherwise, operators of PSO are used as in Equations (10) and (11). This updating process is formulated as:
X i j = E q u a t i o n s ( 1 ) ( 7 ) i f P r i > 0.5 E q u a t i o n s ( 10 ) and ( 11 ) o t h e r w i s e
where P r i [ 0 , 1 ] is the random probability value used to make the operators of HGS and PSO competitive during updating the solutions.
Thereafter, the DOL is applied to the current updated population X. However, since the DOL can take more time, it will be applied if the probability P r D O is smaller than 0.5 as formulated in the following formula:
X i j = X i j i f P r D O > 0.5 X i j D o J o t h e r w i s e
X i j D o J = w × r a n d ( r a n d × X i j o X i j ) + X i j , w > 0
In Equation (21), X i j o is given in Equation (15). In addition, the search space [ U , L ] is dynamically updated during the searching process as:
L j = m i n ( X i j )
U j = m a x ( X i j )
The best solutions from X X D o J are selected based on the fitness value.
The next process is to check the terminal criteria and when they are met then the algorithm return X b . Otherwise, the updating stage are conducted again.

4.3. Framework of the Developed Skin Cancer Detection

The general structure of the developed platform illustrated in this section, In general, the proposed platform comprises two systems (i.e., training and testing). The training system can be used to fine-tune the developed framework from Section 4.1 and Section 4.2. In this case, we use the pre-trained feature extraction model and benefit from the lightweight and fast model to make the process faster. In this study, the mobilenetv3 architecture used for feature extraction is well-known for its compatibility and low resource usage on embedded systems. After the feature extraction phase, the proposed DOLHGS as a light and robust feature selection technique is used to minimize the features representation space and only keep the most relevant features from each processed image. Reducing the dimension of the feature will help the classification system, which is a basic k-nearest neighbors (KNN) model, perform faster training and provide a classification decision in a reasonable time window. The second system in the proposed platform uses the best pre-trained version of the proposed training system to make predictions on the fly without the need to train the system again. As a result, the system will be provided the final decision alongside different evaluation measures such as accuracy, F1-score, and more.

5. Experiments and Results

This section discusses the experiments performed and their results in order to propose a highly effective and efficient approach for melanoma detection.

5.1. Description of Datasets

For our experimental evaluation, two datasets of dermatoscopic images are used to conduct skin cancer classification tasks: the International Skin Imaging Association 2016 challenge dataset (ISIC-2016) [55] and Hospital Pedro Hispano dataset (PH2) [56]. The ISIC-2016 dataset contains 1179 images, divided into two categories: Approximately 80% of the dataset is benign and the rest is malignant. This database is public for download at the website https://challenge.isic-archive.com/data (accessed on 13 March 2023). Detailed information about the ISIC-2016 dataset can be found in [55]. The PH2 database was divided into 3 classes consisting of 200 images, as presented in [56]. This database used is available for free download at http://www.fc.up.pt/addi/ph2%20database.html (accessed on 13 March 2023). The dermoscopic images are 8-bit RGB, compressed in JPEG format with 768 × 560-sized color images. Furthermore, some samples of the images for the two datasets are represented in Figure 3. Table 2 describes more information about the datasets.

5.2. Performance Measures

Our proposed classification method is assessed by measuring Precision (P), Recall (R), Accuracy (AC), F1-measure (F1) and Performance Improvement Rate (PIR).
R = T P T P + F N
P = T P T P + F P
F 1 = 2 P R P + R
A C = T P + T N T P + T N + F P + F N
P I R ( % ) = ( A C A C ) A C 100
In Equations (24)–(27), false Positives (FP) are non-melanoma photos that have been mistakenly classified as melanomas. True Negative (TN) refers to the proportion of accurately identified non-melanoma images. False Negative (FN) images are those in which non-melanoma images are mistakenly classified as melanomas. Recall is a metric used to determine how many labels the system finds. Precision is a way to gauge how many labels the system assigned correctly. Precision and recall are necessary for the F1-measure to produce accurate results. Accuracy determines the identification rate of the system. Finally, the performance improvement rate is an indicator that evaluates, in relation to other literature scheduling methods as defined in Equation (28) the percentage of improved performance on each technique proposed. AC′, and AC are the accuracy ratings derived from the associated by the suggested algorithm and the comparison approach, respectively.

5.3. Results and Discussion

In this study, the two datasets are randomly split into training and testing sets. A comparative study of the proposed work with the existing work for the classification of melanoma on the PH2 dataset and the ISIC-2016 dataset are conducted as described in the following sections.

5.3.1. Comparison with FS Methods

This section confirms the performance of the newly proposed DOLHGS algorithm, which has been experimentally tested on the ISIC-2016 and the PH2 challenge dataset. The developed algorithm is compared to the Multi-Verse Optimizer (MVO) [57], Whale Optimization Algorithm (WOA) [58], Particle Swarm Optimization (PSO) [48], Bat Algorithm (BAT) [59], and Firefly Algorithm (FFA) [60].
Different metrics evaluate these optimization algorithms to address challenging numerical optimization problems. Table 3 illustrates the parameters of each method. The number of search agents is 50 and the number of iterations is 1000.
After training with the ISIC-2016 dataset and the PH2 dataset, the outcomes of feature selection algorithms are summarised in Table 4. For both datasets, the table shows the results of DOLHGS with MVO, PSO, WOA, FFA, BAT, and HGS optimization algorithms using four separate measurements.
For the ISIC-2016 dataset, the results demonstrate the ability of DOLHGS to achieve the best value for all metrics. WOA and BAT are also the second and third most effective, respectively, with the best accuracy, recall, and precision, respectively. On all four measures, MVO had the worst performance. In the second dataset, DOLHGS has the highest level of stability, followed by WOA, PSO, and HGS. On the other hand, MVO is the least efficient. Finally, as can be seen from this table, DOLHGS gives the best performance and outperforms the other algorithms since it received the best results in terms of optimization performances for the two datasets. In this table, the best performances are bolded.
Moreover, the Friedman (FD) test as a nonparametric is used to analyze the difference between the DOLHGS and others, in which the mean rank value is computed. In [61], the FD test is applied to check whether there is a significant difference between other methods overall the datasets. The results of FD are given in Figure 4 for each approach. Based on to the results of FD, the DOLHGS has a better mean rank value than the other approaches according to ISIC and PH2 datasets. In addition, the p-value is smaller than 0.05 for the methods.
In the ISIC dataset, we noticed that the DOLHGS and WOA have a mean rank of 7 and 6, respectively. BAT and PSO have almost the same mean level. However, FFA delivers more significant results than HGS, averaging 3.25, with HGS 2. Lastly, MVO is less than others, with a mean rank of 1. From the FD test results for the PH2 dataset, we also noticed that DOLHGS is better than others in its mean rank of 7, then WOA, with a mean level of 6. The other algorithms, FFA and HGS, equal the mean rank (i.e., 3.75). Finally, the lowest mean ranking is BAT and MVO.
For more analysis, Figure 5 and Figure 6 illustrates the ROC curves and the confusion matrices obtained using the proposed DOLHGS approach for skin cancer classification. It can be noticed the efficiency of the developed DOLHGS for the two datasets.

5.3.2. Comparison with Previous Works

A lot of effort is being put into developing high-accuracy technologies for melanoma diagnosis. We compared our method against other existing state-of-the-art algorithms that have been verified on the same datasets to provide a fair comparison. Table 5 and Table 6 compare the performance of various approaches for melanoma detection on the PH2 and ISIC datasets, respectively. For the ISIC dataset, we conducted a thorough comparison with advanced melanoma detection technologies, which include: Based on separation first and then recognition [62], based on feature fusion [63], Fisher coding and deep residual network are coupled [9], multi-CNN collaborative training model [10], using ensemble method [64], combining Fisher Vector and multi-CNN fusion [54]. A fine-grained categorization principle is used to discriminate features [4]. For the other dataset (i.e., PH2), different methods have been developed. For example, the artificial neural networks [65] is used to construct a decision-support system. Kernel sparse model-based strategy to represent features in a high-dimensional feature space was introduced by [66]. In [67], U-Net was offered as a tool for automatically classifying melanomas. In [6], transfer learning and CNN were used in their framework. For learning features, a hierarchical framework based on two-dimensional superpixels and ResNet-50 was presented in [68].
From another point of view, The PIR(%) based on the accuracy of the proposed DOLHGS approach as it relates to other existing state-of-the-art algorithms is presented in Table 5 and Table 6 on the ISIC and PH2 dataset, respectively. For the execution of the ISIC dataset, DOLHGS shows 3.05%, 3.62%, 1.56%, 2.14%, 0.22%, 1.56%, and 0.67% accuracy improvements over the CUMED, BL-CNN, DCNN-FV, MC-CNN, KNORA-E, MFA, and FUSION methods, respectively. Furthermore, the DOLHGS algorithm outperforms the ANN, Kernel Sparse, Dense-Net201 + SVM, DenseNet201 + KNN, and NB techniques for the PH2 dataset (shown in Table 6) by 4.08%, 3.04%, 4.59%, 3.39%, and 1.07%, respectively. That is, DOLHGS outperforms the other approaches significantly.
In summary, our method is capable of removing extraneous features from high-dimensional images generated by CNN models. However, the major limitation of this framework, both the time and the memory are high levels of complexity. Our future work will focus on lowering complexity and enhancing the efficiency of the suggested framework, among other things. In order to increase our method’s performance, more augmentation approaches can be investigated in the future study. Finally, emphasize the potential role of IoMT in reducing healthcare costs associated with skin cancer management. By enabling early detection and personalized treatment, IoMT can help to reduce the need for costly surgeries and other interventions, and can help to optimize the use of healthcare resources.

6. Conclusions

This paper has presented an alternative skin cancer detection method. The developed method depends on MobileNetV3 architecture for feature extraction tasks using fine-tuning techniques on skin cancer datasets to learn more complex representations. In addition, allocate the relevant features from the extracted image representation according to a novel FS technique based on metaheuristic algorithms named DOLHGS. The improvement of the FS algorithm is performed using the operators of particle swarm optimization as a local search strategy and using dynamic opposite-based learning to improve the diversity of solutions. This leads to enhancing the convergence towards the optimal subset of relevant features. To evaluate the efficiency of the developed method a set of experiments have been performed using two real-world datasets named ISIC-2016 and PH2. The results show that the developed DOLHGS FS method is better than traditional FS methods. In addition, the comparison results with other state-of-the-art skin cancer detection methods illustrated that the developed IoMT approach is an effective method. Furthermore, the framework can be further optimized using other techniques to improve its performance and reduce its complexity such as employing neural architecture search and hyper-parameter optimization. Moreover, the developed technique can be applied to enhance the decision of diagnosis the skin cancer and this will help the expert to early detect the disease and treatments.
In future work, we plan to evaluate our approach on a larger number of datasets and to promote its use in clinical practice. Furthermore, combining numerous classifiers is an exciting field of study that can help researchers improve the effectiveness of their systems.

Author Contributions

Authors have the same contributions. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia grant number IF-PSAU-2022/01/19574.

Data Availability Statement

The data are available from the authors upon request.

Acknowledgments

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this research work through the project number (IF-PSAU-2022/01/19574).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Le, P.T.; Chang, C.C.; Li, Y.H.; Hsu, Y.C.; Wang, J.C. Antialiasing Attention Spatial Convolution Model for Skin Lesion Segmentation with Applications in the Medical IoT. Wirel. Commun. Mob. Comput. 2022, 2022, 1278515. [Google Scholar] [CrossRef]
  2. Thomas, S.M.; Lefevre, J.G.; Baxter, G.; Hamilton, N.A. Interpretable deep learning systems for multi-class segmentation and classification of non-melanoma skin cancer. Med. Image Anal. 2021, 68, 101915. [Google Scholar] [CrossRef] [PubMed]
  3. Kassani, S.H.; Kassani, P.H. A comparative study of deep learning architectures on melanoma detection. Tissue Cell 2019, 58, 76–83. [Google Scholar] [CrossRef] [PubMed]
  4. Wei, L.; Ding, K.; Hu, H. Automatic skin cancer detection in dermoscopy images based on ensemble lightweight deep learning network. IEEE Access 2020, 8, 99633–99647. [Google Scholar] [CrossRef]
  5. Hammad, M.; Iliyasu, A.M.; Subasi, A.; Ho, E.S.; Abd El-Latif, A.A. A multitier deep learning model for arrhythmia detection. IEEE Trans. Instrum. Meas. 2020, 70, 1–9. [Google Scholar] [CrossRef]
  6. Rodrigues, D.D.A.; Ivo, R.F.; Satapathy, S.C.; Wang, S.; Hemanth, J.; Reboucas Filho, P.P. A new approach for classification skin lesion based on transfer learning, deep learning, and IoT system. Pattern Recognit. Lett. 2020, 136, 8–15. [Google Scholar] [CrossRef]
  7. Abbas, F.; Yasmin, M.; Fayyaz, M.; Elaziz, M.A.; Lu, S.; El-Latif, A.A.A. Gender Classification Using Proposed CNN-Based Model and Ant Colony Optimization. Mathematics 2021, 9, 2499. [Google Scholar] [CrossRef]
  8. Jing, H.; He, X.; Han, Q.; Abd El-Latif, A.A.; Niu, X. Saliency detection based on integrated features. Neurocomputing 2014, 129, 114–121. [Google Scholar] [CrossRef]
  9. Yu, Z.; Jiang, X.; Zhou, F.; Qin, J.; Ni, D.; Chen, S.; Lei, B.; Wang, T. Melanoma recognition in dermoscopy images via aggregated deep convolutional features. IEEE Trans. Biomed. Eng. 2018, 66, 1006–1016. [Google Scholar] [CrossRef]
  10. Zhang, J.; Xie, Y.; Wu, Q.; Xia, Y. Medical image classification using synergic deep learning. Med. Image Anal. 2019, 54, 10–19. [Google Scholar] [CrossRef]
  11. Mabrouk, A.; Dahou, A.; Elaziz, M.A.; Díaz Redondo, R.P.; Kayed, M. Medical Image Classification Using Transfer Learning and Chaos Game Optimization on the Internet of Medical Things. Comput. Intell. Neurosci. 2022, 2022, 9112634. [Google Scholar] [CrossRef] [PubMed]
  12. Elaziz, M.A.; Dahou, A.; El-Sappagh, S.; Mabrouk, A.; Gaber, M.M. AHA-AO: Artificial Hummingbird Algorithm with Aquila Optimization for Efficient Feature Selection in Medical Image Classification. Appl. Sci. 2022, 12, 9710. [Google Scholar] [CrossRef]
  13. Adel, H.; Dahou, A.; Mabrouk, A.; Abd Elaziz, M.; Kayed, M.; El-Henawy, I.M.; Alshathri, S.; Amin Ali, A. Improving Crisis Events Detection Using DistilBERT with Hunger Games Search Algorithm. Mathematics 2022, 10, 447. [Google Scholar] [CrossRef]
  14. Mabrouk, A.; Redondo, R.P.D.; Kayed, M. Deep learning-based sentiment classification: A comparative survey. IEEE Access 2020, 8, 85616–85638. [Google Scholar] [CrossRef]
  15. Mabrouk, A.; Redondo, R.P.D.; Kayed, M. SEOpinion: Summarization and Exploration of Opinion from E-Commerce Websites. Sensors 2021, 21, 636. [Google Scholar] [CrossRef]
  16. Kadry, S.; Taniar, D.; Damaševičius, R.; Rajinikanth, V.; Lawal, I.A. Extraction of abnormal skin lesion from dermoscopy image using VGG-SegNet. In Proceedings of the 2021 Seventh International Conference on Bio Signals, Images, and Instrumentation (ICBSII), Chennai, India, 25–27 March 2021; pp. 1–5. [Google Scholar]
  17. Nawaz, M.; Nazir, T.; Masood, M.; Ali, F.; Khan, M.A.; Tariq, U.; Sahar, N.; Damaševičius, R. Melanoma segmentation: A framework of improved DenseNet77 and UNET convolutional neural network. Int. J. Imaging Syst. Technol. 2022, 32, 2137–2153. [Google Scholar] [CrossRef]
  18. Abd Elaziz, M.; Mabrouk, A.; Dahou, A.; Chelloug, S.A. Medical Image Classification Utilizing Ensemble Learning and Levy Flight-Based Honey Badger Algorithm on 6G-Enabled Internet of Things. Comput. Intell. Neurosci. 2022, 2022, 5830766. [Google Scholar] [CrossRef]
  19. Abayomi-Alli, O.O.; Damasevicius, R.; Misra, S.; Maskeliunas, R.; Abayomi-Alli, A. Malignant skin melanoma detection using image augmentation by oversamplingin nonlinear lower-dimensional embedding manifold. Turk. J. Electr. Eng. Comput. Sci. 2021, 29, 2600–2614. [Google Scholar] [CrossRef]
  20. Mabrouk, A.; Díaz Redondo, R.P.; Dahou, A.; Abd Elaziz, M.; Kayed, M. Pneumonia Detection on Chest X-ray Images Using Ensemble of Deep Convolutional Neural Networks. Appl. Sci. 2022, 12, 6448. [Google Scholar] [CrossRef]
  21. Niu, S.; Liu, Y.; Wang, J.; Song, H. A decade survey of transfer learning (2010–2020). IEEE Trans. Artif. Intell. 2020, 1, 151–166. [Google Scholar] [CrossRef]
  22. Niu, S.; Wang, J.; Liu, Y.; Song, H. Transfer learning based data-efficient machine learning enabled classification. In Proceedings of the 2020 IEEE International Conference on Dependable, Autonomic and Secure Computing, International Conference on Pervasive Intelligence and Computing, International Conference on Cloud and Big Data Computing, International Conference on Cyber Science and Technology Congress (DASC/PiCom/CBDCom/CyberSciTech), Calgary, AB, Canada, 17–22 August 2020; pp. 620–626. [Google Scholar]
  23. Niu, S.; Liu, M.; Liu, Y.; Wang, J.; Song, H. Distant domain transfer learning for medical imaging. IEEE J. Biomed. Health Inform. 2021, 25, 3784–3793. [Google Scholar] [CrossRef]
  24. Niu, S.; Hu, Y.; Wang, J.; Liu, Y.; Song, H. Feature-based distant domain transfer learning. In Proceedings of the 2020 IEEE International Conference on Big Data (Big Data), Atlanta, GA, USA, 10–13 December 2020; pp. 5164–5171. [Google Scholar]
  25. Rashid, J.; Ishfaq, M.; Ali, G.; Saeed, M.R.; Hussain, M.; Alkhalifah, T.; Alturise, F.; Samand, N. Skin Cancer Disease Detection using Transfer Learning Technique. Appl. Sci. 2022, 12, 5714. [Google Scholar] [CrossRef]
  26. Lopez, A.R.; Giro-i Nieto, X.; Burdick, J.; Marques, O. Skin lesion classification from dermoscopic images using deep learning techniques. In Proceedings of the 2017 13th IASTED International Conference on Biomedical Engineering (BioMed), Innsbruck, Austria, 20–21 February 2017; pp. 49–54. [Google Scholar]
  27. Ayan, E.; Ünver, H.M. Data augmentation importance for classification of skin lesions via deep learning. In Proceedings of the 2018 Electric Electronics, Computer Science, Biomedical Engineerings’ Meeting (EBBT), Istanbul, Turkey, 18–19 April 2018; pp. 1–4. [Google Scholar]
  28. Tang, P.; Yan, X.; Nan, Y.; Xiang, S.; Krammer, S.; Lasser, T. FusionM4Net: A multi-stage multi-modal learning algorithm for multi-label skin lesion classification. Med. Image Anal. 2022, 76, 102307. [Google Scholar] [CrossRef] [PubMed]
  29. Manickam, P.; Mariappan, S.A.; Murugesan, S.M.; Hansda, S.; Kaushik, A.; Shinde, R.; Thipperudraswamy, S. Artificial Intelligence (AI) and Internet of Medical Things (IoMT) Assisted Biomedical Systems for Intelligent Healthcare. Biosensors 2022, 12, 562. [Google Scholar] [CrossRef] [PubMed]
  30. Tsai, C.W.; Chiang, M.C.; Ksentini, A.; Chen, M. Metaheuristic algorithms for healthcare: Open issues and challenges. Comput. Electr. Eng. 2016, 53, 421–434. [Google Scholar] [CrossRef]
  31. Kang, L.; Chen, R.S.; Cao, W.; Chen, Y.C.; Hu, Y.X. Mechanism analysis of non-inertial particle swarm optimization for Internet of Things in edge computing. Eng. Appl. Artif. Intell. 2020, 94, 103803. [Google Scholar] [CrossRef]
  32. Stephen, V.K.; Sharma, S.; Manalang, A.R.; Al-Harthy, F.R.A. A Multi-hop Energy-Efficient Cluster-Based Routing Using Multi-verse Optimizer in IoT. In Computer Networks and Inventive Communication Technologies; Springer: Berlin/Heidelberg, Germany, 2021; pp. 1–14. [Google Scholar]
  33. Alharbi, A.; Alosaimi, W.; Alyami, H.; Rauf, H.T.; Damaševičius, R. Botnet Attack Detection Using Local Global Best Bat Algorithm for Industrial Internet of Things. Electronics 2021, 10, 1341. [Google Scholar] [CrossRef]
  34. El-Shafeiy, E.; Sallam, K.M.; Chakrabortty, R.K.; Abohany, A.A. A clustering based Swarm Intelligence optimization technique for the Internet of Medical Things. Expert Syst. Appl. 2021, 173, 114648. [Google Scholar] [CrossRef]
  35. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning transferable architectures for scalable image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8697–8710. [Google Scholar]
  36. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  37. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  38. Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q.V. Mnasnet: Platform-aware neural architecture search for mobile. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2820–2828. [Google Scholar]
  39. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar]
  40. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning. PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  41. Ji, J.; Krishna, R.; Fei-Fei, L.; Niebles, J.C. Action genome: Actions as compositions of spatio-temporal scene graphs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10236–10247. [Google Scholar]
  42. Liu, J.; Inkawhich, N.; Nina, O.; Timofte, R. NTIRE 2021 multi-modal aerial view object classification challenge. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 588–595. [Google Scholar]
  43. Ignatov, A.; Romero, A.; Kim, H.; Timofte, R. Real-time video super-resolution on smartphones with deep learning, mobile AI 2021 challenge: Report. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 2535–2544. [Google Scholar]
  44. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for activation functions. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  45. Elfwing, S.; Uchibe, E.; Doya, K. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 2018, 107, 3–11. [Google Scholar] [CrossRef] [PubMed]
  46. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  47. Yang, Y.; Chen, H.; Heidari, A.A.; Gandomi, A.H. Hunger games search: Visions, conception, implementation, deep analysis, perspectives, and towards performance shifts. Expert Syst. Appl. 2021, 177, 114864. [Google Scholar] [CrossRef]
  48. Eberhart, R.; Kennedy, J. A new optimizer using particle swarm theory. In Proceedings of the Sixth International Symposium on Micro Machine and Human Science, Nagoya, Japan, 4–6 October 1995; pp. 39–43. [Google Scholar]
  49. Noman, S.; Shamsuddin, S.M.; Hassanien, A.E. Hybrid learning enhancement of RBF network with particle swarm optimization. In Foundations of Computational, Intelligence; Springer: Berlin/Heidelberg, Germany, 2009; Volume 1, pp. 381–397. [Google Scholar]
  50. Niknam, T.; Amiri, B. An efficient hybrid approach based on PSO, ACO and k-means for cluster analysis. Appl. Soft Comput. 2010, 10, 183–197. [Google Scholar] [CrossRef]
  51. Tizhoosh, H.R. Opposition-based learning: A new scheme for machine intelligence. In Proceedings of the International Conference on Computational Intelligence for Modelling, Control and Automation and International Conference on Intelligent Agents, Web Technologies and Internet Commerce (CIMCA-IAWTIC’06), Vienna, Austria, 28–30 November 2005; Volume 1, pp. 695–701. [Google Scholar]
  52. Ewees, A.A.; Abd Elaziz, M.; Houssein, E.H. Improved grasshopper optimization algorithm using opposition-based learning. Expert Syst. Appl. 2018, 112, 156–172. [Google Scholar] [CrossRef]
  53. Ibrahim, R.A.; Ewees, A.A.; Oliva, D.; Abd Elaziz, M.; Lu, S. Improved salp swarm algorithm based on particle swarm optimization for feature selection. J. Ambient. Intell. Humaniz. Comput. 2019, 10, 3155–3169. [Google Scholar] [CrossRef]
  54. Yu, Z.; Jiang, F.; Zhou, F.; He, X.; Ni, D.; Chen, S.; Wang, T.; Lei, B. Convolutional descriptors aggregation via cross-net for skin lesion recognition. Appl. Soft Comput. 2020, 92, 106281. [Google Scholar] [CrossRef]
  55. Gutman, D.; Codella, N.C.; Celebi, E.; Helba, B.; Marchetti, M.; Mishra, N.; Halpern, A. Skin lesion analysis toward melanoma detection: A challenge at the international symposium on biomedical imaging (ISBI) 2016, hosted by the international skin imaging collaboration (ISIC). arXiv 2016, arXiv:1605.01397. [Google Scholar]
  56. Mendonça, T.; Ferreira, P.M.; Marques, J.S.; Marcal, A.R.; Rozeira, J. PH 2-A dermoscopic image database for research and benchmarking. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 5437–5440. [Google Scholar]
  57. Mirjalili, S.; Mirjalili, S.M.; Hatamlou, A. Multi-verse optimizer: A nature-inspired algorithm for global optimization. Neural Comput. Appl. 2016, 27, 495–513. [Google Scholar] [CrossRef]
  58. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  59. Yang, X.S. A new metaheuristic bat-inspired algorithm. In Nature Inspired Cooperative Strategies for Optimization (NICSO 2010); Springer: Berlin/Heidelberg, Germany, 2010; pp. 65–74. [Google Scholar]
  60. Yang, X.S. Firefly algorithm, Levy flights and global optimization. In Research and Development in Intelligent Systems XXVI; Springer: Berlin/Heidelberg, Germany, 2010; pp. 209–218. [Google Scholar]
  61. Derrac, J.; García, S.; Molina, D.; Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol. Comput. 2011, 1, 3–18. [Google Scholar] [CrossRef]
  62. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.A. Automated melanoma recognition in dermoscopy images via very deep residual networks. IEEE Trans. Med. Imaging 2016, 36, 994–1004. [Google Scholar] [CrossRef] [PubMed]
  63. Ge, Z.; Demyanov, S.; Bozorgtabar, B.; Abedini, M.; Chakravorty, R.; Bowling, A.; Garnavi, R. Exploiting local and generic features for accurate skin lesions classification using clinical and dermoscopy imaging. In Proceedings of the 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), Melbourne, Australia, 18–21 April 2017; pp. 986–990. [Google Scholar]
  64. Pathan, S.; Prabhu, K.G.; Siddalingaswamy, P. Automated detection of melanocytes related pigmented skin lesions: A clinical framework. Biomed. Signal Process. Control. 2019, 51, 59–72. [Google Scholar] [CrossRef]
  65. Ozkan, I.A.; Koklu, M. Skin lesion classification using machine learning algorithms. Int. J. Intell. Syst. Appl. Eng. 2017, 5, 285–289. [Google Scholar] [CrossRef]
  66. Moradi, N.; Mahdavi-Amiri, N. Kernel sparse representation based model for skin lesions segmentation and classification. Comput. Methods Programs Biomed. 2019, 182, 105038. [Google Scholar] [CrossRef]
  67. Al Nazi, Z.; Abir, T.A. Automatic skin lesion segmentation and melanoma detection: Transfer learning approach with u-net and dcnn-svm. In Proceedings of the International Joint Conference on Computational Intelligence, Dhaka, Bangladesh, 20–21 November 2020; pp. 371–381. [Google Scholar]
  68. Afza, F.; Sharif, M.; Mittal, M.; Khan, M.A.; Hemanth, D.J. A hierarchical three-step superpixels and deep learning framework for skin lesion classification. Methods 2021, 202, 88–102. [Google Scholar] [CrossRef]
Figure 1. The structure of MobileNetV3 block.
Figure 1. The structure of MobileNetV3 block.
Diagnostics 13 01579 g001
Figure 2. Diagram showing the suggested melanoma detection methodology.
Figure 2. Diagram showing the suggested melanoma detection methodology.
Diagnostics 13 01579 g002
Figure 3. Example images for melanoma detection task from two datasets. The bottom row shows the ISIC-2016 challenge dataset and the top row displays the PH2 dataset.
Figure 3. Example images for melanoma detection task from two datasets. The bottom row shows the ISIC-2016 challenge dataset and the top row displays the PH2 dataset.
Diagnostics 13 01579 g003
Figure 4. The mean rank for each algorithm.
Figure 4. The mean rank for each algorithm.
Diagnostics 13 01579 g004
Figure 5. Confusion matrices obtained of DOLHGS for skin cancer classification: (a) ISIC-2016 and (b) PH2.
Figure 5. Confusion matrices obtained of DOLHGS for skin cancer classification: (a) ISIC-2016 and (b) PH2.
Diagnostics 13 01579 g005
Figure 6. ROC curves obtained of the DOLHGS for skin cancer classification: (a) ISIC-2016 and (b) PH2.
Figure 6. ROC curves obtained of the DOLHGS for skin cancer classification: (a) ISIC-2016 and (b) PH2.
Diagnostics 13 01579 g006
Table 1. MobileNetV3 architecture as feature extraction backbone.
Table 1. MobileNetV3 architecture as feature extraction backbone.
InputOperatorOutputSENLStride
224 × 224 × 3 3 × 3 2d-Conv16FALSEHS2
112 × 112 × 16 3 × 3 16FALSERE1
112 × 112 × 16 3 × 3 24FALSERE2
56 × 56 × 24 3 × 3 24FALSERE1
56 × 56 × 24 5 × 5 40TRUERE2
28 × 28 × 40 5 × 5 40TRUERE1
28 × 28 × 40 3 × 3 40TRUERE1
28 × 28 × 40 3 × 3 80FALSEHS2
14 × 14 × 80 3 × 3 80FALSEHS1
14 × 14 × 80 3 × 3 80FALSEHS1
14 × 14 × 80 3 × 3 80FALSEHS1
14 × 14 × 80 3 × 3 112TRUEHS1
14 × 14 × 112 3 × 3 112TRUEHS1
14 × 14 × 112 5 × 5 160TRUEHS2
7 × 7 × 160 5 × 5 160TRUEHS1
7 × 7 × 160 5 × 5 160TRUEHS1
7 × 7 × 160 1 × 1 2d-Conv960FALSEHS1
7 × 7 × 960 Adaptive average pooling960FALSE-1
1 × 1 × 960 Image embedding128FALSEHS1
3 × 3 2d-Conv: 3 × 3 depthwise separable convolution. NBN: no batch normalization. SE: Squeeze-And-Excite. NL: nonlinearity (HS denotes h-swish and RE denotes ReLU).
Table 2. Dataset Description.
Table 2. Dataset Description.
DatasetSkin Disease# Training Images# Testing ImagesTotal Images per Category
ISIC-2016Malignant17375248
Benign7273041031
Total images9003791279
P h 2 Common Nevus681280
Atypical Nevus681280
Melanoma34640
Total images17030200
Table 3. Values of the parameters for each approach.
Table 3. Values of the parameters for each approach.
AlgorithmValue of the Parameters
DOLHGSEPSILON = 10 × 10 10 , MIN-PROB = 0, MAX-PROB = −1
WOAa = 2 to 0, a2 = −1 to −2
BATQMin = 0, QMax = 2
MVOWEPMax = 1, WEPMin = 0.2
PSOVMax = 6, WMax = 0.9, WMin = 0.2
FFAAlpha = 0.5, BetaMin = 0.2, Gamma = 1
HGSEPSILON = 10 × 10 10 , POS = 0, F IT = 1
Table 4. Average results for all experimental runs of each algorithm.
Table 4. Average results for all experimental runs of each algorithm.
ISICPH2
ACRPF1ACRPF1
PSO0.8656990.8656990.8569190.8522510.9564290.9564290.9569490.956522
MVO0.8633250.8633250.8539150.8498240.9560710.9560710.9565750.956165
WOA0.867810.867810.8605120.8531410.9571430.9571430.9575920.957233
FFA0.8654350.8654350.8570030.851430.9564290.9564290.9569180.956521
BAT0.8670180.8670180.8601020.8519550.9560710.9560710.9565810.956165
HGS0.8649080.8649080.856520.8509730.9564290.9564290.956940.956513
DOLHGS0.881850.875170.876330.875750.964290.974290.976990.97563
Table 5. Accuracy and PIR (%) results of ISIC-2016 dataset.
Table 5. Accuracy and PIR (%) results of ISIC-2016 dataset.
Ref.YearClassification ModelAC (%)PIR (%)
[62]2016CUMED85.503.05
[63]2017BL-CNN85.003.62
[9]2018DCNN-FV86.811.56
[10]2019MC-CNN86.302.14
[64]2019KNORA-E88.000.22
[54]2020MFA86.811.56
[4]2020FUSION87.600.67
OurpresentDOLHGS88.19-
Table 6. Accuracy and PIR (%) results of the PH2 dataset.
Table 6. Accuracy and PIR (%) results of the PH2 dataset.
Ref.YearClassification ModelAC (%)PIR (%)
[65]2017ANN92.504.08
[66]2019Kernel Sparse93.503.04
[67]2020DenseNet201 + SVM92.004.59
[6]2020DenseNet201 + KNN93.163.39
[68]2021ResNet50 + NB95.401.07
OurpresentDOLHGS96.43-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dahou, A.; Aseeri, A.O.; Mabrouk, A.; Ibrahim, R.A.; Al-Betar, M.A.; Elaziz, M.A. Optimal Skin Cancer Detection Model Using Transfer Learning and Dynamic-Opposite Hunger Games Search. Diagnostics 2023, 13, 1579. https://doi.org/10.3390/diagnostics13091579

AMA Style

Dahou A, Aseeri AO, Mabrouk A, Ibrahim RA, Al-Betar MA, Elaziz MA. Optimal Skin Cancer Detection Model Using Transfer Learning and Dynamic-Opposite Hunger Games Search. Diagnostics. 2023; 13(9):1579. https://doi.org/10.3390/diagnostics13091579

Chicago/Turabian Style

Dahou, Abdelghani, Ahmad O. Aseeri, Alhassan Mabrouk, Rehab Ali Ibrahim, Mohammed Azmi Al-Betar, and Mohamed Abd Elaziz. 2023. "Optimal Skin Cancer Detection Model Using Transfer Learning and Dynamic-Opposite Hunger Games Search" Diagnostics 13, no. 9: 1579. https://doi.org/10.3390/diagnostics13091579

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop