Next Article in Journal
MALDI Imaging, a Powerful Multiplex Approach to Decipher Intratumoral Heterogeneity: Combined Hepato-Cholangiocarcinomas as Proof of Concept
Next Article in Special Issue
Kidney Cancer Diagnosis and Surgery Selection by Machine Learning from CT Scans Combined with Clinical Metadata
Previous Article in Journal
Epigenetic Rewiring of Metastatic Cancer to the Brain: Focus on Lung and Colon Cancers
Previous Article in Special Issue
Predicting Breast Cancer Events in Ductal Carcinoma In Situ (DCIS) Using Generative Adversarial Network Augmented Deep Learning Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Henry Gas Solubility Optimization Algorithm based Feature Extraction in Dermoscopic Images Analysis of Skin Cancer

1
Department of Biomedical Engineering, College of Engineering, Princess Nourah bint Abdulrahman University, Riyadh 11671, Saudi Arabia
2
Department of Accounting, Applied College, King Khalid University, Mohail Asser 63311, Saudi Arabia
3
Department of Software Engineering, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
4
Department of Electrical Engineering, Faculty of Engineering & Technology, Future University in Egypt, New Cairo 11845, Egypt
5
Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam bin Abdulaziz University, Al Kharj 11942, Saudi Arabia
*
Author to whom correspondence should be addressed.
Cancers 2023, 15(7), 2146; https://doi.org/10.3390/cancers15072146
Submission received: 15 February 2023 / Revised: 27 March 2023 / Accepted: 30 March 2023 / Published: 4 April 2023
(This article belongs to the Collection Artificial Intelligence and Machine Learning in Cancer Research)

Abstract

:

Simple Summary

Early diagnosis of skin cancer is vital for providing effective treatment for patients. Dermoscopy is a non-invasive approach that utilizes specific equipment to examine the skin and is helpful in determining the specific patterns and features that might confirm the existence of skin cancer. Recently, Machine Learning (ML) algorithms have been developed to analyze dermoscopic images and classify such images as either benign or malignant. Convolutional Neural Networks (CNNs) and other ML techniques, such as the Support Vector Machine (SVM) and Random Forest classifiers, have been used in the extraction of the features from the dermoscopic images. The extracted features are then used to classify the dermoscopic images as either benign or malignant. Therefore, the current study develops a new Deep Learning-based skin cancer classification method for dermoscopic images, which has the potential to improve the accuracy and efficiency of the skin cancer diagnosis process and produce better outcomes for the patients.

Abstract

Artificial Intelligence (AI) techniques have changed the general perceptions about medical diagnostics, especially after the introduction and development of Convolutional Neural Networks (CNN) and advanced Deep Learning (DL) and Machine Learning (ML) approaches. In general, dermatologists visually inspect the images and assess the morphological variables such as borders, colors, and shapes to diagnose the disease. In this background, AI techniques make use of algorithms and computer systems to mimic the cognitive functions of the human brain and assist clinicians and researchers. In recent years, AI has been applied extensively in the domain of dermatology, especially for the detection and classification of skin cancer and other general skin diseases. In this research article, the authors propose an Optimal Multi-Attention Fusion Convolutional Neural Network-based Skin Cancer Diagnosis (MAFCNN-SCD) technique for the detection of skin cancer in dermoscopic images. The primary aim of the proposed MAFCNN-SCD technique is to classify skin cancer on dermoscopic images. In the presented MAFCNN-SCD technique, the data pre-processing is performed at the initial stage. Next, the MAFNet method is applied as a feature extractor with Henry Gas Solubility Optimization (HGSO) algorithm as a hyperparameter optimizer. Finally, the Deep Belief Network (DBN) method is exploited for the detection and classification of skin cancer. A sequence of simulations was conducted to establish the superior performance of the proposed MAFCNN-SCD approach. The comprehensive comparative analysis outcomes confirmed the supreme performance of the proposed MAFCNN-SCD technique over other methodologies.

1. Introduction

Melanoma is the most severe type of skin cancer that appears on any part of the skin or near a mole. This skin cancer is characterized by an uncontrollable growth of the cells without any apoptosis [1]. In this scenario, such cells of the body parts turn out to be tumorous and start spreading to other parts of the party. Unlike the rest of the skin cancer types, such as basal cell carcinoma and squamous cell carcinoma, melanoma is less common in nature. However, melanoma is highly dangerous compared to the rest of the skin cancer types, as it spreads to distinct body parts if left untreated or undiagnosed in its early stages. Melanoma spreads rapidly across the body and affects almost all body parts [2]. The dermatologists utilize photographic or microscopic instruments to see the additional details that are relevant to lesions. When diagnosing skin cancer, the clinician refers the individual to a tumor expert who performs surgery on the lesions [3]. Dermoscopy is a microscopic technique that inspects the surface of the skin. This technique is utilized to distinguish benign lesions from malignant ones based on the captured images without removing the skin and other maddening tests. This analytical procedure is conducted completely based on the oncologist’s expertise and experience [4]. Such a scenario pushed the current study authors to develop a computer-aided technique by employing the dermoscopic images and displaying the outcomes as assisting apparatuses for dermatologists. Various studies have been conducted so far to attain superior outcomes in the disease diagnosis process.
Numerous methods have been modeled so far for the automatic detection of melanoma-affected skin parts [5]. At first, handcrafted features-related methods were presented for diagnosing melanoma. But, such methods did not yield good outcomes due to dissimilarities in the color, shape, and size of the melanoma moles [6]. Then, segmentation-related methods such as the Iterative Selection Thresholding (ISO) and adaptive thresholding were proposed to enhance the detection accuracy of such automated mechanisms. Such methods tend to work on the segmented part of the melanoma called ‘RoI’ [7]. DL-related techniques have gained more popularity in medical imaging and diagnostics processes in recent years. In techniques such as the CNN, a small portion of the images, with melanoma-affected portions, is considered to train the automatic identification mechanism [8]. There exist numerous potential applications for deep unsupervised learning-based feature extraction from these images, such as object detection, image classification, image retrieval, anomaly detection, generative modeling, etc. Such methods execute the segmentation process on the test images that are related to the trained method. The DL-related techniques exhibit superior performance in detecting and segmenting melanoma images than the handcrafted feature-related methods [9]. Such procedures can automatically calculate the complicated and representative feature set. Furthermore, the DL methods can also easily trace the skin moles of different sizes in the occurrence of noise, blur, the incidence of light, color variations, and intensity [10].
The current research article develops an Optimal Multi-Attention Fusion Convolutional Neural Network-based Skin Cancer Diagnosis (MAFCNN-SCD) technique for the diagnosis of melanoma cancer from dermoscopic images. In the presented MAFCNN-SCD technique, the data pre-processing is performed initially. Next, the MAFNet method is enforced as a feature extractor with Henry Gas Solubility Optimization (HGSO) algorithm as a hyperparameter optimizer. Finally, the Deep Belief Network (DBN) method is exploited for the detection and classification of skin cancer. A sequence of experiments was conducted to validate the improved performance of the proposed MAFCNN-SCD approach. The key contributions of the current research work are listed herewith.
  • An automated MAFCNN-SCD technique has been proposed in this study with pre-processing, MAFNet-based feature extraction, DBN classification, and HGSO-based hyperparameter tuning processes for skin cancer detection and classification. To the best of the authors’ knowledge, the proposed MAFCNN-SCD model is the first of its kind in this domain.
  • The authors employed MAFNet as a feature extractor with DBN as a skin cancer detection and classification classifier.
  • The hyperparameter optimization of the MAFNet model, using the HGSO algorithm with cross-validation, helped to boost the predictive outcomes of the proposed MAFCNN-SCD model for unseen data.

2. Literature Review

Shorfuzzaman [11] presented an interpretable CNN-based stacked ensemble structure for the detection of melanoma skin tumors at earlier stages. In this study, the Transfer Learning (TL) model was employed in the stacked ensemble framework, whereas a distinct number of CNN sub-models that apply similar classifier tasks were also gathered. A novel approach called meta-learner was employed in the prediction of every sub-model, and the last prediction outcomes were attained. Bhimavarapu and Battineni [12] intended to integrate the DL techniques for the automatic classification of melanoma in dermoscopic images. A Fuzzy-based GrabCut-stacked CNN (GC-SCNN) technique was validated using the trained images. The image extraction feature and the lesion classifier were leveraged, and the model’s efficacy was tested using distinct openly-accessible databases. The purpose of the study, conducted by Lafraxo et al. [13], was to automate the procedure of classifying the dermoscopic images comprising skin lesions as either benign or malignant. Thus, an enhanced DL-based solution with CNN was presented in this study. Data augmentation, regularization, and dropout were performed to avoid the over-fitting issue that is generally experienced in the CNN technique.
Banerjee et al. [14] examined a DL-based YOLO technique on the basis of the application of DCNNs. This technique was used to detect melanoma in digital and dermoscopic images. The authors suggested a faster and a precise outcome related to the typical CNNs. But, particular resourceful models were infused under two stages of segmentation. This segmentation is created by combining a graph model using a minimal spanning tree model and an L-type fuzzy number. The latter is related to approximation and mathematical extraction of the actual affected lesion regions in the feature extraction method. Daghrir et al. [15] established a hybrid system for melanoma skin tumor classification, and the method was employed for examining a few suspicious lesions. The presented method was dependent on the prediction of three distinct approaches, such as CNN and two typical ML techniques. These techniques were trained with a group of characteristics that explain the texture, border, and color of the skin tumor.
Tan et al. [16] presented an intelligent Decision Support System (DSS) for skin tumor classification. Specifically, the authors integrated the medically-essential features such as asymmetry, color, border irregularity, and other such dermoscopic structural features with the texture extraction features using the Histogram of Oriented Gradients (HOG), Grey Level Run Length Matrix and LBPs functions for tumor representations. Afterward, the authors presented two improved PSO techniques for the optimization of the features. In literature [17], a Hybrid DL (HDL) system was proposed fusing the sub-band of 3D wavelets. It was a non-invasive and objective system that was used for the inspection of skin images. During the primary phase of the HDL system, an easy Median Filter (MF) was utilized to remove unwanted data like noise and hair. During the secondary phase, the sub-band fusion method was used, and the 3D wavelet transform was executed to obtain the textural data in the dermoscopic images. In the last phase, the HDL system carried out a multiclass classification with the help of the fused sub-band.
Though several models have been proposed in the literature, the existing models do not focus on the hyperparameter selection process. This is a crucial process as it mostly influences classification performance. The hyperparameters such as epoch count, batch size, and learning rate selection are essential to attain effectual outcomes. Since the trial-and-error method for hyperparameter tuning is a tedious and erroneous process, the metaheuristic algorithms are applied. Therefore, in this research work, the HGSO algorithm is employed to select the parameters for the MAFNet model.

3. The Proposed Model

In this study, a new MAFCNN-SCD method has been proposed for the detection and classification of skin cancer from dermoscopic images. The major aim of the proposed MAFCNN-SCD technique is to diagnose and classify the type of skin cancers from dermoscopic images. It encompasses various stages such as image pre-processing, MAFNet feature extraction, HGSO hyperparameter tuning, and DBN classification. Figure 1 defines the overall procedure of the proposed MAFCNN-SCD system.

3.1. Image Pre-Processing

In the presented MAFCNN-SCD approach, the data is pre-processed initially by following the Weiner Filter technique. The corrupted image is determined as I ^ x , y , whereas the local mean is denoted by μ L ^ on a pixel window, the noise variance with the entire value is demonstrated by σ y 2 , whereas the local variance in the window is designated by σ ^ y 2 . Then, the probable method of denoising the image is shown below [18]:
I ^ ^ = I ^ x , y σ y 2 σ ^ y 2 I ^ x , y μ L ^
Here, when the noise variance across the image is equivalent to 0, then σ y 2 = 0 = > I ^ ^ = I ^ x , y . If the global noise variance is smaller, then the local variance is larger than the global variance, according to which the ratio becomes almost equal to 1.
If I ^ ^ y 2 σ y 2 , then I ^ ^ = Î x , y . Whereas a higher local variance portrays the occurrence of the edges in the image windows [19]. In this case, when the local and global variances correspond together, the formulation progresses as follows: I ^ ^ = μ L ^ a s σ ^ y 2 σ y 2 .

3.2. Feature Extraction Model

In this study, the MAFNet method is applied as a feature extractor. MAFNet encompasses 1 × 1 × 1 convolution layers [20]; four convolution models exist in the middle, whereas three convolution structures exist across all the modules. The convolution module is disseminated in a symmetrical structure like [2,2,2,2]. At last, the FC layer is present with a total of 26 convolution modules [21,22]. Afterward, the images are fed as input, and the 1 × 1 convolutional process is initially performed. Next, four convolution models are passed over four convolution models. The 3 × 3 convolutional operation is substituted by the Contextual Transformer (CoT) blocks in the original ResNet convolution blocks. After the preceding convolution, pooling, and excitation operations, the features extracted are then fed as input to the FC layers that perform the role of “classification” in the CNN technique [23]. The architecture of the MAFNet model is shown in Figure 2. The FC layers perform the “classification” process in the CNN method that incorporates the preceding and extremely-abstracted feature and maps the learned feature to a sample space. It also employs the S o f t m a x function to evaluate the probability of classification. At last, the output of the classifier outcome is given as follows.
S o f t m a x z j = e z i i = 1 n e z i ,
Equation (2) demonstrates the number of classes, z j represents the output value of the j - th node, z i indicates the output value of the i - th node, and e shows the natural constant, which can be determined as follows.
L C E = i = 1 N l i l o g p i ,
In Equation (3), l i refers to the unique thermal encoding of the tag i ( i 0 , , N 1 and N indicates the number of tags; when the target label is i , then l i = 1 , whereas the rest of the labels are equivalent to 0 , and p i denotes the predictive probability of the i - th label, i.e., the Softmax value.
It must be noted that the lowest learning rate might result in slower convergence, and the largest learning rate might result in a constant oscillation of the loss functions. In order to construct a model with high accuracy and optimum parameters followed by its quick training, a dynamic-learning-rate approach is utilized. This helps in adjusting the learning rate per 30 epochs, for which the formula is given below [24].
l r = l r 0 × 0.1 e p o c h 30 ,
In Equation (4), l r denotes the present learning rate, l r 0 indicates the primary learning rate, and e p o c h represents the overall number of training rounds.
In this study, the HGSO algorithm is applied as a hyperparameter optimizer. Hashim et al. [25] presented the HGSO technique, which is a newly-found physics-inspired optimization algorithm. This technique has been constructed based on Henry’s gas law which defines the rules for the solubility of a gas in a liquid. In general, temperature and pressure are two major elements that significantly impact the outcome of solubility. In regard to pressure, the ability of a gas to become solvable in the liquid increases when the pressure increases. At high temperatures, the solubility of the solid increases. On the other hand, gases cannot dissolve. By utilizing these two significant characteristics, the HGSO approach comprises eight steps, as listed herewith. Initially, the values of Henry’s constant per group j H j t , the number of gases (population), position, and the partial pressure P i , j of the gas i at every group j are generated and are given as a mathematical form herewith.
X i T + 1 = X m i n + r × X m a x X m i n ,
In Equation (5), X i signifies the location of the i - th gas in population N ,   r is determined by the chaotic number between 0 and 1, and X m i n and X m a x denote the bounds of the search space and (t) shows the iteration.
H j t = l 1 × r a n d ( 0 t 1 ) t P i , j = l 2 × r a n d ( 0 , 1 ) t C j = l 3 × r a n d 0 , 1 ,
In Equation (6), l 1 to l 3 indicate the constant numbers that correspond to the values between 5 × 10−2, 100 and 1 × 10−2.
Next, the clustering process is executed in which the population of the gas is categorized according to the type of gas. In all the groups, each gas has a similar H j .
Then, the evaluation process is conducted to determine the most suitable gas from every cluster j that attains the maximum equilibrium location over the rest of the gases. In this stage, the ranking is used to find the most suitable gas among the entire swarm [26].
Henry’s coefficient is expressed in the following equation.
H j t + 1 = H j t × e C j 1 T t 1 T θ , T t = e τ i t e r
In Equation (7), H j demonstrates the coefficient of Henry’s gas rules in all the groups j ,   T characterizes the temperature, and T θ describes the fixed quantity with a value of 298. Furthermore, i t e r signifies the overall iteration count. The solubility updating formula is given below in which S i , j demonstrates the solvability of a gas i in all the groups j ,   P i , j characterizes the parochial pressure on gas i in all the groups j , and K denotes a fixed value.
S i , j t = K × H j t + 1 × P i , j t .
Step 6 characterizes a formula to upgrade the location of all the gases in Equations (9) and (10):
X i , j t + 1 = X i , j t + F × r × γ × X i , b e s τ t X i , j t   + F × r × α × S i , j t × X i , b e s t t X i , j t
γ = β × e x p ( F b e s t t + ε F i , j t + ε ) t ε = 0.05
In Equation (10), X i , j characterizes the condition of all the gases i in all the groups j ,   r denotes the random numbers between 0 and 1, whereas τ denotes the iteration time, and X i , b e s t illustrates the most suitable gas i in all the groups j . Further, X b e s t characterizes the most suitable gas amongst the entire population [27]. In addition to these, γ signifies the ability of all the gases j in all the groups i to interact through the gases in its group, and α illustrates the effect of the remaining gases on i in every group j and takes the value of 1. Additionally, a certain number is allocated to β .   F i , j shows the fitness of all the gases i in all the groups j at the same time, and F b e s t demonstrates the fitness of the most suitable gas in the entire population. The formula to prevent the local optima situation is given below.
N w = N ( r a n d C 2 C 1 + C 1 , C 1 ; C 1 = 0.1 , C 2 = 0.2 ,
In Equation (11), N w and N indicate the worst agent and the number of search agents.
Lastly, the formula for updating the location of the worst agent is given below.
G i , j = G M i n i , j + r × ( G M a x i , j G M i n i , j ) t
In Equation (12), G i , j demonstrates the condition for all the gases i in group j ,   r denotes a number that is disseminated between 0 and 1, and G M i n i , j and G M a x i , j indicate the bounds for the algorithm. Algorithm 1 illustrates the steps followed in the HGSO approach. The HGSO method derives a Fitness Function (FF) for the enhancement of the classification outcome. It sets a positive value to designate the superior outcomes of the candidate solutions.
Algorithm 1: Pseudocode of HGSO Algorithm.
Initialization: i 1 = 1 , 2 , N , number of gas kinds i , H j , P i , j , C j , l 1 , l 2 , and l 3 .
Split the population agent into a number of gas kinds (cluster) with a similar Henry’ s constant value H j .
Estimate every cluster j .
Obtain the more suitable gas X i , better in all the clusters, and the better search agent X b e s t .
While t < maximal iteration count, do
For every search agent, do
Upgrade the position of each search agent through Equations (9) and (10).
End for
Upgrade Henry’s s coefficient of all the gas kinds based on Equation (7).
Upgrade solubility of all the gases based on Equation (8).
Select and Rank the amount of worst agents based on Equation (11).
Upgrade the position of the worst agent based on Equation (12).
Upgrade the more suitable Xi, better, and the better search agent Xbest.
End while
  t = t + 1
Return Xbest
In this work, the reduced classifier error rate is signified as FF as mentioned in Equation (13).
F i t n e s s = n u m b e r   o f   m i s c l a s s i f i e d   s a m p l e s T o t a l   n u m b e r   o f   s a m p l e s 100

3.3. Skin Cancer Detection Model

In this last stage, the DBN model is exploited for the detection and classification of skin cancer. DBN is a NN that comprises numerous Restricted Boltzmann Machines (RBM) [28]. The input unit specifies the character of the original dataset, whereas the output unit specifies the label of this dataset. From the input to the output layers, the key features of the data are mined from the deep architecture via the layer abstraction process. The DBN can be accomplished by stacking numerous RBMs. The initial-layer RBM is the input of the DBN of pipeline leak detection, whereas the output is characterized by the latter, i.e., RBM Hidden Layer (HL). The DBN is processed as an MLP, and applied to the classifier, after which the LR is added to the output. The DBN model comprises certain RBMs. Every RBM has a visible layer (VL) and an HL. Consider v = { 0 , 1 } n and h = { 0 , 1 } m as the states of VL and HL, correspondingly. The quantity of the RBM joint configuration energy comprises biases and weights.
E v , h ; θ = i = 1 n a j v j j = 1 m b j h j i = 1 j n = 1 m w i j v j h j
Here, θ = a j , b j , w i j represents the model parameter, w i j ˙ indicates the weight between the hidden unit j and the visible unit i ;   a j and b j show the biases of the VL and HL, correspondingly; n and m denote the count of visible and hidden units, correspondingly. The joint likelihood equation for VL and HL is given below.
p v , h ; θ = 1 Z θ e x p E v , h ; θ
In Equation (15), Z θ refers to the normalizing factor that is formulated as follows.
Z θ = v h e x p E v , h ; θ
Since the visible–visible and hidden–hidden cases are independent of each other, the conditional probability of this unit can be formulated using the following equation [29]:
p h j = 1 | θ = 1 1 + exp b j i w i j v i
p v j = 1 | θ = 1 1 + exp a i j w i j h j
The layer-wise learning mechanism of the DBN comprises three HLs. The training dataset originates from a similar pipeline with similar experimental conditions and leakage sizes. Firstly, the training dataset is transferred to the VLs on the initial RBM unit. Then, the hidden unit feeds the input dataset in the VL. At last, the VLs of the 2nd RBM unit obtain a hidden unit in the RBM [30]. The subsequent individual RBM units accomplish the exercises of the DBN structure. Figure 3 demonstrates the infrastructure of the DBN.

4. Performance Evaluation

The proposed model was simulated using Python 3.6.5 tool on a PC with configurations such as i5–8600 k, GeForce 1050 Ti 4 GB, 16 GB RAM, 250 GB SSD, and 1 TB HDD. The parameter settings are given herewith: learning rate: 0.01, dropout: 0.5, batch size: 5, epoch count: 50, and activation: ReLU.

4.1. Dataset Used

The presented MAFCNN-SCD model for the skin cancer classification process was validated using two benchmark databases such as the ISIC 2017 database [31] and the HAM10000 database [32]. The ISIC 2017 dataset comprises 2000 images under three classes, whereas the HAM10000 dataset holds a total of 10,082 samples under seven classes. Table 1 and Table 2 show a detailed description of the datasets under study. Figure 4 demonstrates some of the sample images from the datasets.

4.2. Results Analysis

The confusion matrices generated by the proposed MAFCNN-SCD approach on ISIC 2017 dataset are demonstrated in Figure 5. The figure states that the proposed MAFCNN-SCD system proficiently recognized all three types of skin cancers from the applied dermoscopic images.
Table 3 portrays the overall skin cancer classification outcomes achieved by the proposed MAFCNN-SCD method on ISIC 2017 dataset. On the entire dataset, the MAFCNN-SCD method attained the average a c c u y , s e n s y , s p e c y , F s c o r e , and Mathew Correlation Coefficient (MCC) values, such as 93.47%, 79.92%, 90%, 85.50%, and 79.26%, respectively. Concurrently, on 70% TR database, the proposed MAFCNN-SCD approach achieved the average a c c u y , s e n s y , s p e c y , F s c o r e , and MCC values, such as 94.00%, 81.31%, 90.61%, 86.61%, and 80.66%, correspondingly. In parallel, on 30% of the TS database, the presented MAFCNN-SCD method achieved the average a c c u y , s e n s y , s p e c y , F s c o r e , and MCC values, such as 92.22%, 77.07%, 88.67%, 83.05%, and 76.23%, correspondingly.
Both the Training Accuracy ( T R a c c ) and the Validation Accuracy ( V L a c c ) values acquired by the MAFCNN-SCD approach under ISIC 2017 dataset are presented in Figure 6. The simulation results emphasize that the proposed MAFCNN-SCD algorithm gained increased T R a c c and V L a c c values, while the V L a c c values were better than the T R a c c values.
Both Training Loss ( T R l o s s ) and the Validation Loss ( V L l o s s ) values realized by the proposed MAFCNN-SCD system under ISIC 2017 dataset are exhibited in Figure 7. The simulation results represent that the proposed MAFCNN-SCD approach obtained the least T R l o s s and V L l o s s values, while the V L l o s s values were lesser than the T R l o s s values.
The precision-recall examination outcomes achieved by the proposed MAFCNN-SCD system under ISIC 2017 dataset are shown in Figure 8. The figure portrays that the proposed MAFCNN-SCD method produced high precision-recall values under each class label.
The confusion matrices generated by the proposed MAFCNN-SCD algorithm on the HAM10000 dataset are exhibited in Figure 9. The figure shows that the MAFCNN-SCD approach proficiently recognized all seven types of skin cancer on the applied dermoscopic images.
Table 4 depicts the overall skin cancer classification results attained by the proposed MAFCNN-SCD method on the HAM10000 dataset. On the entire dataset, the proposed MAFCNN-SCD system achieved the average a c c u y , s e n s y , s p e c y , F s c o r e , and MCC values such as 99.39%, 92.25%, 99.49%, 93.08%, and 92.60%, correspondingly. At the same time, on 70% of the TR database, the MAFCNN-SCD algorithm attained the average a c c u y , s e n s y , s p e c y , F s c o r e , and MCC outcomes such as 99.42%, 92.40%, 99.51%, 93.27%, and 92.82%, correspondingly. Concurrently, on 30% of the TS database, the proposed MAFCNN-SCD algorithm achieved the average a c c u y , s e n s y , s p e c y , F s c o r e , and MCC values such as 99.34%, 91.92%, 99.44%, 92.67%, and 92.15%, correspondingly.
Both T R a c c and V L a c c values acquired by the MAFCNN-SCD methodology under the HAM10000 dataset are exhibited in Figure 10. The simulation results confirm that the proposed MAFCNN-SCD approach obtained the maximum T R a c c and V L a c c values, whereas the V L a c c values were better than the T R a c c values.
Both T R l o s s and V L l o s s values realized by the proposed MAFCNN-SCD system under the HAM10000 dataset are shown in Figure 11. The simulation results manifest that the presented MAFCNN-SCD approach obtained the least T R l o s s and V L l o s s values, while the V L l o s s values were lesser than the T R l o s s values.
The precision-recall analysis results achieved by the MAFCNN-SCD system upon the HAM10000 dataset are exhibited in Figure 12. The figure demonstrates that the proposed MAFCNN-SCD method produced the maximum precision-recall values under each class label.

4.3. Discussion

Table 5 and Figure 12 report the comparative study outcomes of the MAFCNN-SCD model and other recent models on the ISIC 2017 dataset. The simulation results imply that the MobileNet method reached a poor performance while the MSVM model gained a slightly raised outcome. Next, the NB, KELM, and DenseNet169 models produced closer skin cancer classification performance. However, the proposed MAFCNN-SCD model accomplished the maximum performance with an accuracy of 92.22%.
Table 6 reports the comparative analysis outcomes achieved by the proposed MAFCNN-SCD method and other recent techniques on the HAM10000 dataset. The simulation outcomes imply that the NB methodology performed poorly, whereas the MobileNet approach acquired a slightly raised outcome.
Then, the KELM, MSVM, and DenseNet169 algorithms achieved closer skin cancer classification performance. But, the proposed MAFCNN-SCD model achieved the maximum performance. The comprehensive comparative analysis outcomes established the enhanced performance of the MAFCNN-SCD technique over other methodologies with maximum accuracy values such as 92.22% and 99.34% on the ISIC 2017 and HAM10000 datasets, respectively. These results establish the effectual skin cancer classification performance of the proposed MAFCNN-SCD method. The enhanced performance of the proposed model is due to the implementation of the HGSO algorithm in the hyperparameter tuning process.

5. Conclusions

In this study, a new MAFCNN-SCD approach has been modeled for skin cancer detection and classification from dermoscopic images. The major aim of the proposed MAFCNN-SCD technique is to identify and classify skin cancer from dermoscopic images. In the presented MAFCNN-SCD technique, the data pre-processing is performed initially. Next, the MAFNet methodology is applied as a feature extractor with the HGSO algorithm as a hyperparameter optimizer. Finally, the DBN method is exploited for skin cancer detection and classification. A sequence of experiments was conducted to showcase the supreme performance of the proposed MAFCNN-SCD approach. The comprehensive comparative analysis outcomes establish the enhanced performance of the MAFCNN-SCD technique over other methodologies with maximum accuracy values, such as 92.22% and 99.34% on the ISIC 2017 and HAM10000 datasets, respectively. Thus, the proposed model can be employed for the automated skin cancer classification process. In the future, the deep instance segmentation process can be incorporated to extend the detection rate of the MAFCNN-SCD technique to achieve low error rates in quantifying the skin lesion’s structure, boundary, and scale. To further increase the system’s performance, huge training datasets should be utilized to avoid under- and over-segmentation cases. Besides, the computation complexity of the proposed model should also be investigated in the future.

Author Contributions

Conceptualization, M.O. and M.M.; methodology, A.M.H.; software, A.S.S.; validation, A.A., M.M. and A.M.H.; formal analysis, A.A.; investigation, M.I.A.; resources, A.A.; data curation, A.S.S.; writing—original draft preparation, M.O., M.M., A.M.H., A.E.O. and A.A.A.; writing—review and editing, M.O., A.A.A., A.S.S. and A.E.O.; visualization, A.A.; supervision, M.O.; project administration, A.M.H.; funding acquisition, M.O., M.M. and A.A. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through small group Research Project under grant number (PGR.1/124/43). Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R203), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. Research Supporting Project number (RSPD2023R787), King Saud University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

This article does not contain any studies with human participants performed by any of the authors.

Data Availability Statement

Data sharing is not applicable to this article as no datasets were generated during the current study.

Conflicts of Interest

The authors declare that they have no conflict of interest. The manuscript was written through the contributions of all authors. All authors have given approval for the final version of the manuscript.

References

  1. Adegun, A.; Viriri, S. Deep learning techniques for skin lesion analysis and melanoma cancer detection: A survey of state-of-the-art. Artif. Intell. Rev. 2021, 54, 811–841. [Google Scholar] [CrossRef]
  2. Kadampur, M.A.; Al Riyaee, S. Skin cancer detection: Applying a deep learning based model driven architecture in the cloud for classifying dermal cell images. Inform. Med. Unlocked 2020, 18, 100282. [Google Scholar] [CrossRef]
  3. Jojoa Acosta, M.F.; Caballero Tovar, L.Y.; Garcia-Zapirain, M.B.; Percybrooks, W.S. Melanoma diagnosis using deep learning techniques on dermatoscopic images. BMC Med. Imaging 2021, 21, 1–11. [Google Scholar] [CrossRef] [PubMed]
  4. Hossin, M.A.; Rupom, F.F.; Mahi, H.R.; Sarker, A.; Ahsan, F.; Warech, S. Melanoma Skin Cancer Detection Using Deep Learning and Advanced Regularizer. In Proceedings of the 2020 International Conference on Advanced Computer Science and Information Systems (ICACSIS), Depok City, Indonesia, 17–18 October 2020; IEEE: New York, NY, USA, 2020; pp. 89–94. [Google Scholar]
  5. Alazzam, M.B.; Alassery, F.; Almulihi, A. Diagnosis of melanoma using deep learning. Math. Probl. Eng. 2021, 2021, 1423605. [Google Scholar] [CrossRef]
  6. Mahbod, A.; Tschandl, P.; Langs, G.; Ecker, R.; Ellinger, I. The effects of skin lesion segmentation on the performance of dermatoscopic image classification. Comput. Methods Programs Biomed. 2020, 197, 105725. [Google Scholar] [CrossRef]
  7. Kassani, S.H.; Kassani, P.H. A comparative study of deep learning architectures on melanoma detection. Tissue Cell 2019, 58, 76–83. [Google Scholar] [CrossRef]
  8. Ningrum, D.N.A.; Yuan, S.P.; Kung, W.M.; Wu, C.C.; Tzeng, I.S.; Huang, C.Y.; Li, J.Y.C.; Wang, Y.C. Deep learning classifier with patient’s metadata of dermoscopic images in malignant melanoma detection. J. Multidiscip. Healthc. 2021, 14, 877. [Google Scholar] [CrossRef]
  9. Mahbod, A.; Schaefer, G.; Ellinger, I.; Ecker, R.; Pitiot, A.; Wang, C. Fusing fine-tuned deep features for skin lesion classification. Comput. Med. Imaging Graph. 2019, 71, 19–29. [Google Scholar] [CrossRef] [Green Version]
  10. Minaee, S.; Wang, Y.; Aygar, A.; Chung, S.; Wang, X.; Lui, Y.W.; Fieremans, E.; Flanagan, S.; Rath, J. MTBI identification from diffusion MR images using bag of adversarial visual features. IEEE Trans. Med. Imaging 2019, 38, 2545–2555. [Google Scholar] [CrossRef]
  11. Shorfuzzaman, M. An explainable stacked ensemble of deep learning models for improved melanoma skin cancer detection. Multimed. Syst. 2022, 28, 1309–1323. [Google Scholar] [CrossRef]
  12. Bhimavarapu, U.; Battineni, G. Skin Lesion Analysis for Melanoma Detection Using the Novel Deep Learning Model Fuzzy GC-SCNN. Healthcare 2022, 10, 962. [Google Scholar] [CrossRef]
  13. Lafraxo, S.; Ansari, M.E.; Charfi, S. MelaNet: An effective deep learning framework for melanoma detection using dermoscopic images. Multimed. Tools Appl. 2022, 81, 16021–16045. [Google Scholar] [CrossRef]
  14. Banerjee, S.; Singh, S.K.; Chakraborty, A.; Das, A.; Bag, R. Melanoma diagnosis using deep learning and fuzzy logic. Diagnostics 2020, 10, 577. [Google Scholar] [CrossRef]
  15. Daghrir, J.; Tlig, L.; Bouchouicha, M.; Sayadi, M. Melanoma Skin Cancer Detection Using Deep Learning and Classical Machine Learning Techniques: A Hybrid Approach. In Proceedings of the 2020 5th International Conference on Advanced Technologies for Signal and Image Processing (ATSIP), Sfax, Tunisia, 2–5 September 2020; IEEE: New York, NY, USA, 2020; pp. 1–5. [Google Scholar]
  16. Tan, T.Y.; Zhang, L.; Lim, C.P. Intelligent skin cancer diagnosis using improved particle swarm optimization and deep learning models. Appl. Soft Comput. 2019, 84, 105725. [Google Scholar] [CrossRef]
  17. Maniraj, S.P.; Maran, P.S. A hybrid deep learning approach for skin cancer diagnosis using subband fusion of 3D wavelets. J. Supercomput. 2022, 78, 12394–12409. [Google Scholar] [CrossRef]
  18. Malibari, A.A.; Alzahrani, J.S.; Eltahir, M.M.; Malik, V.; Obayya, M.; Al Duhayyim, M.; Neto, A.V.L.; de Albuquerque, V.H.C. Optimal deep neural network-driven computer aided diagnosis model for skin cancer. Comput. Electr. Eng. 2022, 103, 108318. [Google Scholar] [CrossRef]
  19. Park, C.R.; Kang, S.H.; Lee, Y. Median modified wiener filter for improving the image quality of gamma camera images. Nucl. Eng. Technol. 2020, 52, 2328–2333. [Google Scholar] [CrossRef]
  20. Xu, H.; Jin, L.; Shen, T.; Huang, F. Skin Cancer Diagnosis Based on Improved Multiattention Convolutional Neural Network. In Proceedings of the 2021 IEEE 5th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 12–14 March 2021; IEEE: New York, NY, USA, 2021; Volume 5, pp. 761–765. [Google Scholar]
  21. Wang, C.; Zhang, S.; Lv, X.; Ding, W.; Fan, X. A Novel Multi-attention Fusion Convolution Neural Network for Blind Image Quality Assessment. Neural Process. Lett. 2022. [Google Scholar] [CrossRef]
  22. Zheng, H.; Song, T.; Gao, C.; Guo, T. M2FN: A Multilayer and Multiattention Fusion Network for Remote Sensing Image Scene Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  23. Chen, J.; Han, J.; Liu, C.; Wang, Y.; Shen, H.; Li, L. A Deep-Learning Method for the Classification of Apple Varieties via Leaf Images from Different Growth Periods in Natural Environment. Symmetry 2022, 14, 1671. [Google Scholar] [CrossRef]
  24. Chen, P.; Gao, J.; Yuan, Y.; Wang, Q. MAFNet: A Multi-Attention Fusion Network for RGB-T Crowd Counting. arXiv 2022, arXiv:2208.06761. [Google Scholar]
  25. Hashim, F.A.; Houssein, E.H.; Mabrouk, M.S.; Al-Atabany, W.; Mirjalili, S. Henry gas solubility optimization: A novel physics-based algorithm. Future Gener. Comput. Syst. 2019, 101, 646–667. [Google Scholar] [CrossRef]
  26. Mohammadi, D.; Abd Elaziz, M.; Moghdani, R.; Demir, E.; Mirjalili, S. Quantum Henry gas solubility optimization algorithm for global optimization. Eng. Comput. 2022, 38, 2329–2348. [Google Scholar] [CrossRef]
  27. Abd Elaziz, M.; Attiya, I. An improved Henry gas solubility optimization algorithm for task scheduling in cloud computing. Artif. Intell. Rev. 2021, 54, 3599–3637. [Google Scholar] [CrossRef]
  28. Yan, Y.; Hu, Z.; Yuan, W.; Wang, J. Pipeline leak detection based on empirical mode decomposition and deep belief network. Meas. Control. 2022, 56, 00202940221088713. [Google Scholar] [CrossRef]
  29. Zhang, Y.; You, D.; Gao, X.; Katayama, S. Online monitoring of welding status based on a DBN model during laser welding. Engineering 2019, 5, 671–678. [Google Scholar] [CrossRef]
  30. Hua, C.; Chen, S.; Xu, G.; Chen, Y. Defect detection method of carbon fiber sucker rod based on multi-sensor information fusion and DBN model. Sensors 2022, 22, 5189. [Google Scholar] [CrossRef]
  31. Codella, N.C.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin Lesion Analysis toward Melanoma Detection: A Challenge at the 2017 International Symposium on Biomedical Imaging (isbi), Hosted by the International Skin Imaging Collaboration (isic). In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; IEEE: New York, NY, USA; pp. 168–172. [Google Scholar]
  32. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 180161. [Google Scholar] [CrossRef]
  33. Kousis, I.; Perikos, I.; Hatzilygeroudis, I.; Virvou, M. Deep Learning Methods for Accurate Skin Cancer Recognition and Mobile Application. Electronics 2022, 11, 1294. [Google Scholar] [CrossRef]
  34. Khan, M.A.; Sharif, M.; Akram, T.; Damaševičius, R.; Maskeliunas, R. Skin Lesion-Segmentation and Multiclass Classification Using Deep Learning Features and Improved Moth Flame Optimization. Diagnostics 2021, 11, 811. [Google Scholar] [CrossRef]
Figure 1. The overall process of the proposed MAFCNN-SCD system.
Figure 1. The overall process of the proposed MAFCNN-SCD system.
Cancers 15 02146 g001
Figure 2. Architecture of MAFNet.
Figure 2. Architecture of MAFNet.
Cancers 15 02146 g002
Figure 3. DBN structure.
Figure 3. DBN structure.
Cancers 15 02146 g003
Figure 4. Sample images.
Figure 4. Sample images.
Cancers 15 02146 g004
Figure 5. Confusion matrices of the proposed MAFCNN-SCD system under ISIC 2017 dataset. (a) Entire database, (b) 70% of TR database, and (c) 30% of TS database.
Figure 5. Confusion matrices of the proposed MAFCNN-SCD system under ISIC 2017 dataset. (a) Entire database, (b) 70% of TR database, and (c) 30% of TS database.
Cancers 15 02146 g005
Figure 6. T R a c c and V L a c c results of the proposed MAFCNN-SCD system upon the ISIC 2017 dataset.
Figure 6. T R a c c and V L a c c results of the proposed MAFCNN-SCD system upon the ISIC 2017 dataset.
Cancers 15 02146 g006
Figure 7. T R l o s s and V L l o s s analyses outcomes of the proposed MAFCNN-SCD system upon the ISIC 2017 dataset.
Figure 7. T R l o s s and V L l o s s analyses outcomes of the proposed MAFCNN-SCD system upon the ISIC 2017 dataset.
Cancers 15 02146 g007
Figure 8. Precision recall analysis outcomes of the proposed MAFCNN-SCD system upon the ISIC 2017 dataset.
Figure 8. Precision recall analysis outcomes of the proposed MAFCNN-SCD system upon the ISIC 2017 dataset.
Cancers 15 02146 g008
Figure 9. Confusion matrices of the proposed MAFCNN-SCD method under HAM10000 dataset (a) Entire database, (b) 70% of TR database, and (c) 30% of TS database.
Figure 9. Confusion matrices of the proposed MAFCNN-SCD method under HAM10000 dataset (a) Entire database, (b) 70% of TR database, and (c) 30% of TS database.
Cancers 15 02146 g009
Figure 10. T R a c c and V L a c c outcomes of the proposed MAFCNN-SCD system upon the HAM10000 dataset.
Figure 10. T R a c c and V L a c c outcomes of the proposed MAFCNN-SCD system upon the HAM10000 dataset.
Cancers 15 02146 g010
Figure 11. T R l o s s and V L l o s s results of the proposed MAFCNN-SCD system upon the HAM10000 dataset.
Figure 11. T R l o s s and V L l o s s results of the proposed MAFCNN-SCD system upon the HAM10000 dataset.
Cancers 15 02146 g011
Figure 12. Precision recall analysis outcomes of the proposed MAFCNN-SCD system upon the HAM10000 dataset.
Figure 12. Precision recall analysis outcomes of the proposed MAFCNN-SCD system upon the HAM10000 dataset.
Cancers 15 02146 g012
Table 1. Details of the ISIC 2017 dataset [31].
Table 1. Details of the ISIC 2017 dataset [31].
ISIC 2017 Dataset
LabelClassNo. of Samples
Class-1Melanoma374
Class-2Seborrheic Keratosis254
Class-3Nevus1372
Total Number of Dataset2000
Table 2. Details on HAM10000 Dataset [32].
Table 2. Details on HAM10000 Dataset [32].
HAM10000 Dataset
DescriptionClassNo. of Samples
Actinic KeratosesAKIEC327
Basal Cell CarcinomaBCC541
Benign KeratosisBKL1099
DermatofibromaDF155
Melanocytic NevusNV6705
MelanomaMEL1113
VascularVASC142
Total No. of Dataset10,082
Table 3. Skin cancer classification outcomes of the proposed MAFCNN-SCD system upon the ISIC 2017 dataset.
Table 3. Skin cancer classification outcomes of the proposed MAFCNN-SCD system upon the ISIC 2017 dataset.
Labels A c c u y S e n s y S p e c y F s c o r e MCC
Entire Dataset
Class-194.471.6699.6382.7280.84
Class-295.7568.999.6680.4679.56
Class-390.2599.270.793.3277.39
Average93.4779.929085.579.26
Training Phase (70%)
Class-194.7172.999.7483.7782
Class-296.2971.8699.5982.1981.22
Class-39199.1872.4993.8678.77
Average9481.3190.6186.6180.66
Testing Phase (30%)
Class-193.6768.7599.3980.2178.1
Class-294.563.2299.8176.9276.28
Class-388.599.2566.8392.0274.32
Average92.2277.0788.6783.0576.23
Table 4. Skin cancer classification outcomes of the proposed MAFCNN-SCD system upon the HAM10000 dataset.
Table 4. Skin cancer classification outcomes of the proposed MAFCNN-SCD system upon the HAM10000 dataset.
Labels A c c u y S e n s y S p e c y F s c o r e MCC
Entire Dataset
AKIEC99.5292.3599.7692.6492.39
BCC99.2792.4299.6593.1192.73
BKL99.569999.6398.0297.78
DF99.6287.199.8287.6687.47
NV98.649997.9398.9896.95
MEL99.5898.3899.7398.1297.88
VASC99.5577.4699.8783.0283.01
Average99.3992.2599.4993.0892.6
Training Phase (70%)
AKIEC99.5393.3399.7492.7292.48
BCC99.3592.7899.793.5693.22
BKL99.699.0999.6798.2197.99
DF99.6687.8599.8488.6888.51
NV98.799.0398.0399.0397.06
MEL99.5298.1999.6897.8197.54
VASC99.5676.5399.8982.8782.94
Average99.4292.499.5193.2792.82
Testing Phase (30%)
AKIEC99.590.299.8392.4692.24
BCC99.0791.7199.5492.2291.73
BKL99.4798.7799.5697.5897.29
DF99.5485.4299.7685.4285.18
NV98.5198.9497.798.8796.7
MEL99.7498.8399.8598.8398.68
VASC99.5479.5599.8383.3383.2
Average99.3491.9299.4492.6792.15
Table 5. Comparative analysis outcomes of the proposed MAFCNN-SCD model and other recent methodologies upon the ISIC 2017 dataset [33,34].
Table 5. Comparative analysis outcomes of the proposed MAFCNN-SCD model and other recent methodologies upon the ISIC 2017 dataset [33,34].
ISIC 2017 Dataset
MethodsAccuracySensitivitySpecificityF-Score
MAFCNN-SCD92.2277.0788.6783.05
Naïve Bayes89.7774.784.0281.37
KELM88.0477.0384.4983.2
MSVM87.1575.4483.1981.45
MobileNet85.0374.1787.9881.18
DenseNet16989.4276.8386.2883.27
Table 6. Comparative analysis outcomes of the proposed MAFCNN-SCD system and other recent algorithms upon the HAM10000 Dataset [33,34].
Table 6. Comparative analysis outcomes of the proposed MAFCNN-SCD system and other recent algorithms upon the HAM10000 Dataset [33,34].
ISIC 2017 Dataset
MethodsAccuracySensitivitySpecificityF-Score
MAFCNN-SCD92.2277.0788.6783.05
Naïve Bayes89.7774.784.0281.37
KELM88.0477.0384.4983.2
MSVM87.1575.4483.1981.45
MobileNet85.0374.1787.9881.18
DenseNet16989.4276.8386.2883.27
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Obayya, M.; Alhebri, A.; Maashi, M.; S. Salama, A.; Mustafa Hilal, A.; Alsaid, M.I.; Osman, A.E.; Alneil, A.A. Henry Gas Solubility Optimization Algorithm based Feature Extraction in Dermoscopic Images Analysis of Skin Cancer. Cancers 2023, 15, 2146. https://doi.org/10.3390/cancers15072146

AMA Style

Obayya M, Alhebri A, Maashi M, S. Salama A, Mustafa Hilal A, Alsaid MI, Osman AE, Alneil AA. Henry Gas Solubility Optimization Algorithm based Feature Extraction in Dermoscopic Images Analysis of Skin Cancer. Cancers. 2023; 15(7):2146. https://doi.org/10.3390/cancers15072146

Chicago/Turabian Style

Obayya, Marwa, Adeeb Alhebri, Mashael Maashi, Ahmed S. Salama, Anwer Mustafa Hilal, Mohamed Ibrahim Alsaid, Azza Elneil Osman, and Amani A. Alneil. 2023. "Henry Gas Solubility Optimization Algorithm based Feature Extraction in Dermoscopic Images Analysis of Skin Cancer" Cancers 15, no. 7: 2146. https://doi.org/10.3390/cancers15072146

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop