Next Article in Journal
Pallidal Recordings in Chronically Implanted Dystonic Patients: Mitigation of Tremor-Related Artifacts
Previous Article in Journal
Response to Mechanical Properties and Physiological Challenges of Fascia: Diagnosis and Rehabilitative Therapeutic Intervention for Myofascial System Disorders
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

WBM-DLNets: Wrapper-Based Metaheuristic Deep Learning Networks Feature Optimization for Enhancing Brain Tumor Detection

by
Muhammad Umair Ali
1,
Shaik Javeed Hussain
2,
Amad Zafar
1,
Muhammad Raheel Bhutta
3,* and
Seung Won Lee
4,*
1
Department of Intelligent Mechatronics Engineering, Sejong University, Seoul 05006, Republic of Korea
2
Department of Electrical and Electronics, Global College of Engineering and Technology, Muscat 112, Oman
3
Department of Electrical and Computer Engineering, University of UTAH Asia Campus, Incheon 21985, Republic of Korea
4
Department of Precision Medicine, Sungkyunkwan University School of Medicine, Suwon 16419, Republic of Korea
*
Authors to whom correspondence should be addressed.
Bioengineering 2023, 10(4), 475; https://doi.org/10.3390/bioengineering10040475
Submission received: 21 March 2023 / Revised: 7 April 2023 / Accepted: 11 April 2023 / Published: 14 April 2023
(This article belongs to the Section Biosignal Processing)

Abstract

:
This study presents wrapper-based metaheuristic deep learning networks (WBM-DLNets) feature optimization algorithms for brain tumor diagnosis using magnetic resonance imaging. Herein, 16 pretrained deep learning networks are used to compute the features. Eight metaheuristic optimization algorithms, namely, the marine predator algorithm, atom search optimization algorithm (ASOA), Harris hawks optimization algorithm, butterfly optimization algorithm, whale optimization algorithm, grey wolf optimization algorithm (GWOA), bat algorithm, and firefly algorithm, are used to evaluate the classification performance using a support vector machine (SVM)-based cost function. A deep-learning network selection approach is applied to determine the best deep-learning network. Finally, all deep features of the best deep learning networks are concatenated to train the SVM model. The proposed WBM-DLNets approach is validated based on an available online dataset. The results reveal that the classification accuracy is significantly improved by utilizing the features selected using WBM-DLNets relative to those obtained using the full set of deep features. DenseNet-201-GWOA and EfficientNet-b0-ASOA yield the best results, with a classification accuracy of 95.7%. Additionally, the results of the WBM-DLNets approach are compared with those reported in the literature.

1. Introduction

Uncontrolled cell development causes brain abnormalities, such as brain tumors. This syndrome has been associated with a considerable number of deaths worldwide and is commonly recognized as a lethal disease [1]. Brain tumors, whether benign or malignant, cause damage to adjacent brain tissue owing to the increased pressure exerted on the skull [2]. In 2022, approximately 88,970 people suffered brain tumors, among which 63,040 were benign, whereas the remaining 25,930 were malignant [3,4]. Gliomas, meningiomas, and pituitary gland tumors are several types of tumors that may affect the human brain [5,6]. Hence, the early detection of brain tumors is critical for timely and efficient treatment.
Neuro-oncologists can effortlessly and rapidly diagnose brain tumors using computer-aided diagnostic (CAD) technologies. The development of new imaging technologies, such as magnetic resonance imaging (MRI), to visualize the characteristics of brain tumors has increased in recent years. MRI is a useful technique that offers detailed scans of tissues and organs. Owing to its noninvasive nature, MRI is the most reliable method for locating and measuring brain tumors [7]. Correctly analyzing multidimensional MRI data can assist in localizing and tracking disease development and advising on treatment. These scans provide vital information regarding the size, shape, and location of brain tumors without exposing the patient to significant levels of ionizing radiation [8]. Radiologists use MRIs to assess and treat brain tumors. Hence, a major concern in radiology is the assessment of brain tumors using imaging techniques for various brain lesions. Several significant obstacles must be overcome to manage brain tumors, including early identification and therapy, which are crucial for patient survival.
In healthcare imaging, machine learning has been utilized for disease diagnosis in breast [9,10], brain [11,12,13], and lung [14,15] tumors. Recently, research on brain tumor segmentation, with numerous segmentation methods using different datasets, has increased considerably. Currently, three types of segmentation models have been developed [16]: supervised machine learning [17], clustering-based segmentation [18,19], and deep learning [13].
The segmentation challenge was reformulated using supervised learning algorithms as a tumor pixel classification problem. This approach produces the desired segmentation classes as the output of the model after feeding the retrieved features of the image as the input. A previous study [17] presented a hybrid random forest and active contour model for brain tumor classification. Another study used support vector machine (SVM) to categorize brain tumors [20,21]. Both the aforementioned studies utilized brain MRI to detect brain tumors. Owing to the similarity of the brain MRI scans, the precision of the model was low for the subcategorization of brain tumors.
Clustering-based segmentation methods divide MRI scans into distinct subcategories and pinpoint the area of interest in each scan. In this method, pixels with a high degree of similarity within each region are categorized as belonging to a particular region. By contrast, pixels that differ from those inside an area are classified as regions that do not belong to the area of interest. K-means clustering is a popular unsupervised machine-learning approach for separating an area of interest from other components of an image. It has been applied for segmenting brain tumors with fair precision as it uses a modest amount of processing time and power [22], and is most effective when large datasets are used. Its flaws include susceptibility to outliers and inadequate definition of the tumor region [23]. In one study, the authors extracted KAZE and speeded up robust local-level features to classify brain MRIs into glioma, meningioma, no tumor, and pituitary classes [18]. Subsequently, they used 8 × 8 pixels for feature extraction, and applied k-means clustering to segment 400 features for each descriptor. The model yielded an accuracy of 95.33% for the multiclass problems. In a subsequent study [19], the authors applied the PSO-based ReliefF algorithm to remove redundant features. They reduced the feature vector from 800 to 169 and achieved an accuracy of 96.30% using the k-fold method. In another study, the tumor area was compared with other brain imaging areas [24] and the expected tumor parts could not be distinguished with certainty. Clustering may lead to imprecise tumor size identification, resulting in ineffective therapy and higher morbidity and fatality rates.
Deep learning segmentation models accurately calculate the features of brain MRI images using the layers of a deep learning network [25]. A deep learning network was developed to segment complete and masked brain MRI images into two categories [26]. The proposed model had classification accuracies of 92.9% and 89.5% for a complete MRI and a mask, respectively. Abiwinanda et al. [27] presented a simplified deep learning network architecture to categorize MRI images into subgroups; the accuracy of this model was only 84.19%. In another study, a 22-layer deep learning network exhibited an accuracy of 95.56% for tertiary class problems (glioma, meningioma, and pituitary gland tumor) [28]. Subsequently, the authors applied data augmentation to balance the data; nevertheless, the use of data augmentation in real-time applications was unreliable. A 25-layer deep learning network reported an accuracy of 92.66% [29]. Pretrained networks, such as GoogleNet and ResNet-50, also reported classification accuracies of 98% and 97.2% for brain tumor detection, respectively [30,31]. Training a pretrained model requires considerable time. To address this problem, a study computed deep features using pretrained networks and utilized them to train conventional classifiers [32]. The concatenated deep features of ShuffleNet V2, DenseNet-169, and MnasNet yielded an accuracy of 93.72% using the SVM model. However, reportedly, the training time of the model was prolonged because of the large size of the training feature vectors. Therefore, the feature removal technique can be applied to improve model accuracy. Dokeroglu et al. [33] recently examined different wrapper-based feature-selection approaches. Wrapper techniques investigate the performance of each feature subset and combine a metaheuristic optimization algorithm with a classifier.
Therefore, in this study, wrapper-based metaheuristic deep-learning network (WBM-DLNet) feature optimization approaches were investigated to improve the classification accuracy of brain MRI scans. First, brain MRI scans were preprocessed to resize and reduce noise. Various pretrained models, such as DarkNet-19, DarkNet-53, DenseNet-201, EfficientNet-b0, GoogLeNet365, GoogLeNet, Inception-ResNet-v2, Inception-v3, MobileNet-v2, NASNet-Mobile, ResNet-101, ResNet-50, ResNet-18, ShuffleNet, SqueezeNet, and Xception, were used to extract information from brain MRI images. Wrapper approaches with various metaheuristic optimization algorithms, such as the marine predators algorithm (MPA), atom search optimization algorithm (ASOA), Harris hawks optimization algorithm (HHOA), butterfly optimization algorithm (BOA), whale optimization algorithm (WOA), grey wolf optimization algorithm (GWOA), bat algorithm (BA), and firefly algorithm (FA), were used to examine the classification performance using SVM. An empirical approach was used to select the best network. Based on the results of various feature subsets, a new concatenated feature vector was formed to evaluate the performance. The performance of the proposed framework was evaluated using an online brain MRI dataset.

2. Methods and Materials

2.1. Brain MRI Dataset

In this study, a collection of brain MRI images available online was used. The dataset utilized herein was retrieved from the Kaggle website [34] and comprised four brain MRI classes. Table 1 presents information on the brain MRI dataset.

2.2. Preprocessing

Preprocessing is required to eliminate unnecessary information and reduce noise from brain MRI scans. Cropping is typically employed for removing nonbrain regions and reducing noise. In this study, a cropping approach was employed to compute the extreme points of the brain area. Dilation and erosion processes, the two fundamental morphological processes, were applied to reduce noise. In an image, dilation involves adding pixels to object borders, whereas erosion involves removing pixels from object boundaries. The size and shape of the structuring element used to process the image determine the number of pixels added to or subtracted from the objects. Additional information on the preprocessing can be found in [32,35]. In addition, the preprocessed images were scaled to meet the input requirements of the pretrained networks employed in this study.

2.3. Deep Feature Extraction

Features are the primary factors when categorizing images. Identifying essential features is vital for improving classification performance. Although brain MRI feature extraction can be performed manually or using a deep learning network, manual extraction is very time-consuming. The considerable diversity in brain MRI scans determines their accuracy. By contrast, deep learning neural networks employ a combination of convolutional, pooling, and fully connected layers to build the model. Pretrained deep learning models can extract deep features using the transfer learning method when the dataset is sufficiently small. Figure 1 illustrates the concept of deep feature extraction using a trained model (GoogLeNet).
Deep neural networks that have been trained on extensive image classification tasks are known as pre-trained deep learning models and are capable of extracting hierarchical features from images. These deep features are acquired by passing an image through the network and analyzing the activation of one or more layers, thus indicating the response of the network to various image features. These activations may be used to categorize or compare images and feed them into other classical machine learning models, such as SVM.
DarkNet-19, DarkNet-53, DenseNet-201, EfficientNet-b0, GoogLeNet365, GoogLeNet, Inception-ResNet-v2, Inception-v3, MobileNet-v2, NASNet-Mobile, ResNet-101, ResNet-50, ResNet-18, ShuffleNet, SqueezeNet, and Xception were the prominent pre-trained CNN models used in this study. Each of these models contains distinguishing features that capture multiple aspects of images, such as local and global patterns or fine-grained details. These models can be used in a range of computer vision applications, such as object identification, image segmentation, and image retrieval.

2.4. Wrapper-Based Feature Selection Approach

Feature selection is essential in numerous machine-learning applications because it affects model accuracy. Therefore, relevant features must be utilized to effectively characterize and classify items. However, the inclusion of irrelevant features may lead to low accuracy. Therefore, the most valuable features must be determined to increase the classification accuracy of the model. During the feature-selection procedure in this study, a subset of a wider set of features was selected to build the machine learning model. Note that a specific criterion is used to assess the quality of the new subset [36]. This can be accomplished using various strategies, such as filter-based, wrapper-based, or embedded feature selection [37]. These approaches can also assist in minimizing model complexity, thus resulting in quicker and more efficient processing.
In this study, wrapper-based algorithms were used to select the most appropriate features for training a machine learning model. Wrapper algorithms are machine learning methods for evaluating the performance of a group of features when used with a particular model (the “wrapper”) [38]. The goal of the wrapper is to assess the impact of the selected features on the accuracy of the model. The wrapper algorithm either selects the current subset of features or seeks an improved subset based on the evaluation results. This procedure is repeated until an optimal subset of features is obtained. Figure 2 illustrates the concept of the wrapper approach. For optimal feature selection, this study employed a variety of metaheuristic algorithms and wrapper-based approaches.
The goal of metaheuristic techniques is to estimate the solutions to complex problems. They are referred to as “meta” because they integrate multiple low-level heuristics to manage high-level optimization tasks [39]. Examples of metaheuristic algorithms include MPA, ASOA, HHOA, BOA, WOA, GWOA, BA, and FA [33,40]. In wrapper feature selection methods, learning algorithms evaluate the performance of the resulting feature subsets during classification. Metaheuristics are used as search algorithms to identify new optimal subsets [41]. The cost functions of all optimization methods are specified in Equation (1).
min ( I ) = ω ( 1 Accuracy ) + σ ( number of selected features Total number   of   features )
where, the value of ω and σ are 0.99 and 0.01, respectively [42].

2.4.1. Marine Predators Algorithm (MPA)

MPA is an optimization technique inspired by nature that adheres to principles that naturally control the best foraging tactics and encounter rates between predators and prey in marine environments. The hunting and foraging strategies of marine predators, such as sharks and dolphins, serve as the basis for MPA [43,44]. The four key phases used for the simulation are predation, reproduction, migration, and exploration. During the predation phase, the algorithm hunts for and catches food; during the reproduction phase, it transfers genetic material to the progeny; during the exploration phase, it searches for new locations in the search space; and during the migration phase, it relocates the predators to new locations.
Numerous optimization problems, including feature selection, image segmentation, and data clustering, have been effectively solved using this technique. MPA is very efficient and helpful in addressing optimization issues in various fields because it can obtain global optima in complicated search spaces [45].

2.4.2. Atom Search Optimization Algorithm (ASOA)

The principle of ASOA is the mobility and interaction of atoms in a physical system [46]. It is a population-based method in which a group of atoms searches for the best answer in a predetermined search space, and it involves the four key steps of initialization, attraction, repulsion, and migration. During the initialization phase, this method creates a collection of atoms with different positions and velocities in the search space; during the attraction stage, the atoms travel toward the best answer; during the repulsion stage, they move away from one another to explore other regions of the search space; and during the migration step, they relocate to a new location in the search space. Numerous optimization issues have been successfully solved using ASOA [47].

2.4.3. Harris Hawks Optimization Algorithm (HHOA)

Bairathi and Gopalani proposed the optimization technique of HHOA, which is based on the cooperative hunting style of Harris hawks [48]. This algorithm is built on a group of individuals called “hawks” who stand in for potential solutions. The four key phases of HHOA are scouting, prey finding, trap setting, and attacks. During the scouting phase, the program randomly creates an initial population of hawks; during the prey-finding phase, the hawks explore the search space to determine the best answers; during the trap-setting phase, they corner the victim; and during the attack phase, they work together to capture it. Numerous optimization issues have been solved successfully using the HHOA [49].

2.4.4. Butterfly Optimization Algorithm (BOA)

The metaheuristic optimization BOA was developed by Arora and Singh after they studied the feeding habits of butterflies in the environment [50]. The BOA mimics butterfly motion and behavior to explore and adapt to changing search space conditions. It has five stages: initialization, search, selection, mutation, and update. During the initialization stage, this technique randomly creates an initial butterfly population; during the search stage, the butterflies explore the search space and identify the best solutions; during the selection stage, they select the best answers from the population; during the mutation stage, the algorithm introduces random modifications to the selected solutions to explore new search space regions; and during the update stage, the algorithm changes the population, based on the latest solutions discovered during the mutation stage. The BOA has been used to solve various complex engineering problems [51].

2.4.5. Whale Optimization Algorithm (WOA)

Mirjalili and Lewis [52] proposed the WOA in 2016, the WOA is based on the hunting behavior of humpback whales, which hunt in groups and employ bubble nets to catch tiny fish or krill. The WOA algorithm can solve complex optimization problems by discovering the prey, enclosing it, and moving it in spiral bubble-net patterns. The process begins with a random solution. Subsequently, other agents alter their locations by selecting the best search agent and selecting a target for attack, which may be the best search agent or a random whale. The WOA has been successfully used for various optimization issues, including function optimization, scheduling, and data clustering [33].

2.4.6. Grey Wolf Optimization Algorithm (GWOA)

Mirjalili et al. developed the GWOA in 2014 based on the grey wolf social structure and hunting skills [53]. Grey wolves hunt in packs, each headed by an alpha wolf, the dominant member of the crew. The group’s second leader, the beta wolf, assists the alpha wolf and conveys the instructions, and the subordinate or delta wolves support. When the prey is within a certain range, the pack surrounds and assaults it. Alpha, beta, and delta wolves are the top three GWOA algorithmic solutions. The remaining wolves are omega wolves and have no bearing on the decisions made in the subsequent iterations. The GWOA employs this social structure and prey-hunting strategies as a mathematical representation to solve optimization issues in domains ranging from computer science to engineering, mathematics, physics, and finance [54,55].

2.4.7. Bat Algorithm (BA)

Yang first presented the BA in 2010 [56] as an optimization technique. This approach is modeled after how bats search for prey by flying randomly, making noise, and listening to echoes [57]. The BA maintains a population of alternative solutions to the optimization problem. It iteratively enhances these solutions using random walk and exploitation procedures. The exploitation operation enables the algorithm to focus on the most promising answers. By contrast, the random-walk operation mimics the unpredictable search behavior of bats. The algorithm also uses a frequency-tuning mechanism motivated by the frequency modulation of bat sounds, which allows it to break out of local optima and discover superior solutions. In addition, the BA is easy to deploy and does not require complicated parameter tuning. Echolocation can be coded as a method for improving the objective function [58].

2.4.8. Firefly Algorithm (FA)

Motivated by the flashing qualities of fireflies, Yang presented the FA in 2010 [59], based on flashes, which attract potential mates and scare away predators. In FA, flashing features are established and used as functions to address combinatorial optimization problems [60]. Fireflies attract fellow flies and move toward brighter fireflies based on their brightness levels. The greater the separation between fireflies, the less enticing they appear to be. In the absence of a brighter light, fireflies travel randomly. Consequently, the intensity of light and appeal levels have key influences on the FA. A specific function can be used to alter brightness at a specific location. Attraction is determined by other fireflies because it is based on the distance and absorption coefficient.

3. Proposed WBM-DLNets Framework for Brain Tumor Detection

MRI is the most reliable tool used by radiologists for the detection of brain tumors. This study aimed to build an intelligent model that can detect brain tumors and classify them into subcategories (glioma, meningioma, and pituitary gland tumor).
As mentioned in Section 2.2, herein, the input MRIs were preprocessed to eliminate noise and irrelevant information. Next, the pre-trained networks were used to calculate the deep features. The most relevant features were retrieved using MPA-, ASOA-, HHOA-, BOA-, WOA-, GWOA-, BA-, and FA-wrapper-based feature selection techniques. A block diagram of the proposed WBM-DLNets brain tumor detection approach and a flow chart of network selection are shown in Figure 3 and Figure 4, respectively.
In the network selection phase, the proposed approach selects deep learning networks with a metaheuristic optimizer having a maximum accuracy of more than 94% for the deep features of an individual deep learning network. Herein, MATLAB’s “Jx-WFST,” a Jxwrapper feature selection library, was used to implement all methods (https://www.mathworks.com/matlabcentral/fileexchange/84139-wrapper-feature-selection-toolbox, accessed on 7 February 2023). Further, the SVM model was used for brain MRI classification [61,62].

4. Results and Discussion

In this study, the WBM-DLNet feature optimization technique was applied to enhance the classification accuracy of brain tumor detection. An online brain tumor classification dataset was used to test the presented WBM-DLNets feature optimization technique [34]. Discrimination between the MRI images of the subcategories of tumors was accomplished by utilizing the deep features of various pre-trained deep learning networks, as discussed above. The WBM algorithms discussed above (MPA, ASOA, HHOA, BOA, WOA, GWOA, BA, and FA) were used to extract valuable information. Further, MATLAB 2022b, operating on a personal computer with the following specifications, was used for all processing and analysis: Core i7, 12th Generation, 32 GB of RAM, NVIDIA GeForce RTX 3050, 1 TB SSD, and 64-bit Windows 11. The extracted feature subset was classified using an SVM and 0.2-holdout validation technique. The parameters of each algorithm are listed in Table 2.
For each brain MRI image, the deep features of the various pretrained networks were extracted before the SoftMax layer. The initial rate, number of epochs, and momentum were 0.001, 100, and 0.9, respectively. The results of the full deep features of various pretrained networks are presented as classification accuracy and feature vector size in Figure 5.
A thorough analysis of the results presented in Figure 4 reveals that the SVM trained using the deep features of DarkNet-53, DenseNet-201, EfficientNet-b0, ResNet-50, and Xception deep-learning networks yielded classification accuracies above 90% for brain MRI images, and the feature vector sizes of these pretrained networks were 1024, 1920, 1280, 2048, and 2048, respectively. EfficientNet-b0 exhibited a classification rate of 92.33% with a training feature vector size of 1280 for brain tumor detection. The results of the various metaheuristic algorithms used to determine the optimal deep features of the various pretrained networks are presented in Table 3. The detailed results of all networks and optimization algorithms in terms of accuracy, feature vector size, and processing time are shown in Figure S1 in the Supplementary Materials.
As presented in Table 3, all optimization algorithms significantly increased and decreased the accuracy and feature vector size, respectively. However, only two deep learning networks (DenseNet-201 and EfficientNet-b0) yielded accuracies greater than 94%. DenseNet-201 with GWOA exhibited almost 3.31% higher accuracy relative to the full features of DenseNet-201 (1920 features). Additionally, it used three times fewer features to train the SVM model (651 features). In the case of EfficientNet-b0, three optimization algorithms, MPA, ASOA, and WOA, classified the brain MRI images with 94.4%, 94.9%, and 94.4% accuracy, respectively. Based on the network selection criteria discussed in Section 3 and Figure 4, the proposed approach selected only the deep features of the EfficientNet-b0-ASOA. The deep features of DenseNet-201-GWOA and EfficientNet-b0-ASOA were concatenated to train the SVM model using the 0.2-holdout and five-fold cross-validation techniques. The confusion metrics of the WBM-DLNets are shown in Figure 6. The results were thoroughly analyzed using the true positive rate (TPR), false negative rate (FNR), positive predictive value (PPV), and false discovery rate (FDR) of the developed machine learning model, as presented in Table 4. Equation (2) can be used to compute TPR, FNR, PPV, FDR, and accuracy.
TPR = True   positive No .   of   real   positive FNR = False   negative No .   of   real   positive PPV = True   positive True   positive + False   positive FDR = False   positive True   positive + False   positive Accuracy   ( % ) = No .   of   correctly   classified   images Total   no .   of   images × 100 }
In the case of holdout validation, 159 of the 165 MRI images of glioma tumors were correctly classified and had a TPR of 96.4% (Figure 6a and Table 4). The no-tumor class contained 80 MRI images to test the model. A total of 77 brain MRI images with a PPV of 97.5% were classified correctly. The proposed WBM-DLNets achieved a high classification accuracy of 95.6% with only 1292 feature vector sizes for holdout validation. In the five-fold cross-validation, the proposed model slightly increased the classification accuracy, as presented in Figure 6b and Table 4. The model correctly predicted 811 out of 824 pituitary gland tumor class MRI images with a TPR of 98.1%.
Another brain MRI dataset comprising 233 patients was used to validate the adaptability and accuracy [63]. A total of 3064 brain MRI scans were obtained at two hospitals in China (Nanfang Hospital and General Hospital), of which 1426 images comprised glioma tumors, 708 comprised meningioma tumors, and 930 comprised pituitary gland tumors [64]. The results of the proposed approach for the new dataset are presented in Figure 7 and Table 5.
The results presented in Figure 7 and Table 5 demonstrate the adaptability and high classification performance of the proposed approach. As presented in Table 6, the results of the proposed WBM-DLNets approach were also compared with those reported in the latest literature.
Recently, the use of machine learning techniques in medical imaging for diagnostic applications has increased. Several researchers have investigated various learning techniques for detecting brain tumors using MRI images [28,29,32,65].
Deep learning networks have a very high level of accuracy in classifying brain MRI images [28,29]. Irmak [29] proposed a brain tumor detection CNN model that yielded an accuracy of 92.66% for subclassification. Additionally, a pretrained deep learning network has been utilized to detect brain tumors [65]; however, the model training it is too time-consuming. Another study [32] concatenated the deep features of various pre-trained models and achieved 93.72% accuracy. The feature vector was excessively large and contained various redundant features. Filter-based feature reduction approaches have also been reported to enhance classification performance [18]. Filtering strategies select features based on their applicability to the dependent variables. However, they do not consider how they affect the model performance. By contrast, wrapper approaches test the effectiveness of a feature subset by training the model with them. Occasionally, a threshold value must be selected to remove unnecessary data in filter-based methods. Wrapper-based strategies have been shown to be more accurate in classifying data. Therefore, in this study, a WBM-DLNet feature optimization approach was proposed to improve the classification performance of brain tumor detection. The WBM-DLNet feature optimization approach and previously published studies are compared in Table 6. The findings revealed that the proposed WBM-DLNet feature optimization technique outperforms competing methods in terms of classification rate. The proposed brain tumor detection approach may help physicians identify tumors quickly and effectively.

5. Conclusions

In this study, WBM-DLNet feature optimization algorithms were utilized to enhance brain tumor detection classification performance. The deep features of the 16 pretrained deep learning networks were computed. Eight metaheuristic optimization algorithms (MPA, ASOA, HHOA, BOA, WOA, GWOA, BA, and FA) were applied to determine the optimal deep features of all networks using the SVM-based cost function. All metaheuristic optimization algorithms significantly enhanced the classification performance and reduced the feature vector size of each pretrained model. Subsequently, a deep-learning network selection approach was applied to determine the best deep features. The best deep features were concatenated to train the SVM model. The model exhibited the highest classification rate of 95.7% with DenseNet-201-GWOA and EfficientNet-b0-ASOA deep feature-trained SVM models. The model was further validated using a new dataset and exhibited a high classification performance of 96.7%. Therefore, the proposed WBM-DLNet feature optimization algorithm can be considered useful for automatic brain tumor detection.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/bioengineering10040475/s1, Figure S1: Detailed results of all the networks and the optimization algorithms in terms of accuracy, feature vector size, and processing time: (a) MPA; (b) ASOA; (c) HHOA; (d) BOA; (e) WOA; (f) GWOA; (g) BA; (h) FA.

Author Contributions

Conceptualization, M.U.A. and A.Z.; formal analysis, A.Z.; funding acquisition, M.R.B. and S.W.L.; investigation, A.Z.; methodology, M.U.A.; resources, M.R.B.; software, S.J.H.; supervision, S.W.L.; validation, S.J.H.; writing—original draft, M.U.A. and A.Z.; writing—review and editing, S.J.H., M.R.B. and S.W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation (NRF) of Korea through the auspices of the Ministry of Science and ICT, Republic of Korea, under Grant NRF-2020R1G1A1012741 (M.R.B.). This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean Government (MOE) (NRF2021R1I1A2059735) (S.W.L).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are included in the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Louis, D.N.; Perry, A.; Reifenberger, G.; von Deimling, A.; Figarella-Branger, D.; Cavenee, W.K.; Ohgaki, H.; Wiestler, O.D.; Kleihues, P.; Ellison, D.W. The 2016 World Health Organization Classification of Tumors of the Central Nervous System: A summary. Acta Neuropathol. 2016, 131, 803–820. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Haj-Hosseini, N.; Milos, P.; Hildesjö, C.; Hallbeck, M.; Richter, J.; Wårdell, K. Fluorescence spectroscopy and optical coherence tomography for brain tumor detection. In Proceedings of the SPIE Photonics Europe, Biophotonics: Photonic Solutions for Better Health Care, Brussels, Belgium, 3–7 April 2016; pp. 9887–9896. [Google Scholar]
  3. Ren, W.; Hasanzade Bashkandi, A.; Afshar Jahanshahi, J.; Qasim Mohammad AlHamad, A.; Javaheri, D.; Mohammadi, M. Brain tumor diagnosis using a step-by-step methodology based on courtship learning-based water strider algorithm. Biomed. Signal Process. Control 2023, 83, 104614. [Google Scholar] [CrossRef]
  4. Brain Tumor Facts. Available online: https://braintumor.org/brain-tumors/about-brain-tumors/brain-tumor-facts/#:~:text=Today%2C%20an%20estimated%20700%2C000%20people,will%20be%20diagnosed%20in%202022 (accessed on 6 March 2023).
  5. American Cancer Society. Available online: www.cancer.org/cancer.html (accessed on 9 September 2021).
  6. American Society of Clinical Oncology. Available online: https://www.cancer.net/cancer-types/brain-tumor/diagnosis (accessed on 9 September 2021).
  7. Wulandari, A.; Sigit, R.; Bachtiar, M.M. Brain tumor segmentation to calculate percentage tumor using MRI. In Proceedings of the 2018 International Electronics Symposium on Knowledge Creation and Intelligent Computing (IES-KCIC), Surabaya, Indonesia, 29–30 October 2018; pp. 292–296. [Google Scholar]
  8. Xu, Z.; Sheykhahmad, F.R.; Ghadimi, N.; Razmjooy, N. Computer-aided diagnosis of skin cancer based on soft computing techniques. Open Med. 2020, 15, 860–871. [Google Scholar] [CrossRef]
  9. Kebede, S.R.; Debelee, T.G.; Schwenker, F.; Yohannes, D. Classifier Based Breast Cancer Segmentation. J. Biomim. Biomater. Biomed. Eng. 2020, 47, 41–61. [Google Scholar] [CrossRef]
  10. Debelee, T.G.; Amirian, M.; Ibenthal, A.; Palm, G.; Schwenker, F. Classification of Mammograms Using Convolutional Neural Network Based Feature Extraction. In Proceedings of the Information and Communication Technology for Development for Africa, Cham, Switzerland, 10–12 December 2018; pp. 89–98. [Google Scholar]
  11. Gab Allah, A.M.; Sarhan, A.M.; Elshennawy, N.M. Classification of Brain MRI Tumor Images Based on Deep Learning PGGAN Augmentation. Diagnostics 2021, 11, 2343. [Google Scholar] [CrossRef] [PubMed]
  12. Alanazi, M.F.; Ali, M.U.; Hussain, S.J.; Zafar, A.; Mohatram, M.; Irfan, M.; AlRuwaili, R.; Alruwaili, M.; Ali, N.H.; Albarrak, A.M. Brain Tumor/Mass Classification Framework Using Magnetic-Resonance-Imaging-Based Isolated and Developed Transfer Deep-Learning Model. Sensors 2022, 22, 372. [Google Scholar] [CrossRef]
  13. Almalki, Y.E.; Ali, M.U.; Kallu, K.D.; Masud, M.; Zafar, A.; Alduraibi, S.K.; Irfan, M.; Basha, M.A.A.; Alshamrani, H.A.; Alduraibi, A.K.; et al. Isolated Convolutional-Neural-Network-Based Deep-Feature Extraction for Brain Tumor Classification Using Shallow Classifier. Diagnostics 2022, 12, 1793. [Google Scholar] [CrossRef]
  14. Debelee, T.G.; Kebede, S.R.; Schwenker, F.; Shewarega, Z.M. Deep Learning in Selected Cancers’ Image Analysis—A Survey. J. Imaging 2020, 6, 121. [Google Scholar] [CrossRef] [PubMed]
  15. Pandian, R.; Vedanarayanan, V.; Ravi Kumar, D.N.S.; Rajakumar, R. Detection and classification of lung cancer using CNN and Google net. Meas. Sens. 2022, 24, 100588. [Google Scholar] [CrossRef]
  16. Allah AM, G.; Sarhan, A.M.; Elshennawy, N.M. Edge U-Net: Brain tumor segmentation using MRI based on deep U-Net model with boundary information. Expert Syst. Appl. 2023, 213, 118833. [Google Scholar] [CrossRef]
  17. Ma, C.; Luo, G.; Wang, K. Concatenated and Connected Random Forests with Multiscale Patch Driven Active Contour Model for Automated Brain Tumor Segmentation of MR Images. IEEE Trans. Med. Imaging 2018, 37, 1943–1954. [Google Scholar] [CrossRef]
  18. Almalki, Y.E.; Ali, M.U.; Ahmed, W.; Kallu, K.D.; Zafar, A.; Alduraibi, S.K.; Irfan, M.; Basha, M.A.A.; Alshamrani, H.A.; Alduraibi, A.K. Robust Gaussian and Nonlinear Hybrid Invariant Clustered Features Aided Approach for Speeded Brain Tumor Diagnosis. Life 2022, 12, 1084. [Google Scholar] [CrossRef] [PubMed]
  19. Ali, M.U.; Kallu, K.D.; Masood, H.; Hussain, S.J.; Ullah, S.; Byun, J.H.; Zafar, A.; Kim, K.S. A Robust Computer-Aided Automated Brain Tumor Diagnosis Approach Using PSO-ReliefF Optimized Gaussian and Non-Linear Feature Space. Life 2022, 12, 2036. [Google Scholar] [CrossRef] [PubMed]
  20. Kumari, R. SVM classification an approach on detecting abnormality in brain MRI images. Int. J. Eng. Res. Appl. 2013, 3, 1686–1690. [Google Scholar]
  21. Ayachi, R.; Ben Amor, N. Brain tumor segmentation using support vector machines. In Proceedings of the Symbolic and Quantitative Approaches to Reasoning with Uncertainty: 10th European Conference, ECSQARU 2009, Verona, Italy, 1–3 July 2009; Proceedings 10, pp. 736–747. [Google Scholar]
  22. Almahfud, M.A.; Setyawan, R.; Sari, C.A.; Setiadi, D.R.I.M.; Rachmawanto, E.H. An Effective MRI Brain Image Segmentation using Joint Clustering (K-Means and Fuzzy C-Means). In Proceedings of the 2018 International Seminar on Research of Information Technology and Intelligent Systems (ISRITI), Yogyakarta, Indonesia, 21–22 November 2018; pp. 11–16. [Google Scholar]
  23. Abdel-Maksoud, E.; Elmogy, M.; Al-Awadi, R. Brain tumor segmentation based on a hybrid clustering technique. Egypt. Inform. J. 2015, 16, 71–81. [Google Scholar] [CrossRef]
  24. Kaya, I.E.; Pehlivanlı, A.Ç.; Sekizkardeş, E.G.; Ibrikci, T. PCA based clustering for brain tumor segmentation of T1w MRI images. Comput. Methods Programs Biomed. 2017, 140, 19–28. [Google Scholar] [CrossRef]
  25. Nazir, M.; Shakil, S.; Khurshid, K. Role of deep learning in brain tumor detection and classification (2015 to 2020): A review. Comput. Med. Imaging Graph. 2021, 91, 101940. [Google Scholar] [CrossRef] [PubMed]
  26. Pereira, S.; Meier, R.; Alves, V.; Reyes, M.; Silva, C.A. Automatic Brain Tumor Grading from MRI Data Using Convolutional Neural Networks and Quality Assessment; Springer: Cham, Switzerland, 2018; pp. 106–114. [Google Scholar]
  27. Abiwinanda, N.; Hanif, M.; Hesaputra, S.T.; Handayani, A.; Mengko, T.R. Brain Tumor Classification Using Convolutional Neural Network; Springer: Singapore, 2019; pp. 183–189. [Google Scholar]
  28. Badža, M.M.; Barjaktarović, M.Č. Classification of Brain Tumors from MRI Images Using a Convolutional Neural Network. Appl. Sci. 2020, 10, 1999. [Google Scholar] [CrossRef] [Green Version]
  29. Irmak, E. Multi-Classification of Brain Tumor MRI Images Using Deep Convolutional Neural Network with Fully Optimized Framework. Iran. J. Sci. Technol. Trans. Electr. Eng. 2021, 45, 1015–1036. [Google Scholar] [CrossRef]
  30. Deepak, S.; Ameer, P.M. Brain tumor classification using deep CNN features via transfer learning. Comput. Biol. Med. 2019, 111, 103345. [Google Scholar] [CrossRef]
  31. Çinar, A.; Yildirim, M. Detection of tumors on brain MRI images using the hybrid convolutional neural network architecture. Med. Hypotheses 2020, 139, 109684. [Google Scholar] [CrossRef]
  32. Kang, J.; Ullah, Z.; Gwak, J. MRI-Based Brain Tumor Classification Using Ensemble of Deep Features and Machine Learning Classifiers. Sensors 2021, 21, 2222. [Google Scholar] [CrossRef] [PubMed]
  33. Dokeroglu, T.; Deniz, A.; Kiziloz, H.E. A comprehensive survey on recent metaheuristics for feature selection. Neurocomputing 2022, 494, 269–296. [Google Scholar] [CrossRef]
  34. Chakrabarty, N.; Kanchan, S. Brain Tumor Classification (MRI). Available online: https://www.kaggle.com/datasets/sartajbhuvaji/brain-tumor-classification-mri?select=Training (accessed on 17 March 2022).
  35. Rosebrock, A. Finding extreme points in contours with Open CV. Available online: https://www.pyimagesearch.com/2016/04/11/finding-extreme-points-in-contours-with-opencv/ (accessed on 9 September 2021).
  36. Dash, M.; Liu, H. Feature selection for classification. Intell. Data Anal. 1997, 1, 131–156. [Google Scholar] [CrossRef]
  37. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  38. Kohavi, R.; John, G.H. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef] [Green Version]
  39. Abdel-Basset, M.; Abdel-Fatah, L.; Sangaiah, A.K. Metaheuristic algorithms: A comprehensive review. Comput. Intell. Multimed. Big Data Cloud Eng. Appl. 2018, 185–231. [Google Scholar] [CrossRef]
  40. Yang, X.-S. Chapter 1—Introduction to Algorithms. In Nature-Inspired Optimization Algorithms, 2nd ed.; Yang, X.-S., Ed.; Academic Press: Cambridge, MA, USA, 2021; pp. 1–22. [Google Scholar]
  41. Liu, W.; Wang, J. A Brief Survey on Nature-Inspired Metaheuristics for Feature Selection in Classification in this Decade. In Proceedings of the 2019 IEEE 16th International Conference on Networking, Sensing and Control (ICNSC), Banff, AB, Canada, 9–11 May 2019; pp. 424–429. [Google Scholar]
  42. Agrawal, P.; Abutarboush, H.F.; Ganesh, T.; Mohamed, A.W. Metaheuristic Algorithms on Feature Selection: A Survey of One Decade of Research (2009–2019). IEEE Access 2021, 9, 26766–26791. [Google Scholar] [CrossRef]
  43. Faramarzi, A.; Heidarinejad, M.; Mirjalili, S.; Gandomi, A.H. Marine Predators Algorithm: A nature-inspired metaheuristic. Expert Syst. Appl. 2020, 152, 113377. [Google Scholar] [CrossRef]
  44. Rai, R.; Dhal, K.G.; Das, A.; Ray, S. An Inclusive Survey on Marine Predators Algorithm: Variants and Applications. Arch. Comput. Methods Eng. 2023. [Google Scholar] [CrossRef]
  45. Ewees, A.A.; Ismail, F.H.; Ghoniem, R.M.; Gaheen, M.A. Enhanced Marine Predators Algorithm for Solving Global Optimization and Feature Selection Problems. Mathematics 2022, 10, 4154. [Google Scholar] [CrossRef]
  46. Zhao, W.; Wang, L.; Zhang, Z. Atom search optimization and its application to solve a hydrogeologic parameter estimation problem. Knowl.-Based Syst. 2019, 163, 283–304. [Google Scholar] [CrossRef]
  47. Kamel, S.; Hamour, H.; Ahmed, M.H.; Nasrat, L. Atom Search optimization Algorithm for Optimal Radial Distribution System Reconfiguration. In Proceedings of the 2019 International Conference on Computer, Control, Electrical, and Electronics Engineering (ICCCEEE), Khartoum, Sudan, 21–23 September 2019; pp. 1–5. [Google Scholar]
  48. Bairathi, D.; Gopalani, D. A novel swarm intelligence based optimization method: Harris' hawk optimization. In Proceedings of the Intelligent Systems Design and Applications: 18th International Conference on Intelligent Systems Design and Applications (ISDA 2018), Vellore, India, 6–8 December 2018; Volume 2, pp. 832–842. [Google Scholar]
  49. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  50. Arora, S.; Singh, S. Butterfly optimization algorithm: A novel approach for global optimization. Soft Comput. 2019, 23, 715–734. [Google Scholar] [CrossRef]
  51. Zhou, H.; Cheng, H.-Y.; Wei, Z.-L.; Zhao, X.; Tang, A.-D.; Xie, L. A Hybrid Butterfly Optimization Algorithm for Numerical Optimization Problems. Comput. Intell. Neurosci. 2021, 2021, 7981670. [Google Scholar] [CrossRef]
  52. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  53. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  54. Emary, E.; Zawbaa, H.M.; Hassanien, A.E. Binary grey wolf optimization approaches for feature selection. Neurocomputing 2016, 172, 371–381. [Google Scholar] [CrossRef]
  55. Tu, Q.; Chen, X.; Liu, X. Multi-strategy ensemble grey wolf optimizer and its application to feature selection. Appl. Soft Comput. 2019, 76, 16–30. [Google Scholar] [CrossRef]
  56. Yang, X.-S. A new metaheuristic bat-inspired algorithm. In Nature Inspired Cooperative Strategies for Optimization (NICSO 2010); Springer: Berlin/Heidelberg, Germany, 2010; pp. 65–74. [Google Scholar]
  57. Yang, X.-S. Chapter 11—Bat Algorithms. In Nature-Inspired Optimization Algorithms, 2nd ed.; Yang, X.-S., Ed.; Academic Press: Cambridge, MA, USA, 2021; pp. 157–173. [Google Scholar]
  58. Yang, X.-S.; He, X. Bat algorithm: Literature review and applications. Int. J. Bio-Inspired Comput. 2013, 5, 141–149. [Google Scholar] [CrossRef] [Green Version]
  59. Yang, X.-S. Firefly algorithm, stochastic test functions and design optimisation. Int. J. Bio-Inspired Comput. 2010, 2, 78–84. [Google Scholar] [CrossRef]
  60. Fister, I.; Fister Jr, I.; Yang, X.-S.; Brest, J. A comprehensive review of firefly algorithms. Swarm Evol. Comput. 2013, 13, 34–46. [Google Scholar] [CrossRef] [Green Version]
  61. Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
  62. Smola, A.J.; Schölkopf, B. A tutorial on support vector regression. Stat. Comput. 2004, 14, 199–222. [Google Scholar] [CrossRef] [Green Version]
  63. Jun, C. Brain Tumor Dataset. 2017. Available online: https://figshare.com/articles/dataset/brain_tumor_dataset/1512427 (accessed on 9 September 2021).
  64. Cheng, J.; Huang, W.; Cao, S.; Yang, R.; Yang, W.; Yun, Z.; Wang, Z.; Feng, Q. Enhanced Performance of Brain Tumor Classification via Tumor Region Augmentation and Partition. PLoS ONE 2015, 10, e0140381. [Google Scholar] [CrossRef]
  65. Rehman, A.; Naz, S.; Razzak, M.I.; Akram, F.; Imran, M. A Deep Learning-Based Framework for Automatic Brain Tumors Classification Using Transfer Learning. Circuits Syst. Signal Process. 2020, 39, 757–775. [Google Scholar] [CrossRef]
Figure 1. Deep feature extraction using GoogLeNet.
Figure 1. Deep feature extraction using GoogLeNet.
Bioengineering 10 00475 g001
Figure 2. Working flow for the wrapper-based approach.
Figure 2. Working flow for the wrapper-based approach.
Bioengineering 10 00475 g002
Figure 3. Block diagram of the proposed WBM-DLNets.
Figure 3. Block diagram of the proposed WBM-DLNets.
Bioengineering 10 00475 g003
Figure 4. Network selection framework.
Figure 4. Network selection framework.
Bioengineering 10 00475 g004
Figure 5. Performance of the SVM classifier trained with full deep features of pre-trained models.
Figure 5. Performance of the SVM classifier trained with full deep features of pre-trained models.
Bioengineering 10 00475 g005
Figure 6. Confusion matrix of the proposed WBM-DLNets for brain tumor detection: (a) 0.2-holdout; (b) five-fold cross-validation.
Figure 6. Confusion matrix of the proposed WBM-DLNets for brain tumor detection: (a) 0.2-holdout; (b) five-fold cross-validation.
Bioengineering 10 00475 g006
Figure 7. Confusion matrix of the proposed WBM-DLNets for brain tumor detection for another dataset [63]: (a) 0.2-holdout; (b) five-fold cross-validation.
Figure 7. Confusion matrix of the proposed WBM-DLNets for brain tumor detection for another dataset [63]: (a) 0.2-holdout; (b) five-fold cross-validation.
Bioengineering 10 00475 g007
Table 1. Details of online brain MRI dataset.
Table 1. Details of online brain MRI dataset.
Glioma TumorMeningioma TumorNo TumorPituitary Tumor
Brain MRI
images
Bioengineering 10 00475 i001Bioengineering 10 00475 i002Bioengineering 10 00475 i003Bioengineering 10 00475 i004
No. of images per class826822395827
Table 2. Parameters for each algorithm.
Table 2. Parameters for each algorithm.
MPAASOAHHOABOAWOAGWOABAFA
Number of iterations = 50Number of iterations = 50Number of iterations = 50Number of iterations = 50Number of iterations = 50Number of iterations = 50Number of iterations = 50Number of iterations = 50
Population size = 10Population size = 10Population size = 10Population size = 10Population size = 10Population size = 10Population size = 10Population size = 10
Fish aggregating devices effect = 0.2Depth weight = 50Levy component = 1.5Modular modality = 0.01Constant = 1 Maximum frequency = 2Absorption coefficient = 1
Constant = 0.5Multiplier weight = 0.2 Switch probability = 0.8 Minimum frequency = 0Constant = 1
Levy component = 1.5 Constant = 0.9 Light amplitude =1
Maximum loudness = 2Control alpha = 0.97
Maximum pulse rate = 1
Table 3. Classification performance of each pre-trained network with metaheuristic algorithms for brain tumor detection.
Table 3. Classification performance of each pre-trained network with metaheuristic algorithms for brain tumor detection.
NetworkMPAASOAHHOABOAWOAGWOABAFA
AccuracyFeature Vector SizeAccuracyFeature Vector SizeAccuracyFeature Vector SizeAccuracyFeature Vector SizeAccuracyFeature Vector SizeAccuracyFeature Vector SizeAccuracyFeature Vector SizeAccuracyFeature Vector Size
DarkNet-190.9062670.8955160.8941720.8804130.8953040.9093030.8944940.892447
DarkNet-530.9307320.9225040.9235750.9134810.9187120.9343620.9134910.920521
DenseNet-2010.9309390.9169790.92212700.8949120.9099700.9466510.9089640.918950
EfficientNet-b00.9448210.9496410.9398620.9255510.9446540.9394510.9346690.936621
GoogLeNet3650.8943880.9014990.8854870.8784720.8825420.8994010.8875220.887479
GoogLeNet0.8896670.8735310.8835010.8525210.8736900.8953380.8645180.871493
Inception-ResNet-v20.9155330.9097530.9087390.9098020.9027820.9254880.9047540.908763
Inception-v30.9048780.9119970.90810230.8957640.89511740.9237680.9019900.895966
MobileNet-v20.9138120.9026470.8907700.8836230.8877380.8945120.8926590.897633
NASNet-Mobile0.8855750.8695050.8717140.8644610.8738540.8873640.8664970.871549
ResNet-1010.92512790.92710570.91512310.9239270.91515360.9278260.9139850.9111005
ResNet-500.93412280.93710130.93912540.9168770.93210680.9396920.92710170.9341016
ResNet-180.8782740.8762420.8753150.8542140.8644280.8872100.8802420.873233
ShuffleNet0.9162170.9042740.9113740.8872710.8953210.9182180.8942860.901256
SqueezeNet0.9045190.9134990.9026190.8855160.8948520.9013850.8894850.890483
Xception 0.9208090.92910350.91611490.9088890.9119780.9307750.91610020.9151009
Table 4. Results of WBM-DLNets for tumor detection.
Table 4. Results of WBM-DLNets for tumor detection.
ValidationClassTPR
(%)
FNR
(%)
PPV
(%)
FDR
(%)
Accuracy
(%)
0.2-holdoutglioma_tumor96.43.697.03.095.6
meningioma_tumor90.99.294.95.1
no_tumor96.33.897.52.5
pituitary_tumor99.40.694.35.7
Five-fold
cross-validation
glioma_tumor95.24.897.82.295.7
meningioma_tumor94.35.792.47.6
no_tumor95.24.895.44.6
pituitary_tumor98.11.997.42.6
Table 5. Results of WBM-DLNets for tumor detection for another dataset [63].
Table 5. Results of WBM-DLNets for tumor detection for another dataset [63].
ValidationClassTPR
(%)
FNR
(%)
PPV
(%)
FDR
(%)
Accuracy
(%)
0.2-holdoutglioma_tumor95.84.298.91.196.7
meningioma_tumor95.74.391.28.8
pituitary_tumor98.91.197.92.1
Five-fold
cross-validation
glioma_tumor96.93.197.72.396.6
meningioma_tumor93.16.992.67.4
pituitary_tumor98.71.398.02.0
Table 6. Comparison of the proposed WBM-DLNets with other studies.
Table 6. Comparison of the proposed WBM-DLNets with other studies.
ReferenceAccuracy (%)
Almalki et al. [18]95.33
Abiwinanda et al. [27]84.19
Irmak [29]92.66
Kang et al. [32]93.72
Rehman et al. [65]95.86
WBM-DLNets (Proposed)95.7 and 96.7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ali, M.U.; Hussain, S.J.; Zafar, A.; Bhutta, M.R.; Lee, S.W. WBM-DLNets: Wrapper-Based Metaheuristic Deep Learning Networks Feature Optimization for Enhancing Brain Tumor Detection. Bioengineering 2023, 10, 475. https://doi.org/10.3390/bioengineering10040475

AMA Style

Ali MU, Hussain SJ, Zafar A, Bhutta MR, Lee SW. WBM-DLNets: Wrapper-Based Metaheuristic Deep Learning Networks Feature Optimization for Enhancing Brain Tumor Detection. Bioengineering. 2023; 10(4):475. https://doi.org/10.3390/bioengineering10040475

Chicago/Turabian Style

Ali, Muhammad Umair, Shaik Javeed Hussain, Amad Zafar, Muhammad Raheel Bhutta, and Seung Won Lee. 2023. "WBM-DLNets: Wrapper-Based Metaheuristic Deep Learning Networks Feature Optimization for Enhancing Brain Tumor Detection" Bioengineering 10, no. 4: 475. https://doi.org/10.3390/bioengineering10040475

APA Style

Ali, M. U., Hussain, S. J., Zafar, A., Bhutta, M. R., & Lee, S. W. (2023). WBM-DLNets: Wrapper-Based Metaheuristic Deep Learning Networks Feature Optimization for Enhancing Brain Tumor Detection. Bioengineering, 10(4), 475. https://doi.org/10.3390/bioengineering10040475

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop