You are currently viewing a new version of our website. To view the old version click .
Bioengineering
  • Article
  • Open Access

22 December 2022

Brain Tumor Detection and Classification Using Deep Learning and Sine-Cosine Fitness Grey Wolf Optimization

,
,
,
,
,
and
1
Computer Engineering and Control Systems Department, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt
2
Department of Communications and Electronics, Delta Higher Institute of Engineering and Technology, Mansoura 35111, Egypt
3
Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
4
Machine Learning & Information Retrieval Department, Faculty of Artificial Intelligence, Kafrelsheikh University, Kafrelsheikh 33511, Egypt
This article belongs to the Section Biochemical Engineering

Abstract

Diagnosing a brain tumor takes a long time and relies heavily on the radiologist’s abilities and experience. The amount of data that must be handled has increased dramatically as the number of patients has increased, making old procedures both costly and ineffective. Many researchers investigated a variety of algorithms for detecting and classifying brain tumors that were both accurate and fast. Deep Learning (DL) approaches have recently been popular in developing automated systems capable of accurately diagnosing or segmenting brain tumors in less time. DL enables a pre-trained Convolutional Neural Network (CNN) model for medical images, specifically for classifying brain cancers. The proposed Brain Tumor Classification Model based on CNN (BCM-CNN) is a CNN hyperparameters optimization using an adaptive dynamic sine-cosine fitness grey wolf optimizer (ADSCFGWO) algorithm. There is an optimization of hyperparameters followed by a training model built with Inception-ResnetV2. The model employs commonly used pre-trained models (Inception-ResnetV2) to improve brain tumor diagnosis, and its output is a binary 0 or 1 (0: Normal, 1: Tumor). There are primarily two types of hyperparameters: (i) hyperparameters that determine the underlying network structure; (ii) a hyperparameter that is responsible for training the network. The ADSCFGWO algorithm draws from both the sine cosine and grey wolf algorithms in an adaptable framework that uses both algorithms’ strengths. The experimental results show that the BCM-CNN as a classifier achieved the best results due to the enhancement of the CNN’s performance by the CNN optimization’s hyperparameters. The BCM-CNN has achieved 99.98% accuracy with the BRaTS 2021 Task 1 dataset.

1. Introduction

Recently, digital medical images have been essential for detecting numerous illnesses. It is additionally used for training and research. The need for electronic medical images is growing dramatically; for example, in 2002, the Department of Radiology at the University Hospital of Geneva produced between 12,000 and 15,000 images daily [1]. An efficient and exact computer-aided diagnostic system is required for medical report creation and medical image research. The old method of manually evaluating medical imaging is time-consuming, inaccurate, and prone to human error. Over the medical diseases, the brain tumor has become a serious issue, ranking 10th among the major causes of death in the US. It is reported that 700,000 persons have brain tumors, of which 80 percent are benign and 20 percent are malignant [2]. According to estimates by the American Cancer Society from 2021, 78,980 adults have been diagnosed with a brain tumor, with 55,150 noncancerous and 24,530 malignant tumors (13,840 men and 10,690 females) [3]. According to studies, brain tumor is the top cause of cancer deaths in children and adults worldwide [4].
The most typical kind of brain disease is a brain tumor. It is an unregulated development of brain cells. Brain tumors are always classified into brain tumors, both primary and secondary. The first starts in the brain and usually stays there, whereas the latter starts as cancer somewhere else in the body and spreads to the brain [5]. There are two different forms of tumors: malignant and benign. A benign tumor is a slow-growing tumor that does not infiltrate nearby tissues; on the other hand, a malignant which is a very aggressive tumor that spreads from one location to another. The World Health Organization (WHO) grades a brain tumor as I-IV. Tumors in categories I and II are regarded as slow-growing, while tumors in categories III and IV are always malignant and have a worse prognosis [6].
In recent decades, many imaging techniques such as X-ray, Magneto Encephalo Graphy (MEG), Computed Tomography (CT), Ultrasonography, Electro Encephalo Graphy (EEG), Single-Photon Emission Computed Tomography (SPECT), Positron Emission Tomography (PET), and Magnetic Resonance Imaging (MRI) have emerged that not only exhibit the detailed and complete facets of brain tumors but also help doctors to accurately diagnose the tumor and determine the correct treatment mechanism [4]. MRI is considered the most popular imaging technique for detecting brain tumors [7]. Without subjecting the patients to excessive ionization radiation, MRI is a non and excellent soft tissue contrast imaging technique that gives essential information about brain tumor shape, location, and size.
The brain tumor diagnosis is highly time intensive and largely depends on the radiologist’s skills and knowledge. Because there are more patients, the amount of data that must be processed has grown significantly, making traditional techniques cost and incorrect [8]. The difficulties are associated with significant brain tumor size, shape, and intensity variations for the same tumor type and similar manifestations of other disease types. A misclassification of a brain tumor can result in major consequences and reduce the patient’s survivability. There is a rise in interest in building automated technologies for processing images to overcome the limitations of manual diagnosis [4,9] and other related applications [10,11,12]. Several systems for computer-aided diagnosis (CAD) have been created recently to diagnose brain tumors automatically.
In recent years, among many other applications, artificial intelligence (AI) has demonstrated promising results as a decision support system to assist in the detection of diseases and the establishment of precise medical diagnoses. In order to address practical problems researchers and governments focus on machine learning, a branch of artificial intelligence [13,14]. Machine learning, for instance, may predict the COVID-19 outbreak in the COVID-19 pandemic challenge by determining how risky the virus is and then scaling up the level of the methods performed. In the realm of medical analysis, machine learning algorithms are frequently used for things such as COVID-19 prediction [15], Alzheimer’s disease progression [16], brain tumor development [17], breast cancer progression [18], and other disorders [19,20,21]. Deep learning and machine learning are essential for identifying diseases and resolving medical problems.
Many researchers investigated numerous algorithms for detecting and classifying brain tumors with high performance and less error. Deep Learning (DL) techniques have recently been widely employed to build automatic systems that can accurately classify or segment brain tumors in less time. DL enables the use of a pre-trained Convolutional Neural Network (CNN) model [22] for medical imagery, specifically for the classification of brain tumors, which has been created for various applications, including GoogLeNet [23], AlexNet, and ResNet-34 [24]. DL is made up of a multi-layered deep neural network [25]. The backpropagation algorithm is used by a neural network (NN) to reduce the error between the target and actual value. Nevertheless, even as the number of layers increases, developing artificial neural network models gets more difficult.
The main contributions of the current work are:
  • Introduce an enhanced model to improve brain tumor diagnosis.
  • It proposes a Brain Tumor Classification Model (BCM-CNN) based on an advanced 3D model using Enhanced Convolutional Neural Network (BCM-CNN).
  • The proposed Brain Tumor Classification Model (BCM-CNN) is based on two submodules; (i) CNN hyperparameters optimization using an adaptive dynamic sine-cosine fitness grey wolf optimizer (ADSCFGWO) algorithm followed by trained Model, and (ii) segmentation model.
  • The ADSCFGWO algorithm draws from both the sine cosine and grey wolf algorithms in an adaptable framework that uses both algorithms’ strengths.
  • The experimental results show that the BCM-CNN as a classifier achieved the best results due to the enhancement of the CNN’s performance by the CNN optimization’s hyperparameters.
The remainder of the paper is structured as follows: A brief review of state-of-the-art deep learning methods for finding brain tumors is discussed in Section 2. The proposed technique is described in detail in Section 3. Section 4 depicts simulation and experimental results. The conclusion and future works are presented in Section 5.

3. Brain Tumor Classification Model Based CNN (BCM-CNN)

This section proposes a Brain Tumor Classification Model (BCM-CNN) based on an advanced model using a Convolutional Neural Network. The overall architecture of the proposed model is shown in Figure 1. The BCM-CNN is used to diagnose a brain tumor. It consists of a hyperparameters optimization, followed by an Inception-ResnetV2 training model. The model’s output is a binary 0 or 1 (0: Normal, 1: Tumor) and uses common pre-trained models (Inception-ResnetV2) to enhance the brain tumor diagnosis process.
Figure 1. The proposed BCM-CNN model steps.

3.1. CNN Hyperparameters Optimization

This subsection discusses the selected hyperparameters. The configuration’s hyperparameters are variables that are not part of the model and whose values cannot be inferred from the data. Two main categories of hyperparameters exist: (i) a network structure-determining hyperparameter; (ii) the network is trained by the hyperparameter. Table 2 contains a list of the hyperparameters that were examined in this study.
Table 2. CNN hyperparameters setup.

3.2. ADSCFGWO for CNN Hyperparameters

To select the most important characteristics from the metamaterial dataset in order to achieve the best possible performance, the adaptive dynamic sine cosine fitness grey wolf optimizer, abbreviated as ADSCFGWO, was initially introduced in [41]. This algorithm draws from both the sine cosine and grey wolf algorithms in an adaptable framework that makes use of both algorithms’ strengths. To estimate the double T-shape monopole antenna properties, the ADSCFGWO algorithm additionally optimizes a bidirectional recurrent neural network (BRNN). In this work, the optimization of the CNN hyperparameters is based on the adaptive dynamic sine cosine fitness grey wolf optimizer (ADSCFGWO) algorithm. The ADSCFGWO algorithm is shown in Algorithm 1.
The population in the potential solution of the ADSCFGWO algorithm, X i ( i = 1 , 2 , , n ) with size n, is split into two groups: the exploration group, n 1 , and exploitation group, n 2 . The exploration group’s job is to use the search space to discover new locations where the greatest possible solution might be located. The exploitation group’s job is to use an objective function to enhance the best solution’s quality. These two groups cooperate in the suggested optimization process to trade responsibilities and required data that can hasten the retrieval of the optimum solution. The effective avoidance of the local optima and the precise exploration of the search space are advantages of this collaboration. The ADSCFGWO optimization technique has two key characteristics: first, it maintains correct control over the equilibrium between the exploitation and exploration groups; and second, it uses a dynamic mechanism to avoid steady regions in the search space.
The fittest solutions are denoted by ( S α ) , ( S β ) , and ( S δ ) . The position update in the direction of the prey position is estimated during the search process as
X ( t + 1 ) = T 1 + T 2 + T 3 3
where T 1 , T 2 , a n d   T 3 are calculated as
T 1 = S α A 1 . D ,   T 2 = S β A 2 . D ,   T 3 = S δ A 3 . D
where D is calculated as | C 1 . ( F α S α + F β S β + F δ S δ ) X ( t ) | . The A and C vectors are defined as A = 2 a . r 1 a and C = 2 r 2 , where the vectors values r 1 and r 2 are randomly selected from the range [ 0 , 1 ] . The values of a are determined in the range [ 0 , 2 ] and is calculated as a = 2 t . 2 T M a x for T M a x iterations. The fitness functions are calculated as
F α = F α F α + F β + F δ ,   F β = F β F α + F β + F δ ,   F δ = F δ F α + F β + F δ
The ADSCFGWO method automatically balances the subgroups of the population’s exploitation and exploration. The algorithm uses a 70/30 system in which two groups—exploration and exploitation groups—represent 70% of the population. A large number of participants in the exploration group early in the optimization process helps with the discovery of novel and intriguing search regions. The overall fitness of agents increases when more exploitative agents can increase their fitness values, but the proportion of agents engaged in exploration falls quickly from 70% to 30%. If a better solution cannot be identified, using an elitism approach ensures convergence by keeping the process leader in consecutive populations. ADSCFGWO may at any point increase the size of the exploration group, provided that the leader’s fitness has not dramatically increased over the course of three consecutive iterations.
The suggested ADSCFGWO algorithm’s computational complexity can be stated as in Table 3 for population n and iterations t m a x . From this analysis, the complexity of computations is O ( t m a x × n ) and O ( t m a x × n × d ) with d dimension.
Algorithm 1 ADSCFGWO algorithm
1:
Initialize ADSCFGWO population X i ( i = 1 , 2 , , n ) with size n, iterations t M a x , fitness function F n , parameters (a, A 1 , A 2 , A 3 , C 1 , C 2 , r 1 , r 2 , r 3 , r 4 )
2:
Calculate fitness function F n for each X i
3:
Find best solutions as S α , S β , S δ
4:
Set t = 1
5:
while   t t M a x do
6:
  Update  r 1 by r 1 = a 1 t M a x i t e r
7:
  for ( i = 1 : i n 1 ) do
8:
      DynamicSearch( F n )
9:
      Update Fitness by Equation (3)
10:
    Update positions from GWO as X ( t + 1 ) = T 1 + T 2 + T 3 3
11:
    if ( r 4 < 0.5 ) then
12:
      Update positions from SCA as X ( t + 1 ) = X ( t ) + r 1 × sin ( r 2 ) × | r 3 S α X ( t ) |
13:
    end if
14:
  end for
15:
  for ( i = 1 : i n 2 ) do
16:
    DynamicSearch( F n )
17:
    Update Fitness by Equation (3)
18:
    Update positions from GWO as X ( t + 1 ) = T 1 + T 2 + T 3 3
19:
    if ( r 4 0.5 ) then
20:
      Update positions from SCA as X ( t + 1 ) = X ( t ) + r 1 × cos ( r 2 ) × | r 3 S α X ( t ) |
21:
    end if
22:
  end for
23:
  Update fitness function F n for each X i
24:
  Update parameters
25:
  Find best solutions as S α , S β , S δ
26:
end while
27:
Return best solution X *
28:
DynamicSearch( F n )
29:
if (Best F n is same for three iterations) then
30:
  Increase exploration group solutions ( n 1 )
31:
  Decrease exploitation group solutions ( n 2 )
33:
end if
Table 3. ADSCFGWO algorithm’s computational complexity.

3.3. 3D U-Net Architecture Segmentation Model

U-Net [42] is a network that is used for fast and accurate image segmentation. It comprises an expanded pathway and a contracting pathway. The contracting pathway adheres to the standard convolutional network design. Two 3 × 3 unpadded convolution layers are applied repeatedly, and after them, a ReLU activation function and a 2 × 2 max-pooling with stride 2 are applied for down-sampling. The number of features at every stage in the down-sampling process is doubled. The expanding pathway consists of an up-sampling process, a 2 × 2 convolution layer that reduces the size of the feature map, a combination with the proportionally clipped feature map from the contracting pathway, and two 3 × 3 convolution layers, each accompanied by a ReLU activation function. All the 64-component extracted features are mapped to the required number of categories in the last layer using a 1 × 1 convolution layer. The model includes 23 convolutional layers overall. The reason for using a U-net network is that it is fast compared to other networks. On a modern GPU, segmentation of a 512 × 512 picture consumes less than a second.
Numerous U-Net-based variant networks have been proposed since U-Net [42]’s extensive research and application in medical image segmentation in 2015; 3D U-Net [43] is the most representative of these. Figure 2 depicts the 3D U-Net’s structure. The encoder-decoder architecture of this model expands on the prior U-Net (2D). The encoder component performs feature extraction from an analysis of the input image. The associated decoder produces a segmented mask. The mask extraction is supervised by this model by minimizing a cost function. 3D U-Net differs from 2D U-Net in that its features are extracted and restored using 3D convolution, 3D max-pooling, and 3D deconvolution blocks in turn after the volume data are input. In addition, batch normalization is added by 3D U-Net to prevent bottlenecks and hasten convergence. For the segmentation process, the dataset is partitioned into a train, validation, and test datasets.
Figure 2. Structure of the 3D U-Net.

4. Experimental Results

This section describes the used dataset, the Performance metrics used in CNN, the implementation of the proposed strategy, and the experiments conducted. The parameters for the ADSCFGWO algorithm’s configuration are shown in Table 4.
Table 4. Parameters for the ADSCFGWO algorithm’s configuration.

4.1. Dataset Description

The used dataset is BRaTS 2021 Task 1 Dataset [44]. As training, validation, and testing data for this year’s BraTS challenge, a sizable number of multi-institutional regular clinically acquired multi-parametric MRI (mpMRI) images of glioma with pathologically confirmed diagnosis and accessible MGMT promoter methylation status are used. For Task 1, the datasets utilized in this year’s competition have been updated with many additional routines clinically collected mpMRI scans since BraTS’20. To quantitatively assess the projected tumor segmentations, expert neuroradiologists create and approve ground truth annotations of tumor sub-regions for each patient in the training, validation, and testing datasets. As shown in Figure 3, the dataset is partitioned into a train, validation, and test datasets. Figure 4 illustrates an example of the dataset.
Figure 3. Dataset distribution.
Figure 4. Example of dataset images.
The data augmentation technique is used in this study to artificially generate fresh training data from the current data. As a sort of data augmentation, picture augmentation produces altered representations of the training dataset’s images. The input dataset is subjected to several image transformations, such as horizontal and vertical shift, horizontal and vertical flip, random rotation, and random zoom. The shift augmentation maintains the same image dimensions while shifting all of the MRI image’s pixels in either a horizontal or vertical direction. When flipping horizontally or vertically, all pixels’ rows and columns are reversed. The MRI image is randomly rotated between 0 and 360 degrees clockwise using the rotation augmentation. The zoom augmentation’s final step involves randomly zooming the MRI image by a factor between [0.9, 1.1].

4.2. Performance Metrics Used in CNN

The conventional computer-aided diagnostic approach may be tested using a variety of key performance metrics, including accuracy, precision, F1-score, recall, specificity, and sensitivity. The number of cases that were accurately identified as defective is shown by the letter T P , which stands for True Positive. False Positive, abbreviated as F P , refers to the number of cases that were incorrectly identified as defective. Additionally, F N stands for False Negative and reflects the number of occurrences that were incorrectly classified as non-defective. T N is for True Negative, which represents the number of cases that were correctly identified as non-defective. The metrics are defined as in Table 5.
Table 5. Performance metrics used in CNN.

4.3. The BCM-CNN Evaluation

As shown in Table 6, the effectiveness of the suggested approach (BCM-CNN) is evaluated in comparison to the previously widely used classifiers CNN [22], Decision Tree (DT) [45], Linear Discriminant (LD) [46], Support Vector Machine (SVM) [47], and K-Nearest Neighbor (K-NN) [48]. The default parameters are used for these methods.
Table 6. The performance of the proposed method (BCM-CNN) versus basic classifiers.
Samples from the texted dataset are employed in the classification experiment. As a result of the BCM-CNN based on the ADSCFGWO algorithm boosting the performance of the CNN after altering its hyperparameters, the BCM-CNN delivered the best results when employed as a classifier, with an accuracy of (0.99980004), Sensitivity (TRP) of (0.99980004), Specificity (TNP) of ( 0.99980004), Pvalue (PPV) of (0.99980004), Nvalue (NPV) of (0.99980004), and F1-score of (0.9998). After the SVM-Linear model, which has an accuracy score of (0.968992248), the K-NN model, which has an accuracy score of (0.965250965), and finally, the LD model, which has an accuracy score of (0.961538462), the simple CNN model gets the second-best accuracy with a score of (0.9765625). The DT model was only able to achieve the lowest level of accuracy with (0.956022945). This came about as a consequence of the fact that the method that was proposed resulted in an improvement in CNN’s overall performance.
Table 7 shows the proposed BCM-CNN-based classifier’s statistical description and a comparison of classifiers based on 11 runs (run the algorithm 11 times) and 80 iterations ( t M a x in Algorithm 1) for 10 agents (Population size n in Algorithm 1) of the ADSCFGWO algorithm. This is to confirm the stability of the proposed method compared to other methods. Table 8 presents the compared and the proposed classifier’s test results using a one-way ANOVA (analysis of variance) test. In contrast, Table 9 discusses the comparison and the proposed classifiers test results using the Wilcoxon Signed-Rank test. With a p-value of less than 0.05, this statistical test demonstrates the significant difference between the suggested BCM-CNN classifier’s results and those of other classifiers.
Table 7. Proposed BCM-CNN classifier’s statistical description and a comparison of classifiers.
Table 8. Compared and proposed classifiers test results for ANOVA.
Table 9. Compared and proposed classifiers test results for Wilcoxon Signed-Rank.
The accuracy of the proposed BCM-CNN and comparative methods is shown by the box plot in Figure 5. This graph demonstrates the maximum accuracy results that the BCM-CNN-based optimization algorithm was able to produce. Based on the number of values with the Bin Center range (0.946–1.0), the accuracy histogram for the algorithms that have been presented and compared is shown in Figure 6, which attests to the stability of the suggested algorithm.
Figure 5. Box plot of accuracy for the BCM-CNN model under consideration and the contrasting models.
Figure 6. Histogram of accuracy for the BCM-CNN model under consideration and the contrasting models.
Figure 7 displays the residual, QQ (quantile-quantile), homoscedasticity plots, and heat map for the proposed and compared techniques. The possible problems can be observed in the residual values and plots as opposed to the plot of the original dataset. The independent variable is plotted on the horizontal axis, while the residual values are plotted on the vertical axis. The ideal situation is achieved if the residual values are scattered randomly and uniformly along the horizontal axis. The residual value is calculated as follows when the mean and the sum of the residuals are both equal to zero: (Actual-Predicted values). Figure 7 displays the residual plot. To determine if a model is linear or nonlinear and which one is best, plot patterns in a residual plot can be used. The projected scores for the dependent variable are examined visually together with the homogeneity of variance or heteroscedasticity. When the error term, also known as noise or random disturbance in the relationship between the dependent and independent variables, is constant across all values of the independent variables, this situation is referred to as homoscedasticity. The heteroscedasticity plot, shown in Figure 7, improves the precision of the research results. Any infraction can be quickly and easily detected.
Figure 7. The heat map, residual, QQ, and homoscedasticity plots of the ADSCFGWO and comparable algorithms.
The QQ plot is also shown in Figure 7. A probability plot is one illustration. By plotting the quantiles against one another, two probability distributions are primarily compared. It is possible to see that the point distributions in the QQ plot fit on the line in the illustration. Since the relationship between the actual and projected residuals is linear, the suggested technique is effective. Figure 7 serves as a tool for data visualization and displays heat maps for the offered and contrasted algorithms. The intensity of a two-dimensional color scale indicates the complexity of an algorithm. The color fluctuation provides obvious visual cues as to how the proposed solution is superior to the comparable algorithms. The ADSCFGWO algorithm’s performance in feature selection, as seen in Figure 7, is supported by these figures.

4.4. 3D U-Net Segmentation Model

There are four classes in the segmentation process. Segmentation classes are NOT tumor, non-enhancing tumor (RED color), EDEMA (Green color), and ENHANCING (yellow color). These classes were converted into three classes later. Figure 8 illustrates samples of images and masks with a positive brain tumor. For more precious and fast detection of brain tumor, the 3D U-net segmentation model has been implemented on the BRaTS 2021 dataset. The dataset is divided into 70% training, 20% validation, and 10% testing. Implementation is constructed online on Kaggle. U-net model enhances segmentation validation accuracy up to 99.33%, and validation loss up to 0.01 as shown in Figure 9. We can conclude that our proposed model can detect brain tumor with high accuracy compared to state-of-the-art techniques in terms of classification and segmentation.
Figure 8. Visualize image with mask of a positive brain tumor.
Figure 9. Training and validation performance parameters; (a) Accuracy, (b) Loss, (c) Dice coef.

5. Conclusions and Future Work

A large number of researchers looked at a wide variety of algorithms with the goal of accurately detecting and classifying brain cancers in a quick and efficient manner. Deep learning (DL) makes it possible to use a Convolutional Neural Network (CNN) model that has already been pre-trained for the analysis of medical pictures, in particular for the categorization of brain tumors. The fundamental objective of this research is to develop an improved model with the intention of making brain tumor diagnosis more accurate. A Convolutional Neural Network (CNN) that is based on a Brain Tumor Classification Model (BCM-CNN) was proposed in this paper. The optimization of the CNN’s hyperparameters was based on an adaptive dynamic sine-cosine fitness grey wolf optimizer (ADSCFGWO) algorithm. The BCM-CNN was used as a classifier in the experiments, and the results reveal that it produced the best results due to the enhancement of the performance of the CNN after the optimization was performed. The BCM-CNN was given the BRaTS 2021 Task 1 dataset, and it performed with an accuracy of 99.99%. The main limitation of the proposed algorithm is that it takes a long time to process due to the extra optimization steps. It may not be applicable as the size of the trained data is limited, so we intend to solve this issue in future work by generalizing more data. We also intend to make a prediction in future work, not just classification.

Author Contributions

Conceptualization, H.Z.; Methodology, S.A.G. and D.S.K.; Software, E.-S.M.E.-K.; Validation, A.H.A. and A.I.; Formal analysis, H.Z.; Investigation, S.A.G., A.I. and F.M.T.; Resources, A.H.A. and D.S.K.; Data curation, F.M.T.; Writing—original draft, H.Z., S.A.G. and D.S.K.; Writing—review & editing, E.-S.M.E.-K. and A.I.; Visualization, A.H.A., D.S.K. and F.M.T.; Supervision, E.-S.M.E.-K.; Project administration, E.-S.M.E.-K.; Funding acquisition, H.Z., S.A.G. and A.I. All authors have read and agreed to the published version of the manuscript.

Funding

Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R120), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the Deanship of Scientific Research at Shaqra University for supporting this work.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial intelligence
DLDeep Learning
CNNConvolutional Neural Network
BCM-CNNBrain Tumor Classification Model based on CNN
ADSCFGWOAdaptive Dynamic Sine-Cosine Fitness Grey Wolf Optimizer
WHOWorld Health Organization
MEGMagneto Encephalon Graph
CTComputed Tomography
EEGUltrasonography, Electro Encephalon Graph
SPECTSingle-Photon Emission Computed Tomography
PETPositron Emission Tomography
MRIMagnetic Resonance Imaging
CADComputer-Aided Diagnosis
NNNeural Network
FCFully Connected
DTDecision Tree
NBNaive Bayes
LDLinear Discrimination
SVMSupport Vector Machine
AUCArea Under the Curve
LDLinear Discriminant
QQQuantile-Quantile

References

  1. Al-Galal, S.A.Y.; Alshaikhli, I.F.T.; Abdulrazzaq, M.M. MRI brain tumor medical images analysis using deep learning techniques: A systematic review. Health Technol. 2021, 11, 267–282. [Google Scholar] [CrossRef]
  2. Rahman, M.L.; Reza, A.W.; Shabuj, S.I. An internet of things-based automatic brain tumor detection system. Indones. J. Electr. Eng. Comput. Sci. 2022, 25, 214–222. [Google Scholar] [CrossRef]
  3. Key Statistics for Brain and Spinal Cord Tumors. Available online: https://www.cancer.org/cancer/brain-spinal-cord-tumors-adults/about/key-statistics.html (accessed on 20 September 2022).
  4. Ayadi, W.; Elhamzi, W.; Charfi, I.; Atri, M. Deep CNN for Brain Tumor Classification. Neural Process. Lett. 2021, 53, 671–700. [Google Scholar] [CrossRef]
  5. Liu, J.; Li, M.; Wang, J.; Wu, F.; Liu, T.; Pan, Y. A survey of MRI-based brain tumor segmentation methods. Tsinghua Sci. Technol. 2014, 19, 578–595. [Google Scholar] [CrossRef]
  6. Amin, J.; Sharif, M.; Haldorai, A.; Yasmin, M.; Nayak, R.S. Brain tumor detection and classification using machine learning: A comprehensive survey. Complex Intell. Syst. 2021, 8, 3161–3183. [Google Scholar] [CrossRef]
  7. Jayade, S.; Ingole, D.T.; Ingole, M.D. Review of Brain Tumor Detection Concept using MRI Images. In Proceedings of the 2019 International Conference on Innovative Trends and Advances in Engineering and Technology (ICITAET), Shegoaon, India, 27–28 December 2019. [Google Scholar] [CrossRef]
  8. Yang, Y.; Yan, L.F.; Zhang, X.; Han, Y.; Nan, H.Y.; Hu, Y.C.; Hu, B.; Yan, S.L.; Zhang, J.; Cheng, D.L.; et al. Glioma Grading on Conventional MR Images: A Deep Learning Study With Transfer Learning. Front. Neurosci. 2018, 12, 804. [Google Scholar] [CrossRef]
  9. Nazir, M.; Shakil, S.; Khurshid, K. Role of deep learning in brain tumor detection and classification (2015 to 2020): A review. Comput. Med. Imaging Graph. 2021, 91, 101940. [Google Scholar] [CrossRef]
  10. El-Kenawy, E.S.M.; Mirjalili, S.; Abdelhamid, A.A.; Ibrahim, A.; Khodadadi, N.; Eid, M.M. Meta-Heuristic Optimization and Keystroke Dynamics for Authentication of Smartphone Users. Mathematics 2022, 10, 2912. [Google Scholar] [CrossRef]
  11. El-kenawy, E.S.M.; Albalawi, F.; Ward, S.A.; Ghoneim, S.S.M.; Eid, M.M.; Abdelhamid, A.A.; Bailek, N.; Ibrahim, A. Feature Selection and Classification of Transformer Faults Based on Novel Meta-Heuristic Algorithm. Mathematics 2022, 10, 3144. [Google Scholar] [CrossRef]
  12. El-Kenawy, E.S.M.; Mirjalili, S.; Alassery, F.; Zhang, Y.D.; Eid, M.M.; El-Mashad, S.Y.; Aloyaydi, B.A.; Ibrahim, A.; Abdelhamid, A.A. Novel Meta-Heuristic Algorithm for Feature Selection, Unconstrained Functions and Engineering Problems. IEEE Access 2022, 10, 40536–40555. [Google Scholar] [CrossRef]
  13. Ibrahim, A.; Mirjalili, S.; El-Said, M.; Ghoneim, S.S.M.; Al-Harthi, M.M.; Ibrahim, T.F.; El-Kenawy, E.S.M. Wind Speed Ensemble Forecasting Based on Deep Learning Using Adaptive Dynamic Optimization Algorithm. IEEE Access 2021, 9, 125787–125804. [Google Scholar] [CrossRef]
  14. El-kenawy, E.S.M.; Abutarboush, H.F.; Mohamed, A.W.; Ibrahim, A. Advance Artificial Intelligence Technique for Designing Double T-shaped Monopole Antenna. Comput. Mater. Contin. 2021, 69, 2983–2995. [Google Scholar] [CrossRef]
  15. Samee, N.A.; El-Kenawy, E.S.M.; Atteia, G.; Jamjoom, M.M.; Ibrahim, A.; Abdelhamid, A.A.; El-Attar, N.E.; Gaber, T.; Slowik, A.; Shams, M.Y. Metaheuristic Optimization Through Deep Learning Classification of COVID-19 in Chest X-Ray Images. Comput. Mater. Contin. 2022, 73, 4193–4210. [Google Scholar] [CrossRef]
  16. Lee, G.; Nho, K.; Kang, B.; Sohn, K.A.; Kim, D. Predicting Alzheimer’s disease progression using multi-modal deep learning approach. Sci. Rep. 2019, 9, 1952. [Google Scholar] [CrossRef] [PubMed]
  17. Sharma, K.; Kaur, A.; Gujral, S. Brain Tumor Detection based on Machine Learning Algorithms. Int. J. Comput. Appl. 2014, 103, 7–11. [Google Scholar] [CrossRef]
  18. Agrawal, M.; Jain, V. Prediction of Breast Cancer based on Various Medical Symptoms Using Machine Learning Algorithms. In Proceedings of the 2022 6th International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 28–30 April 2022; pp. 1242–1245. [Google Scholar] [CrossRef]
  19. Rabbi, M.F.; Mahedy Hasan, S.M.; Champa, A.I.; AsifZaman, M.; Hasan, M.K. Prediction of Liver Disorders using Machine Learning Algorithms: A Comparative Study. In Proceedings of the 2020 2nd International Conference on Advanced Information and Communication Technology (ICAICT), Dhaka, Bangladesh, 28–29 November 2020; pp. 111–116. [Google Scholar] [CrossRef]
  20. Swain, D.; Pani, S.K.; Swain, D. A Metaphoric Investigation on Prediction of Heart Disease using Machine Learning. In Proceedings of the 2018 International Conference on Advanced Computation and Telecommunication (ICACAT), Bhopal, India, 28–29 December 2018; pp. 1–6. [Google Scholar] [CrossRef]
  21. Zhu, W.; Sun, L.; Huang, J.; Han, L.; Zhang, D. Dual Attention Multi-Instance Deep Learning for Alzheimer’s Disease Diagnosis With Structural MRI. IEEE Trans. Med Imaging 2021, 40, 2354–2366. [Google Scholar] [CrossRef]
  22. Abdelhamid, A.A.; El-Kenawy, E.S.M.; Alotaibi, B.; Amer, G.M.; Abdelkader, M.Y.; Ibrahim, A.; Eid, M.M. Robust Speech Emotion Recognition Using CNN+LSTM Based on Stochastic Fractal Search Optimization Algorithm. IEEE Access 2022, 10, 49265–49284. [Google Scholar] [CrossRef]
  23. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar] [CrossRef]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar] [CrossRef]
  25. Alhussan, A.A.; Khafaga, D.S.; El-Kenawy, E.S.M.; Ibrahim, A.; Eid, M.M.; Abdelhamid, A.A. Pothole and Plain Road Classification Using Adaptive Mutation Dipper Throated Optimization and Transfer Learning for Self Driving Cars. IEEE Access 2022, 10, 84188–84211. [Google Scholar] [CrossRef]
  26. Srikanth, B.; Suryanarayana, S.V. Multi-Class classification of brain tumor images using data augmentation with deep neural network. Mater. Today Proc. 2021. [Google Scholar] [CrossRef]
  27. Tandel, G.S.; Balestrieri, A.; Jujaray, T.; Khanna, N.N.; Saba, L.; Suri, J.S. Multiclass magnetic resonance imaging brain tumor classification using artificial intelligence paradigm. Comput. Biol. Med. 2020, 122, 103804. [Google Scholar] [CrossRef]
  28. Deepak, S.; Ameer, P. Brain tumor classification using deep CNN features via transfer learning. Comput. Biol. Med. 2019, 111, 103345. [Google Scholar] [CrossRef] [PubMed]
  29. Kokkalla, S.; Kakarla, J.; Venkateswarlu, I.B.; Singh, M. Three-class brain tumor classification using deep dense inception residual network. Soft Comput. 2021, 25, 8721–8729. [Google Scholar] [CrossRef] [PubMed]
  30. Özlem, P.; Güngen, C. Classification of brain tumors from MR images using deep transfer learning. J. Supercomput. 2021, 77, 7236–7252. [Google Scholar] [CrossRef]
  31. Kesav, N.; Jibukumar, M. Efficient and low complex architecture for detection and classification of Brain Tumor using RCNN with Two Channel CNN. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 6229–6242. [Google Scholar] [CrossRef]
  32. Tummala, S.; Kadry, S.; Bukhari, S.A.C.; Rauf, H.T. Classification of Brain Tumor from Magnetic Resonance Imaging Using Vision Transformers Ensembling. Curr. Oncol. 2022, 29, 7498–7511. [Google Scholar] [CrossRef]
  33. Pareek, M.; Jha, C.K.; Mukherjee, S. Brain Tumor Classification from MRI Images and Calculation of Tumor Area. In Advances in Intelligent Systems and Computing; Springer: Singapore, 2020; pp. 73–83. [Google Scholar] [CrossRef]
  34. Ayadi, W.; Charfi, I.; Elhamzi, W.; Atri, M. Brain tumor classification based on hybrid approach. Vis. Comput. 2020, 38, 107–117. [Google Scholar] [CrossRef]
  35. Konar, D.; Bhattacharyya, S.; Panigrahi, B.K.; Behrman, E.C. Qutrit-Inspired Fully Self-Supervised Shallow Quantum Learning Network for Brain Tumor Segmentation. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 6331–6345. [Google Scholar] [CrossRef]
  36. Khairandish, M.; Sharma, M.; Jain, V.; Chatterjee, J.; Jhanjhi, N. A Hybrid CNN-SVM Threshold Segmentation Approach for Tumor Detection and Classification of MRI Brain Images. IRBM 2022, 43, 290–299. [Google Scholar] [CrossRef]
  37. Öksüz, C.; Urhan, O.; Güllü, M.K. Brain tumor classification using the fused features extracted from expanded tumor region. Biomed. Signal Process. Control 2022, 72, 103356. [Google Scholar] [CrossRef]
  38. Khan, M.A.; Ashraf, I.; Alhaisoni, M.; Damaševičius, R.; Scherer, R.; Rehman, A.; Bukhari, S.A.C. Multimodal Brain Tumor Classification Using Deep Learning and Robust Feature Selection: A Machine Learning Application for Radiologists. Diagnostics 2020, 10, 565. [Google Scholar] [CrossRef]
  39. Kadry, S.; Nam, Y.; Rauf, H.T.; Rajinikanth, V.; Lawal, I.A. Automated Detection of Brain Abnormality using Deep-Learning-Scheme: A Study. In Proceedings of the 2021 Seventh International conference on Bio Signals, Images, and Instrumentation (ICBSII), Chennai, India, 25–27 March 2021. [Google Scholar] [CrossRef]
  40. Irmak, E. Multi-Classification of Brain Tumor MRI Images Using Deep Convolutional Neural Network with Fully Optimized Framework. Iran. J. Sci. Technol. Trans. Electr. Eng. 2021, 45, 1015–1036. [Google Scholar] [CrossRef]
  41. Khafaga, D.S.; Alhussan, A.A.; El-Kenawy, E.S.M.; Ibrahim, A.; Eid, M.M.; Abdelhamid, A.A. Solving Optimization Problems of Metamaterial and Double T-Shape Antennas Using Advanced Meta-Heuristics Algorithms. IEEE Access 2022, 10, 74449–74471. [Google Scholar] [CrossRef]
  42. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Lecture Notes in Computer Science; Springer International Publishing: Berlin, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  43. Özgün, Ç.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016; Springer International Publishing: Berlin, Germany, 2016; pp. 424–432. [Google Scholar] [CrossRef]
  44. BRaTS 2021 Task 1 Dataset, RSNA-ASNR-MICCAI Brain Tumor Segmentation (BraTS) Challenge 2021. Available online: https://www.kaggle.com/datasets/dschettler8845/brats-2021-task1?select=BraTS2021_Training_Data.tar (accessed on 20 September 2022).
  45. Nour, M.; Cömert, Z.; Polat, K. A Novel Medical Diagnosis model for COVID-19 infection detection based on Deep Features and Bayesian Optimization. Appl. Soft Comput. 2020, 97, 106580. [Google Scholar] [CrossRef] [PubMed]
  46. Tharwat, A.; Gaber, T.; Ibrahim, A.; Hassanien, A.E. Linear discriminant analysis: A detailed tutorial. AI Commun. 2017, 30, 169–190. [Google Scholar] [CrossRef]
  47. Benmahamed, Y.; Teguar, M.; Boubakeur, A. Application of SVM and KNN to Duval Pentagon 1 for transformer oil diagnosis. IEEE Trans. Dielectr. Electr. Insul. 2017, 24, 3443–3451. [Google Scholar] [CrossRef]
  48. Benmahamed, Y.; Kemari, Y.; Teguar, M.; Boubakeur, A. Diagnosis of Power Transformer Oil Using KNN and Nave Bayes Classifiers. In Proceedings of the 2018 IEEE 2nd International Conference on Dielectrics (ICD), Budapest, Hungary, 1–5 July 2018. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.