Skip to Content
DiagnosticsDiagnostics
  • Article
  • Open Access

13 June 2023

Dual Deep CNN for Tumor Brain Classification

,
,
,
and
1
Faculty of Computers and Artificial Intelligence, Damietta University, Damietta 34517, Egypt
2
Computer Science Department, Faculty of Computer and Information Science, Mansoura University, Mansoura 35511, Egypt
*
Authors to whom correspondence should be addressed.

Abstract

Brain tumor (BT) is a serious issue and potentially deadly disease that receives much attention. However, early detection and identification of tumor type and location are crucial for effective treatment and saving lives. Manual diagnoses are time-consuming and depend on radiologist experts; the increasing number of new cases of brain tumors makes it difficult to process massive and large amounts of data rapidly, as time is a critical factor in patients’ lives. Hence, artificial intelligence (AI) is vital for understanding disease and its various types. Several studies proposed different techniques for BT detection and classification. These studies are on machine learning (ML) and deep learning (DL). The ML-based method requires handcrafted or automatic feature extraction algorithms; however, DL becomes superior in self-learning and robust in classification and recognition tasks. This research focuses on classifying three types of tumors using MRI imaging: meningioma, glioma, and pituitary tumors. The proposed DCTN model depends on dual convolutional neural networks with VGG-16 architecture concatenated with custom CNN (convolutional neural networks) architecture. After conducting approximately 22 experiments with different architectures and models, our model reached 100% accuracy during training and 99% during testing. The proposed methodology obtained the highest possible improvement over existing research studies. The solution provides a revolution for healthcare providers that can be used as a different disease classification in the future and save human lives.

1. Introduction

Cancer is an alarming neoplastic disorder characterized by uncontrolled cellular proliferation and the potential for metastasis, leading to significant morbidity and mortality. The classification of brain tumors can be one of two categories: noncancerous (benign) or cancerous (malignant). Primary brain tumors develop from the brain’s cells or the surrounding area. Primary tumors are categorized as benign or malignant depending on whether they are nonglial (formed in brain systems such as neurons, blood vessels, and glands) or glial (composed of glial cells). The World Health Organization (WHO) collects and analyzes data on cancer incidence and mortality rates worldwide through its Global Cancer Observatory (GCO) [1]. The GCO provides information on the number of cancer cases and deaths by cancer type, including brain tumors, in countries worldwide. According to the GCO, an estimated 306,000 new cases of brain tumors were diagnosed worldwide in 2020 [2]. However, it is essential to note that this figure is an estimate and may not reflect the full extent of brain tumor incidence, as not all countries have comprehensive cancer registries or reporting systems. In addition to the GCO, other organizations, such as the National Brain Tumor Society and the American Brain Tumor Association, also provide statistics and information on the incidence and prevalence of brain tumors in the United States. According to the American Brain Tumor Association, approximately 87,000 people in the United States live with a primary brain tumor, and approximately 24,530 new cases are diagnosed yearly. However, as with the GCO data, these figures are estimates and may not reflect the full extent of brain tumor incidence in the United States [3].
For many people, there are no differences between tumors and cancers, but in fact, they are different. In other words, cancers are tumors, but all tumors are not cancers, and that does not mean that tumors are safe. Referring to [4], the study stands out because it focuses on the crucial challenge of dissecting glioblastoma’s phenotypic and genetic heterogeneity (GBM) at spatial and temporal levels. GBM is known for its complexity, with variations in genetic mutations, cellular composition, and immune responses observed within and between tumors. Understanding this heterogeneity is vital for developing effective treatment strategies. However, analyzing a single tumor sample fails to capture the diverse characteristics and dynamics of the GBM microenvironment.
Additionally, it is essential to differentiate between tumors and cancers, as noncancerous or benign brain tumors can still be dangerous and cause severe dysfunction. Therefore, this study recognizes the need to study GBM from a multi-dimensional perspective, investigating the interactions between cancer cells and myeloid-lineage cells at different spatial locations within the tumor and considering temporal changes. By addressing the spatial and temporal heterogeneity, the study aims to uncover the mechanisms underlying tumor-immune symbiosis in GBM and identify potential therapeutic targets. This approach has the potential to provide novel insights into the tumor microenvironment and enhance the effectiveness of immunotherapeutic strategies while also considering the broader spectrum of brain tumors, including noncancerous ones. They grow slowly at first and then spread to other parts of the brain, so benign tumors are also dangerous because of their size and location, and a benign tumor can become malignant in sporadic cases. There are many types of tumors; this research focuses on three types of brain tumors: meningioma, glioma, and pituitary tumors [5].
Glioma originates in the supportive tissue cells of the brain or spine and can be classified into various grades based on its growth rate and potential to spread. Its symptoms may include changes in behavior, mood, seizures, and headaches. On the other hand, meningioma starts in the meninges, the protective membranes covering the brain and spinal cord. Although it is usually non-cancerous and slow-growing, it can still cause symptoms such as seizures, vision problems, and headaches [6]. Treatment for a meningioma depends on the tumor’s size and location and can include surgery, radiation therapy, or observation. Lastly, pituitary tumors occur in the pituitary gland, which regulates hormones in the body. They can be benign and cause symptoms such as headaches, vision issues, and hormonal imbalances. Treatment for pituitary tumors can involve surgery, radiation therapy, medication, or a combination of these approaches, depending on the tumor’s type and size. Figure 1 shows the difference between glioma, meningioma, and pituitary tumors [7].
Figure 1. Different brain tumor types.
Magnetic resonance imaging (MRI) is a powerful imaging technique that is particularly effective at detecting brain tumors. This is because it can produce detailed images of the brain with excellent contrast and multiplanar imaging capabilities, which helps doctors distinguish between different types of brain tissue and accurately diagnose tumors. In panic attacks, individuals may undergo subjective distress, but with the aid of an adept radiologist, the majority can triumph over the symptoms through appropriate management strategies and therapeutic interventions [8]. This approach is not only to identify tumors inside the brain but to also check the whole internal structure of the human body for any malignancy. Additionally, MRI is a safe and noninvasive imaging technique that does not require the use of contrast agents or ionizing radiation. Overall, the ability of MRI to provide high-quality images and accurate diagnoses makes it a valuable tool in detecting and diagnosing brain tumors [9].
Several MRI modalities differ in imaging parameters, sequence type, and application. Some common MRI modalities (T1-weighted, T2-weighted, diffusion-weighted, and magnetic resonance spectroscopy). T1-weighted imaging produces images that provide good anatomical detail and contrast between different types of tissues, such as fat, muscle, and bone [10]. T1-weighted images are typically used to estimate the brain, spine, liver, and other organs. T2-weighted images contrast different soft tissue types, such as brain tissue, muscle, and fluid-filled spaces. T2-weighted images are often used to evaluate the brain, spine, and joints. Diffusion-weighted imaging uses a technique that measures the random movement of water molecules in tissue to produce images that highlight changes in the tissue’s microstructure [11]. Diffusion-weighted is often used to evaluate the brain, spine, and prostate gland. Magnetic resonance spectroscopy (MRS) uses the same basic principles as MRI, but instead of producing images, it measures the chemical composition of tissues. MRS is often used to evaluate the brain for metabolic disorders and tumors. Each MRI modality has strengths and limitations and is used for specific clinical indications. The specialist will determine which modality best suits each case [12].
Manual extraction from MRI (magnetic resonance imaging) refers to manually segmenting or identifying structures or regions of interest in medical images obtained through MRI scans. This process involves a human expert, such as a radiologist, manually tracing or drawing boundaries around the structures of interest using specialized software. Manual extraction from an MRI can have several drawbacks, including being a time-consuming and labor-intensive process that requires specialized expertise. This can lead to higher costs and longer patient wait times [13]. The manual extraction process is also prone to human error, resulting in inaccurate results and misdiagnosis.
On the other hand, using an automatic system to extract MRI data has many benefits, including time savings, consistency, cost-effectiveness, improved accuracy, standardization, faster diagnoses and treatments, and improved patient outcomes. Automated systems are less prone to human error and can process large volumes of data quickly, reducing the time needed for extraction compared to manual methods. The accuracy and speed of automated systems can lead to better patient outcomes, including faster diagnoses, better treatment plans, and faster recovery times [14]. Time is a critical factor in saving patients lives with brain tumors. Early detection and treatment can significantly improve the chances of survival, making it essential to be aware of the warning signs and seek medical attention promptly. Furthermore, automated systems outperform manual processes in processing vast amounts of data images [14].
Artificial intelligence and deep learning advancements have significantly aided in the early detection of tumors, potentially reducing reliance on human experts and saving lives. Machine learning and deep learning techniques, particularly convolutional neural networks, have been extensively applied in radiology to automatically extract features from input images. Brain tumors demonstrate notable heterogeneity, comprising diverse cell populations influenced by genetic mutations, epigenetic modifications, and microenvironmental factors. Primary brain tumors originate within the brain, while secondary tumors metastasize from other body parts [15]. Diagnostic imaging methods, such as MRI and CT scans, are invaluable for determining tumor location, size, and extent. Genetic alterations involving genes, such as EGFR, IDH1, TP53, and PTEN contribute to tumor development and progression. Treatment options encompass surgery, radiation therapy, chemotherapy, targeted therapy, and immunotherapy, considering the specific characteristics of tumor subtypes. Challenges include the blood-brain barrier, tumor recurrence, and treatment resistance, necessitating personalized treatment strategies and innovative therapeutic targets. Continuous research and the integration of AI contribute to the ongoing advancement of our understanding and the management of brain tumors. In the past, researchers relied on manually predefined characteristics extracted from medical images to perform the classification task [16]. However, with deep learning, the features extracted from the deeper convolutional layers may need to be more interpretable, making analysis more challenging. Despite this, these techniques have shown high performance and are widely used in radiology. By analyzing these features, we can better understand the capabilities and limitations of deep learning in the classification of tumors [17].
In the medical field, the conventional approach for retrieving and classifying images using machine learning only concentrates on either high-level or low-level features. It relies on features that are manually handcrafted in the feature engineering process. Consequently, there is a critical need to create a technique that integrates high-level and low-level features without manually designed features [18]. Low-level and high-level features are both essential in automatic brain tumor classification. Low-level features, which are crucial image characteristics, such as intensity values and texture descriptors, can be extracted through techniques, such as filtering and segmentation. Additionally, feature extraction algorithms such as local binary patterns (LBP) and Haralick features can be utilized to extract low-level features in automatic brain tumor feature extraction. They serve as a basis for higher-level feature representations [19].
In contrast, high-level features are derived from low-level features and are more abstract in nature. They are learned using deep learning methods, such as CNNs or RNNs. High-level features in brain tumor classification include tumor location, size, shape, and surrounding tissue characteristics. In order to achieve accurate and robust automatic brain tumor classification, it is essential to consider both low-level and high-level features. Low-level features capture fundamental image characteristics, whereas high-level features provide abstract and semantic information [20]. Combining these features is necessary to achieve high accuracy, and deep learning methods can be used to extract these features automatically, which is critical for accurate diagnosis and treatment planning. Despite numerous attempts to create an exact and robust automatic classification system for brain tumors, the task’s difficulty persists.
Artificial intelligence (AI) faces challenges in investigating disease understanding, including various types of cancer, such as data availability, privacy concerns, interpretability, clinical integration, and potential bias. However, AI is vital in addressing these challenges by analyzing large and diverse datasets, providing insights, improving diagnosis, personalizing treatments, and developing novel therapeutic strategies. Bioinformatics analysis, combined with AI, has emerged as a powerful tool for understanding cancer by integrating biological data, aiding in tumor classification, identifying genetic alterations, and facilitating personalized treatment planning. AI-driven bioinformatics analysis also accelerates drug discovery by identifying potential therapeutic targets. However, addressing data quality, privacy, and interpretability challenges is crucial for fully realizing the potential of AI and bioinformatics analysis in advancing cancer research and improving patient outcomes [15].
Artificial intelligence (AI) has a wide-ranging scope in cancer research, diagnosis, and treatment. By analyzing complex datasets, AI can identify patterns and generate insights that aid in cancer diagnosis, personalized treatment planning, and drug discovery. It contributes to improving accuracy, predicting therapeutic responses, and monitoring treatment outcomes. AI also supports cancer research by identifying biomarkers, uncovering molecular mechanisms, and discovering novel relationships. However, addressing challenges related to data quality, privacy, and interpretability is crucial for maximizing the potential of AI in advancing cancer understanding and enhancing patient outcomes [15].

3. Methods

The proposed research framework is highlighted and elaborated on in this section. The proposed framework is introduced in two steps. First, we expand on our deep learning approach and model architecture to detect and classify brain tumor RMI images as meningioma, glioma, or pituitary. Second, we describe each layer of the two CNNs for the DCTN model in detail. Third, we elaborate on the parameters used for training the DCTN model [45].
Our experiments involved the use of multiple CNN models for feature extraction, including EfficientNet, LeNet-5, AlexNet, and various other architectures, in order to achieve high accuracy. Additionally, we explored the application of ViT-based techniques, such as vit-base-patch16, which combines multiple transformer layers to capture the contextual information of input images. The results demonstrated significant improvements in accuracy across all models, with an average increase of 10% compared to the baseline. These findings highlight the effectiveness of both CNN- and ViT-based approaches in our experimental setup.

3.1. Dual Convolution Tumor Network (DCTN)

The DCTN model is built on a dual CNN architecture with a pretrained VGG-16 and a custom CNN. We employ VGG-16 with transfer learning to avoid training the entire pre-trained network. Transfer learning is a technique that allows the use of the same model, which has been learned and tuned with the best weights and biases, from one task to another in the same task domain. The RMI diagnostic task is an image classification task that involves extracting features from images and then providing those features and images to the classifiers. The VGG-16 has already been trained to extract features from the images and then classify the images. The VGG-16 is trained with ImageNet, which contains millions of image samples, and the model should classify the image as one of the 1000 categories. Transfer learning saves time and effort, and many researchers in different fields use it when there are insufficient datasets for training the network from scratch.
We built a small and simple model into a large model by following the trial-and-error approach and observing the accuracy and parameter improvements. The performance of VGG-16 scored low accuracy, and CNN with different architecture helped improve the performance and the accuracy of the VGG-16 model by adding a dual network with a different number of filters to build the DCTN model.
The DCTN has one input layer for both CNNs with a size of 224 × 224 × 3, followed by two branches for each CNN. The first branch provides the input to the first convolution layer of VGG-16, and the second branch provides the input to the first convolution layer of the custom CNN. After each CNN, we used GlobalMaxPooling2D; this allowed us to get the most critical features, lower the dimensionality of the feature maps, concatenate, and then pass the feature maps. The GlobalMaxPooling2D uses the max operation to get the essential values [46].
We concatenated the two CNNs, followed by 3 fully connected layers (FC). The first FC has 1000 neurons, followed by a dropout layer of 0.3 dropping rate. The dropout layer is one technique to tackle overfitting issues by cutting or removing some weights and allowing the model to generalize and make predictions independent of all the weights it has learned during the training process [47]. After the dropout layer, there is a second FC containing 256 layers, followed by another FC with 64 layers, and a SoftMax output layer. Besides SoftMax, there are different classifiers already existing, such as K-nearest neighbor (KNN), support vector machine (SVM), random forest (RF), decision tree (DT), XGBoost, activation functions, etc. [48]. Figure 2 illustrates the DCTN model architecture of the proposed method.
Figure 2. DCTN model architecture.
We used the dataset split as an 80/20 scheme: 80% of the data is used for the training process, and 20% is used for the testing process. We used 50 epochs for the forward pass and backpropagation to train the model, and Figure 3 shows the complete training and testing process for the proposed system.
Figure 3. The training and testing process.
The layers of each CNN are the following: the original pretrained model, VGG-16, has 16 layers. We removed the top of the network, including the FC and SoftMax output layers. The second custom CNN has 12 convolution layers. In the design of the custom CNN, we avoid having a higher number of filters in higher layers to avoid high dimensionality, Due to its capacity to automatically extract features from input images without requiring human intervention, CNN proves to be a time-saving and effort-reducing technique [49]. The input layer for both networks is the same layer with a size of 224 × 224 as shown in Figure 4, which summarizes the layers of the DCTN model.
Figure 4. Summary of DCTN model layers.

3.2. The DCTN Model Layers

The following section describes in detail each layer of DCTN networks. The VGG-16 has 13 layers, and the custom CNN has 12 layers. The VGG-16 has the same kernel or filter size for each convolution layer, 3 × 3 with stride, followed by the MaxPool layer with a kernel size of 2 × 2. The VGG-16 model starts with layer 1, which contains two convolution layers with an input size 224 × 224 with 64 filters. Layer 2 has two convolution layers with 112 × 112 input size with 128 filters. Layer 3 contains three convolution layers with a 56 × 56 input size and 256 filters. Layer 4 has three convolution layers with a 28 × 28 input size and 512 filters. Layer 5 contains three convolution layers with a 14 × 14 input size and 512 filters. Table 1 summarizes the VGG-16 layers with the input size for each layer, number of filters, kernel size, and stride.
Table 1. Summary of the VGG-16 layers.
The custom CNN has the same kernel or filter size for each convolution layer, 3 × 3 with stride, followed by the Maxpool layer with a kernel size of 2 × 2. The custom network starts with layer 1, which contains two convolution layers with input size of 224 × 224 with 7 and 9 filters, respectively. Layer 2 has two convolution layers with 112 × 112 input sizes with 16 and 32 filters, respectively. Layer 3 contains two convolution layers with a 56 × 56 input size and 256 filters. Layer 4 has three convolution layers with 28 × 28 input sizes and 32 and 64 filters, respectively. Layer 5 contains two convolution layers with 14 × 14 input sizes and 64 and 64 filters, respectively. Layer-6 has two convolution layers with input sizes of 7 × 7 with 128 and 128 filters, respectively. Table 2 summarizes the custom CNN layers with the input size for each layer, number of filters, kernel size, and stride.
Table 2. Summary of the custom network layers.

3.2.1. Image Input Layer

The proposed DCTN model starts by retrieving a size 224 × 224 × 3 input image. The original images in the dataset have a resolution of 512 × 512, and the preprocessing step adjusts the image sizes. Once the images are read from the dataset, they are resized and passed to the input layer for processing.

3.2.2. Convolution Layer

The convolution layer is the building block and most crucial component in CNN. It produces feature maps of the low-level and high-level features of the images. Low-level features are essential visual characteristics of an image, such as color, intensity, texture, and edges [50]. These features are typically extracted at the pixel level. In contrast, high-level features are more abstract, complex, and derived from low-level features. High-level features include object recognition, scene classification, and image retrieval based on low-level features such as edges or color [51]. There are three types of convolutions (1D, 2D, and 3D). We use 2D convolution with the same filter or kernel size (3.3), stride with 1, and padding value is the same in Keras. The 2D coevolution mathematical formula is Equation (1):
O n , m = I × F n , m = z c I n + z m + c × F z , c
where O(i, j) represents the output of input image I with a filter or kernel F. Each convolution layer is followed by an activation function, and there are multiple activation functions. There are linear and nonlinear activation functions. Since we are dealing with image classification, we use nonlinear activation functions. Several activation functions exist, such as ReLU, Leaky ReLU, Tanh, etc. We use the ReLU activation function after the convolution layer [52]. The ReLU function converts the output to 0 or 1, neglecting negative values. The following represents the mathematical formula for activation functions in Equations (2)–(4):
ReLU
f x = 1 , x 0 0 , x < 0
Leaky ReLU
f x = x , x 0 s c a l e × x , x < 0
Tanh
f x = 2 1 + e 2 x 1

3.2.3. Pooling Layer

After the convolution layer, the pooling layer removes unnecessary values and keeps the essential feature value (downsampling the convolution layer’s output). The training time for the network can be reduced by downsampling the output of the convolution. There are two types of pooling max and average [53]. We used max pooling: with a size of 2 × 2 filters. The following Equation (5) represents the max pooling process:
m a x r = 0 , , n , c = 0 , , m ( h i + r j + c )
where h is the output of convolution, and n and m are the height and width of convolution. After each CNN, we use 2D GlobalMaxPooling. The GlobalMaxPooling is similar to Maxpooling, except it performs downsampling by computing the maximum height and width in the whole area, while the max pool is in the subarea [46,54]. Figure 5 shows the difference between MaxPooling and GlobalMaxPooling.
Figure 5. Difference between Maxpooling and Globalmaxpooling.

3.2.4. Fully Connected Layer

The proposed model includes a fully connected layer after the convolutional layers, aggregating the features learned from multiple images. This layer then determines the most significant patterns for image classification purposes.

3.2.5. SoftMax Layer

The activation function is used for multiclass classification tasks. It takes the output of FC layers, computes the probability distribution, and produces a range of values between 0 and 1; the sum of probability for all predicted classes is 1. The SoftMax function is Equation (7).
o i = j W i , j X j   +   b i
where zi is the summation of the input weights with neuron value and bias, W is the weight, X is the neuron value, and b is the bias.
y = s o f t m a x = f ( o ) i = exp o i j exp o j
where oi is the predicted probability distribution over the categories.

3.2.6. Loss Function

The sparse categorical loss function is used in machine learning to classify multiple categories where a single integer represents the correct label. The function calculates the difference between the predicted probability distribution and the actual label. It is Equation (1). The function is commonly used in deep learning models for multiclass classification tasks, such as CNNs.
L y , f x = l o g ( f x y )
where y is the correct label and f(x) is the predicted probability distribution over the categories.

3.3. DCTN Training Parameters

We followed a trial-and-error approach to perform experiments and observe the different architectural performances. We constantly monitor the model’s performance during the train validation process for accuracy and converge the record of the results to track the improvement. The following parameters are shown in Table 3 in the DCTN model after conducting 22 experiments [55]. We used the Adam optimizer to train the proposed dual convolution tumor network with an initial and final learning rate of 0.0001, batch size of 64 images, and sparse-categorical-cross-entropy loss function. In the proposed DCTN model, we trained on 50 epochs for brain tumor classification to obtain the optimum results [56].
Table 3. Parameters used in training the DCTN model.

4. Dataset and Preprocessing

The dataset is publicly available on FigShare and has been used in many studies [57]. We preferred to use a public and well-known dataset to compare our model performance with state-of-the-art studies. The dataset has 3064 T1-weighted and contrast-enhanced MRI images from 233 patients with three different types of brain tumors: meningioma, glioma, and pituitary. The dataset images are in MATLAB format (*.mat files), and the data is organized into three subsets. The image resolution is 512 × 512, with voxel spacing of 0.49 × 0.49 mm2 and different RMI views: coronal (or frontal plane), transverse (or axial plane), and sagittal (or lateral plane).
According to Medicine Library [58], the coronal is “any vertical plane that divides the body into anterior and posterior (belly and back) sections”. Transverse is “any plane that divides the body into superior and inferior parts, roughly perpendicular to the spine”, and sagittal is “any imaginary plane parallel to the median plane”. Figure 6 represents the different body planes.
Figure 6. Different brain tumor types [57].
Each brain tumor type has images with different axial planes; for example, meningioma has (209 axial, 268 coronal, and 231 sagittal), glioma has (494 axial, 437 coronal, and 495 sagittal), and pituitary has (291 axial, 319 coronal, and 320 sagittal). The number of sample images for meningioma, glioma, and pituitary is 708, 1426, and 930, respectively. The following Table 4 summarizes the dataset sample of each brain tumor with the axial view.
Table 4. Dataset category description.

5. Results

This section serves as a concise overview of the research, offering a comprehensive analysis of the data and drawing conclusions based on the evidence presented. In addition, this section provides a benchmark for other researchers by comparing the findings with the current state-of-the-art in the field. Tables, graphs, and charts may be included to visualize the data, and statistical analysis is often provided to support the results [59]. The dataset was divided into 80% for training and 20% for testing; 80% has 2451 samples, and 20% has 613 samples. As shown in Figure 7, it illustrates the visualization of a percentage of the data used for training and testing purposes. There are many approaches used in measuring performance [60]. A confusion matrix is a tabular representation used to assess the performance of a classification model. It summarizes the number of correct and incorrect predictions made by the model, categorized by class [61].
Figure 7. Visualization of a percentage of data used for training and testing.
The rows represent the actual classes, while the columns represent the predicted classes. The diagonal elements show the correct predictions for each class, while the nondiagonal elements show incorrect predictions. Several metrics, such as accuracy, precision, recall, and F1 score, can be calculated based on the confusion matrix. As shown in Figure 8, it shows the confusion matrix for a brain tumor MRI from the proposed solution. The correct classifications for each meningioma, glioma, and pituitary tumor are 132, 282, and 193, respectively. The number of glioma samples is 1426, compared with 708 and 930 for meningioma and pituitary tumors, respectively.
Figure 8. Confusion matrix for brain tumor MRI.

Evaluation Metrics

To compare the results of the proposed model with those of the previous model, we need to analyze their performance on a typical set of benchmarks. By doing so, we can determine whether or not the proposed model outperforms the previous model. Performance measurements are crucial for evaluating the effectiveness of a classification model, allowing for comparisons between models to determine the best option. Common performance measurements include accuracy, precision, recall, F1-score, and area under the curve (AUC) [62]. Accuracy measures the proportion of correctly classified instances, as shown in Equation (7), while precision and recall measure the proportion of true positives among all positive predictions and actual positives, respectively. The F1 score balances precision and recall.
A c c u r a c y = T P + T N T P + F P + T N + F N
where these symbols indicate true positive (TP), true negative (TN), false positive (FP), and false negative (FN).
Equation (2) shows how to calculate precision:
P r e c i s i o n = T P T P + F P
Equation (3) presents the calculation of the recall value:
R e c a l l = T P T P + F N
The average between precision and recall is F1-measurement, which is calculated with Equation (4):
F 1 = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n × r e c a l l
The following Table 5 and Figure 9 note our classification results for precision, recall, f1-score, and accuracy.
Table 5. Classification report.
Figure 9. Classification performance for the proposed model.
The AUC chat metric comprehensively assesses the DCTM model’s performance by considering the trade-off between true positive and false positive rates. It measures the model’s ability to correctly distinguish between positive and negative instances in the dataset. A higher AUC value indicates better discriminatory power, suggesting the model achieves higher positive rates while minimizing false positives. This evaluation metric is beneficial in scenarios where imbalanced classes exist, ensuring that a dominant class does not skew the model’s performance. Consequently, the AUC chat enables a reliable comparison of different models and aids in identifying the optimal threshold for classification decisions. The AUC of the proposed model’s receiver operating characteristic (ROC) was achieved at 0.99 as shown in Figure 10, which illustrates the AUC chat for the DCTM model.
Figure 10. The proposed system evaluation.

6. Discussion

We compared our model with state-of-the-art models on the same dataset for consistent performance measurement. Numerous experiments have been conducted on the same dataset that we employed in our study; however, the models utilized varied, with some experiments using machine learning algorithms, such as SVM, Random Forest, Naïve Bayes, and KNN, while others relied on deep learning models based on their respective experimental requirements. Furthermore, there was a distinction in the type of classification employed, with some experiments utilizing binary classification (tumor or no tumor). In contrast, others employed multiclass classification (glioma, meningioma, pituitary tumor, and sometimes non-tumor is a plus) [63]. Notably, multiclass classification was the most commonly used method, consistent with our experimental approach. The following table compares our dual tumor convolution network results enhanced compared with other models in the literature as shown in Table 6.
Table 6. Comparison with our results and related work’s results.
By considering the range of models employed in previous studies, we aimed to comprehensively evaluate our model’s performance within the broader context of existing approaches. This approach allows for a better understanding of the strengths and limitations of our proposed method in comparison to both traditional machine learning algorithms and cutting-edge deep learning models used in the field.

7. Conclusions and Future Work

The Dual Convolution Tumor Network model is designed to achieve high accuracy in classification compared to previous models in the same research field. Initially, several architectures were considered, including EfficientNet, LetNet, Alex Net, and VIT, using different layers. Still, our proposed model, which consists of a pre-trained CNN model VGG16 and a custom convolutional network as the second layer, reached the highest accuracy of 99%. GlobalMaxPooling2D layers were added after each CNN layer before the fully connected layers, allowing for extracting important features and reducing the dimensionality of feature maps. The two CNNs were concatenated and followed by three fully connected layers. The first fully connected layer contained 1000 neurons and a dropout layer with a 0.3 dropping rate to address overfitting. The second fully connected layer contains 256 neurons, followed by another fully connected layer containing 64 neurons and a SoftMax output layer. The ReLU activation function is used in the network, and the loss function used is sparse and categorical. The experiments showed that the proposed dual convolution tumor network model achieved the highest accuracy compared to other state-of-the-art models.
Furthermore, in the future, we aim to propose a model for classifying glioma grades with an auto-segmentation feature used in image processing and computer vision, where an algorithm automatically segments or separates different regions of an image into distinct areas. Glioma grade classification is vital for guiding treatment decisions and predicting patient outcomes. Treatment options for gliomas include surgery, radiation therapy, chemotherapy, and targeted therapy, which vary depending on the tumor grade and other factors such as the patient’s age and overall health.

Author Contributions

Conceptualization, A.M.A.-Z. and A.I.E.; methodology, A.M.; software, A.M. and N.Y.; validation, A.M., N.Y., and E.M.K.A.-A.; formal analysis, A.M.; investigation, A.I.E.; resources, A.M.A.-Z.; data curation, A.M.; writing—original draft preparation, E.M.K.A.-A. and A.M.; writing—review and editing, A.M.A.-Z. and A.I.E.; supervision, A.M.A.-Z. and A.I.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

All data generated or analyzed during this study are included in this published article.

Conflicts of Interest

The authors declare there is no conflict of interest.

References

  1. Mohta, S.; Sharma, S.; Saraya, A. Improvement in Adipocytic Indices as a Predictor of Improved Outcomes after TIPS: Right Conclusion? Liver Int. 2022, 42, 253. [Google Scholar] [CrossRef]
  2. Liu, Y.; Guo, Q.-F.; Chen, J.-L.; Li, X.-R.; Hou, F.; Liu, X.-Y.; Zhang, W.-J.; Zhang, Y.; Gao, F.-F.; Zhang, Y.-Z.; et al. Correlations between Alterations of T-Helper 17 Cells and Treatment Efficacy after Concurrent Radiochemotherapy in Locally Advanced Cervical Cancer (Stage IIB–IIIB): A 3-Year Prospective Study. Chin. Med. J. 2021, 134, 954–962. [Google Scholar] [CrossRef] [PubMed]
  3. Wong, M.C.S.; Huang, J.; Chan, P.S.F.; Choi, P.; Lao, X.Q.; Chan, S.M.; Teoh, A.; Liang, P. Global Incidence and Mortality of Gastric Cancer, 1980–2018. JAMA Netw. Open 2021, 4, e2118457. [Google Scholar] [CrossRef] [PubMed]
  4. Pang, L.; Khan, F.; Heimberger, A.B.; Chen, P. Mechanism and Therapeutic Potential of Tumor-Immune Symbiosis in Glioblastoma. Trends Cancer 2022, 8, 839–854. [Google Scholar] [CrossRef]
  5. Nelson, S.J. Multivoxel Magnetic Resonance Spectroscopy of Brain Tumors. Mol. Cancer Ther. 2003, 2, 497–507. [Google Scholar]
  6. Pfister, S.M.; Reyes-Múgica, M.; Chan, J.K.C.; Hasle, H.; Lazar, A.J.; Rossi, S.; Ferrari, A.; Jarzembowski, J.A.; Pritchard-Jones, K.; Hill, D.A.; et al. A Summary of the Inaugural WHO Classification of Pediatric Tumors: Transitioning from the Optical into the Molecular Era. Cancer Discov. 2021, 12, 331–355. [Google Scholar] [CrossRef] [PubMed]
  7. Ullah, N.; Khan, J.A.; Khan, M.S.; Khan, W.; Hassan, I.; Obayya, M.; Negm, N.; Salama, A.S. An Effective Approach to Detect and Identify Brain Tumors Using Transfer Learning. Appl. Sci. 2022, 12, 5645. [Google Scholar] [CrossRef]
  8. Chen, Y.; Schonlieb, C.-B.; Lio, P.; Leiner, T.; Dragotti, P.L.; Wang, G.; Rueckert, D.; Firmin, D.; Yang, G. AI-Based Reconstruction for Fast MRI—A Systematic Review and Meta-Analysis. Proc. IEEE 2022, 110, 224–245. [Google Scholar] [CrossRef]
  9. Keall, P.J.; Brighi, C.; Glide-Hurst, C.; Liney, G.; Liu, P.Z.Y.; Lydiard, S.; Paganelli, C.; Pham, T.; Shan, S.; Tree, A.C.; et al. Integrated MRI-Guided Radiotherapy—Opportunities and Challenges. Nat. Rev. Clin. Oncol. 2022, 19, 458–470. [Google Scholar] [CrossRef] [PubMed]
  10. Matheus, M.G.; Castillo, M.; Smith, J.K.; Armao, D.; Towle, D.; Muenzer, J. Brain MRI Findings in Patients with Mucopolysaccharidosis Types I and II and Mild Clinical Presentation. Neuroradiology 2004, 46, 666–672. [Google Scholar] [CrossRef]
  11. Singhal, A.B.; Topcuoglu, M.A.; Koroshetz, W.J. Diffusion MRI in Three Types of Anoxic Encephalopathy. J. Neurol. Sci. 2002, 196, 37–40. [Google Scholar] [CrossRef]
  12. Monté-Rubio, G.C.; Falcón, C.; Pomarol-Clotet, E.; Ashburner, J. A Comparison of Various MRI Feature Types for Characterizing Whole Brain Anatomical Differences Using Linear Pattern Recognition Methods. Neuroimage 2018, 178, 753–768. [Google Scholar] [CrossRef] [PubMed]
  13. Cheng, W.; Ping, Y.; Zhang, Y.; Chuang, K.-H.; Liu, Y. Magnetic Resonance Imaging (MRI) Contrast Agents for Tumor Diagnosis. J. Healthc. Eng. 2013, 4, 23–46. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Zacharaki, E.I.; Wang, S.; Chawla, S.; Soo Yoo, D.; Wolf, R.; Melhem, E.R.; Davatzikos, C. Classification of Brain Tumor Type and Grade Using MRI Texture and Shape in a Machine Learning Scheme. Magn. Reson. Med. 2009, 62, 1609–1618. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Khan, F.; Pang, L.; Dunterman, M.; Lesniak, M.S.; Heimberger, A.B.; Chen, P. Macrophages and Microglia in Glioblastoma: Heterogeneity, Plasticity, and Therapy. J. Clin. Investig. 2023, 133, 446–448. [Google Scholar] [CrossRef]
  16. Tandel, G.S.; Balestrieri, A.; Jujaray, T.; Khanna, N.N.; Saba, L.; Suri, J.S. Multiclass magnetic resonance imaging brain tumor classification using artificial intelligence paradigm. Comput. Biol. Med. J. 2020, 112, 103804. [Google Scholar] [CrossRef]
  17. Goyal, L.M.; Saba, T.; Rehman, A.; Larabi-Marie-Sainte, S. Artificial Intelligence and Internet of Things; CRC Press: Boca Raton, FL, USA, 2021. [Google Scholar]
  18. Jena, B.; Saxena, S.; Nayak, G.K.; Balestrieri, A.; Gupta, N.; Khanna, N.N.; Laird, J.R.; Kalra, M.K.; Fouda, M.M.; Saba, L.; et al. Brain Tumor Characterization Using Radiogenomics in Artificial Intelligence Framework. Cancers 2022, 14, 4052. [Google Scholar] [CrossRef]
  19. Di Cataldo, S.; Ficarra, E. Mining Textural Knowledge in Biological Images: Applications, Methods and Trends. Comput. Struct. Biotechnol. J. 2017, 15, 56–67. [Google Scholar] [CrossRef] [Green Version]
  20. Zhou, X.S.; Huang, T.S. CBIR: From Low-Level Features to High-Level Semantics. SPIE 2000, 3974, 426–431. [Google Scholar]
  21. Alzubaidi, L.; Duan, Y.; Al-Dujaili, A.; Ibraheem, I.K.; Alkenani, A.H.; Santamaría, J.; Fadhel, M.A.; Al-Shamma, O.; Zhang, J. Deepening into the Suitability of Using Pre-Trained Models of ImageNet against a Lightweight Convolutional Neural Network in Medical Imaging: An Experimental Study. PeerJ Comput. Sci. 2021, 7, e715. [Google Scholar] [CrossRef]
  22. Noreen, N.; Palaniappan, S.; Qayyum, A.; Ahmad, I.; Alassafi, M.O. Brain Tumor Classification Based on Fine-Tuned Models and the Ensemble Method. Comput. Mater. Contin. 2021, 67, 3967–3982. [Google Scholar] [CrossRef]
  23. Garg, G.; Garg, R. Brain Tumor Detection and Classification Based on Hybrid Ensemble Classifier. arXiv 2021, arXiv:2101.00216. [Google Scholar] [CrossRef]
  24. Shafi, A.; Rahman, M.B.; Anwar, T.; Halder, R.S.; Kays, H.E. Classification of Brain Tumors and Auto-Immune Disease Using Ensemble Learning. Inform. Med. Unlocked 2021, 24, 100608. [Google Scholar] [CrossRef]
  25. Jena, B.; Nayak, G.K.; Saxena, S. An Empirical Study of Different Machine Learning Techniques for Brain Tumor Classification and Subsequent Segmentation Using Hybrid Texture Feature. Mach. Vis. Appl. 2022, 33, 6. [Google Scholar] [CrossRef]
  26. Ahsan, A.; Muhammad, A.; Usman, T.; Yunyoung, N.; Muhammad, N.; Chang-Won, J.; Reham, R.M.; Rasha, H.S. An Ensemble of Optimal Deep Learning Features for Brain Tumor Classification. CMC-Comput. Mater. Contin. 2021, 69, 2653–2670. [Google Scholar] [CrossRef]
  27. Zahoor, M.M.; Qureshi, S.A.; Bibi, S.; Khan, S.H.; Khan, A.; Ghafoor, U.; Bhutta, M.R. A New Deep Hybrid Boosted and Ensemble Learning-Based Brain Tumor Analysis Using MRI. Sensors 2022, 22, 2726. [Google Scholar] [CrossRef] [PubMed]
  28. Benyuan, S.; Jinqiao, D.; Zihao, L.; Cong, L.; Yi, Y.; Bo, B. GPPF: A General Perception Pre-Training Framework via Sparsely Activated Multi-Task Learning. arXiv 2022, arXiv:2208.02148. [Google Scholar] [CrossRef]
  29. Han, K.; Xiao, A.; Wu, E.; Enhua, J.; Xu, C.; Wang, Y. Transformer in Transformer. Adv. Neural Inf. Process. Syst. 2021, 34, 15908–15919. [Google Scholar]
  30. Wang, W.; Chen, C.; Ding, D.; Yu, H.; Zha, S.; Li, J. Transbts: Multimodal Brain Tumor Segmentation Using Transformer. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 109–119. [Google Scholar]
  31. Yurong, G.; Muhammad, A.; Ziaur, R.; Ammara, A.; Waheed Ahmed, A.; Zaheer Ahmed, D.; Muhammad Shoaib, B.; Zhihua, H. A Framework for Efficient Brain Tumor Classification Using MRI Images. Math. Biosci. Eng. 2021, 18, 5790–5815. [Google Scholar]
  32. Zhang, S.; Yuan, G.-C. Deep Transfer Learning for COVID-19 Detection and Lesion Recognition Using Chest CT Images. Comput. Math. Methods Med. 2022, 2022, 4509394. [Google Scholar] [CrossRef]
  33. Ansari, S.U.; Javed, K.; Qaisar, S.M.; Jillani, R.; Haider, U. Multiple Sclerosis Lesion Segmentation in Brain MRI Using Inception Modules Embedded in a Convolutional Neural Network. J. Healthc. Eng. 2021, 2021, 4138137. [Google Scholar] [CrossRef] [PubMed]
  34. Andrearczyk, V.; Oreiller, V.; Hatt, M.; Depeursinge, A. Head and Neck Tumor Segmentation and Outcome Prediction: Second Challenge, HECKTOR 2021, Held in Conjunction with MICCAI 2021, Strasbourg, France, 27 September 2021, Proceedings; Springer: Cham, Switzerland, 2022.
  35. Ranjbarzadeh, R.; Bagherian Kasgari, A.; Jafarzadeh Ghoushchi, S.; Anari, S.; Naseri, M.; Bendechache, M. Brain Tumor Segmentation Based on Deep Learning and an Attention Mechanism Using MRI Multi-Modalities Brain Images. Sci. Rep. 2021, 11, 10930. [Google Scholar] [CrossRef] [PubMed]
  36. Ismail, A.; Abdlerazek, S.; El-Henawy, I.M. Development of smart healthcare system based on speech recognition using support vector machine and dynamic time warping. Sustainability 2020, 12, 2403. [Google Scholar] [CrossRef] [Green Version]
  37. Lhotska, L.; Sukupova, L.; Lacković, I.; Ibbott, G.S. World Congress on Medical Physics and Biomedical Engineering 2018: June 3–8, 2018, Prague, Czech Republic (Vol.1); Springer: Singapore, 2019. [Google Scholar]
  38. Munir, K.; Frezza, F.; Rizzi, A. Deep Learning Hybrid Techniques for Brain Tumor Segmentation. Sensors 2022, 22, 8201. [Google Scholar] [CrossRef] [PubMed]
  39. Gómez-Guzmán, M.A.; Jiménez-Beristaín, L.; García-Guerrero, E.E.; López-Bonilla, O.R.; Tamayo-Perez, U.J.; Esqueda-Elizondo, J.J.; Palomino-Vizcaino, K.; Inzunza-González, E. Classifying Brain Tumors on Magnetic Resonance Imaging by Using Convolutional Neural Networks. Electronics 2023, 12, 955. [Google Scholar] [CrossRef]
  40. Gull, S.; Akbar, S.; Khan, H.U. Automated Detection of Brain Tumor through Magnetic Resonance Images Using Convolutional Neural Network. BioMed Res. Int. 2021, 2021, 3365043. [Google Scholar] [CrossRef]
  41. Latif, G.; Ben Brahim, G.; Iskandar, D.N.F.A.; Bashar, A.; Alghazo, J. Glioma Tumors’ Classification Using Deep-Neural-Network-Based Features with SVM Classifier. Diagnostics 2022, 12, 1018. [Google Scholar] [CrossRef]
  42. Díaz-Pernas, F.J.; Martínez-Zarzuela, M.; Antón-Rodríguez, M.; González-Ortega, D. A Deep Learning Approach for Brain Tumor Classification and Segmentation Using a Multiscale Convolutional Neural Network. Healthcare 2021, 9, 153. [Google Scholar] [CrossRef]
  43. Dequidt, P.; Bourdon, P.; Tremblais, B.; Guillevin, C.; Gianelli, B.; Boutet, C.; Cottier, J.-P.; Vallée, J.-N.; Fernandez-Maloigne, C.; Guillevin, R. Exploring Radiologic Criteria for Glioma Grade Classification on the BraTS Dataset. IRBM 2021, 42, 407–414. [Google Scholar] [CrossRef]
  44. Abd El Kader, I.; Xu, G.; Shuai, Z.; Saminu, S.; Javaid, I.; Salim Ahmad, I. Differential Deep Convolutional Neural Network Model for Brain Tumor Classification. Brain Sci. 2021, 11, 352. [Google Scholar] [CrossRef]
  45. Cheng, J.; Huang, W.; Cao, S.; Yang, R.; Yang, W.; Yun, Z.; Wang, Z.; Feng, Q. Correction: Enhanced Performance of Brain Tumor Classification via Tumor Region Augmentation and Partition. PLoS ONE 2015, 10, e0144479. [Google Scholar] [CrossRef] [PubMed]
  46. Jiang, Z.; Dong, Z.; Wang, L.; Jiang, W. Method for Diagnosis of Acute Lymphoblastic Leukemia Based on ViT-CNN Ensemble Model. Comput. Intell. Neurosci. 2021, 2021, 7529893. [Google Scholar] [CrossRef] [PubMed]
  47. Ali, M.B.; Gu, I.Y.-H.; Berger, M.S.; Pallud, J.; Southwell, D.; Widhalm, G.; Roux, A.; Vecchio, T.G.; Jakola, A.S. Domain Mapping and Deep Learning from Multiple MRI Clinical Datasets for Prediction of Molecular Subtypes in Low Grade Gliomas. Brain Sci. 2020, 10, 463. [Google Scholar] [CrossRef] [PubMed]
  48. Rehman, H.Z.U.; Hwang, H.; Lee, S. Conventional and Deep Learning Methods for Skull Stripping in Brain MRI. Appl. Sci. 2020, 10, 1773. [Google Scholar] [CrossRef] [Green Version]
  49. Mahmoud, A.T.; Awad, W.A.; Behery, G.; Abouhawwash, M.; Masud, M.; Aljuaid, H.; Ebada, A.I. An Automatic Deep Neural Network Model for Fingerprint Classification. Intell. Autom. Soft Comput. 2023, 36, 2007–2023. [Google Scholar] [CrossRef]
  50. Rathore, S.; Niazi, T.; Iftikhar, M.A.; Chaddad, A. Glioma Grading via Analysis of Digital Pathology Images Using Machine Learning. Cancers 2020, 12, 578. [Google Scholar] [CrossRef] [Green Version]
  51. Shaver, M.M.; Kohanteb, P.A.; Chiou, C.; Bardis, M.D.; Chantaduly, C.; Bota, D.; Filippi, C.G.; Weinberg, B.; Grinband, J.; Chow, D.S.; et al. Optimizing Neuro-Oncology Imaging: A Review of Deep Learning Approaches for Glioma Imaging. Cancers 2019, 11, 829. [Google Scholar] [CrossRef] [Green Version]
  52. Shen, Y.; Ma, Y.; Deng, S.; Huang, C.-J.; Kuo, P.-H. An Ensemble Model based on Deep Learning and Data Pre-processing for Short-Term Electrical Load Forecasting. Sustainability 2021, 13, 1694. [Google Scholar] [CrossRef]
  53. Riva, M.; Sciortino, T.; Secoli, R.; D’Amico, E.; Moccia, S.; Fernandes, B.; Conti Nibali, M.; Gay, L.; Rossi, M.; De Momi, E.; et al. Glioma biopsies Classification Using Raman Spectroscopy and Machine Learning Models on Fresh Tissue Samples. Cancers 2021, 13, 1073. [Google Scholar] [CrossRef]
  54. Kim, H.; Jeong, Y.S. Sentiment Classification Using Convolutional Neural Networks. Appl. Sci. 2019, 9, 2347. [Google Scholar] [CrossRef] [Green Version]
  55. Pal, K.; Mitra, K.; Bit, A.; Bhattacharyya, S.; Dey, A. Medical Signal Processing in Biomedical and Clinical Applications. J. Healthc. Eng. 2018, 2018, 3932471. [Google Scholar] [CrossRef]
  56. Bumes, E.; Wirtz, F.-P.; Fellner, C.; Grosse, J.; Hellwig, D.; Oefner, P.J.; Häckl, M.; Linker, R.; Proescholdt, M.; Schmidt, N.O.; et al. Non-Invasive Prediction of IDH Mutation in Patients with Glioma WHO II/III/IV Based on F-18-FET PET-Guided In Vivo 1H-Magnetic Resonance Spectroscopy and Machine Learning. Cancers 2020, 12, 3406. [Google Scholar] [CrossRef]
  57. BT. Data Set. Available online: https://figshare.com/articles/dataset/brain_tumor_dataset/1512427 (accessed on 16 April 2023).
  58. 1.2D: Body Planes and Sections. Medicine LibreTexts. Available online: https://med.libretexts.org/Courses/Okanagan_College/HKIN_110%3A_Human_Anatomy_I_for_Kinesiology/01%3A_Introduction_to_Anatomy/1.02%3A_Mapping_the_Body/1.2D%3A_Body_Planes_and_Sections (accessed on 16 April 2023).
  59. Nwankpa, C.; Ijomah, W.; Gachagan, A.; Marshall, S. Activation Functions: Comparison of Trends in Practice and Research for Deep Learning. arXiv 2018, arXiv:1811.03378. [Google Scholar] [CrossRef]
  60. Yaqub, M.; Feng, J.; Zia, M.S.; Arshid, K.; Jia, K.; Rehman, Z.U.; Mehmood, A. State-of-the-Art CNN Optimizer for Brain Tumor Segmentation in Magnetic Resonance Images. Brain Sci. 2020, 10, 427. [Google Scholar] [CrossRef] [PubMed]
  61. Ghaffari, M.; Sowmya, A.; Oliver, R. Automated Brain Tumour Segmentation Using Multimodal Brain Scans, a Survey Based on Models Submitted to the BraTS 2012–2018 Challenges. IEEE Rev. Biomed. Eng. 2019, 13, 156–168. [Google Scholar] [CrossRef] [PubMed]
  62. Rani, G.; Tiwari, P.K. Handbook of Research on Disease Prediction through Data Analytics and Machine Learning; IGI Global, Medical Information Science Reference: Hershey, PA, USA, 2021. [Google Scholar]
  63. Deepak, S.; Ameer, P.M. Brain Tumor Classification Using Deep CNN Features via Transfer Learning. Comput. Biol. Med. 2019, 111, 103345. [Google Scholar] [CrossRef] [PubMed]
  64. Zulfiqar, F.; Ijaz Bajwa, U.; Mehmood, Y. Multiclass Classification of Brain Tumor Types from MR Images Using EfficientNets. Biomed. Signal Process. Control 2023, 84, 104777. [Google Scholar] [CrossRef]
  65. Zahoor, M.M.; Khan, S.H. Brain Tumor MRI Classification Using a Novel Deep Residual and Regional CNN. arXiv 2022, arXiv:2211.16571. [Google Scholar] [CrossRef]
  66. Verma, A.; Singh, V.P. Design, Analysis and Implementation of Efficient Deep Learning Frameworks for Brain Tumor Classification. Multimed. Tools Appl. 2022, 81, 37541–37567. [Google Scholar] [CrossRef]
  67. Vidyarthi, A.; Agarwal, R.; Gupta, D.; Sharma, R.; Draheim, D.; Tiwari, P. Machine Learning Assisted Methodology for Multiclass Classification of Malignant Brain Tumors. IEEE Access 2022, 10, 50624–50640. [Google Scholar] [CrossRef]
  68. Molder, C.; Lowe, B.; Zhan, J. Learning Medical Materials from Radiography Images. Front. Artif. Intell. 2021, 4, 638299. [Google Scholar] [CrossRef] [PubMed]
  69. Yerukalareddy, D.R.; Pavlovskiy, E. Brain Tumor Classification Based on Mr Images Using GAN as a Pre-Trained Model. In Proceedings of the 2021 IEEE Ural-Siberian Conference on Computational Technologies in Cognitive Science, Genomics and Biomedicine (CSGB), Yekaterinburg, Russia, 26–28 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 380–384. [Google Scholar]
  70. Hilles, S.M.S.; Saleh, N.S. Image Segmentation and Classification Using CNN Model to Detect Brain Tumors. In Proceedings of the 2021 2nd International Informatics and Software Engineering Conference (IISEC), Ankara, Turkey, 16–17 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–7. [Google Scholar]
  71. Chaki, J.; Woźniak, M. A Deep Learning Based Four-Fold Approach to Classify Brain MRI: BTSCNet. Biomed. Signal Process. Control 2023, 85, 104902. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.