Abstract
In medical diagnostics, brain tumor classification remains essential, as accurate and efficient models aid medical professionals in early detection and treatment planning. Deep learning methodologies for brain tumor classification have gained popularity due to their potential to deliver prompt and precise diagnostic results. This article proposes a novel classification technique that integrates the Xception model with a hybrid attention mechanism and progressive image resizing to enhance performance. The methodology is built on a combination of preprocessing techniques, transfer learning architecture reconstruction, and dynamic fine-tuning strategies. To optimize key hyper-parameters, this study employed the Dynamic Chaotic Whale Optimization Algorithm. Additionally, we developed a novel learning rate scheduler that dynamically adjusts the learning rate based on image size at each training phase, improving training efficiency and model adaptability. Batch sizes and layer freezing methods were also adjusted according to image size. We constructed an ensemble approach by preserving models trained on different image sizes and merging their results using weighted averaging, bagging, boosting, stacking, blending, and voting techniques. Our proposed method was evaluated on benchmark datasets achieving remarkable accuracies of 99.67%, 99.09%, and 99.67% compared to the classical algorithms.
1. Introduction
The brain is the most vital organ in the human body, which is enclosed by a hard skull. Brain tumors are one of the biggest challenges of modern medicine since they are so complex and carry such a high fatality rate. In brain tumors, there is abnormal growth of brain cells, and any abnormal growth in this confined space can lead to severe complications []. Brain tumors are either benign or malignant and are also classified into two major categories: primary and metastatic []. Primary brain tumors originate from the brain tissue or its surrounding environment [], while metastatic tumors occur in other parts of the body and metastasize to the brain via the bloodstream. Primary brain tumors can be benign or malignant, whereas metastatic brain tumors are always malignant [].
Malignant tumors grow rapidly and can invade adjacent brain tissue, whereas benign tumors usually grow more slowly. Nonetheless, benign tumors also possess serious risks through the compression of surrounding tissues []. Approximately 70% of brain tumors are benign, and 30% are malignant []. Different primary brain tumors have been identified, including gliomas, astrocytomas, meningiomas, and pituitary tumors []. More than 120 different types of brain tumors have been discovered to date, with the most common being meningioma, glioma, and pituitary tumors. Among them, meningioma is a tumor arising from the meninges, covering membranes of the brain and spinal cord, and is hence one of the most frequently encountered primary intracranial tumors []. Gliomas, however, arise from glial cells named astrocytes. Astrocytomas, one variety of gliomas, are generally low-grade and indolent. Yet high-grade gliomas are among the most virulent brain tumors. Pituitary tumors occur due to the uncontrolled growth of cells in the pituitary gland. Brain tumors need to be identified early since they can be life threatening []. Metastatic brain tumors, the most common form of malignant brain tumors, develop most commonly from cancers of the lung, breasts, or skin (melanoma). To invade the brain, these cancer cells must penetrate the blood–brain barrier (BBB) and acclimatize to the brain-specific microenvironment by secreting cytokines, chemokines, and other mediators to help reorganize the tumor microenvironment (TME) [].
In contemporary medical practice, imaging modalities are being used more by radiologists as they are accurate and pose fewer hazards to patients. For acquiring medical imaging data, various methods have been employed, such as radiography, MRI, tomography, and echocardiography. Among these, MRI is shown to be more beneficial as it provides high-quality images without radioactive exposure. Being a non-surgical technique, MRI gives radiologists useful insights into medical images that contribute to diagnosing brain abnormalities [,]. Radiologists mainly utilize MRI scans to detect brain tumors. But with the exponential increase in medical data, it has become ever more difficult to examine and interpret such large data using traditional techniques [,]. As diagnosis accuracy, to a great extent, is reliant on the experience of the radiologist, computer-based diagnostic support is in high demand for simplifying the process [].
Computer-aided diagnosis (CAD) systems offer an encouraging solution where the early detection of brain tumors can be achieved with minimal human interaction. CAD systems generate diagnostic reports from MRI scans and provide additional guidance to radiologists []. The classification of brain tumors in CAD includes two main steps: feature extraction and classification. The integration of machine learning (ML) and deep learning (DL) has enhanced the CAD process considerably in medical imaging [,,]. Traditional machine learning methods are feature extraction, selection, and classification-based. Feature extraction was widely popular in the early artificial intelligence era, and the methods employed were the gray level co-occurrence matrix (GLCM), bag-of-words (BoW), Gabor filters, thresholding, clustering, contour-based, and intensity histograms [,]. These methods have the tendency to discard valuable information from the source image []. Such conventional ML-based approaches are less suitable for image classification as they heavily depend on manually designed features []. What makes them particularly well-suited for complex biomedical applications is that they are able to automatically extract optimized features from raw data without the necessity of any human intervention, improving classification performance by tailoring the feature extraction process to the problem in question [].
The new breakthrough in deep learning (DL) has resolved the limitations of traditional machine learning (ML) feature extraction using the original image as input. This has put convolutional neural networks (CNNs) into the spotlight, which automatically extract the desired features for image classification using a number of convolutional layers [,]. DL is an advanced approach for prediction and classification with strong performance on problems involving the multi-level processing of information, such as detection, classification, and voice recognition [,].
However, CNNs work optimally with large number data, which are not always available in the case of medical imaging []. Transfer learning is one widely used technique to circumvent this problem. In transfer learning [,], a pre-trained model on a large dataset from another but related problem domain is adapted for the target classification task []. Through the utilization of this pre-trained knowledge, the model is capable of high accuracy even when it has a limited dataset []. Transfer learning also conserves computation, since the model benefits from the pre-trained network’s convolutional weights, with the last dense layer being the only one that needs to be trained [].
The hyperparameters of a CNN play an important role in determining its performance, as they impact the training process of the network. Identifying the best hyperparameters, for instance, learning rate, neurons and layers, batch size, dropout rate, and regularization strength, can significantly enhance CNN accuracy. Manually determining the best values, however, is not a simple task; it requires expertise and time. To address this, metaheuristic techniques are applied to discover the best hyperparameter configurations so that the performance of CNNs is improved [,]. The tunability of hyperparameters is one of the primary issues of contemporary machine and deep learning because differing initial values could profoundly influence performance metrics like accuracy []. Some researchers have turned to metaheuristic algorithms for optimizing deep learning hyperparameters. For instance, ref. [] applied Grey Wolf Optimization (GWO) to CNN hyperparameter optimization in the context of skin cancer classification. Similarly, ref. [] investigated certain metaheuristic algorithms like Genetic Algorithm (GA), Whale Optimization Algorithm (WOA), Multiverse Optimizer (MVO), Satin Bower Optimization (SBO), and Life Choice-Based Optimization (LCBO) to improve CNN performance in the detection of breast cancer abnormalities. Furthermore, ref. [] applied Particle Swarm Optimization (PSO) to the hyperparameter optimization of CNN to identify early lung cancer.
WOA has also been applied in optimizing CNN hyperparameters since it can efficiently explore complex search spaces. According to the study of [], the WOA has traits of simplicity and effectiveness in addressing such optimization issues specifically. Ref. [] demonstrated its ability to enhance neural network performance through the optimization of connection weights, which also has the potential to optimize CNNs. In response, ref. [] showcased WOA’s convergent stability in water demand estimation for water resources, reinforcing its applicability in dynamically tuning CNN hyperparameters. However, like most other metaheuristic algorithms, the WOA is plagued by a slow convergence rate. To solve this, chaos theory has been incorporated into the optimization process, enhancing the global convergence rate and overall performance. In the Chaotic Whale Optimization Algorithm (CWOA), various chaotic maps are explored to regulate WOA’s crucial parameters for balancing exploration and exploitation []. The traditional CWOA, however, relies on a specific chaotic map while optimizing, which might limit its flexibility in different problem environments. The behavior of chaotic maps may vary depending on the optimization landscape, and therefore, a single map may not always have the optimal search performance. To overcome this limitation, the Dynamic Chaotic Whale Optimization Algorithm (DCWOA) was utilized, introducing a dynamic process where the algorithm oscillates between two different chaotic maps rather than adhering to a static one throughout. This dynamic alternation increases the flexibility of the algorithm and helps it navigate various search spaces more effectively.
In this paper, we present an approach involving the combination of the Xception model and hybrid attention, progressive image resizing, and a Dynamic Chaotic Whale Optimization Algorithm (DCWOA) for hyperparameter optimization. We have a dynamic learning rate scheduler that dynamically adjusts the learning rate during runtime depending on the size of images, with batch sizes and layer-unfreezing schedules that are designed to adaptively optimize training. Several versions of models trained on various image sizes are combined using methods such as weighted averaging, bagging, boosting, stacking, and voting. This integrated method was tested on three datasets and achieved state-of-the-art accuracy, exemplifying its power and applicability in medical diagnostics.
The key contributions of this research can be summarized as follows:
- Enhanced the Xception model by integrating hybrid attention mechanisms to enhance the ability of feature extraction and emphasize critical regions in the brain tumor images. Additionally, a progressive image resizing strategy was incorporated, enabling models to learn effectively by gradually increasing image sizes during training.
- Developed a dynamic chaotic variant of the Whale Optimization Algorithm (DCWOA) to optimize key hyperparameters, including the learning rate, the number of neurons, optimizer type, activation function, and dropout rate, ensuring superior model performance and stability.
- Proposed an adaptive learning rate scheduler that dynamically modifies the learning rate according to the image size, facilitating efficient training across different learning phases and preventing premature convergence.
- Designed an intelligent batch size selection and layer-unfreezing strategy aligned with image size variations, enhancing training efficiency, convergence speed, and model adaptability to diverse resolutions.
- Established a comprehensive framework integrating advanced preprocessing techniques, transfer learning-based model reconstruction, fine-tuning, and ensemble strategies, offering a reliable and scalable approach for practical brain tumor diagnosis.
2. Related Works
Classifying brain tumors is crucial for early detection and effective treatment planning. Advances in deep learning, especially CNNs, have greatly improved accuracy and automation. Techniques like transfer learning and ensemble methods further enhance performance, with studies confirming the success of CNN-based models applied in MRI-based tumor categorization for high precision and robustness in clinical settings [,]. This section reviews recent studies highlighting deep learning’s impact on brain tumor categorization.
A study in [] proposed the RBP-CNN technique for brain tumor categorization, incorporating gray standard normalization (GSN) for preprocessing, regional binary pattern (RBP) for feature extraction, and a CNN for training, achieving 96% accuracy. Similarly, ref. [] proposed a multi-level attention network (MANet) for brain tumor identification. The model fuses both cross-channel and spatial attention to augment tumor region discovery without sacrificing temporal dependencies of the semantic features captured from the Xception backbone. MANet produced top accuracies of 94.91% as well as 96.51% on the BraTS and Figshare datasets, respectively.
To improve brain tumor classification, ref. [] developed two deep learning models, a 23-layer CNN and a fine-tuned VGG16, which effectively identified brain abnormalities and tumor grades with a high accuracy of 97.8%, surpassing state-of-the-art models. However, the refined VGG16 was trained, evaluated, and tested on only one dataset, which undermined its generalizability. Ref. [] proposed a CNN-based method using the EfficientNetV2B0 architecture (developed by Google Research, Mountain View, USA) with data preprocessing and transfer learning and achieved an impressive classification accuracy of 99.16%, accompanied by recall, high precision, and F1 score. While this shows the efficacy of EfficientNetV2B0, this study could have explored other models of the EfficientNetV2 family for a more comprehensive comparison.
The work of [] proposed a method whereby features of brain MRI scans from CNNs were distinguished using an SVM, outperforming the CNN using a SoftMax classifier and an accuracy of 95.82%. They found that SoftMax-based CNN classifiers suffered from overfitting due to small dataset sizes. But pre-trained models may have been fine-tuned to address this deficiency. SVMs may also suffer from scalability constraints when handling large datasets and high-dimensional feature sets resulting from CNNs. Similarly, ref. [] suggested an automated brain tumor identification system utilizing a triple-module architecture based on the integration of known pre-trained deep neural networks, PCA, and a Random Forest classifier. This method, even though it leverages deep learning for feature learning, does not make much use of pre-trained models for end-to-end learning. Moreover, PCA dimension reduction could lead to the loss of significant features, which are critical in medical images to correctly classify the tumors.
A pretrained approach using ResNet152, DenseNet169, VGG19, and MobileNetv3 was suggested in the study of [] mainly for brain tumor classification, in which MobileNetv3 achieved the highest accuracy of 98.52%. However, validation and testing on the same data result in data leakage, the inflation of performance metrics, and the failure to assess model generalization correctly. A clear division into training, validation, and test datasets would mitigate this issue. Similarly, ref. [] suggested an ensemble deep learning model by combining a shallow CNN (SCNN) for low-level feature extraction and VGG16 for deeper feature extraction. By feature fusion of the two models to minimize information loss, they achieved 97.77% accuracy. In addition, ref. [] suggested a privacy-preserving federated learning technique for the categorization of brain tumor based on a customized VGG16 structure to achieve decentralized model training without compromising data security.
The research proposed in [] utilizes a dilated PDCNN model with several preprocessing techniques and an average ensemble method for categorizing the brain tumors from the MRI scans. The model is completely trained over three datasets, and the accuracy rate is more than 98%, which reflects the success of the model. The average ensemble method, however, may not give good results if models are non-diverse. With weighted averaging or stacking, performance can be improved as strengths can be merged while reducing weaknesses. The research project in [] foresees an optimization-facilitated hybrid deep learning approach to identify as well as classify brain tumors in the utilized MRI images using the technique exponential deer hunting optimization-based Shepard CNN (ExpDHO-based ShCNN) and also exponential deer hunting optimization-based deep CNN (ExpDHO-based Deep CNN). In their study, ref. [] utilizes seven new deep learning networks for tumor classification and the detection of tumors in the brain like DenseNet121, InceptionResNetV2, ResNet50V2, Xception, InceptionV3, as well as VGG19 and EfficientNetB7, where maximum accuracy is obtained by InceptionResNetV2 above 97%. However, the use of ensemble methods like stacking, weighted average, or bagging to fuse the predictions belonging to individual trained models can enhance robustness by leveraging the strengths of each model. A combined Quantum Dilated Convolutional Neural Network and Deep Maxout Network (QDCNN-DMN) is proposed by the study of [] for classifying brain tumors together with increased specificity, accuracy, and sensitivity than traditional techniques. Similarly, ref. [] presented a WHHO-based deep CNN for the identification of brain tumors using statistical and texture features extracted from MRI images, and optimizing a deep CNN with the WHHO algorithm. The model possessed a best accuracy of 0.816 and sensitivity of 0.974 over other approaches. However, statistical and texture features may lack vital information, whereas feature extraction using deep learning can improve performance by learning important patterns from MRI data directly.
3. Materials and Methods
This section offers a summary of the available resources, experimental setup, including the methodologies used in this study. We utilized a combination of datasets, computational tools, and software frameworks to conduct our experiments. The methods employed cover the complete process, from data preprocessing down to model optimization, ensuring a well-structured and efficient approach to achieving our research goals. Every step was carefully designed and systematically executed to guarantee reliability. Special attention was given to selecting the most appropriate techniques, ensuring that the results obtained are meaningful. Python 3.7.10 (Python Software Foundation, Wilmington, DE, USA) was used, and the model was run on the Kaggle cloud platform (Kaggle Inc., San Francisco, CA, USA).
3.1. Data Source
Three publicly available datasets, Figshare [], SARTAJ [], and Br35H [], containing grayscale and JPG-formatted MRI scans of the human brain were used in this research. The datasets for all three are representative of the three (3) types of brain tumors: meningioma, glioma, and pituitary tumors. The dataset of Figshare contains 3064 images distributed in 708 meningioma, 930 pituitary, and 1426 glioma cases. Similarly, the SARTAJ dataset holds 3264 images that were labeled as 926 glioma, 937 meningioma, 901 pituitaries, and 500 as no tumor. Br35H is a binary dataset with 3000 images, half of them being 1500 tumorous and 1500 non-tumorous brain MRI scans. These multi-class datasets are a good foundation for the evaluation of classification models used in this work.
3.2. Data Partitioning Framework
The datasets employed in this current study were split into three distinct subsets in order to accommodate efficient model creation and testing. Specifically, 80% of each dataset was utilized in training, with the remaining 20% devoted to testing purposes. For even better model performance and prevention against overfitting, the test subset was once again split, with 50% devoted to validation and 50% dedicated to final testing. However, the entire training, validation, and testing procedure was conducted using the TensorFlow 2.11.0 toolkit, taking full advantage of computational resources offered at the Kaggle platform. To prevent data leakage, this study ensured that all data splits were performed on a per-patient basis. This means that images belonging to a single patient were never distributed across training, validation, or test sets. We also used a stratified splitting approach to maintain the class distribution across the different subsets. Furthermore, each dataset was handled independently, and the proposed model was trained separately on each of them to account for differences in class structure and format. This approach ensured that no overlap or contamination occurred between datasets, and patient-level independence was strictly preserved.
3.3. Data Preprocessing Pipeline
To supply the model with a consistent and high-quality input, a complete preprocessing pipeline was implemented to address image size, format, and noise variability in the datasets. The images were cropped to contain just the brain region, which is considered the most significant region to analyze. This was achieved using the popular OpenCV to discover and crop the ultimate top, including bottom, left, as well as right edges of every image. To enhance the saliency of local information and improve feature extraction, contrast-limited adaptive histogram equalization (CLAHE) was employed to boost image contrast. Subsequently, a bilateral filter was applied to reduce noise, particularly from low-contrast areas amplified by CLAHE, without degrading image edges. Progressive image resizing was employed in this research as an integral aspect of the model architecture. This approach employed the step-by-step training of the model using resized images. Each image was resized individually to dimensions of 100 × 100, 128 × 128, 244 × 244, 256 × 256, and 299 × 299 to accommodate the requirements of various training sessions. Resizing was performed as the final step in the preprocessing pipeline to maintain aspect ratio, preserve image features, and avoid loss of salient features. Data augmentation was applied only to two datasets, excluding the Figshare dataset, to tackle the issue of lack of adequately labeled medical image data. No augmentation techniques were applied on Figshare dataset because, during the training phase of the models, we observed the impact of data augmentation on each dataset individually. Our results showed that the application of augmentation techniques (such as rotation, flipping, and scaling) did not lead to any performance improvement on the Figshare dataset. This may be attributed to the dataset’s existing variability and quality, which possibly made further augmentation redundant. However, in contrast, both the SARJAT and Br35H datasets demonstrated improved classification performance when augmentation was applied. This improvement was evident in terms of increased validation accuracy and reduced overfitting, particularly for the smaller and more class-imbalanced Br35H dataset. This technique introduced a huge variety to the training set by adding brightness adjustment, rotation, shearing, horizontal flip, zooming, and width and height shifting. Augmentation was conducted on training subsets after partitioning the datasets to ensure that the CNN model learned invariant features and could differentiate variations. The robustness and model performance improved, and overfitting and underfitting dangers were minimized using this technique. Table 1 provides the data augmentation parameters, and Figure 1 presents the preprocessing flowchart. The preprocessing techniques are also evaluated as shown in Table 2 and Figure 2.
Table 1.
Data augmentation strategy.
Figure 1.
Data preprocessing pipeline.
Table 2.
Preprocessing evaluation.
Figure 2.
Histogram before and after CLAHE.
3.4. Optimization Algorithm
Optimization methods are instrumental to improving deep learning models by finding optimal hyperparameters that improve performance and avoid overfitting. For this study, we employed a variant of the WOA motivated by the well-known humpback whales’ foraging behavior to improve critical hyperparameters that are used by the Xception model architecture. The algorithm was tailored to adjust parameters such as learning rate, activation function, dropout rate, the first dense layer’s number of neurons, and optimizer. The process takes advantage of the capability of the WOA towards efficiently exploring and exploiting the entire search space so as to obtain optimal hyperparameter selection and better model performance.
3.4.1. Whale Optimization Algorithm
The WOA was introduced in []. It is a meta-heuristic optimization method motivated by the hunting habit of humpback whales. The algorithm replicates the bubble net foraging technique, where whales encircle their prey using a combination of circular and spiral bubble patterns to simulate the trapping mechanism. By diving deep and creating spiral-shaped bubbles, the whales effectively contain the prey, preventing its escape. This unique hunting strategy forms the foundation of WOA’s optimization process, as it alternates between chasing prey based on either random or optimal positions and employing a spiral motion to enhance exploitation [].
3.4.2. Encircling Prey
According to the study of [], the Humpback whales find and surround targets, and the WOA mimics this by assuming the best current candidate solution represents or is close to the optimal approach. After identifying the best search agent, the other agents update their places toward it. This behavior is illustrated in Equations (1) and (2).
Here, indicates the present iteration, also along with act as coefficient vectors, denotes the position vector of the optimal solution found so far, and serves as the position vector. The | | symbol represents the value that is absolute, while dot (.) indicates element-wise multiplication. However, it is important to note that is modified in each of the iterations whenever an optimal solution is discovered.
The vectors along with are computed in Equations (3) and (4):
where is gradually reduced from 2 down to 0 throughout the iterations in the exploration and exploitation stages, and stands as a random vector within the range [0, 1].
3.4.3. Bubble-Net Attacking Method
To mathematically model the bubble-net feeding habit of the humpback whales, the following two strategies are proposed []:
- Shrinking encircling mechanism: This behavior is achieved by gradually decreasing the value of a from 2 to 0 throughout the iterations, as illustrated in Equation (3). The new position of the search agent is determined by selecting the random values for within the range of , which enables the agent to be placed anywhere between its starting position and the position of the current best agent.
- Spiral updating position: The equation of the spiral, which simulates the helix-shaped movement of humpback whales between the prey’s position and the whale’s position, is expressed in Equation (5).
represents the distance in between the whale, along with the prey (current optimal approach or solution). Here, serves as a constant that determines the structure of the logarithmic spiral, while similarly is a random number within , and dot (.) signifies element-wise multiplication.
Humpback whales encircle their prey by combining a shrinking circular motion with a spiral path. To represent this behavior, a 50% probability was assigned to choosing either the shrinking encircling mechanism or the spiral model when updating whale positions during optimization using Equation (6).
where p serves as a random number in the range [0, 1].
3.4.4. Search for Prey
As stated by [], the variation in the vector was utilized to facilitate the exploration phase, allowing the questing for prey. By setting to arbitrary values exceeding 1 or falling below −1, search agents were prompted to diverge from a reference whale. The mathematical representation of this phase is formulated as in Equations (7) and (8).
represents a randomly selected position vector (a randomly chosen whale) from the current population.
The pseudocode for the Whale Optimization Algorithm (WOA) is outlined in Algorithm 1 [,].
| Algorithm 1. Pseudo-code of the WOA | |||||
| Calculate the fitness of each search agent | |||||
| = the best search agent | |||||
| while (t < maximum number of iterations) | |||||
| for each search agent | |||||
| Update | |||||
| if1 (p < 0.5) | |||||
| if2 (|A| < 1) | |||||
| Update the position of the current search agent by the Equation (1) | |||||
| else if2 (|A| ≥ 1) | |||||
| Select a random search agent | |||||
| Update the position of the current search agent by Equation (8) | |||||
| end if2 | |||||
| else if1 (p ≥ 0.5) | |||||
| Update the position of the current search by the Equation (5) | |||||
| end if1 | |||||
| end for | |||||
| Check if any search agent goes beyond the search space and amend it | |||||
| Calculate the fitness of each search agent | |||||
| Update if there is a better solution | |||||
| end while | |||||
| return | |||||
3.4.5. Chaotic Whale Optimization Algorithm (CWOA)
The WOA, which is based on the hunting behavior of humpback whales, is a commonly studied algorithm over the past few years due to its simplicity and capability to address challenging optimization problems []. However, as with any other metaheuristic technique, WOA is prone to premature convergence and limited diversity within the population and stagnation at the local optimum, particularly when applied to high-dimensional and complex problems such as hyperparameter tuning in deep learning models []. To overcome such challenges, ref. [] suggested CWOA-based solutions where chaotic systems are applied instead, to substitute the primary parameters belonging to WOA that assists in flipping the exploration and exploitation capability of WOA. In an effort to enhance performance, chaos was incorporated in most of the algorithms of meta-heuristic, which leads to a better convergence rate and helps avoid getting trapped in local optima []. CWOA is an advanced variant of the WOA, which was introduced to boost the convergence rate and optimization precision of the traditional WOA with chaotic maps. Chaos theory, while being deterministic yet non-predictable, has different exploration capabilities through the avoidance of premature convergence as well as enhancing the algorithm’s capacity to escape from local optima. Various chaotic maps, such as the Logistic map, Tent map, Cubic map, Sine map, Singer map, etc., have been incorporated into the traditional WOA to improve its exploration and exploitation []. As stated by [], the pseudocode for the CWOA is presented in Algorithm 2.
| Algorithm 2. Pseudo-code of CWOA | ||||||||
| Initialize | ||||||||
| Evaluate | ||||||||
| Initialize randomly | ||||||||
| while (t < maximum number of iterations) | ||||||||
| Update the chaotic number using the respective chaotic map equation | ||||||||
| for each search agent | ||||||||
| Update a, A, C, l and p | ||||||||
| if1 (p < 0.5) | ||||||||
| if2 (|A| < 1) | ||||||||
| Update the position of the current search agent by the Equation (1) | ||||||||
| else if2 1) | ||||||||
| Select a random search agent (Xrand) | ||||||||
| Update the position of the current search agent by Equation (8) | ||||||||
| end if2 | ||||||||
| else if1 0.5) | ||||||||
| Update the position of the current search by the Equation (5) | ||||||||
| end if1 | ||||||||
| end for | ||||||||
| Check if any search agent goes beyond the search space and amend it | ||||||||
| Calculate the fitness of each search agent | ||||||||
| Update if there is a better solution | ||||||||
| end while | ||||||||
| return | ||||||||
3.4.6. Dynamic Chaotic Whale Optimization Algorithm (DCWOA)
The traditional CWOA utilizes a fixed chaotic map throughout the optimization process, which may limit its adaptability in different problem landscapes. The effectiveness of a chaotic approach can also vary based on the optimization landscape, and a single chaotic map may not always provide the best search dynamics. To address this, we propose a Dynamic Chaotic Whale Optimization Algorithm (DCWOA), which dynamically alternates between different chaotic maps, such as logistic and sine map [] based on the optimization performance. Some other metaheuristic algorithms have also been applied to optimization and feature selection tasks [,]. In this study, Chaotic Whale Optimization Algorithm (CWOA) was selected as the base for hyperparameter tuning because of its adoption in solving optimization problems. CWOA has been utilized in several studies due to its ability to enhance the convergence behavior of the original Whale Optimization Algorithm by introducing chaotic dynamics. Refs. [,,,,,,] have utilized the CWOA method in their respective studies. The objective is to investigate and utilize the impact of chaotic map switching on the effectiveness of the WOA when hyperparameters are tuned in deep learning models.
While traditional methods such as grid search and Bayesian optimization are indeed effective for relatively small or simple hyperparameter spaces, they tend to struggle with high-dimensional and highly non-linear search spaces like the one addressed in our study. Grid search, for instance, becomes computationally prohibitive as dimensionality increases, while Bayesian optimization can be prone to local optima and often lacks the global search capacity needed for complex neural architectures. In contrast, our proposed Dynamic Chaotic Whale Optimization Algorithm (DCWOA) dynamically alternates between different chaotic maps, which enhances its ability to explore the search space broadly and avoid premature convergence. This adaptability and stochastic behavior enable DCWOA to perform robust hyperparameter tuning in deep learning models, offering a level of flexibility and exploratory strength that traditional methods may not easily provide. Additionally, WOA and CWOA, which are the foundation of the DCWOA, have shown promising results in recent studies involving complex optimization scenarios, which further justifies its application in our methodology.
3.4.7. Performance-Based Switching Strategy
The switching mechanism is triggered based on the convergence behavior of the optimization process. A stagnation counter was used to track the number of consecutive iterations where the best fitness (i.e., the loss function in deep learning) remains unchanged. If no improvement is observed for five consecutive iterations, the chaotic map is switched. The updated map governed the chaotic sequence until another stagnation phase was detected. The algorithm alternated between two chaotic maps: logistic map and sine map, depending on the improvement in the best fitness value. Logistic and sine maps were selected through brute-force testing, as they showed the best balance leading to improved convergence in the optimization process. Initially, the logistic map was used as the default chaotic function, defined as in Equation (9):
where represents the chaotic sequence at iteration , and is a control parameter, typically set to 4 for full chaos []. This map ensures a high degree of randomness and exploration during the initial stages of optimization. To monitor the effectiveness of the search process, DCWOA tracks the number of consecutive iterations without improvement in the best fitness value. If no improvement is observed for five consecutive iterations, the algorithm switches from the logistic map to the sine map, as defined in Equation (10):
If improvement resumes, the algorithm resets the stagnation counter and continues with the currently selected chaotic map. The pseudocode of the CWOA is shown in Algorithm 3.
| Algorithm 3. Pseudo-code of DCWOA | ||||||||
| Initialize | ||||||||
| Evaluate | ||||||||
| Initialize randomly | ||||||||
| Set no_improvement_count = 0 and previous_best_fitness = + | ||||||||
| while (t < maximum number of iterations) | ||||||||
| Update the chaotic number using the respective chaotic map equation | ||||||||
| if | ||||||||
| switch from the logistic map to the sine map | ||||||||
| Reset no_improvement_count = 0 after switching | ||||||||
| end if | ||||||||
| for each search agent | ||||||||
| Update a, A, C, l and p based on the chaotic values | ||||||||
| if (p < 0.5) | ||||||||
| if (|A| < 1) | ||||||||
| Update the position of the current search agent by the Equation (1) | ||||||||
| else if 1) | ||||||||
| Select a random search agent (Xrand) | ||||||||
| Update the position of the current search agent by Equation (8) | ||||||||
| end if | ||||||||
| else if 0.5) | ||||||||
| Update the position of the current search by the Equation (5) | ||||||||
| end if | ||||||||
| end for | ||||||||
| Check if any search agent goes beyond the search space and amend it | ||||||||
| Calculate the fitness of each search agent | ||||||||
| If previous_best_fitness) | ||||||||
| increment no_improvement_count by 1 | ||||||||
| Else | ||||||||
| set previous_best_fitness = leader_fitness | ||||||||
| Reset no_improvement_count = 0 | ||||||||
| end if | ||||||||
| Update if there is a better solution | ||||||||
| end while | ||||||||
| return | ||||||||
3.5. Proposed Custom Learning Rate Scheduler
The learning rate is a critical hyperparameter in deep learning that dictates the step size in weight updates during training. The choice of significantly impacts the convergence speed and generalization ability of neural networks. In this work, we propose a custom learning rate scheduler that adapts dynamically based on the image size in progressive resizing strategies. The proposed approach ensures smooth adjustments in while maintaining stable training dynamics. The learning rate adjustment is governed by an exponential scaling function that considers an image-dependent factor to modulate in . The relationship is formulated as Equation (11):
where is the adjusted learning rate, is the base learning rate, represents the image size, and is the scaling coefficient, defined as Equation (12):
Larger image sizes were assigned higher learning rates to accelerate convergence and adapt efficiently to increased feature details. This prevents overfitting at lower resolutions while ensuring stable and efficient training at higher resolutions.
To prevent extreme fluctuations, we impose the following constraints on in Equation (13),
where the adjusted learning rate is clamped within an interval of to , ensuring stability across varying image resolutions. The workflow of the proposed custom learning rate is outlined in Algorithm 4.
| Algorithm 4. Custom learning rate scheduler | |
| Input: Image Size ( | |
| Output: | |
| Define for different image sizes: | |
| Retrieve scaling factor for given image size : | |
| If then | |
| Else, Set (default scaling factor) | |
| Compute the adjusted learning rate: | |
| Apply constraint to maintain stability: | |
| Return | |
3.6. Proposed Model
In this study, we suggest an enhanced brain tumor categorization model based on progressive image resizing and the Xception architecture, optimized using the DCWOA. To enhance feature extraction and classification output, we adjust the Xception pretrained model by eliminating its classification layer and integrating a hybrid attention mechanism, which combines squeeze-and-excitation (SE) blocks and CBAM spatial attention. This is followed by global average pooling (GAP) and batch normalization to stabilize training. To enhance feature learning and regularization, the modified architecture incorporated three fully connected (dense) layers, each with 128 neurons, except for the first dense layer, where the number of neurons was dynamically optimized by DCWOA. Each dense layer was followed by batch normalization and dropout. Additionally, DCWOA was employed to optimize key hyperparameters, including the dropout rate, optimizer (Adam, RMSprop, AdamW, Adamax, or Nadam), initial learning rate, and activation function (ReLU, Tanh, or Sigmoid). For classification, a concluding dense layer was incorporated, utilizing a SoftMax activation function for multi-class datasets, including four neurons for the Sartaj dataset and three neurons for the Figshare dataset, using categorical cross-entropy as the loss function. For the BR35H dataset, a binary classification setup was employed, with a single neuron, sigmoid activation, and binary cross-entropy loss function. The fully fine-tuned Xception model is displayed in Figure 3.
Figure 3.
Finetuned Xception pretrained model.
The training process employed a progressive image resizing method, whereby the model was trained sequentially with progressively larger image sizes: 100, 128, 244, 256, and 299. In each step, the model was trained independently at the given image size, and the weights learned were transferred to the subsequent training step with the subsequent image size. The image sizes were selected based on several practical and methodological considerations. First, we took into account the minimum and maximum input size requirements of the Xception model, which supports input dimensions from 71 × 71 up to 299 × 299. Second, we referenced commonly used image sizes in the recent literature, particularly in brain tumor classification and other medical imaging tasks, where sizes such as 128 × 128, 224 × 224, 256 × 256, and 299 × 299 are frequently employed due to their balance between performance and computational efficiency. Third, we intentionally selected a spread of sizes to support our progressive resizing strategy and to ensure that the ensemble would include models trained at various resolutions, capturing both coarse and fine-grained features. However, other resolutions could also yield competitive results, but our chosen set offers a representative range from low to high resolution. In addition, Xception pretrained layers are unfrozen step by step with a gradually larger image size, allowing the model to learn more and more abstract representations gradually. For promoting efficient convergence, we introduce a learning rate scheduler, which is specially designed to change the learning rate depending on the image size of the input batch, helping the model to learn its training process adaptively. In an attempt to enhance prediction accuracy, generalization, and resilience when trained on different sizes of images, we apply ensemble learning techniques. Specifically, models learned using different image sizes were ensembled using various ensemble strategies, i.e., bagging, boosting, stacking, blending, voting, weighted voting, and weighted ensemble optimized by DCWOA. Finally, the predictions from these ensembles were further hybridized for best classification performance. The complete structure of the proposed model is illustrated in Figure 4. The utilized preprocessing pipeline includes CLAHE for contrast enhancement, bilateral filtering for denoising, and OpenCV cropping to isolate the brain region, which reduces variability caused by artifacts and image borders. Progressive resizing and adaptive training strategies help mitigate differences in resolution. The ensemble approach, especially when optimized using DCWOA, enhances generalization by combining the strengths of diverse models. Each model may handle different types of distortions more effectively, and the dynamic weighting mechanism assigns higher influence to models that perform better under specific conditions. These combined techniques ensure robust performance across heterogeneous datasets.
Figure 4.
The proposed model.
3.7. Performance Evaluation
The effectiveness of the proposed model was evaluated as suggested using four standard classification metrics: recall, F1 score, precision, and accuracy [,], as defined in Equations (14)–(17), respectively. These metrics offer a thorough assessment of the model’s performance in accurately classifying brain tumor images.
4. Results
This section presents findings of experimental testing of the proposed method. Comparisons are based on major performance indicators to measure the efficiency of the method. Comparative evaluation with existing methods is also performed to showcase the improvements achieved.
4.1. Hyperparameter Optimization Based on DCWOA
In this section, the tuned hyperparameters of various parameters utilized in order to optimize the effectiveness of our deep brain tumor classification model are described. We utilized the DCWOA as an efficient exploration strategy in the hyperparameter space that can identify the best setting by optimizing model generalization and convergence. The DCWOA adaptively adjusted its search behavior by changing between two chaotic maps (logistic map and sine map) when it hit a stagnation point to promote a more diverse and adaptive exploration concerning search space. However, the parameter settings for blending DCWOA with the Xception model are presented in Table 3. Each hyperparameter performs a vital role in enhancing the operation of the Xception model architecture. The optimal iteration count for the optimization algorithm is fixed at 20, while the size of the population, or the total count of potential solutions, is set at 10. The choice of 20 iterations with 10 candidate solutions in the DCWOA-based hyperparameter optimization was made to balance computational efficiency and optimization performance. Given the high training cost of deep models like Xception, a limited search space was necessary. Empirical results showed that the DCWOA reliably converged to stable and effective configurations within these constraints, making further expansion unnecessary. Five hyperparameters are used by the optimization process in simultaneous tuning. The learning rate, which influences weight adjustments during training, is restricted to the interval of 1 × 10−4, 1 × 10−2. The dropout rate, responsible for network regularization, varies between 0.2 and 0.5. The optimizer, which adjusts network weights to minimize error or loss, is selected from Adam, RMSprop, AdamW, Adamax, and Nadam. The activating function, which presents non-linearity to enable complex pattern learning, is chosen from ReLU, Tanh, and Sigmoid. Additionally, the count of neurons in the initial dense layer, governing the hidden layer’s size, is adjustable between 256 and 1024. To find the total count of training epochs for Xception, multiple values were tested, and 40 epochs were selected as the optimal setting. Table 4 presents the optimal hyperparameter values identified by the DCWOA across the three popular datasets used: Br35H, SARTAJ, and Figshare. This study did not use a fixed random seed. Instead, this study allowed multiple stochastic initializations to observe the stability and robustness of the algorithm across different runs. This variability was crucial in evaluating the sensitivity of DCWOA to randomness and ensuring that the observed performance improvements were not due to a favorable initialization.
Table 3.
Parameter configurations for using DCWOA with Xception and the corresponding dataset.
Table 4.
Optimal hyperparameter values determined using DCWOA.
4.2. Training the Xception Model Using Optimized Hyperparameters
Here, we present the training process of the proposed model, following progressive image resizing. Scripts for all were implemented in Python 3.7.10 on a Jupyter Notebook 6.3.0 hosted on the Kaggle cloud with a disk capacity of 57.6 GiB, 29 GiB of RAM, and a GPU with 16 GiB of memory using the Keras 2.4.3 and TensorFlow 2.4.1 platforms. The optimized Xception model with the optimal hyperparameters set by the DCWOA was utilized as a method for extracting features during training and assessed on the test data for 80 epochs. While the training used all the optimized hyperparameters fixed, the learning rate was updated dynamically after each epoch based on the image size using the proposed custom learning rate scheduler for performance improvement. Batch sizes were adapted based on image size to prevent memory overflow and maintain training speed. Furthermore, the model layers were progressively unfrozen for increasing image size. The adaptive mechanism ensures that task-specific features are reinforced by fine tuning without losing effective pretrained representations. To further enhance performance, the weights of each trained model at a specific image size were saved and used to initialize the succeeding models for the next image size. Categorical cross-entropy was applied to the Figshare and SARTAJ datasets, and binary cross-entropy was applied to the Br35H dataset. The entire process is presented in Algorithm 5. The proposed model was trained independently on each dataset to handle their respective classification tasks. Specifically, the Br35H dataset, being a binary classification problem, was treated separately from the Figshare and SARTAJ datasets, which are multiclass classification problems. Each dataset was preprocessed, trained, and evaluated individually, and the model architecture and hyperparameters were adjusted accordingly to suit the dataset’s structure. This ensures that the results reported for each dataset are valid and reflect the model’s ability to adapt to different classification challenges.
| Algorithm 5. Training workflow of the proposed model | ||
| Input | ||
| Output: Predicted Class Label C | ||
| Initialize pretrained Xception model M | ||
| Remove classification layer of M | ||
| Integrate hybrid attention mechanism (SE Block + CBAM Spatial Attention) | ||
| Append: | ||
| Global Average Pooling (GAP) | ||
| Batch Normalization | ||
| Three Dense layers of 128 neurons (except first dense layer, which is optimized) | ||
| Batch Normalization (BN) followed by Dropout after each dense layer | ||
| Classification layer | ||
| Softmax activation with 4 neurons (Sartaj dataset) or 3 neurons (Figshare dataset) | ||
| Sigmoid activation with 1 neuron for binary BR3H5 dataset | ||
| Initialize hyperparameter search using DCWOA to optimize: | ||
| Count of neurons in the initial dense layer | ||
| Dropout rate | ||
| For do: | ||
| Set | ||
| Unfreeze additional layers of M based on S | ||
| Train on dataset resized to | ||
| Save | ||
| Load into next stage model | ||
| Perform ensemble learning on models trained at different sizes using: | ||
| Bagging, Boosting, Stacking, Blending, Voting, Weighted Voting, Weighted Ensemble with DCWOA | ||
| Hybridize ensemble outputs to obtain final classification model | ||
| Classification: | ||
| Compute for multi-class datasets. | ||
| Predict as: | ||
| For binary classification, apply a threshold (e.g., 0.5) to the sigmoid output | ||
| End Algorithm | ||
4.3. Evaluating the Performance of the Proposed Model
This very section presents a rigorous testing of the effectiveness of the model proposed at various stages with the improvement achieved by using the proposed techniques. The testing is performed on the model trained on various sizes of images and various ensemble techniques such as stacking, bagging, boosting, blending, voting, weighted voting, and a weighted ensemble optimized by the DCWOA. In addition, a hybrid approach that combines all ensemble methods is explored and tested. These performance measures are applied to each of the three datasets used in this study to achieve a comprehensive performance analysis. The model performance was assessed utilizing precision, accuracy, F1 score, recall, and AUC as performance measures.
4.3.1. Model Performance Based on Image Sizes
To assess the impact of image size (resolution) on classification performance, the model was trained and tested on progressively larger image resolutions: 100 × 100, 128 × 128, 244 × 244, 254 × 254, and 299 × 299. This incremental resize strategy traded computation for feature-extraction ability. At lower resolution (100 × 100, 128 × 128), the model processed images more rapidly but less detailed, generating lower accuracy. As image size increased, the model recognized increasingly accurate structural forms, enhancing classification accuracy. The 299 × 299 resolution used the maximum feature representation, leveraging the Xception architecture to the full extent. The performance was in proportion, with the lowest accuracy acquired at 87.91%, 81.60%, and 93.00% for 100 × 100 images, and the highest accuracy acquired was 98.37%, 97.55%, and 98.33% for 299 × 299 images for the three datasets. The results, in Table 5, Table 6 and Table 7, emphasize the impact of image resolution on model performance and validate progressive resizing as a valuable method for improving classification performance.
Table 5.
Evaluation results based on the Figshare dataset for each of the image resolutions.
Table 6.
Evaluation results based on the SARTAJ dataset for each of the image resolutions.
Table 7.
Evaluation results based on the Br35H dataset for each of the image resolutions.
4.3.2. Model Performance Based on Ensemble Techniques
To further enhance categorization performance, the suggested model was experimented with different ensemble techniques, including bagging, boosting, stacking, blending, voting, weighted voting, and a weighted ensemble optimized using the DCWOA. All of these methods employ different strategies to increase model stability and predictive capability. Bagging decreases variance by averaging predictions from models trained independently, while boosting gradually enhances weak learners by concentrating on misclassified examples. Stacking and blending stack different models using a meta-learner for the final prediction, while voting aggregates multiple models’ output, with weighted voting assigning higher weightage to higher-performing models. Trained with the DCWOA, the weighted ensemble uses the dynamic weighing of model contributions to classify maximally. On three datasets, performance validated the strengths of these ensemble approaches. The lowest accuracy achieved utilizing the Figshare dataset was 98.04%, while the highest accuracy of 99.67% was reached using the weighted ensemble with the DCWOA. For the SARTAJ dataset, the lowest accuracy was 97.55% using bagging, while boosting had the highest accuracy of 99.09%. For the Br35H dataset, accuracy values were quite high on average, i.e., ranging from 96.33% to 99.67%, with most of the ensemble methods reaching the upper bound. To further assess model reliability, confusion matrices (Figure 5) were utilized to analyze classification errors, unmasking true and false predictions. Figure 5a (Figshare) illustrates performance for three-class tumor classification (glioma, meningioma, and pituitary tumor). The model had great accuracy and correctly classified nearly all cases: 71/71 meningiomas, 142/142 gliomas, and 92/93 pituitary tumors. One pituitary tumor was incorrectly classified as a meningioma, with good model performance and little class confusion. Figure 5b (SARTAJ) shows a four-class classification with both tumor cases and no-tumor cases (glioma tumor, meningioma tumor, pituitary tumor, and no tumor). The model classified almost all cases correctly with minimal misclassifications: one glioma misclassified as no tumor, one meningioma misclassified as glioma, and one pituitary tumor misclassified as meningioma. The no-tumor class showed perfect prediction (50/50), which is essential to reduce false alarms during clinical screening. Figure 5c (Br35H) illustrates a binary distinction between no-tumor and tumor samples. The model was nearly perfect in terms of accuracy, correctly classifying 149/150 no-tumor images and 150/150 tumor images. Only one false positive was reported. The confusion matrices in total confirm that the model generalizes extremely well for different tasks of classification, with immense diagnostic potential, particularly in minimizing false negatives, which is of utmost importance in medical applications. In addition, the ROC curves (Figure 6) reveal the discriminative capacity of the model on datasets and accuracy along with the loss curves. Figure 7a–c show training stability and convergence trends. The results, as summarized in Table 8, Table 9 and Table 10, reflect the impact of ensemble learning and optimization in enhancing classification outcomes for Figshare, Br35H, and SARTAJ datasets, respectively. However, the hybrid approach that combines all ensemble techniques did not yield any further improvement in accuracy. The highest classification performance was achieved using individual ensemble methods for all three datasets, indicating that certain ensemble strategies, such as a weighted ensemble optimized by the DCWOA and boosting, were more effective in enhancing model accuracy than a generalized hybrid combination.
Figure 5.
Confusion matrix of the suggested technique based on the three datasets.
Figure 6.
ROC of the suggested technique based on the three datasets.
Figure 7.
Curves for training and validation accuracy and loss.
Table 8.
Evaluation results based on the Figshare dataset for each of the image ensembles.
Table 9.
Evaluation results based on the Br35H dataset for each of the image ensembles.
Table 10.
Evaluation results based on the SARTAJ dataset for each of the image ensembles.
Our proposed classification model, like many deep learning systems, could be vulnerable to several security threats, particularly adversarial and poisoning attacks. For instance, attackers may manipulate input data or training samples in subtle ways to alter the model’s predictions without detection, a concern thoroughly explored in the work by [], which demonstrates how even small perturbations can compromise diagnostic models. Additionally, as healthcare increasingly integrates AI into Internet-of-Things (IoT) frameworks, concerns around data integrity, authentication, and transmission security also emerge. Refs. [,] in studies highlight the risks of model inversion, evasion attacks, and unauthorized data access, especially in resource-constrained or remote environments. Moreover, ref. [] underscores the need for secure and efficient AI deployment in medical monitoring systems, especially where models may be exposed to edge-based or cloud-based threats. While our current work focuses primarily on model accuracy and interpretability, we acknowledge that future extensions should include robust defense mechanisms such as adversarial training, input verification, and secure model deployment strategies to guard against these emerging security threats.
4.4. Comparison of the Proposed Model with Other State-of-the-Art Models in the Literature
In this section, we provide a comparative analysis of our model alongside recently developed state-of-the-art approaches. As evident from Table 11, the proposed model achieved the highest accuracy of 99.67%, which is better compared to most existing methods. The papers were selected based on their relevance to our problem domain, as they address similar challenges and their reflection of recent advancements in the field. Additionally, they use the same dataset as our study, ensuring a consistent and fair comparison of model performance.
Table 11.
Comparative analysis.
4.5. Addressing Data Imbalance
There is a presence of imbalance in class distribution, which can indeed influence model training and evaluation if not properly addressed. To mitigate the impact of this imbalance, we applied data augmentation to the SARTAJ and Br35H datasets, which helped artificially expand the minority classes and reduce potential bias. Although augmentation was not applied to the Figshare dataset due to earlier experimentation showing no improvement, we employed ensemble techniques to all three datasets, which are known to be robust against class imbalance by combining diverse learners and correcting biases of individual models through averaging or majority voting. Importantly, we did not rely solely on accuracy as a performance metric. Instead, we reported Precision, Recall, and F1 score, which are more informative and widely accepted in the literature for evaluating models trained on imbalanced datasets. These metrics allowed us to assess performance across all classes, including minority ones.
4.6. Model Evaluation and Interpretability Using Grad-CAM
To assess the internal decision making of our proposed deep learning model and increase its clinical explainability, we utilized Gradient-weighted Class Activation Mapping (Grad-CAM). Grad-CAM is a common visualization tool that produces class-discriminative heatmaps, enabling the detection of spatial areas within the input image that contribute most significantly to the predictions of the model. In our study, we applied Grad-CAM to both tumor-positive and tumor-negative brain MRI slices to assess whether the model focuses on clinically relevant anatomical structures during classification. For tumor-positive cases, the Grad-CAM heatmaps consistently highlighted regions corresponding to tumor masses. These areas aligned well with known tumor locations by comparing the original images and GradCAM images. For tumor-negative cases, the heatmaps displayed minimal and diffuse activation, with no strong localized focus. This suggests that the model does not falsely fixate on normal structures when no tumor is present, which is indicative of a low false-positive tendency. An example visualization is shown in Figure 8, where the left panel depicts the original MRI image and the right panel overlays the Grad-CAM heatmap. Bright regions in the heatmap (red/yellow) indicate high model attention, clearly corresponding to the tumor area in positive cases. Such interpretability is crucial for building trust in AI-assisted diagnostics, especially in sensitive applications such as neuro-oncology.
Figure 8.
Crad-CAM visualization.
5. Conclusions
This paper suggested a more advanced brain tumor identification model using progressive image resizing, a hybrid attention-based architecture of modified Xception, and an optimized ensemble learning approach. Through progressively improving image resolution in training, the model effectively extracts both high-level and low-level features with improved classification accuracy. The addition of Squeeze-and-Excitation (SE) blocks and CBAM spatial attention enhances feature extraction, while the DCWOA optimizes hyperparameters for better generalizability. Experimental findings reveal that higher image resolution is accountable for enhanced classification accuracy, where peak performance is achieved at 299 × 299 resolution. Ensemble learning approaches also indicate a significant enhancement of model resilience with the top-performing model obtained with the DCWOA-optimized weighted ensemble. The weighted ensemble optimized using the DCWOA outperformed other ensemble methods in the Figshare dataset, achieving up to 99.67% accuracy. It dynamically assigns weights to individual models based on performance. Boosting also outperformed other ensemble methods, particularly on the SARTAJ dataset. Hybrid combinations provided marginal improvements or redundancy, highlighting that performance gains depend on model diversity and dataset characteristics. This research substantiates the effectiveness of the introduced approach towards the enhanced automated detection of brain tumors. This proposed approach improves the precision of brain tumor classification, thereby facilitating more accurate automated diagnosis. Progressive resizing, attention mechanisms, and ensemble learning enhance feature extraction and model stability. Hyperparameter optimization with the DCWOA ensures improved generalization, positioning the model more suitable for real-world medical applications. These findings provide a basis for future AI-powered advancements in medical imaging and clinical decision support. Despite high accuracy, the model was trained and tested on publicly available datasets, which may not reflect real-world clinical variability. Potential risks include overfitting, data distribution shift, and reliance on biased datasets. From an ethical perspective, decisions made by AI systems must be transparent, explainable, and overseen by clinical experts. In clinical settings, our model should serve as a decision-support tool, not a replacement for human judgment.
6. Future Works
The future works can look at the integration of transformer-based model architectures, such as Vision Transformers (ViTs), to enhance feature representation and improve classification efficiency. An investigation of other pretrained models can be looked into, with the possibility of complementary feature extraction capabilities, further enhancing the model’s robustness. Future work will also investigate case-specific error patterns, including the visual inspection of false positives/negatives, and assess their clinical implications to better understand model limitations and areas for improvement
Author Contributions
Conceptualization, A.T.I. and M.A.; data curation, A.F.D.K., A.H.A. and M.K.R.; formal analysis, I.H.H.; funding acquisition, L.A.G.; investigation, A.F.D.K.; methodology, A.T.I., I.H.H. and M.T.M.; project administration, M.A.; resources, L.A.G. and M.K.R.; software, A.F.D.K.; supervision, M.A.; validation, M.T.M.; visualization, A.T.I.; writing—original draft, A.T.I., I.H.H., A.H.A. and H.C.; writing—review and editing, M.T.M., L.A.G. and H.C. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.
Acknowledgments
We acknowledged Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2025R178), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia, for the APC support.
Conflicts of Interest
The authors declare no conflicts of interest to report regarding the present study.
Correction Statement
This article has been republished with a minor correction to the existing affiliation information. This change does not affect the scientific content of the article.
References
- Mondal, A.; Shrivastava, V.K. A novel Parametric Flatten-p Mish activation function based deep CNN model for brain tumor classification. Comput. Biol. Med. 2022, 150, 106183. [Google Scholar] [CrossRef] [PubMed]
- Hill, C.I.; Nixon, C.S.; Ruehmeier, J.L.; Wolf, L.M. Brain tumors. Phys. Ther. 2002, 82, 496–502. [Google Scholar] [CrossRef] [PubMed]
- Lapointe, S.; Perry, A.; Butowski, N.A. Primary brain tumours in adults. Lancet 2018, 392, 432–446. [Google Scholar] [CrossRef] [PubMed]
- Foreman, P.M.; Jackson, B.E.; Singh, K.P.; Romeo, A.K.; Guthrie, B.L.; Fisher, W.S.; Riley, K.O.; Markert, J.M.; Willey, C.D.; Bredel, M.; et al. Postoperative radiosurgery for the treatment of metastatic brain tumor: Evaluation of local failure and leptomeningeal disease. J. Clin. Neurosci. 2018, 49, 48–55. [Google Scholar] [CrossRef]
- Khan, M.S.I.; Rahman, A.; Debnath, T.; Karim, M.R.; Nasir, M.K.; Band, S.S.; Mosavi, A.; Dehzangi, I. Accurate brain tumor detection using deep convolutional neural network. Comput. Struct. Biotechnol. J. 2022, 20, 4733–4745. [Google Scholar] [CrossRef]
- Behin, A.; Hoang-Xuan, K.; Carpentier, A.F.; Delattre, J.Y. Primary brain tumours in adults. Lancet 2003, 361, 323–331. [Google Scholar] [CrossRef]
- Louis, D.N.; Perry, A.; Reifenberger, G.; Von Deimling, A.; Figarella-Branger, D.; Cavenee, W.K.; Ohgaki, H.; Wiestler, O.D.; Kleihues, P.; Ellison, D.W. The 2016 World Health Organization classification of tumors of the central nervous system: A summary. Acta Neuropathol. 2016, 131, 803–820. [Google Scholar] [CrossRef]
- Cacho-Díaz, B.; García-Botello, D.R.; Wegman-Ostrosky, T.; Reyes-Soto, G.; Ortiz-Sánchez, E.; Herrera-Montalvo, L.A. Tumor microenvironment differences between primary tumor and brain metastases. J. Transl. Med. 2020, 18, 1. [Google Scholar] [CrossRef]
- Mabray, M.C.; Cha, S. Advanced MR imaging techniques in daily practice. Neuroimaging Clin. 2016, 26, 647–666. [Google Scholar] [CrossRef]
- Gudigar, A.; Raghavendra, U.; San, T.R.; Ciaccio, E.J.; Acharya, U.R. Application of multiresolution analysis for automated detection of brain abnormality using MR images: A comparative study. Future Gener. Comput. Syst. 2019, 90, 359–367. [Google Scholar] [CrossRef]
- Abiwinanda, N.; Hanif, M.; Hesaputra, S.T.; Handayani, A.; Mengko, T.R. Brain tumor classification using convolutional neural network. In Proceedings of the World Congress on Medical Physics and Biomedical Engineering 2018, Prague, Czech Republic, 3–8 June 2018; Springer: Singapore, 2019; Volume 1, pp. 183–189. [Google Scholar]
- Badža, M.M.; Barjaktarović, M.Č. Classification of brain tumors from MRI images using a convolutional neural network. Appl. Sci. 2020, 10, 1999. [Google Scholar] [CrossRef]
- Chen, Y.; Shao, Y.; Yan, J.; Yuan, T.F.; Qu, Y.; Lee, E.; Wang, S. A feature-free 30-disease pathological brain detection system by linear regression classifier. CNS Neurol. Disord.-Drug Targets (Former. Curr. Drug Targets-CNS Neurol. Disord.) 2017, 16, 5–10. [Google Scholar] [CrossRef] [PubMed]
- Chen, Y.; Yang, M.; Chen, X.; Liu, B.; Wang, H.; Wang, S. Sensorineural hearing loss detection via discrete wavelet transform and principal component analysis combined with generalized eigenvalue proximal support vector machine and Tikhonov regularization. Multimed. Tools Appl. 2018, 77, 3775–3793. [Google Scholar] [CrossRef]
- Wang, S.H.; Zhan, T.M.; Chen, Y.; Zhang, Y.; Yang, M.; Lu, H.M.; Wang, H.N.; Liu, B.; Phillips, P. Multiple sclerosis detection based on biorthogonal wavelet transform, RBF kernel principal component analysis, and logistic regression. IEEE Access 2016, 4, 7567–7576. [Google Scholar] [CrossRef]
- Joseph, A.A.; Abdullahi, M.; Junaidu, S.B.; Ibrahim, H.H.; Chiroma, H. Improved multi-classification of breast cancer histopathologica images using handcrafted features and deep neural network (dense layer). Intell. Syst. Appl. 2022, 14, 200066. [Google Scholar]
- Cheng, J.; Huang, W.; Cao, S.; Yang, R.; Yang, W.; Yun, Z.; Wang, Z.; Feng, Q. Enhanced performance of brain tumor classification via tumor region augmentation and partition. PLoS ONE 2015, 10, e0140381. [Google Scholar] [CrossRef]
- Jiang, J.; Wu, Y.; Huang, M.; Yang, W.; Chen, W.; Feng, Q. 3D brain tumor segmentation in multimodal MR images based on learning population-and patient-specific feature sets. Comput. Med. Imaging Graph. 2013, 37, 512–521. [Google Scholar] [CrossRef]
- Leng, W.Y.; Shamsuddin, S.M. Writer identification for Chinese handwriting. Int. J. Adv. Soft Comput. Appl. 2010, 2, 142–173. [Google Scholar]
- Le, E.P.V.; Wang, Y.; Huang, Y.; Hickman, S.; Gilbert, F.J. Artificial intelligence in breast imaging. Clin. Radiol. 2019, 74, 357–366. [Google Scholar] [CrossRef]
- Kiranyaz, S.; Avci, O.; Abdeljaber, O.; Ince, T.; Gabbouj, M.; Inman, D.J. 1D convolutional neural networks and applications: A survey. Mech. Syst. Signal Process. 2021, 151, 107398. [Google Scholar] [CrossRef]
- Ibrahim, A.T.; Abdullahi, M.; Kana, A.F.D.; Mohammed, M.T.; Hassan, I.H. Categorical classification of skin cancer using a weighted ensemble of transfer learning with test time augmentation. Data Sci. Manag. 2025, 8, 174–184. [Google Scholar] [CrossRef]
- Maijeddah, U.I.; Yusuf, S.A.; Abdullahi, M.; Hassan, I.H. A hybrid transfer learning model with optimized SVM using honey badger optimization algorithm for multi-class lung cancer classification. Sci. World J. 2024, 19, 977–986. [Google Scholar] [CrossRef]
- Pyrkov, T.V.; Slipensky, K.; Barg, M.; Kondrashin, A.; Zhurov, B.; Zenin, A.; Pyatnitskiy, M.; Menshikov, L.; Markov, S.; Fedichev, P.O. Extracting biological age from biomedical data via deep learning: Too much of a good thing? Sci. Rep. 2018, 8, 5210. [Google Scholar] [CrossRef]
- Yusuf, H.M.; Ali, Y.S.; Abubakar, A.H.; Abdullahi, M.; Hassan, I.H. A systematic review of deep learning techniques for rice disease recognition: Current trends and future directions. Frankl. Open 2024, 8, 100154. [Google Scholar] [CrossRef]
- Sa, R.; Owens, W.; Wiegand, R.; Studin, M.; Capoferri, D.; Barooha, K.; Greaux, A.; Rattray, R.; Hutton, A.; Cintineo, J.; et al. Intervertebral disc detection in X-ray images using faster R-CNN. In Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju, Republic of Korea, 11–15 July 2017; pp. 564–567. [Google Scholar]
- Ibrahim, A.T.; Mohammed, M.T.; Suleiman, M.A.; Kofa, A.B.; Ladan, I.M. Classification of Brain Tumors Using MRI Images Based on Fine-Tuned Pretrained Models. Int. J. Res. Sci. Innov. 2024, 11, 353–368. [Google Scholar] [CrossRef]
- Anjum, S.; Hussain, L.; Ali, M.; Alkinani, M.H.; Aziz, W.; Gheller, S.; Abbasi, A.A.; Marchal, A.R.; Suresh, H.; Duong, T.Q. Detecting brain tumors using deep learning convolutional neural network with transfer learning approach. Int. J. Imaging Syst. Technol. 2022, 32, 307–323. [Google Scholar] [CrossRef]
- Alanazi, M.F.; Ali, M.U.; Hussain, S.J.; Zafar, A.; Mohatram, M.; Irfan, M.; AlRuwaili, R.; Alruwaili, M.; Ali, N.H.; Albarrak, A.M. Brain tumor/mass classification framework using magnetic-resonance-imaging-based isolated and developed transfer deep-learning model. Sensors 2022, 22, 372. [Google Scholar] [CrossRef]
- Shin, H.C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef]
- Talo, M.; Baloglu, U.B.; Yıldırım, Ö.; Acharya, U.R. Application of deep transfer learning for automated brain abnormality classification using MR images. Cogn. Syst. Res. 2019, 54, 176–188. [Google Scholar] [CrossRef]
- Nematzadeh, S.; Kiani, F.; Torkamanian-Afshar, M.; Aydin, N. Tuning hyperparameters of machine learning algorithms and deep neural networks using metaheuristics: A bioinformatics study on biomedical and biological cases. Comput. Biol. Chem. 2022, 97, 107619. [Google Scholar] [CrossRef]
- Gaspar, A.; Oliva, D.; Cuevas, E.; Zaldívar, D.; Pérez, M.; Pajares, G. Hyperparameter optimization in a convolutional neural network using metaheuristic algorithms. In Metaheuristics in Machine Learning: Theory and Applications; Springer International Publishing: Cham, Switzerland, 2021; pp. 37–59. [Google Scholar]
- Probst, P.; Boulesteix, A.L.; Bischl, B. Tunability: Importance of hyperparameters of machine learning algorithms. J. Mach. Learn. Res. 2019, 20, 1–32. [Google Scholar]
- Mohakud, R.; Dash, R. Designing a grey wolf optimization based hyper-parameter optimized convolutional neural network classifier for skin cancer detection. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 6280–6291. [Google Scholar] [CrossRef]
- Oyelade, O.N.; Ezugwu, A.E. Characterization of abnormalities in breast cancer images using nature-inspired metaheuristic optimized convolutional neural networks model. Concurr. Comput. Pract. Exp. 2022, 34, e6629. [Google Scholar] [CrossRef]
- Hussain Ali, Y.; Sabu Chooralil, V.; Balasubramanian, K.; Manyam, R.R.; Kidambi Raju, S.; Sadiq, A.T.; Farhan, A.K. Optimization system based on convolutional neural network and internet of medical things for early diagnosis of lung cancer. Bioengineering 2023, 10, 320. [Google Scholar] [CrossRef]
- Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
- Aljarah, I.; Faris, H.; Mirjalili, S. Optimizing connection weights in neural networks using the whale optimization algorithm. Soft Comput. 2018, 22, 1–15. [Google Scholar] [CrossRef]
- Guo, W.; Liu, T.; Dai, F.; Xu, P. An improved whale optimization algorithm for forecasting water resources demand. Appl. Soft Comput. 2020, 86, 105925. [Google Scholar] [CrossRef]
- Kaur, G.; Arora, S. Chaotic whale optimization algorithm. J. Comput. Des. Eng. 2018, 5, 275–284. [Google Scholar] [CrossRef]
- Ravinder, M.; Saluja, G.; Allabun, S.; Alqahtani, M.S.; Abbas, M.; Othman, M.; Soufiene, B.O. Enhanced brain tumor classification using graph convolutional neural network architecture. Sci. Rep. 2023, 13, 14938. [Google Scholar] [CrossRef]
- Asiri, A.A.; Shaf, A.; Ali, T.; Aamir, M.; Irfan, M.; Alqahtani, S.; Mehdar, K.M.; Halawani, H.T.; Alghamdi, A.H.; Alshamrani, A.F.A.; et al. Brain tumor detection and classification using fine-tuned CNN with ResNet50 and U-Net model: A study on TCGA-LGG and TCIA dataset for MRI applications. Life 2023, 13, 1449. [Google Scholar] [CrossRef]
- Ramalakshmi, K.; Rajagopal, S.; Kulkarni, M.B.; Poddar, H. A hyperdimensional framework: Unveiling the interplay of RBP and GSN within CNNs for ultra-precise brain tumor classification. Biomed. Signal Process. Control 2024, 96, 106565. [Google Scholar]
- Shaik, N.S.; Cherukuri, T.K. Multi-level attention network: Application to brain tumor classification. Signal Image Video Process. 2022, 16, 817–824. [Google Scholar] [CrossRef]
- Hassan, E.; Ghadiri, H. Advancing brain tumor classification: A robust framework using EfficientNetV2 transfer learning and statistical analysis. Comput. Biol. Med. 2025, 185, 109542. [Google Scholar] [CrossRef]
- Deepak, S.; Ameer, P.M. Automated categorization of brain tumor from mri using cnn features and svm. J. Ambient Intell. Humaniz. Comput. 2021, 12, 8357–8369. [Google Scholar] [CrossRef]
- Pande, Y.; Chaki, J. Brain tumor detection across diverse MR images: An automated triple-module approach integrating reduced fused deep features and machine learning. Results Eng. 2025, 25, 103832. [Google Scholar] [CrossRef]
- Mathivanan, S.K.; Sonaimuthu, S.; Murugesan, S.; Rajadurai, H.; Shivahare, B.D.; Shah, M.A. Employing deep learning and transfer learning for accurate brain tumor detection. Sci. Rep. 2024, 14, 7232. [Google Scholar] [CrossRef]
- Patil, S.; Kirange, D. Ensemble of deep learning models for brain tumor detection. Procedia Comput. Sci. 2023, 218, 2468–2479. [Google Scholar] [CrossRef]
- Albalawi, E.; TR, M.; Thakur, A.; Kumar, V.V.; Gupta, M.; Khan, S.B.; Almusharraf, A. Integrated approach of federated learning with transfer learning for classification and diagnosis of brain tumor. BMC Med. Imaging 2024, 24, 110. [Google Scholar] [CrossRef]
- Rahman, T.; Islam, M.S.; Uddin, J. MRI-based brain tumor classification using a dilated parallel deep convolutional neural network. Digital 2024, 4, 529–554. [Google Scholar] [CrossRef]
- Kanchanamala, P.; Revathi, K.G.; Ananth, M.B.J. Optimization-enabled hybrid deep learning for brain tumor detection and classification from MRI. Biomed. Signal Process. Control 2023, 84, 104955. [Google Scholar] [CrossRef]
- Anaya-Isaza, A.; Mera-Jiménez, L.; Verdugo-Alejo, L.; Sarasti, L. Optimizing MRI-based brain tumor classification and detection using AI: A comparative analysis of neural networks, transfer learning, data augmentation, and the cross-transformer network. Eur. J. Radiol. Open 2023, 10, 100484. [Google Scholar] [CrossRef] [PubMed]
- Kanchanamala, P.; Kuppusamy, V.; Ganesan, G. QDCNN-DMN: A hybrid deep learning approach for brain tumor classification using MRI images. Biomed. Signal Process. Control 2025, 101, 107199. [Google Scholar] [CrossRef]
- Rammurthy, D.; Mahesh, P.K. Whale Harris hawks optimization based deep learning classifier for brain tumor detection using MRI images. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 3259–3272. [Google Scholar] [CrossRef]
- Cheng, J. Brain magnetic resonance imaging tumor dataset. Figshare MRI Dataset Version 2017, 5. [Google Scholar]
- Bhuvaji, S.; Kadam, A.; Bhumkar, P.; Dedge, S.; Kanchan, S. Brain tumor classification (MRI). Kaggle 2020, 10. Available online: https://www.kaggle.com/datasets/sartajbhuvaji/brain-tumor-classification-mri (accessed on 21 February 2025).
- Hamada, A. Br35H: Brain Tumor Detection 2020. Kaggle. 2020. Available online: https://www.kaggle.com/datasets/ahmedhamada0/brain-tumor-detection (accessed on 25 February 2025).
- Nadimi-Shahraki, M.H.; Zamani, H.; Asghari Varzaneh, Z.; Mirjalili, S. A systematic review of the whale optimization algorithm: Theoretical foundation, improvements, and hybridizations. Arch. Comput. Methods Eng. 2023, 30, 4113–4159. [Google Scholar] [CrossRef]
- Kellert, S.H. The Wake of Chaos: Unpredictable Order in Dynamical Systems; University of Chicago Press: Chicago, IL, USA, 1993. [Google Scholar]
- Yousri, D.; Allam, D.; Eteiba, M.B. Chaotic whale optimizer variants for parameters estimation of the chaotic behavior in Permanent Magnet Synchronous Motor. Appl. Soft Comput. 2019, 74, 479–503. [Google Scholar] [CrossRef]
- Zhu, H.; Zhao, Y.; Song, Y. 2D logistic-modulated-sine-coupling-logistic chaotic map for image encryption. IEEE Access 2019, 7, 14081–14098. [Google Scholar] [CrossRef]
- Isuwa, J.; Abdullahi, M.; Ali, Y.S.; Hassan, I.H.; Buba, J.R.; Aliyu, I.; Kim, J.; Oyelade, O.N. Optimizing microarray cancer gene selection using swarm intelligence: Recent developments and an exploratory study. Egypt. Inform. J. 2023, 24, 100416. [Google Scholar] [CrossRef]
- Hassan, I.H.; Abdullahi, M.; Aliyu, M.M.; Yusuf, S.A.; Abdulrahim, A. An improved binary manta ray foraging optimization algorithm based feature selection and random forest classifier for network intrusion detection. Intell. Syst. Appl. 2022, 16, 200114. [Google Scholar] [CrossRef]
- Li, J.; Guo, L.; Li, Y.; Liu, C.; Wang, L.; Hu, H. Enhancing whale optimization algorithm with chaotic theory for permutation flow shop scheduling problem. Int. J. Comput. Intell. Syst. 2021, 14, 651–675. [Google Scholar] [CrossRef]
- Li, Y.; Han, M.; Guo, Q. Modified whale optimization algorithm based on tent chaotic mapping and its application in structural optimization. KSCE J. Civ. Eng. 2020, 24, 3703–3713. [Google Scholar] [CrossRef]
- Singh, J.P.; Kumar, M. Chaotic whale-atom search optimization-based deep stacked auto encoder for crowd behaviour recognition. J. Exp. Theor. Artif. Intell. 2024, 36, 187–211. [Google Scholar] [CrossRef]
- He, B.; Jia, B.; Zhao, Y.; Wang, X.; Wei, M.; Dietzel, R. Estimate soil moisture of maize by combining support vector machine and chaotic whale optimization algorithm. Agric. Water Manag. 2022, 267, 107618. [Google Scholar] [CrossRef]
- Brodzicki, A.; Piekarski, M.; Jaworek-Korjakowska, J. The whale optimization algorithm approach for deep neural networks. Sensors 2021, 21, 8003. [Google Scholar] [CrossRef]
- Tair, M.; Bacanin, N.; Zivkovic, M.; Venkatachalam, K. A Chaotic Oppositional Whale Optimisation Algorithm with Firefly Search for Medical Diagnostics. Comput. Mater. Contin. 2022, 72. [Google Scholar] [CrossRef]
- Syama, S.; Ramprabhakar, J.; Anand, R.; Guerrero, J.M. A hybrid extreme learning machine model with lévy flight chaotic whale optimization algorithm for wind speed forecasting. Results Eng. 2023, 19, 101274. [Google Scholar] [CrossRef]
- Mozaffari-Kermani, M.; Sur-Kolay, S.; Raghunathan, A.; Jha, N.K. Systematic poisoning attacks on and defenses for machine learning in healthcare. IEEE J. Biomed. Health Inform. 2014, 19, 1893–1905. [Google Scholar] [CrossRef]
- Lv, Z.; Qiao, L.; Kumar Singh, A.; Wang, Q. AI-empowered IoT security for smart cities. ACM Trans. Internet Technol. 2021, 21, 1–21. [Google Scholar] [CrossRef]
- Hu, Y.; Kuang, W.; Qin, Z.; Li, K.; Zhang, J.; Gao, Y.; Li, W.; Li, K. Artificial intelligence security: Threats and countermeasures. ACM Comput. Surv. (CSUR) 2021, 55, 20. [Google Scholar] [CrossRef]
- Nia, A.M.; Mozaffari-Kermani, M.; Sur-Kolay, S.; Raghunathan, A.; Jha, N.K. Energy-efficient long-term continuous personal health monitoring. IEEE Trans. Multi-Scale Comput. Syst. 2015, 1, 85–98. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).