You are currently viewing a new version of our website. To view the old version click .
Diagnostics
  • Article
  • Open Access

11 December 2024

Optimized Hybrid Deep Learning Framework for Early Detection of Alzheimer’s Disease Using Adaptive Weight Selection

,
,
,
,
,
,
and
1
Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakaka 72388, Saudi Arabia
2
Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakaka 72388, Saudi Arabia
3
Cybersecurity Department, College of Engineering and Information Technology, Buraydah Private Colleges, Buraydah 51418, Saudi Arabia
4
College of Computer Engineering and Sciences, Prince Sattam bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
This article belongs to the Special Issue Artificial Intelligence in Alzheimer’s Disease Diagnosis

Abstract

Background: Alzheimer’s disease (AD) is a progressive neurological disorder that significantly affects middle-aged and elderly adults, leading to cognitive deterioration and hindering daily activities. Notwithstanding progress, conventional diagnostic techniques continue to be susceptible to inaccuracies and inefficiencies. Timely and precise diagnosis is essential for early intervention. Methods: We present an enhanced hybrid deep learning framework that amalgamates the EfficientNetV2B3 with Inception-ResNetV2 models. The models were integrated using an adaptive weight selection process informed by the Cuckoo Search optimization algorithm. The procedure commences with the pre-processing of neuroimaging data to guarantee quality and uniformity. Features are subsequently retrieved from the neuroimaging data by utilizing the EfficientNetV2B3 and Inception-ResNetV2 models. The Cuckoo Search algorithm allocates weights to various models dynamically, contingent upon their efficacy in particular diagnostic tasks. The framework achieves balanced usage of the distinct characteristics of both models through the iterative optimization of the weight configuration. This method improves classification accuracy, especially for early-stage Alzheimer’s disease. A thorough assessment was conducted on extensive neuroimaging datasets to verify the framework’s efficacy. Results: The framework attained a Scott’s Pi agreement score of 0.9907, indicating exceptional diagnostic accuracy and dependability, especially in identifying the early stages of Alzheimer’s disease. The results show its superiority over current state-of-the-art techniques.Conclusions: The results indicate the substantial potential of the proposed framework as a reliable and scalable instrument for the identification of Alzheimer’s disease. This method effectively mitigates the shortcomings of conventional diagnostic techniques and current deep learning algorithms by utilizing the complementing capabilities of EfficientNetV2B3 and Inception-ResNetV2 by using an optimized weight selection mechanism. The adaptive characteristics of the Cuckoo Search optimization facilitate its application across many diagnostic circumstances, hence extending its utility to a wider array of neuroimaging datasets. The capacity to accurately identify early-stage Alzheimer’s disease is essential for facilitating prompt therapies, which are crucial for decelerating disease development and enhancing patient outcomes.

1. Introduction

Alzheimer’s disease (AD) is a devastating neurological disorder that gradually impairs memory, thinking, and the ability to carry out simple tasks. Often beginning with mild forgetfulness, AD progressively leads to severe cognitive decline, affecting millions of people worldwide. As of 2020, more than 5.8 million Americans were living with Alzheimer’s disease, mostly older adults. Globally, the number is expected to skyrocket in the coming decades, making AD not just a personal tragedy for those affected and their families but also a massive societal challenge. Caring for people with AD strains healthcare systems, caregivers, and the economy. Early diagnosis is critical for managing symptoms and improving quality of life, but it remains one of the most complex aspects of the disease.
One of the biggest hurdles in diagnosing Alzheimer’s disease is that its symptoms often overlap with normal aging or other forms of dementia, especially in its early stages. While cognitive tests and memory assessments are frequently used, they’re often not definitive. More accurate methods, such as brain imaging techniques like MRI, can detect changes in brain structure associated with AD. However, these procedures are expensive and not always available in all healthcare settings, particularly in underserved regions. As a result, there’s a growing demand for more accessible, efficient diagnostic tools that can provide earlier and more reliable detection of AD.
Artificial intelligence (AI) [1] and machine learning are emerging as powerful tools in the fight against Alzheimer’s disease [2,3]. By training models to recognize patterns in brain images or other biomarkers [4], these technologies hold the promise of automating the diagnosis process, potentially making it faster and more accurate. In particular, deep learning models [5,6,7,8], like convolutional neural networks (CNNs), have shown great promise in analyzing medical images to detect disease patterns. However, these models have limitations: they often require vast amounts of data and computational power, and they can sometimes struggle to generalize across different patient populations or datasets. This is where combining multiple models, an approach known as ensemble learning, can come into play.
Ensemble learning involves using multiple AI models to make predictions and then combining their outputs to create a more accurate final decision. By blending the strengths of different models, this method can reduce errors and improve performance, especially in complex tasks like diagnosing Alzheimer’s disease. Though ensemble learning has been explored in other areas of healthcare, such as cancer detection, it is less commonly applied to AD diagnosis.
This study aims to develop an optimal ensemble learning system for diagnosing Alzheimer’s disease using neuroimaging data from the well-known Alzheimer’s Disease Neuroimaging Initiative (ADNI) database. The focus is on integrating multiple deep learning models, improving the way features from these models are selected and combined, and fine-tuning how the different models’ predictions are weighted to ensure the most accurate results.
The paper is organized as follows: Section 2 reviews related work, Section 3 details the materials and methods employed in this research, Section 4 analyzes the experimental results, and Section 5 concludes the study.

3. Proposed Model for Alzheimer Detection

In this study, we propose an approach, Figure 1, that combines two deep learning models at a time to leverage the strengths of each architecture. Our approach is particularly designed for Alzheimer’s disease detection using medical imaging data, such as MRI scans. The goal is to improve classification accuracy and model robustness by harnessing the complementary features learned by different models.
Figure 1. Proposed model for Alzheimer’s disease detection.
To achieve this, we integrated a weight selection method based on an optimization algorithm. In our approach, each model contributes a probability to the final classification, and the optimization algorithm assigns a weight to each model’s output. These weights are dynamically optimized during training to ensure that the most reliable model in each pair contributes more to the final decision. This allows for a more flexible and adaptive classification system, which is crucial for complex tasks like Alzheimer’s disease detection, where subtle differences in medical images can make a significant impact on diagnostic outcomes.
The optimization algorithm we use ensures that the selected weights for each pair of models are optimal in terms of minimizing classification error. The models are combined in a way that the weighted probabilities from both models produce a final prediction, enhancing the overall performance. For example, when EfficientNetB0 is paired with ResNet50, the optimization algorithm ensures that each model’s contribution to the final decision is weighted according to its performance on the task, leading to improved diagnostic accuracy.
This method of combining models and dynamically adjusting their contributions offers a novel way to handle classification tasks, particularly in medical applications like Alzheimer’s disease detection. The integration of the weight selection method via optimization provides a more tailored and precise classification system capable of adapting to the complexity and variability of medical data.

3.1. Dataset and Data Pre-Processing

The OASIS (Open Access Series of Imaging Studies) (https://www.oasis-brains.org/ (accessed on 15 August 2024)) dataset is a widely used collection of brain imaging data and is primarily used to focus on understanding and diagnosing Alzheimer’s disease. The dataset contains structural brain MRI scans, along with clinical and demographic information, and has been instrumental in advancing research in Alzheimer’s disease detection and neuroimaging.
The dataset, sourced from the OASIS study, consists of 80,000 MRI brain scans taken along the z-axis and includes data from 461 participants. The distribution of scans across the severity levels shows a significant class imbalance, with 67,222 scans in the ‘none’ category, 13,725 in ‘very mild’, 5002 in ‘mild’, and 488 in ‘moderate’. Figure 2 present same examples taken from the dataset.
Figure 2. Examples of images taken from the dataset.
There is a noticeable class imbalance in the dataset presented in Figure 3, where some categories have far more samples than others. This imbalance can create challenges for deep learning models, as they tend to favor the majority class, making it harder for the model to correctly identify and classify cases in the minority classes. To develop a reliable and fair model, it is essential to address this imbalance effectively.
Figure 3. Examples of class imbalance in the dataset.
To address the class imbalance [34], methods like oversampling the minority class, undersampling the majority class, or using more advanced techniques like the Synthetic Minority Over-sampling Technique (SMOTE) can be applied to improve the model’s ability to learn from the under-represented classes. In our case, we experimented with several approaches and ultimately opted for a combination of over- and undersampling.
Furthermore, it is important to use evaluation metrics that reflect the impact of class imbalance. Metrics like precision, recall, and the F1-score provide a more accurate assessment of model performance compared to relying solely on accuracy.

Data Preparation and Handling Class Imbalance

  • Step1: Splitting the Test Set
Before starting any pre-processing and applying oversampling techniques, we first set aside a test set of images from each category. This ensures that the model is evaluated on completely unseen data, giving a more accurate reflection of its performance. Importantly, this test set was not involved in hyperparameter tuning, so it offers an unbiased assessment of the model’s capabilities.
Specifically, 20% of the “moderate dementia” samples and 770 images from each of the other three categories were reserved to form the test set. Although the test set was not proportionally distributed across classes, this approach reflects real-world diagnostic conditions, where certain categories, such as “moderate dementia”, are naturally under-represented. This design ensures that the test set better mirrors realistic clinical scenarios and predictions.
This approach ensured that the training and validation datasets were balanced, facilitating effective model learning, and the test set remained independent and comprised only real images. This separation was critical to provide an unbiased evaluation of the model’s performance.
  • Step 2: Addressing Class Imbalance
Our dataset was highly imbalanced, with the smallest class representing only 0.007% of the largest class. While this imbalance mirrors the actual distribution of Alzheimer’s disease cases (with more severe stages being rarer), it poses a significant challenge for training a neural network. A model trained on imbalanced data can show high accuracy for the majority class but struggle with under-represented categories, resulting in misleading performance metrics like accuracy. Handling this imbalance was a key priority.
A particular complication in this case is that we are working with medical brain scans. Common data augmentation techniques, such as rotations or distortions, are not advisable for medical data since altering these images can remove important structural information [35], potentially leading to inaccurate diagnoses. Given the ethical implications, we avoided aggressive augmentation methods.
We explored several solutions, including undersampling and oversampling. Ultimately, we found that a combination of undersampling the majority classes and oversampling the minority classes provided the best results, which we used in our final model.
For the training and validation sets, we undersampled the majority classes and oversampled the minority ones to achieve 6000 images per class. This gave us a balanced dataset of 24,000 images, which we split into 19,000 for training and 5000 for validation. This balanced set helped train the model effectively while avoiding the pitfalls of class imbalance.

3.2. Deep Learning Model

In our hybrid learning approach, we implemented six distinct models: EfficientNetB0, EfficientNetB3V2, ResNet50, Inception-ResNet50, MobileNet, and ConvNet to enhance Alzheimer’s disease detection. Each model brings unique strengths to feature extraction and image classification, allowing us to capture varying levels of detail in MRI brain scans. By combining these models, we aimed to improve diagnostic accuracy and robustness across multiple disease stages.

3.2.1. EfficientNetB0

EfficientNetB0 [36] is a groundbreaking model that strikes an impressive balance between performance and efficiency. What sets it apart from traditional architectures is its use of compound scaling—a method that scales the depth, width, and resolution of the network in a balanced way. Instead of just making the network deeper or wider, EfficientNetB0 grows in a thoughtful, balanced manner that allows it to achieve great accuracy without consuming too much computational power. This makes it ideal for tasks where speed and efficiency are critical, such as medical imaging on resource-constrained systems. By including EfficientNetB0 in our model combinations, we benefit from its ability to handle tasks that require precision with minimal overhead, which is particularly useful when working with high-resolution scans in Alzheimer’s disease detection.
EfficientNet models are based on compound scaling, which systematically scales the model’s width, depth, and resolution. The scaling formula is
d = α ϕ , w = β ϕ , r = γ ϕ
where
-
d is the depth (number of layers);
-
w is the width (number of channels per layer);
-
r is the image resolution;
-
α , β , and γ are constants that control how depth, width, and resolution are scaled, respectively;
-
ϕ is the scaling coefficient, which controls overall scaling based on available resources.
This equation ensures a balanced increase in network size while maintaining efficiency.

3.2.2. EfficientNetB3V2

A more advanced version of the EfficientNet family [37], EfficientNetB3V2, refines the strengths of its predecessor, incorporating newer techniques like batch normalization and enhanced scaling to boost performance. It is a more powerful model designed to manage higher computational loads while still maintaining the trademark efficiency of the EfficientNet series. For complex image classification tasks, like those needed in Alzheimer’s detection, EfficientNetB3V2 provides superior accuracy while remaining computationally light enough to be practical in real-world applications. By pairing it with other models in our approach, we harness its ability to work through intricate details in the data, offering a deeper layer of analysis without sacrificing speed.

3.2.3. ResNet50

ResNet50 is one of the most iconic models in deep learning, known for its innovative use of residual connections. These connections help the model avoid common training pitfalls, like the vanishing gradient problem, by allowing layers to skip connections and pass information directly across the network. This makes ResNet50 extremely powerful for extracting deep, hierarchical features from images. In the context of Alzheimer’s disease detection, ResNet50 can capture subtle, deep patterns in MRI or CT scans that may be missed by shallower networks. By combining ResNet50 with other models, we tap into its ability to dig deep into the data and find patterns that are crucial for accurate diagnosis, giving our approach a more thorough feature extraction process [38].
ResNet50 uses residual connections to avoid the vanishing gradient problem by skipping layers. The mathematical expression for a residual block is
y = F x , W i + x
where
-
x is the input to the residual block;
-
F x , W i is the learned residual mapping (the result of the convolution layers);
-
y is the output of the block.
These residual connections allow the network to train deeper layers effectively.

3.2.4. Inception-ResNet50

The Inception-ResNet50 model is an exciting fusion of two powerful architectures: Inception and ResNet. The Inception component brings the ability to process features at multiple scales simultaneously, while the ResNet component ensures that the model can be trained to significant depth without running into problems like vanishing gradients. This combination makes Inception-ResNet50 particularly adept at tackling complex visual tasks, where both fine details and broader contextual information are important. In our Alzheimer’s disease detection framework, this model helps by capturing a range of features, from the smallest anomalies to more generalized patterns in the brain scans, adding richness and depth to our predictions [39].

3.2.5. MobileNet

MobileNet is all about lightweight efficiency. Designed for use in mobile and embedded systems, it uses a clever technique called depthwise separable convolutions to reduce the number of parameters and computations while still achieving impressive accuracy. This makes it perfect for real-time applications or settings where computational resources are limited. In our combined model approach, MobileNet brings speed and efficiency to the table, offering a quick and effective way to process large datasets, especially when used alongside heavier models like ResNet50. This combination allows us to build a more adaptive and responsive system for Alzheimer’s disease detection, capable of providing fast preliminary results without compromising accuracy [40].
MobileNet uses depthwise separable convolutions, which significantly reduce computation by splitting the convolution operation into two parts. The depthwise separable convolution is given by
Y = D W X , K d × P W K p
where
-
X is the input;
-
K d and K p are depthwise and pointwise convolution filters, respectively;
-
D W is the depthwise convolution (operating separately on each channel);
-
P W is the pointwise convolution (combining outputs from depthwise convolution).

3.2.6. ConvNet (Convolutional Neural Network)

The convolutional neural network (ConvNet) is the classic workhorse of deep learning, particularly well-suited to image data. ConvNets are composed of layers of filters (or “kernels”) that automatically learn to detect patterns in image edges, textures, shapes, and beyond. What makes ConvNets so powerful is their ability to capture spatial hierarchies in the data, which is crucial for understanding images at both the local and global levels. In our approach, ConvNet serves as a foundational model, offering reliable feature extraction and complementing the more sophisticated architectures we pair it with. Its simplicity and proven effectiveness make it an essential part of our toolset for diagnosing Alzheimer’s disease from complex medical images [41].
A standard convolutional layer in ConvNet can be mathematically expressed as
y i , j , k = m n x i + m , j + n · w m , n , k + b k
where
-
x is the input image or feature map;
-
w represents the learned weights (filters);
-
b k is the bias for filter k;
-
y is the output feature map.

3.3. Combination of Two Models: Average and Weighted Average Techniques

In our proposed approach, we first explore the simple average combination technique, followed by an enhanced method based on a weighted average, where the weights are dynamically selected by an optimization algorithm to improve the overall performance.
  • Step 1: Average Combination Technique
The initial step in our approach is to combine the outputs of two models using the simple average method. In this technique, both models contribute equally to the final classification result. The probabilities predicted by each model are added together and then divided by two, giving equal importance to both. For instance, when combining Model A and Model B to classify Alzheimer’s disease, the final probability output P final would be calculated as
P final = P A + P B 2
This method is easy to implement and can be quite effective, especially when the models perform similarly or have complementary strengths. However, the limitation is that it assumes both models are equally reliable, which may not always be the case, particularly in complex medical tasks like Alzheimer’s detection, where subtle variations in the data can lead to differences in model performance.
  • Step 2: Weighted Average Combination based on a Selection Technique:
To improve on the simple average method, we introduce a weighted average combination, where each model’s contribution to the final decision is weighted based on its performance. The weights for each model are not static; instead, they are selected dynamically by an optimization algorithm that minimizes the classification error.
In this approach, the output from each model is multiplied by a weight, and the final classification is based on the weighted sum of the predictions. The formula for the weighted average combination is
P final = w A · P A + w B · P B
Here, w A and w B represent the weights assigned to Model A and Model B, respectively. The optimization algorithm adjusts these weights during the training process to ensure that the final combined model performs optimally.
The weighted average method offers several advantages. It allows the model that performs better to have more influence on the final decision, while the less accurate model contributes less. This approach is particularly useful in tasks like Alzheimer’s disease detection, where different models may excel at identifying different features in medical images. By optimizing the weights, we ensure that the final model adapts to the data’s complexity and delivers more accurate classifications.
In summary, by starting with a simple averaging approach and then enhancing it with a dynamic, optimized weighted average method, we create a more flexible and precise model. This not only improves the overall performance but also makes the system more adaptable to the unique challenges of Alzheimer’s disease detection.

3.4. Weight Selection Using Cuckoo Search for Combining Deep Learning Models

In this study, we use the Cuckoo Search (CS) algorithm to select the optimal weights when combining the outputs of two deep learning models for Alzheimer’s disease detection. Cuckoo Search is a powerful optimization technique inspired by the natural behavior of cuckoo birds, which lay their eggs in the nests of other birds. This method has gained popularity due to its ability to efficiently explore the solution space while avoiding local optima, making it well-suited for weight optimization in deep learning models, especially in complex tasks like medical image classification [42,43].
The final prediction from our combined models is represented by Equation (6).

3.4.1. Cuckoo Search Algorithm Overview

Cuckoo Search uses random walks known as Lévy flights to search through the solution space. These random steps allow the algorithm to cover a wide area, increasing the chances of finding the optimal solution without getting stuck in sub-optimal regions [42]. The basic principles of CS can be summarized as the following:
Egg Laying: Each cuckoo lays one egg (a new potential solution) in a randomly chosen nest (a solution candidate).
Fitness Evaluation: The quality of the egg (solution) is evaluated; in our case, this is based on the classification accuracy achieved using the current weight combination.
Nest Replacement: The best nests (solutions) are retained for the next iteration, while a portion of the less optimal ones are replaced with new solutions, allowing the algorithm to explore new regions of the solution space [44].

3.4.2. Applying Cuckoo Search for Weight Selection

For our task of combining two deep learning models, we followed a structured process to optimize the weights assigned to each model’s prediction. Here is how we apply Cuckoo Search:
Initialization: A population of candidate weight combinations ( w A , w B ) is initialized. Each combination represents a possible way of combining the outputs of the two models.
Fitness Evaluation: For each candidate combination of weights, we compute the final prediction, P final , and evaluate its accuracy on the validation set. The accuracy serves as the fitness score for each weight combination [45].
Lévy Flights for Exploration: Lévy flights enable the algorithm to take both large and small steps, which helps in searching the solution space efficiently. This ensures that the algorithm does not get stuck in local optima [42,46].
Selection and Replacement: After evaluating the fitness of all solutions, the best-performing weight combinations are retained, and a portion of lower-performing solutions is replaced. This maintains diversity in the solution space and encourages exploration.
Convergence: The algorithm iterates until it converges, i.e., when the improvement in accuracy between generations becomes negligible. At this point, the best weight combination is selected [47].
Final Weight Selection: Once the process is complete, the optimal weight combination ( w A , w B ) is chosen, resulting in the final prediction ( P final [48]).

3.4.3. Advantages of Cuckoo Search for Model Combination

Cuckoo Search offers several key benefits when it comes to selecting optimal weights for combining deep learning models:
Efficient Search: Thanks to the Lévy flights, CS explores the solution space effectively, covering both small local adjustments and large-scale changes.
Avoiding Local Optima: The algorithm’s ability to replace sub-optimal solutions ensures it does not get stuck in local optima, unlike more traditional methods.
Simple and Flexible: CS is straightforward to implement and can easily be adapted for various optimization problems, including weight selection in deep learning.
Cuckoo Search’s ability to balance exploration and exploitation makes it highly suitable for complex optimization problems like medical image analysis [48].

4. Results and Discussions

In the following section, the results of the experiments that were carried out on the model proposed in this study are shown. In addition, we evaluate the outcomes that were achieved by employing a variety of classification strategies within the framework of deep learning models. Python “https://www.python.org/” was used to implement the proposed model, and it was executed on a graphics card with 16 gigabytes of random access memory (RAM) and an RTX 2060 graphics card.
To evaluate the suggested model, we performed multiple experiments using different deep learning models. The design of our model includes three distinct scenarios:
  • Scenario 1: Alzheimer’s disease detection based on deep learning models;
  • Scenario 2: Alzheimer’s disease detection classification based on hybrid deep learning models;
  • Scenario 3: Alzheimer’s disease detection classification based on the weight selection method.

4.1. Evaluation Metrics

We utilized standard metrics, including accuracy, precision, recall, and F-measure, to evaluate our categorization method. To evaluate our model’s performance, we used precision, recall, F1-score, ROC-AUC, and Scott’s Pi. Precision focuses on the accuracy of positive predictions, while recall measures the true positive rate. The F1-score combines both, giving a balanced view of how well the model handles false positives and false negatives. Although these metrics are designed for binary classification, they can be adapted to multi-class and ordinal tasks by calculating a micro-average, which treats each class separately.
A c c a r a c y = T P + T N / ( T P + T N + F P + F N )
P r e c i s i o n = T P / ( T P + F P )
R e c a l l = T P / ( T P + F N )
F 1 s c o r e = 2 / ( 1 / P + 1 / R )
TP denotes true positive, FP denotes false positive, P signifies the precision rate, R indicates the recall rate, TPR represents the true positive rate, and FPR signifies the false positive rate.
We also calculated the ROC-AUC score, which gives an overall picture of the model’s performance by considering the balance between true and false positives. Additionally, we used Scott’s Pi, as it provides a more accurate measure for ordinal classification. Scott’s Pi is typically used to assess inter-rater reliability, but it works well for ordinal tasks since it takes the ordered nature of the data into account, making it ideal for our application, according to the research in [49].

4.2. Performance Evaluation and Comparative Analysis of Deep Learning Models for Alzheimer’s Disease Detection

The performance of six deep learning models—EfficientNetB0, ResNet50V2, ConvNextBase, InceptionResNetV2, MobileNet1, and EfficientNetV2B3—was evaluated based on their accuracy, validation accuracy, loss, and validation loss. These metrics provide valuable insight into how well each model performed in training and its ability to generalize to unseen data. Table 1 presents the results obtained.
Table 1. Performance accurracy of different deep learning models.
Among the models, ResNet50V2 and EfficientNetV2B3 emerged as top performers, exhibiting the highest accuracy on both training and validation datasets. This indicates that these models not only learned the training data effectively but also maintained strong generalization when applied to new, unseen cases. Their architecture allows them to capture detailed and complex patterns, which is crucial for tasks like Alzheimer’s disease classification, where subtle changes in medical images are key.
On the other hand, MobileNet1, known for being lightweight and efficient, showed lower accuracy compared to the deeper models. While its design favors efficiency, it may not be as effective in handling complex tasks like disease classification. However, MobileNet1 remains a viable option for situations where computational resources are limited or when real-time analysis is required, though at the cost of reduced classification accuracy.
The loss values further help us understand the models’ optimization performance. EfficientNetV2B3 and ResNet50V2, again, proved to be the most effective, showing lower loss and validation loss values, which suggests that these models were able to minimize errors effectively during training and perform consistently across the validation set.
Models like InceptionResNetV2 and ConvNextBase, although showing competitive training loss, had higher validation loss values. This points to potential overfitting, where the model performs well on the training data but struggles to maintain the same performance when evaluated on new data. This discrepancy between training and validation loss is an important consideration, as it affects the model’s reliability in real-world scenarios.
While models such as EfficientNetB0 and MobileNet1 performed reasonably well, their lower validation accuracy and higher loss values suggest that they may have limitations in generalizing to more complex cases. Their simpler architecture is less suited for capturing the intricate features required for precise classifications, particularly in medical imaging. Nonetheless, their computational efficiency makes them attractive options for applications where speed and resource constraints are important, even if some performance is sacrificed.
In the context of Alzheimer’s disease detection, the ability to generalize across different stages of the disease is crucial. Models like EfficientNetV2B3 and ResNet50V2 showed the best balance between high accuracy and low error rates, making them strong candidates for this task. Their superior ability to handle complex image patterns ensures more reliable classifications. On the other hand, models such as MobileNet1 and EfficientNetB0, while efficient, may be more suitable in scenarios where real-time processing or limited computational power is prioritized, even though they might not provide the same level of accuracy.

Evaluation of Deep Learning Models for Alzheimer’s Disease Detection Across Multiple Classes

The deep learning models were assessed using key performance metrics such as precision, recall, F1-score, ROC AUC, and Scott’s Pi. These metrics provide insight into both the overall performance of each model and their effectiveness in distinguishing between different stages of dementia. Our results, as detailed in Table 2, demonstrate significant variation in performance metrics.
Table 2. Performance metrics for different deep learning models.
Overall, EfficientNetB0 and EfficientNetV2B3 emerged as top performers, achieving consistently high scores across all metrics. EfficientNetB0 recorded a micro-average precision, recall, and F1-score of 0.9851, along with an impressive ROC AUC of 0.9995, demonstrating its strong ability to differentiate between Alzheimer’s disease stages. Similarly, EfficientNetV2B3 had near-perfect scores, with a micro-average F1-score of 0.9871 and an ROC AUC of 0.9998, confirming its robust generalization across the dataset. These results indicate that both models excel in terms of accuracy and are highly reliable for clinical use, where distinguishing between different stages of dementia is crucial.
ResNet50V2 also delivered a solid performance, with a micro-average F1-score of 0.9673 and an ROC AUC of 0.9986. However, it slightly lagged behind the EfficientNet models, particularly in its precision and recall values for specific stages. For example, its precision for the non-demented class was 0.9816, slightly lower than 0.9664 forEfficientNetB0, indicating some room for improvement in detecting non-demented cases.
InceptionResNetV2 performed well, with a micro-average F1-score of 0.9524 and an ROC AUC of 0.997. It showed high precision and recall, particularly for the more severe stages of dementia, achieving perfect scores for the mild demented class. However, its slightly lower performance on early-stage detection, like the very mild demented class, where it had an F1-score of 0.9267, suggests that further refinement could enhance its ability to detect the subtler changes associated with early Alzheimer’s disease.
In contrast, ConvNextBase and MobileNet displayed more variability in their results. ConvNextBase performed reasonably well in the moderate demented class, with an F1-score of 0.9123, but it struggled with the mild demented class, achieving a significantly lower F1-score of 0.5899. MobileNet, while more efficient, also struggled with early stages, posting a micro-average F1-score of 0.8845, indicating it may not be as effective at identifying milder cases of Alzheimer’s disease, where distinguishing features are less apparent.
A common challenge for ConvNextBase and MobileNet was their handling of class imbalance. These models tended to perform better in the majority classes, whereas their accuracy dropped for the minority classes, such as the moderate demented category. Addressing this imbalance more effectively, perhaps by using better sampling techniques or model adjustments, would likely enhance their overall performance, particularly in detecting earlier stages of Alzheimer’s disease.
Despite these differences, the ROC AUC values for most models remained high, with even ConvNextBase achieving an ROC AUC of 0.9775, indicating its ability to distinguish between Alzheimer’s disease stages, even though precision and recall may vary between categories.
In conclusion, EfficientNetB0 and EfficientNetV2B3 stood out as the most reliable models for detecting Alzheimer’s disease across different stages thanks to their high accuracy, generalization ability, and near-perfect ROC AUC scores. ResNet50V2 and InceptionResNetV2 also performed well but may benefit from further optimization in early-stage detection. ConvNextBase and MobileNet, while efficient, showed limitations in handling class imbalance and detecting early-stage Alzheimer’s disease, suggesting the need for further refinement to improve their accuracy in these categories.

4.3. Alzheimer Detection Based on an Hybrid Deep Learning Model

This evaluation looked at how well different deep learning models—EfficientNetB0, InceptionResNetV2, ResNet50V2, and ConvNeXtBase—performed together by measuring their agreement using Scott’s Pi. This metric helps us understand how well each pair of models aligns in their predictions for Alzheimer’s disease classification, indicating how complementary they are when combined. This evaluation presented in Table 3.
Table 3. Evaluation of Average ensemble learning model for Alzheimer’s disease detection.
The combination of EfficientNetB0 and InceptionResNetV2 showed the highest agreement, with a Scott’s Pi of 0.98860. This strong result indicates that these two models are highly in sync and likely identify similar patterns in the data. Their consistency makes them a great pairing for improving classification accuracy in an ensemble model.
Similarly, InceptionResNetV2 and ResNet50V2 also demonstrated very high agreement, with a Scott’s Pi of 0.98718. This suggests that these models work well together, providing reliable predictions by capturing complementary features of the dataset. Their combination could be particularly effective in ensuring the robust and accurate detection of Alzheimer’s disease stages. The combination of EfficientNetB0 and ResNet50V2 achieved a Scott’s Pi of 0.98575, which is still strong but slightly lower than the top pairings. This suggests that while these models generally align, there may be minor differences in how they approach the classification task. Nonetheless, they remain a solid combination with a high level of consistency.
The pairing of ResNet50V2 and ConvNeXtBase resulted in a Scott’s Pi of 0.96581. Although this score is lower than others, it still indicates a good level of agreement. This combination might face challenges in consistently identifying certain patterns, but it generally performs well, making it a viable pairing for most classification tasks.
The combination of EfficientNetB0 and ConvNeXtBase scored 0.97507 for Scott’s Pi. While this is a decent level of agreement, it suggests that these two models may have some differences in how they handle certain features, which could affect their overall performance when combined.
The lowest agreement was observed between InceptionResNetV2 and ConvNeXtBase, with a Scott’s Pi of 0.94730. Although these models still perform fairly well together, the lower score indicates more variation in their predictions, likely due to differences in their architectures or how they process the data. This combination may not be as reliable as others for precise classification tasks.
Overall, the combinations of EfficientNetB0 with InceptionResNetV2 and InceptionResNetV2 with ResNet50V2 stood out for their high levels of agreement, making them strong candidates for ensemble models that need reliable and consistent predictions. ConvNeXtBase, while showing reasonable performance, displayed slightly lower alignment with other models, suggesting that it may not be the best choice for pairings when maximum consistency is required. These findings highlight the importance of selecting models that complement each other well, especially when aiming for improved classification accuracy in Alzheimer’s disease detection.

Enhanced Alzheimer Detection Through Optimized Weight Selection in Hybrid Deep Learning Models

To further enhance the performance of our model combination, the next step is to refine how we merge the predictions from our best-performing models, EfficientNetV2B3 and InceptionResNetV2. Rather than combining their outputs equally, we will implement a weighted selection method, where each model is assigned a weight based on its accuracy and reliability. This approach allows us to give the model that performs better in specific scenarios more influence, leading to more precise predictions. By optimizing these weights during training, we can create a more balanced and efficient model combination, improving the overall accuracy and making the system more effective at detecting different stages of Alzheimer’s disease. This strategy will ensure that we are using each model’s strengths to their fullest potential.
In this evaluation, we used an optimization algorithm to find the best weight distribution for combining two of our top-performing models, EfficientNetV2B3 and InceptionResNetV2. The aim was to determine how different weight configurations would impact their combined performance, using Scott’s Pi as a measure of agreement between the models. Table 4 presents the results obtained.
Table 4. Performance metrics of optimized weighted average models for Alzheimer’s disease detection using optimized selected alpha ( α ) and Beta ( β ) values.
The results show that the choice of weights plays a significant role in how well the models work together. When both models were given equal weight (0.5; 0.5), the Scott’s Pi score was 0.9893, indicating strong alignment between the two models. However, we found that slightly favoring EfficientNetV2B3 (with a weight of 0.5; 0.45) actually improved the result, giving us the highest Scott’s Pi score of 0.9907. This suggests that while both models contribute well to the classification, giving EfficientNetV2B3 a bit more influence yields better overall performance.
Other weight combinations, like 0.7; 0.3 and 0.6; 0.45, also resulted in high Scott’s Pi values (0.9864 and 0.9893, respectively), showing that these models are generally quite compatible. However, when the weights were skewed too heavily towards one model, such as 0.2; 0.8 in favor of InceptionResNetV2, Scott’s Pi value dropped significantly to 0.9465. This indicates that putting too much emphasis on one model diminishes the ensemble’s overall effectiveness.
What these results highlight is that the best performance comes from a balanced or near-balanced combination of the two models. EfficientNetV2B3 seems to have a slight edge in handling the data, which is why giving it a bit more weight improves the outcome. The optimization algorithm helped identify this sweet spot and keeps an archive of the best-performing weights, ensuring we can maintain this optimal balance in future runs.
The study demonstrates that combining EfficientNetV2B3 and InceptionResNetV2 works best when their contributions are balanced, with a slight preference for EfficientNetV2B3. The optimal weight distribution of 0.5; 0.45 achieved the highest Scott’s Pi score of 0.9907, confirming that fine-tuning the weights can significantly enhance the models’ agreement and overall performance. This method of optimizing and archiving weights provides a strong foundation for future improvements in model accuracy and reliability.

5. Conclusions

This study introduces a novel hybrid deep learning framework designed to enhance the early detection of Alzheimer’s disease through the integration of EfficientNetV2B3 and Inception-ResNetV2 models. By using the Cuckoo Search optimization algorithm, we dynamically adjusted the weighting of these models, achieving a balance (0.5; 0.45) that yielded a Scott’s Pi agreement score of 0.9907, representing near-perfect consistency. This optimized ensemble approach outperformed individual models in terms of accuracy and reliability, particularly in identifying subtle structural changes in the early stages of Alzheimer’s disease. The optimization process not only identified the best-performing configurations but also allowed the system to archive and adapt these configurations for continuous improvement. This adaptability positions the framework as a scalable and practical solution for real-world medical applications. Furthermore, the model’s interpretability techniques provide clinicians with insights into the decision-making process, enhancing its usability and trustworthiness in clinical settings. In conclusion, this optimized hybrid framework represents a significant advancement in AI-assisted medical diagnostics, offering a precise and reliable tool for Alzheimer’s detection. Its ability to detect early-stage Alzheimer’s disease has the potential to improve patient outcomes through timely intervention, paving the way for future developments in neuroimaging technologies and AI-driven healthcare solutions. By addressing a critical need in Alzheimer’s disease diagnosis, this research contributes to a more effective and equitable approach to managing this challenging disease.

Author Contributions

Conceptualization, K.G.; methodology, K.G. and A.A.; software, K.G. and A.A.; validation, L.B.A., M.O.A. and O.R.S.; formal analysis, K.G., A.A., M.O.A. and A.S.; resources, K.G. and H.C.; data curation, O.H. and M.O.A.; writing—original draft preparation, K.G. and A.A.; writing—review and editing, O.H., O.R.S., M.O.A. and A.S.; visualization, K.G., O.H. and A.S.; funding acquisition, K.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No. (DGSSR-2024-02-01226).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data used in this study are openly accessible at the following link: https://www.oasis-brains.org/ (accessed on 15 August 2024).

Acknowledgments

This work was funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No. (DGSSR-2024-02-01226).

Conflicts of Interest

All authors declare that there are no competing interests.

References

  1. Krichen, M.; Abdalzaher, M.S. Performance enhancement of artificial intelligence: A survey. J. Netw. Comput. Appl. 2024, 104034. [Google Scholar] [CrossRef]
  2. Mirzaei, G.; Adeli, H. Machine learning techniques for diagnosis of alzheimer disease, mild cognitive disorder, and other types of dementia. Biomed. Signal Process. Control 2022, 72, 103293. [Google Scholar] [CrossRef]
  3. Alroobaea, R.; Mechti, S.; Haoues, M.; Rubaiee, S.; Ahmed, A.; Andejany, M.; Bragazzi, N.L.; Sharma, D.K.; Kolla, B.P.; Sengan, S. Alzheimer’s Disease Early Detection Using Machine Learning Techniques. Wirel. Pers. Commun. 2021. [Google Scholar] [CrossRef]
  4. Gasmi, K.; Ben Aoun, N.; Alsalem, K.; Ltaifa, I.B.; Alrashdi, I.; Ammar, L.B.; Mrabet, M.; Shehab, A. Enhanced Brain Tumor Diagnosis Using Combined Deep Learning Models and Weight Selection Technique. Front. Neuroinform. 2024, 18, 1444650. [Google Scholar] [CrossRef]
  5. Ding, Y.; Sohn, J.H.; Kawczynski, M.G.; Trivedi, H.; Harnish, R.; Jenkins, N.W.; Lituiev, D.; Copeland, T.P.; Aboian, M.S.; Mari Aparici, C.; et al. A deep learning model to predict a diagnosis of Alzheimer disease by using 18F-FDG PET of the brain. Radiology 2019, 290, 456–464. [Google Scholar] [CrossRef]
  6. Murugan, S.; Venkatesan, C.; Sumithra, M.; Gao, X.Z.; Elakkiya, B.; Akila, M.; Manoharan, S. DEMNET: A deep learning model for early diagnosis of Alzheimer diseases and dementia from MR images. IEEE Access 2021, 9, 90319–90329. [Google Scholar] [CrossRef]
  7. Kundaram, S.S.; Pathak, K.C. Deep learning-based Alzheimer disease detection. In Proceedings of the Fourth International Conference on Microelectronics, Computing and Communication Systems: MCCS, Ranchi, India, 11–12 May 2019; Springer: Berlin/Heidelberg, Germany, 2021; pp. 587–597. [Google Scholar]
  8. Shahin, O.R.; Gasmi, K.; Krichen, M.; Alamro, M.A.; Mihoub, A.; Ammar, L.B.; Tawashi, M.A. Optimized automated blood cells analysis using Enhanced Greywolf Optimization with integrated attention mechanism and YOLOv5. Alex. Eng. J. 2024, 109, 58–70. [Google Scholar] [CrossRef]
  9. Shen, T.; Jiang, J.; Li, Y.; Wu, P.; Zuo, C.; Yan, Z. Decision supporting model for one-year conversion probability from MCI to AD using CNN and SVM. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, Hawaii, 17–21 July 2018; pp. 738–741. [Google Scholar]
  10. Dwivedi, S.; Goel, T.; Sharma, R.; Murugan, R. Structural MRI based Alzheimer’s disease prognosis using 3D convolutional neural network and support vector machine. In Proceedings of the 2021 Advanced Communication Technologies and Signal Processing (ACTS), Virtual, 15–17 December 2021; pp. 1–4. [Google Scholar]
  11. Beheshti, I.; Demirel, H.; Matsuda, H.; Alzheimer’s Disease Neuroimaging Initiative. Classification of Alzheimer’s disease and prediction of mild cognitive impairment-to-Alzheimer’s conversion from structural magnetic resource imaging using feature ranking and a genetic algorithm. Comput. Biol. Med. 2017, 83, 109–119. [Google Scholar] [CrossRef]
  12. Elahifasaee, F. Optimized SVM using AdaBoost and PSO to classify brain images of MR. In Proceedings of the 2022 International Conference on Machine Vision and Image Processing (MVIP), Ahvaz, Iran, 23–24 February 2022; pp. 1–5. [Google Scholar]
  13. Zhang, Y.; Wang, S.; Phillips, P.; Dong, Z.; Ji, G.; Yang, J. Detection of Alzheimer’s disease and mild cognitive impairment based on structural volumetric MR images using 3D-DWT and WTA-KSVM trained by PSOTVAC. Biomed. Signal Process. Control 2015, 21, 58–73. [Google Scholar] [CrossRef]
  14. Hojjati, S.H.; Ebrahimzadeh, A.; Khazaee, A.; Babajani-Feremi, A.; Alzheimer’s Disease Neuroimaging Initiative. Predicting conversion from MCI to AD using resting-state fMRI, graph theoretical approach and SVM. J. Neurosci. Methods 2017, 282, 69–80. [Google Scholar] [CrossRef]
  15. Archana, B.; Kalirajan, K. Alzheimer’s Disease Classification using Convolutional Neural Networks. In Proceedings of the 2023 International Conference on Innovative Data Communication Technologies and Application (ICIDCA), Uttarkhand, India, 14–16 March 2023; pp. 1044–1048. [Google Scholar]
  16. Dimitriadis, S.I.; Liparas, D.; Tsolaki, M.N.; Alzheimer’s Disease Neuroimaging Initiative. Random forest feature selection, fusion and ensemble strategy: Combining multiple morphological MRI measures to discriminate among healhy elderly, MCI, cMCI and alzheimer’s disease patients: From the alzheimer’s disease neuroimaging initiative (ADNI) database. J. Neurosci. Methods 2018, 302, 14–23. [Google Scholar] [PubMed]
  17. Donnelly-Kehoe, P.A.; Pascariello, G.O.; Gómez, J.C.; Alzheimer’s Disease Neuroimaging Initiative. Looking for Alzheimer’s Disease morphometric signatures using machine learning techniques. J. Neurosci. Methods 2018, 302, 24–34. [Google Scholar] [CrossRef] [PubMed]
  18. Farooq, A.; Anwar, S.; Awais, M.; Rehman, S. A deep CNN based multi-class classification of Alzheimer’s disease using MRI. In Proceedings of the 2017 IEEE International Conference on Imaging Systems and Techniques (IST), Beijing, China, 18–20 October 2017; pp. 1–6. [Google Scholar]
  19. Liu, M.; Cheng, D.; Wang, K.; Wang, Y.; Initiative, A.D.N. Multi-modality cascaded convolutional neural networks for Alzheimer’s disease diagnosis. Neuroinformatics 2018, 16, 295–308. [Google Scholar] [CrossRef]
  20. Khan, N.M.; Abraham, N.; Hon, M. Transfer learning with intelligent training data selection for prediction of Alzheimer’s disease. IEEE Access 2019, 7, 72726–72735. [Google Scholar] [CrossRef]
  21. Mehmood, A.; Maqsood, M.; Bashir, M.; Shuyuan, Y. A deep Siamese convolution neural network for multi-class classification of Alzheimer disease. Brain Sci. 2020, 10, 84. [Google Scholar] [CrossRef]
  22. Orouskhani, M.; Zhu, C.; Rostamian, S.; Zadeh, F.S.; Shafiei, M.; Orouskhani, Y. Alzheimer’s disease detection from structural MRI using conditional deep triplet network. Neurosci. Inform. 2022, 2, 100066. [Google Scholar] [CrossRef]
  23. Abrol, A.; Bhattarai, M.; Fedorov, A.; Du, Y.; Plis, S.; Calhoun, V.; for the Alzheimer’s Disease Neuroimaging Initiative. Deep residual learning for neuroimaging: An application to predict progression to Alzheimer’s disease. J. Neurosci. Methods 2020, 339, 108701. [Google Scholar] [CrossRef] [PubMed]
  24. Ghazal, T.M.; Issa, G. Alzheimer disease detection empowered with transfer learning. Comput. Mater. Contin. 2022, 70, 5005–5019. [Google Scholar] [CrossRef]
  25. Mohi ud din dar, G.; Bhagat, A.; Ansarullah, S.I.; Othman, M.T.B.; Hamid, Y.; Alkahtani, H.K.; Ullah, I.; Hamam, H. A novel framework for classification of different Alzheimer’s disease stages using CNN model. Electronics 2023, 12, 469. [Google Scholar] [CrossRef]
  26. Islam, J.; Zhang, Y. Brain MRI analysis for Alzheimer’s disease diagnosis using an ensemble system of deep convolutional neural networks. Brain Inform. 2018, 5, 1–14. [Google Scholar] [CrossRef]
  27. Abd El Hamid, M.M.; Shaheen, M.; Omar, Y.M.; Mabrouk, M.S. Discovering epistasis interactions in Alzheimer’s disease using integrated framework of ensemble learning and multifactor dimensionality reduction (MDR). Ain Shams Eng. J. 2023, 14, 101986. [Google Scholar] [CrossRef]
  28. Ruiz, J.; Mahmud, M.; Modasshir, M.; Shamim Kaiser, M.; Alzheimer’s Disease Neuroimaging Initiative. 3D DenseNet ensemble in 4-way classification of Alzheimer’s Disease. In Proceedings of the Brain Informatics: 13th International Conference, BI 2020, Padua, Italy, 19 September 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 85–96. [Google Scholar]
  29. Pan, D.; Zeng, A.; Jia, L.; Huang, Y.; Frizzell, T.; Song, X. Early detection of Alzheimer’s disease using magnetic resonance imaging: A novel approach combining convolutional neural networks and ensemble learning. Front. Neurosci. 2020, 14, 259. [Google Scholar] [CrossRef] [PubMed]
  30. Razzak, I.; Naz, S.; Ashraf, A.; Khalifa, F.; Bouadjenek, M.R.; Mumtaz, S. Mutliresolutional ensemble PartialNet for Alzheimer detection using magnetic resonance imaging data. Int. J. Intell. Syst. 2022, 37, 6613–6630. [Google Scholar] [CrossRef]
  31. Li, Y.; Luo, J.; Zhang, J. Classification of Alzheimer’s disease in MRI images using knowledge distillation framework: An investigation. Int. J. Comput. Assist. Radiol. Surg. 2022, 17, 1235–1243. [Google Scholar] [CrossRef]
  32. Hazarika, R.A.; Maji, A.K.; Kandar, D.; Jasinska, E.; Krejci, P.; Leonowicz, Z.; Jasinski, M. An approach for classification of Alzheimer’s disease using deep neural network and brain magnetic resonance imaging (MRI). Electronics 2023, 12, 676. [Google Scholar] [CrossRef]
  33. Fu’adah, Y.N.; Wijayanto, I.; Pratiwi, N.; Taliningsih, F.; Rizal, S.; Pramudito, M. Automated classification of Alzheimer’s disease based on MRI image processing using convolutional neural network (CNN) with AlexNet architecture. In Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2021; Volume 1844, p. 012020. [Google Scholar]
  34. Barulina, M.; Okunkov, S.; Ulitin, I.; Sanbaev, A. Sensitivity of Modern Deep Learning Neural Networks to Unbalanced Datasets in Multiclass Classification Problems. Appl. Sci. 2023, 13, 8614. [Google Scholar] [CrossRef]
  35. Parrales-Bravo, F.; Caicedo-Quiroz, R.; Tolozano-Benitez, E.; Gómez-Rodríguez, V.; Cevallos-Torres, L.; Charco-Aguirre, J.; Vasquez-Cevallos, L. OUCH: Oversampling and Undersampling Cannot Help Improve Accuracy in Our Bayesian Classifiers That Predict Preeclampsia. Mathematics 2024, 12, 3351. [Google Scholar] [CrossRef]
  36. Anwar, R.S.S. EfficientNet algorithm for classification of different types of cancer. In Proceedings of the Artificial Intelligence and Applications, Crete, Greece, 17–20 June 2022. [Google Scholar]
  37. Tan, M.; Le, Q. Efficientnetv2: Smaller models and faster training. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 10096–10106. [Google Scholar]
  38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  39. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  40. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications (2017). arXiv 2017, arXiv:1704.04861. [Google Scholar]
  41. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1–9. [Google Scholar] [CrossRef]
  42. Yang, X.S.; Deb, S. Cuckoo search via Lévy flights. In Proceedings of the 2009 World Congress on Nature & Biologically Inspired Computing (NaBIC), Coimbatore, India, 9–11 December 2009; pp. 210–214. [Google Scholar]
  43. Yang, X.S. Nature-Inspired Metaheuristic Algorithms; Luniver Press: Bristol, UK, 2010. [Google Scholar]
  44. Gandomi, A.H.; Yang, X.S.; Alavi, A.H. Cuckoo search algorithm: A metaheuristic approach to solve structural optimization problems. Eng. Comput. 2013, 29, 17–35. [Google Scholar] [CrossRef]
  45. Wang, G.G.; Deb, S.; Cui, Z. Monarch butterfly optimization. Neural Comput. Appl. 2019, 31, 1995–2014. [Google Scholar] [CrossRef]
  46. Yang, X.S. Cuckoo Search and Firefly Algorithm: Theory and Applications; Springer: Berlin/Heidelberg, Germany, 2013; Volume 516. [Google Scholar]
  47. Valian, E.; Mohanna, S.; Tavakoli, S. Improved cuckoo search algorithm for feedforward neural network training. Int. J. Artif. Intell. Appl. 2011, 2, 36–43. [Google Scholar]
  48. Yang, X.S. Recent Advances in Swarm Intelligence and Evolutionary Computation; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  49. Yilmaz, A.E.; Demirhan, H. Weighted kappa measures for ordinal multi-class classification performance. Appl. Soft Comput. 2023, 134, 110020. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.