Next Article in Journal
Improving Diagnostic Performance for Head and Neck Tumors with Simple Diffusion Kurtosis Imaging and Machine Learning Bi-Parameter Analysis
Previous Article in Journal
MRI Detection of Unknown Primary Tumours in the Head and Neck: What Is the Expected Normal Asymmetry in the Size of the Palatine Tonsils?
Previous Article in Special Issue
Secure Hybrid Deep Learning for MRI-Based Brain Tumor Detection in Smart Medical IoT Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Diagnostic Framework with an Optimized Ensemble of Vision Transformers and Convolutional Neural Networks for Enhanced Alzheimer’s Disease Detection in Medical Imaging

1
Department of Computer Science, Westcliff University, 17877 Von Karman Ave 4th Floor, Irvine, CA 92614, USA
2
Department of Computer Science and Engineering, Lovely Professional University, Phagwara 144411, Punjab, India
3
Department of Computer Science, California State University, 5151 State University Dr, Los Angeles, CA 90032, USA
4
Department of Pharmacology, School of Pharmaceutical Sciences, Lovely Professional University, Phagwara 144411, Punjab, India
5
College of Engineering and Technology, Westcliff University, Irvine, CA 92614, USA
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(6), 789; https://doi.org/10.3390/diagnostics15060789
Submission received: 6 February 2025 / Revised: 12 March 2025 / Accepted: 18 March 2025 / Published: 20 March 2025
(This article belongs to the Special Issue Artificial Intelligence in Brain Diseases)

Abstract

:
Background/Objectives: Alzheimer’s disease (AD) is a progressive, neurodegenerative disorder, which causes memory loss and loss of cognitive functioning, along with behavioral changes. Early detection is important to delay disease progression, timely intervention and to increase patients’ and caregivers’ quality of life (QoL). One of the major and primary challenges for preventing any disease is to identify the disease at the initial stage through a quick and reliable detection process. Different researchers across the world are still working relentlessly, coming up with significant solutions. Artificial intelligence-based solutions are putting great importance on identifying the disease efficiently, where deep learning with medical imaging is highly being utilized to develop disease detection frameworks. In this work, a novel and optimized detection framework has been proposed that comes with remarkable performance that can classify the level of Alzheimer’s accurately and efficiently. Methods: A powerful vision transformer model (ViT-B16) with three efficient Convolutional Neural Network (CNN) models (VGG19, ResNet152V2, and EfficientNetV2B3) has been trained with a benchmark dataset, ‘OASIS’, that comes with a high volume of brain Magnetic Resonance Images (MRI). Results: A weighted average ensemble technique with a Grasshopper optimization algorithm has been designed and utilized to ensure maximum performance with high accuracy of 97.31%, precision of 97.32, recall of 97.35, and F1 score of 0.97. Conclusions: The work has been compared with other existing state-of-the-art techniques, where it comes with high efficiency, sensitivity, and reliability. The framework can be utilized in IoMT infrastructure where one can access smart and remote diagnosis services.

1. Introduction

Alzheimer’s disease (AD) is a complex neurodegenerative disorder that significantly affects memory, cognition, and behavior, primarily in older adults [1,2]. It is the leading cause of dementia, accounting for 60–80% of dementia cases globally, including a significant proportion in the United States [3]. They are characterized by the presence of extracellular amyloid deposits, which are amyloid beta (Aβ), and intracellular neurofibrillary tangles, which are tau protein. In addition to that, AD leads to the death of the brain cells and thus a progressive decline in cognitive function [4,5]. Disease symptoms include forgetfulness and difficulty with tasks; later, patients may get lost, have problems with speech and language, become emotionally unbalanced, and lose capacity for self-care [6]. As shown in Figure 1, early-stage symptoms of AD include memory loss, confusion with time and location, difficulty solving problems, withdrawal from social activities, poor judgments, and changes in vision. Additionally, it explains the main risk elements that contribute to the progressive disease, i.e., age, diabetes, genetic predisposition, obesity, smoking, and brain injuries.
In AD, the level of brain damage increases across various degrees of cognitive impairment during the disease course [7]. As for normal cognitive (NC), the brain is functioning well, but certain alterations might start a long time before the first sign appears [8]. Research shows that patients with NC do not lose as much as 1 percent of their brain volume per year [9]. Cases with Mild Cognitive Impairment (MCI) have a faster rate of shrinkage of the brain, which causes them to lose between 1 and 3% of their brain mass annually, commonly in the hippocampal region responsible for memory [10].
As soon as a person enters AD, brain atrophy becomes much more rapid, the losses reaching from 3 to 5% per year, which indicates active neuronal degeneration [11]. Due to its distinctive neuropathologic profile, the damage in AD especially occurs in locations that control memory, language, and other executive abilities, which are thus characteristically impaired in this disease [12]. Figure 2 shows a comparison of the normal brain and the brain affected with AD blood brain barrier (BBB) disruption and neuroinflammation are central to the pathogenesis of AD and are associated with impaired clearances of Aβ peptides. The end result is the appearance of Aβ plaques and neurofibrillary tangles that induce massive neuronal damage. These changes become pathological and over time result in large volumes of brain shrinkage, such as in the hippocampus (which governs memory and functions of the mind). It further shows a deep learning (DL)-based model for AD detection and classification. The computational approach presented here draws synergies from advanced computational techniques to analyze patterns seen in affected brains for accurate detection and reliable performance validation. The potential for better diagnostics and therapeutics arises from the combination of pathophysiological insights and artificial intelligence (AI)-based detection methods.
In the U.S.A., AD is a major public health problem, with 6 million Americans estimated to live with the disease and expected to increase to nearly 14 million by 2060 [13]. While traditional AD diagnosis, with clinical assessment, history, and neuroimaging, is often inaccurate and too late in diagnosis. Nowadays, approximately two-thirds of seniors are care-receiving ones, and approximately one-third of them die with AD or another type of dementia, which is not surprising at all, given the necessity in terms of the development of better diagnostic and therapeutic means. Most cases of AD occur in people who are over the age of 65, albeit the primary risk factor continues to be age [14,15]. This poses an immense problem for the healthcare providers who would like to administer the therapy before the disease starts developing irretrievable stages [16]. In the U.S., approximately 12–18% of those aged 60 and older are living with MCI, with early detection of MCI being important so that we can intervene before AD becomes full blown [17,18].
Additionally, using AI models, medical professionals can detect vulnerable patients, in particular during the MCI stage, when intervention could stop or deter subsequent cognitive decline [19,20]. In the last few years, Machine Learning (ML) and especially DL have made recent advances in the early detection and diagnosis of AD [21]. We have shown that ML techniques have the potential to analyze large datasets and identify patterns that can be difficult to observe by clinicians [22,23]. Newer-generation technologies are being used to analyze many kinds of data, including genetic information, imaging data, and cognitive scores, in order to accurately predict risk and initiate earlier interventions [24]. Furthermore, deep learning models that have exhibited the capability of processing complex datasets like MRI, PET scans, and even other more complex neuroimaging data have substantially improved AD detection accuracy [25]. Despite these advances, challenges remain, particularly in making these models interpretable and explainable to clinicians, as the “black box” nature of many ML models hinders their widespread adoption in real-world clinical settings [26]. In the U.S., numerous research initiatives and studies are focused on using ML and DL models to improve AD detection [27]. Such models can thus discriminate between different cognitive states, for instance, NC, MCI, and AD [28]. There are researchers working on creating explainable artificial intelligence (AI) so these models can not only predict outcomes but also explain why these predictions are so, so that these models can effectively explain the factors leading to them [29]. For instance, we have seen recent studies employ approaches such as SHapley Additive exPlanations (SHAP) to provide both global and local explanations of the predictions made by the models to help the clinician know which features most influenced the decisions [30,31]. However, additional work will be needed to enhance the transparency, interpretability, and clinical applicability of these models.
DL techniques have been explored in this work to classify AD, leveraging a review of the recent works, challenges, and a growing literature in the USA [32]. It will also evaluate the status of current approaches and the future promise of AI-driven diagnostic tools in the application to AD patient management in the clinical setting [33]. By using large datasets, multimodal information, and interpretable AI ways, researchers want to get more accurate, more reliable models to predict the progression of AD and provide early intervention and better outcomes [34]. CNN models are very efficient and are widely being used for image classification-based tasks where various CNN-based diagnosis frameworks have been developed and applied in medical imaging. Although CNN is good at classification tasks, ViT or vision transformer can handle the task more efficiently, where it has more advantages over CNN [13]. ViT uses patches and a self-attention mechanism that allows it to have a long-range relationship. ViT has good scalability with less inductive bias. CNN extracts features from input images in a hierarchical manner where ViT follows global feature learning. Although CNN is widely popular and is highly efficient for image classification applications, ViT comes with more advantages when the dataset is large with a huge volume of instances [31,32]. Ensemble technique helps combine the performance of each model and offers better performance than using the models individually. In a weighted average ensemble, weight is assigned based on the individual performance that can use the combined potential of each model performance-wise and comes with effective results [16].
In this work, the following have been performed:
(1)
An extensive literature study has been performed where related works have been thoroughly studied to know about the existing state-of-the-art techniques and to identify current research gaps.
(2)
A classification framework has been designed based on vision transformer (ViT-B16) and three fine-tuned CNN models (VGG19, ResNet152V2, and EfficientNetV2B3) with modified dense layer architecture trained with almost 0.234 million brain MRI images.
(3)
A weighted ensemble approach with Grasshopper optimization algorithm applied on developed vision transformer and CNN models to build an efficient, robust, and remarkable framework for multiclass classification of Alzheimer’s level.
(4)
An innovative model of AI integrated IoMT architecture has been designed and proposed where the designed classification framework applied in cloud server can play a significant role in telemedicine where one can access smart diagnosis service remotely.
(5)
A comparative analysis with proper facts and figures has been performed with the existing state-of-the-art work that ensures the strengths and novelty of the framework in terms of architecture design and performance.
In Section 2, a literature review is conducted, where recent related work with their key findings and limitations are presented and properly explained. Section 3 deals with the methodology, where design and development of the proposed framework are presented. Section 4 shows the results of the experimental analysis and discussion regarding the facts and figures regarding the performance of the proposed work. Lastly, a conclusion is drawn in Section 5, where a summary of the proposed work is presented with future scope.

2. Related Work

Innovative approaches and technologies that have been adopted for the identification of AD have been illustrated in a recent study. Yogesh et al. [19] analyzed the effectiveness of using ML algorithms in KNN and SVM for diagnosing mental disorders by assessing the output generated by MRI images; moreover, they found that KNN is more accurate than SVM in this process. Such technologies are vital for diagnosis in the early stage [19]. Altogether, the current study contributed to an affordable merged platform that incorporates the cognitive assessments with the noninvasive physiological monitoring to document the advancement of AD by Panda et al. [20]. As Qu et al. [21] suggested, modifying the immunostaining process to a univariate neurodegeneration biomarkers approach that enhances early identification and using it in the GCN model that authors identified provided high percent classification between cognitively impaired and nonimpaired partakers in the study [21]. Tushar et al. [22] proposed a novel hybrid model combining both logistic regression and decision trees that has increased the prediction value greatly, employing data from the OASIS database. The utility of DL techniques was tested by Alshammari et al. [23], who used a CNN model for accurate differentiation between the stages of AD from the MRI data with a high classification rate. Thatere et al. [24] pointed out the important aspects of biomarkers and interventions in ML strategies for AD diagnosis and highlighted data deficits. In another study, Sevcik et al. (2022) [25] were able to note through a systematic review of conversational datasets for AD prediction that increased data quality outweighs increased quantity for predicting the condition using natural language processing. The two related studies of Gurrala et al. [26] and Kumar et al. [35] pointed out that the conventional architecture of CNN, which facilitates MRI data handling, offers extremely high classification rates. Zubair et al. [36] rightly highlighted the role of the early detection strategies; Saxena et al. [37] narrated the recurrent DL progress while also noting the current existing challenges with practitioners. Classification of neuroimaging datasets was the subject of Archana B et al. [38], who recorded an impressive accuracy that serves as an indication that intervention should be promptly initiated. Liu [39] enhanced and expanded the classification by integrating the hippocampal MRI with the whole-brain MRI using an attention-enhanced DenseNet model. Han et al. [40] proposed an optimization in the feature extraction methods using LightGBM because of the high and better classification accuracies noticed in real datasets. Functional MRI is also used, and in a recent study, Maitha Alarjani [41] used a patch-based approach, in which functional MRI is correct; thus, there is a need to find key features for efficient diagnosis urgently. Sanjeev Kumar et al. [42] listed the CNN architectures to outperform the others in stages of AD classification, and Mishra et al. [43] described the early prediction problems of the different datasets for the ML approach. Lu et al. [44] introduced a multiclassification framework to enhance the measurement accuracy, which would in turn minimize the imaging overlap between AD and MCI using the ConvNeXt network. In the work of Prajapati et al. [45], deep neural solutions were identified for binary classification with deep neural networks to achieve over-average validation accuracy rates compared to other methods. Shi et al. [46] attempted to use a generative adversarial network framework to enhance diagnosis for tau PET images, noting that there is individual region analysis that needs to be performed. A very comprehensive overview of the role of AI in qualitative improvement for AD diagnosis using medical imaging was presented by Arulprakash A et al. [29]. According to Neetha et al. [47], the Borderline-DEMNET framework is suggested, and a good degree of accuracy is evaluated where stages of AD are concerned. Chander Prabha [48] also laid emphasis on early detection following an optimal MRI scanning method while achieving the desired % classification. MRI plays an important role in the detection of structural abnormalities, remarked Ruchika Das [49], with hippocampal segmentation providing better patient classification. Yuan et al. [50] compared different conventional ML methods and successfully classified them based on cortical features; Navarro et al. [51] established daily living activity tasks as possible screening tools for the assessment of cognitive impairments in the AD population. In the perspective of continued advancement of identification procedures, Alatrany et al. [52] and Bhargavi et al. [32] also emphasized the important contribution of numerous ML algorithms on enhancing the diagnostic efficiency as well as monitoring of AD stages. Wang et al. [53] attempted to discuss the features of DL in AD detection by using sMRI. In this review, the methodologies proposed were categorized based on input types, and issues regarding the difficulty of spatial modeling and variability of performance were also discussed while addressing future recommendations for further DL diagnostic models. The study by Kayalvizhi et al. [54] has identified that while using the mass VGG16 CNN model, the possibilities of accurate diagnosis of AD from MRI reach 96.75% for the DL model, which means it is applicable in the neuroimaging analysis. According to Akter et al. [55], different types of dementias were identified using the Extreme Gradient Boosting model, with an accuracy of 81% for the feature “ageAtEntry’“’ as the biomarker for the first diagnosis of dementia. In the study of Pallawi et al. [56], extreme emphasis was placed on the early diagnosis using a framework using transfer learning, and combining it with the EfficientNetB0 model achieved 95.78% early efficiency in classifying four stages of AD that achieved general milestones in the dataset for DL applications. Islam et al. [57] studied MRI’s contribution to AD diagnosis through a focus on the hippocampus and put forward YOLO-based models for automated hippo detection whereby positive outcomes have been achieved, helping to advance early recognition. Zhou et al. [58] designed a Chinese game application that combines cognitive tests, which may reflect patterns of cognitive decline useful for early AD identification. Rauniyar et al. [59] used NLP to review the rhetorical fluency in AD patients to set up a referential platform for further comparisons. Feed-forward neural networks have been analyzed by Patra et al. [60] on the OASIS dataset to make assessments that feed-forward neural nets are capable of predicting AD better. The various ML algorithms were reviewed and compared by Varma et al. [61] for their study’s purpose of patient data classification to achieve maximum predictive accuracy in identifying the presence of AD. MCI was split into three subclasses by Jiang et al. [62] using KNN accompanied with an enriched LSTM neural network for higher accurate disease prognosis. Bushra et al. [63] proposed a feature-level fusion approach for which two CNN architectures were used and received accuracies of 94.39% for MCI and 97.90% for AD as compared to the normal control group and were higher than current methods. Subha R et al. [64] have also stressed the need to bring about more efficient early AD diagnosis through the use of a novel hybrid ML model with higher accuracy achieved through the integration of particle swarm optimization. More recently, Buyrukoglu [65] underlined the need for early detection, demonstrating that this is possible with the help of ensemble learning methods, the best of which reached a 92.7% classification rate using the AdaBoost ensemble. Tripathy et al. [33] proposed a new architecture of multilayer feature fusion-based deep CNN for better prediction and found better accuracy of 95.16% with multiple scale features. Chandra et al. [66] tested the Naïve Bayes classifier using the brain volume percentages as extracted biomarkers and reported reasonably good classification performance. A new CNN model for the diagnosis of AD has been developed recently by Talha et al. [67] with sufficient accuracy and F1 score for the early tracking. Bharath et al. [68] have categorized and analyzed the various ML modeling techniques, where they expounded that the utilized SVM model was helpful in early AD and was accompanied with a high accuracy ratio of 98%. According to Sindhu et al. [69], neural network methodologies for AD detection need elaboration, and automated methodologies that improve the degree of precision should be advocated. Jansi et al. [70] analyzed some DL models, InceptionV3, which has achieved an 87.69% accuracy in their maximization of dataset sets. Yin et al. [34] introduced the novel SMIL-DeiT network for classifying stages of AD, reaching an accuracy rate of 93.2%, which is superior to the existing approaches. In Sushmitha et al. [71], a genetic algorithm with multi-instance learning has been used for separate 3D MRI feature extraction and overfitting problems. Kadyan et al. [72] tested speech signal patterns for the early identification of AD and recorded an accuracy of 98.3% using a random forest classifier. In order to understand and overcome the deficiencies of the existing studies about stage-specific brain pattern analysis, Peng et al. [73] put forward the concept of the SF-GCL model. According to STCNN proposed by Anjali et al. [74], the model showed shorter average execution time, superior accuracy, and F1 scores in the identification of AD. Yu et al. [75] have put forward a method known as a supervised deep tree model (SDTree) to solve this problem because current diagnostic technology cannot capture the continually progressive nature of AD. Specifically, by employing the nonlinear reversed graph embedding, their proposed approach deems the progression of AD as a latent tree structure model to enhance the prediction and achieve favorable performances on the ADNI dataset. Finally, Jin et al. [76] described an unsupervised DL model using adversarial autoencoders to identify brain atrophy with 94% accuracy as a classifier. Soni et al. [77] proposed a verb fluency task for the detection of AD with 76% accuracy using random forest, whereas Suttapakti et al. [78] improved feature extraction for AD classification using two-dimensional variational mode decomposition with 94.44% accuracy. Basher et al. [79] used the volumetric features from hippocampi, going up to 94.82% accuracy, and added more to the development of the diagnostic tools for AD. Altogether, the results of recent research investigations signal the paramount importance of early and correct identification, which is the most important for intervention approaches. Further, as the models are improved and new methodologies are devised by researchers, the potential to increase the diagnostic accuracy and ultimately, the benefits of diagnosis to the points of care, increases dramatically. Thus, AD detection now and in the future is laying great emphasis on the integration of the most advanced technologies, such as ML and DL. Table 1 shows a summarized representation of the existing related works.

3. Proposed Methodology

In this section, a brief representation has been given regarding the proposed framework, including dataset information, preprocessing, model architectures, design, and development of a classification framework with functional working methods, etc.

3.1. Dataset Description

To develop deep learning models, a huge amount of data is required. To ensure consistent performance with high sensitivity and reliability, data should be collected from a renowned and trusted source. To train models for developing the proposed framework, ‘OASIS Alzheimer’s Detection’ dataset was used. The collected dataset is considered an authentic and benchmark dataset that has been collected from Kaggle, which is licensed under Apache 2.0. The dataset includes 87,390 MRI images consisting of four classes: Mild dementia (5002 images), Moderate dementia (488 images), Nondementia (67,200 images), and very mild dementia (13,700 images). Figure 3 shows a sample image from the no dementia class from the same dataset of 24 × 24 size.

3.2. Preprocessing

To enhance the performance of the model through reducing data complexity, preprocessing plays a vital role. Data needs to be preprocessed before being fed to models. Preprocessed data ensures promising performance, removing issues like overfitting. The input images from the dataset are of 176 × 208 size, which must be resized before fitting. Moreover, the DL model works well with a large amount of data where augmentation techniques can play vital roles.
Figure 4 shows preprocessed data samples having the same size for all classes. Augmentation technique helps to produce two types of data—synthetic data and augmented data. Synthetic data is a type of artificially created data that has been created through generating algorithms. Augmented data are generated by modifying the existing data. Synthetic data is required to be generated when original data instances are very few in the original dataset. In the selected dataset, almost 87,390 images are already there for which synthetic generation is not required. Augmented data has been generated after applying several image transformation techniques like rotating, zooming, flipping, shearing, shifting, cropping images, etc. The third class has already 67,200 images; that is why augmentation was not performed for that class. Class 4 (Very Mild Dementia) has 13,700 images, which have been increased five times to get 68,200 images. Class 2 (Moderate Dementia) has only 488 images that must generate a class imbalance issue.
Class 4 has been scaled up to 20 times and gets 48,800 images. Class 1 (Mild Dementia) has been increased 10 times and comes with 50,020 images. The comparison has been presented in Figure 5 to indicate the difference between before and after augmentation. Before applying the data augmentation technique, there are a total of 87,390 images, and after applying the augmentation technique, the number becomes 234,220, which is 2.6 times more than the amount of the original dataset offers.

3.3. CNN-Based Classification

Convolutional neural networks are widely used in image classification applications for rich and remarkable performance. CNN architecture is basically divided into two parts—feature extraction and classification. The feature extraction part is responsible for extracting the required features from the input images. The classification part classifies the images based on the received features. Figure 6 shows the basic architecture of CNN, where convolutional and pooling layers are there for the feature extraction part, and the dense layer is responsible for the classification part.
Table 2 shows a comparative analysis of three CNN models- VGG19, ResNet152V2, and EfficientNetV2B3, based on various parameters [35].
To develop the proposed farmwork, besides using a vit-b3 model, three CNN models have been developed, which are VGG19, ResNet152V2, and EfficientNetV2B3.
VGG19, ResNet152V2, and EfficientNetV2B3 have been selected for this study because each model brings unique strengths to medical image classification. VGG19 is known for its deep and straightforward architecture that makes feature extraction effective. ResNet152V2, having its residual connections, overcomes issues like vanishing gradients and assists better training in deep networks. EfficientNetV2B3 offers efficient scaling, maintaining a balance between model speed, size, and accuracy. These three models were found ideal, whereby combining the three models with their complementary strengths—efficiency, depth, and performance—a well-rounded and robust framework can be developed for medical image classification.
VGG19 shows more promising performance than the other two CNN models. Table 3 shows the shape and number of parameters where the feature extraction part was taken from the predefined VGG19 architecture, which is mentioned as functional. The dense layer of the pretrained model was modified according to the desired output requirements, which is shown in Figure 7.

3.4. Architecture of Vision Transformer

Transformers are deep learning models that are designed for handling sequence-to-sequence tasks, being widely and effectively used in natural language processing (NLP) and computer vision. To analyze the relationships between elements in a sequence, transformers use a self-attention technique. Unlike traditional recurrent models, transformers can perform parallel processing to improve efficiency. Application in text generation, language translation, and image recognition, different prominent models like BERT, GPT, and Vision Transformer (ViT) are being used, which were built based on transformers. Their scalability and effectiveness have revolutionized AI applications across multiple domains. Figure 8 shows the basic architecture of the transformer, which has two parts—the upper one is the encoder and the lower one is the decoder. The input data is transformed into vector representations for each word through an embedding layer.
Since transformers do not process sequences in order, positional encodings are added to these embeddings to preserve the word order. The core of the attention mechanism relies on three learnable vectors: Query, Key, and Value, where a query is compared with a key to retrieve a corresponding value. The dot product of Query and Key creates a score matrix, showing how much attention each word should give to others, with higher scores indicating more attention. The scores are scaled to prevent large values that could destabilize gradients. Next, the scores are passed through softmax, turning them into probabilities that highlight the most important words. These probabilities are then multiplied by the value vectors, allowing the keywords to have a stronger influence. The resulting vectors are concatenated and passed through a linear layer for further processing. Since attention is calculated independently for each word, multiple parallel attention heads are used to capture different relationships in the data. Outputs are added to residual connections (to aid gradient flow) and normalized with LayerNorm to stabilize the learning process.
The output then passes through a feed-forward network for a richer representation. Another round of residual connections and LayerNorm ensures the output remains stable. This completes the encoder, which creates a full representation of the input sequence. The decoder uses the encoder’s output and previous time-step inputs to generate the final sequence, applying masked multiheaded attention to prevent “leakage” of future words. During decoding, the attention scores for future words are masked, so the model can only focus on past and present words. Finally, the decoder’s output is processed through a linear layer and softmaxed to produce the final predicted probabilities.
Figure 9 represents the architecture of the vision transformer, where the vision transformer uses the encoding part of the transformer. The input image is broken into small patches. In the proposed work, input images are broken into 64 patches with 8 × 8 dimension, although in the figure, images are broken into 16 patches only just for simple and better understanding. The patches are then embedded with a flattened representation for further processing.
Unlike traditional neural networks, this model does not have a built-in understanding of the position of each sample in the sequence—each sample is a patch from the image. The image is processed along with a learnable positional embedding and passed into the encoder. The positional embeddings are trained during the process, rather than being predefined. Similar to BERT, a special token is included at the start. Each image patch is flattened into a vector and then multiplied by a learnable embedding matrix to create embedded patches. These embedded patches are combined with the positional embeddings and fed into the Transformer. The key distinction is that, instead of using a decoder, the encoder’s output is directly passed into a feed-forward neural network to generate the classification output.
Table 4 presents the number of parameters and shape of the ViT-B16 model. There are a total of 85,772,900 parameters counted, among which 85,771,108 parameters are considered as trainable. Figure 10 shows the layer-based flow diagram where the architecture consists of four dense layers.

3.5. Optimization and Hyperparameter Tuning

Optimization is important to reduce the model complexity as well as eliminate issues like overfitting and underfitting. To get maximum accuracy with minimum loss, hyperparameter tuning plays important roles where each parameter has to be considered carefully, assigning the most optimum values. However, it is difficult to assume which values should be appropriate for which parameters, and here the issue arises. The entire process follows a hit-and-trial approach where each value should be taken and tested the result accordingly. It would be a time-consuming process to check all parameters with all the possible values. Here the optimization algorithm comes that extends the ability to tune the model automatically rather than checking the value manually to get maximum output. Here in this work, the Grasshopper optimization algorithm has been utilized to find the best possible values for certain parameters. The parameters used in the work are batch size, epoch number, dropout rate, learning rate, and patch size. The lower bound and upper bound values for learning rate, dropout rate, batch size, and epoch are
Lower bound, LB = [0.00001, 0.1, 16, 5]
Upper bound, UB = [0.01, 0.6, 128, 30]
Best fitness value represents the optimum solution during the optimization process that finds the minimum or maximum values of the objective function depending on the minimization or maximization problem.
The proposed problem is a minimization one where the smallest fitness value is considered the best. Here the fitness value is calculated with the following formula:
Fitness value = Validation loss − 0.1 × Validation accuracy
To develop vision transformer-based models, patch size needs to be mentioned. Figure 11 shows an input MRI image with respective patch images of (8, 8) where a total of 64 patches are there. Number of patches can be calculated through the following formula:
Number of patch size = (Size of input image // patch size)
Table 5 presents the detailed parameters of each model used in the proposed framework development. From Table 5, it can be observed that the epoch has been taken to 10, 16, 20, and 25, and the learning rate has been taken to 0.001, 0.01, 0.001, and 0.00001 for ViT-B16, VGG19, Reset153V2, and EfficientNetV2B3.

3.6. Weighted Average Ensemble Technique

Weighted Average Ensemble (WAE) involves integrating trained models and assigning weights based on their performance to enhance the system’s overall accuracy. Four models—ViT-B16, VGG19, ResNet152V2, and EfficientNetV2B3—were individually trained, and their outputs were later combined using weighted values (WV1, WV2, WV3, and WV4) ranging from 0.1 to 0.4. The weights were determined based on the performance priority of each model, with the best-performing model receiving the highest weight. The predictions of each model were weighted through a tensor dot operation to produce the final combined output.

3.7. Systematic Designing Approach

The proposed framework has been developed through the combination of three CNN models and 1 ViT. Figure 12 shows the flow diagram for each step of the proposed framework. The very first step is to load data instances from the selected dataset. Once all the data are loaded, preprocessing has to be performed. Preprocessed data has to be split into two parts—train data and test data. Training and testing data are used for training and testing the models. Four models were trained separately with the trained data first. The Grasshopper optimization algorithm helps to find out the best parameters to fetch the best performance. Accordingly, the performance of each model, weight values—WV1, WV2, WV3, and WV4—have been assigned for four models M1 (vit-16), M2 (VGG19), M3 (ResNet152V2), and M4 (EfficientNetV2B3), respectively. After assigning weights, the ensemble technique is applied, where the combination of the performance of each model is utilized. Finally, validation of the ensemble model is performed with test data, and the classification result is obtained as the output.

3.8. Proposed Algorithm

IS: image size, DS: brain MRI Alzheimer dataset, E: number of epoch, BS: batch size, NCL: number of classes, TP: train folder path, PS: patch size, NPS: number of patch size, M1: ViT_b16, M2: VGG19, M3: ResNet152V2, M4: efficientNetV2B3, OPT: optimizer, lb: lower bound, ub: upper bound, LR: learning rate, AV: activation function (Algorithm 1).
Algorithm 1. DL-based Alzheimer’s disease detection approach using brain MRI images
Diagnostics 15 00789 i001
Diagnostics 15 00789 i002
Diagnostics 15 00789 i003

3.9. IoMT Infrastructure with Proposed Classification Framework

The Internet of Medical Things or IoMT, is a network of devices to access healthcare information through the internet. Data from IoMT devices can be shared remotely to anywhere, processed online, and accessed wirelessly. IoMT plays a significant role in telemedicine, where one can get healthcare services remotely.
Figure 13 shows a distributed network connected with a central cloud server, which is accessible to each user through the internet. The cloud server is integrated with an AI module where our proposed algorithm is defined. If any MRI images are received from any sources to the cloud server, it can classify the images instantly and send the images to the respective healthcare facilities accordingly. To make the scenario simple, three cases of data-sending sources have been shown and mentioned in the figure—MRI images from the hospital, MRI images from Personal IoMT devices, and MRI images from diagnostic labs. Thus, in any case, an MRI image is generated and sent to the main cloud server through the internet. The image is received and processed according to the mentioned steps shown in the AI module in the figure and classified into a level of dementia. The AI module is integrated with our proposed classification framework that is able to classify MRI images to find the dementia level. Firstly, the received data is resized as the model’s input size of our proposed classification framework. Then, the processed data is fed to the proposed classification model and generates the targeted classification output. Based on the classification report, images with the report will be sent to the respective healthcare persons, research facilities, or hospitals. Responses or feedback data received from the hospitals, research labs, or individual doctors are delivered to the respective end user accordingly. With the proposed infrastructure, one can get to know the level of dementia as well as get proper information or advice from physicians remotely.

4. Results and Discussion

In this section, the results of each experiment have been analyzed and properly presented to support the strength and novelty of the proposed approach. A brief discussion is presented based on the observation, and the comparative analysis has been performed with the existing state-of-the-art technique.
Table 6 presents the training accuracy, training loss and validation accuracy, and validation loss with each epoch. All models, including CNN and vit are optimized with GOA, where the best performances were obtained with the best parameter values. Table 7 shows the optimum parameters’ values to achieve maximum performance. Figure 14 presents confusion matrices individually of all the models used in the proposed framework, where Figure 15 shows the confusion matrix of the ensemble model.
Table 8 shows the performance metrics individually for each class, and Figure 16 shows the overall performances through a bar graph.
Table 9 shows the resulting values of training and testing accuracies for each model, where it is easily identifiable that the ensemble model with optimization comes with better performance with 97.31% accuracy. The processing time during the training is also less than in other models, which ensures the time efficacy of the proposed approach. Table 10 shows weight distribution, where four weight values (WV) have been assigned to all the four models (ViT-B16, VGG19, ResNet152V2, and EfficientNetV2B3) to combine performances. At the very first combination, it is seen that WV1 For ViT-B16 got a low value, and WV4 for EfficientNetV2B3 received high values where ViT-B16 performed much better than EfficientnetV2B3. The resultant ensemble accuracy was not increased much for the same reason. At the last attempt, WV1 got a high weight value, and WV4 got a low value, presenting the perfect combination of weights where maximum ensemble accuracy was obtained.
Table 11 shows the 5-fold cross-validation method for analyzing the consistency of the performance. The data have been divided into five bins, and for each attempt, each bin is applied to check the result. After getting five different accuracies for each model, mean, variance, and standard deviation are calculated to check the data variability. Table 12 shows the final performance metrics obtained with test data, where accuracy, precision, recall, and F1 score are given. Table 13 shows the accuracy and loss status of the proposed ensemble model during training and validation with and without the optimization technique. It is clearly observed that at the 10th epoch, maximum accuracy and minimum loss are achieved with the optimization technique.
Table 14 shows the comparative analysis between the proposed work and the state-of-the art work.
Figure 17 shows the accuracy and loss graph, where the first plot shows the training and validation accuracy, and the second plot shows the train and validation loss with the epoch increment. Generally, loss gets decreased and accuracy gets increased with epoch increment [83]. Initially, loss was very high and accuracy was very low. After increasing the epoch number, accuracy increased and loss decreased. At the 10th epoch, the maximum accuracy with minimum loss was recorded. ViT-B16 is a heavyweight model that requires a lot of computational time with complex operations. To utilize training time, an early stopping function was applied that stops the epoch iteration at a value of 10, which assumes that at the 10th epoch, models were well trained to perform with good accuracy and less loss.
Figure 18 shows the accuracy and loss performance obtained by the VGG19 model, where at the 16th epoch, the highest accuracy and lowest loss were received. In the first plot, training accuracy performs better than validation accuracy, which indicates issues like model overfitting. In Figure 19, it goes to the 20th epoch and records 89% train accuracy with 0.7% train loss.
In Figure 20, the maximum train accuracy of 84% was received at the 25th epoch, where the minimum loss was received at the same epoch value. Although the training and validation curves in this figure show consistent performance, both the train and validation accuracies achieved at the 25th epoch are lower than those of other models.
Observing the performance from Figure 18, Figure 19 and Figure 20, it is found that both VGG19 and ResNet152V2 offer quite similar performance, where EfficientNetV2B3 comes with lower accuracy for both training and validation.
Figure 21 represents the accuracy loss graph of the proposed ensemble model during training and validation. The comparison has been drawn between optimized and nonoptimized performance, where the optimized ensemble model comes with promising performance, ensuring minimum loss and maximum accuracy.

5. Conclusions

In conclusion, despite being a severe public health threat, AD has a complex pathology and remains challenging for memory, cognition, and behavior. Early detection and intervention, most importantly in the MCI stage, is the best chance to slow the disease’s progression. Applications of ML and DL via the analysis of multimodal data, such as neuroimaging data like MRI and CT scans, have shown promising results for improving diagnostic accuracy. Besides helping to detect such diseases early, these technologies also lead to more personalized and effective interventions. To some extent, these are accomplished, but there remains much to do, especially in developing interpretable and simple-to-deploy AI models in the clinical setting. However, by overcoming these hurdles and further fine-tuning predictive models, we could introduce AI in AD care to transform patient care. Through ongoing research and collaboration, these AI-driven tools can help create better outcomes for those with AD and reduce the burden that this disease potentially can cause. In this work, an optimized classification framework was developed, applying an ensemble technique, utilizing the power of vision transformers and three CNN models. The work offers promising performance, including high accuracy of 97.31%, precision of 97.32, recall of 97.35, and F1 score of 0.97. However, issues like image overlapping may cause feature redundancy where CNN models cannot identify the unique spatial features from images. The study relies solely on secondary data collection, whereas incorporating primary data collection could help address the limitations of the classification framework and enhance its robustness. This study can be found valuable for AI-based healthcare research, wherein, in the future, it can put important significance on the further development of a neuroimaging-based classification approach with novel and innovative features. Embedded devices can be implemented and integrated with MRI machines, through which classification can be performed at the same time as image generation.

Author Contributions

Conceptualization, J.C.B., G.S.C., S.B. and J.D.; Methodology, J.C.B., G.S.C., I.R.N., S.B., J.D. and K.K.B.; Validation, G.S.C., S.B., J.D., K.K.B., M.T.R.T. and A.I.; Formal analysis, G.S.C., J.D., M.T.R.T. and A.I.; Investigation, G.S.C. and A.I.; Data curation, J.C.B., G.S.C., J.D., K.K.B., M.T.R.T. and A.I.; Writing—original draft, J.C.B., G.S.C. and J.D.; Writing—review & editing, J.C.B., G.S.C., I.R.N., S.B., J.D., K.K.B., M.T.R.T. and A.I.; Visualization, J.C.B., G.S.C., I.R.N., S.B., J.D., K.K.B., M.T.R.T. and A.I.; Supervision, S.B. All authors have read and agreed to the published version of the manuscript.

Funding

This study was self-funded by the authors. No external funding sources were involved.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was waived as the study utilized publicly available secondary data.

Data Availability Statement

The data presented in this study are openly available in Kaggle “OASIS Alzheimer’s Detection” at https://doi.org/10.1162/jocn.2007.19.9.1498. “Data were provided 1-12 by OASIS-1: Cross-Sectional: Principal Investigators: D. Marcus, R, Buckner, J, Csernansky J. Morris; P50 AG05681, P01 AG03991, P01 AG026276, R01 AG021910, P20 MH071616, U24 RR021382”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Abubakar, M.B.; Sanusi, K.O.; Ugusman, A.; Mohamed, W.; Kamal, H.; Ibrahim, N.H.; Khoo, C.S.; Kumar, J. Alzheimer’s Disease: An Update and Insights Into Pathophysiology. Front. Aging Neurosci. 2022, 14, 742408. [Google Scholar] [CrossRef]
  2. Sehar, U.; Rawat, P.; Reddy, A.P.; Kopel, J.; Reddy, P.H. Amyloid Beta in Aging and Alzheimer’s Disease. Int. J. Mol. Sci. 2022, 23, 12924. [Google Scholar] [CrossRef]
  3. Gustavsson, A.; Norton, N.; Fast, T.; Frölich, L.; Georges, J.; Holzapfel, D.; Kirabali, T.; Krolak-Salmon, P.; Rossini, P.M.; Ferretti, M.T.; et al. Global estimates on the number of persons across the Alzheimer’s disease continuum. Alzheimer’s Dement. 2023, 19, 658–670. [Google Scholar] [CrossRef]
  4. Rahman, M.M.; Lendel, C. Extracellular protein components of amyloid plaques and their roles in Alzheimer’s disease pathology. Mol. Neurodegener. 2021, 16, 59. [Google Scholar] [CrossRef] [PubMed]
  5. Zhang, H.; Wei, W.; Zhao, M.; Ma, L.; Jiang, X.; Pei, H.; Cao, Y.; Li, H. Interaction between aβ and tau in the pathogenesis of alzheimer’s disease. Int. J. Biol. Sci. 2021, 17, 2181–2192. [Google Scholar] [CrossRef]
  6. Tenchov, R.; Sasso, J.M.; Angela Zhou, Q. Alzheimer’s Disease: Exploring the Landscape of Cognitive Decline. ACS Chem. Neurosci. 2024, 15, 3800–3827. [Google Scholar] [CrossRef]
  7. Rajeev, V.; Fann, D.Y.; Dinh, Q.N.; Kim, H.A.; De Silva, T.M.; Lai, M.K.P.; Chen, C.L.H.; Drummond, G.R.; Sobey, C.G.; Arumugam, T.V. Pathophysiology of blood brain barrier dysfunction during chronic cerebral hypoperfusion in vascular cognitive impairment. Theranostics 2022, 12, 1639–1658. [Google Scholar] [CrossRef]
  8. Xiong, Y.; Chen, X.; Zhao, X.; Fan, Y.; Zhang, Q.; Zhu, W. Altered regional homogeneity and functional brain networks in Type 2 diabetes with and without mild cognitive impairment. Sci. Rep. 2020, 10, 21254. [Google Scholar] [CrossRef]
  9. Fujita, S.; Mori, S.; Onda, K.; Hanaoka, S.; Nomura, Y.; Nakao, T.; Yoshikawa, T.; Takao, H.; Hayashi, N.; Abe, O. Characterization of Brain Volume Changes in Aging Individuals With Normal Cognition Using Serial Magnetic Resonance Imaging. JAMA Netw. Open 2023, 6, E2318153. [Google Scholar] [CrossRef]
  10. Givian, H.; Calbimonte, J.P.; Neuroimaging, D. Early diagnosis of Alzheimer’s disease and mild cognitive impairment using MRI analysis and machine learning algorithms. Discov. Appl. Sci. 2025, 7, 27. [Google Scholar] [CrossRef]
  11. Breijyeh, Z.; Karaman, R. Comprehensive Review on Alzheimer’s Disease: Causes and Treatment. Molecules 2020, 25, 5789. [Google Scholar] [CrossRef] [PubMed]
  12. Zhang, Y.; Wu, K.M.; Yang, L.; Dong, Q.; Yu, J.T. Tauopathies: New perspectives and challenges. Mol. Neurodegener. 2022, 17, 1–29. [Google Scholar] [CrossRef] [PubMed]
  13. Bezprozvanny, I. Alzheimer’s disease–Where do we go from here? Biochem. Biophys. Res. Commun. 2022, 633, 72–76. [Google Scholar] [CrossRef] [PubMed]
  14. Tahami Monfared, A.A.; Byrnes, M.J.; White, L.A.; Zhang, Q. Alzheimer’s Disease: Epidemiology and Clinical Progression. Neurol. Ther. 2022, 11, 553–569. [Google Scholar] [CrossRef]
  15. Fang, E.F.; Xie, C.; Schenkel, J.A.; Wu, C.; Long, Q.; Cui, H.; Aman, Y.; Frank, J.; Liao, J.; Zou, H.; et al. A research agenda for ageing in China in the 21st century (2nd edition): Focusing on basic and translational research, long-term care, policy and social networks. Ageing Res. Rev. 2020, 64, 101174. [Google Scholar] [CrossRef]
  16. Sethi, M.; Rani, S.; Singh, A.; Mazón, J.L.V. A CAD System for Alzheimer’s Disease Classification Using Neuroimaging MRI 2D Slices. Comput. Math. Methods Med. 2022, 2022, 8680737. [Google Scholar] [CrossRef]
  17. Bai, W.; Chen, P.; Cai, H.; Zhang, Q.; Su, Z.; Cheung, T.; Jackson, T.; Sha, S.; Xiang, Y.T. Worldwide prevalence of mild cognitive impairment among community dwellers aged 50 years and older: A meta-analysis and systematic review of epidemiology studies. Age Ageing 2022, 51, afac173. [Google Scholar] [CrossRef]
  18. Rajan, K.B.; Weuve, J.; Barnes, L.L.; McAninch, E.A.; Wilson, R.S.; Evans, D.A. Population estimate of people with clinical Alzheimer’s disease and mild cognitive impairment in the United States (2020–2060). Alzheimer’s Dement. 2021, 17, 1966–1975. [Google Scholar] [CrossRef]
  19. Yogesh; Kumar, K.; Kaur, K.; Thakur, A.; Nazir, N.; Dahiya, O.; Agarwal, S. A Comparative Study of Machine Learning Techniques for the Detection of Alzheimer’s Disease. In Proceedings of the 2023 10th IEEE Uttar Pradesh Section International Conference on Electrical, Electronics and Computer Engineering (UPCON), Gautam Buddha Nagar, India, 1–3 December 2023; Volume 10, pp. 52–56. [Google Scholar] [CrossRef]
  20. Panda, S.; Akbar, A.; Li, C.; Khan, O.T.; Patel, R.; Canevari, J.; Lambropoulos, D.; Haghani, S. A Digital Platform for Early Detection and Monitoring the Progression of Alzheimer’s Disease. In Proceedings of the 2024 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 6–8 January 2024; pp. 1–3. [Google Scholar] [CrossRef]
  21. Qu, Z.; Yao, T.; Liu, X.; Wang, G. A Graph Convolutional Network Based on Univariate Neurodegeneration Biomarker for Alzheimer’s Disease Diagnosis. IEEE J. Transl. Eng. Health Med. 2023, 11, 405–416. [Google Scholar] [CrossRef]
  22. Tushar; Patel, R.K.; Aggarwal, E.; Solanki, K.; Dahiya, O.; Yadav, S.A. A Logistic Regression and Decision Tree Based Hybrid Approach to Predict Alzheimer’s Disease. In Proceedings of the 2023 International Conference on Computational Intelligence and Sustainable Engineering Solutions (CISES), Greater Noida, India, 28–30 April 2023; pp. 722–726. [Google Scholar] [CrossRef]
  23. Alshammari, M.; Mezher, M. A Modified Convolutional Neural Networks for MRI-based Images for Detection and Stage Classification of Alzheimer Disease. In Proceedings of the 2021 National Computing Colleges Conference (NCCC), Taif, Saudi Arabia, 27–28 March 2021. [Google Scholar] [CrossRef]
  24. Thatere, A.; Verma, P.; Reddy, K.T.V.; Umate, L. A Short Survey on Alzheimer’s Disease: Recent Diagnosis and Obstacles. In Proceedings of the 2023 1st DMIHER International Conference on Artificial Intelligence in Education and Industry 4.0 (IDICAIEI), Wardha, India, 27–28 November 2023; pp. 1–6. [Google Scholar] [CrossRef]
  25. Sevcik, A.; Rusko, M. A Systematic Review of Alzheimer’s disease detection based on speech and natural language processing. In Proceedings of the 2022 32nd International Conference Radioelektronika (RADIOELEKTRONIKA), Kosice, Slovakia, 21–22 April 2022; pp. 1–5. [Google Scholar] [CrossRef]
  26. Gurrala, V.K.; Talasila, S.; Medikonda, N.R.; Challa, S.; Sohail, S.; Siddiq, M.A.B. A Web-Based Interface for Comprehensive Staging Classification of Alzheimer’s Disease Diagnosis through Convolutional Neural Networks. In Proceedings of the 2024 5th International Conference for Emerging Technology (INCET), Belgaum, India, 24–26 May 2024; pp. 1–5. [Google Scholar] [CrossRef]
  27. Al-Shoukry, S.; Rassem, T.H.; Makbol, N.M. Alzheimer’s diseases detection by using deep learning algorithms: A mini-review. IEEE Access 2020, 8, 77131–77141. [Google Scholar] [CrossRef]
  28. Grueso, S.; Viejo-Sobera, R. Machine learning methods for predicting progression from mild cognitive impairment to Alzheimer’s disease dementia: A systematic review. Alzheimer’s Res. Ther. 2021, 13, 162. [Google Scholar] [CrossRef] [PubMed]
  29. Arulprakash, A.; Ponmaniraj, S.; Gokul Rajan, V.; Divya, A. Application of Artificial Intelligence Technique to Detect Alzheimer’s Disease. In Proceedings of the 2023 Intelligent Computing and Control for Engineering and Business Systems (ICCEBS), Chennai, India, 14–15 December 2023; pp. 1–5. [Google Scholar] [CrossRef]
  30. Bhattarai, P.; Thakuri, D.S.; Nie, Y.; Chand, G.B. Explainable AI-based Deep-SHAP for mapping the multivariate relationships between regional neuroimaging biomarkers and cognition. Eur. J. Radiol. 2024, 174, 111403. [Google Scholar] [CrossRef] [PubMed]
  31. Alatrany, A.S.; Khan, W.; Hussain, A.; Kolivand, H.; Al-Jumeily, D. An explainable machine learning approach for Alzheimer’s disease classification. Sci. Rep. 2024, 14, 2637. [Google Scholar] [CrossRef] [PubMed]
  32. Bhargavi, M.S.; Prabhakar, B. Deep Learning Approaches for Early Detection of Alzheimer’s Disease using MRI Neuroimaging. In Proceedings of the 2022 International Conference on Connected Systems & Intelligence (CSI), Trivandrum, India, 31 August–2 September 2022; pp. 1–6. [Google Scholar] [CrossRef]
  33. Tripathy, S.K.; Singh, D.; Jaiswal, A. Multi-Layer Feature Fusion-based Deep Multi-layer Depth Separable Convolution Neural Network for Alzheimer’s Disease Detection. In Proceedings of the 2023 International Conference on IoT, Communication and Automation Technology (ICICAT), Gorakhpur, India, 23–24 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  34. Yin, Y.; Jin, W.; Bai, J.; Liu, R.; Zhen, H. SMIL-DeiT:Multiple Instance Learning and Self-supervised Vision Transformer network for Early Alzheimer’s disease classification. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; pp. 1–6. [Google Scholar] [CrossRef]
  35. Kumar, S.; Singh, N.P.; Brahma, B. AI-Based Model for Detection and Classification of Alzheimer Disease. In Proceedings of the 2023 IEEE International Conference on Computer Vision and Machine Intelligence (CVMI), Gwalior, India, 10–11 December 2023; pp. 1–6. [Google Scholar] [CrossRef]
  36. Zubair, L.; Irtaza, S.A.; Nida, N.; Haq, N.U. Alzheimer and Mild Cognitive disease Recognition Using Automated Deep Learning Techniques. In Proceedings of the 2021 International Bhurban Conference on Applied Sciences and Technologies (IBCAST), Islamabad, Pakistan, 12–16 January 2021; pp. 310–315. [Google Scholar] [CrossRef]
  37. Saxena, A.; Kaur, H. Alzheimer’s Disease (AD) Detection Using Various Machine Learning Techniques: A Systematic Review. In Proceedings of the 2023 6th International Conference on Contemporary Computing and Informatics (IC3I), Gautam Buddha Nagar, India, 14–16 September 2023; Volume 6, pp. 77–81. [Google Scholar] [CrossRef]
  38. Archana, B.; Kalirajan, K. Alzheimer’s Disease Classification using Convolutional Neural Networks. In Proceedings of the 2023 International Conference on Innovative Data Communication Technologies and Application (ICIDCA), Uttarakhand, India, 14–16 March 2023; pp. 1044–1048. [Google Scholar] [CrossRef]
  39. Liu, B. Alzheimer’s disease classification using hippocampus and improved DenseNet. In Proceedings of the 2023 International Conference on Image Processing, Computer Vision and Machine Learning (ICICML), Chengdu, China, 3–5 November 2023; pp. 451–454. [Google Scholar] [CrossRef]
  40. Han, L.; Yang, T.; Pu, X.; Sun, L.; Yu, B.; Xi, J. Alzheimer’s Disease Classification using LightGBM and Euclidean Distance Map. IEEE Adv. Inf. Technol. Electron. Autom. Control Conf. 2021, 2021, 1540–1544. [Google Scholar] [CrossRef]
  41. Alarjani, M. Alzheimer’s Disease Detection based on Brain Signals using Computational Modeling. In Proceedings of the 2024 Seventh International Women in Data Science Conference at Prince Sultan University (WiDS PSU), Riyadh, Saudi Arabia, 3–4 March 2024; pp. 77–83. [Google Scholar] [CrossRef]
  42. Sanjeev Kumar, K.; Reddy, B.S.; Ravichandran, M. Alzheimer’s Disease Detection Using Transfer Learning: Performance Analysis of InceptionResNetV2 and ResNet50 Models. In Proceedings of the 2023 Seventh International Conference on Image Information Processing (ICIIP), Solan, India, 22–24 November 2023; pp. 832–837. [Google Scholar] [CrossRef]
  43. Mishra, S.; Sharma, V.; Ramya, G. Alzheimer’s Disease Prediction Using Machine Learning. In Proceedings of the 2024 7th International Conference on Circuit Power and Computing Technologies (ICCPCT), Kollam, India, 8–9 August 2024. [Google Scholar] [CrossRef]
  44. Lu, P.; Tan, Y.; Xing, Y.; Liang, Q.; Yan, X.; Zhang, G. An Alzheimer’s disease classification method based on ConvNeXt. In Proceedings of the 2023 3rd International Symposium on Computer Technology and Information Science (ISCTIS), Chengdu, China, 7–9 July 2023; pp. 884–888. [Google Scholar] [CrossRef]
  45. Prajapati, R.; Khatri, U.; Kwon, G.R. An Efficient Deep Neural Network Binary Classifier for Alzheimer’s Disease Classification. In Proceedings of the 2021 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Jeju Island, Republic of Korea, 13–16 April 2021; pp. 231–234. [Google Scholar] [CrossRef]
  46. Shi, R.; Wang, L.; Jiang, J. An unsupervised region of interest extraction model for tau PET images and its application in the diagnosis of Alzheimer’s disease. In Proceedings of the 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Glasgow, UK, 11–15 July 2022; pp. 2157–2160. [Google Scholar] [CrossRef]
  47. Neetha, P.U.; Simran, S.; Rama Kainthaje, S.; Ujwal; Sunilkumar, G.; Pushpa, C.N.; Thriveni, J.; Venugopal, K.R. Borderline-DEMNET for Multi-Class Alzheimer’s Disease Classification. In Proceedings of the 2023 IEEE 5th International Conference on Cybernetics, Cognition and Machine Learning Applications (ICCCMLA), Hamburg, Germany, 7–8 October 2023; pp. 192–197. [Google Scholar] [CrossRef]
  48. Prabha, C. Classification and Detection of Alzheimer’s Disease: A Brief Analysis. In Proceedings of the 2023 International Conference on Self Sustainable Artificial Intelligence Systems (ICSSAS), Erode, India, 18–20 October 2023; pp. 777–782. [Google Scholar] [CrossRef]
  49. Das, R.; Kalita, S. Classification of Alzheimer’s Disease Stages Through Volumetric Analysis of MRI Data. In Proceedings of the 2022 IEEE Calcutta Conference (CALCON), Kolkata, India, 10–11 December 2022; pp. 165–169. [Google Scholar] [CrossRef]
  50. Yuan, Z.; Yao, X.; Bu, X. Classification of Alzheimer’s Disease Using Conventional Machine Learning Methods with Cortical and Genetic Characteristics. In Proceedings of the 2022 IEEE 2nd International Conference on Power, Electronics and Computer Applications (ICPECA), Shenyang, China, 21–23 January 2022; pp. 303–306. [Google Scholar] [CrossRef]
  51. Palacios-Navarro, G.; Buele, J.; Gimeno Jarque, S.; Bronchal Garcia, A. Cognitive Decline Detection for Alzheimer’s Disease Patients Through an Activity of Daily Living (ADL). IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 30, 2225–2232. [Google Scholar] [CrossRef]
  52. Alatrany, A.S.; Hussain, A.; Alatrany, S.S.J.; Mustafina, J.; Al-Jumeily, D. Comparison of Machine Learning Algorithms for classification of Late Onset Alzheimer’s disease. In Proceedings of the 2023 15th International Conference on Developments in eSystems Engineering (DeSE), Baghdad & Anbar, Iraq, 9–12 January 2023; pp. 60–64. [Google Scholar] [CrossRef]
  53. Wang, T.; Cao, L. Deep Learning Based Diagnosis of Alzheimer’s Disease Using Structural Magnetic Resonance Imaging: A Survey. In Proceedings of the 2021 3rd International Conference on Applied Machine Learning (ICAML), Changsha, China, 23–25 July 2021; pp. 408–412. [Google Scholar] [CrossRef]
  54. Kayalvizhi, M.; Senthil Kumar, G.; Tushal, V.; Yashvanth, M.; Santhosh, G. Deep Learning-Based Severity Detection in Alzheimer’s Disease: A Comprehensive Study on Cognitive Impairment. In Proceedings of the 2023 International Conference on Data Science, Agents & Artificial Intelligence (ICDSAAI), Chennai, India, 21–23 December 2023; pp. 1–6. [Google Scholar] [CrossRef]
  55. Akter, L.; Ferdib-Al-Islam. Dementia Identification for Diagnosing Alzheimer’s Disease using XGBoost Algorithm. In Proceedings of the 2021 International Conference on Information and Communication Technology for Sustainable Development (ICICT4SD), Dhaka, Bangladesh, 27–28 February 2021; pp. 205–209. [Google Scholar] [CrossRef]
  56. Pallawi, S.; Singh, D.K. Detection of Alzheimer’s Disease Stages Using Pre-Trained Deep Learning Approaches. In Proceedings of the 2023 IEEE 5th International Conference on Cybernetics, Cognition and Machine Learning Applications (ICCCMLA), Hamburg, Germany, 7–8 October 2023; pp. 252–256. [Google Scholar] [CrossRef]
  57. Islam, J.; Furqon, E.N.; Farady, I.; Lung, C.W.; Lin, C.Y. Early Alzheimer’s Disease Detection Through YOLO-Based Detection of Hippocampus Region in MRI Images. In Proceedings of the 2023 Sixth International Symposium on Computer, Consumer and Control (IS3C), Taichung, Taiwan, 30 June–3 July 2023; pp. 32–35. [Google Scholar] [CrossRef]
  58. Zhou, Y.; Gao, C.; Zhang, X.; Zhang, W.; Wan, S.; Liu, Y. Early Detection and Intervention of Alzheimer’s disease Based on Game APP. In Proceedings of the 2024 5th International Conference on Information Science, Parallel and Distributed Systems (ISPDS), Guangzhou, China, 31 May–2 June 2024; Volume 2017, pp. 182–188. [Google Scholar] [CrossRef]
  59. Rauniyar, K.; Thakur, S.; Nevatia, A.; Shambharkar, P.G. Early Detection of Alzheimer’s Disease: The Importance of Speech Analysis. In Proceedings of the 2023 2nd International Conference on Applied Artificial Intelligence and Computing (ICAAIC), Salem, India, 4–6 May 2023; pp. 1069–1073. [Google Scholar] [CrossRef]
  60. Patra, B.; Maity, N.P.; Charan Sutar, B. Early Detection of Alzheimer’s Disease using Feed Forward Neural Network. In Proceedings of the 2023 14th International Conference on Computing Communication and Networking Technologies (ICCCNT), Delhi, India, 6–8 July 2023; pp. 1–4. [Google Scholar] [CrossRef]
  61. Varma, B.S.S.; Kalyani, G.; Asish, K.; Bai, M.I. Early Detection of Alzheimer’s Disease Using SVM, Random Forest & FNN Algorithms. In Proceedings of the 2023 2nd International Conference for Innovation in Technology (INOCON), Bangalore, India, 3–5 March 2023; pp. 1–6. [Google Scholar] [CrossRef]
  62. Jiang, Y.; Yu, Z.; Yin, X.; Guo, H. Early Diagnosis and Progression of Alzheimer’s Disease Based on Long Short-Term Memory Model. In Proceedings of the 2023 5th International Conference on Robotics, Intelligent Control and Artificial Intelligence (RICAI), Hangzhou, China, 1–3 December 2023; pp. 620–624. [Google Scholar] [CrossRef]
  63. Bushra, U.H.; Priya, F.C.; Patwary, M.J.A. Fuzziness-Based Semi-Supervised Learning for Early Detection of Alzheimer’s Disease using MRI data. In Proceedings of the 2023 26th International Conference on Computer and Information Technology (ICCIT), Cox’s Bazar, Bangladesh, 13–15 December 2023; pp. 1–6. [Google Scholar] [CrossRef]
  64. Subha, R.; Nayana, B.R.; Selvadass, M. Hybrid Machine Learning Model Using Particle Swarm Optimization for Effectual Diagnosis of Alzheimer’s Disease from Handwriting. In Proceedings of the 2022 4th International Conference on Circuits, Control, Communication and Computing (I4C), Bangalore, India, 21–23 December 2022; pp. 491–495. [Google Scholar] [CrossRef]
  65. Buyrukoglu, S. Improvement of Machine Learning Models Performances based on Ensemble Learning for the detection of Alzheimer Disease. In Proceedings of the 2021 6th International Conference on Computer Science and Engineering (UBMK), Ankara, Turkey, 15–17 September 2021; pp. 102–106. [Google Scholar] [CrossRef]
  66. Chandra, A.; Roy, S. On the Detection of Alzheimer’s Disease using Naïve Bayes Classifier. In Proceedings of the 2023 International Conference on Microwave, Optical, and Communication Engineering (ICMOCE), Optical, Bhubaneswar, India, 26–28 May 2023; pp. 1–4. [Google Scholar] [CrossRef]
  67. Talha, A.; Dhanasree, C.; Divya, E.; Prabhas, K.S.; Syed Abudhagir, U. Performance Evaluation of Deep Learning Models for Alzheimer’s Disease Detection. In Proceedings of the 2024 10th International Conference on Communication and Signal Processing (ICCSP), Melmaruvathur, India, 12–14 April 2024; pp. 317–322. [Google Scholar] [CrossRef]
  68. Bharath, M.; Gowtham, S.; Vedanth, S.; Kodipalli, A.; Rao, T.; Rohini, B.R. Predicting Alzheimer’s Disease Progression through Machine Learning Algorithms. In Proceedings of the 2023 International Conference on Recent Advances in Science and Engineering Technology (ICRASET), B G Nagara, India, 23–24 November 2023; pp. 1–5. [Google Scholar] [CrossRef]
  69. Sindhu, T.S.; Kumaratharan, N.; Anandan, P. Review on Automatic Detection of Alzheimer Disease. In Proceedings of the 2022 International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 25–27 January 2022; pp. 1–6. [Google Scholar] [CrossRef]
  70. Jansi, R.; Gowtham, N.; Ramachandran, S.; Sai Praneeth, V. Revolutionizing Alzheimer’s Disease Prediction using InceptionV3 in Deep Learning. In Proceedings of the 2023 7th International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 22–24 November 2023; pp. 1155–1160. [Google Scholar] [CrossRef]
  71. Sushmitha, S.; Chitrakala, S.; Bharathi, U. sMRI Classification of Alzheimer’s Disease Using Genetic Algorithm and Multi-Instance Learning (GA+MIL). In Proceedings of the 2023 International Conference on Electrical, Electronics, Communication and Computers (ELEXCOM), Roorkee, India, 26–27 August 2023; pp. 1–4. [Google Scholar] [CrossRef]
  72. Kadyan, V.; Bawa, P.; Akhtar, M.M.; Singh, M. Speech-Based Alzheimer’s Disease Classification System with Noise-Resilient Features Optimization. In Proceedings of the 2023 31st Irish Conference on Artificial Intelligence and Cognitive Science (AICS), Letterkenny, Ireland, 7–8 December 2023; pp. 1–4. [Google Scholar] [CrossRef]
  73. Peng, C.; Liu, M.; Meng, C.; Xue, S.; Keogh, K.; Xia, F. Stage-aware Brain Graph Learning for Alzheimer’s Disease. In Proceedings of the 2024 IEEE Conference on Artificial Intelligence (CAI), Singapore, 25–27 June 2024; pp. 1346–1349. [Google Scholar] [CrossRef]
  74. Anjali; Singh, D.; Pandey, O.J.; Dai, H.N. STCNN: Combining SMOTE-TOMEK with CNN for Imbalanced Classification of Alzheimer’s Disease. IEEE Sens. Lett. 2024, 8, 1–4. [Google Scholar] [CrossRef]
  75. Yu, X.; Zhang, L.; Lyu, Y.; Liu, T.; Zhu, D. Supervised Deep Tree in Alzheimer’s Disease. In Proceedings of the 2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI), Cartagena, Colombia, 18–21 April 2023; pp. 1–5. [Google Scholar] [CrossRef]
  76. Jin, S.; Zou, P.; Han, Y.; Jiang, J. Unsupervised detection of individual atrophy in Alzheimer’s disease. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Virtual, 1–5 November 2021; pp. 2647–2650. [Google Scholar] [CrossRef]
  77. Soni, A.; Amrhein, B.; Baucum, M.; Paek, E.J.; Khojandi, A. Using Verb Fluency, Natural Language Processing, and Machine Learning to Detect Alzheimer’s Disease. In Proceedings of the 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Virtual, 1–5 November 2021; pp. 2282–2285. [Google Scholar] [CrossRef]
  78. Suttapakti, U.; Pianprasit, P. Variational Feature Extraction of Two-Dimensional Variational Mode Decomposition for Alzheimer’s Disease Classification. In Proceedings of the 2020 12th International Conference on Information Technology and Electrical Engineering (ICITEE), Yogyakarta, Indonesia, 6–8 October 2020; pp. 344–349. [Google Scholar] [CrossRef]
  79. Basher, A.; Kim, B.C.; Lee, K.H.; Jung, H.Y. Volumetric Feature-Based Alzheimer’s Disease Diagnosis from sMRI Data Using a Convolutional Neural Network and a Deep Neural Network. IEEE Access 2021, 9, 29870–29882. [Google Scholar] [CrossRef]
  80. Butta, R.; Shaik, M.S.; Murthy, G.L.N. Ensemble deep learning approach for early diagnosis of Alzheimer’s disease. Multimed. Tools Appl. 2024, 84, 1403–1428. [Google Scholar] [CrossRef]
  81. Agarwal, D.; Berbís, M.Á.; Luna, A.; Lipari, V.; Ballester, J.B.; de la Torre-Díez, I. Automated Medical Diagnosis of Alzheimer’s Disease Using an Efficient Net Convolutional Neural Network. J. Med. Syst. 2023, 47, 57. [Google Scholar] [CrossRef]
  82. Saim, M.; Feroui, A. Automated Alzheimer’s Disease Detection and Diagnosis Method Based on Bayesian Optimization and CNN—Based Pre-Trained Features; Springer: New York, NY, USA, 2024; ISBN 0123456789. [Google Scholar]
  83. Arafa, D.A.; Moustafa, H.E.D.; Ali, H.A.; Ali-Eldin, A.M.T.; Saraya, S.F. A Deep Learning Framework for Early Diagnosis of Alzheimer’s Disease on MRI Images; Springer: New York, NY, USA, 2024; Volume 83, ISBN 0123456789. [Google Scholar]
Figure 1. An overview of symptoms and risk factors in AD.
Figure 1. An overview of symptoms and risk factors in AD.
Diagnostics 15 00789 g001
Figure 2. Pathophysiology and detection strategies in AD.
Figure 2. Pathophysiology and detection strategies in AD.
Diagnostics 15 00789 g002
Figure 3. Sample instances from all available classes of OASIS Alzheimer’s Detection dataset.
Figure 3. Sample instances from all available classes of OASIS Alzheimer’s Detection dataset.
Diagnostics 15 00789 g003
Figure 4. Preprocessed image samples indicating respective class name.
Figure 4. Preprocessed image samples indicating respective class name.
Diagnostics 15 00789 g004
Figure 5. Dataset description, before and after data augmentation.
Figure 5. Dataset description, before and after data augmentation.
Diagnostics 15 00789 g005
Figure 6. Basic architecture of convolutional neural network.
Figure 6. Basic architecture of convolutional neural network.
Diagnostics 15 00789 g006
Figure 7. Layer-based architecture of the VGG19 model used in the proposed framework.
Figure 7. Layer-based architecture of the VGG19 model used in the proposed framework.
Diagnostics 15 00789 g007
Figure 8. Architecture of a basic transformer.
Figure 8. Architecture of a basic transformer.
Diagnostics 15 00789 g008
Figure 9. Functional architecture of vision transformer (ViT).
Figure 9. Functional architecture of vision transformer (ViT).
Diagnostics 15 00789 g009
Figure 10. Layer-wise architecture of ViT-B16 model.
Figure 10. Layer-wise architecture of ViT-B16 model.
Diagnostics 15 00789 g010
Figure 11. Brain MRI image with respective 64 patches.
Figure 11. Brain MRI image with respective 64 patches.
Diagnostics 15 00789 g011
Figure 12. Workflow diagram of the proposed framework.
Figure 12. Workflow diagram of the proposed framework.
Diagnostics 15 00789 g012
Figure 13. Cloud-AI-based IoMT infrastructure for remote diagnosis of Alzheimer disease.
Figure 13. Cloud-AI-based IoMT infrastructure for remote diagnosis of Alzheimer disease.
Diagnostics 15 00789 g013
Figure 14. Confusion matrix of (a) ViT-B16, (b) VGG19, (c) ResNet152V2, and (d) EfficientNetV2B3.
Figure 14. Confusion matrix of (a) ViT-B16, (b) VGG19, (c) ResNet152V2, and (d) EfficientNetV2B3.
Diagnostics 15 00789 g014
Figure 15. Confusion matrix obtained by the weighted average ensemble technique.
Figure 15. Confusion matrix obtained by the weighted average ensemble technique.
Diagnostics 15 00789 g015
Figure 16. Performance metrics with a weighted average ensemble approach.
Figure 16. Performance metrics with a weighted average ensemble approach.
Diagnostics 15 00789 g016
Figure 17. Accuracy and loss graph for ViT-B16.
Figure 17. Accuracy and loss graph for ViT-B16.
Diagnostics 15 00789 g017
Figure 18. Accuracy and loss graph of VGG19.
Figure 18. Accuracy and loss graph of VGG19.
Diagnostics 15 00789 g018
Figure 19. Accuracy and loss graph for ResNet152V2.
Figure 19. Accuracy and loss graph for ResNet152V2.
Diagnostics 15 00789 g019
Figure 20. Accuracy and loss graph of EfficientNetV2B3.
Figure 20. Accuracy and loss graph of EfficientNetV2B3.
Diagnostics 15 00789 g020
Figure 21. Accuracy and loss status of the proposed model during training and validation.
Figure 21. Accuracy and loss status of the proposed model during training and validation.
Diagnostics 15 00789 g021
Table 1. Summary of Related Works.
Table 1. Summary of Related Works.
Author (s)Model UsedDatasetTotal ImagesMethodologyAccuracyFocus AreaLimitations
Qu et al. (2023) [21]UNB-GCNADNI Dataset928Used univariate neurodegeneration biomarkers with GCN.93.90% (CU vs. AD)
82.05% (MCI vs. AD)
Early AD detectionLacks multimodal validation.
Liu (2023) [39]DenseNet -AMADNI Dataset413Integration of hippocampal and whole-brain MRI92.8%Structural abnormality analysisLimited feature diversity
Han et al. (2021) [40]LightGBMReal Datasets (EADC and ADNI Dataset)78 (41 NC and 37 AD) * 19 Features = 1482 ImagesFeature optimization for improved classification89.7% (AD Classification Feature)
82.1% (Hippocampal Volume)
AD diagnosisInconsistent performance on small datasets
Sanjeev Kumar K et al. (2023) [42]InceptionResNetV2 and ResNet50ADNI dataset6400 MRI ImagesDeep learning architecture for AD stage classification86.84%
90.27%
Staged AD classificationRequires high-quality annotated data
Mishra et al. (2024) [43]ML models (LR, RF, SVM, DT, AdaBoost)OASIS Project Dataset150 SubjectsAnalysis of early prediction problems in AD86.84% (RF) (Highest)Early predictionData inconsistency across datasets
Lu et al. (2023) [44]ConvNeXt-DMLPADNI Dataset33,984Multiclassification framework to reduce imaging overlap of AD and MCI78.95%AD and MCI classificationLimited generalization to external datasets
Prajapati et al. (2021) [45]Deep Neural Networks (DNN)ADNI Dataset58 AD, 48 MCI and 73 CN SubjectsBinary classification using deep neural networks85.19% (AD vs CN)
76.93% (MCI vs. CN)
72.73% (AD vs MCI)
AD stage classificationLimited scope for multiclass classification
Das et al. (2022) [49]Hippocampal Segmentation with CNNADNI and OASIS Dataset (MRI)100 people (Train) and 35 people (Test)Focused on hippocampal segmentation for structural MRI91.43%Structural abnormality detectionLimited generalizability to other brain regions
Pallawi et al. (2023) [56]EfficientNetB0Dataset from Kaggle6400 images (MD-896, MOD-64, ND-3200, and VMD-2240)Transfer learning for early-stage AD classification95.78%Early-stage classificationDataset imbalance
Islam et al. (2023) [57]YOLO (YOLOv3, YOLOv4, YOLOv5, YOLOv6 and YOLOv7ADNI Dataset (MRI data)300 Subjects (2000+ Images)Automated hippocampal detection for early recognitionYOLOv7 (81%-AD vs. MCI, 95%-AD vs. CN, 83%-CN vs. MCI)Early-stage identificationLimited automation coverage
Yin et al. (2022) [34]SMIL-DeiTADNI dataset2032 SubjectsStaged AD classification with enhanced techniques93.2%Staged classificationModel interpretability
Jin et al. (2021) [76]VAE+GANADNI Dataset (sMRI Images)1042 SubjectsIdentification of brain atrophy94%Structural analysisOverfitting potential
Table 2. Comparison of VGG19, ResNet152V2, and EfficientNetV2B3 Architectures for Image Classification.
Table 2. Comparison of VGG19, ResNet152V2, and EfficientNetV2B3 Architectures for Image Classification.
FeatureVGG19ResNet152V2EfficientNetV2B3
Architecture TypeA deep CNN with a straightforward sequential designA deep residual network incorporating skip connectionsAn optimized convolutional network with depth-wise convolutions
Depth (Layers)19 layers152 layersAdaptive depth based scaling
Parameter CountApproximately 143 millionApproximately 60 millionNearly 24 million
AdvantagesSimple yet effective for smaller datasetsOvercomes vanishing gradient issues and excels in deep feature extractionEfficiency in computation while maintaining high accuracy
LimitationsHigh parameter counts leading to increased computational costMore complex training process with longer inference timeRequires sophisticated training techniques for optimal results
Feature Extraction CapabilityCaptures fine-grained spatial details effectivelyExcels at extracting deep semantic representationsStrikes a balance between feature learning and efficiency
Training TimeRelatively longModerateFaster compared to the others
Accuracy PerformanceModerate to highHighVery high
Reason for SelectionIncluded due to its ability to retain detailed spatial informationChosen for its deep feature extraction and improved gradient handlingSelected for its efficiency, optimizing both accuracy and computational demands
Table 3. Shape and Parameter Size of “VGG19” Model.
Table 3. Shape and Parameter Size of “VGG19” Model.
Layer (Type)Output ShapeParam #
vgg19 (Functional)(None, 2, 2, 512)20,024,384
global_average_pooling2d (GlobalAveragePooling2D)(None, 512)0
dense_4 (Dense)(None, 128)65,664
batch_normalization_2 (BatchNormalization)(None, 128)512
dropout (Dropout)(None, 128)0
dense_5 (Dense)(None, 4)516
Total params: 20,091,076
Trainable params: 66,436
Nontrainable params: 20,024,640
Table 4. Shape and Parameter Size of “Vision Transformer (ViT-B16)” Model.
Table 4. Shape and Parameter Size of “Vision Transformer (ViT-B16)” Model.
Layer (Type)Output ShapeParam #
ViT-B16 (Functional)(None, 768)85,660,416
flatten_6 (Flatten)(None, 768)0
batch_normalization_12 (Batch Normalization)(None, 768)3072
dense_24 (Dense)(None, 128)98,432
batch_normalization_13 (Batch Normalization)(None, 128)512
dense_25 (Dense)(None, 64)8256
dense_26 (Dense)(None, 32)2080
dense_27 (Dense)(None, 4)132
Total params: 85,772,900
Trainable params: 85,771,108
Nontrainable params: 1792
Table 5. Parameter details of developed models utilized in the proposed classification frameworks.
Table 5. Parameter details of developed models utilized in the proposed classification frameworks.
ParametersClassification with ViT-B16Classification with VGG19Classification with ResNet152V2Classification with EfficientNetV2B3
Train_loss0.050.080.070.08
Train_accuracy95.2192.1489.1384.15
Training time4 h 17 min2 h 13 min2 h 39 min1 h 46 min
Epoch10162025
OptimizerAdamAdamAdamAdam
Initial learning rate0.0010.00010.0010.00001
Loss functioncategorical cross entropycategorical cross entropycategorical cross entropycategorical cross entropy
Output activation functionSoftmaxSoftmaxSoftmaxSoftmax
Table 6. Accuracy and Loss Values Obtained by the ViT-B16 Model.
Table 6. Accuracy and Loss Values Obtained by the ViT-B16 Model.
Number of EpochTrain AccuracyTrain LossValidation AccuracyValidation Loss
1810.19830.16
2830.18860.15
3840.15850.16
4830.13860.14
5890.12900.12
6910.12930.13
7930.0991 0.07
8920.06940.05
9940.07950.06
10950.05960.04
Table 7. Tuned Parameters with GOA.
Table 7. Tuned Parameters with GOA.
ModelsBest Learning RateBest Dropout RateBest Batch SizeBest Number of EpochsHighest AccuracyMinimum LossBest Fitness Score
ViT-B160.001-81095.210.05−0.03
VGG190.00010.2321692.140.08−0.01
ResNet152V20.0010.2162089.130.07−0.01
EfficientNetV2B30.000010.1322584.150.08−0.004
Table 8. Classification Report of Ensemble Model.
Table 8. Classification Report of Ensemble Model.
PrecisionRecallf1-ScoreSupport
Mild Demented0.95951.00000.9793142
Moderate Demented0.97440.96200.9682158
Nondemented0.98580.95860.9720145
Very Mild Demented0.97320.97320.9732149
accuracy 0.9731594
macro avg0.97320.97350.9732594
weighted avg0.97330.97310.9730594
Table 9. Training Testing Accuracies and Processing Time Obtained by DL models.
Table 9. Training Testing Accuracies and Processing Time Obtained by DL models.
Models Training Accuracy (%) Testing Accuracy (%) Processing Time During Training
ViT-B1695.2196.124 h 17 m
VGG1992.1494.272 h 13 m
ResNet152V289.1389.562 h 39 m
EfficientNetV2B384.1584.171 h 46 m
Ensemble model (Not optimized) 96.3996.431 h 55 m
Ensemble model (Optimized with GOA) 97.2597.311 h 41 m
Table 10. Weight Value Distribution with Respective Obtained Accuracy.
Table 10. Weight Value Distribution with Respective Obtained Accuracy.
WV1WV2WV3WV4Ensemble Accuracy
0.10.20.30.496.42
0.20.10.40.396.68
0.30.40.10.296.83
0.40.30.20.197.19
0.40.30.30.497.04
0.40.30.40.397.25
0.40.30.30.197.31
Table 11. Five-fold cross-validation method for performance analysis of different classification frameworks.
Table 11. Five-fold cross-validation method for performance analysis of different classification frameworks.
Split CountViT-B16VGG19ResNet152V2EfficientNetV2B3Ensemble Model
196.0494.2289.5184.1197.26
295.8694.1489.3684.0497.14
395.9694.0889.2283.9996.99
496.0894.1889.4183.9497.12
595.8894.0589.3184.0997.06
Arithmetic Mean95.9694.1389.3684.0397.14
Variance0.0070.0030.0090.0030.008
Standard deviation (std.)0.080.060.090.060.08
Table 12. Performance Metrics Obtained from the MRI Image-Based Classification Framework.
Table 12. Performance Metrics Obtained from the MRI Image-Based Classification Framework.
Performance metricsViT-B16VGG19ResNet152V2EfficientNetV2B33
Accuracy 96.1294.2789.5684.17
Precision 96.1594.2789.5584.18
Recall 96.1494.3089.5784.17
F1 Score 0.96130.94270.89550.8414
Table 13. Performance Status of Proposed Ensemble Model During Training and Validation.
Table 13. Performance Status of Proposed Ensemble Model During Training and Validation.
EpochTrain Accuracy with GOAValidation Accuracy with GOATrain Accuracy Without GOAValidation Accuracy Without GOATrain Loss with GOAValidation Loss with GOATrain Loss Without GOAValidation Loss Without GOA
183.5184.3781.5482.730.180.170.210.20
286.3488.5482.3887.580.170.140.190.18
385.5489.8682.2887.520.150.160.180.18
486.6888.3684.5684.870.140.130.190.16
590.6789.3489.2587.230.140.110.140.15
694.6892.7593.8691.780.120.080.130.11
792.3694.9792.6492.640.100.070.130.10
895.4393.6594.1694.340.070.050.110.10
996.2896.5295.3195.780.070.040.080.09
1097.2597.3196.3996.430.050.030.080.08
Table 14. Comprehensive Comparison with Other Existing Studies.
Table 14. Comprehensive Comparison with Other Existing Studies.
AuthorsModel UsedDatasetTotal ImagesMethodologyAccuracyFocus AreaLimitations
Butta et al. (2023) [80]Ensemble DL Model (O-ANN, Capsule CNN, Autoencoder)Kaggle Dataset6400 MRI ImagesPreprocessing (histogram equalization, Gaussian filtering, skull stripping), ROI extraction (attention-based U-NET with EEO optimization)94.52%Early and precise AD diagnosis using MRIComputational complexity due to hybrid optimization and ensemble techniques; limited testing on diverse datasets
Agarwal et al. (2023) [81]EfficientNet-b0 CNN with fusion of end-to-end and transfer learning (TL)IXI and ADNI Dataset600 MRI scans
916 MRI Scans (ADNI)
Preprocessing pipeline applied to T1W MRI scans; CNN model trained and evaluated for sMCI vs. AD (binary) and AD vs. CN vs. sMCI (multiclass) classificationBinary: 93.10% (testing), Multiclass: 87.38% (testing)Early detection and staging of AD using MRI biomarkersRequires further validation on larger and more diverse datasets; computational demands of CNN training
Saim et al. (2024) [82]Nine CNN architectures (AlexNet, Inception, EfficientNet-b0, etc.) Kaggle Alzheimer’s dataset ADNI dataset5091 Images (Train)
1273 Images (Test)
Feature extraction using pretrained CNNs (fully connected, inception-based, residual-based, etc.); classification using MSVM, KNN, DT; Kaggle: ADNI: 78.54% (MSVM)Early detection and staging of AD using neuroimaging biomarkersPerformance on ADNI dataset lower; needs further validation on larger and more diverse datasets.
Sanjeev Kumar et al. (2023) [42] InceptionResNetV2 and ResNet50ADNI dataset6400 MRI ImagesDeep learning architecture for AD stage classification86.84%Staged AD classificationRequires high-quality annotated data
Lu et al. (2023) [44]ConvNeXt-DMLPADNI Dataset33,984Multiclassification framework to reduce imaging overlap of AD and MCI78.95%AD and MCI classificationLimited generalization to external datasets
Proposed WorkWeighted ensemble model combined of ViT-B16, VGG19, ResNet152V2, EfficientNetV2B3OASIS Dataset, collected from kaggle234,220Multiclass classification framework, optimized with Grasshopper Optimization Algorithm97.31%Classify dementia status that can be integrated to IoMT infrastructure
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chakra Bortty, J.; Chakraborty, G.S.; Noman, I.R.; Batra, S.; Das, J.; Bishnu, K.K.; Tarafder, M.T.R.; Islam, A. A Novel Diagnostic Framework with an Optimized Ensemble of Vision Transformers and Convolutional Neural Networks for Enhanced Alzheimer’s Disease Detection in Medical Imaging. Diagnostics 2025, 15, 789. https://doi.org/10.3390/diagnostics15060789

AMA Style

Chakra Bortty J, Chakraborty GS, Noman IR, Batra S, Das J, Bishnu KK, Tarafder MTR, Islam A. A Novel Diagnostic Framework with an Optimized Ensemble of Vision Transformers and Convolutional Neural Networks for Enhanced Alzheimer’s Disease Detection in Medical Imaging. Diagnostics. 2025; 15(6):789. https://doi.org/10.3390/diagnostics15060789

Chicago/Turabian Style

Chakra Bortty, Joy, Gouri Shankar Chakraborty, Inshad Rahman Noman, Salil Batra, Joy Das, Kanchon Kumar Bishnu, Md Tanvir Rahman Tarafder, and Araf Islam. 2025. "A Novel Diagnostic Framework with an Optimized Ensemble of Vision Transformers and Convolutional Neural Networks for Enhanced Alzheimer’s Disease Detection in Medical Imaging" Diagnostics 15, no. 6: 789. https://doi.org/10.3390/diagnostics15060789

APA Style

Chakra Bortty, J., Chakraborty, G. S., Noman, I. R., Batra, S., Das, J., Bishnu, K. K., Tarafder, M. T. R., & Islam, A. (2025). A Novel Diagnostic Framework with an Optimized Ensemble of Vision Transformers and Convolutional Neural Networks for Enhanced Alzheimer’s Disease Detection in Medical Imaging. Diagnostics, 15(6), 789. https://doi.org/10.3390/diagnostics15060789

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop