Next Article in Journal
Hypertension and Diabetes as Determinants of Patient-Reported Quality of Life in Permanent Atrial Fibrillation
Previous Article in Journal
Electrophysiological Evidence of Early Auditory Dysfunction in Personal Listening Device Users: Insights from ABR with Ipsilateral Masking
Previous Article in Special Issue
Clinical Application of Machine Learning Models for Early-Stage Chronic Kidney Disease Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Convolutional Neural Network–Long Short-Term Memory (CNN–LSTM)–Attention Model Architecture for Precise Medical Image Analysis and Disease Diagnosis

1
Innovative Skills Ltd., Dhaka 1207, Bangladesh
2
Department of Electrical Engineering, College of Engineering, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
3
Department of Communications and Computer Engineering, Faculty of Engineering, Al-Ahliyya Amman University, Amman 19111, Jordan
4
CS Department, Mutah University, Karak 61710, Jordan
5
College of Engineering and Technology, American University of the Middle East, Egaila 54200, Kuwait
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(21), 2673; https://doi.org/10.3390/diagnostics15212673
Submission received: 21 September 2025 / Revised: 8 October 2025 / Accepted: 17 October 2025 / Published: 23 October 2025
(This article belongs to the Special Issue Machine-Learning-Based Disease Diagnosis and Prediction)

Abstract

Background: Deep learning (DL)-based medical image classification is becoming increasingly reliable, enabling physicians to make faster and more accurate decisions in diagnosis and treatment. A plethora of algorithms have been developed to classify and analyze various types of medical images. Among them, Convolutional Neural Networks (CNNs) have proven highly effective, particularly in medical image analysis and disease detection. Methods: To further enhance these capabilities, this research introduces MediVision, a hybrid DL-based model that integrates a vision backbone based on CNNs for feature extraction, capturing detailed patterns and structures essential for precise classification. These features are then processed through Long Short-Term Memory (LSTM), which identifies sequential dependencies to better recognize disease progression. An attention mechanism is then incorporated that selectively focuses on salient features detected by the LSTM, improving the model’s ability to highlight critical abnormalities. Additionally, MediVision utilizes a skip connection, merging attention outputs with LSTM outputs along with Grad-CAM heatmap to visualize the most important regions of the analyzed medical image and further enhance feature representation and classification accuracy. Results: Tested on ten diverse medical image datasets (including, Alzheimer’s disease, breast ultrasound, blood cell, chest X-ray, chest CT scans, diabetic retinopathy, kidney diseases, bone fracture multi-region, retinal OCT, and brain tumor), MediVision consistently achieved classification accuracies above 95%, with a peak of 98%. Conclusions: The proposed MediVision model offers a robust and effective framework for medical image classification, improving interpretability, reliability, and automated disease diagnosis. To support research reproducibility, the codes and datasets used in this study have been publicly made available through an open-access repository.

1. Introduction

1.1. Overview

The integration of Machine Learning (ML) into medical practice has revolutionized the healthcare field, enabling computers to analyze vast amounts of patients data and derive insights into solving complex clinical problems without explicit programming. As a subset of ML specialized in utilizing artificial neural networks with multiple layers to analyze data, Deep Learning (DL) is playing an increasingly important role in enhancing treatment strategies for chronic conditions, supporting precision medicine, and expediting clinical trials [1,2] and promising significant advances in areas such as diagnostic processes [3,4,5], prognosis prediction [6,7,8], and drug discovery [4,5].
Among these applications, medical image classification presents unique challenges due to various factors. The variability in medical images, due to differences in imaging techniques, patient demographics, and the presence of artifacts, complicates the development of reliable and robust models [9,10]. To elaborate, medical images often contain subtle yet critical details that are essential for accurate diagnosis, requiring highly sensitive and precise models. In parallel, tabular medical data, including patient records, lab results, and other clinical information, presents additional challenges [11,12]. Such data are typically high-dimensional, may contain missing values, and are subject to strict privacy regulations, all of which complicate the classification process [13]. Various ML, DL, and hybrid algorithms, such as Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, Generative Adversarial Networks (GANs), and Support Vector Machine (SVM) [14,15,16,17,18], have advanced automated medical image analysis by improving classification accuracy. However, existing models still face challenges in feature extraction, sequential pattern recognition, generalization across datasets, and computational efficiency [19,20,21].

1.2. Advances in CNN-Based Medical Image Classification

Medical image analysis has significantly benefited from the advancements in CNNs, as they are highly effective for feature extraction, which is essential for accurate disease detection. The work by P. M. de Sousa et al. in [22] explored Wavelet CNNs, demonstrating improved CT image integrity without requiring data augmentation, resizing, or segmentation techniques. Other CNN-based architectures like C-Net [23] and Convolutional Deep Belief Networks (CDBN) [24,25] have also shown promise in medical image classification. The C-Net model consists of a concatenation of three main components: Outer, Middle, and Inner networks. The Outer and Middle networks function as feature extractors, while the Inner network is responsible for classifying the images as malignant or benign; whereas the CDBN model utilizes SVM as a feature classifier dedicated to classifying CT images to enhance feature transfer and reuse while suppressing less useful ones.

1.3. Emerging Hybrid Architectures for Enhanced Feature Representation

Recognizing the inherent limitations of CNNs in capturing global context and sequential dependencies, several hybrid models have been explored for medical image analysis. Among these are the models that proposed combining CNNs with Transformers, such as MambaConvT [26], which utilizes Transformers for enhanced global context. Other hybrid approaches have incorporated techniques like autoencoders and Particle Swarm Optimization (PSO) [27] or ensemble learning methods (e.g., Random Forest and XGBoost) [28] to optimize feature selection and improve classification performance. Further advancements include the integration of EfficientNet with SVM [29] and the development of Hybrid Quantum CNNs (HQCNNs) with adaptive optimizers [30]. Transfer learning with pretrained CNNs has been explored in [31], enabling deep feature extraction that is more abstract and representative of underlying data patterns compared to traditional methods.
Recent advances continue to emphasize the synergy between CNNs and Vision Transformers (ViTs) for improved medical image analysis. Borji et al. [32] introduced a hybrid CNN-ViT-MLP model tailored for osteosarcoma histopathology image classification, achieving high performance on a four-class Cancer Imaging Archive (TCIA) dataset. Their model effectively captured both local and global features without relying on segmentation, highlighting the strength of combining CNNs for fine-grained pattern recognition and ViTs for long-range dependencies. Similarly, the work in [33] proposed the EFFResNet-ViT framework, which fuses EfficientNet-B0 and ResNet-50 backbones with ViT modules. The model demonstrated strong performance on brain tumor MRI and retinal image datasets, while also emphasizing model interpretability via Grad-CAM visualizations and t-SNE analysis. These works underline a growing trend in hybrid architectures that balance accuracy and explainability, which are considered vital requirements for clinical integration. Nonetheless, despite these promising results, the current models have primarily been evaluated on limited datasets and may lack the generalizability and diversity required for broader clinical applications, highlighting the need for further investigation in more varied and extensive data contexts.

1.4. Sequential and Attention-Based Models for Context-Aware Analysis

Recently, advancements in medical image classification resulted in the development of hybrid models that combine CNNs with RNNs, such as LSTM, which are capable of capturing both spatial and temporal features, such that CNNs excel at extracting local spatial features, while LSTM are designed to handle sequential data, making them suitable for learning temporal patterns. For instance, Salehin et al. [34] introduced MedvLSTM, a hybrid model that integrates CNNs and LSTM to enhance feature representation and classification accuracy in medical image analysis. Complementing these efforts, attention-based models have emerged as powerful tools to improve both accuracy and efficiency in medical imaging tasks. These models utilize attention mechanisms to prioritize relevant features within images, leading to superior classification outcomes across diverse datasets as in [35], where the integration of spatial attention in dual-input CNNs has demonstrated the potential to effectively fuse features from original and modified images. Further innovations in attention mechanisms have expanded their applicability to multi-scale and segmentation tasks. Cheng et al. [36] proposed the Attention-based Multi-scale Nested Network (AMNNet), which incorporates Convolutional Block Attention Modules (CBAM) to capture contextual information at varying scales, highlighting the versatility of attention mechanisms in both segmentation and classification. Similarly, Liu et al. [37] introduced Eff-CTNet, a model that combines CNNs and Transformer architectures with a Group Cascade Attention module, capable of reducing computational costs while maintaining high performance on small-scale datasets, and thus balancing between efficiency and accuracy in medical image analysis.

1.5. Motivation and Research Gaps

The comprehensive literature review presented above underscores the growing importance of hybrid approaches in advancing medical image classification. Techniques such as CNN-LSTM integration, attention mechanisms, and multi-scale feature fusion have demonstrated significant potential in addressing the challenges posed by complex medical datasets. However, while these methods have achieved notable progress, limitations remain, particularly in terms of interpretability, computational efficiency, and robustness across diverse datasets. Existing models often struggle to balance these factors effectively, emphasizing the need for a more unified and adaptive solution. In response to these challenges, we propose the MediVision model, a novel framework that seamlessly integrates the strengths of CNNs, LSTM, attention mechanisms, and interpretability tools like Grad-CAM heatmaps. By combining these key components, MediVision not only addresses the limitations of current models but also sets a new benchmark for performance and transparency in medical image analysis. In the following section, we detail the architecture and methodology of MediVision, illustrating how its innovative design achieves state-of-the-art results while ensuring interpretability and scalability.
In this paper, we present MediVision, a hybrid CNN-LSTM-attention model specifically designed for medical image classification. Unlike prior models that simply combine CNNs and LSTMs, MediVision integrates three key components: (1) CNNs for extracting spatial features from medical images, (2) an LSTM module that captures sequential dependencies across spatial feature maps, and (3) a tailored attention mechanism applied post-LSTM to selectively emphasize clinically relevant regions. Additionally, a skip connection preserves original CNN features alongside the LSTM–attention pathway, and Grad-CAM visualization enhances interpretability, allowing clinicians to better understand decision-relevant regions. Our proposed model was extensively evaluated over ten diverse real-world medical image datasets, as illustrated in Figure 1, with performance comparison to existing models. The results demonstrate the superior performance of MediVision in achieving classification accuracies for medical image analysis and disease detection.
The main contributions of this paper are summarized as follows:
  • Present a comprehensive CNN-LSTM-Attention hybrid model for medical image classification;
  • Demonstrate the effectiveness and generalizability of MediVision across ten diverse medical imaging datasets;
  • Investigate the ability of MediVision to seamlessly integrate and exploit spatial, temporal, and attention-based features;
  • Showcase the scalability and adaptability of MediVision for a broad range of medical image analysis tasks;
  • Conduct extensive comparative and statistical analyses across diverse medical imaging datasets to validate the consistency, reliability, and generalizability of MediVision.

2. Materials and Methods

The proposed methodology comprises multiple cascaded stages, such that it begins with dataset partitioning, followed by data augmentation, preprocessing, and then model training. Each medical dataset is first split into training (70%), validation (15%), and testing (15%) subsets to prevent data leakage and ensure unbiased evaluation. Subsequently, data augmentation is applied to the training and validation sets to increase the sample diversity and address class imbalance, while the test set remains untouched. After augmentation, the data undergo preprocessing steps including normalization and label encoding. The proposed hybrid model is then trained on the augmented training set, with performance monitored using the validation set. Final evaluation is conducted on the independent test set using standard metrics (i.e., accuracy, precision, recall, and F1-score) to assess both predictive performance and model resilience.

2.1. Dataset Description and Customization

The effectiveness of DL models in medical image analysis heavily relies on the quality and diversity of the datasets used for training, validation and testing. Ten diverse medical imaging datasets were utilized in this study, as summarized in Table 1, encompassing a wide range of medical conditions, including neuro-degenerative disorders, oncological abnormalities, and pulmonary diseases, which provides a comprehensive foundation for robust model use in medical image classification. These datasets were chosen as representatives of various imaging modalities and diagnostic challenges, enabling a thorough evaluation of the proposed framework. While our experiments focus on these conditions, MediVision is a generalizable framework that can be applied to a broader set of diseases with appropriate training data.

2.2. Data Augmentation and Normalization

One way to artificially expand the amount and complexity of current data is through data augmentation [53]. To give developers access to more representative training data, data augmentation techniques have been used to increase the size of training sets [54,55]. In this study, we have applied different techniques for data augmentation to enhance the diversity of the datasets and to achieve a balanced distribution of classes, as in the example illustrated in Figure 2. The augmentation methods that we have employed include rotation, horizontal and vertical flipping, and other transformations to simulate different perspectives and conditions with the parameters summarized in Table 2.
In image processing and ML-based approaches, normalizing pixel values is a standard preprocessing step. This process scales the pixel values of an image to a standardized range, typically between 0 and 1 [56]. This normalization is essential, because images are often represented in an 8-bit per channel format, where each pixel can range from 0 (black) to 255 (white) [57]. The range of pixel values is given in the closed interval [ 0 , 255 ] . The normalization equation is as follows:
X normalized = X X min X max X min ,
where X represents the pixel value, and  X min and X max are the minimum and maximum pixel values, respectively.

2.3. The Proposed MediVision Model

The MediVision model is a hybrid DL-based architecture, designed to combine the strengths of CNNs, LSTM, and attention mechanism, as shown in Figure 3. This combination allows the model to effectively process and analyze visual data, while capturing sequential dependencies and focusing on the most relevant parts of the input sequence. Additionally, a feature selection component is incorporated to refine the learned representations by prioritizing the most informative features and reducing redundancy, thereby enhancing the classification performance and model interpretability. The proposed MediVision model, with details of each block illustrated in Figure 3, is described as follows.

2.3.1. CNNs Unit

The convolution operation involves sliding a filter (also called a kernel) over an input to produce a feature map. This operation helps detect and preserve spatial hierarchies in images, allowing the learning of features like edges, textures, and more complex patterns [58]. Equation (2) indicates feature maps of a CNN layer.
f i j k = σ m = 0 M 1 n = 0 N 1 w m n k x ( i + m ) ( j + n ) + b k ,
where f i j k is the output feature map at position ( i , j ) for the k-th filter, w m n k is the weight at position ( m , n ) in the k-th filter, x ( i + m ) ( j + n ) is the input at position ( i + m , j + n ) , b k is the bias for the k-th filter, and  σ is the ReLU activation function.

2.3.2. Flattening and Reshaping Unit

Flattening and reshaping are essential preprocessing steps in this model architecture, bridging the gap between convolutional feature extraction and sequential analysis using LSTM and attention mechanisms. Flattening converts multidimensional feature maps into a format suitable for further processing, while reshaping tailors the data for sequential LSTM input, optimizing the model’s ability to capture sequential dependencies within the task of image classification. A tensor x after convolutional and pooling layers has the following shape:
x R ( b a t c h _ s i z e , h e i g h t , w i d t h , c h a n n e l s ) .
The flattening operation reshapes x into a 1D vector V:
V = flatten ( x ) ,
where
V R ( b a t c h _ s i z e , h e i g h t · w i d t h · c h a n n e l s ) .
The reshaping operation prepares the data for input into the LSTM layer, which expects sequences of vectors. After flattening, V is reshaped into a 3D tensor Y:
Y = reshape ( V , b a t c h _ s i z e , 1 , h e i g h t , w i d t h , c h a n n e l s ) ,
where
Y R ( b a t c h _ s i z e , 1 , h e i g h t , w i d t h , c h a n n e l s ) .
Here, 1 indicates that each input sequence to the LSTM consists of a single time step, ensuring compatibility with sequential processing. When replaced by 1 , this will indicate that the reshaping operation automatically determines the remaining dimension size based on the total elements of the flattened vector.

2.3.3. LSTM Unit

The LSTM layer in this model processes reshaped input data Y consisting of sequences of feature vectors. It computes hidden states that capture sequential dependencies within each sequence, thus giving it the ability to remember information over long periods.
H = LSTM ( Y ) ,
where
H R ( b a t c h _ s i z e , 1 , Units ) .
Here, H contains the hidden states for each sequence element (time step). The LSTM computes these hidden states using a set of learned weights and biases, which are updated during training to capture sequential dependencies in the data. The LSTM processes each time step t using the following gate mechanisms:
f t = σ ( W f [ h t 1 , x t ] + b f ) .
The forget gate f t decides how much of the previous cell state C t 1 to retain for the current time step t based on the previous hidden state h t 1 and the current input x t , which takes the input as described in (7) to (9) below:
I t = σ ( W i [ h t 1 , x t ] ) ,
c ˜ t = tanh ( W i [ h t 1 , x t ] + b i ) ,
c t = f t · C t 1 + I t · c ˜ t ,
where the input gate I t determines how much of the candidate values c ˜ t should be added to the cell state c t . The candidate cell state c ˜ t represents the new candidate values that could potentially be added to the cell state c t . The updated cell state c t combines the decision of the forget gate f t on what to forget from C t 1 and the input gate I t on what new information to store in c t .
The output gate O t in (10) determines how much of the updated cell state c t should be exposed as h t , which represents the hidden state computed based on the output gate O t and the updated cell state c t as in (11), such that tanh ( ) is the hyperbolic tangent activation function.
O t = σ ( W o [ h t 1 , x t ] + b o )
h t = O t · tanh ( c t )

2.3.4. Attention Mechanism

The attention mechanism allows the model to focus on the most relevant parts, which are the selected features of the input sequence by computing attention scores. These scores determine the importance of each element in the sequence, enabling the model to highlight and prioritize important features. The context vector produced by the attention mechanism is a weighted summary of the input sequence [59]. The input to the attention mechanism is the output sequence from the LSTM layer h t for t = 1 , 2 , , T , where T is the length of the sequence. The attention mechanism first computes a set of attention scores e t for each time step t based on (12):
e t = tanh ( W · h t + b ) ,
where W is a weight matrix, and b is a bias vector. The score e t can be seen as an intermediate representation that helps to determine the importance of each h t . The scores e t are then normalized using a softmax function to obtain the attention weights α t . These weights sum to 1 and represent the relative importance of each time step.
α t = exp ( e t ) k = 1 T exp ( e k ) ,
such that α t is the attention weight for time step t. The attention weights α t are used to compute a weighted sum of the LSTM outputs, resulting in a context vector C:
C = t = 1 T α t · h t ,
where C represents the weighted combination of the LSTM hidden states.

2.3.5. Skip Connection

The skip connection combines the output of the attention mechanism with the summed output of the LSTM layer as given in (15). This helps retain important information from the original sequence while also incorporating the focused information from the attention mechanism. Skip connections can improve the gradient flow and help in training deeper networks.
Combined Output = C + t = 1 T h t

2.3.6. Prediction Process

This is the final dense layer responsible for producing the classification output. The number of points taken into consideration in this process corresponds to the number of classes in the classification task. The softmax activation function ensures that the output is a probability distribution over the classes, such that
Y = softmax ( W d · Combined Output + b d ) ,
where W d is the weight matrix of the dense layer, b d is the bias of the dense layer, and Y is the final output, which is a probability distribution over the classes. Algorithm A1, provided in Appendix A, summarizes the procedural steps followed by our proposed MediVision model, including input processing, feature extraction using CNNs, LSTM, attention mechanism, skip connection, and predication layers, respectively, to effectively analyze medical images and accurately classify diseases based on learned features.
In this study, we adopted a trial-and-error-based approach to tune hyperparameters, which is a commonly used strategy in DL-based models when automated search methods are found to be of high computational cost [50,56,60]. Specifically, we experimented with various configurations for key parameters such as the number of CNN layers, LSTM units, dropout rates, and learning rates, using validation performance as the guiding metric. Given the heterogeneous nature of the ten medical image datasets investigated in this study, with each having different imaging modality and class distribution, we observed that a uniform configuration often led to suboptimal results on certain datasets. To address this, we performed dataset-specific hyperparameter tuning, and we have documented the codes and final settings for each dataset to enable the reproducibility of our results. While our manual tuning approach yielded competitive results, we acknowledge that a more systematic strategy, such as grid search or Bayesian optimization, could further refine the performance and efficiency. We consider this a promising direction for our future work and plan to explore it in subsequent research efforts. Notably, recent studies, such as the work in [61], have demonstrated the effectiveness of Bayesian optimization in complex model tuning tasks.

3. Results

In this section, the numerical results of the proposed MediVision model are observed and discussed from different perspectives. First, we consider the A c c u r a c y , which measures how often the model is correct across all classes. It is the ratio of the number of correct predictions considering True Positives ( T P ), True Negatives ( T N ), and the Total number of Instances ( T I ).
A c c u r a c y = T P + T N T I
P r e c i s i o n measures the proportion of T P instances among all instances predicted as positive, which comprises T P and False Positives ( F P ). It indicates how many of the predicted positives are truly positives.
P r e c i s i o n = T P T P + F P
R e c a l l measures the proportion of T P instances among all actual positive instances, which comprises T P and False Negatives ( F N ). It shows how well the model identifies positive instances.
R e c a l l = T P T P + F N
The F 1 - S c o r e is the harmonic mean of P r e c i s i o n and R e c a l l , providing a balance between them.
F 1 - S c o r e = 2 × P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l
T P s are the instances that the model correctly identifies as positive, meaning both the prediction and the actual status are positive. T N s represent the instances that the model accurately predicts as negative, where both the prediction and the actual status are negative. Conversely, F P s occur when the model incorrectly predicts a positive outcome for an instance that is negative, while F N s arise when the model mistakenly predicts a negative outcome for an instance that is actually positive.

3.1. Alzheimer’s Disease Dataset

The assessment of the proposed MediVision model for Alzheimer’s disease classification was demonstrated through accuracy-loss trends, confusion matrix and key metrics such as P r e c i s i o n , R e c a l l , and  F 1 - S c o r e . The accuracy and loss curves, illustrated in Figure 4, reveal a steady improvement in MediVision’s performance across training epochs. The training accuracy starts at approximately 60% and gradually increases, reaching around 95% by the final epoch. A similar trend is observed in validation accuracy, which consistently outperforms training accuracy at multiple points, ultimately stabilizing at approximately 96%. This close alignment between the training and validation accuracies suggests strong generalization capabilities, minimizing overfitting risks. Additionally, the loss curves indicate effective learning, with both training and validation losses consistently decreasing and stabilizing around 0.07, which is a large improvement compared to a 40% loss of both the VGG16 and VGG19 models. In addition, the smooth decline, without sharp fluctuations, demonstrates the efficient convergence of our proposed model.
The confusion matrix, which is illustrated in Figure 5, provides an in-depth analysis of the model’s classification performance across four categories: non-demented, very mild demented, mild demented, and moderate demented. The majority of samples are correctly classified, with 521, 527, 517, and 488 correctly classified instances, respectively. Although most samples are accurately classified, a minor misclassification trend is observed in the very mild demented category, where instances are occasionally misclassified as mild demented. This suggests possible feature similarities between these two conditions. However, the overall classification accuracy remains high, underscoring the model’s reliability in distinguishing between different dementia stages. The effectiveness of the model is further validated through the precision, recall, and F1-score metrics. The moderate demented category achieves an exceptional precision and recall of 100%, indicating perfect classification for this group. Similarly, other categories exhibit high performance, with precision values of 96% (mild demented), 98% (non-demented), and 97% (very mild demented). The recall scores remain consistently strong, with mild demented at 99%, moderate demented also at 99%, non-demented at 98%, and very mild demented at 96%. The F1-score, which represents a harmonic mean between precision and recall, follows a similar pattern, with an average score of 98%.

3.2. Blood Cell Dataset

The performance evaluation for the blood cell image dataset follows a similar approach, with the accuracy and loss results depicted in Figure 6, the training accuracy exhibits a consistent upward trend, beginning at lower levels and converging to over 95%, while the validation accuracy follows a similar trajectory, ultimately stabilizing at around 95% with minimal signs of overfitting. The corresponding loss curves reinforce this stability, showing a smooth and steady decrease over the epochs.
The confusion matrix, presented in Figure 7, reveals that the model accurately classifies the majority of samples across all cell types, including basophils, eosinophils, erythroblasts, Ig (immunoglobulins), lymphocytes, monocytes, neutrophils, and platelets, although occasional misclassifications occur, particularly among closely related classes such as Ig and monocytes. Still, the overall classification accuracy remains high, at approximately 95.45%. Furthermore, the detailed metrics underscore this effectiveness, with most classes attaining precision and recall values above 90%, resulting in F1-scores that consistently indicate strong performance. For instance, the eosinophil and neutrophil classes achieve near-perfect precision and recall, whereas Ig shows slightly lower metrics, suggesting a potential area for further refinement. Collectively, the high accuracy, well-balanced precision and recall, and smoothly converging loss function validate the model’s reliability in differentiating various blood cell types, thereby showcasing its utility in clinical diagnostic workflows.

3.3. Bone Fracture Multi-Region Dataset

The loss and accuracy curves for the bone fracture multi-region dataset, depicted in Figure 8, indicate high efficiency of our model in learning and generalization, where the training loss drops steeply from above 0.45 to nearly 0.05 within the first three epochs, demonstrating rapid learning, while the validation loss follows a similar declining trend, reaching below 0.05. By the end of training, both training and validation losses stabilize at very low values, signifying minimal errors. The accuracy curves further reinforce the model’s effectiveness, with the training accuracy rising sharply from approximately 80% to nearly 95% within the first few epochs, and the validation accuracy surpassing 95% by the second epoch, ultimately reaching nearly 98% at the final epoch with minimal overfitting and strong generalization.
At the same time, the confusion matrix in Figure 9 provides additional validation, showing that the first class achieves 687 T P s with only eight misclassifications as the second class, while the second class records 680 T P s with just 12 false positives, demonstrating excellent classification performance. The small number of misclassifications highlights the model’s high predictive accuracy, making it suitable for real-world bone fracture detection. Furthermore, the precision, recall, and F1-score for each class range between 98% and 99%, with an overall model accuracy of 98.56%, confirming its exceptional reliability.

3.4. Brain Tumor Dataset

The accuracy–loss curves for the brain tumor dataset are given in Figure 10, illustrating a consistent decrease in both training and validation loss over the epochs, with the training loss dropping sharply from approximately 0.7 to around 0.05 by the 20th epoch. This steep decline indicates that the model quickly learns the underlying patterns within the dataset. The validation loss follows a similar decreasing trend but remains slightly higher and fluctuates around 10% in later epochs, suggesting mild overfitting, where the model performs slightly better on training data than on unseen validation data. The accuracy curves further reinforce this observation, with the training accuracy rising steadily and reaching nearly 98%, while the validation accuracy, though slightly lower and exhibiting minor fluctuations, stabilizes around 95% by the final epochs.
Looking at the confusion matrix in Figure 11, we get further insights into the model’s classification performance across different brain tumor classes. The glioma class achieves the highest classification accuracy, with 522 T P s and minimal misclassifications into other categories. Similarly, the meningioma and pituitary classes exhibit strong predictive accuracy, with 508 and 495 T P s, respectively, though some samples are misclassified due to overlapping features among the three tumor types. These classification results are further supported by the performance metrics in Table 3, where all three classes exhibit high precision, recall, and F1-scores ranging between 95% and 98%. The overall accuracy of the model stands at 96.76%, signifying its robustness in distinguishing between different tumor types.

3.5. Breast Ultrasound Dataset

The accuracy and loss trends for the case of breast ultrasound image classification are illustrated in Figure 12, in which we notice a steady improvement over the training epochs, indicating effective learning. Initially, the training accuracy starts at approximately 75%, gradually increasing to reach around 98% by the final epoch. The validation accuracy follows a similar trajectory, consistently outperforming the training accuracy at certain points and ultimately stabilizing at approximately 97%, with minimal the risk of overfitting. The loss curves further reinforce this stability, exhibiting a consistent decline without abrupt fluctuations, signifying smooth convergence.
The confusion matrix, presented in Figure 13, provides insights into the classification performance for benign and malignant cases. The model correctly classifies 604 benign cases and 583 malignant cases. However, 14 benign samples are misclassified as malignant, while 17 malignant samples are misclassified as benign. Despite these minor misclassifications, the overall classification accuracy remains at an 97%, demonstrating the model’s robustness in distinguishing between benign and malignant breast tumors. Furthermore, the precision, recall, and F1-score metrics validate the model’s efficiency and reliability, as shown in Table 3. The benign class achieves a precision of 97%, recall of 98%, and  F1-score of 97%, indicating strong classification performance. Similarly, the malignant class exhibits a precision of 97.45% and recall of 98%, reflecting a balanced classification ability across both categories. The high recall values indicate that the model is highly effective in identifying malignant cases, which is critical for medical diagnosis.

3.6. Chest CT Scans Dataset

The loss and accuracy curves, as illustrated in Figure 14 demonstrate strong performance for the chest CT scans dataset. The training loss exhibits a substantial reduction from approximately 1 to 0.4 within the initial epochs, indicating a rapid learning phase. A similar trend is observed in the validation loss, which decreases from approximately 0.8 to 0.4, suggesting a well-balanced learning process and a reasonable level of generalization. As training progresses, both the training and validation losses continue to decline. Notably, the validation loss experiences fluctuations around epochs 10 and 14. However, in general, the losses remain balanced, affirming the model’s stability and robustness. The accuracy curves follow a similar pattern, reinforcing the effectiveness of the model. The training accuracy experiences a sharp rise from approximately 60% to over 90% within the first three epochs, indicating the model’s ability to quickly learn the dataset’s inherent patterns. The validation accuracy also improves steadily, closely following the training accuracy. After approximately five epochs, both the training and validation accuracy stabilize around 95%, with the training accuracy exhibiting minimal variation and the validation accuracy showing slight oscillations. These minor fluctuations in validation accuracy, reflected in the loss curves, indicate that the model maintains a well-calibrated performance without significant overfitting.
The confusion matrix, presented in Figure 15, provides further insights into the model’s classification performance across the four classes. The first class achieves 149  T P s, though eight instances are misclassified as belonging to the fourth class. The second class is correctly classified in 153 instances, with only five misclassified as the first class, suggesting some overlap in features between these two categories. The third class is handled with high precision, achieving 138 T P s with no misclassifications, underscoring the model’s strong ability to distinguish this class. The fourth class records 129 T P s but exhibits minor misclassification errors, with nine instances predicted as the first class and two as the third class. The overall accuracy of the model on this dataset is 94.83%, as shown in Table 3. These results confirm the model’s reliability in distinguishing between different categories within the chest CT scan while maintaining a strong balance between training and validation performance.

3.7. Chest X-Ray Dataset

The performance evaluation for the chest X-ray dataset, also follows the same structured approach. The accuracy–loss curves illustrated in Figure 16 indicate that the training accuracy exhibits a consistent upward trajectory, starting at approximately 84% and steadily increasing to nearly 98% over the epochs. Similarly, the validation accuracy follows a comparable trend, stabilizing around 96–97% with minimal signs of overfitting. The corresponding loss curves further reinforce this stability, showing a steady decline in both training and validation loss. This confirms that the model trains efficiently and achieves robust convergence.
Looking at the confusion matrix, which is presented in Figure 17, shows that the model classifies the majority of samples correctly across all categories: COVID-19, Normal, Pneumonia, and Tuberculosis. Notably, COVID-19 and Pneumonia classifications have a few misclassifications, with some Pneumonia cases being misclassified as Normal or COVID-19. However, the overall misclassification rate remains low, with a strong emphasis on correctly identifying disease conditions. The detailed evaluation metrics presented in Table 3 further validate the model’s high performance. The precision, recall, and F1-score values consistently exceed 97% across all classes, indicating a well-balanced classification model. The overall classification accuracy is 97.31%, confirming that the model is highly reliable in diagnosing chest-related conditions. While minor misclassifications exist, they do not significantly impact the overall performance, making the model highly suitable for clinical diagnostic applications.

3.8. Diabetic Retinopathy Dataset

The loss and accuracy curves for the diabetic retinopathy dataset, as illustrated in Figure 18, exhibit similar characteristics to those observed in other datasets. However, the proposed model demonstrates high efficiency despite some oscillations in validation performance. The training loss steadily decreases from approximately 0.28 to 0.12, indicating a consistent learning process, with all log loss values remaining below 0.05, suggesting that the model effectively learns and fits well to the training data. Although the validation loss follows a similar downward trend, it exhibits some fluctuations, particularly between epochs 3 and 6, before stabilizing and converging toward the training loss, demonstrating improved generalization. The training accuracy rapidly increases from approximately 88% to 94% within the first three epochs, signifying the model’s ability to quickly adapt to the dataset, while the validation accuracy also improves sharply, reaching around 95% by epoch 3, though it shows some minor oscillations between epochs 4 and 6, mirroring the validation loss. Nonetheless, by the final epochs, both the training and validation accuracy stabilize at approximately 95%, confirming that the model generalizes well without significant overfitting.
Additional insight in to the classification ability of the model is given by the confusion matrix, which is illustrated in Figure 19. The matrix indicates a T P value, which is much higher for both classes with 950 correct predictions for the first class, while 904 for the second class were correctly predicted. However, there are some misclassified, where 53 instances of the first class have been classified wrongly as the second class, and 43 instances of the second class were classified as the first class. These numbers are still low enough that it suggests the overall accuracy is good, but it also hints at some of the times where this model must get confused between the two classes, probably because the images making up the set of diabetic retinopathy are nuanced in their features. The balance between the T P s and the misclassifications suggests that while the model is highly effective, further refinement could help reduce these errors and improve the overall accuracy of the classification. The range of precision, recall, and F1-score is in between 95% and 96% for both the classes. The overall accuracy is 95.43%.

3.9. Kidney Diseases Dataset

The loss and accuracy curves for the kidney diseases dataset, given in Figure 20, demonstrate a highly optimized model with strong generalization capabilities. The training loss decreases steadily from 0.6189 to 0.0180, while the validation loss shows some fluctuations, starting at 0.2786 and ending at 0.0771. The accuracy trends confirm the model’s strong performance. The training accuracy increases sharply from 75% to 95% within the first five epochs, showing the model’s quick adaptation, while the validation accuracy closely follows, occasionally surpassing the training accuracy, highlighting strong generalization. After the initial phase, the training accuracy stabilizes around 97% with minor fluctuations, while the validation accuracy oscillates slightly around 94%, maintaining high performance on both sets.
The confusion matrix in Figure 21 further validates the model’s classification effectiveness across different classes. The first class achieves perfect classification with 434 T P s and no F P s. The second class is also well-classified, with 431 T P s, though 18 samples are misclassified as the third class, indicating some difficulty in distinguishing between these two classes. The third class has 424 T P s but shows eight misclassifications as the second class, reinforcing this challenge. The fourth class exhibits excellent classification, with 446 T P s and only three minor misclassifications, with two instances misclassified as the second class. Overall, the model demonstrates strong classification performance, with occasional confusion between the second and third classes. The overall accuracy for the kidney diseases dataset is 98.36%, as summarized in Table 3.

3.10. Retinal OCT Dataset

The loss curves for the retinal OCT dataset are presented in Figure 22, showing a significant and consistent reduction in training loss, which declines from approximately 1.2 to nearly 0.01 by the 25th epoch. The validation loss follows a similar decreasing trend; however, it exhibits minor fluctuations, particularly between the fifth and tenth epochs. This variability in validation loss suggests that, while the model effectively learns feature representations, some inconsistencies in the validation set may contribute to occasional performance fluctuations. Despite these variations, the overall decline in validation loss indicates enhanced generalization capability, and both the training and validation losses eventually converge at low values, signifying a stable and well-optimized model. The accuracy curves further support the model’s strong performance, demonstrating a rapid increase in training accuracy, surpassing 90% within the first five epochs and continuing to improve steadily. The validation accuracy follows a comparable trajectory, albeit slightly lower than the training accuracy, with minor fluctuations mirroring those observed in validation loss. By the 25th epoch, both the training and validation accuracies stabilize at approximately 95%, which indicates minimal overfitting.
The confusion matrix in Figure 23 provides a detailed evaluation across different classes in the retinal OCT dataset. The first class exhibits near-perfect classification, achieving 203 T P s with no observed misclassifications, highlighting the model’s high precision for this category. The second class also demonstrates strong classification performance with 179 T P s; however, six instances are misclassified as the first class, and a few others are incorrectly assigned to different categories. The third class, with 149 T P s, shows a slightly higher misclassification rate, with certain instances incorrectly categorized as either the second or fourth class. The fourth class attains 174 T P s with minimal classification errors, primarily involving minor confusion with the third class. Overall, the model exhibits robust classification performance, with only minor misclassifications observed in the second and third classes, effectively distinguishing between the majority of classes and providing high precision and recall as summarized in Table 3.

4. Discussion

4.1. Model Interpretability via Explainable AI

Additionally, we employed Grad-CAM heatmaps to visually explain MediVision’s decision-making process across representative medical imaging datasets, namely (a) bone fracture multi-region, (b) chest X-ray, and (c) Alzheimer’s disease, as illustrated in Figure 24. These heatmaps highlight the image regions that contribute most significantly to the model’s class predictions. Since MediVision performs classification rather than segmentation, the Grad-CAM heatmaps emphasize discriminative features learned by the model, rather than anatomically precise regions of interest. The intermediate convolutional layers act as feature detectors, and the resulting activations, visualized as intensity-based overlays, reveal where the model focuses during classification. This visual interpretation validates that MediVision accurately identifies significant disease characteristics within the input image. The consistency of highlighted regions across these representative datasets supports the model’s generalizability and its ability to detect subtle disease patterns across diverse imaging modalities. These findings demonstrate MediVision’s potential to enhance diagnostic accuracy and reduce human error in clinical settings.

4.2. Comparative Analysis

The performance comparison between the proposed MediVision model and existing DL-based models, including VGG16, VGG19, and ResNet50, across all ten diverse medical image datasets is summarized in Table 4. Notably, MediVision showed substantial improvements in Alzheimer’s disease, achieving 98% compared to the 75–85% range of the other models. The results also demonstrated significant gains in blood cell classification with 95.45% accuracy and brain tumor detection reaching 96.76%, a considerable advancement over ResNet50’s by 24.44%. Significant gains were also observed in breast ultrasound, chest X-ray/CT, and kidney disease classification. For retinal OCT classification, while the benchmark models already achieve high accuracy, MediVision still demonstrates a considerable improvement in comparison.
In addition, Table 5 demonstrates MediVision’s competitive edge against state-of-the-art DL-based models, including EfficientNet, DenseNet, and Hybrid CNN-LSTM, across diverse medical image datasets. The table shows that MediVision is capable of achieving superior or comparable accuracy in nearly all categories, as a result of its advanced architecture and learning capabilities, particularly for classifying complex medical images like brain tumor and Alzheimer’s disease. These results were achieved through a consistent training strategy based on empirical tuning, where a common set of hyperparameters was identified using a hit-and-trial method. Specifically, we used a batch size of 16, a learning rate of 0.01 with the Adam optimizer, and dropout rates between 0.1 and 0.3. Additionally, early stopping with a patience of three epochs was applied, which helped ensure strong performance across multiple datasets. Overall, the results confirm MediVision’s enhanced generalization and accuracy, highlighting its potential for robust and reliable medical image analysis and disease diagnosis.

4.3. Statistical Analysis

To validate the effectiveness of the proposed MediVision hybrid CNN–LSTM model, statistical significance testing was performed against three baseline convolutional neural network architectures such as VGG16, VGG19, and ResNet50 across ten diverse medical imaging datasets. As summarized in Table 6, MediVision achieved the highest mean classification accuracy of 96.89%, outperforming VGG16 (92.98%), VGG19 (92.79%), and ResNet50 (81.17%). These consistent improvements demonstrate the superior generalization and stability of MediVision across multiple medical imaging modalities.
Pairwise comparisons between MediVision and each baseline model were conducted using both the paired t-test and the non-parametric Wilcoxon signed-rank test, as summarized in Table 7. The Wilcoxon tests yielded statistically significant results (p < 0.01) for all comparisons, confirming that the observed improvements were not due to random variation. Although the paired t-test for the comparison with VGG16 was marginal (p = 0.0597), the non-parametric Wilcoxon result confirmed significance. Furthermore, after applying the Holm–Bonferroni correction to control for multiple comparisons, all adjusted p-values remained below 0.01, reinforcing the robustness of the findings.
Finally, an overall comparison among all four models was performed using the Friedman test, a non-parametric alternative to repeated-measures ANOVA. As shown in Table 8, the test produced a χ 2 value of 27.00 with a p-value of 0.000006 (p < 0.001), indicating a highly significant difference in performance across models. These results collectively demonstrate that MediVision provides statistically and practically significant improvements over conventional CNN architectures for medical image classification tasks.

5. Conclusions

This study introduces a hybrid framework for medical image classification that seamlessly integrates CNNs, LSTMs, and an attention mechanism with a skip connection to enhance the feature extraction and classification accuracy. Trained and tested on ten diverse representative medical image datasets, MediVision outperforms widely used pretrained models such as VGG16, VGG19, and ResNet50 by 3.91%, 4.10%, and 15.72%, respectively, in terms of the mean classification accuracy. The results of both paired statistical tests and the Friedman test confirm that these improvements are statistically significant (p = 0.000006) across multiple medical imaging modalities. The model’s strength lies in its ability to effectively capture both spatial and temporal features, with CNNs extracting spatial information, LSTMs learning sequential dependencies from image sequences, and the attention mechanism refining these features by focusing on the most relevant regions. The inclusion of a skip connection ensures efficient feature propagation, further boosting classification performance. To enhance the model transparency and support the clinical interpretability, Grad-CAM heatmaps were employed to visually highlight the critical image regions influencing classification decisions. Additionally, the datasets were carefully augmented and balanced, ensuring optimal generalization and robustness across different medical imaging domains.
For future work, we plan to conduct a detailed analysis of the computational complexity and runtime performance to evaluate MediVision’s scalability in real-world clinical environments. A comprehensive ablation study will also be undertaken to assess the individual and combined contributions of key architectural components (i.e., CNN, LSTM, attention mechanism, and skip connection). Furthermore, we aim to enhance the feature extraction and interpretability by exploring self-supervised learning and transformer-based architectures. To improve the generalizability, we will extend MediVision to multi-modal medical data, such as integrating MRI, CT scans, and histopathological images. In addition, future work will explore the application of federated learning to facilitate collaborative model training across decentralized medical institutions while preserving data privacy and the integration of MediVision into clinical decision support systems to enable real-time AI-driven diagnostics.

Author Contributions

Conceptualization, M.T.H., Y.M.A., and S.M.S.; methodology, M.T.H., Y.M.A. and A.A.; software, M.T.H., A.A. and H.K.; validation, M.T.H., W.A., S.M.S. and A.I.Z.; formal analysis, M.T.H., W.A. and S.M.S.; investigation, M.T.H., Y.M.A., A.A. and A.I.Z.; resources, M.T.H., W.A., S.M.S., H.K. and A.I.Z.; data curation, M.T.H., S.M.S. and A.A.; writing—original draft preparation, M.T.H., Y.M.A., and S.M.S.; writing—review and editing, M.T.H., Y.M.A., W.A., S.M.S. and H.K.; visualization, M.T.H., Y.M.A., and W.A.; supervision, Y.M.A. and S.M.S.; funding acquisition, Y.M.A. All authors have read and agreed to the published version of the manuscript.

Funding

Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2025R829), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The codes and datasets used within this study are made available through the public repository Zenodo at https://doi.org/10.5281/zenodo.15498860.

Conflicts of Interest

Authors Md. Tanvir Hayat and Salman Md Sultan are employed by the company Innovative Skills Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Appendix A

The following algorithm summarizes the workflow of the proposed MediVision model.
Algorithm A1 Proposed MediVision Procedure
Input: Image I with three channels ( H , W , C ) , number of CNN layers N, LSTM units U, and number of classes K;
  1:
Step 1—CNN Feature Extraction:
  2:
for  i = 1 to N do
  3:
       x Conv2D ( f i l t e r s i , ( 3 , 3 ) , ReLU ) on I;
  4:
       x MaxPooling2D ( 2 , 2 ) ;
  5:
       x Dropout ( d r o p o u t i ) ;
  6:
end for
  7:
x Flatten ( x ) ;
  8:
Step 2—Reshape for LSTM:
  9:
x Reshape ( 1 , 1 ) ;
10:
Step 3—LSTM Processing:
11:
l s t m o u t LSTM ( U , return _ sequences   =   True ) on x;
12:
l s t m o u t Dropout ( d r o p o u t l s t m ) ;
13:
Step 4—Attention Mechanism:
14:
a t t e n t i o n o u t Attention ( l s t m o u t ) ;
15:
Step 5—Skip Connection:
16:
c o m b i n e d o u t ← Add( a t t e n t i o n o u t , ∑   l s t m o u t );
17:
Step 6—Final Classification Layer:
18:
y Dense ( K , softmax or sigmoid ) on c o m b i n e d o u t ;
19:
Step 7—Compilation and Training:
20:
Loss: Sparse Categorical Crossentropy (multiclass) or Binary Crossentropy (binary);
21:
Optimizer: Adam;
22:
Metrics: Accuracy;
23:
Apply Model Checkpoint and Early Stopping Callbacks;
Output: Classified label y;

References

  1. Li, M.; Jiang, Y.; Zhang, Y.; Zhu, H. Medical image analysis using deep learning algorithms. Front. Public Health 2023, 11, 1273253. [Google Scholar] [CrossRef]
  2. Rajula, H.S.R.; Verlato, G.; Manchia, M.; Antonucci, N.; Fanos, V. Comparison of conventional statistical methods with machine learning in medicine: Diagnosis, drug development, and treatment. Medicina 2020, 56, 455. [Google Scholar] [CrossRef] [PubMed]
  3. Kingslake, J.; Dias, R.; Dawson, G.R.; Simon, J.; Goodwin, G.M.; Harmer, C.J.; Morriss, R.; Brown, S.; Guo, B.; Dourish, C.T.; et al. The effects of using the PReDicT Test to guide the antidepressant treatment of depressed patients: Study protocol for a randomised controlled trial. Trials 2017, 18, 558. [Google Scholar] [CrossRef]
  4. Chen, C. Ascent of machine learning in medicine. Nat. Mater. 2019, 18, 407. [Google Scholar] [CrossRef]
  5. Saria, S.; Butte, A.; Sheikh, A. Better medicine through machine learning: What’s real, and what’s artificial? PLoS Med. 2018, 15, e1002721. [Google Scholar] [CrossRef]
  6. Churpek, M.M.; Yuen, T.C.; Winslow, C.; Meltzer, D.O.; Kattan, M.W.; Edelson, D.P. Multicenter comparison of machine learning methods and conventional regression for predicting clinical deterioration on the wards. Crit. Care Med. 2016, 44, 368–374. [Google Scholar] [CrossRef]
  7. Hernesniemi, J.A.; Mahdiani, S.; Tynkkynen, J.A.; Lyytikäinen, L.P.; Mishra, P.P.; Lehtimäki, T.; Eskola, M.; Nikus, K.; Antila, K.; Oksala, N. Extensive phenotype data and machine learning in prediction of mortality in acute coronary syndrome–the MADDEC study. Ann. Med. 2019, 51, 156–163. [Google Scholar] [CrossRef]
  8. Paik, E.S.; Lee, J.W.; Park, J.Y.; Kim, J.H.; Kim, M.; Kim, T.J.; Choi, C.H.; Kim, B.G.; Bae, D.S.; Seo, S.W. Prediction of survival outcomes in patients with epithelial ovarian cancer using machine learning methods. J. Gynecol. Oncol. 2019, 30, e65. [Google Scholar] [CrossRef]
  9. Abràmoff, M.D.; Lou, Y.; Erginay, A.; Clarida, W.; Amelon, R.; Folk, J.C.; Niemeijer, M. Improved automated detection of diabetic retinopathy on a publicly available dataset through integration of deep learning. Investig. Ophthalmol. Vis. Sci. 2016, 57, 5200–5206. [Google Scholar] [CrossRef]
  10. Azizi, S.; Mustafa, B.; Ryan, F.; Beaver, Z.; Freyberg, J.; Deaton, J.; Loh, A.; Karthikesalingam, A.; Kornblith, S.; Chen, T.; et al. Big self-supervised models advance medical image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 3478–3488. [Google Scholar]
  11. Xiao, C.; Choi, E.; Sun, J. Opportunities and challenges in developing deep learning models using electronic health records data: A systematic review. J. Am. Med. Inform. Assoc. 2018, 25, 1419–1428. [Google Scholar] [CrossRef]
  12. Shickel, B.; Tighe, P.J.; Bihorac, A.; Rashidi, P. Deep EHR: A survey of recent advances in deep learning techniques for electronic health record (EHR) analysis. IEEE J. Biomed. Health Inform. 2017, 22, 1589–1604. [Google Scholar] [CrossRef] [PubMed]
  13. Yadav, P.; Steinbach, M.; Kumar, V.; Simon, G. Mining electronic health records (EHRs) A survey. ACM Comput. Surv. (CSUR) 2018, 50, 1–40. [Google Scholar] [CrossRef]
  14. Kim, H.E.; Cosa-Linan, A.; Santhanam, N.; Jannesari, M.; Maros, M.E.; Ganslandt, T. Transfer learning for medical image classification: A literature review. BMC Med. Imaging 2022, 22, 69. [Google Scholar] [CrossRef]
  15. Puttagunta, M.; Ravi, S. Medical image analysis based on deep learning approach. Multimed. Tools Appl. 2021, 80, 24365–24398. [Google Scholar] [CrossRef]
  16. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar] [CrossRef]
  17. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Annual Conference on Neural Information Processing Systems 2014, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  18. Chowdhary, C.L.; Mittal, M.; P, K.; Pattanaik, P.A.; Marszalek, Z. An efficient segmentation and classification system in medical images using intuitionist possibilistic fuzzy C-mean clustering and fuzzy SVM algorithm. Sensors 2020, 20, 3903. [Google Scholar] [CrossRef]
  19. Kumar, R.R.; Shankar, S.V.; Jaiswal, R.; Ray, M.; Budhlakoti, N.; Singh, K. Advances in Deep Learning for Medical Image Analysis: A Comprehensive Investigation. J. Stat. Theory Pract. 2025, 19, 9. [Google Scholar] [CrossRef]
  20. Kourounis, G.; Elmahmudi, A.A.; Thomson, B.; Hunter, J.; Ugail, H.; Wilson, C. Computer image analysis with artificial intelligence: A practical introduction to convolutional neural networks for medical professionals. Postgrad. Med. J. 2023, 99, 1287–1294. [Google Scholar] [CrossRef] [PubMed]
  21. Shi, Z.; Hu, Y.; Mo, G.; Wu, J. Attention-based CNN-LSTM and XGBoost hybrid model for stock prediction. arXiv 2022, arXiv:2204.02623. [Google Scholar]
  22. de Sousa, P.M.; Carneiro, P.C.; Pereira, G.M.; Oliveira, M.M.; da Costa Junior, C.A.; de Moura, L.V.; Mattjie, C.; da Silva, A.M.; Macedo, T.A.A.; Patrocinio, A.C. Towards a Classification Model using CNN and Wavelets applied to COVID-19 CT images. Res. Soc. Dev. 2022, 11, e2411527919. [Google Scholar] [CrossRef]
  23. Barzekar, H.; Yu, Z. C-Net: A reliable convolutional neural network for biomedical image classification. Expert Syst. Appl. 2022, 187, 116003. [Google Scholar] [CrossRef]
  24. Zebari, N.A.; Mohammed, C.N.; Zebari, D.A.; Mohammed, M.A.; Zeebaree, D.Q.; Marhoon, H.A.; Abdulkareem, K.H.; Kadry, S.; Viriyasitavat, W.; Nedoma, J.; et al. A deep learning fusion model for accurate classification of brain tumours in Magnetic Resonance images. CAAI Trans. Intell. Technol. 2024, 9, 790–804. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Wang, S.; Zhao, H.; Guo, Z.; Sun, D. CT image classification based on convolutional neural network. Neural Comput. Appl. 2021, 33, 8191–8200. [Google Scholar] [CrossRef]
  26. Hu, C.; Cao, N.; Zhou, H.; Guo, B. Medical Image Classification with a Hybrid SSM Model Based on CNN and Transformer. Electronics 2024, 13, 3094. [Google Scholar] [CrossRef]
  27. Kadhim, Y.A.; Guzel, M.S.; Mishra, A. A Novel hybrid machine learning-based system using deep learning techniques and meta-heuristic algorithms for various medical datatypes classification. Diagnostics 2024, 14, 1469. [Google Scholar] [CrossRef]
  28. Aydın, M.; Kuş, Z.; Akçelik, Z.K. A Hybrid CNN-Tree Based Model for Enhanced Image Classification Performance. In Proceedings of the 2024 32nd Signal Processing and Communications Applications Conference (SIU), Mersin, Türkiye, 15–18 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–4. [Google Scholar]
  29. Bouslihim, I.; Cherif, W.; Kissi, M. Application of a hybrid EfficientNet-SVM model to medical image classification. In Proceedings of the 2023 14th International Conference on Intelligent Systems: Theories and Applications (SITA), Casablanca, Morocco, 22–23 November 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar]
  30. Ajlouni, N.; Özyavaş, A.; Takaoğlu, M.; Takaoğlu, F.; Ajlouni, F. Medical image diagnosis based on adaptive Hybrid Quantum CNN. BMC Med. Imaging 2023, 23, 126. [Google Scholar] [CrossRef]
  31. Abbas, Q. An intelligent medical image classification system using few-shot learning. Concurr. Comput. Pract. Exp. 2023, 35, e7451. [Google Scholar] [CrossRef]
  32. Borji, A.; Kronreif, G.; Angermayr, B.; Hatamikia, S. Advanced hybrid deep learning model for enhanced evaluation of osteosarcoma histopathology images. Front. Med. 2025, 12, 1555907. [Google Scholar] [CrossRef]
  33. Hussain, T.; Shouno, H.; Hussain, A.; Hussain, D.; Ismail, M.; Mir, T.H.; Hsu, F.R.; Alam, T.; Akhy, S.A. EFFResNet-ViT: A fusion-based convolutional and vision transformer model for explainable medical image classification. IEEE Access 2025, 13, 54040–54068. [Google Scholar] [CrossRef]
  34. Salehin, I.; Islam, M.S.; Amin, N.; Baten, M.A.; Noman, S.; Saifuzzaman, M.; Yazmyradov, S. Real-Time Medical Image Classification with ML Framework and Dedicated CNN–LSTM Architecture. J. Sens. 2023, 2023, 3717035. [Google Scholar] [CrossRef]
  35. Gonçalves, T.; Rio-Torto, I.; Teixeira, L.F.; Cardoso, J.S. A survey on attention mechanisms for medical applications: Are we moving toward better algorithms? IEEE Access 2022, 10, 98909–98935. [Google Scholar] [CrossRef]
  36. Cheng, D.; Deng, J.; Xiao, J.; Yanyan, M.; Kang, J.; Gai, J.; Zhang, B.; Zhao, F. Attention based multi-scale nested network for biomedical image segmentation. Heliyon 2024, 10, e33892. [Google Scholar] [CrossRef]
  37. Liu, S.; Yue, W.; Guo, Z.; Wang, L. Multi-branch CNN and grouping cascade attention for medical image classification. Sci. Rep. 2024, 14, 15013. [Google Scholar] [CrossRef]
  38. Altwijri, O.; Alanazi, R.; Aleid, A.; Alhussaini, K.; Aloqalaa, Z.; Almijalli, M.; Saad, A. Novel deep-learning approach for automatic diagnosis of Alzheimer’s disease from MRI. Appl. Sci. 2023, 13, 13051. [Google Scholar] [CrossRef]
  39. Raptis, C.; Karavasilis, E.; Anastasopoulos, G.; Adamopoulos, A. Comparative Analysis of Conventional CNN v’s ImageNet Pretrained ResNet in Medical Image Classification. Information 2024, 15, 806. [Google Scholar] [CrossRef]
  40. Acevedo, A.; Alférez, S.; Merino, A.; Puigví, L.; Rodellar, J. Recognition of peripheral blood cell images using convolutional neural networks. Comput. Methods Programs Biomed. 2019, 180, 105020. [Google Scholar] [CrossRef]
  41. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 2018, 172, 1122–1131. [Google Scholar] [CrossRef]
  42. Rahman, T.; Khandakar, A.; Kadir, M.A.; Islam, K.R.; Islam, K.F.; Mazhar, R.; Hamid, T.; Islam, M.T.; Kashem, S.; Mahbub, Z.B.; et al. Reliable tuberculosis detection using chest X-ray with deep learning, segmentation and visualization. IEEE Access 2020, 8, 191586–191601. [Google Scholar] [CrossRef]
  43. Bacellar, G.C.; Chandrappa, M.; Kulkarni, R.; Dey, S. COVID-19 chest X-ray image classification using deep learning. medRxiv 2021. [Google Scholar] [CrossRef]
  44. Asif, S.; Wenhui, Y.; Jin, H.; Jinhai, S. Classification of COVID-19 from chest X-ray images using deep convolutional neural network. In Proceedings of the 2020 IEEE 6th International Conference on Computer and Communications (ICCC), Chengdu, China, 11–14 December 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 426–433. [Google Scholar]
  45. Altayeb, M.; Arabiat, A.; Al-Ghraibah, A. Detection and classification of pneumonia using the Orange3 data mining tool. Int. J. Electr. Comput. Eng. (2088-8708) 2024, 14, 6894–6903. [Google Scholar] [CrossRef]
  46. Jassim, O.A.; Abed, M.J.; Saied, Z.H.S. Deep learning techniques in the cancer-related medical domain: A transfer deep learning ensemble model for lung cancer prediction. Baghdad Sci. J. 2024, 21, 1101. [Google Scholar] [CrossRef]
  47. Benítez, V.E.C.; Matto, I.C.; Román, J.C.M.; Noguera, J.L.V.; García-Torres, M.; Ayala, J.; Pinto-Roa, D.P.; Gardel-Sotomayor, P.E.; Facon, J.; Grillo, S.A. Dataset from fundus images for the study of diabetic retinopathy. Data Brief 2021, 36, 107068. [Google Scholar] [CrossRef]
  48. Bhandari, M.; Yogarajah, P.; Kavitha, M.S.; Condell, J. Exploring the capabilities of a lightweight CNN model in accurately identifying renal abnormalities: Cysts, stones, and tumors, using LIME and SHAP. Appl. Sci. 2023, 13, 3125. [Google Scholar] [CrossRef]
  49. AlMahadin, G.; Owida, H.A.; Al-Nabulsi, J.I.; Turab, N.; Al Hawamdeh, N. Automated detection of kidney masses lesions using a deep learning approach. Int. J. Artif. Intell. ISSN 2024, 2252, 2863. [Google Scholar] [CrossRef]
  50. Aldhyani, T.; Ahmed, Z.A.; Alsharbi, B.M.; Ahmad, S.; Al-Adhaileh, M.H.; Kamal, A.H.; Almaiah, M.; Nazeer, J. Diagnosis and detection of bone fracture in radiographic images using deep learning approaches. Front. Med. 2025, 11, 1506686. [Google Scholar] [CrossRef]
  51. Gholami, P.; Roy, P.; Parthasarathy, M.K.; Lakshminarayanan, V. OCTID: Optical coherence tomography image database. Comput. Electr. Eng. 2020, 81, 106532. [Google Scholar] [CrossRef]
  52. Hasan, N.; Ahmed, M.F.; Nasif, M.A.; Haq, M.R.; Rahman, M. Hybrid feature extraction approach for robust brain tumor classification: HOG, GLCM, and artificial neural network. In Proceedings of the 2024 6th International Conference on Electrical Engineering and Information & Communication Technology (ICEEICT), Dhaka, Bangladesh, 2–4 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1292–1297. [Google Scholar]
  53. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  54. Lemley, J.; Bazrafkan, S.; Corcoran, P. Smart augmentation learning an optimal data augmentation strategy. IEEE Access 2017, 5, 5858–5869. [Google Scholar] [CrossRef]
  55. Moreno-Barea, F.J.; Jerez, J.M.; Franco, L. Improving classification accuracy using data augmentation on small data sets. Expert Syst. Appl. 2020, 161, 113696. [Google Scholar] [CrossRef]
  56. Singh, D.; Singh, B. Investigating the impact of data normalization on classification performance. Appl. Soft Comput. 2020, 97, 105524. [Google Scholar] [CrossRef]
  57. Wang, C.; Zhang, L.; Wei, W.; Zhang, Y. Dynamic super-pixel normalization for robust hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5505713. [Google Scholar] [CrossRef]
  58. Chen, B.; Li, H.; Luo, W. Image processing operations identification via convolutional neural network. arXiv 2017, arXiv:1709.02908. [Google Scholar] [CrossRef]
  59. Hao, S.; Lee, D.H.; Zhao, D. Sequence to sequence learning with attention mechanism for short-term passenger flow prediction in large-scale metro system. Transp. Res. Part C Emerg. Technol. 2019, 107, 287–300. [Google Scholar] [CrossRef]
  60. Sethuraman, S.K.; Malaiyappan, N.; Ramalingam, R.; Basheer, S.; Rashid, M.; Ahmad, N. Predicting Alzheimer’s disease using deep neuro-functional networks with resting-state fMRI. Electronics 2023, 12, 1031. [Google Scholar] [CrossRef]
  61. Iqbal, M.A.; Gil, J.M.; Kim, S.K. Attention-Driven Hybrid Ensemble Approach with Bayesian Optimization for Accurate Energy Forecasting in Jeju Island’s Renewable Energy System. IEEE Access 2025, 13, 7986–8010. [Google Scholar] [CrossRef]
  62. Asghar, R.; Kumar, S.; Hynds, P. Automatic classification of 10 blood cell subtypes using transfer learning via pre-trained convolutional neural networks. Informatics Med. Unlocked 2024, 49, 101542. [Google Scholar] [CrossRef]
  63. Singh, D.; Kumar, V.; Yadav, V.; Kaur, M. Deep neural network-based screening model for COVID-19-infected patients using chest X-ray images. Int. J. Pattern Recognit. Artif. Intell. 2021, 35, 2151004. [Google Scholar] [CrossRef]
  64. Mamun, M.; Mahmud, M.I.; Meherin, M.; Abdelgawad, A. Lcdctcnn: Lung cancer diagnosis of ct scan images using cnn based model. In Proceedings of the 2023 10th International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 23–24 March 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 205–212. [Google Scholar]
  65. Wang, Z.; Yin, Y.; Shi, J.; Fang, W.; Li, H.; Wang, X. Zoom-in-net: Deep mining lesions for diabetic retinopathy detection. In Proceedings of the Medical Image Computing and Computer Assisted Intervention- MICCAI 2017: 20th International Conference, Quebec City, QC, Canada, 11–13 September 2017; Proceedings, Part III 20. Springer: Berlin/Heidelberg, Germany, 2017; pp. 267–275. [Google Scholar]
  66. Dai, H.; Yang, Y.; Yue, X.; Chen, S. Improving retinal OCT image classification accuracy using medical pre-training and sample replication methods. Biomed. Signal Process. Control 2024, 91, 106019. [Google Scholar] [CrossRef]
  67. Khan, S.U.R.; Zhao, M.; Asif, S.; Chen, X. Hybrid-NET: A fusion of DenseNet169 and advanced machine learning classifiers for enhanced brain tumor diagnosis. Int. J. Imaging Syst. Technol. 2024, 34, e22975. [Google Scholar] [CrossRef]
Figure 1. Visual representation of medical datasets used in the study, illustrating sample images for (a) Alzheimer’s disease, (b) breast ultrasound, (c) blood cell, (d) chest X-ray, (e) chest CT scans, (f) diabetic retinopathy, (g) kidney diseases, (h) bone fracture multi-region, (i) retinal OCT, and (j) brain tumor.
Figure 1. Visual representation of medical datasets used in the study, illustrating sample images for (a) Alzheimer’s disease, (b) breast ultrasound, (c) blood cell, (d) chest X-ray, (e) chest CT scans, (f) diabetic retinopathy, (g) kidney diseases, (h) bone fracture multi-region, (i) retinal OCT, and (j) brain tumor.
Diagnostics 15 02673 g001
Figure 2. Illustration of applied data augmentation techniques.
Figure 2. Illustration of applied data augmentation techniques.
Diagnostics 15 02673 g002
Figure 3. Proposed MediVision architecture.
Figure 3. Proposed MediVision architecture.
Diagnostics 15 02673 g003
Figure 4. Accuracy and loss graph of Alzheimer’s disease dataset.
Figure 4. Accuracy and loss graph of Alzheimer’s disease dataset.
Diagnostics 15 02673 g004
Figure 5. Confusion matrix of Alzheimer’s disease dataset.
Figure 5. Confusion matrix of Alzheimer’s disease dataset.
Diagnostics 15 02673 g005
Figure 6. Accuracy and loss graph of blood cell dataset.
Figure 6. Accuracy and loss graph of blood cell dataset.
Diagnostics 15 02673 g006
Figure 7. Confusion Matrix of Blood Cell dataset.
Figure 7. Confusion Matrix of Blood Cell dataset.
Diagnostics 15 02673 g007
Figure 8. Accuracy and loss graph of bone fracture multi-region dataset.
Figure 8. Accuracy and loss graph of bone fracture multi-region dataset.
Diagnostics 15 02673 g008
Figure 9. Confusion matrix of bone fracture multi-region dataset.
Figure 9. Confusion matrix of bone fracture multi-region dataset.
Diagnostics 15 02673 g009
Figure 10. Accuracy and loss graph of brain tumor dataset.
Figure 10. Accuracy and loss graph of brain tumor dataset.
Diagnostics 15 02673 g010
Figure 11. Confusion matrix of brain tumor dataset.
Figure 11. Confusion matrix of brain tumor dataset.
Diagnostics 15 02673 g011
Figure 12. Accuracy and loss graph of breast ultrasound dataset.
Figure 12. Accuracy and loss graph of breast ultrasound dataset.
Diagnostics 15 02673 g012
Figure 13. Confusion matrix of breast ultrasound dataset.
Figure 13. Confusion matrix of breast ultrasound dataset.
Diagnostics 15 02673 g013
Figure 14. Accuracy and loss graph of chest CT scans dataset.
Figure 14. Accuracy and loss graph of chest CT scans dataset.
Diagnostics 15 02673 g014
Figure 15. Confusion matrix of chest CT scans dataset.
Figure 15. Confusion matrix of chest CT scans dataset.
Diagnostics 15 02673 g015
Figure 16. Accuracy and loss graph of chest X-ray dataset.
Figure 16. Accuracy and loss graph of chest X-ray dataset.
Diagnostics 15 02673 g016
Figure 17. Confusion matrix of chest X-ray dataset.
Figure 17. Confusion matrix of chest X-ray dataset.
Diagnostics 15 02673 g017
Figure 18. Accuracy and loss graph of diabetic retinopathy dataset.
Figure 18. Accuracy and loss graph of diabetic retinopathy dataset.
Diagnostics 15 02673 g018
Figure 19. Confusion matrix of diabetic retinopathy dataset.
Figure 19. Confusion matrix of diabetic retinopathy dataset.
Diagnostics 15 02673 g019
Figure 20. Accuracy and loss graph of kidney diseases dataset.
Figure 20. Accuracy and loss graph of kidney diseases dataset.
Diagnostics 15 02673 g020
Figure 21. Confusion matrix of kidney diseases dataset.
Figure 21. Confusion matrix of kidney diseases dataset.
Diagnostics 15 02673 g021
Figure 22. Accuracy and loss graph of retinal OCT dataset.
Figure 22. Accuracy and loss graph of retinal OCT dataset.
Diagnostics 15 02673 g022
Figure 23. Confusion matrix of retinal OCT dataset.
Figure 23. Confusion matrix of retinal OCT dataset.
Diagnostics 15 02673 g023
Figure 24. Grad-CAM heatmap visualization showing important regions (in red) for disease classification across representative datasets: (a) bone fracture multi-region, (b) chest X-ray, and (c) Alzheimer’s disease.
Figure 24. Grad-CAM heatmap visualization showing important regions (in red) for disease classification across representative datasets: (a) bone fracture multi-region, (b) chest X-ray, and (c) Alzheimer’s disease.
Diagnostics 15 02673 g024aDiagnostics 15 02673 g024b
Table 1. Summary of the datasets used for training and evaluation.
Table 1. Summary of the datasets used for training and evaluation.
Dataset NameModalityNo. of ImagesClassesSubcategories
Alzheimer’s Disease [38]MRI14,0004Non-demented, Very Mild,
Mild, Moderate
Breast Ultrasound [39]Ultrasound81482Benign, Malignant
Blood Cells [40]Microscopic21,8418Erythroblast, Neutrophil, Basophil, Eosinophil, Lymphocyte,
Monocyte, Immature Granulocyte, Lymphocyte
Chest X-ray [41,42,43,44,45]X-ray14,3754Normal, COVID-19, Pneumonia, Tuberculosis
Chest CT Scans [46]CT40004Normal, Adenocarcinoma, Large Cell, Squamous Cell
Diabetic Retinopathy [47]Fundus13,0002Diabetic, Non-Diabetic
Kidney Diseases [48,49]Ultrasound11,7564Normal, Cyst, Stone, Tumor
Bone Fracture [50]X-ray92062Fractured, Non-Fractured
Retinal OCT [51]OCT60005Normal, AMRD, CSR, DR, MH
Brain Tumor [52]MRI10,5003Glioma, Meningioma, Pituitary
Table 2. Data augmentation parameters.
Table 2. Data augmentation parameters.
Augmentation TechniqueDescription
RotationRotates the image by fixed angles (90°, 180°, 270°).
Horizontal flipFlips the image along the vertical axis.
Vertical flipFlips the image along the horizontal axis.
Brightness adjustmentRandomly increases or decreases brightness. Factor range: 0.7–1.3.
Contrast adjustmentRandomly enhances or reduces contrast. Factor range: 0.8–1.5.
Table 3. Performance of MediVision model across subclasses of all medical image datasets.
Table 3. Performance of MediVision model across subclasses of all medical image datasets.
DatasetsSubclassesPrecision (%)Recall (%)F1-Score (%)Accuracy (%)
Alzheimer’s diseaseMild Demented96.0099.0097.0098.00
Moderate Demented100.0099.00100.00
Non-Demented98.0098.0098.00
Very Mild Demented97.0095.0096.00
Breast ultrasoundBenign97.0098.0097.0097.45
Malignant98.0097.0097.00
Blood cellBasophil93.0095.0094.0095.45
Eosinophil99.0099.0099.00
Erythroblast98.0094.0096.00
Ig86.0090.0088.00
Lymphocyte97.0096.0097.00
Monocyte94.0093.0093.00
Neutrophil97.0096.0097.00
Platelet100.00100.00100.00
Chest X-rayCOVID1997.0097.0097.0097.31
Normal98.0097.0097.00
Pneumonia97.0098.0097.00
Tuberculosis97.0098.0098.00
Chest CT scansAdenocarcinoma91.0093.0092.0094.83
Large cell carcinoma97.0096.0097.00
Normal98.00100.0099.00
Squamous cell carcinoma93.0091.0092.00
Diabetic retinopathyDR96.0095.0095.0095.43
No DR94.0095.0095.00
Kidney diseasesCyst100.00100.00100.0098.36
Normal98.0096.0097.00
Stone96.0098.0097.00
Tumor100.0099.00100.00
Bone fractureFractured98.0099.0099.0098.56
Not Fractured99.0098.0099.00
Retinal OCTAge-related Macular Degeneration97.00100.0098.0095.44
Central Serous Retinopathy98.0092.0095.00
Diabetic Retinopathy89.0096.0092.00
Macular Hole95.0094.0094.00
Normal98.0096.0097.00
Brain tumorGlioma97.0098.0096.0096.76
Meningioma98.0096.0098.00
Pituitary95.0098.0096.00
Table 4. Accuracy benchmarking of the proposed MediVision model.
Table 4. Accuracy benchmarking of the proposed MediVision model.
DatasetModelAccuracy (%)
Alzheimer’s diseaseVGG1685.13
VGG1984.03
ResNet5075.13
MediVision98.00
Blood cellVGG1680.20
VGG1982.73
ResNet5069.09
MediVision95.45
Breast ultrasoundVGG1696.47
VGG1997.29
ResNet5084.65
MediVision97.45
Chest X-rayVGG1696.00
VGG1996.30
ResNet5092.12
MediVision97.31
Chest CT scansVGG1696.00
VGG1995.67
ResNet5080.83
MediVision96.17
Diabetic retinopathyVGG1695.21
VGG1995.18
ResNet5089.13
MediVision95.43
Kidney diseasesVGG1697.05
VGG1996.54
ResNet5090.08
MediVision98.36
Bone fracture multi-regionVGG1698.42
VGG1998.13
ResNet5093.80
MediVision98.56
Retinal OCTVGG1695.44
VGG1995.22
ResNet5064.56
MediVision95.44
Brain tumorVGG1689.90
VGG1986.79
ResNet5072.32
MediVision96.76
Table 5. Accuracy comparison of MediVision against existing methods across medical image datasets.
Table 5. Accuracy comparison of MediVision against existing methods across medical image datasets.
DatasetReferenceMethodBenchmark (%)Proposed (%)
Alzheimer’s disease         Sethuraman et al. [60]CNN96.6198.00
Breast ultrasoundRaptis et al. [39]ResNet-5093.0597.45
Blood cellAsghar et al. [62]EfficientNet94.7095.45
Chest X-raySingh et al. [63]VGG-1694.7097.31
Chest CT scansMamun et al. [64]DenseNet92.0094.83
Diabetic retinopathyWang et al. [65]InceptionV391.1095.43
Kidney diseasesBhandari et al. [48]MobileNet99.3998.36
Bone fracture multi-regionAldhwani et al. [50]Xception97.3598.56
Retinal OCTDai et al. [66]AlexNet95.0095.44
Brain tumorKhan et al. [67]Hybrid CNN-LSTM95.1096.76
Table 6. Mean accuracy across different models.
Table 6. Mean accuracy across different models.
ModelMean Accuracy (%)
VGG1692.98
VGG1992.79
ResNet5081.17
MediVision96.89
Table 7. Statistical significance analysis of MediVision compared with baseline models.
Table 7. Statistical significance analysis of MediVision compared with baseline models.
ComparisonMean Difference (%)Paired t-Test (p)Wilcoxon (p)Holm Adj. (p)
MediVision vs. VGG16+3.910.05970.00390.0039
MediVision vs. VGG19+4.100.04890.00200.0059
MediVision vs. ResNet50+15.720.00060.00200.0039
Table 8. Overall statistical comparison of all models using the Friedman test.
Table 8. Overall statistical comparison of all models using the Friedman test.
Test χ 2 p-Value
Friedman test27.000.000006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hayat, M.T.; Allawi, Y.M.; Alamro, W.; Sultan, S.M.; Abadleh, A.; Kang, H.; Zreikat, A.I. A Hybrid Convolutional Neural Network–Long Short-Term Memory (CNN–LSTM)–Attention Model Architecture for Precise Medical Image Analysis and Disease Diagnosis. Diagnostics 2025, 15, 2673. https://doi.org/10.3390/diagnostics15212673

AMA Style

Hayat MT, Allawi YM, Alamro W, Sultan SM, Abadleh A, Kang H, Zreikat AI. A Hybrid Convolutional Neural Network–Long Short-Term Memory (CNN–LSTM)–Attention Model Architecture for Precise Medical Image Analysis and Disease Diagnosis. Diagnostics. 2025; 15(21):2673. https://doi.org/10.3390/diagnostics15212673

Chicago/Turabian Style

Hayat, Md. Tanvir, Yazan M. Allawi, Wasan Alamro, Salman Md Sultan, Ahmad Abadleh, Hunseok Kang, and Aymen I. Zreikat. 2025. "A Hybrid Convolutional Neural Network–Long Short-Term Memory (CNN–LSTM)–Attention Model Architecture for Precise Medical Image Analysis and Disease Diagnosis" Diagnostics 15, no. 21: 2673. https://doi.org/10.3390/diagnostics15212673

APA Style

Hayat, M. T., Allawi, Y. M., Alamro, W., Sultan, S. M., Abadleh, A., Kang, H., & Zreikat, A. I. (2025). A Hybrid Convolutional Neural Network–Long Short-Term Memory (CNN–LSTM)–Attention Model Architecture for Precise Medical Image Analysis and Disease Diagnosis. Diagnostics, 15(21), 2673. https://doi.org/10.3390/diagnostics15212673

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop