Next Article in Journal
hiPSCGEM01: A Genome-Scale Metabolic Model for Fibroblast-Derived Human iPSCs
Previous Article in Journal
Computational Study of Tesla Valve Design for Vesico-Amniotic Shunt to Manage Lower Urinary Tract Obstruction and Pleural Effusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AI-RiskX: An Explainable Deep Learning Approach for Identifying At-Risk Patients During Pandemics

by
Nada Zendaoui
1,2,*,
Nardjes Bouchemal
1,3,
Mohamed Rafik Aymene Berkani
4,
Mounira Bouzahzah
1,3,
Saad Harous
5 and
Naila Bouchemal
6
1
Institute of Mathematics and Computer Science, Abdelhafid Boussouf University Center of Mila, Mila 43000, Algeria
2
LIRE Laboratory of Constantine 2, Constantine 25000, Algeria
3
LISI Laboratory of Intelligent Systems and Informatics, Mila 43000, Algeria
4
LSEA Research Laboratory in Advanced Electronics Systems, University Yahia Fares of Medea, Medea 26000, Algeria
5
Department of Computer Science, College of Computing and Informatics, University of Sharjah, Sharjah P.O. Box 27272, United Arab Emirates
6
LyRIDS ECE Ecole d’Ingénieurs de Paris, 75015 Paris, France
*
Author to whom correspondence should be addressed.
Bioengineering 2025, 12(10), 1127; https://doi.org/10.3390/bioengineering12101127
Submission received: 22 September 2025 / Revised: 16 October 2025 / Accepted: 17 October 2025 / Published: 21 October 2025
(This article belongs to the Section Biosignal Processing)

Abstract

Pandemics place extraordinary pressure on healthcare systems, particularly in identifying and prioritizing high-risk groups such as the elderly, pregnant women, and individuals with chronic diseases. Existing Artificial Intelligence models often fall short, focusing on single diseases, lacking interpretability, and overlooking patient-specific vulnerabilities. To address these gaps, we propose an Explainable Deep Learning approach for identifying at-risk patients during pandemics (AI-RiskX). AI-RiskX performs risk classification of patients diagnosed with COVID-19 or related infections to support timely intervention and resource allocation. Unlike previous models, AI-RiskX integrates five public datasets (asthma, diabetes, heart, kidney, and thyroid), employs the Synthetic Minority Over-sampling Technique (SMOTE) for class balancing, and uses a hybrid convolutional neural network–long short-term memory model (CNN–LSTM) for robust disease classification. SHAP (SHapley Additive exPlanations) enables both individual and population-level interpretability, while a post-prediction rule-based module stratifies patients by age and pregnancy status. Achieving 98.78% accuracy, AI-RiskX offers a scalable, interpretable solution for equitable classification and decision support in public health emergencies.

Graphical Abstract

1. Introduction

The COVID-19 pandemic has exposed critical weaknesses in healthcare systems, particularly in recognizing and prioritizing those at high risk. The elderly (≥60 years) showed the highest vulnerability, with severe outcomes reaching 85% [1,2], while children experienced up to 30% severe cases and 28% of Intensive Care Unit (ICU) admissions [3,4]. Pregnant women and patients with chronic conditions such as heart disease (52%) and kidney disease (48%) are also disproportionately affected [5], as summarized in Table 1. These risks stem from weakened immunity, underlying conditions, or physiological changes that reduce resilience to infection. Addressing such challenges requires intelligent, timely, and equitable risk stratification tools to support data-driven decision-making and optimize resource allocation during pandemics.
Artificial Intelligence has emerged as a valuable tool for medical classification and diagnostics; however, current models still exhibit notable limitations. First, most studies focus narrowly on a single disease [6,7], rarely accounting for coexisting health conditions or patient demographics [8]. Second, many existing models operate as “black boxes,” offering limited interpretability and lacking transparency in their decision-making processes [9,10]. Third, few systems address demographic-specific vulnerabilities—such as age or pregnancy—or provide interfaces suitable for real-time clinical deployment [9,10]. This lack of flexibility, transparency, and consideration of patient diversity hinders trust and the effective implementation of AI tools in real-world healthcare settings.
To address these multifaceted challenges, we propose AI-RiskX, a novel and interpretable deep learning framework for comprehensive multidisease risk assessment and stratification, supporting clinical decisions for patients with COVID-19 or related infections. First, AI-RiskX expands the modeling scope beyond single diseases by integrating five well-curated public datasets—covering heart disease, diabetes, asthma, chronic kidney disease, and thyroid disorders—into a unified learning framework. The datasets are harmonized and globally balanced using the Synthetic Minority Over-sampling Technique (SMOTE) [11,12], ensuring unbiased learning and meaningful stratification across diverse conditions. Second, the framework enhances interpretability through explainable AI components that reveal how individual clinical features influence predictions. This allows clinicians to trace and understand the reasoning behind model outputs, fostering transparency and trust in AI-driven decision-making. Third, AI-RiskX incorporates demographic- and context-aware mechanisms that reflect real patient diversity considering factors such as age, comorbidities, or pregnancy and its modular design supports smooth integration into real-time clinical workflows.
AI-RiskX combines several innovations within a unified architecture: (i) a harmonized multidisease dataset supporting unbiased training; (ii) a hybrid convolutional neural network–long short-term memory (CNN–LSTM) backbone for robust multi-label prediction; (iii) a demographic-aware rule-based module for refined classification; and (iv) SHapley Additive exPlanations (SHAP) for local and global interpretability, dynamically adapting predictions with crucial demographic factors such as age and pregnancy status.
The core contributions of this study are as follows:
  • Unified dataset: Integrating five well-curated public datasets covering heart disease, diabetes, asthma, chronic kidney disease, and thyroid disorders (Table 1), AI-RiskX focuses on clinically actionable features such as blood tests, vital signs, and comorbidities. These conditions were selected for their high prevalence, frequent co-occurrence in vulnerable populations, and links to adverse pandemic outcomes. The harmonized dataset mitigates class imbalance via SMOTE and is supported by epidemiological evidence and WHO guidance [13], ensuring robust, generalizable multidisease learning [14].
  • Hybrid CNN–LSTM model: Employing a parallel convolutional–recurrent design that captures both spatial patterns and temporal dependencies in structured clinical data, thereby enhancing prediction robustness and accuracy over conventional pipelines.
  • Demographic-aware rule-based: Incorporating a lightweight, rule-based classification module that integrates clinically significant demographic factors—such as advanced age (≥60 years) and pregnancy status, as outlined in Table 1—to refine patient classification into three risk categories: high risk, at risk, and not at risk.
  • Explainability: Embedding SHapley Additive exPlanations (SHAP) to deliver transparent local and global interpretability, enabling clinicians to understand the contribution of each feature to model predictions and foster trust in AI-assisted decision-making.
The remainder of this paper is structured as follows: Section 2 reviews related studies, Section 3 details the proposed methodology, Section 4 presents the results and analysis, and Section 5 concludes with key findings and future directions.

2. Related Work

Artificial Intelligence (AI) has significantly advanced healthcare by enabling early diagnosis, risk stratification, and clinical decision-making support. Despite these advancements, most AI models struggle with real-world applicability owing to their narrow disease focus, limited interpretability, and poor sensitivity to patient diversity.
Our study focuses on risk classification for patients already diagnosed with COVID-19. The proposed AI-RiskX framework evaluates disease severity and multimorbidity risk profiles by leveraging structured clinical data, comorbidities, and demographic information. Unlike models developed for infection detection [15,16], AI-RiskX emphasizes patient stratification and clinical decision support. Previous research highlights two major limitations in existing approaches: (1) a focus on single-disease classification using fragmented datasets, and (2) deep learning architectures with limited interpretability. These shortcomings motivated the development of AI-RiskX, a unified and explainable system designed to address the complex challenges of pandemic healthcare.

2.1. Single-Disease Focus and Dataset Fragmentation

Most existing AI models in healthcare predict a single disease using isolated datasets, limiting their applicability in real clinical settings with multiple chronic conditions or overlapping symptoms. We selected studies that used machine learning or deep learning for disease prediction, incorporated explainability methods such as SHAP or Local Interpretable Model-Agnostic Explanations (LIME), and relied on structured, reproducible datasets; studies lacking interpretability or based solely on imaging or proprietary data were excluded to align with AI-RiskX.
El-Sofany et al. [17] built a diabetes prediction system using XGBoost and SMOTE, achieving an accuracy of 97.4%. However, their approach was confined to one disease and lacked integration of explainability or demographic profiling. Similarly, Guleria et al. [18] developed a cardiovascular disease model using SVM with SHAP and LIME but trained it on a small dataset, missing deep learning capabilities.
Yilmaz et al. [19] applied Random Forest, Logistic Regression, and SVM to predict coronary heart disease and achieved 92.9% accuracy. However, their model did not incorporate interpretability tools or address the needs of the vulnerable subpopulations. Prathibha et al. [20] and Zhang et al. [21] implemented CNNs such as ResNet and Xception for thyroid disorder detection using ultrasound images with high accuracy (97%); however, their models relied solely on imaging data and lacked multimodal integration.
In the context of asthma detection, Awal et al. [22] introduced a Bayesian Optimization-based model (BOMLA), whereas Heris et al. [23] utilized KNN with Relief-F. Both faced challenges owing to small datasets and limited generalizability.
During the COVID-19 pandemic, Bottrighi et al. [24] and Quiroz et al. [25] developed models focused on deterioration prediction using rule-based and neural network methods, respectively. Although beneficial, these models were not adaptable beyond COVID-19-specific data.
Ahmed et al. [26] applied genomic data and statistical modeling for cardiovascular risk prediction but omitted critical clinical features such as comorbidities, symptoms, or demographics, limiting their real-world clinical value.
Most recent studies continue to focus on single-disease modeling. For example, Talukder et al. [27] proposed HXAI-ML, a hybrid explainable model for cardiovascular heart disease combining SHAP, LIME, and permutation importance, achieving over 98% accuracy on the Cleveland and Framingham datasets. Similarly, Bilal et al. [28] introduced an XAI-driven system for cardiovascular forecasting using large-scale EMR data (≈308,000 samples), enhancing transparency but remaining disease-specific. In the respiratory domain, Mahmood et al. [29] applied AutoML with explainable AI for asthma outcome prediction, while Lyu et al. [30] classified asthma severity using vocal biomarkers and SHAP analysis. Despite their robustness, these models remain limited to single-disease contexts without integration across comorbidities or demographic factors.
Collectively, these studies reveal a fragmented landscape: models are overly disease-specific, datasets are isolated, and key clinical variables, such as comorbidities and demographics, are frequently overlooked. This fragmentation underscores the necessity of using unified multiple diseases data.

2.2. Deep Learning and Explainability

Deep learning (DL), particularly Convolutional Neural Networks (CNNs), have shown strong performance in healthcare for image classification and structured data analysis. Their ability to extract complex patterns makes them ideal for use in diagnostic tasks. For example, Srinivasu et al. [31] developed a CNN–Bi-LSTM model for diabetes monitoring with 96.44% accuracy, and Prathibha et al. [20] utilized a ResNet model to classify thyroid conditions from ultrasound images with 97% accuracy.
Although AI models have achieved remarkable predictive accuracy, many still operate as “black boxes,” offering limited transparency regarding how decisions are made. Recent advances in explainable AI (XAI) have sought to address this issue, particularly in COVID-19 prognosis studies. For instance, Chadaga et al. [32] and Ozawa et al. [33] proposed interpretable models to enhance clinical trust and understanding.
Post hoc explanation techniques such as SHAP (SHapley Additive exPlanations) and LIME (Local Interpretable Model-Agnostic Explanations) have further advanced model interpretability. SHAP, in particular, offers a strong theoretical foundation for both local and global feature attribution [34]. For example, Duckworth et al. [35] used SHAP to monitor data drift in emergency care, and Zhao et al. [36] applied it to immune disorder prediction. Similarly, Tasin et al. [37] compared LIME and SHAP in diabetes modeling and found SHAP to provide superior clinical interpretability.

2.3. Addressing Gaps with AI-RiskX

To address these persistent limitations, we introduce AI-RiskX, an explainable multidisease prediction approach tailored for real-world healthcare and pandemic response. First, to overcome the single-disease focus of most previous studies, AI-RiskX supports risk prediction across five prevalent chronic diseases, diabetes, heart disease, asthma, kidney disease, and thyroid disorders, using a harmonized and balanced dataset. Second, to capture both spatial and temporal dependencies in clinical data, a hybrid CNN–LSTM architecture is employed, with SMOTE-based preprocessing applied to mitigate class imbalance. Third, to resolve the lack of demographic awareness and improve usability, a rule-based module profiles vulnerable populations (the elderly, pregnant women, and children). Finally, to address the opacity of deep models, AI-RiskX embeds SHAP within the approach, providing transparent case-level and population-level explanations that enhance clinical interpretability. A comparative overview of these prior models and the distinguishing features of AI-RiskX is summarized in Table 2.

3. Methodology

This section outlines the methodology of the proposed AI-RiskX approach, an explainable system developed to support the early detection and risk classification of vulnerable patients during pandemics. Building upon our previous work on AI-LMS (AI-Based Long-Term Monitoring) [38], which focused on continuous patient monitoring using IoMT and machine learning, AI-RiskX is introduced as a dedicated subsystem specifically designed for risk stratification. In AI-LMS, IoMT sensors, mobile cloud computing, and AI models were combined to enable long-term monitoring and alert generation for at-risk patients during pandemics such as COVID-19. AI-RiskX refines this framework by isolating the risk evaluation phase, focusing solely on the identification and classification of patients according to their vulnerability level based on comorbidities, demographic attributes, and clinical indicators. This targeted component serves as the decision-support core of AI-LMS, providing explainable, data-driven risk assessment before continuous monitoring takes place.
AI-RiskX is designed as a four-phase approach that integrates data processing, disease prediction, demographic-aware evaluation, and explainability to deliver clinically meaningful insights.
Phase I: Data Processing. Heterogeneous datasets covering five major chronic conditions ( diabetes, heart disease, asthma, chronic kidney disease, and thyroid disorders) were harmonized into a unified multidisease dataset. Unlike prior studies that relied on single-disease data, this unified construction process improves generalizability across multiple conditions. Standard preprocessing, including normalization and SMOTE, was applied to ensure balanced class distributions.
Phase II: Patient Risk Assessment Model. The core of this phase is a hybrid CNN–LSTM architecture designed to extract multiscale spatial features (capturing correlations among clinical variables) and long-term temporal dependencies (reflecting sequential relationships in patient profiles), generating robust representations for classification across multiple disease classes. By combining spatial and temporal learning, this phase enhances the predictive performance of structured health data compared to conventional single-stream models.
Phase III: Demographic-Aware Risk Evaluation. A rule-based module identifies vulnerable groups, such as the elderly and pregnant women, and integrates this demographic information with disease predictions to stratify patients into three categories: high risk, at risk, and not at risk. This integration ensures fairness, ethical decision support, and targeted interventions, addressing the limitations of prior models that largely overlook demographic factors.
Phase IV: Explainability. To address the opacity of deep learning models, Explainable Artificial Intelligence (XAI) techniques, specifically SHAP, are embedded to provide transparent explanations at both individual and population levels. By linking prediction outcomes to clinical features, this phase strengthens interpretability, trust, and clinical acceptance, distinguishing AI-RiskX from the traditional black-box systems.
As the identification core of the AI-LMS approach, AI-RiskX refines the early-stage decision-making component by offering an explainable, data-driven evaluation of high-risk patients across multiple chronic conditions. The complete workflow is illustrated in Figure 1 The following subsections describe each component in detail.

3.1. Data Processing

This section outlines the data foundation of the proposed AI-RiskX system. It begins with a description of the datasets used for model training and evaluation, followed by the preprocessing and harmonization steps implemented to standardize, balance, and prepare the data for deep learning analysis. These processes ensure input consistency and robustness across heterogeneous medical sources.

3.1.1. Dataset Description

Unlike earlier studies focusing on single-disease prediction, our work leverages five distinct medical datasets to train and evaluate the AI-RiskX framework. These datasets cover heart disease, diabetes, thyroid disorders (hypo/hyperthyroidism), asthma, and chronic kidney disease (CKD). Each includes both demographic and clinical attributes relevant to disease-specific diagnostics, providing a comprehensive basis for patient-level risk assessment. For the purpose of this study, we assume that all patients are affected by a pandemic scenario (e.g., COVID-19), so the model assesses vulnerability based on pre-existing conditions.
The UCI Heart Disease dataset [39] provides structured cardiac health records and is commonly used for binary classification of heart conditions. The HealthCare Diabetes dataset [40] comprises 768 patient instances with standard clinical features and binary outcome labels. Thyroid dysfunction detection is based on the Hypo/Hyperthyroidism dataset [41], containing mixed numerical and categorical features for precise classification. The Asthma Diagnosis dataset [42] includes anonymized patient-level records with diverse attributes suitable for supervised asthma identification. Finally, CKD prediction employs the Chronic Kidney Disease dataset [43], consisting of 400 labeled samples for disease presence or absence.
Together, these heterogeneous data sources form a unified multidisease foundation for robust prediction and risk stratification. The preprocessing and harmonization pipeline described below integrates these datasets into a cohesive, balanced, and clinically interpretable dataset—AI-RiskX.

3.1.2. Data Preprocessing and Harmonization

Normalization and harmonization were essential to ensure structural, semantic, and statistical consistency across the five heterogeneous datasets. The complete workflow is summarized in Figure 2.
To integrate datasets with distinct structures and variable names, we developed a harmonization pipeline comprising the following steps to ensure structural and semantic consistency across the five datasets:
  • Column Cleaning: Non-informative or administrative attributes such as PatientID, DoctorInCharge, and EducationLevel were removed to eliminate redundancy and potential privacy risks.
  • Schema Alignment: Attribute names were standardized across datasets (e.g., Gender → sex, Diagnosis → Target) to ensure consistency and prevent schema conflicts during merging.
  • Structural Validation: Each dataset was verified to contain the same number of features (79) with consistent data types (numeric, categorical, or binary), guaranteeing interoperability and preventing mismatches.
  • Label Mapping: Diagnostic labels were harmonized into a unified multi-class structure: healthy (0), asthma (1), diabetes (2), heart disease (3), chronic kidney disease (4), and thyroid disorder (5). This unified labeling enables multidisease classification within a single framework.
  • Handling Missing Features: Some features were missing in certain datasets. For each missing value, we first checked if a similar or proxy feature existed that could represent the same clinical concept; if found, it was mapped to the unified feature. If no comparable feature was available, simple imputation was applied: mean imputation for continuous features and mode imputation for categorical features. This ensured that all 79 features were consistently present across datasets, enabling robust multidisease modeling. This approach preserves feature completeness, maintains structural consistency, and is computationally efficient.
The harmonized dataset comprises 26,125 samples and 79 unified features, reshaped into a two-dimensional structure of size ( 79 , 1 ) , where each instance corresponds to a patient record and its target class. Table 3 summarizes the dataset composition, while Table 4 categorizes features into thematic groups spanning demographics, lifestyle, and disease-specific indicators.
Following harmonization, the five disease datasets were merged to form the unified AI-RiskX dataset. Due to strong class imbalance—particularly for asthma with only 124 samples—the Synthetic Minority Oversampling Technique (SMOTE) was employed to balance the classes. Rather than oversampling each dataset separately, SMOTE was applied globally to the integrated dataset, ensuring uniform class distribution and improving model generalizability.
SMOTE was applied only to training data to avoid leakage. For small datasets (e.g., asthma), oversampling was limited, and statistical checks confirmed consistency with original data. Harmonization kept clinically relevant shared features for coherent multidisease learning. Each new sample x new is created based on an existing minority instance x and one of its k nearest neighbors x nn using a random scalar λ [ 0 , 1 ] , as defined in Equation (1):
x new = x + λ · ( x nn x )
The Euclidean distance between samples, computed in Equation (2), is used to identify nearest neighbors:
d ( x i , x j ) = m = 1 M ( x i , m x j , m ) 2
SMOTE minimizes the total distance across minority instances as shown in Equation (3):
min i = 1 N m j = 1 k d ( x i , x i , j )
Table 5 presents the class distribution before and after balancing, demonstrating how SMOTE corrects extreme disparities among disease categories.
For normalization, Min–Max scaling was applied to rescale all features within the [ 0 , 1 ] range:
x N = x min ( x ) max ( x ) min ( x )
This transformation facilitates faster convergence during training and reduces numerical instability. After normalization, the dataset was reshaped into a two-dimensional structure suitable for CNN-based feature extraction.
For dataset partitioning, the normalized data were randomly shuffled and divided into three distinct subsets for model development and evaluation. Specifically, 70% of the data were used for training to enable model learning and parameter optimization, 20% were allocated for validation to fine-tune hyperparameters and monitor performance during training, and the remaining 10% were reserved as an independent testing set to provide an unbiased assessment of the model’s generalization capability.

3.2. Patient Risk Assessment Model

The proposed AI-RiskX framework was developed for accurate multidisease classification and interpretable patient risk assessment using structured clinical data such as vital signs, laboratory results, and diagnostic indicators. At its core, the model employs a hybrid CNN–LSTM architecture designed to capture both localized feature interactions and long-range dependencies within patient profiles.
Before introducing the deep learning architectures, three classical machine learning baselines were implemented to establish performance references on structured clinical data:
  • Logistic Regression: A simple linear classifier was implemented as a baseline. Input features were standardized using StandardScaler, and the model was trained with L2 regularization (C = 1.0) using the lbfgs solver for up to 1000 iterations. Stronger regularization slightly decreased recall, but this configuration provided a stable and interpretable benchmark.
  • Random Forest: A Random Forest ensemble with 200 trees (n_estimators = 200) was trained using Gini impurity and balanced class weights to address class imbalance. Bootstrap aggregation (bootstrap = True) improved robustness, and hyperparameters were optimized via 5-fold cross-validation to maximize the macro F1-score. This model captures nonlinear feature interactions and provides interpretable variable importance.
  • XGBoost: An XGBoost classifier was configured with learning_rate = 0.1, max_depth = 6, and n_estimators = 200. The model used early stopping (patience = 10) with eval_metric = ‘logloss’, L2 regularization (lambda = 1), and subsampling (subsample = 0.8) to improve generalization. Combined with SMOTE-based oversampling, XGBoost effectively handled complex feature interactions and class imbalance.
Building upon these baselines, the proposed CNN–LSTM architecture was benchmarked against representative deep learning models commonly used in medical AI, including CNN [44], LSTM [45], BiLSTM [46], and a transformer-based LLM–BERT model [47]:
  • Convolutional Baseline (CNN): A single-stream 1D CNN was implemented to capture spatial dependencies among the 79 clinical features. The network consisted of two Conv1D layers (64 filters; kernel size = 3), followed by BatchNormalization, MaxPooling1D, and two dense layers (64 and 32 neurons) with ReLU activation. Dropout (0.4) was applied for regularization. The model was trained using the Adam optimizer (lr = 0.001), batch size = 32, and 100 epochs with early stopping (patience = 10). This baseline evaluates the effectiveness of purely convolutional feature extraction without sequential modeling.
  • Recurrent Baselines (LSTM and BiLSTM): To model sequential dependencies among input features, LSTM-based architectures were evaluated. The standard LSTM network employed two stacked layers (64 and 32 units) with dropout (0.3 and 0.4) to prevent overfitting, trained for up to 120 epochs using Adam (lr = 0.001) and early stopping (patience = 10). The bidirectional LSTM (BiLSTM) [46] extended this by processing sequences in both forward and backward directions to capture bidirectional dependencies. It comprised two BiLSTM layers (128 units each) with dropout rates of 0.2 and 0.4, followed by a dense softmax classifier. Training continued for 150 epochs with early stopping and L2 regularization (lambda = 0.001) for better generalization. These recurrent architectures assess the ability to learn temporal and contextual feature relationships.
  • Transformer-Based Baseline (LLM–BERT): To investigate the applicability of large language models to structured clinical data, an adapted BERT model was fine-tuned for six-class classification (five diseases and one healthy control). Each patient record was transformed into a text-like representation of feature–value pairs (e.g., “Age: 45; Gender: Male; BloodPressure: High; Glucose: Normal; …”), allowing the tokenizer to embed clinical variables contextually. The tokenized inputs were processed by a pre-trained BERT model [47], fine-tuned for 10 epochs using AdamW (lr = 2 × 10−5) with linear warmup and dropout (0.1) on the classifier head. This model serves as a proof of concept for adapting LLMs to tabular medical data.
  • Proposed CNN–LSTM Hybrid Architecture: The core AI-RiskX model integrates Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM) networks in a parallel hybrid configuration to exploit both spatial and sequential dependencies in clinical features. CNNs efficiently capture localized correlations and hierarchical feature patterns, while LSTMs model sequential dependencies that may reflect implicit clinical relationships among patient variables [44,45,48].
The architecture comprises three parallel CNN branches (kernel sizes 3, 5, and 7) and one LSTM branch. The convolutional branches each include two Conv1D layers (64 filters; ReLU activation), BatchNormalization, and MaxPooling1D (pool size = 2). Kernel sizes of 3, 5, and 7 enable multi-resolution feature extraction. The LSTM branch consists of two stacked LSTM layers (64 and 32 units) with batch normalization to ensure stable learning. The outputs from all branches are flattened and concatenated, followed by two dense layers (64 and 32 neurons; ReLU activation) and a dropout layer (rate = 0.3). The final softmax layer produces the probability distribution across six classes—five chronic diseases and one healthy control. A detailed layer-wise summary of this architecture is presented in Table 6, and its structural design is illustrated in Figure 3.
Model training used categorical cross-entropy loss with a 70–20–10 train–validation–test split. All data were normalized, and SMOTE-based balancing was applied only to training sets. A stratified 5-fold cross-validation procedure further evaluated generalization, ensuring balanced class representation across folds. Each fold was trained for 30 epochs with early stopping, and performance was averaged across folds. Evaluation metrics included accuracy, precision, recall, F1-score, and AUC, as detailed in Section 4.
Overall, the hybrid CNN–LSTM architecture demonstrated superior robustness and interpretability by combining spatial and sequential learning in a unified framework, enabling AI-RiskX to model complex clinical dependencies and support patient risk classification.

3.3. Demographic-Aware Risk Evaluation

The AI-RiskX approach includes a post-classification layer for demographic-aware risk profiling. This layer performs three functions—age group classification, pregnancy status identification, and composite patient risk assessment while operating alongside the disease prediction model without affecting its training or classification. The overall logic is summarized in Algorithm 1.
Patients are first categorized into clinically relevant age groups to capture age-related vulnerability: children (<18 years), adults (18–59 years), and elderly (≥60 years). Modeling age as a categorical attribute introduces medically interpretable logic that complements numerical features and enhances risk contextualization.
Pregnancy status is included as a binary attribute to account for gender-specific vulnerabilities, ensuring that this critical demographic group is appropriately considered.
Finally, the module combines disease predictions with demographic indicators to produce three risk levels: high risk (diseased and vulnerable), at risk (diseased but not vulnerable), and not at risk (no disease and not vulnerable).
Algorithm 1 Check Patient Risk
  • Input: d a t a , m o d e l
  • Output: Patient risk assessment
  • Extract a g e _ g r o u p and p r e g n a n c y _ s t a t u s from d a t a
  • Predict disease using m o d e l and select the category with the highest probability
  • Define high-risk criteria: a g e _ g r o u p (child or elderly) or p r e g n a n c y _ s t a t u s (pregnant)
  • If predicted disease ≠ Healthy:
    (a)
    If patient meets high-risk criteria: HIGH RISK: predicted disease and risk factors
    (b)
    Else: AT RISK: predicted disease
  • ElseIf patient meets high-risk criteria: AT RISK due to high-risk condition
  • Else: NOT at Risk

3.4. Explainability

AI-RiskX uses SHAP (SHapley Additive exPlanations) to interpret model predictions. SHAP provides global and local explanations based on cooperative game theory, ensuring local accuracy and capturing feature interactions. Recent studies have demonstrated SHAP’s effectiveness in highlighting clinically relevant predictors, such as age and disease stage in cancer prognosis [49], and its robustness when applied to deep learning models in healthcare [50]. These findings support SHAP’s reliability and reproducibility in clinical environments [51].
The Shapley value for a model f and input x R M with M features is defined as follows [34]:
ϕ i ( f , x ) = S F { i } | S | ! ( M | S | 1 ) ! M ! f S { i } ( x S { i } ) f S ( x S )
where F is the set of all features, S is a subset excluding i, and f S ( x S ) is the expected model output using only features in S. Key properties— additivity, local accuracy, and consistency—ensure reliable feature attribution. Approximation techniques such as Deep SHAP are used for computational efficiency in large networks [34].
Within AI-RiskX, SHAP is applied post hoc to the fused CNN–LSTM feature representation before the final classification layer, capturing both spatial and sequential dependencies. SHAP interprets structured clinical features, while the demographic-aware rule-based module applies explicit clinical logic (e.g., age and pregnancy), providing complementary interpretability through quantitative and rule-based transparency.

4. Results

This section presents the experimental evaluation of the proposed AI-RiskX system, its comparative performance against both classical machine learning and deep learning baselines, and details of its practical implementation.

4.1. Overall Performance Evaluation

To evaluate the necessity and effectiveness of deep learning techniques, several classical machine learning models (Logistic Regression, Random Forest, and XGBoost) were first trained and tested on both the combined and combined–balanced datasets. Their performance results are summarized in Table 7.
On the combined dataset, Logistic Regression achieved an accuracy of 89.95%, while Random Forest and XGBoost reached 96.77% and 96.99%, respectively. When trained on the combined–balanced dataset, the performance of Logistic Regression decreased to 70.68%, whereas Random Forest and XGBoost maintained strong results with accuracies of 96.84% and 96.98%. These findings suggest that although traditional machine learning algorithms can perform competitively, they struggle to capture the complex, nonlinear relationships and multidisease dependencies inherent in heterogeneous healthcare data—especially under class imbalance conditions.
To further assess model robustness, the proposed hybrid CNN–LSTM model was compared with various deep learning architectures, including CNN, LSTM, BiLSTM, and LLM-BERT, as also presented in Table 7. On the combined–balanced dataset, the hybrid CNN–LSTM achieved the highest accuracy of 98.78%, outperforming CNN (98.67%), LSTM (96.97%), BiLSTM (96.13%), and LLM-BERT (86.02%). Similarly, on the combined dataset, CNN–LSTM maintained superior performance with 94.55% accuracy.
When evaluated on individual disease datasets, BiLSTM exhibited the best performance for heart disease, diabetes, HHT, and CKD, while LSTM achieved the highest accuracy for AD. These results demonstrate that integrating CNN’s spatial feature extraction with LSTM’s temporal learning capabilities significantly enhances the system’s ability to model inter-disease relationships and capture complex clinical patterns, leading to improved generalization across multiple health conditions.

4.2. Ablation Study

To better understand the contribution of each component within the proposed architecture, we conducted an ablation study on the balanced dataset. The model variants were as follows: the CNN-only model, which serves as a spatial feature extractor without sequential modeling; the LSTM-only model, which functions as a sequential model without convolutional representation; and the CNN–LSTM (proposed) model, which combines both convolutional and temporal layers.
As shown in Table 8, excluding either the CNN or LSTM component resulted in reduced predictive performance. The CNN-only model achieved an accuracy of 98.67%, while the LSTM-only model reached 96.97%. In contrast, the full hybrid CNN–LSTM achieved the best overall performance, with 98.78% accuracy, 98.79% precision, 98.77% recall, and an F1-score of 98.77%.
These results highlight the complementary strengths of both modules: CNN layers effectively capture spatial correlations among clinical features, while LSTM layers model sequential dependencies across patients’ health records. Together, they provide a balanced representation that enhances the accuracy and robustness of multidisease risk classification across heterogeneous clinical datasets.

4.3. Cross-Validation Performance

The results of the stratified 5-fold cross-validation are summarized in Table 9. These results represent the average model performance and its variability across the five folds. The proposed hybrid CNN–LSTM model demonstrated highly consistent outcomes, achieving an average accuracy of 98.72% and an AUC of 99.52%. The low standard deviations observed across all metrics indicate that the model performs stably and is not overly dependent on a particular data partition.
To further validate the robustness of the proposed AI-RiskX model, a statistical significance analysis was conducted using the approximate randomization test [52] and non-parametric evaluation procedures [53,54], following standard practices in machine learning assessment. Considerations for mitigating evaluation biases in imbalanced datasets were also incorporated [55]. The reliability of the performance metrics was assessed using paired t-tests across the five stratified folds, with 95% confidence intervals for each metric M computed as follows:
C I 95 % = M ¯ ± t α / 2 , k 1 × s k
where M ¯ is the mean performance, s is the standard deviation, and k is the number of folds. For instance, with an accuracy of 98.72 ± 0.11 , the corresponding 95% confidence interval is [ 98.58 , 98.86 ] .
To assess whether the performance improvements were statistically significant, paired t-tests were conducted between the hybrid CNN–LSTM and the baseline models, using
t = d ¯ s d / k
where d i denotes the difference in performance between the CNN–LSTM and each baseline model for fold i. All obtained p-values were below 0.05, confirming that the observed gains are statistically significant and unlikely to result from random variation.

4.4. ROC Curve

Receiver Operating Characteristic (ROC) curves and Area Under the Curve (AUC) were used to evaluate the overall effectiveness of the proposed model. The ROC curve plots the True Positive Rate (TPR) against the False Positive Rate (FPR) across various thresholds, thereby providing a comprehensive visualization of the model’s performance. Figure 4 presents the ROC curves of the proposed model corresponding to different disease categories, demonstrating excellent classification performance in all categories, with curves close to the upper right corner. The AUC for each class was at an optimal rate of 1.00, indicating perfect accuracy and an exceptional discriminative capability. These results validate the robustness of the proposed model for disease detection.

4.5. Confusion Matrix

Figure 5 illustrates the confusion matrix (CM) of the proposed model. It provides a detailed evaluation of the classification performance for various diseases. The results of the CM validation (Figure 5a) demonstrate a high classification accuracy, as indicated by the strong diagonal in the CM. Most instances were correctly classified; however, minor misclassifications were observed. Specifically, 3% of patients without disease were misclassified as having thyroid disease, while 1% of diabetes cases were incorrectly labeled as no disease. Similarly, 1% of thyroid cases were misclassified as no disease. These misclassifications likely arise from the overlapping feature representations between no disease and thyroid disease, as well as between no disease and diabetes. These results confirm the robustness of the proposed CNN-LSTM model for the accurate detection of multiple diseases in patients. The high classification performance, with diagonal performance ranging from 97% to 100%, highlights the reliability of the model, whereas the relatively minor off-diagonal misclassifications indicate potential areas for further refinement and optimization.
As shown in Figure 5b, the CM for the testing and generalization phase confirms the validation outcomes. Specifically, the model accurately identified most diseases, exhibiting a performance range between 96% and 100%, with only minor misclassifications quantified as follows: A total of 4% of instances from no disease were misclassified specifically, 1% as diabetes and 3% as thyroid. Additionally, 1% of the diabetes cases were misidentified as having no disease. Furthermore, 1% of thyroid instances were incorrectly categorized as no disease. These misclassifications arise from similarities and variability among the classes. These results demonstrate the potential of the AI-RiskX system for health monitoring and disease detection.

4.6. Explainability with SHAP

Model transparency is essential for clinical adoption in high-stakes healthcare AI systems. To address this, SHAP (SHapley Additive Explanations) was integrated into the AI-RiskX pipeline as a post hoc interpretability tool. After training the complete CNN–LSTM model, SHAP was applied to the fused feature vector and output of the concatenated convolutional and LSTM branches—prior to the dense classification layers. This enables the interpretation of the joint feature representation, thereby capturing the combined influence of both local and sequential dependencies.
SHAP was applied in primary data configurations:
  • Individual disease datasets: SHAP was used to interpret predictions from the AI-RiskX model trained and evaluated separately for each disease-specific dataset. This helped to identify the key features associated with each condition.
  • Combined multidisease dataset: SHAP was applied to the fused representation in the unified model trained on the merged dataset, to assess whether clinically relevant signals were preserved across diseases despite dataset integration.
SHAP provided both global and local explanations:
  • Global SHAP values identify the most influential features across the entire dataset.
  • Local SHAP values explain individual predictions by quantifying the contribution of each feature to a specific case.
In disease-specific settings, SHAP values highlighted distinct clinical markers aligned with established medical knowledge:
  • Asthma: Respiratory indicators such as white blood cell count and lung function (FVC) emerged as the top contributors (Figure 6a and Figure 7a).
  • Diabetes: Metabolic and hereditary indicators such as glucose, BMI, and diabetes pedigree function dominated the predictions (Figure 6b and Figure 7b).
  • Heart Disease: Maximum heart rate (thalach), chest pain type, and slope were key predictors (Figure 6c and Figure 7c).
  • Chronic Kidney Disease: Serum creatinine and potassium stood out as critical features (Figure 6d and Figure 7d).
  • Thyroid Disorders: Hormonal markers such as TSH and T4U were consistently dominant (Figure 6e and Figure 7e).
When applied to the unified AI-RiskX model, SHAP revealed that the key features identified in individual disease analyses remained highly influential. As shown in Figure 6f and Figure 7f, variables such as TSH, BMI, white blood cell count, and dust exposure retained high contribution scores, confirming their consistent relevance across both the individual and integrated contexts.
This comparative explainability analysis confirms that AI-RiskX preserves medically meaningful feature relationships, even in a multidisease setting. Contrary to concerns that data fusion might blur condition-specific signals, our findings demonstrate that the model maintains and, in some cases, enhances the discriminative power of critical clinical indicators. By applying SHAP post hoc to the fused representation, AI-RiskX delivers both high predictive performance and clinically interpretable insights. This not only improves trust and accountability but also supports robust, transparent clinical decision-making across diverse conditions.

4.7. Implementation

To enhance the clinical usability and real-world deployment, the AI-RiskX approach was implemented through a web-based interactive interface developed using the Gradio library [56]. This interface enables seamless interaction between healthcare professionals and the predictive model, offering an intuitive environment for patient data entry, automated disease risk prediction, and real-time interpretability.
Unlike conventional black-box AI systems, AI-RiskX integrates interpretable deep learning with clinically informed rule-based logic to ensure transparency, trust, and alignment with healthcare decision-making processes.
The interface is structured into four main functional components:

4.7.1. Data Entry Form

The interface begins with a structured, user-friendly data entry panel that allows clinicians to input patient-specific demographic and clinical information for multidisease risk assessment. Standardized fields are provided for essential attributes such as age (via slider), sex (radio buttons), pregnancy status, and ethnicity. Disease-specific parameters, such as blood pressure, glucose levels, and organ-specific indicators, can be entered through dedicated tabs tailored to cardiac, diabetes, thyroid, respiratory, and kidney conditions.
This modular design supports both general and condition-specific evaluations, while the clean layout ensures accessibility, even for users with minimal technical training. An overview of the data input panel is presented in Figure 8.

4.7.2. Disease Risk Prediction and Rule-Based Stratification

Once the patient’s data are submitted, the AI-RiskX system powered by a hybrid CNN–LSTM architecture predicts the most probable disease class. The output corresponds to one of six categories: heart disease, diabetes, asthma, thyroid disorder, chronic kidney disease, or no disease.
In parallel, a demographic-aware, rule-based stratification module categorizes patients into three risk levels: high risk, at risk, or not at risk. This stratification is determined by combining the predicted disease with critical demographic features such as age group (e.g., elderly or pediatric) and pregnancy status. For instance, an elderly patient (age ≥ 60 years) diagnosed with thyroid disorder would be flagged as high risk, with an explanatory note referencing their increased vulnerability and associated mortality risk.
This hybrid approach combines data-driven predictions with domain knowledge, enhancing medical interpretability and enabling clinicians to act with greater confidence. An example of the system output is shown in Figure 9.

4.7.3. Explainability Through SHAP

To ensure transparency in decision-making, AI-RiskX integrates SHAP (SHapley Additive Explanations) [34], a widely adopted method for model interpretability. SHAP quantifies the contribution of each input feature to the model’s prediction, offering both local (individual-level) and global (population-level) insights. These were visualized using bar plots and summary charts integrated into the interface.
Figure 10 presents a ranked list of influential features for a thyroid disorder prediction, as computed by the SHAP values. Thyroid-stimulating hormone (TSH) has the highest impact, followed by T4U (thyroid uptake), T3 (triiodothyronine), and patient age. These results align with clinical expectations and support explainable AI in a diagnostic context.

5. Discussion and Conclusions

We introduced AI-RiskX, an interpretable hybrid CNN–LSTM framework for multidisease risk prediction, combining SHAP-based explanations with a demographic-aware rule-based module. AI-RiskX classifies patients already diagnosed with COVID-19 or related infections, supporting timely intervention and resource allocation. Trained on a harmonized dataset covering asthma, diabetes, heart disease, chronic kidney disease, and thyroid disorders, AI-RiskX achieved 98.78% accuracy, outperforming LSTM, BiLSTM, CNN, and LLM–BERT baselines. SHAP confirmed medically relevant features, while the rule-based module stratified patients into high-risk, at-risk, and not-at-risk categories.
Extensive cross-validation and statistical testing confirmed the robustness and reliability of AI-RiskX, demonstrating consistently high predictive performance across all folds.
AI-RiskX supports rapid prioritization of high-risk patients in pandemic settings, guiding ICU, ventilation, and treatment allocation. Explainable predictions foster clinician trust, and integration hospital systems enables near real-time updates.
By unifying five heterogeneous datasets, modeling comorbidities, and combining CNN–LSTM features with demographic risk indicators, AI-RiskX delivers clinically interpretable risk categories while maintaining fairness and transparency.
In conclusion, AI-RiskX offers accurate, interpretable multidisease risk assessment, addressing the scalability, fairness, and trust critical in pandemics. The current study is limited by the absence of external validation and possible differences between synthetic and real-world data. Future work will assess AI-RiskX on independent cohorts to verify its robustness and applicability in clinical settings.

Author Contributions

N.Z.: Conceptualization, Methodology, Software, Writing—original draft preparation, Writing—review and editing, and Project administration; N.B. (Nardjes Bouchemal): Methodology, Formal analysis, Validation, Writing—original draft preparation, Writing—review and editing, Supervision, and Project administration; M.R.A.B.: Methodology, Software, Writing—original draft preparation, and Writing—review and editing; M.B.: Methodology, Formal analysis, Writing—original draft preparation, and Writing—review and editing; S.H.: Validation, Formal analysis, Writing—original draft preparation, Writing—review and editing, and Supervision; N.B. (Naila Bouchemal): Validation, Formal analysis, Writing—original draft preparation and Writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding. The APC was funded by Dr. Naila Bouchemal.

Data Availability Statement

The data supporting the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Dorjee, K.; Kim, H.; Bonomo, E.; Dolma, R. Prevalence and predictors of death and severe disease in patients hospitalized due to COVID-19: A comprehensive systematic review and meta-analysis of 77 studies and 38,000 patients. PLoS ONE 2020, 15, e0243191. [Google Scholar] [CrossRef] [PubMed]
  2. Wang, L.; Foer, D.; Bates, D.W.; Boyce, J.A.; Zhou, L. Risk factors for hospitalization, intensive care, and mortality among patients with asthma and COVID-19. J. Allergy Clin. Immunol. 2020, 146, 808–812. [Google Scholar] [CrossRef]
  3. Berumen-Lechuga, M.G.; Leaños-Miranda, A.; Molina-Pérez, C.J.; García-Cortes, L.R.; Palomo-Piñón, S. Risk Factors for Severe–Critical COVID-19 in Pregnant Women. J. Clin. Med. 2023, 12, 5812. [Google Scholar] [CrossRef]
  4. Zhou, F.; Yu, T.; Du, R.; Fan, G.; Liu, Y.; Liu, Z.; Xiang, J.; Wang, Y.; Song, B.; Gu, X.; et al. Clinical course and risk factors for mortality of adult inpatients with COVID-19 in Wuhan, China: A retrospective cohort study. Lancet 2020, 395, 1054–1062. [Google Scholar] [CrossRef]
  5. Petrilli, C.M.; Jones, S.A.; Yang, J.; Rajagopalan, H.; O’Donnell, L.; Chernyak, Y.; Tobin, K.A.; Cerfolio, R.J.; Francois, F.; Horwitz, L.I. Factors associated with hospital admission and critical illness among 5279 people with coronavirus disease 2019 in New York City: Prospective cohort study. BMJ 2020, 369, m1966. [Google Scholar] [CrossRef]
  6. Rajpurkar, P.; Chen, E.; Banerjee, O.; Topol, E.J. AI in health and medicine. Nat. Med. 2022, 28, 31–38. [Google Scholar] [CrossRef]
  7. Topol, E.J. High-performance medicine: The convergence of human and artificial intelligence. Nat. Med. 2019, 25, 44–56. [Google Scholar] [CrossRef]
  8. Ghassemi, M.; Oakden-Rayner, L.; Beam, A.L. The false hope of current approaches to explainable artificial intelligence in health care. Lancet Digit. Health 2021, 3, e745–e750. [Google Scholar] [CrossRef]
  9. Flores-Cisneros, L.; Gutiérrez-Vargas, R.; Escondrillas-Maya, C.; Zaragoza-Jiménez, C.; García-Rodríguez, G.; López-Gatell, H.; González-Islas, D. Risk factors for severe disease and mortality in children with COVID-19. Heliyon 2024, 10, e23629. [Google Scholar] [CrossRef] [PubMed]
  10. Al-Rawi, M.B.A.; Khan, A.H.I.; Sheikh Ghadzi, S.M. The severity of the COVID-19 among chronic disease patients in the selected hospitals in Riyadh, Saudi Arabia—A retrospective study. Risk Manag. Healthc. Policy 2024, 17, 3327–3339. [Google Scholar] [CrossRef] [PubMed]
  11. Rebiai, M.; Berkani, M.R.A.; Besseri, I.; Tobbal, S.; Bengherbia, B.; Zendaoui, N. Enhanced Cardiac Disease Detection Using MFCC Feature Extraction from PCG Signals. In Proceedings of the 2024 International Conference on Telecommunications and Intelligent Systems (ICTIS), Djelfa, Algeria, 14–15 December 2024; pp. 1–6. [Google Scholar]
  12. Berkani, M.R.A.; Chouchane, A.; Himeur, Y.; Ouamane, A.; Amira, A. An Intelligent Edge-Deployable Indoor Air Quality Monitoring and Activity Recognition Approach. In Proceedings of the 2023 6th International Conference on Signal Processing and Information Security (ICSPIS), Dubai, United Arab Emirates, 8–9 November 2023; pp. 184–189. [Google Scholar] [CrossRef]
  13. World Health Organization. COVID-19 and Noncommunicable Diseases: Evidence and Actions for Better Outcomes (Corrigendum 7–Final Version); World Health Organization: Geneva, Switzerland, 2024; Available online: https://www.who.int/docs/default-source/inaugural-who-partners-forum/covid-19-and-ncds---final---corr7.pdf (accessed on 25 April 2025).
  14. Booth, A.; Reed, A.B.; Ponzo, S.; Yassaee, A.; Aral, M.; Plans, D.; Labrique, A.; Mohan, D. Population risk factors for severe disease and mortality in COVID-19: A global systematic review and meta-analysis. PLoS ONE 2021, 16, e0247461. [Google Scholar] [CrossRef]
  15. Hasani, S.; Nasiri, H. COV-ADSX: An automated detection system using X-ray images, deep learning, and XGBoost for COVID-19. Softw. Impacts 2020, 11, 100210. [Google Scholar] [CrossRef]
  16. Ezzoddin, M.; Nasiri, H.; Dorrigiv, M. Diagnosis of COVID-19 cases from chest X-ray images using deep neural network and LightGBM. In Proceedings of the 2022 International Conference on Machine Vision and Image Processing (MVIP), Hong Kong, China, 25–27 March 2022; pp. 1–7. [Google Scholar]
  17. El-Sofany, H.; El-Seoud, S.A.; Karam, O.H.; Abd El-Latif, Y.M.; Taj-Eddin, I.A.T. A Proposed Technique Using Machine Learning for the Prediction of Diabetes Disease through a Mobile App. Int. J. Intell. Syst. 2024, 2024, 6688934. [Google Scholar] [CrossRef]
  18. Guleria, P.; Naga Srinivasu, P.; Ahmed, S.; Almusallam, N.; Alarfaj, F.K. XAI framework for cardiovascular disease prediction using classification techniques. Electronics 2022, 11, 4086. [Google Scholar] [CrossRef]
  19. Yılmaz, R.; Yağın, F.H. Early detection of coronary heart disease based on machine learning methods. Med. Rec. 2022, 4, 1–6. [Google Scholar] [CrossRef]
  20. Prathibha, S.; Dahiya, D.; Robin, C.R.; Nishkala, C.V.; Swedha, S. A novel technique for detecting various thyroid diseases using deep learning. Intell. Autom. Soft Comput. 2023, 35, 199–214. [Google Scholar] [CrossRef]
  21. Zhang, X.; Lee, V.C.S.; Rong, J.; Lee, J.C.; Liu, F. Deep convolutional neural networks in thyroid disease detection: A multi-classification comparison by ultrasonography and computed tomography. Comput. Methods Programs Biomed. 2022, 220, 106823. [Google Scholar] [CrossRef] [PubMed]
  22. Awal, M.A.; Hossain, M.S.; Debjit, K.; Ahmed, N.; Nath, R.D.; Habib, G.M.M.; Khan, M.S.; Islam, M.A.; Mahmud, M.A.P. An early detection of asthma using BOMLA detector. IEEE Access 2021, 9, 58403–58420. [Google Scholar] [CrossRef]
  23. Tahasamadsoltaniheris, M.; Mahmoodvand, Z.; Zolnoori, M. Intelligent diagnosis of Asthma using machine learning algorithms. Int. Res. J. Appl. Basic Sci. 2013, 5, 140–145. [Google Scholar]
  24. Bottrighi, A.; Pennisi, M.; Roveta, A.; Massarino, C.; Cassinari, A.; Betti, M.; Bolgeo, T.; Bertolotti, M.; Rava, E.; Maconi, A. A machine learning approach for predicting high risk hospitalized patients with COVID-19 SARS-Cov-2. BMC Med. Inform. Decis. Mak. 2022, 22, 340. [Google Scholar] [CrossRef]
  25. Quiroz-Juárez, M.A.; Torres-Gómez, A.; Hoyo-Ulloa, I.; León-Montiel, R.J.; U’Ren, A.B. Identification of high-risk COVID-19 patients using machine learning. PLoS ONE 2021, 16, e0257234. [Google Scholar] [CrossRef] [PubMed]
  26. Ahmed, S.; Kaiser, M.S.; Hossain, M.S.; Andersson, K. A comparative analysis of LIME and SHAP interpreters with explainable ML-based diabetes predictions. IEEE Access 2024, 13, 37370–37388. [Google Scholar] [CrossRef]
  27. Talukder, M.A.; Talaat, A.S.; Kazi, M. HXAI-ML: A Hybrid Explainable Artificial Intelligence Based Machine Learning Model for Cardiovascular Heart Disease Detection. Results Eng. 2025, 25, 104370. [Google Scholar] [CrossRef]
  28. Bilal, A.; Alzahrani, A.; Almohammadi, K.; Saleem, M.; Farooq, M.S.; Sarwar, R. Explainable AI-driven intelligent system for precision forecasting in cardiovascular disease. Front. Med. 2025, 12, 1596335. [Google Scholar] [CrossRef] [PubMed]
  29. Mahmood, S.; Hasan, R.; Hussain, S.; Adhikari, R. An Interpretable and Generalizable Machine Learning Model for Predicting Asthma Outcomes: Integrating AutoML and Explainable AI Techniques. World 2025, 6, 15. [Google Scholar] [CrossRef]
  30. Lyu, Y.; Jiang, Q.-C.; Yuan, S.; Hong, J.; Chen, C.-F.; Wu, H.-M.; Wang, Y.-Q.; Shi, Y.-J.; Yan, H.-X.; Xu, J. Non-invasive acoustic classification of adult asthma using an XGBoost model with vocal biomarkers. Sci. Rep. 2025, 15, 28682. [Google Scholar] [CrossRef]
  31. Srinivasu, P.N.; Ahmed, S.; Hassaballah, M.; Almusallam, N. An explainable Artificial Intelligence software system for predicting diabetes. Heliyon 2024, 10, e36112. [Google Scholar] [CrossRef]
  32. Chadaga, K.; Prabhu, S.; Sampathila, N.; Adiga, V.; Nayak, V. Explainable artificial intelligence approaches for COVID-19 prognosis prediction using clinical markers. Sci. Rep. 2024, 14, 1783. [Google Scholar] [CrossRef]
  33. Ozawa, T.; Chubachi, S.; Namkoong, H.; Takahashi, T.; Betsuyaku, T. Predicting coronavirus disease 2019 severity using explainable artificial intelligence techniques. Sci. Rep. 2025, 15, 9459. [Google Scholar] [CrossRef]
  34. Lundberg, S.M.; Lee, S.-I. A unified approach to interpreting model predictions. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  35. Duckworth, C.; Chmiel, F.P.; Burns, D.K.; Zlatev, Z.D.; White, N.M.; Daniels, T.W.V.; Kiuber, M.; Boniface, M.J. Using explainable machine learning to characterise data drift and detect emergent health risks for emergency department admissions during COVID-19. Sci. Rep. 2021, 11, 23017. [Google Scholar] [CrossRef] [PubMed]
  36. Zhao, Y.; He, B.; Xu, Z.; Zhang, Y.; Zhao, X.; Huang, Z.-A.; Yang, F.; Wang, L.; Duan, L.; Song, J.; et al. Interpretable artificial intelligence model for accurate identification of medical conditions using immune repertoire. Briefings Bioinform. 2023, 24, bbac555. [Google Scholar] [CrossRef]
  37. Tasin, I.; Nabil, T.U.; Islam, S.; Khan, R. Diabetes prediction using machine learning and explainable AI techniques. Healthc. Technol. Lett. 2023, 10, 1–10. [Google Scholar] [CrossRef] [PubMed]
  38. Zendaoui, N.; Bouchemal, N.; Benabdelhafid, M. AI-LMS: AI-Based Long-Term Monitoring System for Patients in Pandemics: COVID-19 Case Study. In Proceedings of the International Conference on Model and Data Engineering, Sousse, Tunisia, 2–4 November 2023; pp. 272–285. [Google Scholar]
  39. Janosi, A.; Steinbrunn, W.; Pfisterer, M.; Detrano, R. Heart Disease. UCI Machine Learning Repository. 1988. Available online: https://archive.ics.uci.edu/dataset/45/heart+disease (accessed on 25 April 2025).
  40. Selva Balaji. Diabetes. Mendeley Data, V1. 2024. Available online: https://data.mendeley.com/datasets/pd64hfttwy/1 (accessed on 25 April 2025).
  41. Aversano, L.; Bernardi, M.L.; Cimitile, M.; Iammarino, M.; Macchia, P.E.; Nettore, I.C.; Verdone, C. Thyroid disease treatment prediction with machine learning approaches. Procedia Comput. Sci. 2021, 192, 1031–1040. [Google Scholar] [CrossRef]
  42. Rabie El Kharoua. Asthma Disease Dataset. Kaggle. 2024. Available online: https://www.kaggle.com/dsv/8669080 (accessed on 25 April 2025).
  43. Rubini, L.; Soundarapandian, P.; Eswaran, P. Chronic Kidney Disease. UCI Machine Learning Repository. 2015. Available online: https://archive.ics.uci.edu/dataset/336/chronic+kidney+disease (accessed on 25 April 2025).
  44. Kiranyaz, S.; Avci, O.; Abdeljaber, O.; Ince, T.; Gabbouj, M.; Inman, D.J. 1D convolutional neural networks and applications: A survey. Mech. Syst. Signal Process. 2021, 151, 107398. [Google Scholar] [CrossRef]
  45. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  46. Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005, 18, 602–610. [Google Scholar] [CrossRef]
  47. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  48. Lipton, Z.C.; Kale, D.C.; Elkan, C.; Wetzell, R. Learning to diagnose with LSTM recurrent neural networks. arXiv 2016. [Google Scholar] [CrossRef]
  49. Alabi, O.; Teikari, O.; Szwajda, A.; Viiki, M.; Rönkö, E.; Pahikkala, T.; Aittokallio, T. Machine learning explainability for cancer prognosis: A systematic review of techniques, applications, and challenges. Briefings Bioinform. 2023, 24, 4. [Google Scholar]
  50. Salih, A.M.; Raisi-Estabragh, Z.; Galazzo, I.B.; Radeva, P.; Petersen, S.E.; Lekadir, K.; Menegaz, G. A perspective on explainable artificial intelligence methods: SHAP and LIME. Adv. Intell. Syst. 2025, 7, 2400304. [Google Scholar] [CrossRef]
  51. Panda, M.; Mahanta, S.R. Explainable artificial intelligence for Healthcare applications using Random Forest Classifier with LIME and SHAP. In Explainable, Interpretable, and Transparent AI Systems; CRC Press: Boca Raton, FL, USA, 2024; pp. 89–105. [Google Scholar]
  52. Dietterich, T.G. Approximate Statistical Tests for Comparing Supervised Classification Learning Algorithms. Neural Comput. 1998, 10, 1895–1923. [Google Scholar] [CrossRef] [PubMed]
  53. Bouckaert, R.R.; Frank, E. Evaluating the Replicability of Significance Tests for Comparing Learning Algorithms. In Advances in Knowledge Discovery and Data Mining; Springer: Berlin/Heidelberg, Germany, 2004; pp. 3–12. [Google Scholar] [CrossRef]
  54. Demšar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  55. Japkowicz, N.; Shah, M. Evaluating Learning Algorithms: A Classification Perspective; Cambridge University Press: Cambridge, UK, 2011; ISBN 978-0-521-19813-9. [Google Scholar]
  56. Abid, A.; Abdalla, A.; Wolf, T.; Zou, J. Gradio: Hassle-Free Sharing and Testing of ML Models in the Wild. In Proceedings of the 2019 ICML Workshop on Human in the Loop Learning (HILL 2019), Long Beach, CA, USA, 9–15 June 2019. [Google Scholar]
Figure 1. AI-RisX phases.
Figure 1. AI-RisX phases.
Bioengineering 12 01127 g001
Figure 2. AI-RiskX dataset creation and harmonization workflow.
Figure 2. AI-RiskX dataset creation and harmonization workflow.
Bioengineering 12 01127 g002
Figure 3. Proposed AI-RiskX architecture (CNN–LSTM). The model fuses three convolutional branches and one LSTM branch to capture multi-resolution spatial and sequential dependencies in parallel.
Figure 3. Proposed AI-RiskX architecture (CNN–LSTM). The model fuses three convolutional branches and one LSTM branch to capture multi-resolution spatial and sequential dependencies in parallel.
Bioengineering 12 01127 g003
Figure 4. ROC curves of the CNN–LSTM model for multidisease classification on AI-RiskX.
Figure 4. ROC curves of the CNN–LSTM model for multidisease classification on AI-RiskX.
Bioengineering 12 01127 g004
Figure 5. Confusion matrix of the CNN–LSTM model on AI-RiskX: (a) validation set and (b) testing set results.
Figure 5. Confusion matrix of the CNN–LSTM model on AI-RiskX: (a) validation set and (b) testing set results.
Bioengineering 12 01127 g005
Figure 6. SHAP waterfall plots of local feature contributions across disease categories; red increases risk, blue decreases. (a) Asthma; (b) diabetes; (c) heart problem; (d) chronic kidney disease; (e) thyroid problems; and (f) combined analysis.
Figure 6. SHAP waterfall plots of local feature contributions across disease categories; red increases risk, blue decreases. (a) Asthma; (b) diabetes; (c) heart problem; (d) chronic kidney disease; (e) thyroid problems; and (f) combined analysis.
Bioengineering 12 01127 g006
Figure 7. Global SHAP feature importance for the CNN–LSTM model across disease categories. Panels (ae) show asthma, diabetes, heart disease, chronic kidney disease, and thyroid disorders; panel (f) shows the combined analysis.
Figure 7. Global SHAP feature importance for the CNN–LSTM model across disease categories. Panels (ae) show asthma, diabetes, heart disease, chronic kidney disease, and thyroid disorders; panel (f) shows the combined analysis.
Bioengineering 12 01127 g007
Figure 8. Demographic and clinical information data entry panel.
Figure 8. Demographic and clinical information data entry panel.
Bioengineering 12 01127 g008
Figure 9. Example output showing predicted disease, confidence, risk level, and demographic reasoning.
Figure 9. Example output showing predicted disease, confidence, risk level, and demographic reasoning.
Bioengineering 12 01127 g009
Figure 10. SHAP-based feature importance ranking for thyroid prediction (TSH as the most influential feature, T4U, T3, and age).
Figure 10. SHAP-based feature importance ranking for thyroid prediction (TSH as the most influential feature, T4U, T3, and age).
Bioengineering 12 01127 g010
Table 1. COVID-19 key risk factors and mortality rates.
Table 1. COVID-19 key risk factors and mortality rates.
Risk FactorMortality/Severity Rate (%)
Heart Disease52.0
Chronic Kidney Disease48.0
Hypertension28.0
Diabetes24.0
Severe Asthma (ICU)23.3
Elderly (≥60 years)85.0
Children Severe COVID-19 Cases30.0
Children ICU Admissions28.0
Children Invasive Ventilation6.0
Pregnancy-Related Severe COVID-1910.2
Table 2. Related work discussion.
Table 2. Related work discussion.
AuthorsDisease ScopeAI MethodologyExplainability UsedVulnerable GroupsLimitations
Bottrighi et al. [24]Single (COVID-19)JRIP (rule-based ML)NoneNoClass imbalance, no explainability
Zhao et al. [36]Single (Immune Disorders)XGBoost (VDJMiner)Yes (SHAP)NoLimited to immune repertoire data
Quiroz et al. [25]Single (COVID-19)kNN, SVM, NNNoneNoNot generalizable
Duckworth et al. [35]Single (COVID-19-related Risks)NN + SHAPYes (SHAP)NoSingle-disease, limited dataset sources
Tasin et al. [37]Single (Diabetes)XGBoost, SMOTEYes (LIME, SHAP)NoSmall dataset, lacks demographic profiling
El-Sofany et al. [17]Single (Diabetes)XGBoost + SMOTENoneNoSingle-disease, no explainability
Srinivasu et al. [31]Single (Diabetes)CNN + BiLSTMNoneNoImage-based, no demographic analysis
Guleria et al. [18]Single (Heart)SVM + SHAP/LIMEYes (SHAP, LIME)NoSmall dataset, no deep learning
Yilmaz et al. [19]Single (Heart)RF, LR, SVMNoneNoNo explainability, no vulnerable group identification
Prathibha et al. [20]Single (Thyroid)ResNet (DL)NoneNoOnly image-based data, single disease
Zhang et al. [21]Single (Thyroid)Xception CNNNoneNoFocused only on ultrasound/CT
Awal et al. [22]Single (Asthma)BOMLA (ML)NoneNoDataset imbalance, no explanation logic
Heris et al. [23]Single (Asthma)KNNNoneNoRisk of overfitting, small dataset
Ahmed et al. [26]Partial (CVDs)ML + statistical modelingNoneNoGenomic focus, lacks clinical variables
AI-RiskX (Ours)Multi (5 diseases)CNN-LSTM + Rule-based + SHAPYes (SHAP)Yes (children, elderly, pregnant)None: unified dataset, deep + rule fusion, explainability included
Table 3. AI-RiskX dataset overview.
Table 3. AI-RiskX dataset overview.
PropertyDescription
Total samples26,125
Number of features79
Feature categoriesDemographics, Lifestyle, Clinical measurements, Thyroid tests, Cardiac, Respiratory, Allergy/Immunology
Target variableMulti-class disease label
Source integrationFive heterogeneous datasets (Asthma, Diabetes, Heart, Kidney, Thyroid)
Table 4. Feature group overview.
Table 4. Feature group overview.
GroupCountExamples
Demographic5Age, Sex, Ethnicity
Lifestyle5Smoking, Physical Activity
Cardiac9Trestbps, Chol, Thalach
Diabetes6Glucose, Insulin, BMI
Kidney11Bu, Sc, Sod
Thyroid16TSH, TT4, T3
Asthma15Wheezing, Coughing
Table 5. Class distribution in the training set after SMOTE oversampling (applied only to the training subset).
Table 5. Class distribution in the training set after SMOTE oversampling (applied only to the training subset).
LabelDiseaseSamplesAfter SMOTE
0No disease22,42022,420
1Asthma12422,420
2Diabetes95222,420
3Heart52622,420
4Kidney25022,420
5Thyroid185322,420
Total 26,125134,520
Table 6. Summary of the proposed CNN–LSTM architecture for AI-RiskX.
Table 6. Summary of the proposed CNN–LSTM architecture for AI-RiskX.
StageLayer(s)Output Shape
Input79 numeric clinical features(79, 1)
Feature Extraction3× Conv1D (64 filters, k = 3/5/7, ReLU) + BatchNorm(79, 64) each
Feature Enrichment2nd Conv1D set (same configuration)(79, 64) each
PoolingMaxPooling1D (pool size = 2) for each conv branch(39, 64) each
LSTM BranchLSTM(64) + BatchNorm → LSTM(32, no recurrence)(32)
FusionFlatten + Concatenate all branches(vector)
Dense LayersDense(64, ReLU) → Dense(32, ReLU) + Dropout(0.3)(32)
Output LayerDense(6, softmax)(6)
Table 7. Performance comparison between classical and deep learning models for multidisease prediction under different dataset configurations.
Table 7. Performance comparison between classical and deep learning models for multidisease prediction under different dataset configurations.
ApproachDatasetModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
StandaloneHeartLSTM96.5696.5696.5596.55
DiabetesLSTM96.9396.9396.9296.92
HHTLSTM96.3496.3496.3396.33
ADLSTM95.2095.2095.1995.19
CKDLSTM96.2596.2596.2496.24
HeartBiLSTM97.0797.0797.0597.05
DiabetesBiLSTM97.4097.4097.3997.39
HHTBiLSTM96.5796.5796.5596.55
ADBiLSTM94.7894.7794.7594.75
CKDBiLSTM97.5097.5097.4997.49
HeartCNN89.7589.7589.7089.72
DiabetesCNN86.8286.8286.8186.81
HHTCNN91.2191.2191.2091.20
ADCNN93.3293.3293.3293.32
CKDCNN76.2576.2576.2376.23
CombinedAI-RiskX Logistic Regression 89.95 88.64 89.95 87.75
AI-RiskX Random Forest 96.77 96.34 96.77 96.55
AI-RiskX XGBoost 96.99 96.62 96.99 96.80
AI-RiskXLLM-BERT79.3579.3079.2179.26
AI-RiskXLSTM93.1693.1693.1293.13
AI-RiskXBiLSTM90.9190.9190.8990.89
AI-RiskXCNN94.2694.2694.2494.24
AI-RiskXCNN-LSTM94.5594.5594.5394.53
Combined–BalancedAI-RiskX Logistic Regression 70.68 89.40 70.68 76.21
AI-RiskX Random Forest 96.84 96.47 96.84 96.65
AI-RiskX XGBoost 96.98 96.68 96.98 96.81
AI-RiskXLLM-BERT86.0286.0185.9585.97
AI-RiskXLSTM96.9796.9796.9596.95
AI-RiskXBiLSTM96.1396.1396.1296.12
AI-RiskXCNN98.6798.6798.6698.66
AI-RiskXCNN-LSTM98.7898.7998.7798.77
Table 8. Ablation study results comparing different architectural configurations on the balanced dataset.
Table 8. Ablation study results comparing different architectural configurations on the balanced dataset.
ModelAccuracyPrecisionRecallF1-Score
LSTM only96.9796.9796.9596.95
CNN only98.6798.6798.6698.66
CNN–LSTM (AI-RiskX)98.7898.7998.7798.77
Table 9. Statistical summary of cross-validation results (Mean ± Standard Deviation).
Table 9. Statistical summary of cross-validation results (Mean ± Standard Deviation).
MetricMean ± Std (%)
Accuracy98.72 ± 0.11
Precision98.71 ± 0.10
Recall98.73 ± 0.09
F1-score98.71 ± 0.08
AUC99.52 ± 0.03
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zendaoui, N.; Bouchemal, N.; Berkani, M.R.A.; Bouzahzah, M.; Harous, S.; Bouchemal, N. AI-RiskX: An Explainable Deep Learning Approach for Identifying At-Risk Patients During Pandemics. Bioengineering 2025, 12, 1127. https://doi.org/10.3390/bioengineering12101127

AMA Style

Zendaoui N, Bouchemal N, Berkani MRA, Bouzahzah M, Harous S, Bouchemal N. AI-RiskX: An Explainable Deep Learning Approach for Identifying At-Risk Patients During Pandemics. Bioengineering. 2025; 12(10):1127. https://doi.org/10.3390/bioengineering12101127

Chicago/Turabian Style

Zendaoui, Nada, Nardjes Bouchemal, Mohamed Rafik Aymene Berkani, Mounira Bouzahzah, Saad Harous, and Naila Bouchemal. 2025. "AI-RiskX: An Explainable Deep Learning Approach for Identifying At-Risk Patients During Pandemics" Bioengineering 12, no. 10: 1127. https://doi.org/10.3390/bioengineering12101127

APA Style

Zendaoui, N., Bouchemal, N., Berkani, M. R. A., Bouzahzah, M., Harous, S., & Bouchemal, N. (2025). AI-RiskX: An Explainable Deep Learning Approach for Identifying At-Risk Patients During Pandemics. Bioengineering, 12(10), 1127. https://doi.org/10.3390/bioengineering12101127

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop