Next Article in Journal
Advancing Diabetic Retinopathy Screening: A Systematic Review of Artificial Intelligence and Optical Coherence Tomography Angiography Innovations
Previous Article in Journal
Artificial Intelligence in Relation to Accurate Information and Tasks in Gynecologic Oncology and Clinical Medicine—Dunning–Kruger Effects and Ultracrepidarianism
Previous Article in Special Issue
Leveraging Attention-Based Deep Learning in Binary Classification for Early-Stage Breast Cancer Diagnosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Diabetic Foot Ulcers Detection Model Using a Hybrid Convolutional Neural Networks–Vision Transformers

by
Abdul Rahaman Wahab Sait
1,* and
Ramprasad Nagaraj
2
1
Department of Archives and Communication, Center of Documentation and Administrative Communication, King Faisal University, P.O. Box 400, Hofuf 31982, Al-Ahsa, Saudi Arabia
2
Department of Biochemistry, S S Hospital, S S Institute of Medical Sciences & Research Centre, Rajiv Gandhi University of Health Sciences, Davangere 577005, Karnataka, India
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(6), 736; https://doi.org/10.3390/diagnostics15060736
Submission received: 10 February 2025 / Revised: 13 March 2025 / Accepted: 14 March 2025 / Published: 15 March 2025

Abstract

:
Background: Diabetic foot ulcers (DFUs) are severe and common complications of diabetes. Early and accurate DFUs classification is essential for effective treatment and prevention of severe complications. The existing DFUs classification methods have certain limitations, including limited performance, poor generalization, and lack of interpretability, restricting their use in clinical settings. Objectives: To overcome these limitations, this study proposes an innovative model to achieve robust and interpretable DFUs classification. Methodology: The proposed DFUs classification integrates MobileNet V3-SWIN, LeViT-Peformer, Tensor-based feature fusion, and ensemble splines-based Kolmogorov–Arnold Networks (KANs) with Shapley Additive exPlanations (SHAP) values to classify DFUs severities into ischemia and infection classes. In order to train and generalize the proposed model, the authors utilized the DFUs challenge (DFUC) 2021 and 2020 datasets. Findings: The proposed model achieved state-of-the-art performance, outperforming the existing approaches by obtaining an average accuracy of 98.7%, precision of 97.3%, recall of 97.4%, and F1-score of 97.3% on DFUC 2021. On DFUC 2020, it maintained a robust generalization accuracy of 96.9%, demonstrating superiority over standalone and baseline models. The study findings have significant implications for research and clinical practice. The findings offer an effective platform for scalable and explainable automated DFUs treatment and management, improving patient outcomes and clinical practices.

1. Introduction

A significant consequence of diabetes mellitus, diabetic foot ulcers (DFUs) affect millions of individuals across the globe [1]. These ulcers pose a significant public health risk, leading to morbidity, amputations, and mortality. Factors including neuropathy, ischemia, and infection delay the healing process, resulting in the formation of these ulcers [2]. DFUs have a substantial impact on individuals’ quality of life [3]. Individuals with chronic wounds generally experience intense pain, restricted movement, and a lack of social interaction [4]. These wounds require extended medical consultations, dressing changes, and occasionally surgery interventions. Due to their limited physical ability and dependency on caretakers, individuals encounter depression and anxiety [5]. Hospitalization or amputations may increase the length of stay and cause substantial healthcare expenses.
Effective DFU therapy requires timely and precise ischemia and infection identification [6]. Traditionally, DFUs diagnosis relied on laboratory testing, imaging, and clinical evaluations [7]. Clinicians identify infections using clinical symptoms, bacterial cultures, and inflammatory markers, whereas perfusion levels determine ischemia. The dependence on clinical knowledge renders these procedures laborious, subjective, and inconsistent. Imaging modalities, including thermal, hyperspectral, and fluorescence imaging have gained traction in diagnosing DFUs in the initial stages [8]. These modalities offer detailed insights into blood flow, tissue composition, and infection biomarkers [9]. However, extensive implementation costs and a steep learning curve for healthcare professionals prevent the widespread adoption of these modalities. In contrast, photographic images capture visible wound features, including size, shape, and discoloration, with sufficient clarity [10]. These features can be shared for telemedicine consultations, supporting early DFUs diagnosis.
In the absence of prompt medical attention, ulcers have the potential to penetrate deep into soft tissue, tendons, and potentially bone, elevating the likelihood of developing osteomyelitis and systemic infections leading to the amputation of the lower extremities [10]. By using an artificial intelligence-powered DFUs detection system, healthcare personnel may initiate debridement and infection management in the early stages, preventing severe consequences [11]. The advent of deep learning (DL) techniques has enabled automated DFUs detection, minimizing human errors. These models can capture crucial features associated with ischemia and infection classes using photographic images [11]. They are ideal for high-throughput diagnostics, handling large volumes of images in limited time [12]. Convolutional neural networks (CNNs) and vision transformers (ViTs) are robust DL architectures, playing a significant role in extracting features from the images [13]. Each architecture has the potential to offer unique DFUs features, contributing to reliable DFUs detection. CNNs use convolution layers to capture the size, shape, and texture of ulcers with surrounding skin abnormalities [14]. The hierarchical structure detects subtle features of ischemia and infection. Similarly, ViTs identify global relationships within an image [15]. These models process an image as a sequence of patches. The overall condition and spatial relationships between multiple ulcers can be extracted through the ViTs architecture. The inherent adaptability of ViTs enables model interpretability, highlighting regions influencing their predictions [16]. Although CNNs and ViTs architectures demonstrate remarkable outcomes, they fail to provide deeper insights into their decision-making processes [17]. Accurate DFUs detection and classification requires diverse meaningful features [18]. The feature fusion technique enriches the data representation. The absence of reliable feature fusion techniques integrating local and global features leads to suboptimal classification outcomes. The lack of model interpretability and inability to combine the features of multiple architectures cause challenges in identifying ischemia and infection images [19].
The ensemble learning (EL) approach combines the predictions of multiple models [20]. It reduces the risk of overfitting the training data. It enhances the model’s generalization capability [21]. In the context of DFUs classification, an individual classification model may struggle to maintain a trade-off between specificity and sensitivity. Combining feature fusion and EL approaches can significantly improve classification performance. These features motivated the authors to develop an EL-based DFUs classification model. The proposed model integrates the advanced feature extraction and EL approaches to identify DFUs in the initial stages. In addition, it offers interpretable outcomes to support clinicians in making effective decisions. In this study, the authors provide the following unique contributions to the DFUs detection and classification literature:
  • Hybrid feature extraction framework using the potentials of CNNs and ViTs:
The proposed framework leverages the strengths of MobileNet V3-SWIN and LeViT-Performer models to extract fine-grained localized features and spatial arrangement of ulcers and healthy tissues. It addresses the limitations of individual approaches by developing a robust and comprehensive feature extraction process.
2.
Tensor fusion technique-based critical feature identification:
An innovative feature fusion is introduced to combine key features associated with ischemia and infection images. By leveraging the potential of tensor fusion techniques, the proposed feature fusion overcomes the limitations of existing techniques. In addition, it preserves the integrity of diverse features.
3.
Ensembled splines-driven Kolmogorov–Arnold Networks (KANs)-based DFUs classification.
The proposed study offers a technique to ensemble the outcomes of spline functions within the KANs. It enhances the interpretability and classification performance. The KANs-based classification uses Shapley additive explanations (SHAP) to build Explainable AI, providing insights into the classification process.
The remaining part of this study is structured into four key sections, presenting a comprehensive understanding of the development and evaluation of an interpretable DFUs classification model. Section 2 outlines the data preparation and architectural design of the proposed model. It reveals the hybrid feature extraction and KANs-based classification approaches. The study findings are outlined in Section 3. Section 4 contains the interpretation of the findings and emphasizes the model’s contributions to DFUs classification. It covers the challenges and future model improvements. Lastly, Section 5 summarizes the study’s outcomes, underscoring the potential clinical impact of the proposed DFUs classification.

2. Materials and Methods

This study introduces a model integrating hybrid feature extraction, tensor fusion-based feature fusion, and KANs-based DFUs classification. The synergy of MobileNet V3-SWIN and LeViT-Performer overcomes the shortcomings of individual CNNs and ViTs-based feature extraction models. To optimally integrate the diverse DFUs features, the authors employed tensor fusion-based feature fusion. This approach was used to enhance the feature representation and improve model robustness. The KANs-based classifier with SHAP values decomposes complex decision boundaries into smooth and explainable sub-functions, achieving reliable and interpretable classification. The proposed research methodology is outlined in Figure 1.

2.1. Data Acquisition and Preparation

The effectiveness of DL-based DFUs classification depends on the quality and diversity of the dataset. In this study, the authors utilized the diabetic foot ulcer challenge (DFUC) 2021 and DFUC 2020 datasets [22] to develop a robust and interpretable model. They selected these datasets due to high-quality annotations, balanced class distributions, and diverse image conditions. The images were captured using three digital cameras, including Kodak DX4560, Nikon D3300, and Nikon Coolpix P100. Proper authorization was obtained from the dataset providers to use the foot images for this study. DFUC 2021 [22] contains 15,683 images, whereas DFUC 2020 [22] includes 4000 foot images. These are designed for DFUs classification. They cover ulcer images collected over different time frames. As a result, the proposed model can understand the temporal evolution of DFUs, distinguishing between early-stage and chronic ulcers. It can evaluate the progression of the disease using the dataset’s time-series characteristics. Expert wound care specialists clinically evaluated the DFUC datasets. They manually annotated and classified the ulcers based on medical guidelines and diagnostic criteria. The datasets serve as a gold standard benchmark for evaluating DL-based DFUs classification models. By leveraging the DFUC 2021 and DFUC 2020 longitudinal datasets, the proposed model can support dynamic, personalized, and clinically relevant ulcer management in real-world settings. Figure 2a,b show the sample infection and ischemia images of the DFUC 2021 dataset. The use of multiple datasets offers an environment for comprehensive and unbiased evaluation. Table 1 reveals the key features of the datasets.
Although CNNs and ViTs are designed for robust feature extraction using the raw images, data preprocessing remains crucial to guarantee generalization and stability in DFUs classification. The authors standardized the images to 224 × 224 pixels. Mean subtraction and standard deviation scaling were used to scale the pixel intensity values between 0 and 1. Additionally, Gaussian filtering was employed to smoothen artifacts in order to overcome unnecessary distortions in feature representations. The authors used the LAB color space conversion technique to enhance ischemia detection. To improve model generalization, the authors employed data augmentation techniques. The data augmentation techniques introduce controlled perturbations into the dataset to train the model to recognize perturbation patterns and adjust accordingly. Geometric transformations, including random rotations, horizontal flips, and perspective transformations, were used to train the model to learn orientation-invariant features. Random rotation supports the model in identifying ulcers from different orientations. Horizontal and vertical flipping addresses variability in foot positioning. It assists the model in maintaining stability across real-time images. Random clipping was performed to prevent the model from being dependent on specific ulcer positions within the image. Color augmentations, including brightness, contrast, saturation, and hue adjustments, were used to simulate variations in clinical lighting conditions. The simulation of real-world distortions can improve the model’s resilience to noise, occlusions, and imaging inconsistencies. In order to generate synthetic ulcer images, the authors used a cycle-consistent generative adversarial network (CycleGAN). CycleGAN applies image-to-image translation, allowing the transformation of images between domains without requiring paired training data. To counteract the class imbalances, the authors selectively applyed data augmentation to minority classes while maintaining relatively unchanged non-infected ulcer images.

2.2. MobileNet V3-SWIN Feature Extraction

MobileNet V3 [23] is an optimized CNN backbone using depthwise separable convolutions to extract the fine-grained texture, ulcer boundaries, and local tissue characteristics. The effective spatial encoding technique captures ulcer-specific features better than a standard CNN layer. SWIN transformer typically processes the image with a convolutional layer. This layer is limited in its ability to capture complex textures, edges, and fine-grained patterns of ulcers. Thus, the authors replaced the SWIN transformer’s initial convolutional stem with the MobileNet V3 backbone. Figure 3 illustrates the feature extraction process using the synergy of the MobileNet V3-SWIN model.
The MobileNet V3 model facilitates early-stage feature extraction while the SWIN transformer hierarchically handles the extracted feature maps. Equation (1) shows the feature extraction approach using the MobileNet V3 model.
F M = σ W d X W P
where F M is the MobileNet V3 feature, W d is the depthwise convolution kernel, W P is the pointwise convolution kernel, σ is the hard-swish activation, is the matrix multiplication, and is the element-wise multiplication.
SWIN divides the extracted feature maps into small non-overlapping windows and applies self-attention within each window. Without substantial memory, it refines high-dimensional features. Equation (2) outlines the process of feature representation using multi-head self-attention (MSA).
F R l = M S A Q , K , V + F R l 1
where   F R l is the feature representation at layer l, and M S A Q , K , V is computed using Equation (3).
M S A Q , K , V = S o f t m a x Q K T d V  
where Q , K , a n d   V are the query, key, and value matrices, and d is the scaling factor.
SWIN [24] employs the shifted windows function in the subsequent layers, maintaining continuity in feature representation. By leveraging multiple transformer blocks, the final global feature representation ( F S ) is obtained, encapsulating high-level spatial relationships and structural abnormalities over the DFUs region. Gradient SHAP values are used to highlight the feature contributions. Although SWIN guarantees optimal feature utilization, the fine-grained local details of the MobileNet V3 may not be preserved. Thus, the authors retained the MobileNet V3 features and fused them with SWIN hierarchical features using the tensor fusion-based feature fusion.

2.3. LeViT-Performer Feature Extraction

To extract diverse DFUs features, the authors employed a hybrid ViTs-based feature extraction. They combined the potential of LeViT [25] and Performer [26] models. By leveraging this strategy, fine details such as ulcer textures, boundaries, and infection patterns with global structural dependencies, including ulcer distribution across foot regions, are extracted. Figure 4 reveals the recommended LeViT-Performer model.
Initially, LeViT captures fine details using its CNN-based embedding. Equation (4) outlines the feature extraction process.
F L = σ W c X + b
where F L is the LeViT features, X   R H × W × C is the image, W c is the convolutional weight matrix, represents the convolutional operation, R is the image tensor with height (H), weight (W), and channels (C), and σ is the GELU activation function.
To refine the extracted features, LeViT uses MSA. This approach generates the ulcer-specific textual variations. However, it lacks the ability to effectively model long-range feature dependencies. To overcome this limitation, the Performer transformer is used to represent features with global ulcer patterns. The Kernel-based approximation is employed to encode the global features. The transformer SHAP value is used to identify the critical features using their weights. The projection layer refines the encoded features as shown in Equation (5).
F P = w P F R P F L
where F R P F L is the final feature representation from Performer using F L , w P is a projection weight matrix, and F P is the extracted features using the LeViT-Performer model.

2.4. Tensor Fusion-Based Feature Fusion

Effective integration of multi-source feature representation is essential for achieving high classification accuracy. In order to retain local and hierarchical feature representation, the authors employed a tensor fusion approach [27,28]. Unlike conventional feature fusion approaches, tensor fusion uses adaptive weights. It allows the model to prioritize relevant features of ischemia and infection classes. Dimensionality reduction is used to remove redundant features, reducing computational overhead. By leveraging this approach, memory footprint can be reduced, enabling real-time implementation with limited resources. Figure 5 presents the suggested feature fusion approach.
The authors applied feature normalization to prevent feature dominance. The feature maps were extracted from different architectures with distinct dimensions. The authors applied a learnable projection transformation to ensure compatibility. Equation (6) presents the normalization of feature maps.
F M = Normalization W M F M F S = Normalization W S F S F P = Normalization W P F P
where W M , W S , and W P are learnable weight matrices, projecting feature maps into a common latent space to guarantee dimensional compatibility before fusion.
The trainable fusion weights (α, β, and γ) are used to dynamically adjust the contributions of each feature. Equation (7) shows the feature fusion process.
F f u s e d = α   F M + β F S + γ F P
where F f u s e d is the fused feature, F M is the MobileNet V3 features, F S is the SWIN features, F P is the LeViT-Performer features, and α, β, and γ are adaptive fusion weights optimized through gradient descent. These weights are normalized to ensure stability, as shown in Equation (8).
α + β + γ = 1 ,   α ,   β ,   γ   0 .
Equation (9) highlights the loss function, optimizing fusion weights for accurate DFUs classification.
L f u s e d = i = 1 N y i f F f u s e d 2
where y i is the ground truth label, f F f u s e d is the outcome after feature fusion, and N represents the number of samples.
Furthermore, the authors applied substantial dimensionality reduction using principal component analysis (PCA) to minimize the computational overheads.

2.5. Ensembled Splines Driven KANs-Based Classification

The KANs [29] classifier utilizes three splines, B-splines, natural splines, and polynomial splines, to classify ischemia and infection classes. Each spline provides a different level of smoothness and flexibility in decision boundaries. An ensemble voting mechanism is applied to combine the outcomes of three splines, achieving a stable classification. B-splines provide smooth and stable approximations. It can handle subtle variations in ulcer shape and color. Natural splines identify gradual variations in ulcer texture better than B-splines. Polynomial splines fit high-order polynomials to data, capturing ulcer region complexity. A voting-based ensemble mechanism is used to improve prediction accuracy and generalization. The authors employed a soft voting (weighted voting) approach to reduce the impact of single misclassification. Equation (10) presents the voting approach for classifying the images.
P F y = C j F f u s e d = i = 1 N w i P i y = C j F f u s e d  
where P F is the final prediction, y is the class label, C j is the class with probability ( P i ), and w i is the weight assigned to classifier i based on BOHB optimization.
To interpret the feature contributions, the SHAP [30] value is used. It analyzes the importance of each fused feature. Equation (11) outlines the computation of the SHAP values.
S i = Z K A N s F f u s e d Z K A N s F f u s e d i
where S i is the SHAP values of different splines, Z K A N s F f u s e d is the prediction with all features, and Z K A N s F f u s e d i is the prediction when feature i is removed.

3. Results

The proposed model was developed and evaluated in a high-performance computing environment to ensure efficient training and inference. The experiment was performed on Windows 11 Pro (64-bit), powered by an Intel i7+ processor with NVIDI GeForce RTX 3080 Ti (12 GB GDDR 6X VRAM). The tensor cores of the RTX 3080 Ti were used to optimize matrix operations, batch normalizations, and self-attention mechanisms. The authors implemented the model using Python 3.9., providing a stable ecosystem for feature extraction and classification. The TIMM (Version 0.6.12.) library was employed for MobileNet V3-SWIN and LeViT-Performer-based feature extractions. In addition, different libraries, including Torchvision (version 0.12.0), NumPy (version 1.21.25.), Albumentations (version 1.1.0.), MatPlotlib (version 3.5.1.), and Seaborn (version 0.11.2.), were used for the model development. Table 2 highlights the configuration details for the model implementation.
The authors evaluated the model’s performance using well-structured dataset partitioning and comprehensive evaluation metrics, ensuring robustness and statistical reliability. The DFUC 2021 dataset was divided into three subsets: train set (70%), validation set (15%), and test set (15%). To assess the generalization capability across different datasets, the DFUC 2020 dataset was utilized, using 20% of its images. This strategy was used to guarantee that the model was not overfitted to a single dataset. The performance metrics, including accuracy (Acc), precision (Prec), Recall (Rec), F1-score (F1), and specificity (Spec), were used to determine the model’s classification ability. The area under the receiver operating characteristic curve (AUROC) is used to ensure clinical reliability and real-world applicability. In addition, standard deviation (SD) and confidence interval (CI) are analyzed to quantify model consistency.
Figure 6a,b highlight the accuracy and loss progression over 60 epochs, reflecting the model’s convergence behavior. Figure 6a indicates a consistent convergence rate without overfitting. The convergence behavior reflects the well-optimized training process. The absence of a widening gap between the training and validation losses ensures that the model learns patterns without overfitting or underfitting. Feature normalization, dropout, or weight regularization support the model in reducing overfitting and minimizing the training and validation losses.
Table 3 underscores the model’s state-of-the-art performance in classifying images across datasets. The findings of the DFUC 2020 dataset reinforce the model’s generalization capabilities, maintaining a high accuracy of 97.2% for ischemia and 96.7% for infection. The consistent performance demonstrates the model’s generalization, reliability, and clinical applicability in a real-world environment. The multi-scale representation enhances the potential of the proposed model in determining ischemia and infection features.
Figure 7 displays the comparative results of the proposed model across DFUC 2020 and DFUC 2021 datasets. It highlights the robustness and effectiveness of the proposed model in DFUs classification. The model achieves an average accuracy of 98.7% on DFUC 2021, indicating consistent and improved classification performance. F1-Score of 97.3% (DFUC 2021) and 96.0% (DFUC 2020) reflects balanced precision and recall scores. The optimal feature representation and robust classification model are the factors for the successful DFUs classification.
The authors systematically evaluated the impact of each component on classification performance to validate the effectiveness of the fusion process. Table 4 reveals the findings of the ablation study. The removal of feature extractors leads to a significant drop in accuracy, showing the contribution of local, hierarchical, and hybrid features. The elimination of tensor fusion and usage of direct concatenation reduces accuracy by 3.6%. This indicates that trainable fusion dynamically balances feature contributions effectively. The replacement of KANs with a standard fully concatenated layer results in an accuracy drop of 4.1%, demonstrating the KANs model’s significance.
Table 5 provides a detailed evaluation of the proposed model on the DFUC 2021 dataset. Compared to baseline and individual models, the proposed model achieves the highest accuracy, 98.7%, with a low SD of 0.0004 and a narrow CI of 96.3–97.1. Comparatively, SWIN and Linformer achieve better accuracy. However, the limitations in the feature extraction reduced their overall performance. The pre-trained models were ideal for larger datasets, affecting their recall and specificity on domain-specific datasets. The innovative feature extraction and classification render the proposed model suitable for real-world DFUs classification.
Figure 8 reveals the comparative analysis of the proposed DFUs classification against baseline models. It highlights the superior generalization across standard evaluation metrics. The specificity of over 95% underscores the potential of the proposed model in identifying positive cases. The suggested tensor fusion produced compact and informative features, enhancing generalization to novel data. Regularization strategies prevent model overfitting during training, contributing to successfully classifying ischemia and infection cases. In contrast, the pre-trained models lack advanced attention mechanisms, causing challenges in detecting unique patterns of DFUs.
Table 6 illustrates the findings of the specific classes with respective SHAP values. SHAP 1 offers local spatial details, capturing ulcer textures, color, and edge to the prediction. SHAP 2 presents the hierarchical contextual relationships, including surrounding tissue conditions or ulcer size. SHAP 3 represents the combination of local and global features, exhibiting overlapping characteristics of ischemia and infection. A higher SHAP value for a specific feature region indicates that the feature strongly influences the prediction. In contrast, a negative SHAP value represents that the feature reduces the likelihood of the prediction.
Figure 9a,b demonstrate the model’s ability to classify ischemia and infection cases accurately. The matrices provide insights into the number of correctly and incorrectly classified samples. In clinical settings, false positives (incorrectly classifying a non-severe ulcer as severe) lead to unnecessary interventions, such as aggressive treatments or hospitalization, increasing healthcare costs and patient distress. Likewise, false negatives (failing to detect a severe ulcer) may result in infection, gangrene, and potential amputation. By being trained and validated on clinically evaluated datasets, the model predictions inherently reflect expert medical assessments, rendering it suitable for real-world deployment. The proposed model integrates adaptive fusion, fine-tuned KANs, and SHAP-based interpretability, reducing false positives and negatives. The absence of false negatives indicates the model’s sensitivity, covering critical DFUs cases. The low number of false positives and negatives across datasets presents the model’s balanced performance. The robustness of the proposed EL approach in handling diverse clinical data yielded minimal misclassification rates.
Figure 10 presents the AUROC curves for DFUs classification. The results demonstrate the significance of the model in balancing true positive and false positive rates. The near-perfect AUROC values reflect the remarkable classification performance of the proposed model. The findings guarantee the model’s adaptability and generalization capability. The consistent AUROC values across DFUC 2020 and DFUC 2021 datasets highlight the model’s adaptability to variations, such as differences in image resolution, lighting, and ulcer complexities.
Table 7 presents the effectiveness and superiority of the proposed model compared to existing approaches. The built-in interpretability addresses the limitations in existing models, making it a valuable tool for clinical applications. The AUROC values showcase the exceptional discriminative power of the proposed model. The existing studies [31,32,33,34] obtained lower accuracy due to the limited feature extraction capabilities. Toofanee et al. (2022) [35] achieved an accuracy of 95.0% on the DFUC 2021 dataset. In contrast, the proposed model maintains superior performance across datasets, indicating robust generalization. Previous models, including those presented by Qayyam et al. (2022) [36] and Sarmun et al. (2024) [37], obtained lower F1, demonstrating challenges in handling class imbalances. The dynamic integration of multi-source features and effective modelling of non-linear relationships enables the proposed model to deliver exceptional outcomes. The combination of advanced techniques through the innovative approach supports the proposed model in addressing the shortcomings of existing methods, such as lower performance, lack of robustness, and limited interpretability.

4. Discussion

This study proposes a unique DFUs classification model, improving feature extraction, fusion, and interpretability. The model delivers state-of-the-art performance through the integration of advanced techniques, including MobileNet V3, SWIN Transformer, LeViT-Performer, tensor fusion, and ensemble splines-based KANs. According to the findings of comparative analysis, the suggested techniques improve the model’s resilience, interpretability, and reliability. One of the novel contributions of this study is the hybrid feature extraction framework. The combination of CNNs and ViTs allows the model to capture global and local features. In order to differentiate ischemia and infection, MobileNet V3-SWIN captures subtle patterns, including ulcer margins, granulation, texture patterns, adjacent tissue conditions, and extensive inflammatory regions. LeViT-Performer offers hybrid feature representations integrating the benefits of CNNs and transformers to uncover spatial and global relationships. This multi-source feature extraction addresses the shortcomings of pre-trained models emphasizing either local or global features.
The comparison analysis demonstrates that the suggested model outperforms current methodologies, attaining accuracies of 98.7% on DFUC 2021 and 96.9% on DFUC 2020. These findings demonstrate its capacity to generalize across datasets, guaranteeing reliability in clinical applications. The combination of hybrid feature extraction, adaptive tensor fusion, interpretable and explainable KANs-based classification, and robust generalization strategies enable the proposed model to achieve state-of-the-art classification accuracy, outperforming existing models in terms of accuracy, interpretability, and real-world applicability. The key factor for the successful classification is tensor fusion. Unlike concatenation and early fusion techniques, tensor fusion uses adaptable and trainable weights to integrate the extracted features dynamically. By using this strategy, the model optimizes the hierarchical and local characteristics of DFUs images. This dynamic and efficient fusion technique supports the model’s exceptional performance, illustrating its efficacy in managing unbalanced classes and varied DFUs representations. The standalone CNNs encounter challenges with global context awareness, leading to misclassification in complex DFU cases. In contrast, the application of hybrid feature extraction and KANs-based DFUs classification supported the proposed model to encode hierarchical relationships and long-range dependencies, enhancing DFUs classification.
The use of ensemble splines-based KANs for classification is an additional distinctive characteristic of this research. KANs classification allows the model to approximate decision boundaries accurately. It guarantees clinical relevance and strong classification performance, in contrast to the existing DFUs classification model. Employing adaptive spline functions in KANs improves model interpretability, revealing the impact of individual characteristics on predictions. The interpretability of DFUs classification is crucial in therapeutic applications. Aligning with clinician demands and promoting trust, the model clearly explains its predictions by including SHAP values in the pipeline. The findings of the comparative analysis reinforce the significance of the proposed model’s capability in classifying DFUs images. The proposed model surpassed the existing approaches by achieving superior performance across different evaluation metrics. It can deliver reliable and adaptable real-time solutions for DFUs classification, supporting clinicians in diagnosing and managing DFUs with confidence.
The inference time of the proposed DFUs classification model depends on different factors, including hardware configurations and computational optimization. On a high-performance GPU, the model’s average inference time per image is ~30–50 ms. The inference time increases to ~220–570 ms on a CPU-only system. These outcomes indicate the capability of the model to deliver real-time predictions when deployed on GPU-enabled environments, making it suitable for real-time diagnostic workflows. The model can be integrated into electronic medical records and telemedicine settings, assisting clinicians in rapid DFUs management. Incorporating it on telemedicine platforms can reduce the need for frequent in-person visits. This strategy can enable seamless communications between patients, general practitioners, and wound care specialists, facilitating comprehensive DFUs management.
The model can be hosted on cloud platforms, supporting healthcare providers to upload DFUs images through a secure online portal. Using this approach, individuals in remote locations can receive expert medical opinions without visiting healthcare centers. The deployment of the model in mobile health applications can allow patients to capture DFUs images using their smartphones, leading to early-stage DFUs assessment. The SHAP value interpretability reveals that the model addresses adversarial attacks. The proposed MobileNet V3-SWIN-LeViT-Performer-based feature extraction framework with tensor fusion incorporates several mechanisms that improve the proposed DFUs classification model’s resilience to adversarial perturbations. The combination of local and global features reduces sensitivity to pixel-level perturbations. Reducing redundancy in feature extraction prevents an adversarial attack from misleading the model. The tensor fusion strategy maintains the model’s resilience through the dynamic modification of feature weights. The trainable fusion weights prioritize reliable features when an adversarial attack or noise affects feature channels.
To mitigate potential biases, several key approaches were incorporated into the proposed model. The model was trained and validated on DFUC 2021 and 2020 datasets. The cross-dataset validation indicates the model’s generalizability across multiple settings. The pre-trained weights of MobileNet V3, SWIN, LeViT, and Performer benefit the model from a broader feature representation before fine-tuning using the DFU-specific features. The SHAP-value-based interpretation allows the model to verify whether ulcer predictions were based on clinically relevant areas rather than dataset-specific biases. A direct comparison between the proposed DFUs classification and clinician-based diagnosis can be analyzed in terms of accuracy, consistency, detection of subtle features, and inference time. Clinicians use visual inspection, patient history, and clinical tests to classify DFUs. Subjective clinical evaluations may overlook subtle patterns of ischemic and infection. In addition, clinicians require substantial time to assess an ulcer. In contrast, the proposed model achieves an accuracy of 98.7% using DFUs images in limited inference time, superior to inter-rater variability among human experts.
The authors encountered several challenges during the model development. Integrating diverse features requires dimensionality reduction and batch-wise processing in order to reduce memory overhead. To maintain a balance between local and global features, careful tuning of fusion weights is required. This approach prevents overemphasis on specific feature types. In addition, advanced regularization techniques and data augmentation strategies are essential to handle class imbalances in the DFUs datasets. Implementing the proposed model in a resource-constrained environment may face challenges due to the complexities of the tensor fusion strategy and ensemble splines-based KANs classification. The use of adaptive splines provides a high-level understanding of the decision-making process. This limitation could hinder the adoption of the model in clinical workflows. The reliance on a single modality may not fully capture the complexities of DFUs diagnosis. The dependency on pre-trained models may limit the adaptability to entirely novel datasets or unseen clinical variations. The suggested modular approach may lead to suboptimal performance on highly diverse datasets. Substantial training is required to improve the overall model’s performance. Despite the high classification accuracy, certain misclassifications may occur due to inherent challenges in DFUs imaging. Ischemic ulcers may have subtle color variations, reducing the model’s ability to differentiate the ulcers from normal skin tones. Lighting inconsistencies may obscure vascular deficiencies. This limitation may affect the model in capturing microvascular patterns associated with ischemia detection. The absence of diverse ethnic and demographic representations may reduce the applicability of the proposed model to diverse patient populations. Salam et al. [39] demonstrated the importance of integrating diverse demographic data, including ethnicity and age, into DL models to enhance their robustness and generalizability. The training data encompassing diverse demographic variables can enhance the model’s effectiveness. Using these data, the proposed model can learn to recognize a wider range of ulcer manifestations. This strategy can improve the model’s clinical applicability.
By quantifying the contribution of each feature, SHAP values provide a deeper analysis of feature importance. In a clinical setting, the value highlights specific ulcers, tissue textures, and vascular structures, influencing the predictions of ischemia or infection. The longitudinal DFUC datasets ensure the stability of the proposed model across diverse datasets, improving real-world adoption. To enhance SHAP’s clinical utility, higher-order feature interactions can be incorporated in order to provide context-aware explanations. Using visual heatmaps, clinicians can understand the model’s decision-making process. This transparency can enable model validation, ensuring its alignment with expert diagnostic patterns. Validating the proposed model across diverse clinical settings can guarantee its real-world applicability. By following the findings of Khan et al. [40], the model’s generalizability can be improved. Multi-center data benchmarking can be performed to maintain the model’s performance across different imaging sources. Novel test sets can be used to ensure better generalization. Data privacy is a significant challenge in medical AI development. Asif et al. [41] discussed the significance of the federated learning framework for secure model generalization. The federated learning framework can facilitate secure model generalization. By adopting federated learning, the proposed model can generalize across diverse clinical settings without compromising data privacy.
The authors utilized TensorRT for NVIDIA GPU in order to deploy the proposed model in resource-constrained environments. The utilization of linearized self-attention reduces computational overhead while maintaining feature extraction efficiency. The integration of MobileNet V3-SWIN and LeViT-Performer may pose challenges for deployment in resource-constrained environments, such as mobile devices and edge computing systems. Future research avenues can extend the applicability and interpretability of the proposed model. Developing lightweight architectures for tensor fusion and KANs can optimize computational efficiency. The integration of pruning strategies with the proposed approach can enable model deployment in resource-constrained clinical settings. By applying model pruning, redundant or less significant weights can be systematically removed without affecting performance. Quantization can be used to convert the model parameters from floating-point to lower-bit representations, reducing memory usage and computational costs. Low-rank decompositions can be employed in ViTs architecture to minimize computational complexity. Enriching the dataset with early-stage ulcers and mixed ischemia and infection samples can improve the model’s adaptability. The inclusion of counterfactual explanations or attention heatmaps can enhance the model’s transparency and acceptance in clinical environments.

5. Conclusions

This study introduces a novel and innovative paradigm for treating and managing DFUs. It addresses key challenges in detecting and classifying DFUs severities. The integration of novel feature extraction and classification outperforms existing state-of-the-art approaches. The study findings highlight the strength and adaptability of the proposed model, providing a valuable tool for early and accurate DFUs severity classification. By leveraging SHAP values, the proposed model offers insights into feature importance, allowing clinicians to understand the model’s decision. This transparency strengthens the model and supports healthcare’s demand for explainable healthcare solutions. The generalization performance on the DFUC 2020 dataset demonstrates the potential of the proposed model on unseen data. It underscores the model’s ability to handle the complexities of DFUs classification. However, the model has certain limitations. The dependency on pre-trained models may require additional fine-tuning in real-time implementation. The computational demands of tensor fusion and KANs classification may cause challenges in deploying the proposed model in resource-constrained environments. Future avenues of this research should focus on incorporating multi-modality data to provide a holistic understanding of DFUs severities. Further advancements in explainability can offer deeper insights into the model’s decision-making process, enhancing clinical trust.

Author Contributions

Conceptualization, A.R.W.S. and R.N.; Methodology, A.R.W.S. and R.N.; Software, A.R.W.S. and R.N.; Validation, A.R.W.S. and R.N.; Formal analysis, A.R.W.S. and R.N.; Investigation, A.R.W.S. and R.N.; Resources, A.R.W.S. and R.N.; Data curation, A.R.W.S. and R.N.; Writing—original draft, A.R.W.S.; Writing—review & editing, R.N.; Visualization, A.R.W.S.; Project administration, A.R.W.S.; Funding acquisition, A.R.W.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia (Grant No. KFU251050).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Diabetic Foot Ulcers Dataset is publicly available https://dfu-challenge.github.io/dfuc2021.html, accessed on 22 October 2024.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yap, M.H.; Hachiuma, R.; Alavi, A.; Brüngel, R.; Cassidy, B.; Goyal, M.; Zhu, H.; Rückert, J.; Olshansky, M.; Huang, X.; et al. Deep learning in diabetic foot ulcers detection: A comprehensive evaluation. Comput. Biol. Med. 2021, 135, 104596. [Google Scholar] [CrossRef] [PubMed]
  2. Tulloch, J.; Zamani, R.; Akrami, M. Machine learning in the prevention, diagnosis and management of diabetic foot ulcers: A systematic review. IEEE Access 2020, 8, 198977–199000. [Google Scholar] [CrossRef]
  3. Guan, H.; Wang, Y.; Niu, P.; Zhang, Y.; Zhang, Y.; Miao, R.; Fang, X.; Yin, R.; Zhao, S.; Liu, J.; et al. The role of machine learning in advancing diabetic foot: A review. Front. Endocrinol. 2024, 15, 1325434. [Google Scholar] [CrossRef] [PubMed]
  4. Wang, X.; Zhu, S.; Wang, B.Q.; Li, W.; Gu, S.; Chen, H.; Qin, C.; Dai, Y.; Li, J. Enhancing diabetic foot ulcer prediction with machine learning: A focus on Localized examinations. Heliyon 2024, 10, e37635. [Google Scholar]
  5. Goyal, M.; Reeves, N.D.; Rajbhandari, S.; Yap, M.H. Robust methods for real-time diabetic foot ulcer detection and localization on mobile devices. IEEE J. Biomed. Health Inform. 2018, 23, 1730–1741. [Google Scholar] [CrossRef]
  6. Weatherall, T.; Avsar, P.; Nugent, L.; Moore, Z.; McDermott, J.H.; Sreenan, S.; Wilson, H.; McEvoy, N.L.; Derwin, R.; Chadwick, P.; et al. The impact of machine learning on the prediction of diabetic foot ulcers–a systematic review. J. Tissue Viability 2024, 33, 853–863. [Google Scholar] [CrossRef]
  7. Nanda, R.; Nath, A.; Patel, S.; Mohapatra, E. Machine learning algorithm to evaluate risk factors of diabetic foot ulcers and its severity. Med. Biol. Eng. Comput. 2022, 60, 2349–2357. [Google Scholar] [CrossRef]
  8. Mousa, K.M.; Mousa, F.A.; Mohamed, H.S.; Elsawy, M.M. Prediction of foot ulcers using artificial intelligence for diabetic patients at Cairo university hospital, Egypt. SAGE Open Nurs. 2023, 9, 23779608231185873. [Google Scholar] [CrossRef]
  9. Yazdanpanah, L.; Nasiri, M.; Adarvishi, S. Literature review on the management of diabetic foot ulcer. World J. Diabetes 2015, 6, 13. [Google Scholar] [CrossRef]
  10. Anbarasi, L.J.; Jawahar, M.; Jayakumari, R.B.; Narendra, M.; Ravi, V.; Neeraja, R. An overview of current developments and methods for identifying diabetic foot ulcers: A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2024, 14, e1562. [Google Scholar] [CrossRef]
  11. Das, S.K.; Roy, P.; Singh, P.; Diwakar, M.; Singh, V.; Maurya, A.; Kumar, S.; Kadry, S.; Kim, J. Diabetic foot ulcer identification: A review. Diagnostics 2023, 13, 1998. [Google Scholar] [CrossRef] [PubMed]
  12. Goyal, M.; Reeves, N.D.; Rajbhandari, S.; Ahmad, N.; Wang, C.; Yap, M.H. Recognition of ischaemia and infection in diabetic foot ulcers: Dataset and techniques. Comput. Biol. Med. 2020, 117, 103616. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, J.; Qiu, Y.; Peng, L.; Zhou, Q.; Wang, Z.; Qi, M. A comprehensive review of methods based on deep learning for diabetes-related foot ulcers. Front. Endocrinol. 2022, 13, 945020. [Google Scholar] [CrossRef] [PubMed]
  14. Fadhel, M.A.; Alzubaidi, L.; Gu, Y.; Santamaría, J.; Duan, Y. Real-time diabetic foot ulcer classification based on deep learning & parallel hardware computational tools. Multimedia Tools Appl. 2024, 83, 70369–70394. [Google Scholar]
  15. Khandakar, A.; Chowdhury, M.E.; Reaz, M.B.I.; Ali, S.H.M.; Hasan, M.A.; Kiranyaz, S.; Rahman, T.; Alfkey, R.; Bakar, A.A.A.; Malik, R.A. A machine-learning model for early detection of diabetic foot using thermogram images. Comput. Biol. Med. 2021, 137, 104838. [Google Scholar] [CrossRef]
  16. Biswas, S.; Mostafiz, R.; Uddin, M.S.; Paul, B.K. XAI-FusionNet: Diabetic foot ulcer detection based on multi-scale feature fusion with explainable artificial intelligence. Heliyon 2024, 10, e31228. [Google Scholar] [CrossRef]
  17. Prabhu, M.S.; Verma, S. A deep learning framework and its implementation for diabetic Foot ulcer classification. In Proceedings of the 2021 9th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO), Noida, India, 3–4 September 2021; pp. 1–5. [Google Scholar]
  18. Reyes-Luévano, J.; Guerrero-Viramontes, J.A.; Romo-Andrade, J.R.; Funes-Gallanzi, M. DFU_VIRNet: A novel Visible-InfraRed CNN to improve diabetic foot ulcer classification and early detection of ulcer risk zones. Biomed. Signal Process. Control 2023, 86, 105341. [Google Scholar] [CrossRef]
  19. Sharma, A.; Kaushal, A.; Dogra, K.; Mohana, R. Deep Learning Perspectives for Prediction of Diabetic Foot Ulcers. In Metaverse Applications for Intelligent Healthcare; IGI Global: Hershey, PA, USA, 2024; pp. 203–228. [Google Scholar]
  20. Bansal, N.; Vidyarthi, A. DFootNet: A Domain Adaptive Classification Framework for Diabetic Foot Ulcers Using Dense Neural Network Architecture. Cogn. Comput. 2024, 16, 2511–2527. [Google Scholar] [CrossRef]
  21. Amin, J.; Sharif, M.; Anjum, M.A.; Khan, H.U.; Malik, M.S.A.; Kadry, S. An integrated design for classification and localization of diabetic foot ulcer based on CNN and YOLOv2-DFU models. IEEE Access 2020, 8, 228586–228597. [Google Scholar] [CrossRef]
  22. Diabetic foot Ulcers Dataset. Available online: https://dfu-challenge.github.io/dfuc2021.html (accessed on 22 October 2024).
  23. Koonce, B. MobileNetV3. In Convolutional Neural Networks with Swift for Tensorflow: Image Recognition and Dataset Categorization; O′Reilly Media: Sevastopol, CA, USA, 2021; pp. 125–144. [Google Scholar]
  24. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference On Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  25. Graham, B.; El-Nouby, A.; Touvron, H.; Stock, P.; Joulin, A.; Jégou, H.; Douze, M. Levit: A vision transformer in convnet’s clothing for faster inference. In Proceedings of the IEEE/CVF International Conference On Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 12259–12269. [Google Scholar]
  26. Choromanski, K.; Likhosherstov, V.; Dohan, D.; Song, X.; Gane, A.; Sarlos, T.; Hawkins, P.; Davis, J.; Mohiuddin, A.; Kaiser, L.; et al. Rethinking attention with performers. arXiv 2020, arXiv:2009.14794. [Google Scholar]
  27. Mungoli, N. Adaptive Feature Fusion: Enhancing Generalization in Deep Learning Models. arXiv 2023, arXiv:2304.03290. [Google Scholar]
  28. Wang, T.; Li, M.; Chen, Z.; Chen, X.; Chen, H.; Wu, B.; Wu, X.; Li, Y.; Lu, Y. Tensor based multi-modality fusion for prediction of postoperative early recurrence of single hepatocellular carcinoma. World J. Gastroenterol. 2023, 12465, 1246524. [Google Scholar] [CrossRef]
  29. Liu, Z.; Wang, Y.; Vaidya, S.; Ruehle, F.; Halverson, J.; Soljačić, M.; Hou, T.Y.; Tegmark, M. Kan: Kolmogorov-arnold networks. arXiv 2024, arXiv:2404.19756. [Google Scholar]
  30. Marcílio, W.E.; Eler, D.M. From explanations to feature selection: Assessing SHAP values as feature selection mechanism. In Proceedings of the 2020 33rd SIBGRAPI conference on Graphics, Patterns and Images (SIBGRAPI), Porto de Galinhas, Brazil, 7–10 November 2020; pp. 340–347. [Google Scholar]
  31. Al-Garaawi, N.; Ebsim, R.; Alharan, A.F.; Yap, M.H. Diabetic foot ulcer classification using mapped binary patterns and convolutional neural networks. Comput. Biol. Med. 2022, 140, 105055. [Google Scholar] [CrossRef]
  32. Galdran, A.; Carneiro, G.; Ballester, M.A. Convolutional nets versus vision transformers for diabetic foot ulcer classification. In Diabetic Foot Ulcers Grand Challenge; Springer: Cham, Switzerland, 2022; pp. 21–29. [Google Scholar]
  33. Bloch, L.; Brungel, R.; Friedrich, C.M. Boosting efficientnets ensemble performance via pseudo-labels and synthetic images by pix2pixHD for infection and ischaemia classification in diabetic foot ulcers. In Diabetic Foot Ulcers Grand Challenge; Springer: Cham, Switzerland, 2022; pp. 30–49. 2p. [Google Scholar]
  34. Ahmed, S.; Naveed, H. Bias adjustable activation network for imbalanced data. In Diabetic Foot Ulcers Grand Challenge; Yap, M.H., Cassidy, B., Kendrick, C., Eds.; Springer: Cham, Switzerland, 2022; pp. 50–61. [Google Scholar]
  35. Toofanee, M.S.A.; Dowlut, S.; Hamroun, M.; Tamine, K.; Petit, V.; Duong, A.K.; Sauveron, D. Dfu-Siam a novel diabetic foot ulcer classification with deep learning. IEEE Access 2023, 11, 98315–98332. [Google Scholar] [CrossRef]
  36. Qayyum, A.; Benzinou, A.; Mazher, M.; Meriaudeau, F. Efficient multi-model vision transformer based on feature fusion for classification of DFUC 2021 challenge. In Diabetic Foot Ulcers Grand Challenge; Yap, M.H., Cassidy, B., Kendrick, C., Eds.; Springer: Cham, Switzerland, 2022; pp. 62–75. [Google Scholar]
  37. Sarmun, R.; Chowdhury, M.E.; Murugappan, M.; Aqel, A.; Ezzuddin, M.; Rahman, S.M.; Khandakar, A.; Akter, S.; Alfkey, R.; Hasan, A. Diabetic Foot Ulcer Detection: Combining Deep Learning Models for Improved Localization. Cogn. Comput. 2024, 16, 1413–1431. [Google Scholar] [CrossRef]
  38. Ahsan, M.; Naz, S.; Ahmad, R.; Ehsan, H.; Sikandar, A. A deep learning approach for diabetic foot ulcer classification and recognition. Information 2023, 14, 36. [Google Scholar] [CrossRef]
  39. Salam, A.; Ullah, F.; Amin, F.; Khan, I.A.; Villena, E.G.; Castilla, A.K.; de la Torre, I. Efficient prediction of anticancer peptides through deep learning. PeerJ Comput. Sci. 2024, 10, e2171. [Google Scholar] [CrossRef]
  40. Khan, I.; Akbar, W.; Soomro, A.; Hussain, T.; Khalil, I.; Khan, M.N.; Salam, A. Enhancing Ocular Health Precision: Cataract Detection Using Fundus Images and ResNet-50. IECE Trans. Intell. Syst. 2024, 1, 145–160. [Google Scholar] [CrossRef]
  41. Asif, M.; Naz, S.; Ali, F.; Salam, A.; Amin, F.; Ullah, F.; Alabrah, A. Advanced Zero-Shot Learning (AZSL) Framework for Secure Model Generalization in Federated Learning. IEEE Access 2024, 12, 184393–184407. [Google Scholar] [CrossRef]
Figure 1. The proposed DFUs classification.
Figure 1. The proposed DFUs classification.
Diagnostics 15 00736 g001
Figure 2. (a) Infection and (b) Ischemia.
Figure 2. (a) Infection and (b) Ischemia.
Diagnostics 15 00736 g002
Figure 3. MobileNet V3-SWIN-based feature extraction.
Figure 3. MobileNet V3-SWIN-based feature extraction.
Diagnostics 15 00736 g003
Figure 4. LeViT-Performer-based Feature Extraction.
Figure 4. LeViT-Performer-based Feature Extraction.
Diagnostics 15 00736 g004
Figure 5. Feature fusion approach.
Figure 5. Feature fusion approach.
Diagnostics 15 00736 g005
Figure 6. (a) Training and (b) validation analysis.
Figure 6. (a) Training and (b) validation analysis.
Diagnostics 15 00736 g006
Figure 7. Findings of the performance evaluation.
Figure 7. Findings of the performance evaluation.
Diagnostics 15 00736 g007
Figure 8. Findings of the comparative analysis—DFUC 2020.
Figure 8. Findings of the comparative analysis—DFUC 2020.
Diagnostics 15 00736 g008
Figure 9. Confusion matrix (a) DFUC 2021 (b) DFUC 2020.
Figure 9. Confusion matrix (a) DFUC 2021 (b) DFUC 2020.
Diagnostics 15 00736 g009
Figure 10. Findings of the AUROC analysis.
Figure 10. Findings of the AUROC analysis.
Diagnostics 15 00736 g010
Table 1. Dataset features supporting model’s interpretability.
Table 1. Dataset features supporting model’s interpretability.
Feature NameDescriptionImportance For Model’s Interpretability
Ulcer shapeGeometric structure of the ulcerDistinguishing ischemia from infection
Ulcer textureGranulation tissue pattern, surface roughness, and irregularitiesDifferentiating healthy and ulcer patterns
Ulcer colorRGB distribution of ulcer regionDistinguishing necrotic and infected ulcers
Wound exudatePresence of fluid in the ulcerInfection indicator
Depth informationShallow or deep ulcers based on color gradientsSignificant for assessing ulcer severities
Lighting conditionBrightness, contrast, and exposure variationEnsuring model robustness under clinical image variability
Foot regionForefoot, mid-foot, and heel ulcer regionSignificant for making clinical decisions
Table 2. Computational strategies.
Table 2. Computational strategies.
ModelParametersValues
MobileNet V3- SWINBackboneMobileNet V3
Transformer headSWIN transformer with windowed self attention
Batch size64
OptimizerAdamW
Learning rate 1   ×   10 4
Weight delay 1   ×   10 2
LeViT-PerformerCNN-Transformer fusionLeViT combines CNN’s spatial efficiency with Performer optimized self-attention
Self-attentionLinearized attention
ActivationGELU
Batch size32
OptimizerSGD
Learning rate 5   ×   10 4
Dropout0.3
Tensor fusion-based fusionFeature attention scalingWeighted sum with trainable co-efficient
Feature reductionPCA
KANs-based classificationHyperparameter tuningBOHB
Regularization0.007
Number of hidden nodes64
Learning rate 4.8   ×   10 4
Batch size32
Splines functional order3rd order splines
Output layerSigmoid
Table 3. Performance analysis.
Table 3. Performance analysis.
AccPrecRecF1Spec
DFUC 2021
Ischemia98.697.197.397.296.5
Infection98.997.597.697.596.7
DFUC 2020
Ischemia97.295.896.395.995.9
Infection96.796.196.196.095.8
Table 4. Ablation study (DFUC 2021).
Table 4. Ablation study (DFUC 2021).
Model VariantAccPrecRecF1Spec
Proposed Model98.797.397.497.396.6
Without LeviT-Performer and Fusion (MobileNet V3-SWIN + KANs)96.494.594.994.794.1
Without MobileNet V3-SWIN and Fusion (LeViT-Performer + KANs)95.893.994.294.093.7
Without Tensor Fusion (MobileNet V3-SWIN + LeViT-Performer + Concatenation + KANs)95.193.192.893.092.1
Without KANs (MobileNet V3-SWIN + LeViT-Performer + Fusion + Fully Connected Layer)94.792.592.392.491.8
Table 5. Comparative analysis—DFUC 2021.
Table 5. Comparative analysis—DFUC 2021.
AccPrecRecF1SpecSDCI
Proposed Model98.797.397.497.396.60.0004[96.3–97.1]
SWIN93.893.493.793.593.90.0003[96.1–96.8]
MobileNet V394.994.194.394.294.10.0005[95.3–95.9]
RegNet X92.191.390.891.090.20.0005[95.8–96.7]
LeViT93.792.491.992.191.40.0005[96.1–97.4]
Performer93.192.993.493.193.10.0003[96.3–96.7]
Linformer94.593.694.193.893.90.0003[95.9–96.4]
Table 6. Sample outcomes.
Table 6. Sample outcomes.
Input ImageGround TruthPredictionSHAP ValueClinical Relevance
Diagnostics 15 00736 i001InfectionInfection0.35, 0.45, and 0.20Highlighted regions indicate redness, swelling, and potential pus, leading to infection-specific patterns.
Diagnostics 15 00736 i002InfectionInfection0.40, 0.50, and 0.10Yellowish discharge and wound edges are the crucial indicators of infection. In addition, swelling and wound patterns influenced the prediction.
Diagnostics 15 00736 i003InfectionInfection0.38, 0.47, and 0.15Central ulcer discharge and edge discoloration indicate a bacterial infection.
Diagnostics 15 00736 i004IschemiaIschemia0.45, 0.40, 0.18Dark necrotic tissue and reduced blood supply areas confirm ischemic interpretation.
Diagnostics 15 00736 i005IschemiaIschemia0.50, 0.35, and 0.11Yellowish dry tissue with no surrounding swelling, indicating ischemia diagnosis.
Diagnostics 15 00736 i006IschemiaIschemia0.42, 0.16, and 0.37 Shrunken tissue regions represent the ischemia patterns.
Table 7. Findings of the comparative analysis.
Table 7. Findings of the comparative analysis.
ModelDatasetFeature ExtractionClassificationInterpretabilityPerformance
Proposed Model DFUC 2021 [22]MobileNet V3-SWIN and LeViT-PerformerEnsembled Splines-based KANsAcc: 98.7%
Prec: 97.3%
Rec: 97.4%
F1: 97.3%
Spec: 96.6%
AUROC: 0.93
Proposed Model DFUC 2020 [22]Acc: 96.9%
Prec: 95.9%
Rec: 96.2%
F1: 96.0%
Spec: 95.8%
AUROC: 0.91
AlGarawi et al. (2022) [31]DFUC 2020 [22]Customized CNNsFully connected layer×Average Acc: 86.7%
Average F1: 86.7%
AUC: 0.90
Galdran et al. (2022) [32]DFUC 2020 [22]ResNeXt 50—EfficientNet-ViT-based DeiTFully connected layer×Average Prec: 70.0%
Average Rec: 68.5%
Average F1: 75.7%
AUC: 0.88
Bloch et al. (2022) [33]DFUC 2021 [22]EfficientNet B0-B6Fully connected layer×Average F1: 53.5%
Average Prec: 68.5%
Average Rec: 66.5%
AUC: 0.86
Ahmed and Naveed (2022) [34]DFUC 2021 [22]EfficientNet B0-B6—ResNet 50Fully connected layer×Average F1: 53.9%
Average Prec: 67.3%
Average Rec: 67.1%
AUC: 0.59
Toofanee et al. (2023) [35]DFUC 2021 [22]EfficientNet-BeiTKNN-based classification×Acc: 95.0%
AUC: 0.8298
Prec: 93.9%
Rec: 93.9%
F1: 93.7%
Qayyam et al. (2022) [36]DFUC 2021 [22]Customized ViTsFully connected layer×Average F1: 43.47%
Average Prec: 68.0%
Average Rec: 66.3%
AUC: 0.84
Sarmun et al. (2024) [37]DFUC 2021
[22]
YOLO V8XFully connected layer×Prec: 89.7%
Rec: 74.0%
F1: 81.1%
YOLO V8X + ResNetEL approach×Prec: 83.5%
Rec: 75.2%
F1: 79.1%
Ahsan et al. (2023) [38]DFUC 2020
[22]
ResNet 50Fully connected layer×Average Acc: 92.12%
Average F1: 92.24%
AUC: 0.87
Note: ✓ (Presence of Model’s interpretability) and × (Absence of Model’s interpretability).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sait, A.R.W.; Nagaraj, R. Diabetic Foot Ulcers Detection Model Using a Hybrid Convolutional Neural Networks–Vision Transformers. Diagnostics 2025, 15, 736. https://doi.org/10.3390/diagnostics15060736

AMA Style

Sait ARW, Nagaraj R. Diabetic Foot Ulcers Detection Model Using a Hybrid Convolutional Neural Networks–Vision Transformers. Diagnostics. 2025; 15(6):736. https://doi.org/10.3390/diagnostics15060736

Chicago/Turabian Style

Sait, Abdul Rahaman Wahab, and Ramprasad Nagaraj. 2025. "Diabetic Foot Ulcers Detection Model Using a Hybrid Convolutional Neural Networks–Vision Transformers" Diagnostics 15, no. 6: 736. https://doi.org/10.3390/diagnostics15060736

APA Style

Sait, A. R. W., & Nagaraj, R. (2025). Diabetic Foot Ulcers Detection Model Using a Hybrid Convolutional Neural Networks–Vision Transformers. Diagnostics, 15(6), 736. https://doi.org/10.3390/diagnostics15060736

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop