Next Article in Journal
Evaluating the Effect of Bile Acid Levels on Maternal and Perinatal Outcomes in Intrahepatic Cholestasis of Pregnancy: A Retrospective Study
Previous Article in Journal
The Relationship Between OCT and VEP Parameters with Disability and Disease Duration in Relapsing–Remitting Multiple Sclerosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interpreting Venous and Arterial Ulcer Images Through the Grad-CAM Lens: Insights and Implications in CNN-Based Wound Image Classification

1
Department of Plastic, Hand, and Reconstructive Surgery, University Hospital Regensburg, 93053 Regensburg, Germany
2
Ambient Assisted Living & Medical Assistance Systems, Department of Computer Science, University of Bayreuth, 95447 Bayreuth, Germany
3
Machine Learning and Data Analytics Lab (MaD Lab), Department Artificial Intelligence in Biomedical Engineering (AIBE), Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), 91052 Erlangen, Germany
4
Imaging Solutions, PreSens Precision Sensing GmbH, 93053 Regensburg, Germany
5
Department of Dermatology, University Hospital Regensburg, 93053 Regensburg, Germany
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(17), 2184; https://doi.org/10.3390/diagnostics15172184
Submission received: 4 July 2025 / Revised: 1 August 2025 / Accepted: 25 August 2025 / Published: 28 August 2025
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

Background/Objectives: Chronic wounds of the lower extremities, particularly arterial and venous ulcers, represent a significant and costly challenge in medical care. To assist in differential diagnosis, we aim to evaluate various advanced deep-learning models for classifying arterial and venous ulcers and visualize their decision-making processes. Methods: A retrospective dataset of 607 images (198 arterial and 409 venous ulcers) was used to train five convolutional neural networks: ResNet50, ResNeXt50, ConvNeXt, EfficientNetB2, and EfficientNetV2. Model performance was assessed using accuracy, precision, recall, F1-score, and ROC-AUC. Grad-CAM was applied to visualize image regions contributing to classification decisions. Results: The models demonstrated high classification performance, with accuracy ranging from 72% (ConvNeXt) to 98% (ResNeXt50). Precision and recall values indicated strong discrimination between arterial and venous ulcers, with EfficientNetV2 achieving the highest precision. Conclusions: AI-assisted classification of venous and arterial ulcers offers a valuable method for enhancing diagnostic efficiency.

1. Introduction

Chronic wounds, particularly venous and arterial ulcers of the lower extremities, represent a global problem with a prevalence estimated at between 0.1% and 2% in the general population [1,2]. As society continues to age and the obesity rate increases, vascular ulcers are also anticipated to rise because of obesity-associated comorbidities. Timely diagnosis of this condition is crucial for selecting the appropriate therapy and preventing further progression of the wound, as well as avoiding amputation of the limb, which is a not uncommon complication of this wound type [3]. In general, arterial and venous ulcers have numerous classic features that help with differential diagnosis [4]. The cause of an arterial ulcer is peripheral arterial disease (PAD), i.e., the arteries supplying the affected extremity are either stenosed or completely occluded. Due to this reduced blood flow, minor traumas heal poorly and thus lead to the classic clinical picture of a well-defined ulcer with deep (“punched-out”) wound edges. Due to tissue hypoxia, the wounds tend to be dry and often show black necrotic tissue or exposed tendons, suggesting a non-venous condition. In contrast, venous ulcers tend to be more irregular in shape and have a rather moist wound environment. While venous ulcers often occur above or behind the medial malleolus, arterial ulcers tend to occur on the edge of the tibia, above the lateral malleolus, or on the feet. The surrounding skin also differs (Table 1).
Clinical signs, patient history, and vascular imaging are recommended to determine the etiology of the ulcer. The primary tests for diagnosing arterial and venous ulcers are the ankle-brachial pressure index (ABPI), duplex ultrasonography, and color flow Doppler imaging (CFDI) [5,6]. Further examinations, such as venography, angiography, microbiological and/or histological examination of a biopsy, may be necessary in some cases to clarify the etiology of the ulcer. While this offers essential insights into the etiology of the wound, it is also time-consuming, subjective, and requires specialized training, delaying proper therapy beyond local wound care. Furthermore, not all care facilities have access to comprehensive diagnostic tools like duplex ultrasound. In such settings, AI-based wound classification could offer a quick and reliable way to determine the cause, supported by clear visual explanations such as Grad-CAM. This could, for example, enable early identification of patients likely to have venous ulcers, allowing them to benefit immediately from compression therapy. Therefore, AI could serve as a practical decision-support tool that helps clinicians in the diagnostic process and may speed up the initiation of appropriate treatment.
Artificial intelligence (AI), Machine learning (ML), and Deep learning (DL) have revolutionized medical image analysis. For example, in dermatology, the diagnosis and risk assessment of skin changes in preventive care are automated and specific [7]. In radiological imaging, AI can enhance image quality [8] and help detect and classify anomalies [9,10]. DL, a subfield of ML, employs deep neural networks to identify patterns. Convolutional neural networks, a prominent area in DL, effectively process two-dimensional input data.
Recent studies increasingly move from single-word outputs toward paragraph-level report generation, enabled by vision-language models that couple strong visual backbones with text decoders. These systems highlight clinically aligned evaluation and factuality and outline clear pathways from robust classifiers to structured findings and auto-drafted reports [11,12,13]. In this context, our focus on reliable ulcer classification serves as a practical building block for future report-level outputs.
Deep convolutional neural networks (DCNNs) are a variant of convolutional neural networks (CNNs) that can recognize relevant image details and more complicated patterns in the analysis of images. In wound assessment, there are various approaches for identifying wound boundaries (localization or segmentation) and classifying different wound types [14,15] and tissue types (e.g., granulation, slough, and eschar) [14] using DL.
So far, however, the classification has been broader between different wound types, such as surgical wounds, venous leg ulcers, and pressure ulcers [14]. A decision support system for a detailed distinction between venous and arterial ulcers could contribute to a faster diagnosis and, thus, earlier targeted therapy. Additionally, employing explainable AI (XAI) modeling can facilitate its integration and acceptance within healthcare systems [16,17].
Our study aims to apply and evaluate various state-of-the-art deep learning models for the image-based classification of vascular ulcers, specifically distinguishing between venous and arterial ulcers. Beyond classification performance, our clinical focus lies in visualizing and interpreting the decision-making process of the model using explainable AI (XAI) techniques such as Gradient-weighted Class Activation Mapping (Grad-CAM), to ensure transparency and facilitate clinical trust and integration.

2. Materials and Methods

2.1. Patients

We utilized the hospital database from University Hospital Regensburg (Germany) to retrospectively filter all patients by the relevant ICD codes (International Statistical Classification of Diseases and Related Health Problems) for chronic ulcers of the lower extremities. Next, we manually verified their diagnoses in the patient charts based on vascular assessments (e.g., Doppler ultrasound, duplex ultrasound, ankle-brachial pressure index), identifying 72 patients with purely arterial or purely venous ulcers of the lower extremities. We included only those with standardized photographic wound documentation in the study (see Table 2).
The final dataset included 607 images in .jpg format (198 arterial and 409 venous ulcer images). The images were of varying sizes, with the average height and average width being 833 and 742 pixels, respectively. There was a repetition of patients with wound images taken at various stages of wound healing. However, no wound in its form was repeated twice. Images from the same patients were taken on different dates. Example images of unprocessed wound documentation for each ulcer class are shown in Figure 1. The images were randomly selected and represent the original format without any preprocessing.

2.2. Preprocessing

To enhance the focus on the region of interest (ROI), all images were manually cropped by delineating the wound and applying a 10% offset on all sides. This process effectively removed identifiable patient information while retaining a narrow margin of surrounding skin to preserve clinically relevant context. By eliminating unnecessary background information, we ensured that the analysis remained focused on the wound.
Given the color-dependent nature of wound images and their capture under controlled conditions, we employed a data augmentation process designed to enhance dataset variety without increasing its size, thereby supporting model robustness. We opted for weak augmentations to maintain the integrity of the visual data. Each image was resized to 224 × 224 pixels and normalized using a mean of 0.5 and a standard deviation of 0.5. During training, augmentations such as horizontal flip, vertical flip, and random cropping were applied with a 10% probability for each transformation. These augmentations were implemented dynamically at runtime, meaning they were randomly applied each time an image was accessed, modifying the sample without creating additional entries in the dataset. It is important to note that these augmentations, except for normalization and resizing, were not applied to the validation set or during inference.
A stratified train-test split was performed, with 10% of the dataset allocated to the test set to preserve class distribution. This split was held constant across all model training pipelines to ensure comparability of results.

2.3. Models

State-of-the-art deep learning models for image classification, namely ResNet50 [16], ResNeXt50 [18], ConvNeXt [19], EfficientNet B2 [20], and EfficientNet V2 [21], were selected for the task of ulcer classification, specifically to distinguish between venous and arterial ulcers. These models were chosen due to their proven performance in medical imaging and their architectural diversity, which allows for a comprehensive evaluation of classification performance across different CNN design principles. Each model was trained independently on the training dataset.
ResNet50 [20] is a deep convolutional neural network with 50 layers, introduced as part of the ResNet family, which uses residual learning to enable the effective training of very deep architectures. Its key innovation is the use of skip connections that allow gradients to flow more easily through the network, alleviating the vanishing gradient problem common in deep networks. This architecture enables ResNet50 to achieve high accuracy on image classification tasks while being more efficient and stable to train than traditional deep CNNs. ResNet50 has been widely used in applications ranging from transfer learning in general vision tasks [20] to specialized domains such as COVID-19 detection from chest X-rays [22] and skin cancer diagnosis from dermoscopy images [23].
ResNeXt50 [18] is a deep convolutional neural network that builds on the ResNet architecture by introducing the concept of cardinality, the number of parallel paths within a residual block, allowing it to capture more diverse feature representations without greatly increasing complexity. Instead of simply increasing depth or width, ResNeXt uses grouped convolutions to achieve a better trade-off between computational efficiency and model capacity. This design makes ResNeXt50 more accurate and flexible than standard ResNet50, particularly on large-scale image classification tasks like ImageNet. The model has been successfully applied in various classification problems, including skin lesion analysis and plant disease detection [24].
ConvNeXt [19] is a modern CNN architecture that updates the classic ResNet backbone using design strategies borrowed from Vision Transformers, such as large kernel sizes, LayerNorm, inverted bottlenecks, GELU activations, and the AdamW optimizer. These refinements help ConvNeXt achieve state-of-the-art performance on benchmarks like ImageNet while maintaining the efficiency and inductive bias of convolutional models. It has demonstrated its performance on a variety of tasks like detecting periapical lesions in radiographs [25], detecting vascular leukoencephalopathy from CT images [26], and domain-specific tasks such as rice grain type classification [27]. ConvNeXt variants. ConvNeXt is released in multiple capacity levels: Tiny (T), Small (S), Base (B), Large (L), and X-Large (XL). We use the Tiny (T) variant as our default backbone throughout this work.
EfficientNet-B2 [20] is a mid-sized member of the EfficientNet family introduced by Google AI, featuring compound scaling of network depth, width, and input resolution to optimize both accuracy and efficiency. The architecture uses MobileNetV2-style MBConv blocks [28] and was discovered via neural architecture search, making EfficientNet-B2 particularly efficient in parameter usage and FLOPs while delivering strong classification performance. EfficientNetB2 is known to outperform comparative models on various tasks like skin cancer classification and COVID-19 detection from ultrasound images [29].
EfficientNet-V2 [21] extends the original EfficientNet family by introducing several training- and architecture-level enhancements aimed at improving both convergence speed and predictive performance. It incorporates Fused-MBConv blocks, which combine depthwise and regular convolutions in early stages, and uses progressive learning—gradually increasing image resolution and regularization strength during training—to accelerate convergence and improve generalization. These innovations result in significantly faster training times (5×–11×) and higher accuracy, with EfficientNet-V2 models achieving up to 87.3% top-1 accuracy on ImageNet while maintaining parameter efficiency. EfficientNet-V2 variants have demonstrated superior performance across a range of classification tasks, including acute lymphoblastic leukemia detection [30], and liver fibrosis staging from MRI [31].
The architectures of the models were downloaded from Hugging Face in the configurations detailed in Table 3. To ensure a fair comparison and facilitate reproducibility, all deep-learning classifiers were trained using a unified protocol summarized in Table 4. A batch size of 32 and an initial learning rate of 1 × 10−4 were used, and models were trained for 50 epochs. Random weight initialization was carried out. We evaluated their performance using standard classification metrics, as described below. To explore model interpretability and better understand the decision-making process, we also applied Grad CAM to visualize the image regions most influential in each model’s predictions.

2.4. Performance Evaluation

In evaluating the performance of our Convolutional Neural Network (CNN) models for wound image classification, we employed several key metrics to ensure a comprehensive assessment. Accuracy was used as a primary indicator, measuring the proportion of correctly classified images out of the total number of images. This metric provides a straightforward measure of overall model performance. However, given the potential imbalance in wound image datasets, we also calculated Precision, which focuses on the correctness of positive predictions. Precision is particularly important in medical applications, as it reflects the model’s ability to minimize false positives, ensuring that when a wound is identified, it is indeed that wound type.
To further evaluate the model’s effectiveness, we incorporated the F1 Score, which balances Precision and Recall, the ability to find all relevant instances. The F1 Score is crucial in scenarios where both false positives and false negatives are costly, providing a single metric that captures the trade-off between these two aspects. To address class imbalance, we employed the ROC-AUC metric, which effectively evaluates the ability of the model to differentiate between classes by considering the trade-off between true positive and false positive rates. This metric is especially useful for imbalanced datasets, providing a comprehensive assessment of the model’s performance across all classes. Additionally, we utilized Gradient weighted Class Activation Mapping, Grad CAM, to enhance the explainability of our model’s decisions. Grad CAM generates visual explanations that highlight important regions in the image that contribute to the model’s prediction, offering insights into how the CNN interprets wound features. This approach not only aids in understanding model behavior but also builds trust in the model’s outputs, which is essential for clinical adoption.

3. Results

3.1. Performance Metrics

In our evaluation of various Convolutional Neural Network (CNN) architectures for wound image classification, we observed distinct performance metrics across different models. Results highlighted in Table 4 showcase the effectiveness of advanced CNN architectures in accurately classifying wound images. While there is room for improvement with the ConvNeXt model, the models from the EfficientNet family have exhibited superior performance even though they are relatively lightweight. The deeper models ResNet50 and ResNeXt50 also exhibited performance on par with EfficientNet family models.
Further analysis of precision and recall for all five models confirms strong performance across classes (Table 5)

3.2. Deep Learning Interpretability with Gradient-Weighted Class Activation Mapping (Grad-CAM), Precision Venous (%) and Precision Arterial (%) Refer to the Accuracy of the Model in Identifying Venous or Arterial Wounds Among All Its Predictions for Each Category

The heatmaps generated by Grad-CAM are displayed in the figures below. The color ranges from blue to red and represents the significance value of the region, ranging from low (blue) to high (red), respectively, for predicting the etiology of the ulcer. Figure 2A illustrates the results for venous ulcers, and Figure 2B displays the model outputs for arterial ulcers.
As can be seen, only the models from the EfficientNet family have demonstrated dynamic focus, examining different parts of the images rather than fixating on the center. This dynamic attention suggests that these models are better equipped to adapt to variations in wound appearances, potentially contributing to their higher performance metrics. The insights gained from GradCAM underscore the importance of interpretability in medical image analysis, as understanding model behavior builds trust and aids in clinical decision-making.

4. Discussion

Our study demonstrates the significant potential of deep learning-based classification models for the differential diagnosis of arterial and venous ulcers. In particular, the EfficientNetV2 and ResNeXt50 CNN models exhibited a highly accurate differentiation between these two wound categories (97% and 98% accuracies). Huang et al. [33] presented a CNN model that classified five wound types (including arterial and venous ulcer wounds), achieving 96% and 86% accuracy in the venous and arterial ulcer classification tasks with an AUC (area under the curve) of 0.924 and 0.897, respectively. Although that impressive venous ulcer accuracy was based on only 26 venous wound images (with a total dataset of 2149 images), their five-class distinction was a much more complex task than our binary classification. Interestingly, their model outperformed the medical personnel significantly in all five wound categories [33]. Other binary classification studies [34,35] for distinct wound types or wound complications (e.g., maceration [35] using a CNN-based method similar to ours) can yield high accuracies if domain-appropriate data preprocessing is combined with modern CNN architectures. Patel et al. [36] used a multi-modal approach, demonstrating that including location data (exact lower leg region of the wound) can enhance performance in multi-class scenarios.
To make CNN-based models more transparent to explain AI and visualize relevant regions for CNN-based decision-making, we used Gradient-weighted Class Activation Mapping (Grad-CAM) [37]. Explainable AI (XAI) is undoubtedly a key aspect on the path to implementing AI in clinical routine, as it provides an intuitive insight into the decision-making process of neural networks. Visualization via Grad-CAM, in particular, makes it easier to understand why the AI assumes a venous or arterial etiology, for example. However, we have not conducted a direct comparison with experienced clinicians (e.g., dermatologists). In everyday practice, physicians do not make their decisions solely based on a “heat map.” They consider other clinical factors (e.g., Table 1) and assess the patient holistically. Pure image-based diagnosis with Grad-CAM analysis seems almost “limited” compared to this multifactorial clinical routine, but at the same time, it is even more impressive, as it achieves high levels of accuracy without additional information. Therefore, when faced with an image classification problem, a natural question arises about whether the model is genuinely identifying the etiology of the wound. A good visual explanation should be both class-discriminative (i.e., localizing the relevant regions necessary for a prediction) and high-resolution (i.e., finding the fine details in an image) [38]. Our analysis of the Grad-CAM results indicated that certain models, such as EfficientNetV2, dynamically concentrated on the ulcer borders and surrounding tissue, suggesting they incorporate contextual features like “punched-out” wound borders or skin discoloration in their decision-making. In contrast, other models, such as ResNeXt50, focused on the central wound bed, possibly highlighting ulcer depth and tissue composition. While Grad-CAM and LIME (Locally Interpretable Explanations and Model-Independent) [16] are qualitative heat-map approaches to highlight regions of the image that strongly influence the final prediction, Lo et al. found SHAP (SHapley Additive exPlanation) [17,39], the most effective and more detailed for providing quantitative per-pixel importance scores [14]. Therefore, the use of qualitative and quantitative XAI techniques to interpret deep learning models is a crucial part of validating and trusting these models in clinical practice [40].
Because our data was collected from a single-center cohort, future research should collect larger, multicenter datasets that include diverse imaging conditions (lighting, camera angle, background, distance) and patient demographics to improve generalizability and robustness. Additionally, we only considered the simplest scenario, binary classification of arterial versus venous ulcers. Real-world wound care involves distinguishing among mixed arterial–venous ulcers, diabetic foot ulcers, pressure ulcers, and other causes. Therefore, expanding and balancing the label set will be crucial. Furthermore, our analysis involved manually cropping images prior to processing, based on the assumption that images would primarily contain the wound and some immediate background. This approach is supported by existing literature and algorithms capable of automatic wound detection and cropping [15,41]. Additionally, we employed EfficientNet in its standard configuration to establish a robust and reproducible baseline for comparison, focusing on the effects of dataset quality, preprocessing, and augmentation without introducing architecture-specific confounders. Although attention modules like Squeeze-and-Excitation (SE) blocks [42] could improve feature weighting, they also add additional parameters and complexity, which may lead to overfitting—particularly with our small and unbalanced dataset. Finally, our models relied solely on wound photographs, excluding additional clinical data such as lesion localization, vascular imaging, patient characteristics, ankle–brachial index, or laboratory markers. Incorporating these data streams would provide richer context and likely enhance diagnostic accuracy. Addressing these limitations will help develop more accurate, robust, and clinically useful decision-support systems for ulcer assessment.

5. Conclusions

Chronic vascular ulcers impose a significant clinical and economic burden, and timely, precise differentiation among various etiologies is crucial for selecting effective treatment. This study aimed to determine whether modern convolutional neural networks can provide reliable, interpretable image-based support for this decision. We evaluated five CNNs (ResNet50, ResNeXt50, ConvNeXt, EfficientNetB2, and EfficientNetV2) for image-based differentiation of venous and arterial ulcers and used Grad-CAM to visualize the reasoning of the model. ResNeXt50 performed best with a macro-average accuracy of 97.85% and ROC-AUC of 0.9966, followed by EfficientNetV2 (96.77%) and ResNet50/EfficientNetB2 (95.70%). ConvNeXt (Tiny) lagged behind at 72.04%. Grad-CAM revealed model-specific attention patterns consistent with our metrics. The EfficientNet models displayed dynamic, context-aware focus, shifting between ulcer borders, perilesional skin changes, and the wound bed, while ResNext50 tended to fixate mainly on the center of the wound bed. This behavior aligns with the highest precision of EfficientNetV2 and supports the clinical plausibility of its decisions. These findings indicate that lightweight, explainable classifiers can support decision-making in the clinical setting. Limitations include a retrospective, single-center dataset of 607 images and a binary classification task. Future work will expand to multicenter and multiclass settings, incorporate segmentation, localization, and conduct prospective external validation.

Author Contributions

Conceptualization, H.N., S.K. (Sally Kempa), N.V.S.J.J., R.J.M., G.L. and H.L.; methodology, O.F., N.V.S.J.J. and H.L.; validation, N.V.S.J.J., H.N. and H.L.; formal analysis, N.V.S.J.J., H.N., S.K. (Silvan Klein) and H.L.; investigation, H.N. and N.V.S.J.J.; resources, S.K. (Sally Kempa), L.P., S.S. and M.B.; data curation, H.N., N.V.S.J.J. and S.K. (Sally Kempa); writing—original draft preparation, H.N., N.V.S.J.J. and S.K. (Sally Kempa); writing—review and editing, S.K. (Sally Kempa), S.K. (Silvan Klein)., S.S., M.B., L.P., H.L., R.J.M., G.L. and O.F.; visualization, N.V.S.J.J., H.N.; supervision, S.K. (Sally Kempa)., H.L. and L.P.; project administration, R.J.M., G.L. and H.L.; funding acquisition, L.P., H.L., O.F., G.L. and R.J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Bayerische Forschungsstiftung (AZ 1582-23).

Institutional Review Board Statement

Formal and documented ethical approval was obtained (reference number 23-3562-104, University of Regensburg, 2 November 2023).

Informed Consent Statement

Patient consent was waived due to the retrospective nature of the study and the use of fully anonymized image data.

Data Availability Statement

For data supporting reported results, please contact the corresponding author.

Conflicts of Interest

Authors Robert Johannes Meier, Gregor Liebsch were employed by PreSens Precision Sensing GmbH. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
CNNConvolutional neural networks
DCNNDeep convolutional neural networks
Grad-CAMGradient-weighted Class Activation Mapping
PADPeripheral artery disease
CVIChronic venous insufficiency

References

  1. O’Brien, J.F.; Grace, P.A.; Perry, I.J.; Burke, P.E. Prevalence and aetiology of leg ulcers in Ireland. Ir. J. Med. Sci. 2000, 169, 110–112. [Google Scholar] [CrossRef]
  2. Wong, M.; Daly, S.; Katanas, J.; Shea, H.; Upton, L.; Puyk, K.; Thayaparan, A.; Kern, J.S.; Gibbs, H.; Aung, A.K. Putting skin in the game: A descriptive study of lower extremity ulcers in general medical inpatients. Intern. Med. J. 2024, 54, 2067–2071. [Google Scholar] [CrossRef] [PubMed]
  3. Star, A. Differentiating Lower Extremity Wounds: Arterial, Venous, Neurotrophic. Semin. Interv. Radiol 2018, 35, 399–405. [Google Scholar] [CrossRef]
  4. Grey, J.E.; Harding, K.G.; Enoch, S. Venous and arterial leg ulcers. BMJ 2006, 332, 347–350. [Google Scholar] [CrossRef] [PubMed]
  5. Pannier, F.; Eberhard, R. Differential diagnosis of leg ulcers. Phlebology 2013, 28, 55–60. [Google Scholar] [CrossRef]
  6. Lazarides, M.K.; Athanasios, D.G. The role of hemodynamic measurements in the management of venous and ischemic ulcers. Int. J. Low. Extrem. Wounds 2007, 6, 254–261. [Google Scholar] [CrossRef] [PubMed]
  7. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  8. Liu, F.; Jang, H.; Kijowski, R.; Bradshaw, T.; McMillan, A.B. Deep Learning MR Imaging-based Attenuation Correction for PET/MR Imaging. Radiology 2018, 286, 676–684. [Google Scholar] [CrossRef]
  9. Lakhani, P.; Sundaram, B. Deep Learning at Chest Radiography: Automated Classification of Pulmonary Tuberculosis by Using Convolutional Neural Networks. Radiology 2017, 284, 574–582. [Google Scholar] [CrossRef]
  10. Schaffter, T.; Buist, D.S.M.; Lee, C.I.; Nikulin, Y.; Ribli, D.; Guan, Y.; Lotter, W.; Jie, Z.; Du, H.; Wang, S.; et al. Evaluation of Combined Artificial Intelligence and Radiologist Assessment to Interpret Screening Mammograms. JAMA Netw. Open 2020, 3, e200265. [Google Scholar] [CrossRef]
  11. Uçan, M.; Kaya, B.; Kaya, M. Generating Medical Reports With a Novel Deep Learning Architecture. Int. J. Imaging Syst. Technol. 2025, 35, e70062. [Google Scholar] [CrossRef]
  12. Tanno, R.; Barrett, D.G.T.; Sellergren, A.; Ghaisas, S.; Dathathri, S.; See, A.; Welbl, J.; Lau, C.; Tu, T.; Azizi, S.; et al. Collaboration between clinicians and vision-language models in radiology report generation. Nat. Med. 2025, 31, 599–608. [Google Scholar] [CrossRef]
  13. Wang, X.; Figueredo, G.; Li, R.; Zhang, W.E.; Chen, W.; Chen, X. A survey of deep-learning-based radiology report generation using multimodal inputs. Med. Image Anal. 2025, 103, 103627. [Google Scholar] [CrossRef]
  14. Lo, Z.J.; Mak, M.H.W.; Liang, S.; Chan, Y.M.; Goh, C.C.; Lai, T.; Tan, A.; Thng, P.; Rodriguez, J.; Weyde, T.; et al. Development of an explainable artificial intelligence model for Asian vascular wound images. Int. Wound J. 2023, 21, e14565. [Google Scholar] [CrossRef] [PubMed]
  15. Anisuzzaman, D.M.; Patel, Y.; Rostami, B.; Niezgoda, J.; Gopalakrishnan, S.; Yu, Z. Multi-modal wound classification using wound image and location by dee p neural network. Sci. Rep. 2022, 12, 20057. [Google Scholar] [CrossRef] [PubMed]
  16. Sarp, S.; Kuzlu, M.; Wilson, E.; Cali, U.; Guler, O. The Enlightening Role of Explainable Artificial Intelligence in Chronic Wound Classification. Electronics 2021, 10, 1406. [Google Scholar] [CrossRef]
  17. Srinivasu, P.N.; Sandhya, N.; Jhaveri, R.H.; Raut, R. From Blackbox to Explainable AI in Healthcare: Existing Tools and Case Studies. Mob. Inf. Syst. 2022, 2022, 8167821. [Google Scholar] [CrossRef]
  18. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
  19. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11966–11976. [Google Scholar]
  20. Mingxing, T.; Quoc, V.L. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2019, arXiv:1905.11946. [Google Scholar]
  21. Mingxing, T.; Quoc, V.L. EfficientNetV2: Smaller Models and Faster Training. arXiv 2021, arXiv:2104.00298. [Google Scholar]
  22. Alam, N.-A.; Ahsan, M.; Based, M.A.; Haider, J.; Kowalski, M. COVID-19 Detection from Chest X-ray Images Using Feature Fusion and Deep Learning. Sensors 2021, 21, 1480. [Google Scholar] [CrossRef]
  23. Zhang, N.; Cai, Y.-X.; Wang, Y.-Y.; Tian, Y.-T.; Wang, X.-L.; Badami, B. Skin cancer diagnosis based on optimized convolutional neural network. Artif. Intell. Med. 2020, 102, 101756. [Google Scholar] [CrossRef]
  24. Tanwar, S.; Singh, J. ResNext50 based convolution neural network-long short term memory model for plant disease classification. Multimed. Tools Appl. 2023, 82, 29527–29545. [Google Scholar] [CrossRef]
  25. Liu, J.; Jin, C.; Wang, X.; Pan, K.; Li, Z.; Yi, X.; Shao, Y.; Sun, X.; Yu, X. A comparative analysis of deep learning models for assisting in the diagnosis of periapical lesions in periapical radiographs. BMC Oral Health 2025, 25, 801. [Google Scholar] [CrossRef] [PubMed]
  26. Cernekova, Z.; Sisik, V.; Jafari, F. Detection of Vascular Leukoencephalopathy in CT Images. In Proceedings of the Artificial Intelligence XLI; Springer: Cham, Switzerland, 2024; pp. 158–172. [Google Scholar]
  27. Lv, P.; Xu, H.; Zhang, Q.; Shi, L.; Li, H.; Chen, Y.; Zhang, Y.; Cao, D.; Liu, Z.; Liu, Y.; et al. An improved lightweight ConvNeXt for rice classification. Alex. Eng. J. 2025, 112, 84–97. [Google Scholar] [CrossRef]
  28. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  29. Rahhal, M.M.A.; Bazi, Y.; Jomaa, R.M.; Zuair, M.; Melgani, F. Contrasting EfficientNet, ViT, and gMLP for COVID-19 Detection in Ultrasound Imagery. J. Pers. Med. 2022, 12, 1707. [Google Scholar] [CrossRef]
  30. Abd El-Aziz, A.A.; Mahmood, M.A.; Abd El-Ghany, S. A Robust EfficientNetV2-S Classifier for Predicting Acute Lymphoblastic Leukemia Based on Cross Validation. Symmetry 2025, 17, 24. [Google Scholar] [CrossRef]
  31. Zhao, H.; Zhang, X.; Gao, Y.; Wang, L.; Xiao, L.; Liu, S.; Huang, B.; Li, Z. Diagnostic performance of EfficientNetV2-S method for staging liver fibrosis based on multiparametric MRI. Heliyon 2024, 10, e35115. [Google Scholar] [CrossRef]
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  33. Huang, P.-H.; Pan, Y.-H.; Luo, Y.-S.; Chen, Y.-F.; Lo, Y.-C.; Chen, T.P.-C.; Perng, C.-K. Development of a deep learning-based tool to assist wound classificati on. J. Plast. Reconstr. Amp Aesthetic Surg. 2023, 79, 89–97. [Google Scholar] [CrossRef]
  34. Malihi, L.; Hüsers, J.; Richter, M.L.; Moelleken, M.; Przysucha, M.; Busch, D.; Heggemann, J.; Hafer, G.; Wiemeyer, S.; Heidemann, G.; et al. Automatic Wound Type Classification with Convolutional Neural Networks. In Studies in Health Technology and Informatics; IOS Press: Amsterdam, The Netherlands, 2022. [Google Scholar]
  35. Dührkoop, E.; Malihi, L.; Erfurt-Berge, C.; Heidemann, G.; Przysucha, M.; Busch, D.; Hübner, U. Automatic Classification of Wound Images Showing Healing Complications: Towards an Optimised Approach for Detecting Maceration. In Studies in Health Technology and Informatics; IOS Press: Amsterdam, The Netherlands, 2024. [Google Scholar]
  36. Patel, Y.; Shah, T.; Dhar, M.K.; Zhang, T.; Niezgoda, J.; Gopalakrishnan, S.; Yu, Z. Integrated image and location analysis for wound classification: A dee p learning approach. Sci. Rep. 2024, 14, 7043. [Google Scholar] [CrossRef]
  37. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  38. Amin, J.; Sharif, M.; Anjum, M.A.; Khan, H.U.; Malik, M.S.A.; Kadry, S. An Integrated Design for Classification and Localization of Diabetic F oot Ulcer Based on CNN and YOLOv2-DFU Models. IEEE Access 2020, 8, 228586–228597. [Google Scholar] [CrossRef]
  39. Nohara, Y.; Matsumoto, K.; Soejima, H.; Nakashima, N. Explanation of machine learning models using shapley additive explanation and application for real data in hospital. Comput. Methods Programs Biomed. 2022, 214, 106584. [Google Scholar] [CrossRef] [PubMed]
  40. Rathore, P.S.; Kumar, A.; Nandal, A.; Dhaka, A.; Sharma, A.K. A feature explainability-based deep learning technique for diabetic foot ulcer identification. Sci. Rep. 2025, 15, 6758. [Google Scholar] [CrossRef]
  41. Rostami, B.; Anisuzzaman, D.M.; Wang, C.; Gopalakrishnan, S.; Niezgoda, J.; Yu, Z. Multiclass wound image classification using an ensemble deep CNN-based classifier. Comput. Biol. Med. 2021, 134, 104536. [Google Scholar] [CrossRef]
  42. Ucan, S.; Ucan, M.; Kaya, M. Deep Learning Based Approach with EfficientNet and SE Block Attention Mechanism for Multiclass Alzheimer′s Disease Detection. In Proceedings of the 2023 4th International Conference on Data Analytics for Business and Industry (ICDABI), Zallaq, Bahrain, 25–26 October 2023; pp. 285–289. [Google Scholar]
Figure 1. Example images of (A) an arterial ulcer and (B) a venous ulcer of the lower extremities.
Figure 1. Example images of (A) an arterial ulcer and (B) a venous ulcer of the lower extremities.
Diagnostics 15 02184 g001
Figure 2. Grad-CAM visualizations for ulcer classification. (A) Venous ulcer; (B) Arterial ulcer: representative original images (far left) and Grad-CAM outputs for five different DL models: ResNet50, ResNeXt50, ConvNeXt, EfficientNetB2, EfficientNetV2 (from left to right). Red and yellow areas indicate regions highly relevant to the model decision, while blue and purple areas are less relevant.
Figure 2. Grad-CAM visualizations for ulcer classification. (A) Venous ulcer; (B) Arterial ulcer: representative original images (far left) and Grad-CAM outputs for five different DL models: ResNet50, ResNeXt50, ConvNeXt, EfficientNetB2, EfficientNetV2 (from left to right). Red and yellow areas indicate regions highly relevant to the model decision, while blue and purple areas are less relevant.
Diagnostics 15 02184 g002
Table 1. Differential diagnosis between venous and arterial ulcers.
Table 1. Differential diagnosis between venous and arterial ulcers.
CharacteristicsVenous UlcersArterial Ulcers
EtiologySustained venous hypertensionArterial insufficiency (ischemia)
LocalizationMedial/lower legToes, feet, lateral ankle, pressure points
Wound edgeGently slopingPunched-out, well-demarcated, deep
Wound bedRed, granulating tissue with sloughPale, dry, necrotic, exposed tendons
ExudateModerate-to-heavyMinimal
Wound shapeIrregularRegular, round
Pedal pulsesPresent and normalDiminished or absent
Surrounding skinEdema, hemosiderin staining, lipodermatosclerosisShiny, cool, hair loss, pallor, cyanosis
Table 2. Inclusion and exclusion criteria.
Table 2. Inclusion and exclusion criteria.
Inclusion CriteriaExclusion Criteria
Patients with a confirmed diagnosis of arterial or purely venous ulcers of the lower extremitiesMixed arterial–venous ulcers.
Availability of photographic wound documentation under standardized conditionsUlcers of other etiologies.
Table 3. Selected Convolutional neural network architectures and weight references.
Table 3. Selected Convolutional neural network architectures and weight references.
ModelModel VersionHugging Face Configuration
ResNet50 [32]Vanillaresnet50.a1_in1k
ResNeXt50 [18]Vanillaresnext50_32x4d.a1h_in1k
ConvNeXt [19]Tinyconvnextv2_tiny.fcmae_ft_in1k
EfficientNetB2 [20]Vanillaefficientnet_b2.ra_in1k
EfficientNetV2 [21]Smallefficientnetv2_rw_s.ra2_in1
Table 4. Hyper-parameters for training classification models.
Table 4. Hyper-parameters for training classification models.
Hyper-ParameterValue
Image Size224 × 224
Batch Size32
Learning Rate1 × 10−4
Epochs50
Table 5. Classification performance in %.
Table 5. Classification performance in %.
CNN ModelMacro-Avg Acc.Precision (%)Macro-Avg F1ROC-AUC
VenousArterial
ResNet5095.7098.1892.1195.510.9883
ResNeXt5097.8598.2597.2297.730.9966
ConvNeXt (Tiny)72.0469.6285.7164.440.9381
EfficientNetB295.7098.1892.1195.510.9744
EfficientNetV296.7795.00100.0096.770.9810
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Neuwieser, H.; Jami, N.V.S.J.; Meier, R.J.; Liebsch, G.; Felthaus, O.; Klein, S.; Schreml, S.; Berneburg, M.; Prantl, L.; Leutheuser, H.; et al. Interpreting Venous and Arterial Ulcer Images Through the Grad-CAM Lens: Insights and Implications in CNN-Based Wound Image Classification. Diagnostics 2025, 15, 2184. https://doi.org/10.3390/diagnostics15172184

AMA Style

Neuwieser H, Jami NVSJ, Meier RJ, Liebsch G, Felthaus O, Klein S, Schreml S, Berneburg M, Prantl L, Leutheuser H, et al. Interpreting Venous and Arterial Ulcer Images Through the Grad-CAM Lens: Insights and Implications in CNN-Based Wound Image Classification. Diagnostics. 2025; 15(17):2184. https://doi.org/10.3390/diagnostics15172184

Chicago/Turabian Style

Neuwieser, Hannah, Naga Venkata Sai Jitin Jami, Robert Johannes Meier, Gregor Liebsch, Oliver Felthaus, Silvan Klein, Stephan Schreml, Mark Berneburg, Lukas Prantl, Heike Leutheuser, and et al. 2025. "Interpreting Venous and Arterial Ulcer Images Through the Grad-CAM Lens: Insights and Implications in CNN-Based Wound Image Classification" Diagnostics 15, no. 17: 2184. https://doi.org/10.3390/diagnostics15172184

APA Style

Neuwieser, H., Jami, N. V. S. J., Meier, R. J., Liebsch, G., Felthaus, O., Klein, S., Schreml, S., Berneburg, M., Prantl, L., Leutheuser, H., & Kempa, S. (2025). Interpreting Venous and Arterial Ulcer Images Through the Grad-CAM Lens: Insights and Implications in CNN-Based Wound Image Classification. Diagnostics, 15(17), 2184. https://doi.org/10.3390/diagnostics15172184

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop