1. Introduction
A hard-to-heal wound is usually caused by an underlying disease where trauma might or might not be involved; examples are diabetic foot-, venous leg-, and pressure ulcers. These wounds heal by secondary wound healing and are classified as hard-to-heal when healing is delayed past certain rate thresholds [
1]. The prevalence of hard-to-heal wounds is growing and currently estimated at 1–2% across developed countries [
2]. The costs and resources demanded to take care of these wounds have been studied in various healthcare systems. In wounds that do not require hospitalization, there are still significant costs of primary care visits, nurse working time, drugs, other interventions as well as dressing materials [
3]. One study estimated the US Medicare costs for wound care to be at least
$32 billion annually [
4]. Preventive measures that target the etiology of hard-to-heal wounds are carried out by various disciplines of medicine, yet the prevalence is growing, and therefore so is the need for more efficient wound management strategies. To refine treatments, an individualized treatment strategy is needed, and this necessitates an advanced wound bed evaluation.
Clinical Problem Addressed and Innovation
Health care professionals need to overall accurately assess the wound to be able to evaluate whether the healing process is adequate and implemented treatment strategies are functional. This includes identifying wound size, i.e., the wound surface area and depth, together with an adequate assessment of the wound bed characteristics, most commonly involving slough and necrosis.
The conventional clinical approach to measuring wound dimensions typically involves the utilization of disposable plastic rulers for surface area assessment. The evaluation of wound depth is comparatively less prevalent and less documented [
5,
6,
7,
8,
9]. Wound depth measurements encompasses either (a) a subjective appraisal conducted without the aid of instruments, or (b) a method wherein healthcare practitioners uses a Q-tip for approximating depth by measuring the penetration of the Q-tip into the wound with a disposable ruler. Measurements of the wound surface area and depth acquired through these methods are well known to be imprecise and subjective [
10]. Accurate estimations of wound surface area and depth are central for evaluating wound progression. Due to the absence of easy and practical clinical methods for assessing especially wound depth, there is low reliability (or even absence) of measurements documented in the electronic health records (EHRs) today. For the assessment of wound bed characteristics, such as slough and necrosis, this is evaluated by ocular inspection based on the skills of the observer. This evaluation then underlies the subsequent debridement, a very important intervention for the outcome of the healing process.
This study evaluates accuracy and precision of a mobile, automated, at the point of care device, SeeWound©2, which through AI methodology provides image-based wound surface area and depth assessments, together with a quantitative wound bed slough and necrosis characterization. A comparison to current conventional techniques, such as digital planimetry and best practices, is also undertaken.
2. Materials and Methods
SeeWound©2 (Dermacut AB, Norr Mälarstrand, 60 11235 Stockholm, Sweden)uses a LiDAR sensor to assess wound dimensions, see
Figure 1 below. The LiDAR sensor emits light that underlies the estimate of the distance between camera and wound by calculating the time it takes for light to travel from the sensor to the wound and back. In a previous publication, the influence of light interference, such as the need to use of flash, light settings in the room, and evaluation of optimal distance between the wound and the camera, has been undertaken [
11]. LIDAR technology (iPad 11 Pro) and a newly developed algorithm, (SeeWound© 1(Dermacut AB, Norr Mälarstrand, 60 11235 Stockholm, Sweden)) which also incorporates an angle adjustment function that ensures control over the angular distortion during image capture, is used [
11]. If the wound can be seen in the images, the previous publication shows that the wound area size can be properly depicted. Further, the paper shows that, with normal ambient light, the flash function is not needed [
11]. The development of the surface area measurement algorithm has been previously described (SeeWound© 1). In short, it involved a training, validation, and a test set according to general model producing practices and includes an angle adjustment function [
11]. For the wound surface estimation in the present study, the SeeWound© 1 algorithm is identical to that of SeeWound© 2. For a comparison of wound size measurements using SeeWound© 1 and 2 on the same wound and using different mobile platforms, see
Table 1. The current application (SeeWound©2) is currently available for research through both TestFlight and Google Play by a written request and regulatory approval.
Segmentation Process and Wound Depth Estimation
The SeeWound©2 uses two sequential convolutional neural networks based on the U-Net architecture to perform wound analysis. U-Net 1 is responsible for segmenting the wound bed. The output is subsequently used for wound area calculation, depth estimation, and serving as the input mask for the second U-Net model. U-Net 2 performs wound tissue segmentation, specifically identifying slough and necrotic tissue within the wound area previously delineated by U-Net 1. An illustration of this process is provided in
Figure 2.
2.1. Wound Surface Area Determination
At the outpatient wound clinic of the National Burn Centre of Linköping University Hospital, wound specialists (nurses) examined the wound size in 11 patients with hard-to-heal wounds with the use of the SeeWound©2. For the accuracy estimate, wound images were captured in 11 patients (
Table 2). These measurements were then compared to wound sizes obtained through manual pixel-by-pixel segmentation of a wound image performed by a wound specialist using digital planimetry, serving as the true size (digital planimetry) for the accuracy evaluation. Repeated images were also taken during the patient visit for the precision estimate in eight patients (
Table 3). These paired measurements were used to assess the device’s reproducibility and precision.
2.2. Wound Depth Assessment
Wound depth assessment is conducted in two steps. First, the AI model is used to segment the wound area. In the second step, the boundary between healthy and wounded tissue as a reference surface is calculated, representing the estimated level of intact skin prior to injury. From this reconstructed surface, the vertical distance down to the bottom of the wound bed is measured by the LiDAR© system. This method enables estimation of wound depth by calculating the difference between the hypothetical healthy skin surface and the actual wound base (
Figure 2): where D1 and D2 serve as clarification positions for the flip of the wound into portrait view, and where DF is the calculated depth between the reconstructed wound surface and the vertical distance down to actual wound bed.
The wound depth assessment functionality of SeeWound 2 was evaluated in an “in vitro” model (
Figure 3 and
Figure 4) (Part A) and in a clinical setting (Part B), the latter similar as described for the clinical wound surface evaluation, see above. Issues thought to affect the accuracy and precision of the wound depth estimate, such as the wound border specifics and absolute depth were examined.
The vitro model (Part A) was constructed to assess the effects of wound aperture and wound border effects on the wound depth estimate. The model includes an adjustable wound aperture in relation to the wound depth (see
Figure 4). It is built on a stationary structure with two adjustable cubes, whereupon the in vitro wound model [
12] (see
Figure 3) is placed (in between). The structure then allowed the two adjustable objects to be moved towards each other, decreasing the wound aperture and at the same time increasing the depth of the wound. This model allows the examination of the effects of wound aperture and wound edge angle on the wound depth assessment. This is performed while the two blocks are adjusted manually and in accordance with the study protocol. It was realized early on that, given the resolution of the SeeWound© 2 depth measurement ± 3 mm, the shallowest wounds could not be differentiated from each other (see discussion section).
All in all, 51 wounds of different sizes, depth, and form were evaluated (Part A). Each wound was examined 8–10 times. The results were compared to the true depth of each (probe and ruler). Furthermore, limitations regarding technical constraints when undertaking a depth estimate were identified, and then possible constraints related to the specific dimensions of the wound (for example, wound size and depth particularly in relation to the size of the wound opening, angle of wound edges) were examined.
Clinical evaluation (Part B). In the outpatient clinic of the Burn Centre of Linköping University Hospital, wound nurse specialists first measured wound depth using a sterile plastic probe and a ruler. This manual measurement was documented, after which two images of the patient’s wound were captured using the SeeWound© 2 device. The images were captured with the camera held perpendicular to the wound surface, in standard luminance patient room. SeeWound© 2 automatically detects the wound contours and calculates the deepest point based on its segmentation (see
Figure 5). The automatic depth measurements from SeeWound© 2 were then compared to the manual probe measurements, which served as the ground truth (probe). Two images and measurements were performed to determine the device’s accuracy against the probe and for the reproducibility in repeated measurements.
2.3. Development of AI Model for the Identification of Slough and Necrosis
The U-Net architecture is well suited for identifying small objects and local differences in images, making it ideal for segmenting slough and necrosis such as in a wound bed image. As common within AI model medical applications, there are often limited data available for training, which makes the U-Net a good choice due to its ability to train efficiently on smaller datasets [
13,
14]. The U-Net is divided into two parts: the left side (contracting path) and the right side (expanding path). In the contracting path, the image’s height and width decrease, while the number of filters increases. This is performed through pooling operations, which reduce the spatial dimensions of the image. The goal of the contracting path is to locate and extract higher level features, such as colors, edges, and textures, through convolutions. Once the image has been reduced in size and the key features have been extracted, the expanding path performs the opposite task by upsampling the image back to its original size The expanding path uses the features identified in the contracting path (via skip connections) to recreate the full-sized image and make pixel-by-pixel predictions of the classes slough and necrosis [
13,
14].
Skip connections are links between the contracting and expanding paths, shown by the gray arrows in
Figure 6 below. These connections help the network use high-resolution information from the contracting path to improve upsampling and produce a more accurate output. This prevents the model from losing spatial information during the shrinking process when rebuilding the image to its original size [
13,
14].
The data used for the estimate of slough and necrosis consist of images, collected from clinics using the SeeWound 1 device, of hard-to-heal wounds. In these images (n = 3383), the ground truths of the wound, for slough- and necrotic, segmentation were made by an experienced wound physician. Thereafter, this data set was divided into three categories: training 70%, validation 20%, and test set of 10%. The training data are entirely separated from the test set. In the training process, we used a randomization for partitioning the training data into training and validation subsets.
2.4. Model Training Performance on Training Validation Set for Slough and Necrosis
The developed model demonstrated a consistent decrease in training loss over the epochs, which indicatively means that the model is successful in its learning of the task to segment slough and necrosis from in the wound bed. The training accuracy, dice coefficient, and intersection-over-union (iou) increase steadily throughout the model training process and reflect the model’s capability of learning to accurately classify pixels to the image’s designated classifications, and the iou reflects how good the ground truth resemblance is in the predictive and segmentation performance. Validation accuracy, dice coefficient, and the iou remained high and stable throughout the model training, indicating stable performance against the validation data set and continuous improvement in the training. See
Figure 7, a and b below, for a visual representation of the training data per epoch and, respectively, for the validation performance of each epoch. The model training was stopped based on time at 10 epochs, as seen in the figure, meaning that when the model did not improve over 10 epochs, the training session was terminated to avoid over-fitting.
2.5. Clinical Evaluation
The final clinical evaluation of the accuracy and precision of the model was performed on two patient cohorts and included 466 and 25 patients, respectively, which were not included in the original training data. The patients included in this evaluation stem from units including vascular, dermatology, endocrinological, primary care, and municipal care. This implies a heterogeneous data set including, but not limited to, diabetic foot ulcers, venous ulcers, pressure ulcers, and ischemic ulcers. The output was evaluated by a wound care physician with ground truth segmentations (planimetry) to determine if there were correct identifications of slough and necrosis in the image. A binary comparison was also made, identifying if slough and necrosis were present or not in the wound as compared to ground truth (n = 466).
2.6. Ethics
Ethical approval was obtained from the relevant ethics committee (The National Swedish Ethics committee, Gothenburg, Dnr. 2021-04913) The ethics committee waived the need from informed consent from the patients as all the images examined did not contain any identifiable personal data and were considered de-identified/anonymized data. All patients were informed prior to image acquisition about the intended use of the images for research purposes.
2.7. Statistics
Data are presented as mean and standard deviation, standard error or confidence intervals. A p < 0.05 is considered statistically significant.
4. Discussion
4.1. Summary of Findings
This study aimed to evaluate the accuracy and precision of the AI-powered tool SeeWound©2, for assessing the wound surface area, the wound depth, and the wound bed characteristics (slough and necrosis). The results show that SeeWound© 2, which is currently applying for a MDR certification, demonstrated robust performance metrics: for the wound surface area, the accuracy was 96.28% and the precision was 90.00%, with a coefficient of variation (CV) of 5.56%. For the wound depth, the accuracy was 90.75% and the precision was 89.55%, with a CV of 3.07%. For the wound bed characteristics, the AI model achieved accuracy values of 80.30% for slough and 84.73% for necrosis, with corresponding precision values of 88.61% (slough) and 85.74% (necrosis). These results demonstrate that SeeWound© 2 is a useful tool for wound assessment, documentation, and clinical decision support, particularly for wound size and characterization. Furthermore, comparisons with traditional methods, such as digital planimetry, indicate that SeeWound© 2 can be used as a decision support for wound size and characterization in clinical practice. The SeeWound© 2 device has an automatic export of the wound metrics to present to the major Swedish Clinical documentation system (COSMIC) by a facilitated capture algorithm, which also can be modified as needed to other specific clinical documentation systems.
4.2. Wound Surface Area
Among the parameters evaluated, the accuracy and precision for estimating the wound surface area were the highest. This was expected, given that the technical complexity of surface area estimation is relatively lower compared to the depth and the wound bed characteristics. The algorithm underlying the surface area model was developed in 2019 and has undergone continuous refinement, including adjustments to the camera-to-wound angle, significantly improving precision. This model, which is consistent across different mobile devices (see
Table 1 above), produces results comparable to traditional planimetry. A comparison with the ASURA AI model [
15] showed that SeeWound© 2 outperforms conventional ruler-based methods, which often overestimate the wound area due to simplified length × width calculations, with error margins ranging from 12% to 87% [
8,
16,
17,
18,
19]. SeeWound© 2 achieved an accuracy of 96% and precision of 90%, surpassing the performance of both ASURA and traditional methods. Notably, ASURA’s model demonstrated an accuracy of 87.9% when measuring objects of known size [
15], yet SeeWound© 2 was able to outperform these models in wound surface area estimations, confirming its high accuracy and precision.
4.3. Wound Depth
Assessing wound depth presents a greater challenge than surface measurements, primarily due to the inherent complexity of wound structures. Although high accuracy has been achieved with other systems, such as the WoundVue system (with 12.9% precision) and the method proposed by Virginia et al. (with 2% accuracy), these systems require additional hardware [
20,
21]. In contrast, SeeWound© 2 leverages LiDAR technology in modern smartphones, which eliminates the need for external equipment and simplifies its use in clinical settings. However, certain limitations remain. The LiDAR sensor’s depth map resolution (192 × 256 pixels) has limitations when capturing small depth variations, particularly for shallow wounds with depths under 5 mm. Additionally, the performance of the system weakens when the wound edges are steep or when the wound opening is less than 40 degrees. This is due to the light from the LiDAR sensor having difficulty reaching and reflecting on the respective wound surface. Despite these limitations, the system remains a valuable tool in clinical practice where precise depth measurements for very shallow wounds hold limited clinical value.
A review by Lasschuit et al. [
10] examined available devices and methods for measuring wound depth, concluding that, despite the availability of several 3D imaging systems, depth measurements are still most reliably obtained using a sterile probe. This underscores the challenge in accurately assessing wound depth, as manual methods, such as ruler-based estimates, have demonstrated significant inaccuracies, with overestimations ranging from 12% to 87% [
16,
17,
18,
19], and inter-rater variability of estimating the wound size from −44% up to +75% [
7]. These limitations are particularly pronounced in complex wound geometries or cases involving undermining [
10].
4.4. Wound Characteristics (Slough and Necrosis)
Assessing wound bed characteristics such as slough and necrosis is particularly challenging due to the subjective nature of tissue identification, especially at the borders between tissue types. In binary evaluations, the AI model demonstrated high accuracy and precision in identifying the presence or absence of slough and necrosis. However, pixel-by-pixel evaluations of slough and necrosis were less accurate, with slough accuracy at 80.30% and necrosis accuracy at 84.73%. These results are consistent with the limitations inherent in the task of defining slough and necrosis, especially at their borders, where observer subjectivity plays a significant role. The precision for digital planimetry in slough and necrosis was 88.61% for slough and 85.74% for necrosis, with relatively high CV values (7.00% for slough and 7.54% for necrosis), suggesting that inter- and intra-rater variability, rather than the AI algorithm, is largely responsible for discrepancies in accuracy.
Mukherjee et al. [
22] trained a Support Vector Machine (SVM) model using clinician-annotated images as the ground truth during the training phase. The model achieved an overall tissue segmentation accuracy of 86.13% across granulation, slough, and necrosis classes. Similarly, Morgado et al. [
23] utilized a DeepLabV3-R50 segmentation model to classify wound tissue types, achieving mean absolute errors of 14.31% for slough and 8.84% for eschar. Zoppo et al. [
24] evaluated several wound assessment tools and reported mean relative errors of approximately 14% for WoundViewer, Silhouette, and Visitrack, while the MOWA system exhibited a higher relative error of 23%. Fauzi et al. [
25] evaluated a color-based segmentation approach using hue, saturation, and value (HSV) color space, reporting an overall accuracy of 75%, with class-specific accuracies of 63% for slough and 75.1% for eschar. They also reported inter-expert agreement scores of 67.4% and 84.3%, respectively. These studies demonstrate the ongoing challenges and variability in accurately segmenting wound tissue types, especially for slough and necrosis, underscoring the potential of AI in reducing inter-expert variability and improving wound assessment.
Interestingly, a study by Mohammed et al. [
26] found that clinicians achieved 84% and 87% agreement for the presence of slough and necrosis, respectively, in nine wound images. This finding aligns with the current study, where the AI model showed a higher degree of consistency compared to clinician-based estimates. Further comparative studies by Ramachandram et al. [
27] and Mukherjee et al. [
22] also reported high variability among clinicians in quantifying tissue proportions, with Ramachandram et al.’s deep learning model achieving a mean intersection over union (IoU) of 86.44% for wound segmentation and 71.92% for tissue segmentation across a dataset of 58 wound images. Similarly, SeeWound© 2 demonstrated better precision and accuracy than digital planimetry, outperforming manual tracing and segmentations, which have average intra-rater ICCs of 81.02% and 78.16% for slough and necrosis respectively [
27].
In clinical practice, ocular methods for estimating slough and necrosis, such as the quarter-based percentage estimation method, result in significant estimation errors. A typical absolute error in tissue estimations is ±12.5%, based on the 25% range used for proportion estimates of slough and necrosis (Bates–Jensen Wound Assessment Tool, BWAT) [
28]. In contrast, SeeWound© 2 provides a more objective and consistent method for classifying tissue types, outperforming ocular assessments, which exhibit poor inter-rater agreement (ICC of 0.37–0.38) [
26].
4.5. Limitations
4.5.1. Wound Surface Area
The primary challenge is that traditional methods, such as ruler-based techniques, often lead to overestimations, with error margins ranging from 12% to 87% and inter-variations between −44% to +75% on the same wounds [
7,
8,
16,
17,
18], particularly when estimating wound areas with simple length × width calculations. However, these issues are mitigated with SeeWound©2, which provides accurate estimations without the limitations associated with traditional methods.
4.5.2. Wound Depth
The most significant and obvious limitation is when parts of the wound cavity are not visible to the Lidar camera. Another limitation for wound depth measurement lies in the resolution of the LiDAR depth map and the angle of image capture. The latter is due to the image depth map resolution provided by the LiDAR technology presently uses 192 × 256 pixels. Combined with its sensitivity to the distance between the camera and the object—requiring more than 30 cm for optimal performance—this results in a limited ability to accurately capture steep edges. Even when the depth map is morphed to a higher resolution (e.g., 512 × 683 pixels), the system still challenges the sensitivity needed to detect small surface variations or subtle depth differences. Additionally, steep wound edges and non-visible wound portions in undermined or deeply cavernous wounds result in underestimated depth measurements. The light from the LiDAR sensor has difficulty reaching and reflecting on such wounds, reducing its performance in these clinical scenarios (see
Figure 9 for visual explanation of the challenges). Future work to facilitate a 3-D wound bed reconstruction is currently being planned.
4.5.3. Wound Bed Characteristics
Evaluating wound bed characteristics such as slough and necrosis remain challenging due to the subjective nature of tissue identification, especially at the borders between tissue types. However, SeeWound© 2 has demonstrated better consistency and higher accuracy than manual, ocular assessments, which are subject to significant observer variability. Furthermore, studies comparing AI-based segmentation with clinician-annotated images have shown that the AI model outperforms clinicians in both consistency and accuracy, highlighting the potential of AI-powered tools in wound care. We therefore suggest that the present accuracy is sufficient for clinical use. This, however, needs to be further documented in future studies.