Skip to Content
DiagnosticsDiagnostics
  • Article
  • Open Access

6 September 2023

Validating Wound Severity Assessment via Region-Anchored Convolutional Neural Network Model for Mobile Image-Based Size and Tissue Classification

,
,
and
1
Department of Computer Science and Engineering, Kongunadu College of Engineering and Technology, Trichy 621215, India
2
Department of Computer Science, Prince Sattam Bin Abdulaziz University, Wadi al dwassir 1190, Saudi Arabia
3
Department of Computer Sciences, College of Arts and Sciences, Prince Sattam Bin Abdulaziz University, Wadi al dwassir 1190, Saudi Arabia
*
Authors to whom correspondence should be addressed.

Abstract

Evaluating and tracking the size of a wound is a crucial step in wound assessment. The measurement of various indicators on wounds over time plays a vital role in treating and managing crucial wounds. This article introduces the concept of utilizing mobile device-captured photographs to address this challenge. The research explores the application of digital technologies in the treatment of chronic wounds, offering tools to assist healthcare professionals in enhancing patient care and decision-making. Additionally, it investigates the use of deep learning (DL) algorithms along with the use of computer vision techniques to enhance the validation results of wounds. The proposed method involves tissue classification as well as visual recognition system. The wound’s region of interest (RoI) is determined using superpixel techniques, enabling the calculation of its wounded zone. A classification model based on the Region Anchored CNN framework is employed to detect and differentiate wounds and classify their tissues. The outcome demonstrates that the suggested method of DL, with visual methodologies to detect the shape of a wound and measure its size, achieves exceptional results. By utilizing Resnet50, an accuracy of 0.85 percent is obtained, while the Tissue Classification CNN exhibits a Median Deviation Error of 2.91 and a precision range of 0.96%. These outcomes highlight the effectiveness of the methodology in real-world scenarios and its potential to enhance therapeutic treatments for patients with chronic wounds.

1. Introduction

Skin ulcers, a leading source of disease and mortality around the globe, are brought on by a number of illnesses and disorders, including insulin resistance, peripheral nerve damage, motionlessness, stress, artery disease, illnesses, and vein insufficiency. These ulcers are characterised by their inability to heal properly within the expected timeframe of 4 weeks to 3 months, leading to a lack of functional and anatomical restoration. This is particularly relevant in the case of diabetic foot ulcers. Underlying pathological factors often hinder or delay the healing process, resulting in significant negative impacts on the physical, emotional, and social well-being of affected individuals. Chronic ulceration and foot diabetes amputated ulcer demand a lot of time and supplies for treatment, which has economic ramifications [1]. These variables contribute to the worldwide difficulty in treating complicated wounds like pressure ulcers, lesions of the lower extremities (such as arterial, venous, and diabetic foot ulcers), and neoplasm lesions [2]. Due to the absence of standardised treatment recommendations and the abundance of care items, it is possible that the root cause of complicated injuries is not adequately addressed during therapy. The engagement of healthcare experts is critical for the medical care of complicated injuries in order to encourage regeneration and enhance the use of both physical and mental assets. Clinical practise rules for managing wounds have been modified by medical organisations by adding additional assistance systems. These systems act as tools to aid healthcare providers in their organisational and behavioural decisions [3]. These recommendations’ main goal is to provide doctors with the information they need to identify, avoid, and manage injuries as effectively as possible. In order to improve the efficiency of the recovery process, they also advocate for the careful allocation of organisational capabilities. Medical practise recommendations often concentrate on resolving problems associated with the avoidance and therapy of wounds caused by pressure in those that are vulnerable or have been diagnosed with them presently. These guidelines outline the various stages of clinical practice involved in managing chronic wounds, encompassing prevention, treatment, and healing [4].
Chronic wounds contribute significantly to illness and death rates, placing a substantial financial burden of around $25 billion per year on the U.S. healthcare system. Effective management of patients with chronic wounds involves regular visits by an interprofessional team. The primary measure of wound healing is the reduction in the surface area of the lesion. This measurement is essential for clinicians to track progress and make informed decisions regarding treatment options. Therefore, precise and accurate wound area measurements play a vital role in maximizing patient outcomes for individuals with chronic wounds. Many medical facilities still use mechanical ruler-based measures [5,6], which are inconsistent and might underestimate the wound’s extent by 40%. Reach acetic tracer is a different frequently utilised technique, although it may change the boundaries of the wound, put individuals at a greater chance for disease at danger, and be uncomfortable. Automated analogue planimetry, which analyses wound pictures, increases precision but is time-consuming and may not be suitable for busier healing facilities. The proportion of normal granulation tissue in the site of the wound, in addition to the overall size of the lesion, is a key factor in deciding when a wound will start to recover or if it has to be definitively closed with a skin transplant or flap [7,8,9]. Clinicians use colour to determine the proportion of tissue that granulates (PGT). Dark red tissue with granules may signify disease, whereas pale tissues could indicate inadequate neurogenesis and blood flow to the wound bed. Visual assessment of PGT is, however, unreliable and might vary greatly amongst doctors. According to current studies, colour picture analysis of granulation tissues can forecast how the ulcers will heal, offering a more precise evaluation than perceptual judgement alone.
This has yet to be a widely accepted methodology for assessing injuries. However, developments in AI are making automatic diagnostics evaluation of images possible. As wound diagnostic tools and instruments like Silhouette (Aranz Medicine) and insight (Ekare) get better, doctors are able to evaluate injuries more quickly, accurately, and consistently. This helps doctors make better decisions and gives patients better results. However, rigorous validation is necessary to ensure that these tools perform comparably to standard measurement techniques. AI-based visual injury detection solutions are not clinically evaluated.
This study seeks to evaluate the dependability of AI-driven injury assessment technology by juxtaposing it against the judgments made by medical professionals through manual evaluations. The effectiveness of the AI in accurately identifying wound areas featuring granulation tissue was measured by thoroughly analysing the error metrics pattern between the AI-generated tracings and a benchmark established by two individual human tracings. However, we acknowledge that relying solely on two human evaluators might have limitations, and we recognize the importance of considering alternative, clinically validated strategies for a more comprehensive validation process. We also created a qualitative technique for assessing the AIs efficiency since wound evaluation is biased. This featured skilled wound healing, presenting doctors masked-reviewing AI and humans tracings.

1.1. Problem Statement

The problem addressed in the research paper titled “Wound Image Validation using Region Anchored CNN Model for tracking size and tissue classification” revolves around the accurate and reliable validation of wound images, specifically concerning two critical aspects: wound size tracking and tissue classification. In order to deal with injuries successfully and to plan treatments, it is crucial to identify the various kinds of tissue present in an injury and to assess them precisely. This study intends to address the issues associated with injury imagery analysis.
Existing wound image validation methods may suffer from inconsistencies, subjective assessments, and manual errors, making it difficult to obtain accurate and standardised measurements. Manual tracing and visual estimation can lead to variations between clinicians, impacting the reliability of wound assessments and decisions regarding treatment options. Therefore, there is a need for an advanced and automated approach that can consistently and objectively track wound size and classify different tissue types in wound images.

1.2. Major Contribution of the Proposed Research Work

The effective adoption of the Region Anchored CNNs Algorithm for injuries’ picture evaluation might impact wound treatment and care. Major contributions of the above research work can be summarised as follows:
  • Leveraging mobile device-captured photographs, our approach offers a convenient and accessible way to monitor wound size changes over time.
  • By integrating deep learning algorithms and computer vision techniques, our method aims to improve wound validation outcomes, aiding healthcare professionals in making well-informed decisions for patient care.
  • The proposed approach employs a Region Anchored Convolutional Neural Network (CNN) framework, utilizing superpixel techniques to accurately identify the wound’s region of interest (RoI), leading to precise wound size calculations.
  • Beyond wound size, our research incorporates a Tissue Classification CNN that effectively categorises different tissue types within the wound.
  • The model’s Median Deviation Error of 2.91 and precision range of 0.96% highlight its success in accurately detecting and classifying diverse tissue types, including granulation tissue, necrotic tissue, and epithelial tissue.

3. Taxonomy of the Proposed Methodology

This research introduces a digital approach for determining the area of a lesion and categorising its tissues. A computerised device selects the location that is important (defining the wound), a computerised method measures space utilising a third-party calibrator, and an experienced convolutional neural structure classifies tissue. Figure 1 shows the information gathering and sorting stages, which includes evaluation and tissues grouping. These processes were integrated into clinical practice using the clinic gram® mobile application.
Figure 1. Pipeline of the proposed model flow.

3.1. Wound Identification Mechanism

This method uses neighbouring pixels’ similarity to determine a wound’s boundaries. When recording a picture of an injury utilising a smartphone or tablet, the individual creates a scribble inside the injury zone (refer to Figure 2) to assist the technology in recognising the injury. The procedure is shown using an example picture in Figure 3. After the machine locates the shape of the injury, a surgical mask [18] is used to separate the area that is important from the backdrop and epidermal around it.
Figure 2. Capturing, highlighting, and extracting the wound mask from a test image allows the system to detect a wound.
Figure 3. Original, superpixel segmented, scraw, and identified region of interest images.
In our proposed approach, the identification of the Region of Interest (ROI) within wound images is a pivotal step for accurate wound assessment. The ROI corresponds to the area of the image that encompasses the wound itself, excluding surrounding healthy tissue. This process involves the utilization of a Region Anchored Convolutional Neural Network (CNN) framework that employs superpixel techniques to pinpoint the exact boundaries of the wound within the image.
Superpixel Segmentation: Superpixel segmentation is a technique that groups pixels with similar characteristics, such as colour and texture, into coherent regions. This approach assists in effectively identifying the boundaries of the wound, as it considers both local and global image features. By segmenting the image into superpixels, the CNN can then focus its attention on these localised regions, aiding in the precise delineation of the wound area.
Region Anchoring: The Region Anchored CNN model leverages the information from superpixels to anchor and align the CNN’s receptive field with the wound’s boundaries. This anchoring process ensures the CNN’s analysis to be concentrated on the wound area, which subsequently enables accurate identification of the ROI. By integrating superpixel segmentation and region anchoring within the CNN framework, our approach ensures that the wound’s ROI is precisely identified. This is paramount for accurate wound size calculations, tissue classification, and overall effective wound assessment.
The approach uses a mix of the super pixel approach and k-means algorithm [19] to figure out the shape of the injury and describe the ROIs after trying out approaches like Felzenszwalb [4], Mean shift, as well as Rapid change [6]. The super pixel technique dynamically divides the wound picture into many segments that have comparable characteristics or traits. In contrast, the unsupervised categorisation (clustering) method k-means organises objects according to their features. By minimising the sum of similarities among each item to the centre of its cluster, grouping is accomplished. Quadratic proximity is often used in this process. The system applies those algorithms and splits the picture into several megapixels to determine the shape of the injury. FNRs comprised the reference tracing region included located in the specimen’s testing observe established by referencing traced overall region utilising calculation 1.
F N R = 100 × R O / R
The healthcare expert sketches a tiny portion within the injury utilising the identical cellphone programme that was employed for picture collection, and the outline is created by relating several megapixels of the injury. The SP algorithm operates automatically and autonomously to calculate the various superpixelblocks. This section highlights the importance of adjusting the different operating variables of SP for optimal performance. These elements consist of N sections: The approximate number of sections needed. When grouping images in SP, compact influences how colour and closeness are compared. Greatest repetitions are permitted by the k-means algorithm. The Gaussian picture blurring filter’s symbol is sigma. Enforce connectivity: A variable indicating whether or not the segments should be connected. The adjustment of these variables is crucial to achieving the best results in detecting the wound’s contour.

3.2. Wound Area Calculation

The ROIs and an indicator are used to measure the incision. A blue rectangle (see Figure 4) serves as our identifier because of its high contrast with the background. Each side of the square marker is 2 cm long, giving it a set size. We can figure out how big the cut is by compared the sizes of the line on the picture to how big it is in real life. The doctor places the marker perpendicularly close to the incision before taking the picture. This step is essential since the marker’s precise placement affects the measured results. The measurement is deemed to be inaccurate if the indicator continues to be inclined, partly concealed, or impacted by light or shadow. In the first step (see Figure 5), the picture is read and converted from RGBs to HSVs colour space. The picture is then restricted to a range of colours, including many shades of blue, producing a mask which just contains pixels inside that spectrum. The mask is next checked for object features, and only squares with lengths greater than twice their breadth, like our indicator, are picked. An artificial eye algorithm is then used to roughly convert all of the outlines into square forms. The function determines which chosen rectangular forms most closely resemble the real contour through assessing the regions’ sizes. If numerous square forms are discovered, the form that is nearest to a rectangular shape, indicating the ones with the smallest distinction among the object’s region after performing the function and the originally identified region with the mask being used, is kept.
Figure 4. Left: original wound and marker, right: marker identified with green outline.
Figure 5. Left to right: marker detection and corrected picture processing.
Furthermore, an evaluation system for the marker has been implemented to ensure accurate measurements. The technology assesses the marker’s proportionality and asymmetricity to find its right place. The measurement of assessment is created by dividing the lesser identified half by the bigger identified sides in order to measure disparity. A score close to 0 denotes extreme disproportion, whereas an amount close to 1 represents a perfect structure with optimal proportions. Calculating vertical as well as horizontal disparity requires the marker’s vertex’ tilt. If the marker [20] is positioned entirely along the X and Y axes, this measurement tends towards 0. If the Z-axis is horizontal, the imbalanced values on both sides ought to be identical. The dimensions of the indicated incision may be determined by calculating the ratio received from the indicator, which connects the total amount of pixels to the real measurement of 4 cm2, after the indicator has been identified. After being found, the precision of the percentage and location are verified.
When the sample tracing’s non-regional portions are adjusted for normalisation by the benchmark tracing’s overall size, the samples tracing’s non-regional portions are called the false-positive region (FPR), calculated using Equation (2):
F P R = 100 × T O / R
The absolute variation in trace region, normalised by the original tracing total region, was the relative variance (RV), calculated using Equation (3):
R V = 100 × T R / R
To assess different aspects of tracing disparities, we utilised three error metrics. The ARVs evaluated the relative disparity between area measures without taking injury border location into account. However, the ARV may be modest or nil if commentators tracked totally independent regions of the same dimension. In contrast, FNR and FPR contrasted the experimental tracing and comparison tracing by assessing overlapping and underlapping, giving more information about the source of a mistake than Jaccard’s and Dice parameters, that incorporate both forms of faults. Two recordings for an image have been contrasted where only one tracing determined full skin epithelization or absence of granular tissue. However, if two or all three tracers indicated full epithelialization, no comparisons could be made for that image.

3.3. Identification of Tissue from Wound Imaging

Wound detection requires precise tissue separation. Complex lesions can exhibit irregular shapes, imprecise borders, and highly heterogeneous colouration [7]. To address this challenge, a convolutional network was employed to develop a system capable of classifying different tissues present in the wound. Multiple models were trained using a wound dataset consisting of 726 samples. The tissue samples in the dataset were distributed proportionally and evenly. To ensure successful outcomes, the images underwent preprocessing and augmentation steps. A black mask was applied around the ROIs; the images were resized to dimensions of 200 × 200 pixels, and they were divided into sections measuring 5 × 5 pixels. In our approach, we applied transformations such as rescaling the zoom range and performing horizontal flips during the augmentation phase. These techniques were utilized to enhance the dataset and improve the effectiveness of deep learning algorithms.
The information was then divided between learning groups, which included 80% of the specimens, as well as test collections, which included the balance, or 20%, of the sample sizes. This division allowed for training the model on one set and evaluating its performance on the other. The network’s precision, memory capacity, and F1 score were analysed to assess its effectiveness. To extract features from the data, transfer learning was utilised as it is specifically designed for small training databases. Compared to models built from scratch, transfer learning offers advantages, such as requiring less data, reducing training time, and improving overall performance. This study utilised Region Based CNN, which is recognised for its outstanding image classification capability due to its many small filters that considerably decrease the total amount of parameters, though at a less rapid processing pace. It additionally introduces additional blocks through the use of intermediate layers present. To tailor this CNN algorithm to the job at hand, the information set’s picture input size was changed, and four more layers were added. These layers comprised a layer of flatten to transform the outcome of the model into a vector of attributes, a dense layer having ReLU stimulation, a layer for Dropout containing a proportion of 0.3 to avoid excessive fitting, and a heavy barrier with activation using softmax to provide a category probabilities matrix as the end product. Accurate identification of samples is a highly significant step in creating a supervised learning system [21]. Recognizing the significance of correctly categorising the dataset, a support system was introduced to assist in the labelling process. During this stage, clinical experts collaborated to assign labels to all the generated images (refer to Figure 6).
Figure 6. Flow chart of RA-CNN model.

Region Anchored CNN

The procedure flow of the Region Anchored-CNN model framework (Figure 4) involves the following steps: ResNet101, a well-known CNNs framework in MLs, is used in the foundation to build a feature pyramidal network (FPN) to produce characteristic maps of varied sizes. The feature map from the preceding phase is used with anchoring with bounding box regression analysis to create a network of proposed regions (RPN). Utilising the results of the regression, classifier, and segmented obligations, all three sections of ROI Align—classification, regression, and segmented—are carried out. The processes of ROI alignment, loss parameter export, and image export were conducted [2]. ResNet is a commonly utilised architecture in RA-CNN for object detection and segmentation [7]. Previous research has shown that RA-CNN outperforms other deep learning architectures for wound segmentation. Extensive hyperparameter tuning was carried out to optimize the performance of the CNN model.
In this work, we did not investigate the usage of different CNNs; instead, we were primarily concerned with assessing the performance of RA-CNN for injury separation. An enhanced neural network model for categorising objects is called RA-CNN. It has the capacity to carry out examples of segmentation, regression evaluation, and categorisation all at once. As a consequence, several kinds of losses are generated throughout the period of training, making it difficult to choose the best epoch for assessing the model. RA-CNN losses fall into instruction and verification classifications, each with six subgroups. Mrcnn_class, Mrcnn_Bbox, and Mrcnn_Anch all have a direct impact on the assessment results. The best mrcnn_class efficiency was assessed in the research’s 45th period in a training of RA-CNNs representation. In the event of a tie, the period with the most contemporary period was used. Because classifications have an established baseline but extrapolation may add mistakes as a consequence of hand annotations, mrcnn_bbox and the period with the best mrcnn_anch effectiveness were chosen.
Furthermore, a method of early stopping was employed to choose participants, and epoch 33 was identified as the point to halt training. The best accuracy in categorisation and a fair degree of reliability in analysis were both shown during this specific period. Despite the fact that there may be ambiguity in calculating wound area, the choice of period is meticulously addressed to reduce mistakes when recognising sicknesses and therapeutic strategies. According to the study’s results, the appropriate assessment period may significantly affect the RA-CNN prediction accuracy and therapeutic usefulness.

3.4. Detecting ROI and Measuring the Area

As part of implementing and validating the area calculation algorithm using a calibrator, an expert collected 30 samples. These samples were acquired by placing a blue adhesive marker measuring 2 cm by 2 cm in the same plane as the lesion. Conventional planimetry, the Kundin technique, and digital planimetry, which makes use of photo editing software, were all used to test the preciseness of the technique. Conventional planimetry involves drawing the injury on clear acrylic. The technique’s precision was calculated by measuring the discrepancies among the tests acquired using various approaches.
Before evaluating the algorithm’s performance with real images, a laboratory study was conducted using pre-made samples. Geometric shapes were replicated on a sheet of paper using a marker to simulate these samples. Multiple captures were taken from different viewpoints to evaluate the achievable precision and error rate. An analysis among the estimated region, the region acquired using Visitrak, with data space generated through digitised planimetry, additionally employing a graphic editor, was performed in order to verify the preciseness of the readings taken. The marker’s location was further verified by contrasting its own edges and slopes.
Let us represent the wound image data as D, where
D = { ( x i , y i , z i ) }   f o r   i = 1,2 , , N
wherein:
  • xi represents the i-th wound image,
  • yi represents the ground truth size category label for wound i (e.g., small, medium, large), and
  • zi denotes the ground truth tissue type label for wound i (e.g., granulation tissue, epithelial tissue).
The Region Anchored CNN Model can be represented by the following mathematical formulation:
  • Feature Extraction: F(xi) denotes the feature extraction process, which transforms the input wound image x_i into a feature map F(xi).
  • Region Proposal Network (RPN): R (F(xi)) generates region proposals (ROI) based on the featured maps F(xi), denoted as R (F(xi)).
  • ROI Alignment: The featured mapping F(xi) and ROI proposals R(F(xi)) undergo ROI alignment to extract aligned features for further processing.
  • Size Classification: S(R(F(xi))) represents the size classification branch, which takes the aligned features and predicts the size category label yi for wound i.
  • Tissue Classification: T(R(F(xi))) represents the tissue classification branch, which takes the aligned features and predicts the tissue type label zi for wound i.
  • Loss Functions: The model is trained using appropriate loss functions for size and tissue classification tasks. Let Ls and Lt represent the size and tissue classification loss functions, respectively.
The overall objective function for training the Region Anchored CNN Model can be expressed as follows:
J = Σ [ i = 1   t o   N ] ( L s ( S ( R ( F ( x i ) ) ) , y i ) + L t ( T ( R ( F ( x i ) ) ) , z i ) )
where J represents the overall loss function to be minimized during training, and the summation is taken over all N wound images in the dataset D.
The equations for the size and tissue classification tasks using cross-entropy loss functions are as follows:
Let xi be the input wound image, FR(xi) be the aligned features obtained after ROI alignment, and Ps(yi | F_R(xi)) be the predicted probability distribution over size categories for wound i.
The cross-entropy losing a function:
L s ( y i , P s ( y i | F R ( x i ) ) ) = Σ [ j = 1   t o   N s ] y i , j × l o g ( P s ( y i , j | F R ( x i ) ) )
wherein:
  • yi is the ground truth one-hot encoded vector representing the size category of wound i.
  • Ns is the no. of size categories.
  • yi,j is the j-th element of the ground reality size category vector yi.
  • Ps(yi,j | FR(xi)) is the predict prospect for the j-th size category for wound i.
For tissue, Classification representation: let Pt(zi | FR(xi)) be the predicted probability distribution over tissue types for wound i.
The cross-entropy losing a function:
L t ( z i , P t ( z i | F R ( x i ) ) ) = Σ [ k = 1   t o   N t ] z i , k * l o g ( P t ( z i , k | F R ( x i ) ) )
wherein:
  • zi is the ground truth one-hot encoded vector representing the tissue type of wound i.
  • Nt is the no. of tissue types.
  • zi,k is the k-th element of the ground reality tissue type vector zi.
  • Pt(zi,k | FR(xi)) is the predicted probability for the k-th tissue type for wound i.
These equations represent the standard cross-entropy loss functions for multi-class classification tasks, such as size and tissue classification.

4. Empirical Analysis of the Proposed Methodology

The subsequent sections will present the results obtained from different stages of the research. To ensure the credibility of the findings, a comparison was made with results from previously established and validated methodologies.

4.1. Detection of ROI and Region Quantity

Super cells and K-mean methods were compared for rapidity, precision, and efficacy in segmentation.
Initially, a projection analysis was conducted, utilizing a predetermined number of segments (N) and a fixed parameter (Sigma), through manual assignment. After evaluating the obtained outcomes, we recommend that polynomial regression be employed for optimal calculation in different scenarios. Specifically, using 50, 100, and 200 segments (N) and 1, 3, and 5 Sigma values, it was found that the combination of 100 Super pixels (SP) and Sigma 3 produced the most favourable results. For the inadequately segmented images, a selection was made to reprocess them using Super pixels with alternative values of 150, 200, and 300. To estimate the appropriate N segments and Sigma values, a polynomial regression approach is proposed, taking into consideration the diverse types of wounds and their resolutions. Initially, a projection analysis was carried out by manually assigning a fixed number of segments (N) and a parameter (Sigma). After analysing the obtained outcomes, polynomial regression is recommended for optimal calculation in each scenario. Specifically, using 50, 100, and 200 segments (N) and 1, 3, and 5 Sigma values, it was determined that the combination of 100 Super pixels (SP) and Sigma 3 achieved the most favourable results as in Table 2.
Table 2. Analysis of the suggested algorithm with regard to trained algorithms.
For the images that were not adequately segmented, a subset was chosen for reprocessing using Super pixels with alternative values of 150, 200, and 300. To estimate the appropriate number of segments (N) and Sigma values, a polynomial regression approach is proposed, taking into consideration the various types of wounds and their resolutions.
A desktop application was employed by a clinical expert to accurately outline the shape under high-resolution conditions. The medical expert measured the identical incisions utilising planimetry and a Visitrak instrument as in Figure 7, which were then utilised to determine the area and to verify the technique’s efficacy. Height and width were measured using the Kundin measure. When calculating the REs of the measurements, the total variance was subtracted from the result of the expert’s Digital Tool. Groups of data were established. The null hypothesis asserts that the median variations are insignificant whereas the alternative explanation argues that some are of statistical importance.
Figure 7. Proposed model, planimetry, Kundin, and photo editor comparison.
We can dismiss the null argument that the variances in the median of REs are not statistically important if the p-value is 0.05 or lower. This proves that the system utilised affects the region being measured numerically. We ran tests with p-values larger than 0.05 to determine whether there was a statistically significant disparity among the mean of various groups, that were categorised by injury length as in Figure 8. Due to the data’s non-normal dispersion and different group dimensions, we used Matlab’s Kruskal–Wallis statistic and ANOVAs. The data acquired using conventional planimetry, the Kundin technique, and our Digital approaches methodology all showed substantial variations in the findings (p-value 0.04 in the case in all instances). High reliability between ratings, with values ranging 0.99 to 1.00, appeared as well for every metrics. Consequently, considering digital planimetry as the most accurate reference, our proposed method demonstrates the closest relative variance, as in Table 3. For qualitative factors, absolute percentages were calculated. In this research, repeated measurements of injury pictures are symbolised by the ICC, which stands for (Intraclass Reliability Correlation). p-values under 0.05 were taken into account when determining statistical significance (shown in Figure 9).
Figure 8. Outcome of the proposed result. (a) Normal wound of trauma, (b) Ground truth, (c) Superpixel contour, (d) RoI phase, (e) Tissue classification, (f) RA-CNN results.
Table 3. Computation of errors.
Figure 9. Correlating Contour vs. area result. (a) Proposed vs. digital planimetry of n = 28. (b) Proposed vs. manual planimetry of n = 28.

4.2. Tissue Classification

Evaluating the equation used in the pre-trained boxes and our network’s parameters, the table below shows tissues categorisation accuracy. The whole collection was separated into a training collection that had 80% of the information (n = 727) with a set to be tested that contained 20% of the information after being normalised to guarantee a comparable proportion of every tissue type. The proposed model exhibited the highest performance for both training and testing, achieving an accuracy of 85% during training and 99% in predicting necrotic tissue. Except for InceptionResNetV2, that produced a success rate of 49%, each of the other models had the best achievement, with an efficiency of more than 70%. The ROC curves indicated that necrotic tissue consistently had the highest degree of accuracy in every instance, meaning it was the tissue class most accurately predicted by the models as in Figure 10.
Figure 10. Comparison of ROCs, (a) VGG16, (b) InceptionV2, (c) InceptionV3, (d) proposed model.
In terms of classification performance, the enhanced RA-CNN architecture utilizing the LRELU activation function demonstrates superior overall performance when applied to the transfer learning of the small wound dataset.

4.3. Discussion

Various methods [7,22,23,24,25,26] have been examined to simplify and expedite the assessment of wounds by medical professionals compared to alternative approaches. A decision-support technology [27,28,29] that incorporates artificial vision with AIs features may aid in the categorisation of the injury and cautious, assisting in avoiding complicated operations even if the healthcare professional continues to have the ultimate say. In terms of accuracy, noncontact techniques for evaluating patient lacerations were found to outperform other methods, including conventional techniques. Our approach combines the identification of relevant areas, calculation of their size, and classification of the tissue involved. Supplementing the evaluation with photographs taken from different perspectives is essential. Placing the marker incorrectly can also introduce challenges in accurately measuring the wound. The doctor must take the shot at a 90-degree angle and keep the marker in the identical planes as the image, even if systems could identify aberrations. There may be major computing challenges if the marker shifts and ends up on another surface. In addition, brightness has to be considered. Tissue categorisation algorithms may make mistakes if specimens are taken under different illumination circumstances.
Despite these hazards, we think improved artificial intelligence and computational learning techniques will ultimately have an integral part in the practice of medicine. This is primarily due to the value these tools bring to healthcare professionals and the overall improvement they offer in patient care.

5. Conclusions

In this study, different computer vision-based algorithms were evaluated to perform area calculation, boundary determination, and tissue classification of wounds using cost-effective hardware. The results showed that a smartphone can produce reliable findings, which can be beneficial in clinical practice. The categorisation of the material around the injury and appraisal of the injured region provides medical experts with in-depth knowledge about the injury’s present condition and probable future development. These variables are crucial in defining the concept of healing time, understanding the factors influencing the healing process, and assessing treatment effectiveness. Furthermore, the RA-CNN learning model was trained and successfully produced tissue categorisation results. The digitization of tissue quantity in a wound represents a significant advancement in wound treatment. Identifying necrotic cells is particularly vital during the healing process, as it determines the need for appropriate treatment. The trained RA-CNN model achieved an RV score of 0.01–15.1 during evaluation, with overall testing results indicating high recall (0.90), precision (0.96), F1 score (0.90), and Overall Accuracy (96.7). These results show the RA-CNN model’s capability to identify and categorise diseases, resulting in an efficient and convenient tool for doctors to raise the bar for therapeutic healing of wounds.
Future work: To implement AI-assisted surgical wound closure systems that analyse wound characteristics in real-time during surgery, guiding surgeons to achieve better wound closure and reduce post-operative complications.

Author Contributions

Conceptualization, Y.J.; Methodology, S.S. and Y.J.; Software, S.S. and Y.J.; Investigation and Resources, S.M.A.A.; S.S.; Data curation, H.A.; Writing—original draft, Y.J.; Writing—review and editing, S.S.; Visualization, S.S. and Y.J.; Project administration, S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

Princess Prince sattam bin Abdulaziz University Researchers Supporting Project number (PSAU/2023/R/1444), Prince Sattam Bin Abdul Aziz University, Saudi Arabia, Wadi Aldwassir, 1190.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

FNRFalse Negative Rate
FPRFalse Positive Rate
RVRelative Variance
DDataset containing wound images
NTotal number of wound images in dataset
xi, yi, ziInput wound image, Ground truth size category, Ground truth tissue type for wound i
RNumber of relevant instances in dataset
ONumber of observed instances in dataset
TTotal number of instances in dataset
JOverall loss function to be minimized during training
Ps(yi | FR(xi))Predicted probability distribution over size categories for wound i
Ls(yi,Ps(yi | FR(xi)))Cross-entropy loss function for size classification
yi,jj-th element of the ground truth size category vector yi
NsNumber of size categories
Pt(zi | FR(xi))Predicted probability distribution over tissue types for wound i
Lt(zi,Pt(zi | FR(xi)))Cross-entropy loss function for tissue classification
zi,kk-th element of the ground truth tissue type vector zi
NtNumber of tissue types

References

  1. He, J.; Baxter, S.L.; Xu, J.; Xu, J.; Zhou, X.; Zhang, K. The practical implementation of artificial intelligence technologies in medicine. Nat. Med. 2019, 25, 30–36. [Google Scholar] [CrossRef]
  2. Mamone, V.; Fonzo, M.D.; Esposito, N.; Ferrari, M.; Ferrari, V. Monitoring wound healing with contactless measurements and augmented reality. IEEE J. Transl. Eng. Health Med. 2020, 8, 2700412. [Google Scholar] [CrossRef]
  3. Dong, H.; Yang, G.; Liu, F.; Mo, Y.; Guo, Y. Automatic brain tumor detection and segmentation using U-net based fully convolutional networks. In Medical Image Understanding and Analysis; Springer: Berlin/Heidelberg, Germany, 2017; pp. 506–517. [Google Scholar] [CrossRef]
  4. Novikov, A.A.; Lenis, D.; Major, D.; Hladuvka, J.; Wimmer, M.; Buhler, K. Fully convolutional architectures for multiclass segmentation in chest radiographs. IEEE Trans. Med. Imaging 2018, 37, 1865–1876. [Google Scholar] [CrossRef]
  5. Kulurkar, P.; kumar Dixit, C.; Bharathi, V.C.; Monikavishnuvarthini, A.; Dhakne, A.; Preethi, P. AI based elderly fall prediction system using wearable sensors: A smart home-care technology with IOT. Meas. Sens. 2023, 25, 100614. [Google Scholar] [CrossRef]
  6. Lindholm, C.; Searle, R. Wound management for the 21st century: Combining effectiveness and efficiency. Int. Wound J. 2016, 13 (Suppl. S2), 5–15. [Google Scholar] [CrossRef]
  7. Palanisamy, P.; Padmanabhan, A.; Ramasamy, A.; Subramaniam, S. Remote Patient Activity Monitoring System by Integrating IoT Sensors and Artificial Intelligence Techniques. Sensors 2023, 23, 5869. [Google Scholar] [CrossRef]
  8. Jung, K.; Covington, S.; Sen, C.K.; Januszyk, M.; Kirsner, R.S.; Gurtner, G.C.; Shah, N.H. Rapid identification of slow healing wounds. Wound Repair. Regen. 2016, 24, 181–188. [Google Scholar] [CrossRef]
  9. Dabas, M.; Schwartz, D.; Beeckman, D.; Gefen, A. Application of artificial intelligence methodologies to chronic wound care and management: A scoping review. Adv. Wound Care 2023, 12, 205–240. [Google Scholar] [CrossRef] [PubMed]
  10. Abubakar, A.; Ugail, H.; Bukar, A.M. Can machine learning be used to discriminate between burns and pressure ulcer? In Intelligent Systems and Applications: Proceedings of the 2019 Intelligent Systems Conference (IntelliSys); Springer International Publishing: Belfast, UK, 2020; Volume 2, pp. 870–880. [Google Scholar]
  11. Goyal, M.; Reeves, N.D.; Rajbhandari, S.; Ahmad, N.; Wang, C.; Yap, M.H. Recognition of ischaemia and infection in diabetic foot ulcers: Dataset and techniques. Comput. Biol. Med. 2020, 117, 103616. [Google Scholar] [CrossRef] [PubMed]
  12. Dang, L.; Li, J.; Bai, X.; Liu, M.; Li, N.; Ren, K.; Sun, J. Novel Prediction Method Applied to Wound Age Estimation: Developing a Stacking Ensemble Model to Improve Predictive Performance Based on Multi-mRNA. Diagnostics 2023, 13, 395. [Google Scholar] [CrossRef] [PubMed]
  13. Wang, S.H.; Shyu VB, H.; Chiu, W.K.; Huang, R.W.; Lai, B.R.; Tsai, C.H. An Overview of Clinical Examinations in the Evaluation and Assessment of Arterial and Venous Insufficiency Wounds. Diagnostics 2023, 13, 2494. [Google Scholar] [CrossRef]
  14. Shenoy, V.N.; Foster, E.; Aalami, L.; Majeed, B.; Aalami, O. Deepwound: Automated postoperative wound assessment and surgical site surveillance through convolutional neural networks. In Proceedings of the 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018; pp. 1017–1021. [Google Scholar]
  15. Alzubaidi, L.; Fadhel, M.A.; Oleiwi, S.R.; Al-Shamma, O.; Zhang, J. DFU_QUTNet: Diabetic foot ulcer classification using novel deep convolutional neural network. Multimed. Tools Appl. 2019, 79, 15655–15677. [Google Scholar] [CrossRef]
  16. Rostami, B.; Anisuzzaman, D.M.; Wang, C.; Gopalakrishnan, S.; Niezgoda, J.; Yu, Z. Multi-class wound image classification using an ensemble deep CNN-based classifier. Comput. Biol. Med. 2021, 134, 104536. [Google Scholar] [CrossRef]
  17. Sarp, S.; Kuzlu, M.; Wilson, E.; Cali, U.; Guler, O. A highly transparent and explainable artificial intelligence tool for chronicwound classification: XAI-CWC. Preprints 2021, 2021010346. [Google Scholar] [CrossRef]
  18. Bahar, E.; Yoon, H. Modeling and Predicting the Cell Migration Properties from Scratch Wound Healing Assay on Cisplatin-Resistant Ovarian Cancer Cell Lines Using Artificial Neural Network. Healthcare 2021, 9, 911. [Google Scholar] [CrossRef] [PubMed]
  19. Asokan, R.; Preethi, P. Deep learning with conceptual view in meta data for content categorization. In Deep Learning Applications and Intelligent Decision Making in Engineering; IGI Global: Hershey, PA, USA, 2021; pp. 176–191. [Google Scholar]
  20. Marijanović, D.; Nyarko, E.K.; Filko, D. Wound detection by simple feedforward neural network. Electronics 2022, 11, 329. [Google Scholar] [CrossRef]
  21. Adnan, M.; Asif, M.; Ahmad, M.B.; Mahmood, T.; Masood, K.; Ashraf, R.; Faisal, C.N. An Automatic Wound Detection System Empowered by Deep Learning. J. Phys. Conf. Ser. 2023, 2547, 012005. [Google Scholar] [CrossRef]
  22. Carrión, H.; Jafari, M.; Bagood, M.D.; Yang, H.Y.; Isseroff, R.R.; Gomez, M. Automatic wound detection and size estimation using deep learning algorithms. PLoS Comput. Biol. 2022, 18, e1009852. [Google Scholar] [CrossRef]
  23. Scebba, G.; Zhang, J.; Catanzaro, S.; Mihai, C.; Distler, O.; Berli, M.; Karlen, W. Detect-and-segment: A deep learning approach to automate wound image segmentation. Inform. Med. Unlocked 2022, 29, 100884. [Google Scholar] [CrossRef]
  24. Preethi, P.; Asokan, R. Neural network oriented roni prediction for embedding process with hex code encryption in dicom images. In Proceedings of the 2nd International Conference on Advances in Computing, Communication Control and Networking (ICACCCN), Greater Noida, India, 18–19 December 2020; pp. 18–19. [Google Scholar]
  25. Anisuzzaman, D.M.; Wang, C.; Rostami, B.; Gopalakrishnan, S.; Niezgoda, J.; Yu, Z. Image-based artificial intelligence in wound assessment: A systematic review. Adv. Wound Care 2022, 11, 687–709. [Google Scholar] [CrossRef] [PubMed]
  26. Wu, J.M.; Tsai, C.J.; Ho, T.W.; Lai, F.; Tai, H.C.; Lin, M.T. A unified framework for automatic detection of wound infection with artificial intelligence. Appl. Sci. 2020, 10, 5353. [Google Scholar] [CrossRef]
  27. Cukjati, D.; Reberšek, S.; Miklavčič, D. A reliable method of determining wound healing rate. Med. Biol. Eng. Comput. 2001, 39, 263–271. [Google Scholar] [CrossRef] [PubMed]
  28. Georgilakis, P.S.; Hatziargyriou, N.D.; Doulamis, N.D.; Doulamis, A.D.; Kollias, S.D. Prediction of iron losses of wound core distribution transformers based on artificial neural networks. Neurocomputing 1998, 23, 15–29. [Google Scholar] [CrossRef]
  29. Chin, C.L.; Li, C.Y.; Lai, Y.M.; Chen, T.; Sun, T.Y.; Lin, J.C. CWD2GAN: Generative Adversarial Network of Chronic Wound Depth Detection for Predicting Chronic Wound Depth. Int. J. Pattern Recognit. Artif. Intell. 2023, 37, 2351004. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.