Next Article in Journal
Two Birds with One Stone: Hepatocellular Carcinoma and Small Cell Lung Cancer Imaged with [18F]Fluorocholine Positron Emission Tomography/Computed Tomography
Next Article in Special Issue
Accurate Intervertebral Disc Segmentation Approach Based on Deep Learning
Previous Article in Journal
Differential Protein Expression Patterns of HOXA13 and HOXB13 Are Associated with Bladder Cancer Progression
Previous Article in Special Issue
An Anatomical Template for the Normalization of Medical Images of Adult Human Hands
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Diabetic Foot Ulcer Recognition Using Multi-Level Thermographic Image Data

1
Department of Electrical and Computer Engineering, COMSATS University Islamabad, Lahore Campus, Lahore 54000, Pakistan
2
Department of Computer Engineering, Aligarh Muslim University, Aligarh 202002, India
3
Department of Computer Science and Engineering, Meerut Institute of Engineering and Technology, Meerut 250005, India
4
Department of Business Administration, College of Business Administration, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(16), 2637; https://doi.org/10.3390/diagnostics13162637
Submission received: 27 June 2023 / Revised: 29 July 2023 / Accepted: 6 August 2023 / Published: 10 August 2023

Abstract

:
Lower extremity diabetic foot ulcers (DFUs) are a severe consequence of diabetes mellitus (DM). It has been estimated that people with diabetes have a 15% to 25% lifetime risk of acquiring DFUs which leads to the risk of lower limb amputations up to 85% due to poor diagnosis and treatment. Diabetic foot develops planter ulcers where thermography is used to detect the changes in the planter temperature. In this study, publicly available thermographic image data including both control group and diabetic group patients are used. Thermograms at image level as well as patch level are utilized for DFU detection. For DFU recognition, several machine-learning-based classification approaches are employed with hand-crafted features. Moreover, a couple of convolutional neural network models including ResNet50 and DenseNet121 are evaluated for DFU recognition. Finally, a CNN-based custom-developed model is proposed for the recognition task. The results are produced using image-level data, patch-level data, and image–patch combination data. The proposed CNN-based model outperformed the utilized models as well as the state-of-the-art models in terms of the AUC and accuracy. Moreover, the recognition accuracy for both the machine-learning and deep-learning approaches was higher for the image-level thermogram data in comparison to the patch-level or combination of image–patch thermograms.

1. Introduction

Insulin insufficiency in the body causes diabetes mellitus (DM), which results in high blood glucose (hyperglycemia) for an extended period. Uncontrolled diabetes for a long period of time can lead to complications such as nephropathy, retinopathy, Charcot foot development, amputation, or even death [1]. Uncontrolled DM damages the nerves; if the nerves in the legs or feet are damaged, it causes a lack of feeling called sensory diabetic neuropathy. When a patient does not feel a sore or cut in his foot due to neuropathy, that cut causes infection and worsens the foot condition. The other situation is the low flow of blood. Peripheral vascular disease causes low blood flow in the arms and legs. If the cut is not healing due to low blood flow, there is a risk of developing ulcers. DFU (diabetic foot ulcer) is most common in diabetic patients; more than 15% of patients face this problem [2]. An illustration of DFU is shown in Figure 1.
Diabetic foot issues are expensive and have a negative impact on one’s quality of life. This may be prevented or considerably delayed in many situations by undertaking a risk assessment and inspection of diabetes patients’ foot health at an early stage. For that purpose, temperature may have an impact. Diabetics’ plantar foot temperature may fluctuate due to neuropathy, ischemia, or infection. Temperature differences of more than 2.2 °C (4 °F) between the right and left foot are considered abnormal, where the normal difference is typically less than 1°C [3,4,5]. With the use of a thermal-imaging camera, problems may be identified early, saving time and money in the long run. Infrared thermography may be utilized to produce a clear image of the thermal energy released by the site being monitored in real time if the temperature is above absolute zero [6,7,8]. Thermography is a non-invasive, non-contact, cost-effective, rapid, and painless means of screening the patient’s skin temperature. This imaging technique can detect temperature changes on human skin.
To detect the ulcer, the thermogram needs to be assessed by a professional expert. The availability of such experts is a challenge, particularly in remote areas. Therefore, several research studies have targeted the provision of an automatic DFU recognition system. Few among them utilized foot thermograms [9,10] while many considered visible-band (RGB camera) images [11,12,13,14,15,16,17,18]. In this study, the thermogram images of diabetic foot are considered for DFU detection. For the recognition of DFU, the experiments are carried out at three levels of thermogram data: image-level, patch-level, and a combination of image–patch thermogram data. To detect the DFU foot, pre-trained deep-learning models are employed using transfer learning. Moreover, a problem-oriented, custom, CNN-based, computationally light model is developed to compare with the state-of-the-art results. For a comprehensive and comparative analysis, classical feature-based recognition using machine-learning techniques is also carried out. The details of the rest of the paper are as follows: Section 2 presents the background work and related studies; Section 3 includes the dataset and augmentation details. The methodology is discussed in Section 4. The results and discussions are presented in Section 5. The conclusion is added in Section 6.

2. Background

Machine-learning and deep-learning techniques have gained a lot of interest in recent years for diabetic foot ulcer monitoring and diagnosis in patients with neuropathic diabetes. Much research has been carried out recently for DFU recognition and classification [12,13,15,16,18,19]. However, those studies considered the visible-band images of diabetic feet. Most of these studies employed deep-learning approaches for DFU diagnosis. In contrast, fewer studies judged the thermographic image data. Since this study is focused on the thermographic image data, the literature related to DFU classification in thermograms is discussed.
In the study [9], the researchers compared different machine-learning and deep-learning models. Automatic segmentation and ROI with feature extraction were represented by a fuzzy entropy set with a histogram-based segmentation method for optimization. After augmentation, the data were trained and tested. The SVM produced the lowest AUC (area under the curve), sensitivity, and accuracy. The proposed network model DFTNet and the common ANN had the best performances, using AUC values of 0.8533 and 0.8333%, respectively. In a study by [20], the quantity of pixels with a temperature exceeding 2.2 °C was measured to compute the region of interest. The produced segmentation masks were 99.25% accurate in detecting the absence of a foot sole, 98.83% accurate in constructing a bounding box, and 94.95% accurate in detecting the presence of a foot sole. In another study [14], the authors collected the IRT (infrared thermograph) images and made a dataset with 39 ischemic DFU patients, of which 14 had active ischemic wounds and the remainder had healing wounds. The image ROI abstract was classified by ANN, kNN (k-nearest neighbor), and SVM (support vector machines), while the image was decomposed using DST (discrete wavelength transform) and HOS (higher-order spectra). The best achieved accuracy reported was 98.39% using the SVM classifier. Researchers in study [21] took thermograms in controlled environments with a homogeneous background and used k-means clustering, and a further approach on every foot for foot segmentation. The identification of ulceration on the image was different pixel to pixel and a thresholding technique was used. They successfully differentiated between ulcer and non-ulcer wounds with the help of a classifier and segmented based on a machine-learning model with 91.8% sensitivity, 98.4% specificity, and 91.1% accuracy. In another study [22], the authors used a controlled environment with a room temperature of 20 °C for thermogram images and foot segmentation. A temperature threshold matrix was created, as well as an additional method established to label each foot. A pattern spectrum with thresholding techniques was used to identify the ulceration in the image. They claimed results between the risk and non-risk zones with the help of a classifier and performed segmentation based on a machine-learning model with a sensitivity of 97.33% and specificity of 91.33%. In another study [23], analysis of DFU thermograms with machine learning was performed where the researcher used controlled environments at a room temperature of 20 °C and humidity of 55% for thermogram and foot segmentation. The health-care expert identified the ulceration in the image. The SVM classifier and wavelet characteristics vector were used. They produced results between DF and non-DF with the help of a classifier with an accuracy of 89.39%, sensitivity of 81.81%, and specificity of 96.97%. The detection of DFU thermograms with machine learning has been presented in [24], in an uncontrolled environment with respect to room temperature, illumination, and close-ups for the image of the thermogram. With the temperature filter, the average temperature and threshold were used for diagnostic purposes with the help of a machine-learning-based classifier. They claimed results between DF and non-DF based on a machine-learning model with a non-risk class sensitivity of 91.32% and specificity of 91.84% and ulcer class sensitivity of 90.29%, accuracy of 90.28, and specificity of 90.28%. The authors in study [10] used infrared imaging to detect abnormalities in foot segmentation and registration. They concluded that the ACWE (active contour without edges) method produced quite good results. Automatic pre-symptomatic ulcer detection was performed to determine the clinically relevant difference in temperature between the feet, which was 2.2 °C. The researchers in [7] used infrared imaging and for image decomposition, they examined the ROI of complete feet and mean temperatures. In individuals with localized difficulties, the ipsilateral and contralateral foot and mean temperatures are the same. When compared to a similar area in the contralateral foot and the mean of the entire ipsilateral foot, the ROI temperature was greater than 2 °C. The average temperature difference between both ipsilateral and contralateral feet was greater than 3 °C in patients with widespread problems. In study [25], infrared imaging and clinical foot assessments were presented. For image decomposition, they examined the ROC curve. With a 76% sensitivity and 40% specificity, the contralateral locations had a difference of 2.2 °C between each other, which showed the best cut-off value for diagnosing diabetic foot. The variation of 3.5 °C between the mean temperature of the right and left foot was shown to be the best cut-off value for determining the urgency for treatment, with an 89% sensitivity. In this study [26], the authors employed infrared imaging to detect anomalies, followed by grayscale characterization and temperature pattern foot segmentation. Then, to pattern the spectrum, mathematical morphology was used as well as a multi-layer perceptron with k-fold validation. The subjects had a butterfly pattern, and the pattern spectrum was like that of ovals and rounds. Quadrant 4 had the greatest mean percentage of pixels for the control group, at 88.05%. Due to the different patterns, the pattern spectrum was abnormal. In quadrant 3, the mean proportion of pixels for the diabetic group was 28.87%, while the authors achieved an average classification rate of 94.33 percent. In this paper [27], they examined a database of dynamic IRT plantar diagnostic images with 39 current diabetic foot ulcer patients. The mean temperature of the region of interest, which corresponds to the important change places of diabetic foot ulcer, was assessed and the images were examined by assessing the mean temperature of the region of interest, which relates to some of the important change places of diabetic foot ulcer. The statistics found no evidence of a significant difference between the thermal asymmetry values and thermal recovery differences in any region of interest, except the one at the medial forefoot. The regions of interest were assessed on both feet, with the value of the thermal asymmetry factored into each one. A decision support system was constructed using the database and analytical results to classify the data and examine the accurate identification of the DFU using machine-learning methods such as ANN, kNN, and SVM. The best overall results were achieved with a kNN of 5 neighbors.

3. Data and Augmentation

A public dataset of thermograms is used in this study [3]. These data comprise 334 plantar thermograms obtained from 122 individuals diagnosed with diabetes mellitus (DM) and 45 individuals not diagnosed with diabetes (control group). In the DM group, there were 16 female and 29 male subjects aged between 20 and 35 years. Moreover, there were 89 females and 33 males with their age ranging from 45–65 years in the control group. The subjects were recruited as volunteers from the city of Puebla, Mexico, and thermogram acquisition was carried out over a period of three years (2012–2014). A sample pair of feet from each group is shown in Figure 2.
Each thermogram includes four more images representing plantar angiosomes. Those are considered as patches in this study. A sample of full images and corresponding patch images (plantar angiosomes) from both categories are shown in Figure 3. For infrared (IR) image acquisition, the subject was laid on a bed with an IR camera at a distance of one meter from the feet [3]. To avoid sensing the temperature from the rest of the body, an IR obstructive material was placed. Two IR camera FLIR E60 and FLIR E6 were used at a room temperature of 20 ± 1 C [3]. Since the database includes the segmented foot and patch RGB images, those are used as they are without any pre-processing. However, data augmentation is utilized to increase the dataset size as well as to balance the classes. Augmentation is carried out by rotating the images at 90°, 180°, and 270°, as well as by horizontal flip, vertical flip, and both horizontal and vertical flip simultaneously. Image-level augmentation is performed by making 500 samples for each class and patch-level augmentation is performed to prepare 1500 samples per class. The class-wise detail of the thermograms is summarized in Table 1.

4. Methodology

Computer-aided diagnostic techniques assist medical practitioners in being able to diagnose with a higher confidence. Machine-learning and deep-learning techniques have been of interest for utilization as a decision support system.

4.1. Machine-Learning Approaches

Traditional machine-learning techniques have been considered for the classification of thermograms in the literature. Based on their proven performance in the literature, including medical diagnosis, several of them were considered for this study for comprehensive comparative analysis including SVM, random forest (RF), multi-layer perceptron (MLP), naive Bayes, kNN, XGBoost, AdaBoost, and bagging.
  • Support Vector Machines: Support vector machine may be utilized both for regression and classification problems [28]. For classification, it is a classifier with a goal to locate a hyperplane separating the two classes with a large margin.
  • Random Forest: It is a classifier with several decision trees on various subsets of the provided dataset that takes the average to enhance the predicted accuracy of that dataset [29]. To anticipate the ultimate output, the random forest collects guesses from all of its trees and combines them into a single prediction. Overfitting can be avoided by having a larger number of trees to choose from when making a model.
  • kNN: k-nearest neighbor algorithm assigns the class to the test sample based on the nearest neighbors with the largest majority [30]. Being nearest depends on the distance metric which is normally the Euclidean distance or absolute distance.
  • Naive Bayes: To categorize the data, a naive Bayes classifier applies concepts from probability theory [31]. The theorem developed by Bayes is utilized by the naive Bayes classification algorithms. The most important takeaway from Bayes’ theorem is that the probability of an event can be recalculated whenever new evidence is added to the mix.
  • XGBoost: Extreme gradient boosting is built on supervised machine learning, decision trees, ensemble learning, and gradient boosting [32].
  • AdaBoost: As part of an ensemble method in machine learning, adaptive boosting is a technique known as AdaBoost [33]. AdaBoost’s most frequent algorithm is a decision tree with only one split, known as a decision tree with only one level. Decision stumps are another name for these trees. This algorithm creates a model and equally weighs all the input data points in that model.
  • Bagging: An ensemble meta-estimator, bagging classifiers fit base classifiers on random subsets of the original dataset and then aggregate their individual predictions (either by voting or average) to generate a final forecast [34].

4.2. Feature Extraction

The machine-learning classifiers discussed above require feature extraction. There are many features in the literature that have been used for computer vision and pattern recognition tasks; a few of the popular ones among them are used in this study including local binary pattern, gray level cooccurrence matrix, histogram of oriented gradients, and Gabor features to be used with machine-learning classifiers.
  • Local Binary Patterns (LBP): Each pixel in a picture is labelled using the local binary patterns operator by thresholding a 3x3 neighborhood surrounding each pixel with the center value [35]. These classes are used to label pixels. Each result is assigned a binary value, which is either a 1 or a 0, depending on whether the surrounding pixels are equal or greater than the center pixels.
  • Histogram of Oriented Gradients (HOG): The purpose of HOG is to detect the presence of a particular object oriented at a specified direction [36]. The magnitude of pixel orientation data is weighted to establish the criteria for characterizing an item in these attributes.
  • Gabor Filters: They are linear Gabor filters that detect if an image has a certain frequency content within a given region of interest for texture research [37]. Many current vision experts believe that the frequency and orientation representations of Gabor filters are like those perceived in the human eye.
  • Gray Level Cooccurrence Matrix (GLCM): A GLCM is a matrix representing the frequency of cooccurrence of a pair of pixel intensities at a specified distance and angle [38]. The GLCM is computed to extract the texture features from images. Cluster prominence, cluster shade, dissimilarity, energy, entropy, homogeneity, and maximum probability are the GLCM features used in this study.

4.3. Deep-Learning Approaches

Currently, deep-learning approaches, particularly convolutional neural networks, have been extensively employed in computer-aided medical diagnostics. In this study, two deep-learning models are employed via transfer learning including ResNet50 and DenseNet. Moreover, a custom-developed CNN model is proposed.
  • ResNet50: This model [39] was originally trained for 1000 classes using the ImageNet database [40]. The ResNet50 CNN model is adequately efficient on vision tasks, and particularly fits well considering the efficiency as well as complexity. Moreover, this has been widely used as the default choice for deep transfer learning in computer vision tasks. It has 48 convolutional layers, one max-pool, and one average-pool layer. For diabetic foot ulcer recognition, the last three layers’ parameters were updated via transfer learning by network training using thermogram image data. The input image was resized as 224 × 224 to match the ResNet input image resolution requirement. The output was restricted to one neuron, providing the probability for the sample to be recognized as a diabetic group.
  • DenseNet121: DenseNet was developed with the aim to obtain benefit from a deep network while keeping fewer parameters [41]. It improves the accuracy by minimizing the problem of a vanishing gradient. It has 120 convolutional layers and 4 average-pool layers. To use it with diabetic foot thermogram data, the last two layers were used for the parameter update via training. The input image was resized at 224 × 224 for this network also.
  • Proposed CNN Model: In addition to pre-trained networks, a custom CNN model was developed specifically for DF classification. For this purpose, the DFTNet model was adopted as the base model [9]; however, there were major differences. This study utilized the input volume size 180 × 80 × 3; however, the DFTNet used an input volume of 227 × 227 × 3. The learning rate was chosen as 0.001 with the Adam optimizer. The batch size was set to 64. The network architecture diagram is shown in Figure 4. The detailed working of the proposed model is shown in Figure 5.

5. Results and Discussion

For training the classifiers, 80% of the data were used for training and the remaining 20% for test purposes. For performance evaluation, classification measures such as the sensitivity, specificity, precision, accuracy, and F1-score were considered. The mathematical expressions of these measures are as follows:
Sensitivity = TP/(TP + FN)
Specificity = TN/(TN + FP)
Accuracy = (TP + TN)/(TP + TN + FP + FN)
F1 Score = 2TP/(2TP + FP + FN)
where TP: true positive, TN: true negative, FP: false positive, FN: false negative.

5.1. Results of Machine-Learning Approaches

As discussed in the methodology section, the machine-learning approaches used for the classification of thermogram data include SVM, RF, XGBoost, naive Bayes, ADABoost, kNN, and bagging. To compute the results for each of these algorithms, all individual features including HOG, Gabor, GLCM, and LBP were employed. Moreover, a combination of all features was also used.

5.1.1. Classification of Combined Data (Image+Patch)

In this section, the images and patches are combined to form the total data. Then, the features are extracted from these data and used for classification. The receiver operative characteristic (ROC) curve and AUC (area under the curve) results of the machine-learning classifiers are shown in Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10, using individual as well as combined features. Table 2 shows the results in terms of the evaluation metrics for the individual and combined features. It can be observed that a sensitivity score of 0.71 was recorded using the SVM, RF, XGBoost, and bagging classifiers with HOG features. The SVM and bagging produced the best results using both the HOG features and Gabor features. In the case of specificity, again, the SVM produced 0.95 using the GLCM features. The overall best accuracy of 78% was achieved by the SVM classifier while combining all four kinds of features. The highest F1-score was recorded by both the RF and XGBoost classifiers of 0.71. By comparing the ROC curves of classifiers employing different features, it can be observed that the HOG features produced the best results among the other individual features. The overall best AUC of 0.93 was recorded by employing combined features from multiple classifiers including RF, bagging, and naive Bayes. Concretely, the SVM happened to be the best machine-learning classifier regarding recognizing the diabetes class (with the highest sensitivity value observation), while the combined features gave the best accuracy as well as AUC.

5.1.2. Full-Image Thermogram-Based Classification Results

In this section, thermogram image-level data are considered to record the classification results. Considering the results presented in the previous section, further experiments with machine-learning classifiers were restricted to using combined features only. Therefore, the HOG, Gabor, GLCM, and LBP features were extracted from the images and combined to be fed into the classifiers. Table 3 shows the image-level classification results using the combined features. The ROC curves are shown in Figure 11. The best sensitivity 0.642 was recorded using the kNN classifier while the SVM produced the best specificity value of 0.968. The XGBoost classifier achieved the best results for image-level classification with an 85.6% accuracy and F1-score of 0.688. The highest AUC of 0.84 was recorded with the naive Bayes classifier.

5.1.3. Thermogram Patch-Based Classification Results

In this section, only thermogram patch-level data are considered to compute the results. The HOG, Gabor, GLCM, and LBP features were extracted from the image patches and combined to be fed into the classifiers. Table 4 shows the image-level classification results using the combined features. The AdaBoost classifier produced the best sensitivity score and F1-score of 0.638 and 0.609. In contrast, the naive Bayes classifier achieved the highest specificity and accuracy of 0.935 and 78.7%, respectively. The ROC curves of the patch-level results are shown in Figure 12. The highest AUC of 0.84 was observed by the naive Bayes classifier.

5.2. Results of Deep-Learning Approaches

In this section, the results of the CNN-based models are presented. As discussed earlier in the methodology section, the ResNet50 and DesnseNet121 models are used via transfer learning to classify the thermogram data. Moreover, the customized CNN model was developed for DFU classification. For the deep-learning approaches, the results are recorded at image level, at patch level, and by combining the images and patches. The results of both pre-trained models as well as the customized model are shown in Table 5. It can be observed that the proposed CNN model outperformed the pre-trained models in terms of sensitivity at each data representation level while achieving a best value of 0.97 at image level. The ResNet50 model produced the best specificity at patch level and at image–patch level of 0.913 and 0.902, respectively. However, the highest specificity of 0.958 was achieved by the proposed model again. The proposed model achieved the best accuracy among the three deep-learning models with a value of 97.1%, 93%, and 93.3% at image level, patch level, and image-patch level, respectively. The best F1-score of 0.891 was again recorded for the proposed model for all data representation levels. In medical diagnosis, it is critical to reduce false negatives so that the potential patient may be treated on time and the risk of amputation may be reduced. Table 5 presents the sensitivity and specificity analysis of the pre-trained models as well as the custom model. It can be observed that the custom model achieved the highest sensitivity rate and ensured the minimization of false negatives at all three levels. The graphical representation of the results using deep-learning-based models are shown in Figure 13. The accuracy and loss observations for the proposed CNN model are illustrated in Figure 14, Figure 15 and Figure 16 at image level, patch level, and image-patch level, respectively. The proposed custom-developed CNN model was trained on the original foot thermogram data from scratch. In comparison, the pre-trained model was utilized with pre-trained weights originally tuned on ImageNet data; only the weights of the last couple of layers were tuned based on the thermogram data. Therefore, the accuracy of the custom-developed model was better than the pre-trained models.
When comparing the results of the proposed model with the state-of-the-art existing solution [9], the comparison is not straight forward. Firstly, the number of samples used after augmentation with the DFTNet were 10× than the original samples; however, the samples after augmentation were almost 3× which means that only 1/3 of data were used in this study in comparison to [9]. Next, only the patch-level thermogram data were considered in the study [9]. However, this study used data representation and result computation at image level, patch level, and image-patch level. Moreover, this study presents a binary classification between the diabetic foot and control group foot. However, the study [9] segregated the data as five distinct classes, and performed binary classification by taking two classes at a time randomly. Yet, they [9] achieved the best AUC score of 0.8533 and 0.8333 using the DFTNet and ANN, respectively. In comparison, we achieved an AUC score of 0.976, 0.932, and 0.938 for the image-level, patch-level, and image-patch level data, respectively. The DFTNet [9] achieved the best accuracy of 0.853 while the proposed model produced a 0.97 accuracy. Concretely, the comparison with the study [9] is not simple and there are many differences with respect to data selection, data quantity, data representation, and class representation. However, the proposed model outperforms by looking at several possible simpler comparisons.
Regarding the limitations of this study, it is worth mentioning that in all the experiments, including machine-learning and deep-learning classification, the thermogram images of the database and the augmented samples were utilized without any pre-processing. However, since the augmentation was performed only by rotation and flipping, the content of the image was retained. Speaking about the practical applicability of the proposed model, it cannot replace a human expert completely; however, it can provide a reliable second opinion while diagnosing the diabetic foot ulcer diagnosis. In contrast, in remote areas where a medical expert is not available, such a system can play a significant role in decision-making for DFU diagnosis.

6. Conclusions

In this study, the classification of DFU foot is presented via the comprehensive and comparative analysis of machine-learning as well as deep-learning approaches. The thermogram data of DFU and non-DFU feet are used at image level, patch level, and combined image–patch levels. The results of the machine-learning approaches are presented by employing several features extracted from thermograms. The SVM classifier among the machine-learning techniques proved to be the best. The classification results of the machine-leaning models were best when the data at the full-image level were used. Among the pre-trained and the proposed deep-leaning models, the proposed model produced the best results. The comparison of the proposed model with an existing solution based on the used thermogram data was not straight forward. However, while comparing at an abstract level, the proposed model performed better given the constraints.

Author Contributions

Conceptualization, I.K.; methodology, I.K. and A.R.; software, W.A. and A.R.; validation, M.A., S.S. and W.A.; formal analysis, I.K, A.R, M.A, S.S. and W.A..; investigation, I.K., M.A. and S.S.; resources, I.K. and A.R.; writing—original draft preparation, I.K. and A.R.; writing—review and editing, M.A., S.S. and W.A.; supervision, I.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R259), Princess Nourah bint Abdulrahman University, Saudi Arabia.

Institutional Review Board Statement

The study only used a public dataset which was originally collected in another study. No human was involved in this study.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are openly available online: https://ieee-dataport.org/open-access/plantar-thermogram-database-study-diabetic-foot-complications, accessed on 13 February 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, S.; Zhang, K.; Chao, L.; Li, D.; Tian, X.; Bao, H.; Chen, G.; Xia, Y. Exploring the Utility of Radar and Satellite-Sensed Precipitation and Their Dynamic Bias Correction for Integrated Prediction of Flood and Landslide Hazards. J. Hydrol. 2021, 603, 126964. [Google Scholar] [CrossRef]
  2. Iversen, M.M.; Tell, G.S.; Riise, T.; Hanestad, B.R.; Østbye, T.; Graue, M.; Midthjell, K. History of Foot Ulcer Increases Mortality Among Individuals with DiabetesTen-Year Follow-up of the Nord-Trøndelag Health Study, Norway. Diabetes Care 2009, 32, 2193–2199. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Hernandez-Contreras, D.A.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.D.J.; Renero-Carrillo, F.J. Plantar Thermogram Database for the Study of Diabetic Foot Complications. IEEE Access 2019, 7, 161296–161307. [Google Scholar] [CrossRef]
  4. Ring, F. Thermal Imaging Today and Its Relevance to Diabetes. J. Diabetes Sci. Technol. 2010, 4, 857–862. [Google Scholar] [CrossRef] [Green Version]
  5. Bagavathiappan, S.; Philip, J.; Jayakumar, T.; Raj, B.; Rao, P.N.S.; Varalakshmi, M.; Mohan, V. Correlation between Plantar Foot Temperature and Diabetic Neuropathy: A Case Study by Using an Infrared Thermal Imaging Technique. J. Diabetes Sci. Technol. 2010, 4, 1386–1392. [Google Scholar] [CrossRef] [Green Version]
  6. Hernandez-Contreras, D.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.; Gonzalez-Bernal, J. Narrative Review: Diabetic Foot and Infrared Thermography. Infrared Phys. Technol. 2016, 78, 105–117. [Google Scholar] [CrossRef]
  7. Liu, C.; van der Heijden, F.; Klein, M.E.; van Baal, J.G.; Bus, S.A.; van Netten, J.J. Infrared Dermal Thermography on Diabetic Feet Soles to Predict Ulcerations: A Case Study. Adv. Biomed. Clin. Diagn. Syst. XI 2013, 8572, 102–110. [Google Scholar] [CrossRef]
  8. Faust, O.; Rajendra Acharya, U.; Ng, E.Y.K.; Hong, T.J.; Yu, W. Application of Infrared Thermography in Computer Aided Diagnosis. Infrared Phys. Technol. 2014, 66, 160–175. [Google Scholar] [CrossRef]
  9. Cruz-Vega, I.; Hernandez-Contreras, D.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.d.J.; Ramirez-Cortes, J.M. Deep Learning Classification for Diabetic Foot Thermograms. Sensors 2020, 20, 1762. [Google Scholar] [CrossRef] [Green Version]
  10. Adam, M.; Ng, E.Y.K.; Oh, S.L.; Heng, M.L.; Hagiwara, Y.; Tan, J.H.; Tong, J.W.K.; Acharya, U.R. Automated Detection of Diabetic Foot with and without Neuropathy Using Double Density-Dual Tree-Complex Wavelet Transform on Foot Thermograms. Infrared Phys. Technol. 2018, 92, 270–279. [Google Scholar] [CrossRef]
  11. Goyal, M.; Reeves, N.D.; Davison, A.K.; Rajbhandari, S.; Spragg, J.; Yap, M.H. DFUNet: Convolutional Neural Networks for Diabetic Foot Ulcer Classification. IEEE Trans. Emerg. Top. Comput. Intell. 2018, 4, 728–739. [Google Scholar] [CrossRef] [Green Version]
  12. Nanda, R.; Nath, A.; Patel, S.; Mohapatra, E. Machine Learning Algorithm to Evaluate Risk Factors of Diabetic Foot Ulcers and Its Severity. Med. Biol. Eng. Comput. 2022, 60, 2349–2357. [Google Scholar] [CrossRef] [PubMed]
  13. Cao, C.; Qiu, Y.; Wang, Z.; Ou, J.; Wang, J.; Hounye, A.H.; Hou, M.; Zhou, Q.; Zhang, J. Nested Segmentation and Multi-Level Classification of Diabetic Foot Ulcer Based on Mask R-CNN. Multimed. Tools Appl. 2022, 82, 18887–18906. [Google Scholar] [CrossRef]
  14. Fraiwan, L.; AlKhodari, M.; Ninan, J.; Mustafa, B.; Saleh, A.; Ghazal, M. Diabetic Foot Ulcer Mobile Detection System Using Smart Phone Thermal Camera: A Feasibility Study. Biomed. Eng. Online 2017, 16, 117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Xu, Y.; Han, K.; Zhou, Y.; Wu, J.; Xie, X.; Xiang, W. Classification of Diabetic Foot Ulcers Using Class Knowledge Banks. Front. Bioeng. Biotechnol. 2022, 9, 811028. [Google Scholar] [CrossRef]
  16. Ai, L.; Yang, M.; Xie, Z. ACTNet: Asymmetric Convolutional Transformer Network for Diabetic Foot Ulcers Classification. Phys. Eng. Sci. Med. 2022, 45, 1175–1181. [Google Scholar] [CrossRef]
  17. Goyal, M.; Reeves, N.D.; Rajbhandari, S.; Yap, M.H. Robust Methods for Real-Time Diabetic Foot Ulcer Detection and Localization on Mobile Devices. IEEE J. Biomed. Health Inform. 2019, 23, 1730–1741. [Google Scholar] [CrossRef] [Green Version]
  18. Santos, F.; Santos, E.; Vogado, L.H.; Ito, M.; Bianchi, A.; Tavares, J.M.; Veras, R. DFU-VGG, a Novel and Improved VGG-19 Network for Diabetic Foot Ulcer Classification. In Proceedings of the 2022 29th International Conference on Systems, Signals and Image Processing (IWSSIP), Sofia, Bulgaria, 1–3 June 2022. [Google Scholar] [CrossRef]
  19. Al-Garaawi, N.; Ebsim, R.; Alharan, A.F.H.; Yap, M.H. Diabetic Foot Ulcer Classification Using Mapped Binary Patterns and Convolutional Neural Networks. Comput. Biol. Med. 2022, 140, 105055. [Google Scholar] [CrossRef]
  20. Tulloch, J.; Zamani, R.; Akrami, M. Machine Learning in the Prevention, Diagnosis and Management of Diabetic Foot Ulcers: A Systematic Review. IEEE Access 2020, 8, 198977–199000. [Google Scholar] [CrossRef]
  21. Niri, R.; Lucas, Y.; Treuillet, S.; Douzi, H. Smartphone-Based Thermal Imaging System for Diabetic Foot Ulcer Assessment. In Journées d’Etude sur la TéléSanté; Sorbonne Universités: Paris, France, 2019. [Google Scholar]
  22. Liu, C.; van Netten, J.J.; van Baal, J.G.; Bus, S.A.; van der Heijden, F. Automatic Detection of Diabetic Foot Complications with Infrared Thermography by Asymmetric Analysis. J. Biomed. Opt. 2015, 20, 026003. [Google Scholar] [CrossRef] [Green Version]
  23. Vardasca, R.; Magalhaes, C.; Seixas, A.; Carvalho, R.; Mendes, J. Diabetic Foot Monitoring Using Dynamic Thermography and AI Classifiers. In Proceedings of the 3rd Quantitative InfraRed Thermography Asia Conference (QIRT), Tokyo, Japan, 1–5 July 2019. [Google Scholar] [CrossRef]
  24. Adam, M.; Ng, E.Y.K.; Oh, S.L.; Heng, M.L.; Hagiwara, Y.; Tan, J.H.; Tong, J.W.K.; Acharya, U.R. Automated Characterization of Diabetic Foot Using Nonlinear Features Extracted from Thermograms. Infrared Phys. Technol. 2018, 89, 325–337. [Google Scholar] [CrossRef]
  25. Van Netten, J.J.; Van Baal, J.G.; Liu, C.; Van Der Heijden, F.; Bus, S.A. Infrared Thermal Imaging for Automated Detection of Diabetic Foot Complications. J. Diabetes Sci. Technol. 2013, 7, 1122–1129. [Google Scholar] [CrossRef] [Green Version]
  26. Van Netten, J.J.; Prijs, M.; Van Baal, J.G.; Liu, C.; Van Der Heijden, F.; Bus, S.A. Diagnostic Values for Skin Temperature Assessment to Detect Diabetes-Related Foot Complications. Diabetes Technol. Ther. 2014, 16, 714–721. [Google Scholar] [CrossRef] [Green Version]
  27. Khosa, I.; Pasero, E. Defect Detection in Food Ingredients Using Multilayer Perceptron Neural Network; IEEE: Piscataway, NJ, USA, 2014; ISBN 9781479928057. [Google Scholar]
  28. Schölkopf, B. SVMs-A Practical Consequence of Learning Theory. IEEE Intell. Syst. Their Appl. 1998, 13, 18–21. [Google Scholar] [CrossRef] [Green Version]
  29. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  30. Baskett, F.; Shustek, L.J. An Algorithm for Finding Nearest Neighbors. IEEE Trans. Comput. 1975, C-24, 1000–1006. [Google Scholar] [CrossRef]
  31. Cichosz, P. Naïve Bayes Classifier. In Data Min. Algorithms; Wiley Data and Cybersecurity: Hoboken, NJ, USA, 2015; pp. 118–133. [Google Scholar] [CrossRef]
  32. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 13–17 August 2016. [Google Scholar] [CrossRef]
  33. Rätsch, G.; Onoda, T.; Müller, K.R. Soft Margins for AdaBoost. Mach. Learn. 2001, 42, 287–320. [Google Scholar] [CrossRef]
  34. Breiman, L. Bagging Predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  35. Pietikäinen, M. Image Analysis with Local Binary Patterns. Lect. Notes Comput. Sci. 2005, 3540, 115–118. [Google Scholar] [CrossRef] [Green Version]
  36. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume I, pp. 886–893. [Google Scholar] [CrossRef] [Green Version]
  37. Fogel, I.; Sagi, D. Gabor Filters as Texture Discriminator. Biol. Cybern. 1989, 61, 103–113. [Google Scholar] [CrossRef]
  38. Haralick, R.M.; Dinstein, I.; Shanmugam, K. Textural Features for Image Classification. IEEE Trans. Syst. Man Cybern. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef] [Green Version]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar]
  40. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2010; pp. 248–255. [Google Scholar]
  41. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
Figure 1. An illustration of foot ulcers.
Figure 1. An illustration of foot ulcers.
Diagnostics 13 02637 g001
Figure 2. Sample images of healthy foot and DM foot.
Figure 2. Sample images of healthy foot and DM foot.
Diagnostics 13 02637 g002
Figure 3. A sample of thermograms from diabetic and control group and corresponding patch images (plantar angiosomes). MPA: medial plantar artery, LPA: lateral plantar artery, MCA: medial calcaneal artery, LCA: lateral calcaneal artery.
Figure 3. A sample of thermograms from diabetic and control group and corresponding patch images (plantar angiosomes). MPA: medial plantar artery, LPA: lateral plantar artery, MCA: medial calcaneal artery, LCA: lateral calcaneal artery.
Diagnostics 13 02637 g003
Figure 4. The architecture diagram of proposed model for DFU classification.
Figure 4. The architecture diagram of proposed model for DFU classification.
Diagnostics 13 02637 g004
Figure 5. The complete architecture working map of the proposed model for DFU classification.
Figure 5. The complete architecture working map of the proposed model for DFU classification.
Diagnostics 13 02637 g005
Figure 6. The ROC curves of classifiers using HOG feature (image + patch).
Figure 6. The ROC curves of classifiers using HOG feature (image + patch).
Diagnostics 13 02637 g006
Figure 7. The ROC curves of classifiers using Gabor feature (image + patch).
Figure 7. The ROC curves of classifiers using Gabor feature (image + patch).
Diagnostics 13 02637 g007
Figure 8. The ROC curves of classifiers using GLCM feature (image + patch).
Figure 8. The ROC curves of classifiers using GLCM feature (image + patch).
Diagnostics 13 02637 g008
Figure 9. The ROC curves of classifiers using LBP feature (image + patch).
Figure 9. The ROC curves of classifiers using LBP feature (image + patch).
Diagnostics 13 02637 g009
Figure 10. The ROC curves of classifiers using HOG + Gabor + GLCM + LBP feature (image + patch).
Figure 10. The ROC curves of classifiers using HOG + Gabor + GLCM + LBP feature (image + patch).
Diagnostics 13 02637 g010
Figure 11. The ROC curves of classifiers using HOG + Gabor + GLCM + LBP feature (full images only).
Figure 11. The ROC curves of classifiers using HOG + Gabor + GLCM + LBP feature (full images only).
Diagnostics 13 02637 g011
Figure 12. The ROC curves of classifiers using HOG + Gabor + GLCM + LBP feature (patches only).
Figure 12. The ROC curves of classifiers using HOG + Gabor + GLCM + LBP feature (patches only).
Diagnostics 13 02637 g012
Figure 13. DFU classification results using deep-learning approaches.
Figure 13. DFU classification results using deep-learning approaches.
Diagnostics 13 02637 g013
Figure 14. Full-image-level results of the proposed model.
Figure 14. Full-image-level results of the proposed model.
Diagnostics 13 02637 g014
Figure 15. Patch-level results of the proposed model.
Figure 15. Patch-level results of the proposed model.
Diagnostics 13 02637 g015
Figure 16. Image–patch level results of the proposed model.
Figure 16. Image–patch level results of the proposed model.
Diagnostics 13 02637 g016
Table 1. Thermogram image data detail: original and after augmentation.
Table 1. Thermogram image data detail: original and after augmentation.
CategoryDiabetic GroupControl GroupTotal
No. of cases12245167
Original full images24490334
Images after augmentation5005001000
Original patches9763601336
Patches after augmentation150015003000
Table 2. Machine-learning classifier results using different features for combined image+patch data.
Table 2. Machine-learning classifier results using different features for combined image+patch data.
Evaluation MetricFeaturesSVMRFXGBoostNaive BayesAdaBoostkNNBagging
SensitivityHOG0.710.710.710.640.570.670.71
Gabor0.710.140.160.710.280.140.71
GLCM0.260.280.360.140.140.360.4
LBP0.070.140.140.070.070.140.19
Combined0.530.510.510.460.640.570.55
SpecificityHoG0.850.90.90.850.870.70.87
Gabor0.90.770.870.470.80.90.87
GLCM0.950.750.750.850.90.750.77
LBP0.90.780.780.80.850.730.88
Combined0.890.830.820.940.790.810.83
AccuracyHoG0.710.650.690.660.590.70.73
Gabor0.680.610.650.540.670.70.73
GLCM0.70.630.650.670.60.650.67
LBP0.690.590.60.610.650.570.59
Combined0.780.730.720.790.750.730.74
F1-scoreHoG0.670.710.710.620.590.60.69
Gabor0.70.160.10.440.30.20.69
GLCM0.340.290.340.180.20.340.36
LBP0.10.160.190.080.090.150.17
Combined0.590.540.530.570.610.570.57
Table 3. Machine-learning classifier results using different features for full-image-level data.
Table 3. Machine-learning classifier results using different features for full-image-level data.
ModelsSensitivitySpecificityAccuracyF1-Score
SVM0.3450.9680.8110.479
RF0.5590.9480.8500.652
XGB0.6300.9320.8560.688
NB0.3920.90.7720.464
ADA0.6300.8960.8290.650
kNN0.6420.9160.8470.679
BAG0.6190.9080.8350.654
Table 4. Machine-learning classifier results using different features for patch-level data.
Table 4. Machine-learning classifier results using different features for patch-level data.
ModelsSensitivitySpecificityAccuracyF1-Score
SVM0.5300.8860.7760.594
RF0.5060.8270.7270.535
XGB0.5060.8210.7230.531
NB0.4570.9350.7870.571
ADA0.6380.7940.7460.609
kNN0.5660.8050.7310.566
BAG0.5540.8270.7420.571
Table 5. Deep-learning classifier results using data at different levels.
Table 5. Deep-learning classifier results using data at different levels.
ModelsSensitivitySpecificityAccuracyF1-ScoreAUC
Image
Level
ResNet500.7780.5710.6270.8750.623
DenseNet1210.50.5420.5210.5110.52
Custom Model0.970.9580.970.8910.976
Patch
Level
ResNet500.5650.9130.8320.8740.834
DenseNet1210.6610.8220.7910.80.788
Custom Model0.8390.8890.930.8910.932
Image
+ Patch
Level
ResNet500.6970.9020.8810.750.879
DenseNet1210.750.8750.8430.80.841
Custom Model0.8390.8890.9330.8910.938
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khosa, I.; Raza, A.; Anjum, M.; Ahmad, W.; Shahab, S. Automatic Diabetic Foot Ulcer Recognition Using Multi-Level Thermographic Image Data. Diagnostics 2023, 13, 2637. https://doi.org/10.3390/diagnostics13162637

AMA Style

Khosa I, Raza A, Anjum M, Ahmad W, Shahab S. Automatic Diabetic Foot Ulcer Recognition Using Multi-Level Thermographic Image Data. Diagnostics. 2023; 13(16):2637. https://doi.org/10.3390/diagnostics13162637

Chicago/Turabian Style

Khosa, Ikramullah, Awais Raza, Mohd Anjum, Waseem Ahmad, and Sana Shahab. 2023. "Automatic Diabetic Foot Ulcer Recognition Using Multi-Level Thermographic Image Data" Diagnostics 13, no. 16: 2637. https://doi.org/10.3390/diagnostics13162637

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop