Next Article in Journal
A3C-TL-GTO: Alzheimer Automatic Accurate Classification Using Transfer Learning and Artificial Gorilla Troops Optimizer
Previous Article in Journal
Monitoring the Early Strength Development of Cement Mortar with Piezoelectric Transducers Based on Eigenfrequency Analysis Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Machine Learning Approach for Severity Classification of Diabetic Foot Complications Using Thermogram Images

by
Amith Khandakar
1,2,
Muhammad E. H. Chowdhury
1,*,
Mamun Bin Ibne Reaz
2,*,
Sawal Hamid Md Ali
2,
Serkan Kiranyaz
1,
Tawsifur Rahman
1,
Moajjem Hossain Chowdhury
2,
Mohamed Arselene Ayari
3,4,
Rashad Alfkey
5,
Ahmad Ashrif A. Bakar
2,
Rayaz A. Malik
6 and
Anwarul Hasan
7
1
Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
2
Department of Electrical, Electronics and Systems Engineering, Universiti Kebangsaan Malaysia, Bangi 43600, Selangor, Malaysia
3
Department of Civil and Architectural Engineering, Qatar University, Doha 2713, Qatar
4
Technology Innovation and Engineering Education Unit, Qatar University, Doha 2713, Qatar
5
Acute Care Surgery and General Surgery, Hamad Medical Corporation, Doha 3050, Qatar
6
Weill Cornell Medicine-Qatar, Ar-Rayyan 24144, Qatar
7
Department of Industrial and Mechanical Engineering, Qatar University, Doha 2713, Qatar
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(11), 4249; https://doi.org/10.3390/s22114249
Submission received: 19 February 2022 / Revised: 27 April 2022 / Accepted: 9 May 2022 / Published: 2 June 2022
(This article belongs to the Special Issue Thermal & Infrared Imaging and Sensing in Biomedical Applications)

Abstract

:
Diabetes mellitus (DM) is one of the most prevalent diseases in the world, and is correlated to a high index of mortality. One of its major complications is diabetic foot, leading to plantar ulcers, amputation, and death. Several studies report that a thermogram helps to detect changes in the plantar temperature of the foot, which may lead to a higher risk of ulceration. However, in diabetic patients, the distribution of plantar temperature does not follow a standard pattern, thereby making it difficult to quantify the changes. The abnormal temperature distribution in infrared (IR) foot thermogram images can be used for the early detection of diabetic foot before ulceration to avoid complications. There is no machine learning-based technique reported in the literature to classify these thermograms based on the severity of diabetic foot complications. This paper uses an available labeled diabetic thermogram dataset and uses the k-mean clustering technique to cluster the severity risk of diabetic foot ulcers using an unsupervised approach. Using the plantar foot temperature, the new clustered dataset is verified by expert medical doctors in terms of risk for the development of foot ulcers. The newly labeled dataset is then investigated in terms of robustness to be classified by any machine learning network. Classical machine learning algorithms with feature engineering and a convolutional neural network (CNN) with image-enhancement techniques are investigated to provide the best-performing network in classifying thermograms based on severity. It is found that the popular VGG 19 CNN model shows an accuracy, precision, sensitivity, F1-score, and specificity of 95.08%, 95.08%, 95.09%, 95.08%, and 97.2%, respectively, in the stratification of severity. A stacking classifier is proposed using extracted features of the thermogram, which is created using the trained gradient boost classifier, XGBoost classifier, and random forest classifier. This provides a comparable performance of 94.47%, 94.45%, 94.47%, 94.43%, and 93.25% for accuracy, precision, sensitivity, F1-score, and specificity, respectively.

1. Introduction

Diabetes mellitus (DM) is a chronic medical condition resulting from high amounts of sugar in the blood, which often leads to severe health complications such as heart-related diseases, kidney failure, blindness, and lower limb amputation [1]. Diabetes mellitus leads to foot ulcers, which might not heal properly due to poor blood circulation [2]. This might result in the spreading of infection and, eventually, can lead to amputation [3]. Amputations of lower limbs are more common in diabetic individuals, and according to statistics, diabetic foot lesions account for 25% of hospital admissions, and 40% of individuals presenting with diabetic foot necessitate amputation [4]. The reoccurrence rate of diabetic foot ulcers (DFUs) is also significantly high, with approximately 40% after the first year and 60% within three years of occurrence [5]. In an advanced and developed country such as Qatar, with more than 20% of the population being diabetic, it is reported that one of the most common reasons for people being admitted to hospitals is diabetic foot complications [6]. Every year in the USA alone, over one million diabetic patients suffer an amputation due to failure of recognizing and treating diabetic foot ulcers properly [7]. In Europe, 250,000 diabetic patients have their legs amputated with a death rate of 30% in one month and 50% in a year [8]. A diabetic patient under a ‘high risk’ category needs regular check-ups, hygienic personal care, and continuous expensive medication to avoid unwanted consequences. Diabetic foot ulcers lead to increased healthcare costs, decreased quality of life, infections, amputations, and death. Early detection and better DFU classification tools would enable a correct diagnosis, effective treatment, and timely intervention to prevent further consequences.
Self-diagnosis at home, i.e., self-care such as monitoring without medical assistance, for early signs of ulcers could be useful in preventing severe after-effects. However, the easiest monitoring technique, visual inspection, has its limitations, such as people with obesity or visual impairment not being able to see their sites of ulcers easily. Some systems that allow for easy monitoring are TeleDiaFoS [9] and Bludrop [10]. TeleDiaFoS is a telemedicine system for the home telemonitoring and telecare of diabetic patients. In this system, the affected foot is scanned using a specialized patient module, which is operated with a simple two-button remote controller. While this method has shown great results that are comparable to Visitrak [11] and Silhouette [12], it requires a specialized module specifically for it. This might increase the cost. Bludrop has also introduced a patented system that promises high accuracy, but their system also requires specialized hardware. According to recent studies, a temperature monitoring system at home has been able to spot 97% of diabetic foot ulcers well in advance [13,14,15,16]. It has also been confirmed that patients going through the continuous monitoring of foot temperature had a low risk of foot complications [17]. There have been studies that have tried to find the relation between temperature and diabetes mortality [18,19], where they stressed that exposure to high and low temperatures increases the chances of diabetic mortality, especially in the aged population. Foltynski et al. in [20] concluded that ambient temperature does influence foot temperature even during foot ulceration and, thus, it should be taken into consideration, especially during the assessment for diabetic foot ulceration. As early as the 1970s, skin-temperature monitoring emerged as a useful tool for identifying patients at risk for ulceration. In the literature, the temperature monitoring approach uses the plantar foot temperature asymmetry between both feet, which is referred to as an “asymmetry analysis”, to find ulcers at an early stage [21]. A temperature difference of 2.22 °C (4 °F) over at least two consecutive days could be used as a threshold to indicate the starting time of preventive therapy for foot ulcers [14]. It has also been found that the system correctly identified 97% of observed DFUs, with an average lead time of 37 days [22]. Therefore, a foot with an elevated temperature (>2.2 °C compared to the opposite foot) can be considered to be “at-risk” of ulceration due to inflammation at the site of measurement.
Thermography is a popular technique used to examine thermal changes in a diabetic foot [22]. This is a pain-free, no-touch, and non-invasive technique. Several studies have proposed thermogram-based techniques for the study of diabetic feet [2,3,23], which include the identification of characteristic patterns in infrared images and the measurement of changes in thermal distribution. It was reported that the control group showed a specific butterfly pattern [24], while the DM group showed a large variety of spatial patterns [25,26]. It is possible to calculate and determine an estimation of thermal changes with respect to one foot as a reference [27,28,29,30] using the contralateral comparison of temperatures. However, if both feet have temperature changes, but neither of them has the butterfly pattern, one of the feet cannot act as a reference. Asymmetry cannot be measured even if there is a large temperature difference and identical spatial distributions present in both feet. An alternative approach is to calculate the temperature change using a control group butterfly pattern [31,32,33].
Machine learning (ML) techniques are gaining popularity in biomedical applications in assisting medical experts in early diagnosis [34,35,36]. The authors conducted an extensive investigation and developed a trained AdaBoost classifier, which achieved an F1-score of 97% in classifying diabetic and healthy patients using thermogram images [37]. However, the temperature distribution of a diabetic foot does not have a specific spatial pattern and it is, therefore, important to devise a method to distinguish the diabetic feet with different temperature distributions that do not depend on a spatial pattern. The spatial distribution may change after a while and the temperature changes are not significant in some cases. Moreover, the detection of irregular temperature rises in the plantar region is important for diabetic patients. The deep learning technique using thermogram images to classify the control and diabetic patients is not a well-studied domain. Moreover, the severity grading of diabetic foot is also lacking in the literature, which might help to establish an early warning tool for diabetic foot ulcer detection. A simple, effective, and accurate machine learning technique for diabetic foot severity stratification before foot ulcer development using thermogram images would be very useful.
Several studies [31,32,38,39,40,41,42,43,44,45] have attempted to extract features that can be used to identify the hot region in the plantar thermogram, which could be a sign of tissue damage or inflammation. In all of the works, the plantar region was split into six areas and different statistical features were extracted. In order to obtain various coefficients from texture and entropy characteristics, Adam et al. in [43] employed a discrete wavelet transformation (DWT) and higher-order spectra (HOS). In another work, the foot image was decomposed using a double-density dual-tree complex wavelet transform (DD-DT-CWT), and many features were retrieved from the decomposed images [42]. To categorize patients as normal or in the ulcer group, Saminathan et al. in [41] segmented the plantar area into 11 regions using region raising and retrieved texture characteristics. Maldonado et al. in [40] employed the DL approach to segment a visible foot picture, which was then used to segment the plantar area of the same patient’s thermogram image to identify ulceration or necrosis based on temperature differences. Hernandez et al. in [32] presented the thermal change index (TCI) as a quantitative indicator for detecting thermal changes in the plantar region of diabetic patients in comparison to a reference control group. They published a public database of diabetic foot thermogram pictures and named it the “Plantar Thermogram Database”, and used TCI to classify the participants into classes one to five based on the spatial temperature distribution and temperature range. The authors conducted considerable research and produced a trained AdaBoost classifier that classified diabetic and healthy patients using thermogram images with an F1-score of 97 percent [37]. Cruz-Vega et al. in [38] proposed a deep learning technique to classify the plantar thermogram database images in a non-convenient classification scheme, where the results were shown by taking two classes at a time and then averaging the results after ten folds of a different combination of two set classes. For the classification of class three and class four, a new diabetic foot thermogram network (DFTNet) was proposed, with a sensitivity and accuracy of 0.9167 and 0.853, respectively. However, the ground-truth classification of the patients in this database was performed using the TCI score, which takes the butterfly pattern of the control group which is individually used to compare each affected foot. Therefore, the classification entirely relies on the reliability of the TCI scoring technique, which was found questionable in two different aspects. Firstly, it was observed that thermogram images with a butterfly pattern even present in higher classes and the temperature distribution of severe diabetic patients as reported in other articles were incorrectly classified to lower classes based on the TCI score. Secondly, the state-of-the-art deep learning techniques failed to reliably classify the thermogram images into different classes, which were graded based on the TCI score. If a database is publicly available, it is easy to re-evaluate the labels in datasets if it is found that the labels are questionable [46]. Aradillas et al. [47] mentioned scenarios where they found errors in the labeling of the training samples in databases and proposed cross-validation techniques to remove them. Hu et al. [48] mentioned wrong labels in face-recognition databases and proposed a training paradigm that would take the dataset size into account and train itself by probabilistically removing some datasets that could have been incorrectly labeled. Ding et al. [49] proposed a semi-supervised two-stage approach in learning from noisy labels.
As per the above discussion, it can be seen that there is a lack of severity grading of the risk of ulcer development. The authors, including a set of expert medical doctors working in a diabetic foot clinic, stated that a foot plantar temperature distribution can help in the early detection of a foot ulcer development [50,51,52]. The scientific hypothesis of this work is whether machine learning approaches can be used to categorize thermograms into severity classes of patients with their risk of ulcer development. This severity classification can help in early treatment that could prevent diabetic foot ulcers. The improved performance of a machine learning model could help in early detection from the convenience of the patient’s home and reduce the burden on the healthcare systems, considering diabetic foot complications are very common and lead to expensive follow-up and treatment procedures throughout the world. All the above studies motivated the authors to carry out an extensive investigation on dataset re-labeling, using an unsupervised machine learning and pre-trained convolutional neural network (CNN) to extract features automatically from thermogram images, reduce feature dimensionality using the principal component analysis (PCA), and, finally, to classify thermograms using k-mean clustering [53] to revise the labels. The new labels identified by the unsupervised clusters were fine-tuned by a pool of medical doctors (MDs). Then, the authors investigated classical machine learning techniques using feature engineering and 2D CNNs with image enhancement techniques on the images to develop the best-performing classification network. The major contributions of this paper can be stated as follows:
The revision of labels of the thermogram dataset for the diabetic severity grading of a publicly available thermogram dataset using unsupervised machine learning (K-mean clustering), which is fine-tuned by medical doctors (MDs).
The extraction and ranking of relevant features from temperature pixels for classifying thermograms into diabetic severity groups.
The exploration of the effect of various image enhancement techniques on thermogram images in improving the performance of 2D CNN models in diabetic severity classification.
This work is the first of its kind to propose a machine learning technique to classify diabetic thermograms into different severity classes: mild, moderate, and severe.
The manuscript is organized into five sections, where Section 1 discusses the introduction and the related works and key contributions, Section 2 discusses the detailed methodology used in this work, and Section 3 and Section 4 presents the results and discussion. Finally, Section 5 presents the conclusions.

2. Methodology

The methodology adopted in the work is presented in Figure 1. In the study, the thermogram was applied to a pre-trained CNN model to extract useful features, and then the feature dimensionality was reduced using principal component analysis (PCA), and the sparse feature space was then applied as an input to the k-mean clustering algorithm to provide unsupervised revised classes, which was then verified by medical doctors. The newly revised classes—mild, moderate, and severe—were then tested in terms of classification performance by 2D CNN approach using different image enhancement techniques and classical machine learning algorithms on the extracted features from the thermograms. The following sections went over the dataset used in the study, the K-mean clustering method used for unsupervised clustering, and the investigation performed using (i) thermogram images and 2D CNN networks and various image enhancement techniques, and (ii) classical machine learning algorithms with feature engineering (feature extraction, and feature reduction) on the thermogram images. This section also went through the performance indicators that were utilized to choose the best machine learning algorithm.

2.1. Dataset

In this study, 167 foot-pair thermograms of DM (122) and control (45) subjects were obtained from a public database from the General Hospital of the North, the General Hospital of the South, the BIOCARE clinic, and the National Institute of Astrophysics, Optics and Electronics (INAOE) for 3 years (from 2012 to 2014). All of these places were located in the city of Puebla, Mexico. The dataset had demographic information such as age (control group—27.76 ± 8.09; diabetic group—55.98 ± 10.57 years), gender, height, and weight. The participants who participated in the study were asked to lay in the supine position for 15 min to reach thermodynamic equilibrium to improve the accuracy of temperature variation due to diabetic complications and had avoided prolonged sun exposure, intense physical activity, and any effort that could affect blood pressure [32,39]. The participants were asked to remove their shoes and socks and clean their feet with a damp towel.
The data collection performed during resting time was to avoid exerting any effort that would impact the subjects’ blood pressure and, as a result, their plantar temperature. The dataset included segmented thermograms of the medial plantar artery (MPA), lateral plantar artery (LPA), medial calcaneal artery (MCA), and lateral calcaneal artery (LCA), a concept suggested by Taylor and Palmer [54], in addition to the segmented foot thermograms (Figure 2). Because it was utilized to compute local temperature distribution, the information gained utilizing angiosomes was related not only to the damage caused by DM in arteries, but also to the accompanying ulceration risk. The dataset, which is the largest diabetic thermogram public dataset, also included pixelated temperature measurements for complete feet and four angiosomes for both feet.

2.2. K-Mean Clustering Unsupervised Classification

In this part of the investigation, the underlying features of the images extracted using a pre-trained CNN model were used to cluster them in different groups using k-mean clustering. K-mean is the most popular method for clustering data in an unsupervised manner, and was proposed in 1967 [53]. K-mean is an unsupervised, non-deterministic, numerical, iterative method of clustering. In k-mean, each cluster is represented by the mean value of objects in the cluster. Here, we partitioned a set of n objects into k clusters, so that intercluster similarity was low and intracluster similarity was high. The similarity was measured in terms of the mean value of objects in a cluster. A similar concept was used in our investigation and could be divided into 4 steps:
  • Pre-processing: Preparing the image so that it could be fed properly to the CNN model.
  • Feature extraction: Using a pre-trained CNN model to extract the underlying features from a specific layer.
  • Dimensionality reduction: Using principal component analysis (PCA) [55] to reduce the noise in the feature space and reduce the dimensionality
  • Clustering: Using K-mean to cluster the images based on similar features.
In the pre-processing step, the image was first resized to 224 × 224 size. This was because the CNN model available in Keras [56], pretrained on ImageNet Dataset [57], required the input image to be of that size. Interpolation was applied to images before resizing to allow for rescaling by a non-integer scaling factor. This step did not change the properties of the image, and instead ensured that the image was formatted correctly.
For feature extraction in our study, the popular CNN network—VGG19 was used [58]. The VGG19 network developed by the Oxford Visual Geometry Group is a popular CNN for computer vision tasks because of its high performance and relative simplicity. The detailed framework, along with network architecture, can be seen in Figure 3. The pre-trained VGG19 model was used to extract 4096 features from the ‘FC1’ layer from each thermogram image. The selection of the network layer for feature extraction was a hyperparameter in this experiment, and it was tuned to obtain the parameter that provided the best result, which is a common exercise. We empirically found out that the FC1 layer gave us the best result similar to other studies that investigated k-means clustering [59,60]. The number of features needed to be reduced so that the k-mean algorithm did not overfit and it remained robust to noise. PCA was used to transform and reduce the features such that only the most important features were retained [51]. Figure 4 shows the cumulative variance versus the number of components after PCA. It can be seen that the first 10 components provided a cumulative variance of 0.5, and after adding 90 more components, the variance rose to 0.9. Thus, significant information was retained in the initial components. In our study, we used 45 components, which helped to achieve 75% cumulative variance, which is usually an acceptable threshold in various k-mean clustering studies [61,62,63,64].
The reduced and transformed features were separated into different groups using k-mean clustering [65]. This is an unsupervised machine learning algorithm, which popularized dealing with non-labeled data [66,67]. K-mean’s aim was to group data that were similar in feature space. The algorithm flow can be seen in Algorithm 1 below. In our study, max_iter was set to 500 and the break condition that happened only after line 4 was fulfilled for all clusters.
Algorithm 1: K-mean clustering
Input:Feature matrix, number of centroids (k)
Output:Trained model
1:for i 1 to m a x _ i t e r do
2:Assign each point with the centroid that it is closest to in latent space;
3:Recalculate the position of the clusters ( c e n t r o i d _ p o s i t i o n ) to be equal to the mean position of all of its associated points;
4:if c e n t r o i d _ p o s i t i o n i = = c e n t r o i d _ p o s i t i o n i 1 then
5:break;
6: i ++;
7:end for
The algorithm aimed to reduce inertia. Inertia is the sum of squared Euclidean distances from a centroid to its associated data points and can be seen mathematically in Equation (1).
I n e r t i a = j = 1 k i = 1 n x i ( j ) c j 2
where k is the total number of clusters, n is the number of samples associated with a cluster, x is the position of the sample in the feature space, and c is the position of the centroid in the feature space.
K-means++ [68] was used to initialize the centroid position. This method helped to achieve good clustering performance and reduced computational complexity. In this method, the first centroid was selected from the data with uniform probability. The other centroids were selected from the data with a probability proportionate to their distance from the nearest centroid. Hence, the initial centroids were close to the data points but were far apart from each other.
As stated earlier, the labeled dataset provided by k-mean clustering was used for the investigation using (i) transfer learning, image enhancement, and transfer learning 2D CNN and (ii) classical machine learning techniques and feature engineering on the extracted features from the thermograms.

2.3. Two-Dimensional CNN-Based Classification

Two-dimensional CNN is widely used in biomedical applications for automatic and early diagnosis of anomalies such as COVID-19 pneumonia, tuberculosis, and other diseases [69,70]. A labeled dataset can be divided into training and testing datasets, with the training dataset being used to train the network and the unseen testing dataset being used to verify its performance. During the training process, a portion of the training dataset is used for validation to avoid overfitting. Five-fold cross-validation was employed in this study, which divided the dataset five-fold and found the performance metric for the testing dataset five times. Each time, one of the folds was used as testing dataset, and the remaining folds were used for training and validation. This approach helped in stating results considering the complete dataset and making sure the test data were always unseen. The final results were the overall and weighted result of the five folds; detailed performance metric was shown later. As larger amount of data used for training always helps in obtaining better trained model, the authors used popular augmentation techniques (rotation and translation) to increase the training data size. The rotation operation used for image augmentation was performed by rotating the images in clockwise and counter-clockwise directions with angles from 5 to 30 in increments of 2°. Image translation was conducted by translating images horizontally and vertically from −15 to 15%. The details of the training, validation and testing dataset for 2D binary and severity classifier are shown in Table 1.
Transfer learning: As we had a limited dataset, which can be seen in Table 1, we could make use of pre-trained models which were trained on a large ImageNet database [71]. These pre-trained networks were trained on very large ImageNet database [71], and had good classification performance. These networks could be further trained on any other classification problem, and this is known as transfer learning. Based on extensive literature review and previous performances [37], in this study, seven well-known pre-trained deep CNN models were used for thermograms’ classification: ResNet18, ResNet50, ResNet100 [72], DenseNet201 [72], InceptionV3 [73], VGG19 [58], and MobileNetV2 [74].
Image enhancement: It was found that image enhancement techniques such as adaptive histogram equalization (AHE) [75] and gamma correction [70,76] can help the 2D CNN in improving its classification performance for thermograms [37]. Some samples of the image enhancement on DM and CG can be seen in Figure 5. The authors investigated the improvement of performance made possible by the different image enhancement techniques, and they were reported in the Experimental Results section.

2.4. Classical Machine Learning Approach

This section discusses the features extracted for classical ML techniques, feature reduction techniques, feature ranking techniques, machine learning classifiers, and the extensive investigations performed using two approaches.

2.4.1. Extracted Features and Feature Reduction

The authors looked through the literature carefully to summarize the features that are used in clinical practice and machine learning approaches to analyze the foot thermograms for diabetic foot diagnosis. The details of the final list of features identified by the authors were mentioned in their previous work in [37] and also mentioned below:
T h e r m a l   C h a n g e   I n d e x   ( T C I ) = C G a n g D M a n g 4
where CGang and DMang are the temperature values of the angiosome for the control group and a DM subject, respectively.
E s t i m a t e d   T e m p e r a t u r e   ( E T ) = a j 1 C j 1   + a j C j + a j + 1 C j + 1 a j 1 + a j + a j + 1
E s t i m a t e d   T e m p e r a t u r e   D i f f e r e n c e   ( E T D ) = E T l e f t   A n g i o s o m e E T r i g h t   a n g i o s o m e
H o t   S p o t   E s t i m a t o r   ( H S E ) = c l E T
A histogram for the percentage of pixels in the thermogram (either complete foot or angiosomes) in the different classmark temperatures (C0 = 26.5 °C, C1 = 28.5 °C, C2 = 29.5 °C, C3 = 30.5 °C, C4 = 31 °C, C5 = 32.5 °C, C6 = 33.5 °C, and C7 = 34.5 °C) was generated to equate the parameters in Equations (3)–(5). The classmark temperature and the associated percentage of pixels in that region were denoted by the terms Cj and aj, respectively. The percentage of pixels in the surrounding classmark temperatures Cj−1 and Cj+1 were represented by the values aj−1 and aj+1, respectively.
In addition to these parameters, the authors formulated NRTclass j, which was the number of pixels in the class j temperature range over the total number of non-zero pixels, where class j could be class 1 to 5. This parameter is visually very important for distinguishing the variation in the plantar temperature distribution, and was also reported in the authors’ previous work [37].
Thus, TCI, highest temperature value, NTR (class 1–5), HSE, ET, ETD, mean, median, SD of temperature for the distinct angiosomes, LPA, LCA, MPA, MCA, and full foot were among the 37 features that could be employed for early diabetic foot identification. In their prior study [37], the authors published the statistics of the data provided by the source [39].
By determining the association between the various features, the final list of features was streamlined to eliminate redundant features. Features with a correlation of greater than 95% were deleted, improving overall performance by lowering the number of redundant features and preventing overfitting [77,78].

2.4.2. Machine Learning Classifiers

The authors also explored different machine learning (ML) classifiers in this study to compare the performances. The popular ML classifiers used in the study were multilayer perceptron (MLP) [79], support vector machine (SVM) [80], random forest [81], extra tree [82], GradientBoost [83], logistic regression [84], K-nearest neighbor (KNN) [85], XGBoost [86], AdaBoost [87], and linear discriminant analysis (LDA) [88].

2.4.3. Classical Machine Learning Approach 1: Optimal Combination of Feature Ranking, Number of Features

After optimization, the dataset’s shortlisted parameters were evaluated to determine decisions and the best features for severity classification. The multi-tree extreme gradient boost (XGBoost) [89], random forest [90], extra tree [91], chi-squared [92], pearson correlation coefficient [93], recursive feature elimination (RFE) [94], logistic regression [95], and LightGBM [96] were used to identify three different sets of feature ranking. Rigorous research determined the greatest combination of features that offered the best performance, and the best-performing top-ranked features from the different feature-ranking methodologies were used to identify the best combination of features.

2.4.4. Classical Machine Learning Approach 2: Stacking-Based Classification

We presented a stacking-based classifier that worked by merging numerous best-performing classifiers created by different learning algorithms L 1 , , L N on a single dataset S, which consisted of examples s i = ( x i , y i ) , i.e., pairs of feature vectors ( x i ) and their classifications ( y i ). A collection of base-level classifiers M 1 , , M N was trained in the first phase, where M i = L i ( S ) . A meta-level classifier was trained in the second step, which integrated the outputs of the base-level classifiers (Figure 6).
A cross-validation approach was used to build a training set for learning the meta-level classifier. We applied each of the base-level learning algorithms to four folds of dataset for five-fold cross-validation, leaving one fold for testing. A probability distribution over the possible class values was predicted by each base-level classifier. When applied to example x, the prediction of the base-level classifier M was a probability distribution:
P M ( x ) = ( P M ( c 1 | x ) , P M ( c 2 | x ) , . , P M ( c m | x ) ) ,
P M ( c i | x ) signifies the probability that example x belonged to class c i as estimated (and predicted) by classifier M, where ( c 1 , c 2 , , c m ) is the set of possible class values. Classifier M predicted the class c j with the highest class probability P M ( c j | x ) . The meta-level characteristics are the probabilities predicted by each of the base-level classifiers for each conceivable class, i.e., P M j ( c i | x ) for i = 1……m and j = 1……N, where ( c 1 , c 2 , , c m ) is the set of possible class values and P M ( c i | x ) denotes:

2.5. Performance Evaluation and Classification Scheme

In all of our experiments, we reported sensitivity, specificity, precision, accuracy, F1-score, and area under the curve (AUC) for five folds, as our evaluation metrics.
S e n s i t i v i t y = ( T P ) ( T P + F N )
S p e c i f i c i t y = ( T N ) ( T N + F P )
P r e c i s i o n = ( T P ) ( T P + F P )
A c c u r a c y = T P + T N ( T P + F N ) + ( F P + T N )
F 1   S c o r e = ( 2 P r e c i s i o n S e n s i t i v i t y ) ( P r e c i s i o n + S e n s i t i v i t y )
Here, TP, FP, TN, and FN are true positive, false positive, true negative, and False negative. In any classification, TP is the number of correctly identified thermograms, TN is the number of correctly identified thermograms of the other class, FP is the number of thermograms misclassified and FN is the number of thermograms of the other class misclassified. In this paper, weighted performance metric with 95% confidence interval was reported for sensitivity, specificity, precision, and F1-score, and for the accuracy, the overall accuracy was reported.
All the experiments were done on google colab platform and used GPU GeForce RTX2070 Super which is manufactured by NVIDIA, it is located in Santa Clara, CA, USA. The other software was Matlab from Mathworks, Natick, MA, USA.

3. Experimental Results

This section of the paper provides the results of the various important experiments of the paper. The details are below:

3.1. K-Mean Clustering Unsupervised Classification

As stated in the previous section, k-mean clustering was used to classify the thermogram images in an unsupervised fashion. This was conducted by using the features extracted from the VGG-19 network, which were later reduced and transformed using PCA. K-mean clustering was applied for various cluster numbers, i.e., k in Algorithm 1. The optimal number of clusters would be equivalent to the correct severity class of the thermograms, later verified by medical experts. The T-distributed stochastic neighbor embedding (t-SNE) plots, which are very useful in visualizing the clustering of data in two dimensions [97,98], for various clusters can be seen in Figure 7. The t-SNE is a dimensionality reduction method that is used to visualize high-dimensional data. In this work, the high-dimensional feature space was visualized by reducing it to two-dimensions using t-SNE, and then plotting it in a scatterplot. It was evident that there was a clear distinction in clustering for cluster sizes two and three compared to a cluster size of five (original classification performed using TCI scores, i.e., class one–class five). As stated in Figure 4, 45 components were used in the k-mean clustering.
The clusters found using the k-mean clustering approach were confirmed by experts in the diabetic foot complication domain as thermograms of patients with mild, moderate, and severe severity, and can be seen in Figure 8. The experts used their experience with plantar foot temperature distribution amongst diabetic patients as reference for confirming the severity classification. The clusters found by k-mean clustering provided three differentiable sets of thermograms. Plantar foot thermograms with slightly a deviated butterfly pattern were labelled as mild by medical experts. The second set of plantar foot thermograms where there was an abnormal and high-temperature distribution was labelled as moderate by medical experts, as the temperature distributions were still not very high (which is usually indicated by the red color distribution in thermograms) and could indicate the possibility of ulcer development in the near future. The third set of plantar foot thermograms had extreme high temperature (which is usually indicated by the red color distribution in thermograms) throughout the foot, indicating the possibility of ulcer development in the immediate future, and were labelled as severe. Such clustering was not found in the original classification using the TCI-based method (i.e., class one–class five). The clustering algorithm separated the images into three clusters. The mild class had 82 images, the moderate class had 84 images, and the severe class had 78 images.
The output of the k-mean clustering could be further understood with the help of the pie charts in Figure 9, which showcase the distribution of the classes in the different clusters found by the k-mean clustering approach. It can be clearly seen that many class one thermograms were very similar to class five thermogram (refer to the severe class in the pie chart) and many class three, class four, and class five were similar to class one and class two (refer to the mild class in the pie chart).

3.2. Classical Machine Learning-Based Classification

Once the new labeled datasets were confirmed using the k-mean clustering approach and verified by the medical experts, the highlighted correlated features from the 37 extracted were reduced to 27 features—NRT (class one), NRT (class two), NRT (class three), NRT (class four), NRT (class five), highest temperature, TCI, HSE, ETD, mean, and STD of MPA angiosome; HSE, ET, ETD, mean, and STD of LPA angiosome; HSE, ET, ETD, mean, and STD of LCA angiosome; HSE, ETD, and STD of MCA angiosome; HSE, ETD, and STD of full feet. The heatmaps of the correlation matrix with all features and after removing the highly correlated features are shown in Supplementary Figure S1.

3.2.1. Classical Machine Learning Approach 1: Optimal Combination of Feature Ranking, Number of Features

In this experiment, 3 feature selection techniques with 10 machine learning models were investigated with 27 optimized features to identify the best-combined results in 810 investigations. The best-performing combination is presented in Table 2. It can be seen that the XGboost classifier with the random forest feature selection technique and the top 25 features showed the best performance of 92.63% weighted F1-score in the diabetic severity classification.

3.2.2. Classical Machine Learning Approach 2: Stacking-Based Classification

As stated earlier, we shortlisted the major contributing features from the 29 finalized features after a feature reduction. The shortlisting was performed based on finding the features which were categorized as highly correlated with the output based on all the techniques—Pearson, Chi-square, RFE, logistic, random forest, and LightGBM. Table 3 showcases the features which had a total of more than four, and it was found that only eight features (TCI, NRT (class four), NRT (class three), the mean of MPA, mean of LPA, ET of LPA, mean of LCA, and highest temperature) were top ranked features by all the techniques (Pearson, Chi-Square, RFE, logistics, random forest, and LightGBM). These eight features were used to check the performance using different machine learning classifiers and stacking performed on the top three classifiers, providing the best result, as can be seen in Table 4. The stacking classifier created using the trained gradient boost classifier, XGBoost classifier, and random forest classifier was the best-performing classifier, with 94.47%, 94.45%, 94.47%, 94.43%, and 93.25% for accuracy, precision, sensitivity, F1-score, and specificity, respectively. The performance of this approach was better than Approach 1, as shown in Table 2, with the only difference in the inference time, which was found more in the stacking classifier, as expected.

3.3. Two-Dimensional CNN-Based Classification

As discussed earlier, the authors investigated state-of-the-art transfer learning networks—ResNet18, ResNet50, VGG19, DenseNet201, InceptionV3, and MobileNetv2, along with popular image enhancement techniques. The best-performing network and enhancement types (e.g., AHE) were used for severity classification. Gamma correction did not improve the performance, as it did not help in sharpening the distinguishing features. Further discussion on the image enhancement techniques was provided in the Discussion section. Independent foot images were used to check if the different pre-trained networks could classify them into different severity levels. Table 5 reports the best-performing AHE, with VGG19 showcasing the best performance. It is to be noted that the improved performance came with a tradeoff of a slightly higher inference time. The increase in inference time to classify the thermogram image was caused due to the addition of the image enhancement technique as a pre-processing step. However, the increase in inference time due to this additional pre-processing step was only 1 ms, which would not affect even real-time classification applications. It is more important to have improved performance than to have a faster response. In our study, the image enhancement helped in the performance with a 1 ms excess delay. The ROC curves for the original and AHE thermograms are also shown in Figure 10.

4. Discussion

As mentioned earlier, the authors conducted an extensive investigation and developed a trained AdaBoost classifier to achieve an F1-score of 97% in classifying diabetic and healthy patients using thermogram images [37]. The diabetic foot thermogram network (DFTNet) proposed a deep learning technique to classify the images of the plantar thermogram database in a non-convenient classification scheme, where the results were shown by taking two classes at a time and then averaging the results after ten folds of a different combination of two set classes [38]. As discussed earlier, the ground-truth classification of the patients in this database was performed using the TCI score and entirely relied on the reliability of the TCI scoring technique. Previous works and the lack of severity grading of diabetic foot complications motivated this work.
Again, to the best of the author’s knowledge, no previous studies have investigated image enhancement techniques for diabetic foot severity classification using thermograms for 2D CNNs. The authors investigated different pre-trained networks and also found that the image enhancement techniques helped in the classification performance. Figure 11 confirms the improved performance of the image enhancement technique using the F1-score performance metric.
Even the classical machine learning approach provided better classification performance using TCI, NRT (Class 4), NRT (Class 3), the mean of MPA, mean of LPA, ET of LPA, mean of LCA, and highest temperature extracted features from the thermogram. The stacked classifiers using popular gradient boost, XGBoost, and random forest classifiers provided a comparable classification performance with a much lower inference time (Table 4 and Table 5). The following interesting results could be summarized:
AHE, due to its special equalization, helped with the severity classification using the pre-trained VGG 19 network.
A deeper layer network such as DensetNet201 did not improve the performance for image enhancement, which could be attributed to the simplistic nature of the thermogram images, which did not require a very deep network to extract meaningful features.
The proposed features [37] and new classification technique helped in classifying the diabetic thermograms into different severity groups.

5. Conclusions

Diabetic foot problems are an issue that has great consequences not only in terms of mortality, but also in terms of the expense needed for monitoring and controlling the disease. Thus, early detection and severity classification can help in preventing such complications. The deployment of machine learning in biomedical applications could help in preparing easy-to-use solutions for early detection, not only for medical experts to save their time, but also since such solutions can be useful for patients in a home setting. They can use it in their homes, especially during pandemic times, where visits to healthcare services are preferred to be limited, avoiding stress on the healthcare system. The authors in the paper proposed a novel framework to cluster diabetic thermograms based on severity, which was not present in the literature, and the ones that were available needed verification. The trained 2D CNN and classical machine learning models could help in severity stratification using foot thermograms, which can be captured using infrared (IR) cameras. To the best of the author’s knowledge, this was the first study analyzing such a diabetic foot condition severity classifier for the early and reliable stratification of diabetic foot. The machine learning classifier performance was comparable to the 2D CNN performance using image enhancement. In conclusion, such a system could be easily deployed as a web application, and patients could benefit from remote health care using just an infrared camera and a mobile application, which is a future direction of our research.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/s22114249/s1: Figure S1: heatmap of the correlation matrix with all the features (A) and after removing the highly correlated features (B).

Author Contributions

Formal analysis, A.K., M.E.H.C., M.B.I.R., S.H.M.A., M.A.A. and A.A.A.B.; Investigation, A.K., M.E.H.C., S.K., A.A.A.B. and A.H.; Methodology, A.K., M.E.H.C., S.K., A.A.A.B. and A.H.; Resources, S.K., T.R. and M.H.C.; Software, T.R., M.H.C. and M.A.A.; Supervision, M.B.I.R., S.H.M.A. and A.A.A.B.; Validation, R.A. and R.A.M.; Visualization, R.A.M.; Writing—original draft, A.K.; Writing—review & editing, M.B.I.R., S.H.M.A., M.A.A., R.A., R.A.M. and A.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was made possible by the Qatar National Research Fund (QNRF) NPRP12S-0227-190164 and the International Research Collaboration Co-Fund (IRCC) grant IRCC-2021-001 and Universiti Kebangsaan Malaysia under grant DPK-2021-001. The statements made herein are solely the responsibility of the authors.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cho, N.; Kirigia, J.; Mbanya, J.; Ogurstova, K.; Guariguata, L.; Rathmann, W. IDF Diabetes Atlas, 8th ed.; International Diabetes Federation: Brussels, Belgium, 2015; p. 160. [Google Scholar]
  2. Sims, D.S., Jr.; Cavanagh, P.R.; Ulbrecht, J.S. Risk factors in the diabetic foot: Recognition and management. Phys. Ther. 1988, 68, 1887–1902. [Google Scholar] [CrossRef] [PubMed]
  3. Iversen, M.M.; Tell, G.S.; Riise, T.; Hanestad, B.R.; Østbye, T.; Graue, M.; Midthjell, K. History of foot ulcer increases mortality among individuals with diabetes: Ten-year follow-up of the Nord-Trøndelag Health Study, Norway. Diabetes Care 2009, 32, 2193–2199. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Singh, G.; Chawla, S. Amputation in diabetic patients. Med. J. Armed Forces India 2006, 62, 36–39. [Google Scholar] [CrossRef] [Green Version]
  5. Armstrong, D.G.; Boulton, A.J.; Bus, S.A. Diabetic foot ulcers and their recurrence. N. Engl. J. Med. 2017, 376, 2367–2375. [Google Scholar] [CrossRef]
  6. Ponirakis, G.; Elhadd, T.; Chinnaiyan, S.; Dabbous, Z.; Siddiqui, M.; Al-muhannadi, H.; Petropoulos, I.N.; Khan, A.; Ashawesh, K.A.; Dukhan, K.M.O. Prevalence and management of diabetic neuropathy in secondary care in Qatar. Diabetes/Metab. Res. Rev. 2020, 36, e3286. [Google Scholar] [CrossRef]
  7. Ananian, C.E.; Dhillon, Y.S.; Van Gils, C.C.; Lindsey, D.C.; Otto, R.J.; Dove, C.R.; Pierce, J.T.; Saunders, M.C. A multicenter, randomized, single-blind trial comparing the efficacy of viable cryopreserved placental membrane to human fibroblast-derived dermal substitute for the treatment of chronic diabetic foot ulcers. Wound Repair Regen. 2018, 26, 274–283. [Google Scholar] [CrossRef]
  8. Peter-Riesch, B. The diabetic foot: The never-ending challenge. Nov. Diabetes 2016, 31, 108–134. [Google Scholar]
  9. Ladyzynski, P.; Foltynski, P.; Molik, M.; Tarwacka, J.; Migalska-Musial, K.; Mlynarczuk, M.; Wojcicki, J.M.; Krzymien, J.; Karnafel, W. Area of the diabetic ulcers estimated applying a foot scanner–based home telecare system and three reference methods. Diabetes Technol. Ther. 2011, 13, 1101–1107. [Google Scholar] [CrossRef]
  10. Bluedrop Medical. Available online: https://bluedropmedical.com/ (accessed on 1 April 2022).
  11. Sugama, J.; Matsui, Y.; Sanada, H.; Konya, C.; Okuwa, M.; Kitagawa, A. A study of the efficiency and convenience of an advanced portable Wound Measurement System (VISITRAKTM). J. Clin. Nurs. 2007, 16, 1265–1269. [Google Scholar] [CrossRef]
  12. Molik, M.; Foltynski, P.; Ladyzynski, P.; Tarwacka, J.; Migalska-Musial, K.; Ciechanowska, A.; Sabalinska, S.; Mlynarczuk, M.; Wojcicki, J.M.; Krzymien, J. Comparison of the wound area assessment methods in the diabetic foot syndrome. Biocybernet. Biomed. Eng. 2010, 30, 3–15. [Google Scholar]
  13. Reyzelman, A.M.; Koelewyn, K.; Murphy, M.; Shen, X.; Yu, E.; Pillai, R.; Fu, J.; Scholten, H.J.; Ma, R. Continuous temperature-monitoring socks for home use in patients with diabetes: Observational study. J. Med. Internet Res. 2018, 20, e12460. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Frykberg, R.G.; Gordon, I.L.; Reyzelman, A.M.; Cazzell, S.M.; Fitzgerald, R.H.; Rothenberg, G.M.; Bloom, J.D.; Petersen, B.J.; Linders, D.R.; Nouvong, A. Feasibility and efficacy of a smart mat technology to predict development of diabetic plantar ulcers. Diabetes Care 2017, 40, 973–980. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Nagase Inagaki, F.N. The Impact of Diabetic Foot Problems on Health-Related Quality of Life of People with Diabetes. Master’s Thesis, University of Alberta, Edmonton, AB, Canada, 2017. [Google Scholar]
  16. van Doremalen, R.F.; van Netten, J.J.; van Baal, J.G.; Vollenbroek-Hutten, M.M.; van der Heijden, F. Infrared 3D thermography for inflammation detection in diabetic foot disease: A proof of concept. J. Diabetes Sci. Technol. 2020, 14, 46–54. [Google Scholar] [CrossRef]
  17. Crisologo, P.A.; Lavery, L.A. Remote home monitoring to identify and prevent diabetic foot ulceration. Ann. Transl. Med. 2017, 5, 430. [Google Scholar] [CrossRef]
  18. Yang, J.; Yin, P.; Zhou, M.; Ou, C.-Q.; Li, M.; Liu, Y.; Gao, J.; Chen, B.; Liu, J.; Bai, L. The effect of ambient temperature on diabetes mortality in China: A multi-city time series study. Sci. Total Environ. 2016, 543, 75–82. [Google Scholar] [CrossRef]
  19. Song, X.; Jiang, L.; Zhang, D.; Wang, X.; Ma, Y.; Hu, Y.; Tang, J.; Li, X.; Huang, W.; Meng, Y. Impact of short-term exposure to extreme temperatures on diabetes mellitus morbidity and mortality? A systematic review and meta-analysis. Environ. Sci. Pollut. Res. 2021, 28, 58035–58049. [Google Scholar] [CrossRef]
  20. Foltyński, P.; Mrozikiewicz-Rakowska, B.; Ładyżyński, P.; Wójcicki, J.M.; Karnafel, W. The influence of ambient temperature on foot temperature in patients with diabetic foot ulceration. Biocybern. Biomed. Eng. 2014, 34, 178–183. [Google Scholar] [CrossRef]
  21. Albers, J.W.; Jacobson, R. Decompression nerve surgery for diabetic neuropathy: A structured review of published clinical trials. Diabetes Metab. Syndr. Obes. Targets Ther. 2018, 11, 493. [Google Scholar] [CrossRef] [Green Version]
  22. Hernandez-Contreras, D.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.; Gonzalez-Bernal, J. Narrative review: Diabetic foot and infrared thermography. Infrared Phys. Technol. 2016, 78, 105–117. [Google Scholar] [CrossRef]
  23. Ring, F. Thermal imaging today and its relevance to diabetes. J. Diabetes Sci. Technol. 2010, 4, 857–862. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Chan, A.W.; MacFarlane, I.A.; Bowsher, D.R. Contact thermography of painful diabetic neuropathic foot. Diabetes Care 1991, 14, 918–922. [Google Scholar] [CrossRef] [PubMed]
  25. Nagase, T.; Sanada, H.; Takehara, K.; Oe, M.; Iizaka, S.; Ohashi, Y.; Oba, M.; Kadowaki, T.; Nakagami, G. Variations of plantar thermographic patterns in normal controls and non-ulcer diabetic patients: Novel classification using angiosome concept. J. Plast. Reconstr. Aesthetic Surg. 2011, 64, 860–866. [Google Scholar] [CrossRef] [PubMed]
  26. Mori, T.; Nagase, T.; Takehara, K.; Oe, M.; Ohashi, Y.; Amemiya, A.; Noguchi, H.; Ueki, K.; Kadowaki, T.; Sanada, H. Morphological Pattern Classification System for Plantar Thermography of Patients with Diabetes; SAGE Publications Sage CA: Los Angeles, CA, USA, 2013. [Google Scholar]
  27. Jones, B.F. A reappraisal of the use of infrared thermal image analysis in medicine. IEEE Trans. Med. Imaging 1998, 17, 1019–1027. [Google Scholar] [CrossRef] [PubMed]
  28. Kaabouch, N.; Chen, Y.; Anderson, J.; Ames, F.; Paulson, R. Asymmetry analysis based on genetic algorithms for the prediction of foot ulcers. In Visualization and Data Analysis 2009; International Society for Optics and Photonics: Bellingham, WA, USA, 2009; p. 724304. [Google Scholar]
  29. Kaabouch, N.; Chen, Y.; Hu, W.-C.; Anderson, J.W.; Ames, F.; Paulson, R. Enhancement of the asymmetry-based overlapping analysis through features extraction. J. Electron. Imaging 2011, 20, 013012. [Google Scholar] [CrossRef]
  30. Liu, C.; van Netten, J.J.; Van Baal, J.G.; Bus, S.A.; van Der Heijden, F. Automatic detection of diabetic foot complications with infrared thermography by asymmetric analysis. J. Biomed. Opt. 2015, 20, 026003. [Google Scholar] [CrossRef] [Green Version]
  31. Hernandez-Contreras, D.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.; Ramirez-Cortes, J.; Renero-Carrillo, F. Automatic classification of thermal patterns in diabetic foot based on morphological pattern spectrum. Infrared Phys. Technol. 2015, 73, 149–157. [Google Scholar] [CrossRef]
  32. Hernandez-Contreras, D.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.; Gonzalez-Bernal, J.; Altamirano-Robles, L. A quantitative index for classification of plantar thermal changes in the diabetic foot. Infrared Phys. Technol. 2017, 81, 242–249. [Google Scholar] [CrossRef]
  33. Hernandez-Contreras, D.A.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.D.J.; Orihuela-Espina, F. Statistical approximation of plantar temperature distribution on diabetic subjects based on beta mixture model. IEEE Access 2019, 7, 28383–28391. [Google Scholar] [CrossRef]
  34. Kamavisdar, P.; Saluja, S.; Agrawal, S. A survey on image classification approaches and techniques. Int. J. Adv. Res. Comput. Commun. Eng. 2013, 2, 1005–1009. [Google Scholar]
  35. Ren, J. ANN vs. SVM: Which one performs better in classification of MCCs in mammogram imaging. Knowl.-Based Syst. 2012, 26, 144–153. [Google Scholar] [CrossRef] [Green Version]
  36. Lu, D.; Weng, Q. A survey of image classification methods and techniques for improving classification performance. Int. J. Remote Sens. 2007, 28, 823–870. [Google Scholar] [CrossRef]
  37. Khandakar, A.; Chowdhury, M.E.; Reaz, M.B.I.; Ali, S.H.M.; Hasan, M.A.; Kiranyaz, S.; Rahman, T.; Alfkey, R.; Bakar, A.A.A.; Malik, R.A. A machine learning model for early detection of diabetic foot using thermogram images. Comput. Biol. Med. 2021, 137, 104838. [Google Scholar] [CrossRef] [PubMed]
  38. Cruz-Vega, I.; Hernandez-Contreras, D.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.d.J.; Ramirez-Cortes, J.M. Deep Learning Classification for Diabetic Foot Thermograms. Sensors 2020, 20, 1762. [Google Scholar] [CrossRef] [Green Version]
  39. Hernandez-Contreras, D.A.; Peregrina-Barreto, H.; de Jesus Rangel-Magdaleno, J.; Renero-Carrillo, F.J. Plantar thermogram database for the study of diabetic foot complications. IEEE Access 2019, 7, 161296–161307. [Google Scholar] [CrossRef]
  40. Maldonado, H.; Bayareh, R.; Torres, I.; Vera, A.; Gutiérrez, J.; Leija, L. Automatic detection of risk zones in diabetic foot soles by processing thermographic images taken in an uncontrolled environment. Infrared Phys. Technol. 2020, 105, 103187. [Google Scholar] [CrossRef]
  41. Saminathan, J.; Sasikala, M.; Narayanamurthy, V.; Rajesh, K.; Arvind, R. Computer aided detection of diabetic foot ulcer using asymmetry analysis of texture and temperature features. Infrared Phys. Technol. 2020, 105, 103219. [Google Scholar] [CrossRef]
  42. Adam, M.; Ng, E.Y.; Oh, S.L.; Heng, M.L.; Hagiwara, Y.; Tan, J.H.; Tong, J.W.; Acharya, U.R. Automated detection of diabetic foot with and without neuropathy using double density-dual tree-complex wavelet transform on foot thermograms. Infrared Phys. Technol. 2018, 92, 270–279. [Google Scholar] [CrossRef]
  43. Adam, M.; Ng, E.Y.; Oh, S.L.; Heng, M.L.; Hagiwara, Y.; Tan, J.H.; Tong, J.W.; Acharya, U.R. Automated characterization of diabetic foot using nonlinear features extracted from thermograms. Infrared Phys. Technol. 2018, 89, 325–337. [Google Scholar] [CrossRef]
  44. Gururajarao, S.B.; Venkatappa, U.; Shivaram, J.M.; Sikkandar, M.Y.; Al Amoudi, A. Infrared thermography and soft computing for diabetic foot assessment. In Machine Learning in Bio-Signal Analysis and Diagnostic Imaging; Elsevier: Amsterdam, The Netherlands, 2019; pp. 73–97. [Google Scholar]
  45. Etehadtavakol, M.; Emrani, Z.; Ng, E.Y.K. Rapid extraction of the hottest or coldest regions of medical thermographic images. Med. Biol. Eng. Comput. 2019, 57, 379–388. [Google Scholar] [CrossRef]
  46. Wang, Y.; Huang, R.; Huang, G.; Song, S.; Wu, C. Collaborative learning with corrupted labels. Neural Netw. 2020, 125, 205–213. [Google Scholar] [CrossRef] [PubMed]
  47. Aradillas, J.C.; Murillo-Fuentes, J.J.; Olmos, P.M. Improving offline HTR in small datasets by purging unreliable labels. In Proceedings of the 2020 17th International Conference on Frontiers in Handwriting Recognition (ICFHR), Dortmund, Germany, 8–10 September 2020; pp. 25–30. [Google Scholar]
  48. Hu, W.; Huang, Y.; Zhang, F.; Li, R. Noise-tolerant paradigm for training face recognition CNNs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11887–11896. [Google Scholar]
  49. Ding, Y.; Wang, L.; Fan, D.; Gong, B. A semi-supervised two-stage approach to learning from noisy labels. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1215–1224. [Google Scholar]
  50. Fraiwan, L.; AlKhodari, M.; Ninan, J.; Mustafa, B.; Saleh, A.; Ghazal, M. Diabetic foot ulcer mobile detection system using smart phone thermal camera: A feasibility study. Biomed. Eng. Online 2017, 16, 117. [Google Scholar] [CrossRef] [Green Version]
  51. Alzubaidi, L.; Fadhel, M.A.; Oleiwi, S.R.; Al-Shamma, O.; Zhang, J. DFU_QUTNet: Diabetic foot ulcer classification using novel deep convolutional neural network. Multimed. Tools Appl. 2020, 79, 15655–15677. [Google Scholar] [CrossRef]
  52. Tulloch, J.; Zamani, R.; Akrami, M. Machine learning in the prevention, diagnosis and management of diabetic foot ulcers: A systematic review. IEEE Access 2020, 8, 198977–199000. [Google Scholar] [CrossRef]
  53. Yadav, J.; Sharma, M. A Review of K-mean Algorithm. Int. J. Eng. Trends Technol. 2013, 4, 2972–2976. [Google Scholar]
  54. Taylor, G.I.; Palmer, J.H. Angiosome theory. Br. J. Plast. Surg. 1992, 45, 327–328. [Google Scholar] [CrossRef]
  55. Xie, X. Principal Component Analysis. 2019. Available online: https://www.google.com/url?sa=t&rct=j&q=&esrc=s&source=web&cd=&ved=2ahUKEwi25YTAgtn3AhUGmlYBHdtqDwYQFnoECAMQAQ&url=https%3A%2F%2Fwww.ics.uci.edu%2F~xhx%2Fcourses%2FCS273P%2F12-pca-273p.pdf&usg=AOvVaw1xc-uIUhsccvWmCWGL411_ (accessed on 1 August 2021).
  56. Keras. Available online: https://keras.io/ (accessed on 1 August 2021).
  57. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  58. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  59. Cohn, R.; Holm, E. Unsupervised machine learning via transfer learning and k-means clustering to classify materials image data. Integr. Mater. Manuf. Innov. 2021, 10, 231–244. [Google Scholar] [CrossRef]
  60. Song, K.; Yan, Y. A noise robust method based on completed local binary patterns for hot-rolled steel strip surface defects. Appl. Surf. Sci. 2013, 285, 858–864. [Google Scholar] [CrossRef]
  61. Blankenship, S.M.; Campbell, M.R.; Hess, J.E.; Hess, M.A.; Kassler, T.W.; Kozfkay, C.C.; Matala, A.P.; Narum, S.R.; Paquin, M.M.; Small, M.P. Major lineages and metapopulations in Columbia River Oncorhynchus mykiss are structured by dynamic landscape features and environments. Trans. Am. Fish. Soc. 2011, 140, 665–684. [Google Scholar] [CrossRef]
  62. Hamed, M.A.R. Application of Surface Water Quality Classification Models Using PRINCIPAL Components Analysis and Cluster Analysis. 2019. Available online: https://ssrn.com/abstract=3364401 (accessed on 1 August 2021).
  63. Malik, H.; Hemmati, H.; Hassan, A.E. Automatic detection of performance deviations in the load testing of large scale systems. In Proceedings of the 2013 35th International Conference on Software Engineering (ICSE), San Francisco, CA, USA, 18–26 May 2013; pp. 1012–1021. [Google Scholar]
  64. Toe, M.T.; Kanzaki, M.; Lien, T.-H.; Cheng, K.-S. Spatial and temporal rainfall patterns in Central Dry Zone, Myanmar-A hydrological cross-scale analysis. Terr. Atmos. Ocean. Sci. 2017, 28, 425–436. [Google Scholar] [CrossRef] [Green Version]
  65. Lloyd, S. Least squares quantization in PCM. IEEE Trans. Inf. Theory 1982, 28, 129–137. [Google Scholar] [CrossRef] [Green Version]
  66. Bock, H.-H. Clustering methods: A history of k-means algorithms. In Selected Contributions in Data Analysis and Classification; Springer: Berlin/Heidelberg, Germany, 2007; pp. 161–172. [Google Scholar]
  67. Jain, A.K. Data clustering: 50 years beyond K-means. Pattern Recognit. Lett. 2010, 31, 651–666. [Google Scholar] [CrossRef]
  68. Arthur, D.; Vassilvitskii, S. k-Means++: The Advantages of Careful Seeding; ACM Digital Library: Stanford, CA, USA, 2006. [Google Scholar]
  69. Anwar, S.M.; Majid, M.; Qayyum, A.; Awais, M.; Alnowami, M.; Khan, M.K. Medical image analysis using convolutional neural networks: A review. J. Med. Syst. 2018, 42, 226. [Google Scholar] [CrossRef] [Green Version]
  70. Tahir, A.M.; Qiblawey, Y.; Khandakar, A.; Rahman, T.; Khurshid, U.; Musharavati, F.; Islam, M.; Kiranyaz, S.; Al-Maadeed, S.; Chowdhury, M.E. Deep learning for reliable classification of COVID-19, MERS, and SARS from chest X-ray images. Cogn. Comput. 2022, 1–21. [Google Scholar] [CrossRef]
  71. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  72. Mishra, M.; Menon, H.; Mukherjee, A. Characterization of S1 and S2 Heart Sounds Using Stacked Autoencoder and Convolutional Neural Network. IEEE Trans. Instrum. Meas. 2018, 68, 3211–3220. [Google Scholar] [CrossRef]
  73. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  74. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  75. Zimmerman, J.B.; Pizer, S.M.; Staab, E.V.; Perry, J.R.; McCartney, W.; Brenton, B.C. An evaluation of the effectiveness of adaptive histogram equalization for contrast enhancement. IEEE Trans. Med. Imaging 1988, 7, 304–312. [Google Scholar] [CrossRef] [Green Version]
  76. Tawsifur Rahman, A.K.; Qiblawey, Y.; Tahir, A.; Kiranyaz, S.; Saad, M.T.I.; Kashem, B.A.; Al Maadeed, S.; Zughaier, S.M.; Khan, M.E.H.C.M.S. Exploring the Effect of Image Enhancement Techniques on COVID-19 Detection using Chest X-rays Images. arXiv 2020, arXiv:2012.02238. [Google Scholar]
  77. Chowdhury, M.H.; Shuzan, M.N.I.; Chowdhury, M.E.; Mahbub, Z.B.; Uddin, M.M.; Khandakar, A.; Reaz, M.B.I. Estimating blood pressure from the photoplethysmogram signal and demographic features using machine learning techniques. Sensors 2020, 20, 3127. [Google Scholar] [CrossRef]
  78. Hall, M.A. Correlation-Based Feature Selection for Machine Learning. Ph.D. Thesis, The University of Waikato, Hamilton, New Zealand, 1999. [Google Scholar]
  79. Multilayer Perceptron. Available online: https://en.wikipedia.org/wiki/Multilayer_perceptron (accessed on 2 March 2022).
  80. Zhang, Y. Support vector machine classification algorithm and its application. In Proceedings of the International Conference on Information Computing and Applications; Springer: Berlin/Heidelberg, Germany, 2012; pp. 179–186. [Google Scholar]
  81. Pal, M. Random forest classifier for remote sensing classification. Int. J. Remote Sens. 2005, 26, 217–222. [Google Scholar] [CrossRef]
  82. Sharaff, A.; Gupta, H. Extra-tree classifier with metaheuristics approach for email classification. In Advances in Computer Communication and Computational Sciences; Springer: Berlin/Heidelberg, Germany, 2019; pp. 189–197. [Google Scholar]
  83. Bahad, P.; Saxena, P. Study of adaboost and gradient boosting algorithms for predictive analytics. In Proceedings of the International Conference on Intelligent Computing and Smart Communication 2019; Springer: Singapore, 2020; pp. 235–244. [Google Scholar]
  84. Logistic Regression. Available online: https://en.wikipedia.org/wiki/Logistic_regression (accessed on 2 March 2022).
  85. Liao, Y.; Vemuri, V.R. Use of k-nearest neighbor classifier for intrusion detection. Comput. Secur. 2002, 21, 439–448. [Google Scholar] [CrossRef]
  86. Shi, X.; Li, Q.; Qi, Y.; Huang, T.; Li, J. An accident prediction approach based on XGBoost. In Proceedings of the 2017 12th International Conference on Intelligent Systems and Knowledge Engineering (ISKE), Nanjing, China, 24–26 November 2017; pp. 1–7. [Google Scholar]
  87. Bobkov, V.; Bobkova, A.; Porshnev, S.; Zuzin, V. The application of ensemble learning for delineation of the left ventricle on echocardiographic records. In Proceedings of the 2016 Dynamics of Systems, Mechanisms and Machines (Dynamics), Omsk, Russia, 15–17 November 2016; pp. 1–5. [Google Scholar]
  88. Gu, Q.; Li, Z.; Han, J. Linear discriminant dimensionality reduction. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Springer: Berlin/Heidelberg, Germany, 2011; pp. 549–564. [Google Scholar]
  89. Chen, T.; He, T.; Benesty, M.; Khotilovich, V.; Tang, Y. Xgboost: Extreme Gradient Boosting; R Package Version 0.4-2; 2015; pp. 1–4. Available online: https://www.google.com/url?sa=t&rct=j&q=&esrc=s&source=web&cd=&cad=rja&uact=8&ved=2ahUKEwjGh7GmhNn3AhU6tlYBHRGJBQYQFnoECAMQAQ&url=https%3A%2F%2Fcran.microsoft.com%2Fsnapshot%2F2015-10-20%2Fweb%2Fpackages%2Fxgboost%2Fxgboost.pdf&usg=AOvVaw1w25OwKdxLEfpj0rZsvL6J (accessed on 1 March 2022).
  90. Saeys, Y.; Abeel, T.; Van de Peer, Y. Robust feature selection using ensemble feature selection techniques. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Springer: Berlin/Heidelberg, Germany, 2008; pp. 313–325. [Google Scholar]
  91. Petković, M.; Kocev, D.; Džeroski, S. Feature ranking for multi-target regression. Mach. Learn. 2020, 109, 1179–1204. [Google Scholar] [CrossRef]
  92. Yusof, A.R.; Udzir, N.I.; Selamat, A.; Hamdan, H.; Abdullah, M.T. Adaptive feature selection for denial of services (DoS) attack. In Proceedings of the 2017 IEEE Conference on Application, Information and Network Security (AINS), Miri, Malaysia, 13–14 November 2017; pp. 81–84. [Google Scholar]
  93. Saidi, R.; Bouaguel, W.; Essoussi, N. Hybrid feature selection method based on the genetic algorithm and pearson correlation coefficient. In Machine Learning Paradigms: Theory and Application; Springer: Berlin/Heidelberg, Germany, 2019; pp. 3–24. [Google Scholar]
  94. Lin, X.; Yang, F.; Zhou, L.; Yin, P.; Kong, H.; Xing, W.; Lu, X.; Jia, L.; Wang, Q.; Xu, G. A support vector machine-recursive feature elimination feature selection method based on artificial contrast variables and mutual information. J. Chromatogr. B 2012, 910, 149–155. [Google Scholar] [CrossRef]
  95. Bursac, Z.; Gauss, C.H.; Williams, D.K.; Hosmer, D.W. Purposeful selection of variables in logistic regression. Source Code Biol. Med. 2008, 3, 17. [Google Scholar] [CrossRef] [Green Version]
  96. Leevy, J.L.; Hancock, J.; Zuech, R.; Khoshgoftaar, T.M. Detecting cybersecurity attacks using different network features with lightgbm and xgboost learners. In Proceedings of the 2020 IEEE Second International Conference on Cognitive Machine Intelligence (CogMI), Atlanta, GA, USA, 28–31 October 2020; pp. 190–197. [Google Scholar]
  97. Taha, A.A.; Hanbury, A. Metrics for evaluating 3D medical image segmentation: Analysis, selection, and tool. BMC Med. Imaging 2015, 15, 29. [Google Scholar] [CrossRef] [Green Version]
  98. Rahman, T.; Khandakar, A.; Kadir, M.A.; Islam, K.R.; Islam, K.F.; Mazhar, R.; Hamid, T.; Islam, M.T.; Kashem, S.; Mahbub, Z.B. Reliable tuberculosis detection using chest X-ray with deep learning, segmentation and visualization. IEEE Access 2020, 8, 191586–191601. [Google Scholar] [CrossRef]
Figure 1. Illustration of the study methodology.
Figure 1. Illustration of the study methodology.
Sensors 22 04249 g001
Figure 2. Sample of MPA, LPA, MCA, and LCA angiosomes of the foot for control and diabetic groups [37].
Figure 2. Sample of MPA, LPA, MCA, and LCA angiosomes of the foot for control and diabetic groups [37].
Sensors 22 04249 g002
Figure 3. Detailed framework of thermogram image clustering.
Figure 3. Detailed framework of thermogram image clustering.
Sensors 22 04249 g003
Figure 4. Cumulative variance vs. the number of PCA components.
Figure 4. Cumulative variance vs. the number of PCA components.
Sensors 22 04249 g004
Figure 5. Original thermogram versus enhanced thermogram using AHE and gamma correction for DM and CG [37].
Figure 5. Original thermogram versus enhanced thermogram using AHE and gamma correction for DM and CG [37].
Sensors 22 04249 g005
Figure 6. Schematic representation of the stacking model architecture.
Figure 6. Schematic representation of the stacking model architecture.
Sensors 22 04249 g006
Figure 7. t-SNE plots with the (a) TCI-based classes (class 1–class 5) and output from k-mean clustering with (b) k = 2, (c) k = 3, (d) k = 4, and (e) k = 5.
Figure 7. t-SNE plots with the (a) TCI-based classes (class 1–class 5) and output from k-mean clustering with (b) k = 2, (c) k = 3, (d) k = 4, and (e) k = 5.
Sensors 22 04249 g007
Figure 8. Diabetic thermograms classified into three severities: mild, moderate, and severe.
Figure 8. Diabetic thermograms classified into three severities: mild, moderate, and severe.
Sensors 22 04249 g008
Figure 9. Class 1, 2, 3, 4, and 5 distribution in the K-mean clustering categories—mild, moderate, and severe.
Figure 9. Class 1, 2, 3, 4, and 5 distribution in the K-mean clustering categories—mild, moderate, and severe.
Sensors 22 04249 g009
Figure 10. AUC for the original and best-performing AHE thermogram in severity classification.
Figure 10. AUC for the original and best-performing AHE thermogram in severity classification.
Sensors 22 04249 g010
Figure 11. Comparison of F1-score between original and AHE-enhanced thermogram images using 2D severity classifier.
Figure 11. Comparison of F1-score between original and AHE-enhanced thermogram images using 2D severity classifier.
Sensors 22 04249 g011
Table 1. Details of the dataset used for training (with and without augmentation), validation, and testing.
Table 1. Details of the dataset used for training (with and without augmentation), validation, and testing.
ClassifierDatasetCount of Diabetic Thermograms/Cluster
Identified in the Paper
Training Dataset Details
Training
(60% of the Data) Thermogram
/Fold
Augmented Train Thermogram
/Fold
Validation
(20% of the Data) Thermogram
/Fold
Test
(20 % of the Data) Image/Fold
SeverityContreras et al. [39]Mild432040311
Moderate482244411
Severe931806724
Table 2. Performance metrics for the best-performing combinations using Approach 1.
Table 2. Performance metrics for the best-performing combinations using Approach 1.
ClassifierFeature Selection# of FeatureClassAccuracyPrecisionSensitivityF1-ScoreSpecificityInference Time (ms)
XGBoostRandom Forest25Mild92.59 ± 6.8091.53 ± 7.2394.74 ± 5.8093.1 ± 6.5891.94 ± 7.070.397
Moderate92.59 ± 6.4786.15 ± 8.5388.89 ± 7.7687.5 ± 8.1793.89 ± 5.91
Severe92.59 ± 4.6196.64 ± 3.1793.50 ± 4.3495.04 ± 3.8291.67 ± 4.86
Overall92.59 ± 3.2992.72 ± 3.2692.59 ± 3.2992.63 ± 3.2892.31 ± 3.34
Table 3. Shortlisted features based on the six feature ranking techniques.
Table 3. Shortlisted features based on the six feature ranking techniques.
Feature PearsonChi-SquareRFELogisticsRandom ForestLightGBMTotal
TCI6
NRT (Class 4)6
NRT (Class 3)6
Mean of MPA6
Mean of LPA6
ET of LPA6
Mean of LCA6
Highest Temperature6
NRT (Class 2) 5
NRT (Class 1) 5
ET of LCA 5
NRT (Class 5) 4
STD of MPA 4
ETD of MPA 4
STD of MCA 4
ETD of MCA 4
STD of LPA 4
HSE of LCA 4
ETD of Full foot 4
Table 4. Class-wise performance metrics for the top 3 machine learning classifiers and the stacked classifier.
Table 4. Class-wise performance metrics for the top 3 machine learning classifiers and the stacked classifier.
ClassifierClassAccuracyPrecisionSensitivityF1-ScoreSpecificityInference Time (ms)
Gradient BoostMild92.01 ± 4.9891.38 ± 5.1592.98 ± 4.6992.17 ± 4.9391.71 ± 5.060.379
Moderate92.01 ± 4.7384.50 ± 6.3286.51 ± 5.9785.49 ± 6.1593.92 ± 4.17
Severe92.01 ± 3.3796.30 ± 2.3594.35 ± 2.8795.32 ± 2.6389.58 ± 3.80
Overall92.01 ± 3.4092.10 ± 3.3892.01 ± 3.4092.04 ± 3.4091.20 ± 3.55
XGBoostMild93.24 ± 4.6190.08 ± 5.4995.61 ± 3.7692.77 ± 4.7692.51 ± 4.830.336
Moderate93.24 ± 4.3889.26 ± 5.4185.71 ± 6.1187.45 ± 5.7895.86 ± 3.48
Severe93.24 ± 3.1396.75 ± 2.2195.97 ± 2.4596.36 ± 2.3390.42 ± 3.66
Overall93.24 ± 3.1593.26 ± 3.1593.24 ± 3.1593.22 ± 3.1592.31 ± 3.34
Random ForestMild91.80 ± 5.0489.19 ± 5.786.84 ± 6.2188.00 ± 5.9793.32 ± 4.580.327
Moderate91.80 ± 4.7990.43 ± 5.1482.54 ± 6.6386.31 ± 6.0095.03 ± 3.80
Severe91.80 ± 3.4193.51 ± 3.0798.79 ± 1.3696.08 ± 2.4284.58 ± 4.49
Overall91.80 ± 3.4491.71 ± 3.4691.80 ± 3.4491.67 ± 3.4789.32 ± 3.88
Stacking (Gradient Boost + XGBoost + Random Forest)Mild94.47 ± 4.2091.53 ± 5.1194.74 ± 4.1093.10 ± 4.6594.39 ± 4.230.379 + 0.336 + 0.327 = 1.042
Moderate94.47 ± 3.9992.44 ± 4.6287.30 ± 5.8189.80 ± 5.2996.96 ± 3.00
Severe94.47 ± 2.8596.81 ± 2.1997.98 ± 1.7597.39 ± 1.9890.83 ± 3.59
Overall94.47 ± 2.8794.45 ± 2.8794.47 ± 2.8794.43 ± 2.8893.25 ± 3.15
Table 5. Two-dimensional CNN five-fold testing performance of severity classifier.
Table 5. Two-dimensional CNN five-fold testing performance of severity classifier.
EnhancementNetworkClassAccuracyPrecisionSensitivityF1-ScoreSpecificityInference Time (ms)
OriginalVGG 19Mild98.77 ± 2.8698.21 ± 3.4496.49 ± 4.7897.34 ± 4.1899.47 ± 1.887.271
Moderate94.67 ± 5.5586.76 ± 8.3793.65 ± 6.0290.07 ± 7.3995.03 ± 5.37
Severe95.90 ± 3.4997.50 ± 2.7594.35 ± 4.0695.90 ± 3.4997.5 ± 2.75
Overall94.76 ± 2.8294.89 ± 2.7694.67 ± 2.8294.73 ± 2.8097.32 ± 2.03
AHEVGG 19Mild98.77 ± 2.8696.55 ± 4.7498.25 ± 3.497.39 ± 4.1498.93 ± 2.678.161
Moderate95.08 ± 5.3490.48 ± 7.2590.48 ± 7.2590.48 ± 7.2596.69 ± 4.42
Severe96.31 ± 3.3296.75 ± 3.1295.97 ± 3.4696.36 ± 3.396.67 ± 3.16
Overall95.08 ± 2.7195.08 ± 2.7195.09 ± 2.7195.08 ± 2.7197.2 ± 2.07
Gamma CorrectionVGG 19Mild88.11 ± 8.4090.91 ± 7.4687.72 ± 8.5289.29 ± 8.0388.24 ± 8.369.651
Moderate88.11 ± 7.9975.71 ± 10.5984.13 ± 9.0279.70 ± 9.9389.50 ± 7.57
Severe88.11 ± 5.7094.12 ± 4.1490.32 ± 5.2092.18 ± 4.7385.83 ± 6.14
Overall88.11 ± 4.0688.62 ± 3.9988.11 ± 4.0688.28 ± 4.0487.34 ± 4.17
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Khandakar, A.; Chowdhury, M.E.H.; Reaz, M.B.I.; Ali, S.H.M.; Kiranyaz, S.; Rahman, T.; Chowdhury, M.H.; Ayari, M.A.; Alfkey, R.; Bakar, A.A.A.; et al. A Novel Machine Learning Approach for Severity Classification of Diabetic Foot Complications Using Thermogram Images. Sensors 2022, 22, 4249. https://doi.org/10.3390/s22114249

AMA Style

Khandakar A, Chowdhury MEH, Reaz MBI, Ali SHM, Kiranyaz S, Rahman T, Chowdhury MH, Ayari MA, Alfkey R, Bakar AAA, et al. A Novel Machine Learning Approach for Severity Classification of Diabetic Foot Complications Using Thermogram Images. Sensors. 2022; 22(11):4249. https://doi.org/10.3390/s22114249

Chicago/Turabian Style

Khandakar, Amith, Muhammad E. H. Chowdhury, Mamun Bin Ibne Reaz, Sawal Hamid Md Ali, Serkan Kiranyaz, Tawsifur Rahman, Moajjem Hossain Chowdhury, Mohamed Arselene Ayari, Rashad Alfkey, Ahmad Ashrif A. Bakar, and et al. 2022. "A Novel Machine Learning Approach for Severity Classification of Diabetic Foot Complications Using Thermogram Images" Sensors 22, no. 11: 4249. https://doi.org/10.3390/s22114249

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop