Next Article in Journal
Intelligent Driving Assistant Based on Road Accident Risk Map Analysis and Vehicle Telemetry
Next Article in Special Issue
End-To-End Deep Learning Architecture for Continuous Blood Pressure Estimation Using Attention Mechanism
Previous Article in Journal
Sensing Performance Investigations on Two-Photon Fluorescent Probes for Detecting β-Amyloid in Alzheimer’s Disease
Previous Article in Special Issue
Exploration of Brain Connectivity during Human Inhibitory Control Using Inter-Trial Coherence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Classification for Diabetic Foot Thermograms †

by
Israel Cruz-Vega
1,
Daniel Hernandez-Contreras
2,
Hayde Peregrina-Barreto
3,*,
Jose de Jesus Rangel-Magdaleno
2 and
Juan Manuel Ramirez-Cortes
2
1
CONACYT Research Fellow—National Institute of Astrophysics, Optics, and Electronics, Santa Maria Tonantzintla, Puebla 72840, Mexico
2
Department of Electronics, National Institute of Astrophysics, Optics, and Electronics, Santa Maria Tonantzintla, Puebla 72840, Mexico
3
Department of Computational Science, National Institute of Astrophysics, Optics, and Electronics, Santa Maria Tonantzintla, Puebla 72840, Mexico
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in 2019 IEEE International Instrumentation and Measurement Technology Conference (I2MTC).
Sensors 2020, 20(6), 1762; https://doi.org/10.3390/s20061762
Submission received: 31 January 2020 / Revised: 6 March 2020 / Accepted: 13 March 2020 / Published: 22 March 2020
(This article belongs to the Special Issue Biomedical Signal Processing)

Abstract

:
According to the World Health Organization (WHO), Diabetes Mellitus (DM) is one of the most prevalent diseases in the world. It is also associated with a high mortality index. Diabetic foot is one of its main complications, and it comprises the development of plantar ulcers that could result in an amputation. Several works report that thermography is useful to detect changes in the plantar temperature, which could give rise to a higher risk of ulceration. However, the plantar temperature distribution does not follow a particular pattern in diabetic patients, thereby making it difficult to measure the changes. Thus, there is an interest in improving the success of the analysis and classification methods that help to detect abnormal changes in the plantar temperature. All this leads to the use of computer-aided systems, such as those involved in artificial intelligence (AI), which operate with highly complex data structures. This paper compares machine learning-based techniques with Deep Learning (DL) structures. We tested common structures in the mode of transfer learning, including AlexNet and GoogleNet. Moreover, we designed a new DL-structure, which is trained from scratch and is able to reach higher values in terms of accuracy and other quality measures. The main goal of this work is to analyze the use of AI and DL for the classification of diabetic foot thermograms, highlighting their advantages and limitations. To the best of our knowledge, this is the first proposal of DL networks applied to the classification of diabetic foot thermograms. The experiments are conducted over thermograms of DM and control groups. After that, a multi-level classification is performed based on a previously reported thermal change index. The high accuracy obtained shows the usefulness of AI and DL as auxiliary tools to aid during the medical diagnosis.

1. Introduction

Diabetes Mellitus (DM) is one of the leading worldwide causes of death and degrading elements in the quality of life of those affected by the disease [1]. There are several complications associated with DM, including heart attacks, vision loss, kidney failure, and amputations of inferior limbs. Such difficulties not only affect people’s health, but they also have a significant impact on their personal and working life. Diabetic foot is one of the main complications. It has been reported that a loss of sensitivity in the diabetic foot, along with mechanical stress in the plantar region, may increase the risk of ulceration [2], which can lead to an amputation [3]. It is also known that an increase in temperature in the plantar region of diabetic patients is associated with a higher risk of ulceration [4]. Hence, the interest in monitoring the temperature frequently through different approaches has arisen [5,6].
Thermography is a technique that has been applied to the study of diabetic foot by analyzing the thermal changes that occur in the affected foot [5]. This technique presents two main advantages. Firstly, contact is not required, and, secondly, it is non-invasive. Several works concerning its use to study diabetic foot have been reported [7,8,9,10]. Two main approaches have been proposed for thermogram analysis. These involve identifying characteristic patterns and measuring thermal changes. On one hand, a control group has been demonstrated to show a particular spatial pattern, called butterfly pattern [11]. However, there is a wide variation of spatial patterns in the DM group [12,13]. On the other hand, it is possible to measure the thermal changes and to make an assessment of these with respect to a reference. Some works propose a contralateral comparison of temperature by assuming that one foot serves as a reference to the other [14,15,16,17]. However, the contralateral comparison is limited when one of the feet cannot serve as reference. For example, if both feet have changes in temperature, but neither one of them has the butterfly pattern, then one cannot be a reference of the other. If changes in both feet have similar spatial distributions, asymmetry will not be detected, even when there is a significant change in temperature. An alternative to this approach is to measure the changes by computing a representative value for each foot of the DM group and taking the butterfly pattern as reference. Therefore, the measurement depends on the temperature distribution, and not on a spatial pattern [18,19,20]. This kind of analysis can help to describe thermograms to improve their automatic classification and bring additional information to the medical expert.
Medical information systems are changing from traditional manual data analysis to expert computer-based analysis or Computer-Aided Diagnosis (CAD) for detection purposes and in support decision systems. The primary objective is to provide accurate diagnostic support tools for clinicians. Human visual capacity for providing a diagnosis based on clinical imaging is known to have limitations, and certain effects, such as optical illusions, could affect the diagnosis accuracy [21]. Other factors to take into account are errors due to negligence, fatigue, and sensory overload caused by the massive amount of information [22]. In addition, the shortage of specialists for the diagnostic task in many healthcare institutions is also a current problem.
Traditional techniques for automatic classification, such as Artificial Neural Networks (ANN) and Support Vector Machines (SVM), are commonly used for image classification and feature extraction [23,24,25]. Both techniques are black boxes but have relevant properties that include parallel multiprocessing and the use of data-transformation with kernels. They provide high-accuracy classification results, especially in the classification and processing of images. However, prior to these processes, they require previous steps, namely preprocessing and the subsequent extraction and selection of features. Currently, there is an interest in the use of Deep Learning Networks for several tasks in medical applications, such as feature extraction, semantic segmentation and classification, among others. The shift of attention from conventional paradigms in machine learning to DL is a result of the high accuracy achieved through its massive-learning-structures, which allow DL to obtain deeper traits of the data. However, some issues need to be addressed when DL is used; these include the dataset size, the appropriate labeling of the samples, the segmentation and selection of Regions Of Interest (ROIs), the use of pre-trained structures in the mode of transfer learning, or the design of a proper new learning-structure from scratch, among others.
Although it could be relatively easy to identify when a foot thermogram does not belong to the control group at a glance, it is difficult to say how much the distribution has changed since there is no spatial pattern that describes the changes in the diabetic foot. Hence, it is important to provide a measurement or description of the thermal change that does not depend on a spatial pattern. The following considerations must be taken into account: (i) the spatial distribution could change after some time, and, (ii) in certain cases, the thermal changes are not so evident. In fact, the distribution could be visually perceived as slightly different from the control. Thus, the visual assessment of the expert can be complemented with an analysis to identify and measure the thermal changes. Moreover, a trained net specialized in diabetic foot thermograms can help to automatize this task.
Regarding the temperature changes in the diabetic foot, we are interested in detecting abnormal temperature increases in thermograms of the plantar region for their classification. This provides relevant considerations for the comparison of the three techniques: Machine Learning (ML), Singular-Value Decompositon(SVD), and DL. For this, feature extraction will be performed by an automatic segmentation process derived from an evolutionary-computing algorithm and histogram-based thresholding method using fuzzy entropy measures. Moreover, the design of a new DL-structure called Diabetic Foot Thermogram Network (DFTNet) is proposed, presenting the best results in accuracy and other performance measures, such as specificity, Area Under the Curve (AUC), and precision, among others. In the case of classifying multiple levels of temperature, the neighboring classes are complicated to classify due to similarities of the thermograms; however, DFTNet outperforms the results of such complicated cases, even in the case of the complex structures of AlexNet and GoogLeNet. With these new proposals, we are extending our previous research on comparing the Multilayer Perceptron (MLP), SVM, and AlexNet for the classification of just DM patients against the healthy ones [26]. The new DFTNet and the classification of the additional five multiple levels of the disease represent this further improvement. In addition, the use of patches instead of complete images, allowed us to increase the database for at least five times, obtaining a sufficient size of the dataset to train the DFTNet from scratch. All of these results are backed up by the incorporation of new quality measures. The DFTNet is available as supplementary material of this work.
The content of the paper is described as follows. In Section 2, relevant findings in automatic segmentation and in the use of computational intelligence-based classifiers are addressed. The methodology in Section 3 describes the steps followed in this research and the description of the proposed DL-structure. Section 4 presents the final results and Section 5 describes a discussion. Finally, the conclusions and future directions are presented in Section 6.

2. Related Work

CAD systems are generally classified into offline and online methods [27]. Most of these CAD systems follow an offline process, which consists of preprocessing the images, extracting relevant features, analyzing statistical techniques, determining the highly significant features, and finally classifying them with either some manual techniques (like thresholding or the point-to-point mean difference technique) or by using computational intelligent methods (like Artificial Neural Networks or Fuzzy Logic). In the online system, the same distinct features are obtained, and the system tries to classify as the data comes in. The latter marks a recent use of applications, like those in smartphones, as an indicative tool, not as a diagnostic one [28,29].
In the next subsections, we mention some related works of the segmentation process and the classification based on computational intelligence. Both of these are of interest to this work.

2.1. Segmentation and Feature Extraction

Image segmentation and feature extraction are two crucial steps in computer-aided diagnosis systems. The selection of a suitable segmentation technique and the extraction of relevant features can increase the accuracy of the CAD systems. Scientific literature reports several thermogram segmentation processes to isolate the plantar region from the background. Kaabouch et al. [16] evaluated five groups of auto-thresholding techniques, namely histogram shaped-based methods, clustering-based methods, entropy-based methods, object attribute-based methods, and complex genetic algorithms, and concluded that the genetic algorithms based on the thresholding technique give the best results. In 2013, Liu et al. [30] implemented the methods described in [16], but the results obtained were not satisfactory in image since there was no clear intensity difference between the foot and the background. They implemented an Active Contours Without Edges (ACWE) method; however, the images without high contrast and visible parts of the ankles and legs still needed manual adjustment. Nandagopan et al. [31] presented a comparison between two segmentation methods applied in 10 thermograms of healthy feet and 10 of diabetic ones. The results showed that the edge detection technique was more reliable than the watershed method. In [32], the snake algorithm is used to isolate the plantar region from the thermograms. The snake algorithm serves, for both purposes, to separate the feet from the background and to separate the right foot from the left one into different segmented clusters according to their temperatures. The authors claim that this snake algorithm is more robust than conventional algorithms, such as Canny, Sobel, and Otsu-thresholding. Another work [33] presents a method based on a modified active contour model that includes prior shape information. The proposed method outperforms the classical snake method, as well as other segmentation methods, such as region growing, active contour approaches, level-based methods [34,35,36], and the graphic cut approach [37].
Image segmentation techniques can also play a crucial role to isolate the hottest region in plantar thermograms, which can be used to extract relevant features. A hot region in diabetic subjects can be a sign of tissue damage or inflammation. Etehadtavakol et al. [38] demonstrated the importance of extracting the hottest/coldest regions from thermographic images using lazy snapping. Lazy snapping is an interactive image algorithm that divides coarse and fine scale processing, accomplishing object condition and detailed adjustment effortlessly. In addition, lazy snapping contributes instant visual feedback, separating the divided contour from the accurate object boundary conveniently, regardless of the existence of low contrast edges. In [18], the authors used a histogram-shape based thresholding method to extract the hottest temperatures from the plantar region. After that, they formed a feature vector based on the components of the morphological pattern spectrum, including a position criterion. In [39], Gururajarao et al. used an active contour model for plantar segmentation and a thresholding method to extract the hottest region. They divided the plantar region into six areas and extracted features, such as correlation, mean temperature difference, contrast and homogeneity, among others. In 2018, Adam et al. [40] proposed a CAD system where the segmented images are submitted to discrete wavelet transform (DWT) and higher order spectra (HOS) techniques, and then several texture and entropy features are extracted from the coefficients. The extracted features are ranked using t-values and classified using a SVM classifier. In a subsequent work [41], they use a double density-dual tree-complex wavelet transform (DD-DT-CWT) to decompose the image. Several entropy and texture features are extracted from the decomposed images, namely, Hu’s invariant moment, gray level co-occurrence matrix, Shannon, Renyi, Kapur, Yager, Fuzzy, and Vadja. The features obtained were as follows: 4032 for the left foot, 4032 for the right foot, and 8064 for the bilateral foot. Saminathan et al. [42] segmented the plantar region by using region growing and extracted texture features to 11 regions within the foot. A symmetrical analysis was performed in such regions to classify them as either normal or ulcer. Maldonado et al. [43] performed a foot segmentation from the visible image by using DL, which was used afterwards to extract the plantar region in the thermogram. They classify the risk of ulceration or necrosis according to temperature differences.

2.2. Classification with Computational Intelligence Methods

The use of computational intelligence algorithms (CI) has been extended successfully in applications that involve pattern recognition, classification, automatic control and optimization, among others. Within the black-boxes methodologies, ANN is considered to be the most popular one, where the vast majority of the reported article reaches an accuracy of 80% in the prediction of DM. An alternative that has improved such level of accuracy is SVM, rising as the most successful algorithm in both biological and clinical datasets in DM [44]. The reason is that ANN uses derivative-based methods to update the weights, undergoing a slow convergence rate and often yielding suboptimal solutions [45]. Another outstanding methodology that has been gaining relevance is the use of massive neural structures to learn at several levels of abstraction; this is called Deep Learning. We will now mention some related works using CI methodologies in several tasks of DM’s foot classification. Kavakiotis et al. [44] present a systematic review of machine learning applications and data mining. The tasks include prediction, diagnosis, complications and health care, among others related to diabetes. After mentioning a series of different types of DM and examples of data, the paper compares algorithms, like Logistic regression (LR), ANN, Random Forest, and k-NN. We would like to emphasize that the accuracy of an algorithm depends on the type of data (dimensionality, origin, and kind); however, SVM is the most successful and widely used classifier. In [46], the work makes emphasis on efficient coding data by decreasing input data redundancy using independent component analysis algorithms (ICA). The results are obtained by testing the algorithm in the Pima Indians Diabetes database, where the SVM algorithm classifies the diabetics with a 98% accuracy rate. ICA improves the ability of the classifier by finding proper class boundaries; the reason is that ICA reduces the statistical dependence of the collected data. An exciting work mentioned in [47] goes beyond categorizing the types of DBs. The work tries to envisage the side-effects or other chronic diseases a patient should anticipate. For such purpose, the prediction/classification tasks consider more descriptive information composed of independent variables from former and current consultants of DM. Moreover, they add symptoms of the patients to the algorithms. Other works like [48] present an ensemble-type classifier, where algorithms like ANN, Naive Bayes, k-NN, and Random Forest, among others, are combined and perform better than all other individual counterparts.
All the works mentioned above and many others follow a crucial step for the automated analysis: the extraction of discriminant features from the images. This process is still done by human researchers. The real tendency is to let computers learn the features by themselves, trying to optimally represent the data for the problem at hand. This is the basic concept of many deep learning algorithms, where the massive structure of the ANN transforms the data while it learns increasingly higher level features. The tendency in the use of DL is to provide an increase in the accuracy level in classification processes; however, DL usually requires a large amount of data in the training process and the difficult task of labeling such training set. Some promising alternatives have emerged to deal with these drawbacks. There are two main ways to train DL architectures, either by training from scratch (or full training), which requires a large quantity of labeled training data, or by using extensive computational and memory resources, like Graphics Processing Units (GPUs), to obtain high velocities of processing. In the medical domain, this requirement may be difficult in the task of expert annotation and in the availability of patients to get a considerable number of images. A promising alternative is to fine-tune a pre-trained DL structure using a large labeled dataset from a different application. In the case of medical images, fine-tuning proved to either outperform or, in the worst case, perform as well as a Convolutional Neural Network (CNN) trained from scratch. The experiments considered four different medical imaging applications in three specialties: radiology, cardiology, and gastroenterology, involving classification and segmentation [49].
Litjens et al. present in [50] a work surveying the use of DL in medical image analysis. They study the implementation for image classification, object detection, segmentation, registration, and other tasks, mentioning its application in neural, retinal, pulmonary, digital pathology, breast, cardiac, abdominal, and musculoskeletal medical areas. The work presented in [51] proposes the use of traditional computer vision techniques in the detection of diabetic foot ulcers (DFU). They emphasize the importance of these methods, which represent a cost-effective, remote, and convenient healthcare solution over the traditional costly clinical approaches that rely on patient and clinical vigilance. Their proposal consists of improving the extraction of essential features for DFU classification based on a neural architecture, which is a combination of essential aspects of CNNs in their depth and parallel convolutional layers. The network is called diabetic foot ulcer network (DFUNet), and it reaches an accuracy of 94 % , outperforming DL pre-trained structures like LeNet, AlexNet, and GoogleLeNet. The same author extends his research using DL in multi-class semantic segmentation for melanomas, benign lesions, and the pre-cancer stage called seborrhoeic keratoses [52].
Finally, we can conclude that, in the field of medical images, DL techniques are permeating and improving traditional processes, including the analysis, diagnosis, detection, classification, and segmentation. There are still challenges to be met, such as the automatic annotation to delineate and classify the images without the help of a specialist. Other issues include multi-class classification, as well as the improvement of the automatic classifiers in the detection, recognition, segmentation, and monitoring of the disease, among others.

3. Methodology

This section describes all the steps followed to obtain a final comparison of the machine learning-based classifiers mentioned above. For purposes of this work, 110 thermograms of DM subjects obtained from a public thermogram database were used [53].
In order to use the MLP and SVM algorithms, a process of selection of the region of interest (ROI) for segmentation is required in addition to the further extraction of relevant features. In this case, we use a histogram-based segmentation method represented by fuzzy sets and optimized with an evolutionary optimization technique. This process is mentioned briefly in this section, but it is expanded in the appendices at the end of the work. Finally, we briefly mention the description of the three machine learning-based classifiers and the new proposed DL structure.

3.1. Dataset and Data Augmentation

When DL structures are trained from scratch, they require a large number of images because of the enormous number of parameters trained in them. Data augmentation is an affordable technique to obtain such quantity of data when we do not dispose of it. Data augmentation consists of a combination of various processing techniques, like rotation, flipping, contrast enhancement, using different color, space, and random scaling. In this work, rotation is performed at angles of 90 , 180 , and 270 . We used three types of flipping (horizontal flip, vertical flip, and horizontal+vertical flip) performed on the original patches. We also obtained several patches of each image, allowing us to increase the data set tenfold. Figure 1 shows five classes of thermal change that can be found in the database and an example of the extracted patches.

3.2. Automatic Segmentation

The MLP and SVM require the extraction of relevant features to be introduced to the classifiers before training them. In this paper, the ROI of DM patients is segmented before the feature extraction using a histogram-based method. In the process of obtaining the partition of a digital image into multiple segments, a variety of image segmentation methods have been developed, such as thresholding, clustering-based methods, compression-based methods, histogram-based methods, and edge detection, among others. This work uses a histogram-based method, mainly using fuzzy logic, representing the segments of the image. The fuzzy logic approach for image processing allows us to use membership functions, defining the degree to which a pixel belongs to one segment or another. Furthermore, we obtain a better definition of segments by using fuzzy logic according to the measure of entropy. The optimized parameters are obtained by using a heuristic optimization technique based on Differential Evolution. The Appendices Appendix A and Appendix B contain an expanded explanation of this, and Figure 2 represents the basic steps of this process.

3.3. Machine Learning Classifiers

Computer aids have become an indispensable necessity in image-based medical decision-making processes, such as in the detection and diagnosis, automated segmentation, automated image annotation, and image retrieval [54]. The use of computational intelligence algorithms tries to overcome some drawbacks, including the exhaustive task of interpreting a large number of images. Other algorithms involve achieving more accurate diagnosis systems, thereby providing a higher level of reliability, which the patient needs. The advance of medical imaging technologies, which provide new imaging modalities and methodologies, require new computational algorithms according to the characteristics of the images and the high quantity of handled data. Some of the most widely used algorithms are ANN and SVM. However, DL has been used in the classification of medical images as well as in other fields, as a result of the high accuracy rate obtained from its complex learning structure, and, in many cases, the vast quantity of data processed, which allows these structure to obtain multiple levels of feature abstractions from the data. In this work, the former two algorithms are compared with DL, presenting the advantages and disadvantages of their use. For a better explanation of these algorithms, please refer to the sections of the appendix.

3.4. Multiple Classes

Previous research of some authors in this work showed that it is possible to estimate local plantar temperatures based on the angiosome concept [19]. An angiosome is a composite unit of tissues supplied by an artery, providing valuable information relating temperature data to artery damage. For this purpose, the foot is divided into four angiosomes: medial plantar artery (MPA), lateral plantar artery (LPA), medial calcaneal artery (MCA), and lateral calcaneal artery (LCA). The obtained information using the angiosomes is related not only to the damage generated by DM in arteries but also to the associated ulceration risk since it is used to compute local temperatures. See Figure 3 for the illustration of the angiosomes.
Additionally, they proposed a new thermal change index (TCI) as an attempt to overcome some drawbacks in the measurement of thermal measure-based algorithms in the diabetic foot. These include differences in thermal pattern assumptions when both feet are compared, which occur, for instance, when the patient presents a partial or total amputation or ulceration. TCI takes advantage of the Control Group (CG) well-established butterfly pattern that is used individually to compare each foot affected by the disease. The TCI value is the mean difference between corresponding angiosomes from a DM subject and the reference values obtained from angiosomes of the CG:
T C I = C a n g D M a n g 4
where C a n g and D M a n g are the temperature values of the angiosome for the control group and for a DM subject, respectively.
Five categories of the change degree of the plantar regions were obtained based on the results in [19]. These are used in this work to test the classification algorithms with multiple-classes of the DM-thermal images.

3.5. Performance Evaluation and Classification Scheme

In all our experiments, we report Sensitivity, Specificity, Precision, Accuracy, F-measure, and Area under the curve (AUC) as our evaluation metrics. It is well known that, in medical imaging, Sensitivity and Specificity are relevant metrics to evaluate classifier completeness [56]:
S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N F P + T N
P r e c i s i o n = T P T P + F P
A c c u r a c y = T P + F N T P + T N + + F P + F N
F m e a s u r e = 2 T P 2 T P + F P + F N

3.6. Proposed Deep Learning Network

After not being able to obtain satisfactory results with SVM, MLP, AlexNet, and GoogLeNet, especially with the five levels of classification, we propose a new type of DL structure. The name of this network is Diabetic Foot Thermograms Network (DFTNet). With this proposal, we considerably reduce the number of layers, compared with the 22 layers of GoogLeNet, which also result in a decreasing training time.
The parameters used from training the DFTNet are a maximum of 100 epochs, a minibatch size of 64, and the Adam solver with a learning rate of 0.001. The configuration of the computer is: CPU Intel i7–7700 HQ @2.8 GHz, GPU NVIDIA GeForce GTX 1060, RAM 16 GB, Software Matlab. The structure of DFTNet is shown in Table 1.

4. Experimental Results

We divide the experimental results into two parts. The first part presents the results of the process of automatic segmentation, using the measure of fuzzy entropy and optimizing with meta-heuristics like DE. We show the relevant results from this procedure that let us obtain the best region of interest (ROI). The second part presents a comparison of the most pertinent classifiers in the area of computational intelligence; these are ANN, SVM, and DL. The first two require the additional stage, which involves segmenting the ROI and extracting relevant features, whereas the last one “learns” directly from the dataset.

4.1. Automatic Segmentation with DE

As mentioned earlier, diabetic foot classification follows an automatic segmentation to obtain the ROI. In this case, DE is used as an optimization procedure of the fuzzy entropy measure. Once a normalized histogram represents the image with L = 255 gray levels, the segmentation levels determining the thresholds are optimized with DE. The objective is to obtain the maximum value of the total fuzzy entropy function (See Appendix), where each n-level segment is represented by fuzzy sets and its corresponding membership function values. After the optimal value of entropy is reached, the n 1 number of thresholds is extracted from the fuzzy parameters.
In our case, the ROI for a healthy subject as part of the control group (CG) is given in the Butterfly Pattern distribution, which presents the highest temperature in the arch, and decreases as it moves away from this area. In the case of the DM group, this distribution of highest temperatures defining the ROI is shown in different regions of the foot [18]. In the histogram of the thermogram with 32 gray levels, the global maximum represents the image background, while the other represents the plantar region (PR). Thus, by using thresholding, which is one of the most common pixel-based methods to segment an image [57], the ROI is obtained for posterior classification with three of the most common and relevant classifiers.
Following the procedure of automatic segmentation, several thresholds are considered to maximize the proposed fuzzy entropy measure, and, at the same time, obtain the desired pattern, distinguishing the CG from the DB group. We can consider that, for each proposed number of thresholds ( n 1 ), we will have n-level segments or classes.
For instance, Figure 4 shows the results of one segmented image of the DB group for two, three, and four thresholds. The entropy values obtained were 0.78106 , 0.52698 , and 0.45723 , respectively. We also tried the one-threshold case, defining two segments or classes, obtaining the highest level of entropy: 0.96938 . However, the segmented pattern does not satisfy a defined pattern for the ROI. Thus, our criteria lie between having a high value of entropy and obtaining a defined ROI pattern at the same time.
This automatic segmentation based on DE defines the thresholds of fuzzy sets. Figure 5 and Figure 6 show the initial and final fuzzy sets after the DE-optimization process for one case of the DM and CG, respectively. The location of the selected thresholds is in the middle part of the overlap of fuzzy sets. We can observe the modification not only in the parameters and form of the fuzzy sets by the DE, but also in the location of the thresholds. The initial thresholds are located in T 1 = 64 and T 2 = 243 within the 255 values of the histogram. After the evolutive optimization, these values are: T 1 = 56 and T 2 = 167 . The next section presents the results of classification with SVM and ANN. The extracted features for these classifiers were obtained with this automatic segmentation process.

4.2. Classification of Multiple Classes

For multi-class classification (see Section 3.4), three different machine learning algorithms were used: MLP, SVM, and DL. The first two machine learning classifiers require an extracted vector of features from the segmented images by the DE histogram-based algorithm, whereas DL does not need this step due to its massive structure. The extracted feature’s vector for the first two algorithms contains five elements per image, and these are related to the segmented area: number of pixels, mean value, variance, maximum entropy value, and the index value.
We begin by presenting the results of ANN and SVM, which share a similar classification procedure in the sense that they use extracted features from the ROI. The training process of both techniques requires a split of the training set into 85% for training, 5% for validation, and 10% for testing. We adopted the 10-fold cross-validation technique. Both algorithms run on Matlab. The results of the MLP and SVM with the required measures are shown in Table 2. The best results were obtained with the 3rd segmentation level. This is why we only present these results. In addition, the best results were achieved using a 3-layer ANN structure and a linear kernel for the SVM classifier.
In these cases, the multiple classes are obtained from the TCI-values. The classification is a bit complicated because of the similarities in thermograms among DM-patients. Regarding ANN and SVM, SVM has some advantages in improving some performance measures due to the high-dimensional mapping and the use of nonlinear kernels, enhancing the classification accuracy.
In addition, we use two pre-trained DL models (AlexNet and GoogLeNet), and our proposed DL-structure, the DFTNet. AlexNet is a convolutional neural network that has been trained on over a million images and can classify images into 1000 object categories (such as a keyboard, coffee mug, pencil, and many animals) [58]. The training set is divided into 70 % for training and 30 % for validation. AlexNet has five convolutional layers and three fully connected layers. The fine-tuning process in our case consists of extracting the last three layers that were previously configured to classify 1000 images and to be re-trained with our data set of the thermal images of diabetic and control groups. We replace these three layers by a fully connected layer, a soft-max layer, and a classification output layer. The first layer requires input images of 227-by-227-by-3, where 3 is the number of color channels. In addition, the process requires additional augmentation on the images: performing operations like randomly flipping along the vertical axis and randomly translating them up to 30 pixels, horizontally and vertically. This step of data augmentation prevents the network from overfitting and memorizing the exact details of the training images.
GoogLeNet is another state-of-the-art CNN architecture with 22 deep network layers. GoogLeNet proposes a new module called inception, which acts as a multiple convolution filter input, and it is processed on the same input while doing pooling at the same time. All the outcomes are merged into a single feature layer, allowing the model to take advantage of multiple level feature extraction from each input [59].
The performance measures of the pre-trained AlexNet and GoogLeNet models and those obtained by our proposed DFTNet model, which was trained from scratch, are shown in Table 3 and Table 4, respectively. A 10-fold cross-validation technique was used for each case.
We achieved some good results in DL-structures with well-separated classes like 1-5, 2-5, 4-1, and so on. However, neighboring classes, especially 3-4, present the lowest values in precision, accuracy, and other measures. The problem is that, in neighboring classes, and especially in classes like 3-4 and 4-5, the patterns are so similar and the training images do not provide sufficient information to learn in the DL-structures.

5. Discussion

Two traditional machine learning classifiers and three models based on CNNs were used for classification of the thermograms. We present a five-level classification procedure, where the thermograms are divided into five classes according to the TCI mentioned in Section 3.4.
The first two classifiers required an additional step of feature extraction, obtaining the ROI with the histogram-based segmentation procedure, which uses a fuzzy entropy measure and optimization of the segments with DE. The extracted feature vector contains relevant information from the thermogram’s ROIs, like the number of pixels, mean value, variance, maximum entropy value, and the index value of the grayscale (see Figure 4). This step is a little bit cumbersome due to the optimization procedure and the attempt with different thresholds in the segments to obtain the best ROIs. It is a method that involves more the participation of the expert. In regard to the dataset comments, the number of samples is not large, and we did not even increase the number of samples by dividing each image into patches. The CNN classifiers require the additional step of data augmentation for the three structures (GoogLeNet, AlexNet, and DFTNet). This is why we chose to divide each image into patches, like the one shown in Figure 1, increasing our data set about ten times in addition to the data augmentation process of the CNN structures. The classification is carried out comparing one-vs.-one classes (OVO) to use regular performance measures. We obtained the highest values in sensitivity, specificity, AUC, and Accuracy with classifiers like SVM, and, in the CNN structures, in almost all the pairs of compared classes, especially in well-separated classes like 1-3, 1-4, 1-5, 2-4, 2-5, 3-5, etc. In consecutive classes, the best results were obtained in cases 1-2 and 4-5. Because distributions without thermal change are found in class 1, the transition to class 2 is important as this may indicate the initial step in a thermal change. This transition is remarked by the loss of the butterfly pattern with slightly higher temperatures [19]. However, classes 3 and 4 present the lowest values of performance measures due to the similarities between them. In these two classes, we can find larger hot regions, but the difference is not as obvious as in the transition to class 5, where the hot regions already cover almost the entire plantar region. Figure 7 shows the ROC curves for this worst case of comparison between classes 3 and 4. In this case, SVM achieved the lowest values of AUC and other measures, like sensitivity and accuracy. The two highest results are of our DFTNet and the common ANN, with AUC-values of 0.8533 and 0.8333, respectively. ANN has 0.66 in sensitivity and 0.833 in accuracy, while DFNet has 0.9167 in sensitivity and 0.853 in accuracy. We can attribute the highest results of our DFNet to the specific design of the network to this type of images. Although the structures of GoogLeNet and AlexNet are more complex and are supposedly better classifiers, they were trained with a different type of images.
In general, we can comment that CNNs have advantages when they are trained with a sufficient number of images if they are trained from scratch. The other option is using pre-trained networks, but it works well if datasets are not too difficult to classify. In this research, the similarities between classes when we classified the five levels made it difficult for the pre-trained CNNs and SVM. In addition, dividing the images into five levels using the TCI values decreases the number of samples for each class, which is not optimal for the classifiers, especially for the CNNs. The tactics of dividing each image into patches selected by the expert, along with the operations of data augmentation for CNNs and the design of our own CNN structure, provided the best results for the easiest and worst cases of classification.

6. Conclusions

This work presents a comparison of conventional classifiers like ANN, SVM, and those of current importance like CNNs. The aim is to classify patterns in thermal images of patients with DM. The work includes the classification of five levels within DM-patients. The results of the first simulations using traditional ANN and SVM classifiers obtained satisfactory results after a feature extraction process. However, the tendency of using DL structures is not only to gain an increase in the accuracy of the classification, but also to avoid, in some cases, the exhaustive task of feature extraction and segmentation of the desired patterns. An advantage of supporting our work in the use of such DL structures is the use of pre-trained networks. The results obtained with DL are better and have low training time.
We can observe in the results that consecutive classes were affected by similarities in temperature distributions of the thermograms. In this case, CNNs like GoogLeNet and AlexNet obtained the same accuracy results, which were not satisfactory. We proposed a new design of CNN with a simple structure but a better design. The proposed DFTNet lets us obtain satisfactory results using measures of sensitivity, specificity, accuracy, and AUC-values, among others, even in the worst cases of classification.
As future work, we are working on obtaining more images of thermograms, and improving the structure of our CNN, perhaps with parallel convolution operations, like the GoogLeNet. We aim to get better results of classification with less participation from the expert in the selection of patches and ROIs.
This research is aimed to classify thermograms and predict the presence of ulceration in feet. Another extension of this work could be to classify feet with ulcerations, which contain quite distinctive texture and color features when compared with normal healthy skin.

Supplementary Materials

The proposed DFTNet is available online at https://www.mdpi.com/1424-8220/20/6/1762/s1.

Author Contributions

Investigation, D.H.-C., H.P.-B., and J.d.J.R.-M.; Methodology, I.C.-V.; Writing—original draft, I.C.-V.and D.H.-C.; Writing—review and editing, H.P.-B., J.d.J.R.-M., and J.M.R.-C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Segmentation: Multilevel Shannon Entropy and Fuzzy Entropy

Regarding Information Theory, entropy is the measure of the amount of information that is missing before the reception of a message (information) [60]. Information entropy measures the average amount of information in a message in terms of discrete probabilities p i .
In terms of image segmentation, the concept of Shannon entropy is based on the maximum amount of information concerning the distribution of the target and background of the image. By measuring the entropy of the gray histogram of the image, the best threshold is found [61]. Thus, the entropy of the total image is defined as follows:
H ( I ) = i = 1 n p i log 2 p i
where p i is the probability of occurrence of the i-th gray level, and I denotes an 8-bit gray level image of dimension M × N . P is the normalized histogram for the image with L = 255 gray levels. In this case, we obtain the n 1 thresholds (t) partitioning the normalized histogram into n classes, that is, by maximizing the total entropy. This measure constitutes individual measures of entropy for each one of the proposed classes in the image [62].
However, a more precise measure of entropy that deals with imprecise, vague, or noisy data relies on Fuzzy Logic, having fuzzy sets as essential elements, and the corresponding membership value within such sets.
Fuzzy sets are a generalization of classical sets, that is, instead of using a sharp definition of belongingness of an element to a certain set, such element can partially belong to this set and another with certain degrees. A fuzzy set is defined as A = x , μ A ( x ) ) | x X , where 0 < μ A ( x ) < 1 is the membership function measuring the closeness (or belongingness) of x to A.
In this case, fuzzy sets represent the segmented regions defining the image classes. For instance, using trapezoidal membership functions, the estimation of the membership parameters of the n segmented regions, μ 1 , μ 2 , , μ n for the n-level thresholding, namely a 1 , c 1 , , a n 1 , c n 1 , where 0 a 1 c 1 , , a n 1 c n 1 L 1 . See Figure A1.
Figure A1. Fuzzy membership function for n-level thresholding [26].
Figure A1. Fuzzy membership function for n-level thresholding [26].
Sensors 20 01762 g0a1
According to [63], the membership function for the n-level thresholding is:
μ 1 ( k ) = 1 k a 1 k c 1 a 1 c 1 a 1 k c 1 0 k > c 1
μ n 1 ( k ) = 0 k a n 2 k a n 2 c n 2 a n 2 a n 2 < k c n 2 1 c n 2 < k a n 2 k c n 1 a n 1 c n 1 a n 1 < k c n 1 0 k > c n 1
μ n ( k ) = 1 k a n 1 k a n c n a n a n 1 k c n 1 1 k > c n 1
The maximum fuzzy entropy for each of the n-level segments is:
H 1 ( t ) = i = 0 L 1 p i μ 1 ( i ) P 1 ln p i μ 1 ( i ) P 1 H 2 ( t ) = i = 0 L 1 p i μ 2 ( i ) P 2 ln p i μ 2 ( i ) P 2 H n ( t ) = i = 0 L 1 p i μ n ( i ) P n ln p i μ n ( i ) P n
where
P 1 ( t ) = i = 0 L 1 p i μ 1 ( i ) ; P 2 ( t ) = i = 0 L 1 p i μ 2 ( i ) ; ; P n ( t ) = i = 0 L 1 p i μ n ( i )
Maximizing the total entropy allows us to obtain the optimal value of parameters:
ϕ ( a 1 , c 1 , , a n 1 , c n 1 ) = A r g m a x ( H 1 ( t ) + H 2 ( t ) + + H n ( t ) )
Then, extracting the n 1 number of thresholds from the fuzzy parameters as follows:
t 1 = a 1 + c 1 2 ; t 2 = a 2 + c 2 2 ; t n 1 = a n 1 + c n 1 2 .
The purpose of this work is to modify these parameters, trying to obtain an optimal segmentation of the image. In this work, we use heuristic optimization based on differential evolution.

Appendix B. Differential Evolution

Evolutionary computing (EC) draws inspiration from the process of natural evolution. A common task in the use of EC algorithms is the optimization of specific parameters of the proposed model, delivering the correct output for each known input [64]. Some of the existing techniques in EC include genetic algorithms (GA) or particle swarm optimization (PSO). We use differential evolution (DE), which has been shown to outperform GA and PSO for multilevel thresholding based image segmentation problems. Sarkar et al. [62,65] presented a comparison between these algorithms in terms of computational time, mean objective value, and standard deviation. The results favor DE in both visual and statistical comparisons of the segmented images.
The main idea with DE is that, given a population of candidate solution vectors in R n , the i-th individual (parameter vector) of the population at generation (time) t is a D-dimensional vector containing a set of D optimization parameters:
X ¯ i ( t ) = x 1 , x 2 , x D
A change in the members of the population is performed by the creation of a donor vector P ¯ . In this work, the donor vector P ¯ for each i-th member is created using the DE/rand/1 variant [66], that is, it is created by randomly choosing three other parameter vectors ( ( r 1 , r 2 , r 3 ) 1 , N P ), and r 1 r 2 r 3 . NP is the number of population members. Then, P ¯ is obtained by multiplying a scalar number F with the difference of any two of the three r i as follows:
P ¯ i , j = X r 1 , j + F · ( X r 2 , j X r 3 , j )
where j-th is the component of the i-th vector [66]. Henceforth, for simplicity of the equations, we have omitted the use of t. A binomial crossover operation is used to increase the potential diversity of the population. The binomial crossover is performed for each of the D variables, subject to the crossover parameter C r 0 , 1 . In addition, the number of parameters inherited from the mutant has a (nearly) binomial distribution. Thus, for each target vector X ¯ i , a trial vector R ¯ i is created as follows:
R ¯ i , j = P i , j if r a n d i ( 0 , 1 ) C r X i , j o t h e r w i s e T h e
where r a n d is a number in a uniform random distribution. Finally, the selection is performed to determine which one, whether the target vector or the trial vector, will survive in the next generation. If the trial vector yields a better value of the fitness function, it replaces the target vector in the next generation; otherwise, the parent remains without change:
X ¯ i , j = R ¯ i , j if f ( R ¯ i , j ) > f ( X ¯ i , j ) X ¯ i , j o t h e r w i s e
where f ( · ) is the fitness function.

Appendix B.1. Artificial Neural Networks

The ability to learn from input data, with or without a teacher, has prioritized ANN as one of the primary tools in classification, segmentation, regression, and extraction of outstanding characteristics. A common ANN type for pattern recognition is the Multilayer Perceptron (MLP) feed-forward neural network. The success in the learning process of an ANN is related mainly to two aspects: the ANN architecture and the learning algorithm. For this case, the MLP is trained using the error-back-propagation algorithm. This is based on a generalized delta learning rule as an iterative gradient-based algorithm that minimizes the root-mean-squared error between the actual output of the ANN and the training set or desired output. The learning process consists of two steps. The first step is the feed-forward step, without modifications to the network’s parameters, that is: (1) Given a vector of characteristics X = x 1 , x 2 , , x N , defining the ANN-input, it is pondered by a vector W of weights between neurons; (2) The induced local field of the ANN is produced by summing the products of the previous step, that is: v = W X ; (3) Finally, the ANN-output is computed as y = ϕ ( v ) , where ϕ is the activation function of the neuron.
The second step, which is the back-forward step, modifies W , in such a way that the error function ϵ = 1 N k = 1 n e ( k ) 2 is minimized, e ( k ) = d ( k ) y ( k ) , and N is the number of samples. In general, the weights’ update follows a gradient descend learning rule, which is also called delta-rule:
w ( k + 1 ) = w ( k ) η E W
where η is the learning rate. Some other algorithms that improve the convergence in the learning process have been proposed. One of these is the called Levenberg–Marquardt.

Appendix B.2. Support Vector Machines

Support Vector Machines (SVMs), like MLP and radial-basis-function networks, are another category of universal feed-forward networks. They are usually used for pattern classification and nonlinear regression [67]. SVMs like the MLP provide a hyperplane between two classes, functioning as a decision surface, but with the difference that SVMs maximize the margin of separation between positive and negative examples of such classes. Considering the training example ( x i , y i ) i = 1 N , where x i is the input pattern, and y i = ± 1 the target output, the equation of the decision surface in the form of a hyperplane is:
f ( x ) = w T x + b = 0
where w is a weight vector and b is a bias. This best separating hyperplane is the result of finding w and b, minimizing w , such that, for all points ( x i , y i ) i = 1 N ,
y i f ( x i ) 1
The support vectors are the x i on the boundary which y i f ( x i ) = 1 . The problem of minimizing w is treated as a quadratic programming problem, obtaining the optimal solution by means of a Lagrange formulation [68].
In cases of nonlinearly separable data, SVMs allow the construction of hyperplanes in high dimensional feature spaces, retaining all the simplicity of an SVM-separating hyperplane. This is possible by using kernels and the mapping to these high dimensional spaces with inner-products (dot products). Depending on the kernel of the inner-product, different learning machines like polynomials, Gaussians and sigmoids functions, among others, characterize different nonlinear decision surfaces. The improvement in the performance of an SVM is by selecting the features that match a particular kernel, choosing an appropriate kernel that fits the features, or linearizing the features to a higher dimensional space and matching them with linear kernels [69].

Appendix B.3. Deep Learning

Deep learning (DL) are computational models composed of multiple processing layers. These models are characterized by learning data representations with various levels of abstraction and discovering complex structures in large datasets. DL methods have dramatically improved the state-of-the-art in speech recognition, visual object recognition, object detection, and genomics, among others [70]. In this work, we use a Convolutional Neural Network (CNN) that is a subtype of deep discriminative architecture with satisfactory performance in processing two-dimensional data with grid-like topologies, such as images and videos [71].
In CNNs, the convolutional layers carry out the detection of certain local features in all the locations of their input images. Each node connects to only a small subset of spatially connected neurons in the input image channels. The sets of connection weights shared between the nodes, namely the kernels, enable the search for the same local features in the input channels. A convolutional layer has n kernels that learn to detect n local features, whose strength across the input images is visible in the n resulting feature maps. The pooling layer follows the convolutional layer and is used in order to reduce the computational complexity to achieve a hierarchical set of image features. CNNs contain several pairs of convolutional and pooling layers, followed by some consecutive fully connected layers, and, finally, a softmax layer to generate the desired outputs.
CNNs are trained with the back-propagation algorithm minimizing a cost function of the unknown weights. The cost function is an averaged value of the logarithmic representation of the probability by which each training image is correctly classified over the whole training set. It is represented as:
L = 1 X i X l n p y i X i
where X denotes the number of training images, X i is the i t h training image with the corresponding label y i , and p y i X i is the probability value indicating the correct classification of X i . A typical training algorithm is stochastic gradient descent for minimizing the cost function, the action of reducing the cost over the entire training set is approximated with the cost over mini-batches of data. The process of updating the weights follows the next formulations [49]:
γ t = γ t N X V l t + 1 = μ V l t γ t α l L ^ W l W l t + 1 = W l t + V l t + 1
where W l t denotes the weights in the l-th convolutional layer at iteration t, L ^ means the cost over a mini-batch of size N, α l is the learning rate of the l-th layer, μ is the momentum indicating the contribution of the previous weight update in the current iteration, and γ is the scheduling rate that decreases the learning rate α at the end of each epoch.

References

  1. Cho, N.H.; Kirigia, J.; Mbanya, J.C.; Ogurstova, K.; Guariguata, L.; Rathmann, W.; Roglic, G.; Forouhi, N.; Dajani, R.; Esteghamatil, A.; et al. IDF Diabetes Atlas, 8th ed.; IDF: Brussels, Belgium, 2017. [Google Scholar]
  2. Sims, D.S., Jr.; Cavanagh, P.R.; Ulbrecht, J.S. Risk factors in the diabetic foot: recognition and management. Phys. Ther. 1998, 68, 1887–1902. [Google Scholar] [CrossRef] [PubMed]
  3. Iversen, M.; Tell, G.; Riise, T.; Hanestad, B.; Østbye, T.; Graue, M.; Midthjell, K. History of foot ulcer increases mortality among individuals with diabetes: ten-year follow-up of the Nord-Trøndelag Health Study, Norway. Diabetes Care 2009, 32, 2193–2199. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Ring, F. Thermal imaging today and its relevance to diabetes. J. Diabetes Sci. Technol. 2010, 4, 857–862. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Hernandez-Contreras, D.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.; Gonzalez-Bernal, J. Narrative review: Diabetic foot and infrared thermography. Infrared Phys. Technol. 2016, 78, 105–117. [Google Scholar] [CrossRef]
  6. Martín-Vaquero, J.; Hernández Encinas, A.; Queiruga-Dios, A.; José Bullón, J.; Martínez-Nova, A.; Torreblanca González, J.; Bullón-Carbajo, C. Review on wearables to monitor foot temperature in diabetic patients. Sensors 2019, 19, 776. [Google Scholar] [CrossRef] [Green Version]
  7. Armstrong, D.G.; Holtz-Neiderer, K.; Wendel, C.; Mohler, M.J.; Kimbriel, H.R.; Lavery, L.A. Skin temperature monitoring reduces the risk for diabetic foot ulceration in high-risk patients. Am. J. Med. 2007, 120, 1042–1046. [Google Scholar] [CrossRef]
  8. Bagavathiappan, S.; Philip, J.; Jayakumar, T.; Raj, B.; Rao, P.N.S.; Varalakshmi, M.; Mohan, V. Correlation between plantar foot temperature and diabetic neuropathy: a case study by using an infrared thermal imaging technique. J. Diabetes Sci. Technol. 2010, 4, 1386–1392. [Google Scholar] [CrossRef] [Green Version]
  9. Roback, K.; Johansson, M.; Starkhammar, A. Feasibility of a thermographic method for early detection of foot disorders in diabetes. Diabetes Technol. Therap. 2009, 11, 663–667. [Google Scholar] [CrossRef] [Green Version]
  10. Lavery, L.A. and Higgins, K.R. and Lanctot, D.R.and Constantinides, G.P. and Zamorano, R.G.; Athanasiou, K.A.; Armstrong, D.G.; Agrawal, C.M. Preventing diabetic foot ulcer recurrence in high-risk patients use of temperature monitoring as a selfassessment tool. Diabetes Care 2007, 30, 14–20. [Google Scholar] [CrossRef] [Green Version]
  11. Chan, A.W.; MacFarlane, I.A.; Bowsher, D.R. Contact thermography of painful diabetic neuropathic foot. Diabetes Care 1991, 14, 918–922. [Google Scholar] [CrossRef]
  12. Nagase, T.; Sanada, H.; Takehara, K.; Oe, M.; Iizaka, S.; Ohashi, Y.; Oba, M.; Kadowaki, T.; Nakagami, G. Variations of plantar thermographic patterns in normal controls and non-ulcer diabetic patients: novel classification using angiosome concept. J. Plast. Reconstr. Aesthet. Surg. 2011, 64, 860–866. [Google Scholar] [CrossRef] [PubMed]
  13. Mori, T.; Nagase, T.; Takehara, K.; Oe, M.; Ohashi, Y.; Amemiya, A.; Noguchi, H.; Ueki, K.; Kadowaki, T.; Sanada, H. Morphological pattern classification system for plantar thermography of patients with diabetes. J. Diabetes Sci. Technol. 2013, 7, 1102–1112. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Jones, B.F. A reappraisal of the use of infrared thermal image analysis in medicine. IEEE Trans. Med. Imag. 1998, 17, 1019–1027. [Google Scholar] [CrossRef] [PubMed]
  15. Kaabouch, N.; Chen, Y.; Anderson, J.; Ames, F.; Paulson, R. Asymmetry analysis based on genetic algorithms for the prediction of foot ulcers. In Proceedings of the IS&T/SPIE Electronic Imaging, Visualization and Data Analysis, San Jose, CA, USA, 18–22 January 2009; p. 724304. [Google Scholar]
  16. Kaabouch, N.; Chen, Y.; Hu, W.-C.; Anderson, J.W.; Ames, F.; Paulson, R. Enhancement of the asymmetry-based overlapping analysis rough features extraction. J. Electron. Imag. 2011, 20, 013012. [Google Scholar] [CrossRef]
  17. Liu, C.; van Netten, J.J.; van Baal, J.G.; Bus, S.A.; van der Heijden, F. Automatic detection of diabetic foot complications with infrared thermography by asymmetric analysis. J. Biomed. Opt. 2015, 20, 026003. [Google Scholar] [CrossRef] [Green Version]
  18. Hernandez-Contreras, D.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.; Ramirez-Cortes, J.; Renero-Carrillo, F. Automatic classification of thermal patterns in diabetic foot based on morphological pattern spectrum. Infrared Phys. Technol. 2015, 73, 149–157. [Google Scholar] [CrossRef]
  19. Hernandez-Contreras, D.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.; Gonzalez-Bernal, J.A.; Altamirano-Robles, L. A quantitative index for classification of plantar thermal changes in the diabetic foot. Infrared Phys. Technol. 2017, 81, 242–249. [Google Scholar] [CrossRef]
  20. Hernandez-Contreras, D.A.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.; Orihuela-Espina, F. Statistical Approximation of Plantar Temperature Distribution on Diabetic Subjects Based on Beta Mixture Model. IEEE Access 2019, 7, 28383–28391. [Google Scholar] [CrossRef]
  21. Suissa, S.; Ernst, P. Optical illusions from visual data analysis: example of the new zealand asthma mortality epidemic. J. Chronic Dis. 1997, 50, 1079–1088. [Google Scholar] [CrossRef]
  22. Faust, O.; Acharya, U.R.; Ng, E.; Hong, T.J.; Yu, W. Application of infrared thermography in computer aided diagnosis. Infrared Phys. Technol. 2014, 66, 160–175. [Google Scholar] [CrossRef]
  23. Kamavisdar, P.; Saluja, S.; Agrawal, S. A survey on image classification approaches and techniques. Int. J. Adv. Res. Com. Commun. Eng. 2013, 2, 1005–1009. [Google Scholar]
  24. Ren, J. ANN vs. SVM: Which one performs better in classification of MCCs in mammogram imaging. Knowl-Based Syst. 2012, 26, 144–153. [Google Scholar] [CrossRef] [Green Version]
  25. Lu, D.; Weng, Q. A survey of image classification methods and techniques for improving classification performance. Int. J. Remote Sens. 2007, 28, 823–870. [Google Scholar] [CrossRef]
  26. Cruz-Vega, I.; Peregrina-Barreto, H.; de Jesus Rangel-Magdaleno, J.; Manuel Ramirez-Cortes, J. A comparison of intelligent classifiers of thermal patterns in diabetic foot. In Proceedings of the 2019 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Auckland, New Zealand, 20–23 May 2019; pp. 1–6. [Google Scholar]
  27. Adam, M.; Ng, E.Y.; Tan, J.H.; Heng, M.L.; Tong, J.W.; Acharya, U.R. Computer aided diagnosis of diabetic foot using infrared thermography: A review. Comput. Biol. Med. 2017, 91, 326–336. [Google Scholar] [CrossRef]
  28. Fraiwan, L.; AlKhodari, M.; Ninan, J.; Mustafa, B.; Saleh, A.; Ghazal, M. Diabetic foot ulcer mobile detection system using smart phone thermal camera: a feasibility study. Biomed. Eng. online 2017, 16, 117. [Google Scholar] [CrossRef] [Green Version]
  29. Karan, O.; Bayraktar, C.; Gümüşkaya, H.; Karlık, B. Diagnosing diabetes using neural networks on small mobile devices. Expert Syst. Appl. 2012, 39, 54–60. [Google Scholar] [CrossRef]
  30. Liu, C.; van der Heijden, F.; Klein, M.E.; van Baal, J.G.; Bus, S.A.; van Netten, J.J. Infrared dermal thermography on diabetic feet soles to predict ulcerations: a case study. In Proceedings of the SPIEBiOS, Advanced Biomedical and Clinical Diagnostic Systems XI, San Francisco, CA, USA, 2–7 February 2013. [Google Scholar]
  31. Nandagopan, G.L.; Bhargavi, A.B. Implementation and Comparison of Two Image Segmentation Techniques on Thermal Foot Images and Detection of Ulceration Using Asymmetry Presented at ICCSP 2016. Available online: https://ieeexplore.ieee.org/abstract/document/7754155 (accessed on 15 March 2020).
  32. Etehadtavakol, M.; Ng, E.; Kaabouch, N. Automatic segmentation of thermal images of diabetic-at-risk feet using the snakes algorithm. Infrared Phys. Technol. 2017, 86, 66–76. [Google Scholar] [CrossRef]
  33. Asma, B.; Harba, R.; Canals, R.; Ledee, R.; Jabloun, R. A joint snake and atlas-based segmentation of plantar foot thermal images. In Proceedings of the 2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA), Montreal, QC, Canada, 28 November–1 December 2017. [Google Scholar]
  34. Zhang, K.; Zhang, L.; Song, H.; Zhou, W. Active contours with selective local or global segmentation: A new formulation and level set method. Image Vision Comput. 2010, 28, 668–676. [Google Scholar] [CrossRef]
  35. Dong, F.; Chen, Z.; Wang, J. A new level set method for inhomogeneous image segmentation. Image Vision Comput. 2013, 31, 809–822. [Google Scholar] [CrossRef]
  36. Li, B.N.; Qin, J.; Wang, R.; Wang, M.; Li, X. Selective level set segmentation using fuzzy region competition. IEEE Access 2016, 4, 4777–4788. [Google Scholar] [CrossRef]
  37. Salah, M.B.; Mitiche, A.; Ismail, I.B. Multiregion image segmentation by parametric kernel graph cuts. IEEE T. Image Process. 2010, 20, 545–557. [Google Scholar] [CrossRef] [PubMed]
  38. Etehadtavakol, M.; Emrani, Z.; Ng, E.Y.K. Rapid extraction of the hottest or coldest regions of medical thermographic images. Med. Biol. Eng. Comput. 2019, 57, 379–388. [Google Scholar] [CrossRef] [PubMed]
  39. Gururajarao, S.B.; Venkatappa, U.; Shivaram, J.M.; Sikkandar, M.Y.; Amoudi, A.A. Infrared Thermography and Soft Computing for Diabetic Foot Assessment. Mach. Learn. Bio-Signal Anal. Diagn. Imaging 2019, 73–97. [Google Scholar] [CrossRef]
  40. Adam, M.; Ng, E.Y.K.; Oh, S.L.; Heng, M.L.; Hagiwara, Y.; Tan, J.H.; Tong, J.W.K.; Acharya, U.R. Automated characterization of diabetic foot using nonlinear features extracted from thermograms. Infrared Phys. Technol. 2018, 89, 325–337. [Google Scholar] [CrossRef]
  41. Adam, M.; Ng, E.Y.K.; Oh, S.L.; Heng, M.L.; Hagiwara, Y.; Tan, J.H.; Tong, J.W.K.; Acharya, U.R. Automated detection of diabetic foot with and without neuropathy using double density-dual tree-complex wavelet transform on foot thermograms. Infrared Phys. Technol. 2018, 92, 270–279. [Google Scholar] [CrossRef]
  42. Saminathan, J.; Sasikala, M.; Narayanamurthy, V.B.; Rajesh, K.; Arvind, R. Computer aided detection of diabetic foot ulcer using asymmetry analysis of texture and temperature features. Infrared Phys. Technol. 2020, 105, 103219. [Google Scholar] [CrossRef]
  43. Maldonado, H.; Bayareh, R.; Torres, I.A.; Vera, A.; Gutiérrez, J.; Leija, L. Automatic detection of risk zones in diabetic foot soles by processing thermographic images taken in an uncontrolled environment. Infrared Phys. Technol. 2020, 105, 103187. [Google Scholar] [CrossRef]
  44. Kavakiotis, I.; Tsave, O.; Salifoglou, A.; Maglaveras, N.; Vlahavas, I.; Chouvarda, I. Machine learning and data mining methods in diabetes research. Comput. Struct. Biotechnol. J. 2017, 15, 104–116. [Google Scholar] [CrossRef]
  45. Temurtas, H.; Yumusak, N.; Temurtas, F. A comparative study on diabetes disease diagnosis using neural networks. Expert Syts. Appl. 2009, 36, 8610–8615. [Google Scholar] [CrossRef]
  46. Ribeiro, Á.C.; Barros, A.K.; Santana, E.; Príncipe, J.C. Diabetes classification using a redundancy reduction preprocessor. Research on Biomed. Eng. 2015, 31, 97–106. [Google Scholar] [CrossRef] [Green Version]
  47. Rashid, T.A.; Abdulla, S.M.; Abdulla, R.M. Decision Support System for Diabetes Mellitus through Machine Learning Techniques. Available online: https://www.researchgate.net/profile/Tarik_Rashid/publication/305730658_Decision_Support_System_for_Diabetes_Mellitus_through_Machine_Learning_Techniques.pdf (accessed on 15 March 2020).
  48. Sethi, H.; Goraya, A.; Sharma, V. Artificial intelligence based ensemble model for diagnosis of diabetes. Int. J. Adv. Res. Com. Sci. 2017, 8, 1540. [Google Scholar]
  49. Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE T. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Litjens, G.; Kooi, T.B.; Bejnordi, E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Goyal, M.; Reeves, N.D.; Davison, A.K.; Rajbhandari, S.; Spragg, J.; Yap, M.H. Dfunet: Convolutional neural networks for diabetic foot ulcer classification. IEEE Trans. Emerging Topics Comput. Intell. 2018, 1–12. [Google Scholar] [CrossRef] [Green Version]
  52. Goyal, M.; Yap, M.H. Multi-class semantic segmentation of skin lesions via fully convolutional networks. arXiv 2017, arXiv:1711.10449. Available online: https://arxiv.org/abs/1711.10449 (accessed on 15 March 2020).
  53. Hernandez-Contreras, D.A.; Peregrina-Barreto, H.; de Jesus Rangel-Magdaleno, J.; Renero-Carrillo, F.J. Plantar Thermogram Database for the Study of Diabetic Foot Complications. IEEE Access 2019, 7, 161296–161307. [Google Scholar] [CrossRef]
  54. Suzuki, K. Computational Intelligence in Biomedical Imaging; Springer: Berlin, Germany, 2016. [Google Scholar]
  55. Taylor, G.I.; Palmer, J.H. The vascular territories (angiosomes) of the body: experimental study and clinical applications. Br. J. Plast. Surg. 1987, 40, 113–141. [Google Scholar] [CrossRef]
  56. Taha, A.A.; Hanbury, A. Metrics for evaluating 3D medical image segmentation: analysis, selection, and tool. BMC Med. Imaging 2015, 15, 29. [Google Scholar] [CrossRef] [Green Version]
  57. Russ, J.C. The Image Processing Handbook; CRC Press: Boca Raton, FL, USA, 2016. [Google Scholar]
  58. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems 25 (NIPS 2012), Lake Tahoe, Nevada, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  59. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  60. Balian, R. Entropy, a protean concept. In Poincaré Seminar 2003; Birkhauser Verlag: Basel, Spain, 2004; pp. 119–144. [Google Scholar]
  61. Yu, X.; Zhou, Z.; Gao, Q.; Li, D.; Ríha, K. Infrared image segmentation using growing immune field and clone threshold. Infrared Phys. Technol. 2018, 88, 184–193. [Google Scholar] [CrossRef]
  62. Sarkar, S.; Das, S.; Chaudhuri, S.S. Multilevel image thresholding based on tsallis entropy and differential evolution. In Proceedings of the International Conference on Swarm, Evolutionary, and Memetic Computing, Bhubaneswar, India, 20–22 December 2012; pp. 17–24. [Google Scholar]
  63. Sarkar, S.; Paul, S.; Burman, R.; Das, S.; Chaudhuri, S.S. A fuzzy entropy based multi-level image thresholding using differential evolution. In Proceedings of the International Conference on Swarm, Evolutionary, and Memetic Computing, Bhubaneswar, India, 18–20 December 2014; pp. 386–395. [Google Scholar]
  64. Eiben, A.E.; Smith, J.E. Introduction to Evolutionary Computing; Springer: Berlin, Germany, 2003. [Google Scholar]
  65. Sarkar, S.; Patra, G.R.; Das, S. A differential evolution based approach formultilevel image segmentation using minimum cross entropy thresholding. In Proceedings of the International Conference on Swarm, Evolutionary, and Memetic Computing, Visakhapatnam, India, 19–21 December 2011; pp. 51–58. [Google Scholar]
  66. Das, S.; Suganthan, P.N. Differential evolution: A survey of the state-of-the-art. IEEE T. Evolut. Comput. 2011, 15, 4–31. [Google Scholar] [CrossRef]
  67. Haykin, S. A comprehensive foundation. Neural Netw. 2004, 2, 41. [Google Scholar]
  68. Friedman, J.; Hastie, T.; Tibshirani, R. The Elements of Statistical Learning; Springer: New York, NY, USA, 2001. [Google Scholar]
  69. Gopinath, R.; Kumar, C.S.; Vishnuprasad, K.; Ramachandran, K. Feature mapping techniques for improving the performance of fault diagnosis of synchronous generator. Int. J. Progn. Health Manag. 2015, 6, 12. [Google Scholar]
  70. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  71. Liu, W.; Wang, Z.; Liu, X.; Zeng, N.; Liu, Y.; Alsaadi, F.E. A survey of deep neural network architectures and their applications. Neurocomputing 2017, 234, 11–26. [Google Scholar] [CrossRef]
Figure 1. Images of the five level grades of the thermograms.
Figure 1. Images of the five level grades of the thermograms.
Sensors 20 01762 g001
Figure 2. Automatic segmentation process [26].
Figure 2. Automatic segmentation process [26].
Sensors 20 01762 g002
Figure 3. Angiosomes suggested by Taylor and Palmer [55] for temperature analysis.
Figure 3. Angiosomes suggested by Taylor and Palmer [55] for temperature analysis.
Sensors 20 01762 g003
Figure 4. Images of the automatic segmentation process of one case of the DM group, including the RGB original image, the grayscale representation, and the obtained images from one to four thresholds [26].
Figure 4. Images of the automatic segmentation process of one case of the DM group, including the RGB original image, the grayscale representation, and the obtained images from one to four thresholds [26].
Sensors 20 01762 g004
Figure 5. Fuzzy sets of the segmentation process in a DM foot with the optimal value of two thresholds [26].
Figure 5. Fuzzy sets of the segmentation process in a DM foot with the optimal value of two thresholds [26].
Sensors 20 01762 g005
Figure 6. Fuzzy sets of the segmentation process in a CG foot with the optimal value of three thresholds [26].
Figure 6. Fuzzy sets of the segmentation process in a CG foot with the optimal value of three thresholds [26].
Sensors 20 01762 g006
Figure 7. ROC curves of the worst 5-level classification case (Classes 3 and 4).
Figure 7. ROC curves of the worst 5-level classification case (Classes 3 and 4).
Sensors 20 01762 g007
Table 1. Description of the DFTNet Architecture.
Table 1. Description of the DFTNet Architecture.
Layer No.Layer TypeFilter SizeStrideNo. of FiltersFC Units
Layer 1Conv.7 × 71 ×132-
Layer 2Max-Pool3×32×2--
Layer 3Conv.1×11×164-
Layer 4Conv.3×31×164
Layer 5Max-Pool3×32×2--
Layer 6Conv.3×31×132
Layer 7Max-Pool2×22×2--
Layer 8Conv.3×31×132-
Layer 9Full Conn.---No. Classes
Table 2. The performance measures of SVM (left) and ANN (right) in classification of the five classes.
Table 2. The performance measures of SVM (left) and ANN (right) in classification of the five classes.
Case1st Class2nd ClassSensitivitySpecificityPrecisionAccuracyF-measureAUC
1151.0000–0.91671.0000–0.91671.0000–0.91671.0000–0.91671.0000–0.91671.0000–0.9167
2251.0000–0.91671.0000–0.83331.0000–0.84621.0000–0.87501.0000–0.88001.0000–0.8750
3140.9167–0.66671.0000–1.00001.0000–1.00000.9583–0.83330.9565–0.80000.9167–0.8333
4351.0000–1.00000.9167–0.58330.9231–0.70590.9583–0.79170.9600–0.82760.9826–0.7917
5240.9167–0.66670.5000–0.91670.6471–0.88890.7083–0.79170.7586–0.76190.9167–0.7917
6131.0000–0.50001.0000–1.00001.0000–1.00001.0000–0.75001.0000–0.66671.0000–0.7500
7451.0000–1.00000.9167–0.41670.9231–0.63160.9583–0.70830.9600–0.77421.0000–0.7083
8340.2500–0.66670.5833–1.00000.3750–1.00000.4167–0.83330.3000–0.80000.3194–0.8333
9231.0000–0.83331.0000–1.00001.0000–1.00001.0000–0.91671.0000–0.90911.0000–0.9167
10120.8333–0.50000.5833–1.00000.6667–1.00000.7083–0.75000.7407–0.66670.7639–0.7500
Average 0.8917–0.76670.8500–0.86670.8535–0.89890.8708–0.81670.8676–0.80030.8899–0.8167
Table 3. The performance measures of AlexNet (left) and GoogleNet (right) in classification of the five classes.
Table 3. The performance measures of AlexNet (left) and GoogleNet (right) in classification of the five classes.
Case1st Class2nd ClassSensitivitySpecificityPrecisionAccuracyF-measureAUC
1150.9545–0.90911.0000–1.00001.0000–1.00000.9783–0.95650.9767–0.95240.9773–0.9545
2251.0000–0.95830.9583–0.95830.9600–0.95830.9792–0.95830.9796–0.95830.9792–0.9583
3140.7727–0.77271.0000–0.95831.0000–0.94440.8913–0.86960.8718–0.85000.8864–0.8655
4350.9583–0.91670.7917–0.79170.8214–0.81480.8750–0.85420.8846–0.86270.8750–0.8542
5240.9167–0.79170.8333–0.91670.8462–0.90480.8750–0.85420.8800–0.84440.8750–0.8542
6130.9545–0.90910.5417–1.00000.6563–1.00000.7391–0.95650.7778–0.95240.7481–0.9545
7450.7500–0.83330.8750–0.75000.8571–0.76920.8125–0.79170.8000–0.80000.8125–0.7917
8340.5000–0.58330.5417–0.62500.5217–0.60870.5208–0.60420.5106–0.59570.5208–0.6042
9230.8333–0.45830.6250–0.95830.6897–0.91670.7292–0.70830.7547–0.61110.7292–0.7083
10120.8182–0.31820.6250–0.91670.6667–0.77780.7174–0.63040.7347–0.45160.7216–0.6174
Average 0.8458–0.74510.7792–0.88750.8019–0.86950.8118–0.81840.8171–0.78790.8125–0.8163
Table 4. The performance measures of our proposed DFTNet structure classifying the five classes.
Table 4. The performance measures of our proposed DFTNet structure classifying the five classes.
CaseClass 1Class 2SensitivitySpecificityPrecisionAccuracyF-measureAUC
115111111
225111111
31410.95830.95650.97830.97780.9792
4350.83330.95830.95240.89580.88890.8958
524111111
6130.95450.95830.95450.95650.95450.9564
7450.95830.95830.95830.95830.95830.9583
8340.91670.750.78570.83330.84620.8533
9230.91670.8750.880.89580.8980.8958
10120.95450.91670.9130.93480.93330.9356
Average 0.95340.93750.94010.94530.94570.9455

Share and Cite

MDPI and ACS Style

Cruz-Vega, I.; Hernandez-Contreras, D.; Peregrina-Barreto, H.; Rangel-Magdaleno, J.d.J.; Ramirez-Cortes, J.M. Deep Learning Classification for Diabetic Foot Thermograms. Sensors 2020, 20, 1762. https://doi.org/10.3390/s20061762

AMA Style

Cruz-Vega I, Hernandez-Contreras D, Peregrina-Barreto H, Rangel-Magdaleno JdJ, Ramirez-Cortes JM. Deep Learning Classification for Diabetic Foot Thermograms. Sensors. 2020; 20(6):1762. https://doi.org/10.3390/s20061762

Chicago/Turabian Style

Cruz-Vega, Israel, Daniel Hernandez-Contreras, Hayde Peregrina-Barreto, Jose de Jesus Rangel-Magdaleno, and Juan Manuel Ramirez-Cortes. 2020. "Deep Learning Classification for Diabetic Foot Thermograms" Sensors 20, no. 6: 1762. https://doi.org/10.3390/s20061762

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop