Next Article in Journal
Extended Online DMD and Weighted Modifications for Streaming Data Analysis
Next Article in Special Issue
Machine Learning in X-ray Diagnosis for Oral Health: A Review of Recent Progress
Previous Article in Journal
Mathematical Modeling of Multi-Phase Filtration in a Deformable Porous Medium
Previous Article in Special Issue
Survey of Recent Deep Neural Networks with Strong Annotated Supervision in Histopathology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feet Segmentation for Regional Analgesia Monitoring Using Convolutional RFF and Layer-Wise Weighted CAM Interpretability

by
Juan Carlos Aguirre-Arango
*,
Andrés Marino Álvarez-Meza
and
German Castellanos-Dominguez
Signal Processing and Recognition Group, Universidad Nacional de Colombia, Manizales 170003, Colombia
*
Author to whom correspondence should be addressed.
Computation 2023, 11(6), 113; https://doi.org/10.3390/computation11060113
Submission received: 9 May 2023 / Revised: 28 May 2023 / Accepted: 31 May 2023 / Published: 8 June 2023
(This article belongs to the Special Issue Computational Medical Image Analysis)

Abstract

:
Regional neuraxial analgesia for pain relief during labor is a universally accepted, safe, and effective procedure involving administering medication into the epidural. Still, an adequate assessment requires continuous patient monitoring after catheter placement. This research introduces a cutting-edge semantic thermal image segmentation method emphasizing superior interpretability for regional neuraxial analgesia monitoring. Namely, we propose a novel Convolutional Random Fourier Features-based approach, termed CRFFg, and custom-designed layer-wise weighted class-activation maps created explicitly for foot segmentation. Our method aims to enhance three well-known semantic segmentation (FCN, UNet, and ResUNet). We have rigorously evaluated our methodology on a challenging dataset of foot thermal images from pregnant women who underwent epidural anesthesia. Its limited size and significant variability distinguish this dataset. Furthermore, our validation results indicate that our proposed methodology not only delivers competitive results in foot segmentation but also significantly improves the explainability of the process.

1. Introduction

The use of regional neuraxial analgesia for pain relief during labor is widely acknowledged as a safe method [1]. It involves the administration of medication into the epidural or subarachnoid space in the lower back. This procedure blocks pain signals from the uterus and cervix to the brain. This method is considered safe and effective for most women and is associated with lower rates of complications than other forms of pain relief [1,2]. Electrophysiological testing measures nerve fiber reactions to painful stimuli with electromyography, excitatory or inhibitory reflexes, evoked potentials, electroencephalography, and magnetoencephalography [3]. In addition, imaging techniques objectively measure relevant bodily function patterns (such as blood flow, oxygen use, and sugar metabolism) using positron emission tomography (PET), single-photon emission computed tomography (SPECT), and functional magnetic resonance imaging (fMRI) [4].
Nonetheless, imaging techniques can be costly and are generally prohibited in obstetric patients, limiting their use. A cost-effective alternative approach is utilizing thermographic skin images to measure body temperature and predict the distribution and efficacy of epidural anesthesia [5]. This approach is achieved by identifying areas of cold sensation [5]. The use of thermal imaging provides an objective and non-invasive solution to assess warm modifications resulting from blood flow redistribution after catheter placement [6]. However, an adequate assessment requires temperature measurements from the patient’s foot soles at various times after catheter placement to accurately characterize early thermal modifications [7,8]. Regarding this, semantic segmentation of feet in infrared thermal images in obstetric environments is challenging due to various factors. Firstly, thermal images possess inherent characteristics such as low contrast, blurred edges, and uneven intensity distribution, making it difficult to identify objects accurately [9,10]. The second challenge is the high variability of foot position in clinical settings. Additionally, the specialized equipment required for collecting these images and the limited willingness of mothers to participate in research studies resulted in a need for more available samples and the challenge of acquiring annotated data, which is crucial for developing effective segmentation techniques.
Semantic segmentation is crucial in medical image analysis, with deep learning widely used. Fully Convolutional Networks (FCN) [11] is a popular approach that uses Convolutional layers for pixel-wise classification but produces coarse Region of Interest (ROI) and poor boundary definitions for medical images [12]. Likewise, U-Net [13] consists of encoders and decoders that handle objects of varying scales but have difficulty dealing with opaque or unclear goal masks [14]. U-Net++ [15] extends U-Net with nested skip connections for highly accurate segmentation but with increased complexity and overfitting risk. Besides, SegNet [16] is an encoder–decoder architecture that handles objects of different scales but cannot handle fine details. Mask R-CNN [17] extends Faster R-CNN [18] for instance segmentation with high accuracy but requires a large amount of training data and has high computational complexity. On the other hand, PSPNet uses a pyramid pooling module for multi-scale contextual information and increased accuracy but with high computational complexity and a tendency to produce fragmented segmentation maps for small objects [19].
Specifically for semantic segmentation of feet from infrared thermal images, most works were developed in the context of diabetic foot disorders. In [20], the authors combine RGB, infrared, and depth images to perform plantar foot segmentation based on a U-Net architecture together with RANdom SAmple Consensus (RANSAC) [21], which relies too much on depth information. The authors in [22] use a similar approach to integrating thermal and RGB images to be fed into a U-Net model. Their experiments show that RGB images help in more complex cases. In [23], the authors compare multiple models on thermal images, including U-Net, Segnet, FCN, and prior shape active contour-based methodology, proving Segnet outperforms them all. Similarily, in [24], the authors compare multiple infrared thermographic feet segmentation models using transfer learning and removal algorithms based on morphological operations on U-Net, FCN, and Segnet, showing that Segnet outperforms the rest of the models but with high computational cost.
On the other hand, Visual Transformers (VIT) [25] have revolutionized self-attention mechanisms to identify long-range image dependencies. Several recent works have leveraged VIT capabilities to enhance global image representation. For instance, in [26], a U-Net architecture fused with a VIT-based transformer significantly improves model performance. However, this approach requires a pre-trained model and many iterations. Similarly, in [27], a pure U-Net-like transformer is proposed to capture long-range dependencies. Another recent work [28] suggests parallel branches, one based on transformers to capture long-range dependencies and the other on CNN to conserve high resolution. The authors of [29] propose a squeeze-and-expansion transformer that combines local and global information to handle diverse representations effectively. This method has unlimited practical receptive fields, even at high feature resolutions. However, it relies on a large dataset and has higher computational costs than conventional methods. To address the data-hungry nature of transformer-based models, the work in [30] proposes a semi-supervised cross-teaching approach between CNN and Transformers. The most recent work in this field, Meta Segment Anything [31], relies on an extensive natural database (around 1B images) for general segmentation. However, medical and natural images have noticeable differences, including color and blurriness. It is also pertinent to note that accepting ambiguity can incorporate regions that may not be part of the regions of interest. Specifically, while transformers excel at capturing long-range dependencies, they still face challenges in scenarios where data is scarce [32].
Likewise, transfer learning-based strategies in medical image segmentation is a powerful technique that utilizes pre-trained models to enhance performance, minimize data requirements, and optimize computational resources [33]. Nevertheless, choosing an appropriate and representative pre-trained model is crucial to avoid suboptimal results and potential bias [34,35]. Nevertheless, in our study, we aim to assess the effectiveness of our proposal independently, thus excluding the use of transfer learning.
Here, we present a cutting-edge Convolutional Random Fourier Features (CRFFg) technique for foot segmentation in thermal images, leveraging layer-wise weighted class activation maps. Our proposed data-driven method is twofold. First, it integrates Random Fourier Features within a convolutional framework, enabling weight updates through gradient descent. To assess the efficacy of our approach, we benchmark it against three widely-used architectures: U-Net [13], FCN [11], and ResUNet [36]. We enhance these architectures by incorporating CRFFg at the skip connections, bolsters representation, and facilitate the fusion of low-level semantics from the decoder to the encoder. Second, we introduce a layer-wise strategy for quantitatively analyzing Class Activation Maps (CAMs) for semantic segmentation tasks [37]. Our experimental findings showcase the competitive performance of our models and the accurate quantitative assessment of CAMs. The proposed CRFFg method offers a promising solution for foot segmentation in thermal images, tailored explicitly for regional analgesia monitoring. Additionally, layer-wise weighted class activation maps contribute to a more comprehensive understanding of feature representations within neural networks.
The paper is organized as follows: Section 2 describes the materials and methods used in the study. Section 3 and Section 4 present the experimental setup and results, respectively, followed by Section 5, which provides the concluding remarks.

2. Material and Methods

2.1. Deep Learning for Semantic Segmentation

Provided an image set, { I n R H × W ˜ × C : n N } , we will call a label mask the corresponding matrix M n that encodes the membership of each n-th image pixel to a particular class, where H is height, W ˜ is width, and C holds the color channels of the image set. For simplicity, C = 1 is assumed. As regards the semantic segmentation task under consideration, each mask is binary, M { 0 , 1 } H × W ˜ , representing either the background or the foreground.
An estimate for matrix mask M ^ [ 0 , 1 ] H × W ˜ can be obtained through deep learning models for semantic segmentation, stacking convolutional layers as follows:
M ^ = ( φ L φ L ) ( I )
where φ l : R H l 1 × W ˜ l 1 × D l 1 R H l × W ˜ l × D l denotes a function composition for the l-th layer ( l L ), which comprises learnable parameters represented by W l R k ˜ l × k ˜ l × D l 1 × D l and b l R D l ( k ˜ l holds the l-th convolutional kernel size). Of note, the feature map F l = φ l F l 1 = ς l ( W l F l 1 + b l ) R H l × W ˜ l × D l , is comprised of D l distinct features extracted, ς l ( · ) is a nonlinear activiation function, and ⊗ stands for image-based convolution. Essentially, the function composition in Equation (1) transforms the input feature map from the previous layer, ( l 1 ) , into the output feature map for the current layer, l, by employing the learnable parameters W l and b l . The resulting F l captures the salient information within the l-th network layer.
The parameter set Θ = { W l , b l : l L } is estimated within the following optimizing framework [38]:
Θ * = arg min Θ E L { M n , M ^ n | Θ } : n N ,
where L : { 0 , 1 } H × W ˜ × [ 0 , 1 ] H × W ˜ R in Equation (2) is a given loss function and notation E { · } stands for the expectation operator.

2.2. Convolutional Random Fourier Features Gradient—CRFFg

Random Fourier Features establish a finite-dimensional, explicit mapping that approximates shift-invariant kernels k ( · ) as described in Rahimi et al. (2009) [39]. This explicit mapping, denoted by z : R Q ˜ R Q , serves to transform the input space into a finite-dimensional space H R Q , where the inner product can be obtained as:
k ( x x ) = ϕ ( x ) , ϕ ( x ) H z ( x ) z ( x ) .
The mapping z in Equation (3) is defined through Bochner’s theorem [40]:
k ( x x ) = R Q ˜ p ( ω ) exp ( i ω ( x x ) ) d ω = E ω exp ( i ω ( x x ) ) ,
where x , x R Q ˜ , p ( w ) is the probability density function of w R Q ˜ that defines the type of kernel. Specifically, the Gaussian kernel, favored for its universal approximating properties and mathematical tractability [41], is achieved from Equation (4) by setting p ( w ) = N ( 0 , σ 2 I ^ ) ; σ R + is a length-scale and I ^ is an identity matrix of proper size.
As both the kernel and the probability are real values, the imaginary component can be disregarded by employing the Euler equation. This leads to the use of a cosine function rather than an exponential, ensuring the following relationship:
z ( x ) = 2 Q cos ( ω 1 x + b 1 ) , , cos ( ω Q x + b Q ) ,
where ω q R Q ˜ , b q R , and q Q .
We aim to extend the kernel-based mapping depicted in Equation (5) for application to spatial data, such as images, by utilizing the power of convolutional operations. These operations have garnered significant attention for their efficacy in processing grid data [42]. Convolutional operations exhibit two crucial properties—translation equivariance and locality—that render them particularly suitable for handling spatial data [42]. In order to integrate these properties into the Random Fourier Features framework, we adapt the z mapping to operate within local regions of the grid input space. This results in the computation of the feature map F l R H l × W ˜ l × Q l , where the mapping is defined as z : R H l 1 × W ˜ l 1 × D l 1 R H l × W ˜ l × Q l , yielding:
F l = z ( F l 1 ) = cos W l Δ l F l 1 + b l ,
where Δ l R + is a scale parameter. The parameters W l R k ˜ l × k ˜ l × D l 1 × Q l and b l R Q l are initialized as in Equations (4) and (5), and updated through gradient descent under a back-propagation-based optimization of Equation (2) [38]. Consequently, we refer to the layers in Equation (6) as Convolutional Random Fourier Features Gradient (CRFFg).
The conceptual depiction of the proposed CRFFg layer is shown in Figure 1. Using this approach, we aim to integrate the advantageous attributes of kernel methods into a deep learning-based feature representation enhancement. In addition, using convolutions for local and equivariant representation of spatial data provides a robust and efficient strategy for image processing.

2.3. Layer-Wise Weighted Class Activation Maps for Semantic Segmentation

Class Activation Maps (CAMs) are a powerful tool to enhance the interpretability of outcomes derived from deep learning models. They achieve this by emphasizing the critical image regions in determining the model’s predicted output. To evaluate the contribution of these regions to a specific class r { 0 , 1 } , a linear combination of feature maps from a designated convolutional neural network layer l can be employed [37]. Here, given an input image I and a target class r, the salient input spatial information coded by the l-th layer into a trained deep learning semantic segmentation model with parameter set Θ * , as in Equation (2), is gathered through the Layer-CAM algorithm, yielding [43]:
S l r = ( Λ R e L U ) d D l α l r d F l r d
where S l r R H × W ˜ holds the Layer-CAM for class r at layer l, Λ : R H l × W ˜ l R H × W ˜ is the up-sampling operator, R e L U ( x ) = max ( 0 , x ) is the Rectified Linear activation function, and ⊙ stands for Hadamard product. Besides, F l r d R H l × W ˜ l collects the d-th feature map and α l r d R H l × W ˜ l is a weighting matrix holding elements:
α l r d [ i , j ] = R e L U y r / F l r d [ i , j ] ,
with α l r d [ i , j ] α l r d and F l r d [ i , j ] F l r d . y r is the score for class r that is computed using the approach in [44] adopted for the semantic segmentation tasks, as follows:
y r = E F ˜ L [ i , j ] : i , j | M [ i , j ] = r
where F ˜ L [ i , j ] F ˜ L holds the feature map elements for layer L in Equation (1) fixing a linear activation function.
As previously mentioned, the use of CAM-based representations enhances the explainability of deep learning models for segmentation tasks. To evaluate the interpretability of CAMs for a given model, we propose the following semantic segmentation measures, where higher scores indicate better interpretability:
CAM-based Cumulative Relevance ( ρ r ) : It involves computing the cumulative contribution from each CAM representation to detect class r within the segmented region of interest. This can be expressed as follows:
ρ r = E l E n { 1 ( M ˜ n r S n l r ) 1 1 S n l r 1 : n N } l L , ρ r [ 0 , 1 ] ,
where M ˜ n r { 0 , 1 } H × W ˜ collects a binary mask that identifies the pixel locations associated with the class r, and S n l r holds the Layer-CAM for image n with respect to layer l (see Equation (7)).
Mask-based Cumulative Relevance ( ϱ r ): It assesses the relevance averaged across the class pixel set related to the target mask of interest. Then, each class-based cumulative relevance is computed as follows:
ϱ r = E l E n { 1 ( M ˜ n r S n l r ) l l M ˜ n r l : n N } l L , ϱ r R + .
The normalized Mask-based Cumulative Relevance can be computed as:
ρ r = ρ r max r { 0 , 1 } ρ r , ρ r [ 0 , 1 ] .
CAM-Dice ( D ): A version of the Dice measure that quantifies mask thickness and how the extracted CAM is densely filled:
D r = E l E n { 2 l M ˜ n r S n l r l l M ˜ n r l + l S n l r l : n N } : l L , D r [ 0 , 1 ] .
The proposed measures enable the weighting of each layer’s contribution to a given class across the model by adjusting the normalization term related to the target mask, the estimated CAM, or both pixel-based salient activations. Figure 2 depicts a graphical representation of the proposed measures. The green circle represents the CAM generated for a specific region, as indicated by the white circle. These measures are designed to capture the relationship between the CAMs and the regions of interest. Furthermore, Figure 3 presents some exemplary scenarios. For instance, the ρ measure is associated with the proportion of the CAM inside the region of interest. On the other hand, ρ is based on the proportion of CAMs that, on average, belong to each pixel of the region of interest while maintaining the relationship between the classes (in this case, green for the foreground and red for the background). Additionally, D r follows a similar concept as the Dice coefficient used in segmentation, assessing the homogeneity of the intersection of the regions. In this case, we want to determine if the CAM is uniformly distributed.

2.4. Feet Segmentation Pipeline from Thermal Images

In a nutshell, the proposed methodology is evaluated using the pipeline shown in Figure 4, including the following testing stages:
(i)
Foot Infrared Thermal Data Acquisition and Preprocessing.
(ii)
Architecture Set-Up of tested Deep models for foot segmentation. Three DL architectures are contrasted using our CRFFg: U-Net, Fully Convolutional Network (FCN), and ResUNet.
(iii)
Assessment of semantic segmentation accuracy. In this study, we examine how data augmentation affects the performance of tested deep learning algorithms.
(iv)
Relevance-maps extraction from our Layer-Wise weighted CAMs to provide interpretability.

3. Experimental Set-Up

The proposed deep learning model for semantic segmentation enhances foot thermal images’ interpretability, achieving competitive segmentation performance. To this end, we evaluate the impact of incorporating a convolutional representation of CRFFg and layer-wise weighted CAM into three well-known deep-learning architectures.

3.1. Protocol for Infrared Thermal Data Acquisition: ThermalFeet Dataset

The protocol for data acquisition was designed by the physician staff at “SES Hospital Universitario de Caldas” to standardize the data collection of infrared thermal images acquired from pregnant women who underwent epidural anesthesia during labor. This protocol is in accordance with the occupational risks associated with assisting local anesthetics via epidural neuraxial as specified by the hospital’s administration, following previously implemented protocols [8,45,46,47,48].
Patient monitoring includes the necessary equipment for taking vital signs and a metal stretcher with foam cushion and plastic exterior covered only with a white sheet. The continuous monitoring device is placed 1.5 m from the stretcher in the same room, as shown in Figure 5. Before the epidural procedure, anesthesiologists assess each patient clinically and provide written and verbal information about the trial before obtaining her written consent. The patient’s body temperature, heart rate, oxygen saturation, and non-invasive blood pressure are monitored every five minutes. Skin temperature values are recorded during the procedure. Sensitivity responses are evaluated using superficial touch and cold tests with cotton wool soaked in water applied to the previously determined dermatomes. The temperature test records the verbal response as Yes or No for superficial touch and Cold or No Cold.
The protocol timeline for acquiring infrared thermal images is as follows: Initially, the woman is asked to be in a supine position before the first thermal image (T0) is captured once the first dose of the analgesic mixture is administered. A single thermal picture is taken at the placement of the operated catheter (0.45 mm; Perifix, Braun®, Kronberg, Germany) positioned within the space selected for injecting epidural anesthesia in the cervical region (at L2 to L3 or L3 to L4), measuring a few millimeters.
Within the next 25 min, one thermographic recording of the lower extremity is taken every five minutes (T1–T5). The catheter remains in the epidural space taped to the skin so that one image is captured every five minutes until six pictures have been collected. Though the clinical protocol demands images of both feet taken in a fixed corporal position, this condition is barely achievable due to the difficulty of labor procedures and contractions.
The data was collected under two different hardware specifications: (i) A set of 196 images captured from 22 pregnant women during labour using a FLIR A320 infrared camera with a resolution of 640 × 480 and a spectral range within 7.5 to 13 μ m. (ii) A set of 128 images with improved sensitivity and flexibility taken using a FLIR E95 thermal camera, having a resolution of 640 × 480 and spectral range within 7.5 to 14 μ m. In this study, 166 thermal images are selected from both sets as fulfilling the quality criteria of validation, as detailed in [24]. An anesthesiologist manually segmented the region of interest. The dataset is publicly available at https://gcpds-image-segmentation.readthedocs.io/en/latest/notebooks/02-datasets.html (accessed on 5 April 2023).

3.2. Set-Up of Compared Deep Learning Architectures

The following deep learning architectures are contrasted and enhanced using our CRFFg approach:
Fully Convolutional Network (FCN) [11]: This architecture is based on the VGG (Very Deep Convolutional Network) [49] model to recognize large-scale images. By using only convolutional layers, FCN models can deliver a segmentation map with pixel-level accuracy while reducing the computational burden.
U-Net [13]: This architecture unfolds into two parts: The encoder consists of convolutional layers to reduce the spatial image dimensions. The decoder holds layers to upsample the encoded features back to the original image size.
ResUNet [36]: This model extends the U-Net architecture by incorporating residual connections to improve performance. Deep learning training is improved by residual connections, which allow gradients to flow directly through the network.
Figure 6 presents the mentioned architectures, illustrating their unique layers, blocks, and the dimensions and filters associated. Different colors represent the different blocks or layers, and the spatial dimension of each level is also indicated. We estimate the effectiveness of incorporating the CRFFg layer for comparison purposes in FCN, U-Net, and ResUNet architectures. However, each evaluated CRFFg layer arrangement differs from another in the semantic segmentation features that feed the decoder, as detailed in [50,51,52]. Then, the CRFFg layer is placed at skip connections to enhance the feature fusion between encoders and decoders.
To evaluate the performance difference with the proposed CRFFg-layer strategy, we utilize a standard convolutional layer featuring an equal number of filters and a ReLU activation function at the same position within the architecture. In particular, we analyze the influence of the CRFFg layer dimension on segmentation performance, testing two multiplication values (one and three). Besides, to study the impact of CRFFg, we set the hyperparameters of all models variation from FCN, U-Net, and ResUNet architectures the same. The number of epochs is 200, and the batch size is 16. Additionally, the scale value of the CRFFg, Δ , is set as described in the standard RFF’s Tensorflow implementation for simplicity. Regarding the weights, they are trained using gradient descent with backpropagation. The selected optimizer is Adam due to its faster convergence, adaptive learning rate, reduced sensitivity to hyperparameters, and combining benefits of convex optimization [53]. The learning rate is initialized as 1 e 3 , and a dice-based loss is employed in Equation (2), as follows:
L D i c e ( M n , M ^ n ) = 2 1 ( M n M ^ n ) 1 + ϵ 1 M n 1 + 1 M ^ n 1 + ϵ ,
where ϵ = 1 avoids numerical instability. All experiments are carried out in Python 3.8, with the Tensorflow 2.4.1 API, on a Google Colaboratory environment (code repository: https://github.com/aguirrejuan/Foot-segmentation-CRFFg, accessed on 25 April 2023).

3.3. Training Details and Quantitative Assessment

With the aim to prevent overfitting and improve the generalization of trained models, the data augmentation procedure is performed on each image with horizontal flip enabled since feet are mostly symmetrical on the horizontal axis, specifically left-right and right-left on each foot. Hence, vertical overturn is disabled to prevent unrealistic upside-down foot representations. In the augmentation procedure, the images are rotated seven times within a range of −15 to 15 degrees, translated by 10% right to left, and zoomed in and out by 15%, as described in [20].
Moreover, the following metrics are used to measure segmentation performance [54]:
D = 2 | M M ^ | | M | + | M ^ | = 2 T P 2 T P + F P + F N
J = | M M ^ | | M M ^ | = T P F N + F P + T P
S e = T P T P + F N
S p = T N T N + F P
where T P , F N , and F P represent the true positive, false negative, and false positive predictions, respectively, for comparing the actual and estimated label masks M n and M ^ n for a given input image I n . In addition, the introduced layer-wise, weighted CAM-based interpretability measures are computed for CAM-Dice, CAM-based Cumulative Relevance, and Mask-based Cumulative Relevance (see Equations (12) and (13)).
As for the validation strategy, we selected the hold-out cross-validation strategy with the following partitions: 80% of the samples for training, 10% for validation, and 10% for testing.

4. Results and Discussion

4.1. Visual Inspection Results

Figure 7 shows results obtained from thermalFeet database without data augmentation, where each row represents a different architecture: FCN in the first row, U-Net in the second row, and ResUNet in the third row. As expected, the performance of the models under a small-size dataset is poor. The regions of faster change in temperature, which characterize the dataset, are where the models struggle more. At first glance, we observe that the FCN architecture is the one that struggles the most, having high false positives regions in regions that exhibit low-high temperatures.
Figure 8 shows results obtained incorporating data augmentation. The positive impact of the data augmentation on the resulting segmentation of all the models is visible. Moreover, FCN architectures produce smoother borders and fewer false positives than other architectures. This can be explained due to the high receptive field that possesses the FCN architecture, allowing it to capture complex and heterogeneous regions (the variability of the temperatures) that compose the feet.
Notably, when comparing FCN models with a multiplication factor of 1 (M1), the model with our CRFFg (blue) generally outperforms in terms of pixel membership prediction (sensitivity). However, this trend only holds when the multiplication factor is increased to 3 (M3), probably because the large model is a propensity to overfit, making the prediction less confident in new data points. On the other hand, U-Net models blunder with regions that exhibit fast temperature changes. The same characteristic the FCN possesses can explain this, but the U-Net does not have a high receptive field that allows it to characterize high heterogeneous feet. As a result, among the U-Net approaches, U-Net CRFFg S-M1 performs satisfactorily with low false positives and high false negatives. At the same time, its direct competitor, U-Net S-M1, shows the opposite trend. Similarly, using CRFFg in the other U-Net alternatives reduces the number of false positives. Finally, the ResUNet architecture has the same behavior as the U-Net but with smoother borders, which can be explained due to the multiple stack layers at the ResBlock, which increase multiple steps of representation, allowing to capture of helpful representation. The ResUNet S-M1 works better on average; adding layers at the skip connections appears to reduce performance, creating false positives and false negatives. The latter can be explained due to the small size of the dataset. Specifically, using CRFFg with ResUNet does not result in noteworthy improvements.

4.2. Method Comparison Results of Semantic Segmentation Performance

Figure 9 illustrates the learning curves, e.g., training loss vs. epochs, of the compared models. Upon visual inspection, notable differences between the curves with and without data augmentation can be observed. When data augmentation is not applied, the algorithms exhibit higher validation loss in the initial 40 epochs. Regardless, they subsequently demonstrate a downward trend in validation loss. It is essential to mention that the learning curves exhibit increased noise, likely due to the limited size of the dataset. The limited dataset challenges the models to capture generalized features early in training. Moreover, in the validation partition without data augmentation, some models display a phenomenon known as double descent [55], where layers at different locations in the networks may learn at different rates [56]. In contrast, the training and validation losses consistently decrease in the data augmentation scenario, albeit with minor noise in the validation partition.
It is worth noting that both the FCN CRFFg S-M3 and, to a lesser extent, the FCN CRFFg S-M1 tend to exhibit faster decreases in validation loss during early iterations. This conduct can be attributed to the generalization capabilities of the RFF from kernel methods. On the other hand, in the ResUNet architectures, although it needs to be clarified, the ResUNet S-M3 tends to experience an early decline, even though it also reaches its minimum early, which is not the minimum among the approaches. Conversely, no apparent differences are observed within the U-Net architectures. Notably, the models in the data augmentation scenario are similar.
In turn, Figure 10a displays the values of semantic segmentation performance for thermalFeet dataset achieved by each compared deep learning architecture: FCN (colored in blue), ResNet (red), U-Net (green). For interpretation purposes, the results are presented for the evaluation measures separately. As seen, the specificity estimates are very close to the maximal value and show the lowest variability. This result can be explained by the relatively small feet sizes compared with the background, making their correct detection and segmentation more difficult. On the contrary, sensitivity assessments are of less value and have much more variability, accounting for the diversity in the regions of interest (i.e., size, shape, and location). Due to the changing behavior of thermal patterns and the limited datasets available, learners have difficulty obtaining an accurate model.
Regarding overlapping between estimated thermal masks, the Dice value is acceptable but with higher variance values for FCN, implying that other tested models segment complex shapes more accurately. As expected, the Jaccard index mean values resemble the Dice assessments, although with increased variance, which highlights the mismatch between the ground truth and the predicted mask even more.
A comparison between the segmentation metric value achieved by the baseline architecture (without any modifications) and the value estimated for every evaluated semantic segmentation strategy is presented in Figure 10b. Note that specificity is removed because its estimates are obtained with minimal variations.
As seen, the performance improvement depends on the learner model size (also called algorithm complexity). Namely, the baseline architecture of FCN holds 1,197,375 parameters, baseline ResUnet— 643,549, and baseline Unet—494,093. Thus, the FCN model contains the largest tuning parameter set and achieves the poorest performance, but it benefits the most from the evaluated architectures. As data augmentation is also applied, this finding becomes more evident. It may be pointed out that adding new data decreases model overfitting inherent to massive model sizes. Likewise, the following ResUnet model takes advantage of the enhanced architecture strategy using our CRFFg and improves performance. It increases more by generating new data points, however, to a lesser extent. Lastly, the learner with the lowest parameter set gets almost no benefits or is negatively affected by the strategies considered for architecture enhancement. Still, the strategies taken into account combined with expanded training data sizes can be improved, though very modestly. See Table A1, Appendix A, for the detailed segmentation performance results concerning the studied approaches.

4.3. Results of Assessing the Proposed CAM-Based Relevance Analysis Measures

We aim to evaluate the tested deep learning models for assessing the contribution of CAM-based representations to interpretability. To this end, we plot the pairwise relationship between the essential explanation elements (background and foreground) and the above-proposed measure for assessing the CAM-based relevance of performed image segmentation masks. Figure 11 displays the scatter plots obtained by each segmentation learner. CAMs extracted by the learner contribute more to the interpretability of regions of interest if the measure value tends toward the top-right corner. Moreover, we focus on the contribution of CAM representations to segmenting between background and foreground, utilizing the patient’s feet as critical identification features.
The findings from the modified CAM-Dice results can be split into two groups (refer to the left plot in Figure 11). One group involves ResUnet and UNet architectures, and the other showcases the better performance, featuring FCN architectures. It is also important to mention that the data augmentation strategy does not significantly boost interpretability as much as it enhances segmentation performance measures. Looking at the CAM-based Cumulative Relevance (refer to the middle plot in Figure 11), it is apparent that models with refined representations at skip connections surpass the baseline models. Even though there is no substantial difference between models with these enhancements, most models are situated in the top-right corner. This position suggests that the primary relevance is focused on the area of interest. Significantly, relevance seems to accumulate more in the background than in the foreground, which is logical, considering the relative sizes of both areas. In Figure 11, the Mask-based Cumulative Relevance plot on the right side demonstrates that most models tend to exhibit high-foreground-low-background relevance. This pattern leads to a bias favoring the foreground class, as reflected in the more robust activation of CAMs for the foreground class. However, it is interesting that models employing CRFFg perform better in separating classes situated towards the top-right corner, suggesting superior capabilities in differentiating foreground and background classes.
Figure 12 displays examples of CAMs extracted by the best models per architecture under the Mask-based Cumulative Relevance for feet (colored in green) and background (red color), respectively. As seen, the higher weight is located at the last part of the decoder, where the higher values of semantic information are found. Besides, the weights for the background class are also less than for the foreground class, showing that the models emphasize the latter while preserving the relevance weights for the former.
In particular, FCN CRFFg S-M3 is the best FCN model, as shown in Figure 12a, and extracts most of the weights in three layers (i.e., l3, l4, and l5), meaning that other layers do not contribute to the class foreground. On the other hand, this architecture leads to CAMs with lower values for background class (see examples on the right). This behavior can be explained because the FCN architecture holds an extensive receptive field. Hence, the FCN CRFFg S-M3 model enables capturing more global information crucial for segmentation and concentrating weights in a few layers.
In the case of ResUNet, ResUNet CRFFg S-M3 performs the most efficiently, as shown in Figure 12b. Since the receptive field decreases, the ResUNet architecture distributes the contribution more evenly among the extracted CAM representations. However, the more significant values remain in the l3, l4, and l5 layers. There is also activation of weights for the background class that can be explained, firstly, since the CRFFg configuration helps capture complex non-linear dependencies. Secondly, the local receptive field allows class separation.
Lastly, the CRFFg S-M3 model is the most effective for the U-Net architecture, with a performance similar to the outperforming ResUNet architecture, as shown in Figure 12c. However, several differences in the Fusion CAMs extracted by U-Net CRFFg S-M3 show high activation within the feet, suggesting that this model is not only sensitive to the foreground class. In addition, it captures more global features from feet.

5. Concluding Remarks

We introduce an innovative semantic segmentation approach that enhances interpretability by incorporating Convolutional Random Fourier Features and layer-wise weighted class activation maps. Our approach has been tested on a unique dataset of thermal foot images from pregnant women who have received epidural anesthesia, which is small but exhibits considerable variability. Besides, our strategy is two-pronged. Firstly, we introduce a novel Random Fourier Features layer, CRFFg, for handling image data, aiming to enhance three renowned architectures - FCN, UNet, and ResUNet. Secondly, we introduce three new quantitative measures to assess the interpretability of any deep learning model used for segmentation tasks. Our validation results indicate that the proposed approach boosts explainability and maintains competitive foot segmentation performance. In addition, the dataset used is tailored explicitly for epidural insertion during childbirth, reinforcing the practical relevance of our methodology.
There are, however, several observations worth mentioning:
Data acquisition tailored for Epidural. Epidural anesthesia involves the delivery of medicines that numb body parts to relieve pain, and the acquisition of data is usually performed under uncontrolled conditions with strong maternal artifacts. Moreover, it is impossible to fix a timeline for data collection. In addition, a timeline for gathering data cannot be set correctly. To the extent of our knowledge, this is the first time a protocol has been presented to regulate the data collection of infrared thermal images acquired from pregnant women who underwent epidural anesthesia during labor. As a result, data were assembled under real-world conditions that contained 196 thermal images fulfilling validation quality criteria.
Deep learning models for image semantic segmentation. Combined with machine learning, thermal imaging has proven helpful for performing semantic segmentation as a powerful method of dense prediction to adverse lighting conditions, providing better performance compared to their traditional counterparts. State-of-the-art medical image segmentation models include variants of U-Net models. A major reason for their success is that they employ skip connections, combining deep, semantic, and coarse-grained feature maps from the decoder subnetwork with shallow, low-level, fine-grained feature maps from the encoder subnetwork. They recover fine-grained details of target objects despite complex backgrounds [57]. Nevertheless, the collected image data from epidural anesthesia is insufficient for training the most commonly-known deep learners, which may result in overfitness to the training set. We address this issue by employing data augmentation addresses that artificially increase training data inputs to feed three tested architectures of deep learning models (FCN, U-Net, ResUNet), thus improving segmentation accuracy results. As seen in Figure 10b, the segmentation accuracy gain depends on the learner model complexity used: The fewer parameters the learner holds, the more the effectivity of data augmentation. Thus, the UNet learner with the lowest parameter set gets almost no benefit.
Strategies for enhancing the performance of deep learning-based segmentation. Three deep-learning architectures are explored to increase the interpretability of semantic segmentation results at competitive accuracy, ranked in decreased order of computational complexity as follows: FCN, ResUNet, and U-Net. Regarding the accuracy of semantic models, the data augmentation yields a sensibility metric value dependent on the model complexity: the more parameters the architecture holds, the higher the segmentation accuracy improvement. Thus, FCN benefits more from artificial data than ResUNet and U-Net. In the same way, both overlapping metrics (Jaccard and Dice) depend on the complexity of models. By contrast, the specificity reaches very high values regardless of trained deep learning because the background texture’s homogeneity saturates most captured thermal images. Nonetheless, the proposed modifications to architectures are not a solid argument for influencing their performed accuracy of semantic segmentation. In terms of enhancing explainability, the weak influence of data augmentation is the first finding to be drawn, as seen in the scatterplots of Figure 11. All tested models produce more significant CAM activations from layers with a wider receptive field. Moreover, the CRFFg layer also improves the representation of the foreground and background. It is also important to note the metrics developed for assessing the explainability of CAM representations, allowing scalability to larger image sets without visual inspection.
In terms of future research, the authors intend to integrate Vision Transformers and attention mechanisms for semantic segmentation into the CRFFg-based representation [58]. Besides, we propose to include variational autoencoders and transfer learning strategies within our framework to prevent overfitting and enhance data interpretability [33,59].

Author Contributions

Conceptualization, J.C.A.-A., A.M.Á.-M. and G.C.-D.; methodology, J.C.A.-A., A.M.Á.-M. and G.C.-D.; software, J.C.A.-A.; validation, J.C.A.-A., A.M.Á.-M. and G.C.-D.; formal analysis, J.C.A.-A. and G.C.-D.; investigation, J.C.A.-A., A.M.Á.-M. and G.C.-D.; resources, A.M.Á.-M. and G.C.-D.; data curation, J.C.A.-A.; writing–original draft preparation, J.C.A.-A., A.M.Á.-M. and G.C.-D.; writing–review and editing, A.M.Á.-M. and G.C.-D.; visualization, J.C.A.-A.; supervision, A.M.Á.-M. and G.C.-D.; project administration, A.M.Á.-M.; funding acquisition, A.M.Á.-M. and G.C.-D. All authors have read and agreed to the published version of the manuscript.

Funding

Under grants provided by the projects: “Prototipo de visión por computador para la identificación de problemas fitosanitarios en cultivos de plátano en el departamento de Caldas” (Hermes 51175) funded by Universidad Nacional de Colombia, and “Desarrollo de una herramienta de visión por computador para el análisis de plantas orientado al fortalecimiento de la seguridad alimentaria” (Hermes 54339) funded by Universidad Nacional de Colombia and Universidad de Caldas.

Institutional Review Board Statement

This study uses anonymized public datasets with institutional review board statement as presented in https://gcpds-image-segmentation.readthedocs.io/en/latest/notebooks/02-datasets.html (accessed on 5 April 2023).

Informed Consent Statement

This study uses anonymized public datasets as presented in https://gcpds-image-segmentation.readthedocs.io/en/latest/notebooks/02-datasets.html (accessed on 5 April 2023).

Data Availability Statement

Dataset is publicly available at: https://gcpds-image-segmentation.readthedocs.io/en/latest/notebooks/02-datasets.html (accessed on 5 April 2023).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Method Comparison from Absolute Semantic Segmentation Performances

Table A1 presents the absolute semantic segmentation results acquired from thermalFeet database. For clarity, the rank position of each method is also included. As can be seen, our enhancement based on CRFFg boosts the segmentation performance. Notably, ResUNet CRFFg S-M1 outperforms the tested approaches concerning the measured quantitative assessments.
Table A1. Absolute Semantic Segmentation Results on thermalFeet database. WODA: Without Data Augmentation, WDA: With Data Augmentation. The average ± standard deviation performance is displayed regarding the test partitions. M1 and M3 stand for CRFFg’s dimension as a multiplication factor of the enhanced layer’s size.
Table A1. Absolute Semantic Segmentation Results on thermalFeet database. WODA: Without Data Augmentation, WDA: With Data Augmentation. The average ± standard deviation performance is displayed regarding the test partitions. M1 and M3 stand for CRFFg’s dimension as a multiplication factor of the enhanced layer’s size.
ApproachMeasureWODARankWDARank
FCNDice0.9527 ± 0.02383.00.8646 ± 0.062410.0
Jaccard0.9106 ± 0.04243.00.7668 ± 0.096910.0
Sensitivity0.9352 ± 0.04824.00.8260 ± 0.10986.0
Specificity0.9857 ± 0.01057.00.9697 ± 0.018613.0
FCN CRFFg S-M1Dice0.9530 ± 0.02572.00.8510 ± 0.062312.0
Jaccard0.9113 ± 0.04562.00.7456 ± 0.091312.0
Sensitivity0.9424 ± 0.05263.00.8016 ± 0.099913.0
Specificity0.9810 ± 0.015812.00.9697 ± 0.023314.0
FCN CRFFg S-M3Dice0.9480 ± 0.02245.00.8346 ± 0.091615.0
Jaccard0.9021 ± 0.04035.00.7262 ± 0.128415.0
Sensitivity0.9340 ± 0.04236.00.7771 ± 0.132515.0
Specificity0.9804 ± 0.016813.00.9714 ± 0.024610.0
FCN S-M1Dice0.9469 ± 0.02736.00.8421 ± 0.087014.0
Jaccard0.9003 ± 0.04866.00.7367 ± 0.125414.0
Sensitivity0.9286 ± 0.05187.00.7867 ± 0.142214.0
Specificity0.9843 ± 0.01099.00.9714 ± 0.02079.0
FCN S-M3Dice0.9519 ± 0.02814.00.8470 ± 0.073713.0
Jaccard0.9096 ± 0.04994.00.7414 ± 0.107013.0
Sensitivity0.9341 ± 0.05435.00.8160 ± 0.11529.0
Specificity0.9865 ± 0.01076.00.9604 ± 0.030015.0
ResUNetDice0.9348 ± 0.050211.00.8569 ± 0.077911.0
Jaccard0.8816 ± 0.086811.00.7575 ± 0.115211.0
Sensitivity0.9029 ± 0.082512.00.8152 ± 0.131611.0
Specificity0.9896 ± 0.00672.00.9712 ± 0.018012.0
ResUNet CRFFg S-M1Dice0.9456 ± 0.03177.00.8851 ± 0.04494.0
Jaccard0.8984 ± 0.05607.00.7968 ± 0.07094.0
Sensitivity0.9472 ± 0.05401.00.8283 ± 0.08535.0
Specificity0.9725 ± 0.023014.00.9841 ± 0.01233.0
ResUNet CRFFg S-M3Dice0.9111 ± 0.060215.00.8969 ± 0.04441.0
Jaccard0.8420 ± 0.095115.00.8160 ± 0.07371.0
Sensitivity0.9075 ± 0.060711.00.8675 ± 0.08031.0
Specificity0.9663 ± 0.034615.00.9712 ± 0.024411.0
ResUNet S-M1Dice0.9558 ± 0.02791.00.8865 ± 0.06763.0
Jaccard0.9167 ± 0.04981.00.8026 ± 0.10613.0
Sensitivity0.9459 ± 0.04822.00.8403 ± 0.11232.0
Specificity0.9831 ± 0.015210.00.9750 ± 0.02878.0
ResUNet S-M3Dice0.9237 ± 0.041114.00.8677 ± 0.08949.0
Jaccard0.8610 ± 0.071314.00.7763 ± 0.12819.0
Sensitivity0.8875 ± 0.075614.00.8179 ± 0.13338.0
Specificity0.9846 ± 0.01288.00.9755 ± 0.02177.0
U-NetDice0.9371 ± 0.031210.00.8713 ± 0.07568.0
Jaccard0.8832 ± 0.055110.00.7796 ± 0.11458.0
Sensitivity0.9120 ± 0.057110.00.8107 ± 0.124812.0
Specificity0.9811 ± 0.019911.00.9847 ± 0.01302.0
U-Net CRFFg S-M1Dice0.9448 ± 0.02978.00.8827 ± 0.06175.0
Jaccard0.8969 ± 0.05288.00.7954 ± 0.09655.0
Sensitivity0.9160 ± 0.05619.00.8383 ± 0.10624.0
Specificity0.9902 ± 0.00571.00.9780 ± 0.01245.0
U-Net CRFFg S-M3Dice0.9252 ± 0.040413.00.8821 ± 0.06456.0
Jaccard0.8634 ± 0.069413.00.7948 ± 0.10046.0
Sensitivity0.8831 ± 0.073015.00.8231 ± 0.11107.0
Specificity0.9893 ± 0.00663.00.9873 ± 0.00881.0
U-Net S-M1Dice0.9400 ± 0.03649.00.8898 ± 0.05362.0
Jaccard0.8890 ± 0.06359.00.8056 ± 0.08612.0
Sensitivity0.9162 ± 0.06198.00.8384 ± 0.09043.0
Specificity0.9866 ± 0.00865.00.9777 ± 0.02086.0
U-Net S-M3Dice0.9293 ± 0.041912.00.8767 ± 0.07727.0
Jaccard0.8707 ± 0.072812.00.7883 ± 0.11527.0
Sensitivity0.8934 ± 0.079213.00.8152 ± 0.118110.0
Specificity0.9878 ± 0.00984.00.9805 ± 0.01894.0

References

  1. Brown, D.T.; Wildsmith, J.A.W.; Covino, B.G.; Scott, D.B. Effect of Baricity on Spinal Anaesthesia with Amethocaine. BJA Br. J. Anaesth. 1980, 52, 589–596. [Google Scholar] [CrossRef] [PubMed]
  2. McCombe, K.; Bogod, D. Regional anaesthesia: Risk, consent and complications. Anaesthesia 2021, 76, 18–26. [Google Scholar] [CrossRef] [PubMed]
  3. Chae, Y.; Park, H.J.; Lee, I.S. Pain modalities in the body and brain: Current knowledge and future perspectives. Neurosci. Biobehav. Rev. 2022, 139, 104744. [Google Scholar] [CrossRef] [PubMed]
  4. Curatolo, M.; Petersen-Felix, S.; Arendt-Nielsen, L. Assessment of regional analgesia in clinical practice and research. Br. Med Bull. 2005, 71, 61–76. [Google Scholar] [CrossRef] [Green Version]
  5. Bruins, A.; Kistemaker, K.; Boom, A.; Klaessens, J.; Verdaasdonk, R.; Boer, C. Thermographic skin temperature measurement compared with cold sensation in predicting the efficacy and distribution of epidural anesthesia. J. Clin. Monit. Comput. 2018, 32, 335–341. [Google Scholar] [CrossRef] [PubMed]
  6. Haren, F.; Kadic, L.; Driessen, J. Skin temperature measured by infrared thermography after ultrasound-guided blockade of the sciatic nerve. Acta Anaesthesiol. Scand. 2013, 57. [Google Scholar] [CrossRef]
  7. Stevens, M.F.; Werdehausen, R.; Hermanns, H.; Lipfert, P. Skin temperature during regional anesthesia of the lower extremity. Anesth. Analg. 2006, 102, 1247–1251. [Google Scholar] [CrossRef]
  8. Werdehausen, R.; Braun, S.; Hermanns, H.; Freynhagen, R.; Lipfert, P.; Stevens, M.F. Uniform Distribution of Skin-Temperature Increase After Different Regional-Anesthesia Techniques of the Lower Extremity. Reg. Anesth. Pain Med. 2007, 32, 73–78. [Google Scholar] [CrossRef]
  9. Zhang, L.; Nan, Q.; Bian, S.; Liu, T.; Xu, Z. Real-time segmentation method of billet infrared image based on multi-scale feature fusion. Sci. Rep. 2022, 12, 6879. [Google Scholar] [CrossRef]
  10. Kütük, Z.; Algan, G. Semantic Segmentation for Thermal Images: A Comparative Survey. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, New Orleans, LA, USA, 19–20 June 2022. [Google Scholar] [CrossRef]
  11. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. arXiv 2014, arXiv:1411.4038. [Google Scholar] [CrossRef]
  12. Bi, L.; Kim, J.; Kumar, A.; Fulham, M.J.; Feng, D. Stacked fully convolutional networks with multi-channel learning: Application to medical image segmentation. Vis. Comput. 2017, 33, 1061–1071. [Google Scholar] [CrossRef] [Green Version]
  13. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar] [CrossRef]
  14. Kumar, V.; Webb, J.M.; Gregory, A.; Denis, M.; Meixner, D.D.; Bayat, M.; Whaley, D.H.; Fatemi, M.; Alizad, A. Automated and real-time segmentation of suspicious breast masses using convolutional neural network. PLoS ONE 2018, 13, e0195816. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. arXiv 2018, arXiv:1807.10165. [Google Scholar] [CrossRef]
  16. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder–decoder Architecture for Image Segmentation. arXiv 2016, arXiv:1511.00561. [Google Scholar] [CrossRef]
  17. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. arXiv 2018, arXiv:1703.06870. [Google Scholar] [CrossRef]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. arXiv 2016, arXiv:1506.01497. [Google Scholar] [CrossRef] [Green Version]
  19. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. arXiv 2017, arXiv:1612.01105. [Google Scholar] [CrossRef]
  20. Arteaga-Marrero, N.; Hernández, A.; Villa, E.; González-Pérez, S.; Luque, C.; Ruiz-Alzola, J. Segmentation Approaches for Diabetic Foot Disorders. Sensors 2021, 21, 934. [Google Scholar] [CrossRef]
  21. Fischler, M.A.; Bolles, R.C. Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  22. Bouallal, D.; Bougrine, A.; Douzi, H.; Harba, R.; Canals, R.; Vilcahuaman, L.; Arbanil, H. Segmentation of plantar foot thermal images: Application to diabetic foot diagnosis. In Proceedings of the 2020 International Conference on Systems, Signals and Image Processing (IWSSIP), Niteroi, Brazil, 1–3 July 2020; pp. 116–121. [Google Scholar] [CrossRef]
  23. Bougrine, A.; Harba, R.; Canals, R.; Ledee, R.; Jabloun, M. On the segmentation of plantar foot thermal images with deep learning. In Proceedings of the 2019 27th European Signal Processing Conference (EUSIPCO), A Coruna, Spain, 2–6 September 2019. [Google Scholar] [CrossRef]
  24. Mejia-Zuluaga, R.; Aguirre-Arango, J.C.; Collazos-Huertas, D.; Daza-Castillo, J.; Valencia-Marulanda, N.; Calderón-Marulanda, M.; Aguirre-Ospina, Ó.; Alvarez-Meza, A.; Castellanos-Dominguez, G. Deep Learning Semantic Segmentation of Feet Using Infrared Thermal Images. In Proceedings of the Advances in Artificial Intelligence—IBERAMIA, Cartagena de Indias, Colombia, 23–25 November 2022; Bicharra Garcia, A.C., Ferro, M., Rodríguez Ribón, J.C., Eds.; Springer International Publishing: Cham, Switzerland, 2022; pp. 342–352. [Google Scholar]
  25. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar] [CrossRef]
  26. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar] [CrossRef]
  27. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation. In Computer Vision—ECCV 2022 Workshops; Springer: Cham, Switzerland, 2021; pp. 205–218. [Google Scholar] [CrossRef]
  28. Zhang, Y.; Liu, H.; Hu, Q. TransFuse: Fusing Transformers and CNNs for Medical Image Segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2021; pp. 14–24. [Google Scholar] [CrossRef]
  29. Li, S.; Sui, X.; Luo, X.; Xu, X.; Liu, Y.; Goh, R. Medical Image Segmentation Using Squeeze-and-Expansion Transformers. Ijcai Int. Jt. Conf. Artif. Intell. 2021, 807–815. [Google Scholar] [CrossRef]
  30. Luo, X.; Hu, M.; Song, T.; Wang, G.; Zhang, S. Semi-Supervised Medical Image Segmentation via Cross Teaching between CNN and Transformer. In Proceedings of the International Conference on Medical Imaging with Deep Learning, Zurich, Germany, 6–8 July 2022; pp. 820–833. [Google Scholar]
  31. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment Anything. arXiv 2023, arXiv:2304.02643. [Google Scholar] [CrossRef]
  32. Ruan, B.K.; Shuai, H.H.; Cheng, W.H. Vision Transformers: State of the Art and Research Challenges. arXiv 2022, arXiv:2207.03041. [Google Scholar] [CrossRef]
  33. Karimi, D.; Warfield, S.K.; Gholipour, A. Critical Assessment of Transfer Learning for Medical Image Segmentation with Fully Convolutional Neural Networks. arXiv 2022, arXiv:2006.00356. [Google Scholar] [CrossRef]
  34. Alzubaidi, L.; Fadhel, M.A.; Al-Shamma, O.; Zhang, J.; Santamaría, J.; Duan, Y.; R. Oleiwi, S. Towards a better understanding of transfer learning for medical imaging: A case study. Appl. Sci. 2020, 10, 4523. [Google Scholar] [CrossRef]
  35. Guan, H.; Liu, M. Domain Adaptation for Medical Image Analysis: A Survey. IEEE Trans. Biomed. Eng. 2022, 69, 1173–1185. [Google Scholar] [CrossRef]
  36. Anas, E.M.A.; Nouranian, S.; Mahdavi, S.S.; Spadinger, I.; Morris, W.J.; Salcudean, S.E.; Mousavi, P.; Abolmaesumi, P. Clinical Target-Volume Delineation in Prostate Brachytherapy Using Residual Neural Networks. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2017, Quebec City, QC, Canada, 11–13 September 2017; Descoteaux, M., Maier-Hein, L., Franz, A., Jannin, P., Collins, D.L., Duchesne, S., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 365–373. [Google Scholar]
  37. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. arXiv 2015, arXiv:1512.04150. [Google Scholar] [CrossRef]
  38. Zhang, A.; Lipton, Z.C.; Li, M.; Smola, A.J. Dive into deep learning. arXiv Prepr. 2021, arXiv:2106.11342. [Google Scholar]
  39. Rahimi, A.; Recht, B. Random features for large-scale kernel machines. Adv. Neural Inf. Process. Syst. 2007, 20, 1–8. [Google Scholar]
  40. Rudin, W.; Bers, L.; Courant, R.; Stoker, J.J.; Henney, D.R. Fourier Analysis on Groups; Interscience Tracts in Pure and Applied Mathematics; Interscience: New York, NY, USA, 1976; p. 19. [Google Scholar]
  41. Álvarez-Meza, A.M.; Cárdenas-Pe na, D.; Castellanos-Dominguez, G. Unsupervised Kernel Function Building Using Maximization of Information Potential Variability. In Proceedings of the Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications, Puerto Vallarta, Mexico, 2–5 November 2014; Bayro-Corrochano, E., Hancock, E., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 335–342. [Google Scholar]
  42. Bronstein, M.M.; Bruna, J.; Cohen, T.; Velickovic, P. Geometric Deep Learning: Grids, Groups, Graphs, Geodesics, and Gauges. arXiv 2021, arXiv:2104.13478. [Google Scholar] [CrossRef]
  43. Jiang, P.T.; Zhang, C.B.; Hou, Q.; Cheng, M.M.; Wei, Y. LayerCAM: Exploring hierarchical class activation maps for localization. IEEE Trans. Image Process. 2021, 30, 5875–5888. [Google Scholar] [CrossRef] [PubMed]
  44. Jimenez-Casta no, C.A.; Álvarez-Meza, A.M.; Aguirre-Ospina, O.D.; Cárdenas-Pe na, D.A.; Orozco-Gutiérrez, Á.A. Random fourier features-based deep learning improvement with class activation interpretability for nerve structure segmentation. Sensors 2021, 21, 7741. [Google Scholar] [CrossRef]
  45. Galvin, E.M.; Niehof, S.; Medina, H.J.; Zijlstra, F.J.; van Bommel, J.; Klein, J.; Verbrugge, S.J.C. Thermographic temperature measurement compared with pinprick and cold sensation in predicting the effectiveness of regional blocks. Anesth. Analg. 2006, 102, 598–604. [Google Scholar] [CrossRef] [PubMed]
  46. Chestnut, D.H.; Wong, C.A.; Tsen, L.C.; Kee, W.M.D.N.; Beilin, Y.; Mhyre, J. Chestnut’s Obstetric Anesthesia: Principles and Practice E-Book: Expert Consult-Online and Print; Saunders: Philadelphia, PA, USA, 2014. [Google Scholar]
  47. Asghar, S.; Lundstrøm, L.H.; Bjerregaard, L.S.; Lange, K.H.W. Ultrasound-guided lateral infraclavicular block evaluated by infrared thermography and distal skin temperature. Acta Anaesthesiol. Scand. 2014, 58, 867–874. [Google Scholar] [CrossRef]
  48. Lange, K.H.; Jansen, T.; Asghar, S.; Kristensen, P.; Skjønnemand, M.; Nørgaard, P. Skin temperature measured by infrared thermography after specific ultrasound-guided blocking of the musculocutaneous, radial, ulnar, and median nerves in the upper extremity. Br. J. Anaesth. 2011, 106, 887–895. [Google Scholar] [CrossRef] [Green Version]
  49. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar] [CrossRef]
  50. Ibtehaz, N.; Rahman, M.S. MultiResUNet: Rethinking the U-Net architecture for multimodal biomedical image segmentation. Neural Netw. 2020, 121, 74–87. [Google Scholar] [CrossRef]
  51. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In DLMIA 2018, ML-CDS 2018: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar] [CrossRef] [Green Version]
  52. Wang, X.; Wang, L.; Zhong, X.; Bai, C.; Huang, X.; Zhao, R.; Xia, M. PaI-Net: A modified U-Net of reducing semantic gap for surgical instrument segmentation. IET Image Process. 2021, 15, 2959–2969. [Google Scholar] [CrossRef]
  53. Haji, S.H.; Abdulazeez, A.M. Comparison of optimization techniques based on gradient descent algorithm: A review. PalArch J. Archaeol. Egypt/Egyptol. 2021, 18, 2715–2743. [Google Scholar]
  54. Van der Walt, S.; Schönberger, J.L.; Nunez-Iglesias, J.; Boulogne, F.; Warner, J.D.; Yager, N.; Gouillart, E.; Yu, T. scikit-image: Image processing in Python. PeerJ 2014, 2, e453. [Google Scholar] [CrossRef]
  55. Nakkiran, P.; Kaplun, G.; Bansal, Y.; Yang, T.; Barak, B.; Sutskever, I. Deep Double Descent: Where Bigger Models and More Data Hurt. arXiv 2019, arXiv:1912.02292. [Google Scholar] [CrossRef]
  56. Heckel, R.; Yilmaz, F.F. Early Stopping in Deep Networks: Double Descent and How to Eliminate it. arXiv 2020, arXiv:2007.10099. [Google Scholar] [CrossRef]
  57. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: Redesigning Skip Connections to Exploit Multiscale Features in Image Segmentation. IEEE Trans. Med. Imaging 2020, 39, 1856–1867. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  58. Peng, H.; Pappas, N.; Yogatama, D.; Schwartz, R.; Smith, N.A.; Kong, L. Random feature attention. arXiv 2021, arXiv:2103.02143. [Google Scholar]
  59. Nguyen, T.P.; Pham, T.T.; Nguyen, T.; Le, H.; Nguyen, D.; Lam, H.; Nguyen, P.; Fowler, J.; Tran, M.T.; Le, N. EmbryosFormer: Deformable Transformer and Collaborative Encoding-Decoding for Embryos Stage Development Classification. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikola, HI, USA, 2–7 January 2023; pp. 1981–1990. [Google Scholar]
Figure 1. The Convolutional Random Fourier Features Gradient (CRFFg) mapping, grounded in kernel methods, is employed for image-based data examination within deep learning frameworks.
Figure 1. The Convolutional Random Fourier Features Gradient (CRFFg) mapping, grounded in kernel methods, is employed for image-based data examination within deep learning frameworks.
Computation 11 00113 g001
Figure 2. Graphic depiction of the proposed relevance measures for Layer-Wise Class Activation Maps used in semantic segmentation tasks.
Figure 2. Graphic depiction of the proposed relevance measures for Layer-Wise Class Activation Maps used in semantic segmentation tasks.
Computation 11 00113 g002
Figure 3. Illustrative scenarios regarding our novel Layer-Wise Class Activation Maps for semantic segmentation.
Figure 3. Illustrative scenarios regarding our novel Layer-Wise Class Activation Maps for semantic segmentation.
Computation 11 00113 g003
Figure 4. Foot segmentation from thermal images using our CRFFg-based deep learning enhancement holding layer-wise weighted CAM interpretability.
Figure 4. Foot segmentation from thermal images using our CRFFg-based deep learning enhancement holding layer-wise weighted CAM interpretability.
Computation 11 00113 g004
Figure 5. Regional analgesia monitoring protocol using local anesthetics via epidural neuraxial and thermal images.
Figure 5. Regional analgesia monitoring protocol using local anesthetics via epidural neuraxial and thermal images.
Computation 11 00113 g005
Figure 6. Tested semantic segmentation architectures. Our CRFFg approach aims to enhance the data representation (see red dots).
Figure 6. Tested semantic segmentation architectures. Our CRFFg approach aims to enhance the data representation (see red dots).
Computation 11 00113 g006aComputation 11 00113 g006b
Figure 7. Visual inspection of the results on thermalFeet database without data augmentation. Our CRFFg-based enhancements are also presented. The first row shows the results for the FCN architecture, the second row for U-Net, and the third row for ResUNet. A unique color differentiates each model within an architecture. M1 and M3 represent CRFFg’s dimension as a multiplication factor of the enhanced layer’s size.
Figure 7. Visual inspection of the results on thermalFeet database without data augmentation. Our CRFFg-based enhancements are also presented. The first row shows the results for the FCN architecture, the second row for U-Net, and the third row for ResUNet. A unique color differentiates each model within an architecture. M1 and M3 represent CRFFg’s dimension as a multiplication factor of the enhanced layer’s size.
Computation 11 00113 g007
Figure 8. Visual inspection of the results on thermalFeet database with data augmentation. Our CRFFg-based enhancements are also presented. The first row shows the results for the FCN architecture, the second row for U-Net, and the third row for ResUNet. A unique color differentiates each model within an architecture. M1 and M3 represent CRFFg’s dimension as a multiplication factor of the enhanced layer’s size.
Figure 8. Visual inspection of the results on thermalFeet database with data augmentation. Our CRFFg-based enhancements are also presented. The first row shows the results for the FCN architecture, the second row for U-Net, and the third row for ResUNet. A unique color differentiates each model within an architecture. M1 and M3 represent CRFFg’s dimension as a multiplication factor of the enhanced layer’s size.
Computation 11 00113 g008
Figure 9. Training neural network loss vs. epochs corresponding to the various models examined are presented. M1 and M3 signify the dimensions of the CRFFg layer, expressed as multiplication factors of the enhanced layer’s size.
Figure 9. Training neural network loss vs. epochs corresponding to the various models examined are presented. M1 and M3 signify the dimensions of the CRFFg layer, expressed as multiplication factors of the enhanced layer’s size.
Computation 11 00113 g009
Figure 10. Results of the comparison between methods. The segmentation performance of ThermalFeet is evaluated using baseline models FCN, UNet, and ResUNet, and compared to our proposal that incorporates CRFFg-based enhancements. M1 and M3 represent CRFFg’s dimension as a multiplication factor of the enhanced layer’s size. Aug stands for data augmentation. (a) Segmentation performance results on ThermalFeet database. The three types of architecture used in this study (FCN, U-Net, ResUNet) are differentiated by color. The type of variation in the architecture is indicated by the marker used; (b) The improvement of each strategy, normalized with respect to the baseline performance of each architecture.
Figure 10. Results of the comparison between methods. The segmentation performance of ThermalFeet is evaluated using baseline models FCN, UNet, and ResUNet, and compared to our proposal that incorporates CRFFg-based enhancements. M1 and M3 represent CRFFg’s dimension as a multiplication factor of the enhanced layer’s size. Aug stands for data augmentation. (a) Segmentation performance results on ThermalFeet database. The three types of architecture used in this study (FCN, U-Net, ResUNet) are differentiated by color. The type of variation in the architecture is indicated by the marker used; (b) The improvement of each strategy, normalized with respect to the baseline performance of each architecture.
Computation 11 00113 g010aComputation 11 00113 g010b
Figure 11. Results of Interpretability Measures on ThermalFeet. The three types of architecture used in this study (FCN, U-Net, ResUNet) are differentiated by color. The type of variation in the architecture is indicated by the marker used. M1 and M3 represent CRFFg’s dimension as a multiplication factor of the enhanced layer’s size.
Figure 11. Results of Interpretability Measures on ThermalFeet. The three types of architecture used in this study (FCN, U-Net, ResUNet) are differentiated by color. The type of variation in the architecture is indicated by the marker used. M1 and M3 represent CRFFg’s dimension as a multiplication factor of the enhanced layer’s size.
Computation 11 00113 g011
Figure 12. Salient relevance analysis results. Best models concerning the Mask-based Cumulative Relevance, ϱ r measure, are presented for FCN, UNet, and ResUNet with our CRFFg-based enhancement.
Figure 12. Salient relevance analysis results. Best models concerning the Mask-based Cumulative Relevance, ϱ r measure, are presented for FCN, UNet, and ResUNet with our CRFFg-based enhancement.
Computation 11 00113 g012
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aguirre-Arango, J.C.; Álvarez-Meza, A.M.; Castellanos-Dominguez, G. Feet Segmentation for Regional Analgesia Monitoring Using Convolutional RFF and Layer-Wise Weighted CAM Interpretability. Computation 2023, 11, 113. https://doi.org/10.3390/computation11060113

AMA Style

Aguirre-Arango JC, Álvarez-Meza AM, Castellanos-Dominguez G. Feet Segmentation for Regional Analgesia Monitoring Using Convolutional RFF and Layer-Wise Weighted CAM Interpretability. Computation. 2023; 11(6):113. https://doi.org/10.3390/computation11060113

Chicago/Turabian Style

Aguirre-Arango, Juan Carlos, Andrés Marino Álvarez-Meza, and German Castellanos-Dominguez. 2023. "Feet Segmentation for Regional Analgesia Monitoring Using Convolutional RFF and Layer-Wise Weighted CAM Interpretability" Computation 11, no. 6: 113. https://doi.org/10.3390/computation11060113

APA Style

Aguirre-Arango, J. C., Álvarez-Meza, A. M., & Castellanos-Dominguez, G. (2023). Feet Segmentation for Regional Analgesia Monitoring Using Convolutional RFF and Layer-Wise Weighted CAM Interpretability. Computation, 11(6), 113. https://doi.org/10.3390/computation11060113

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop