Next Article in Journal
Plasma-Treated Nanostructured Resistive Gas Sensors: A Review
Previous Article in Journal
The Riemannian Means Field Classifier for EEG-Based BCI Data
Previous Article in Special Issue
SDA-Net: A Spatially Optimized Dual-Stream Network with Adaptive Global Attention for Building Extraction in Multi-Modal Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advancing Textile Damage Segmentation: A Novel RGBT Dataset and Thermal Frequency Normalization

by
Farshid Rayhan
1,
Jitesh Joshi
1,
Guangyu Ren
1,
Lucie Hernandez
2,
Bruna Petreca
2,
Sharon Baurley
2,
Nadia Berthouze
1 and
Youngjun Cho
1,*
1
Department of Computer Science, University College London, London NW1 2AE, UK
2
Materials Science Research Centre, Royal College of Art, London SW7 2EU, UK
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(7), 2306; https://doi.org/10.3390/s25072306
Submission received: 28 February 2025 / Revised: 28 March 2025 / Accepted: 28 March 2025 / Published: 5 April 2025

Abstract

:
RGB-Thermal (RGBT) semantic segmentation is an emerging technology for identifying objects and materials in high dynamic range scenes. Thermal imaging particularly enhances feature extraction at close range for applications such as textile damage detection. In this paper, we present RGBT-Textile, a novel dataset specifically developed for close-range textile and damage segmentation. We meticulously designed the data collection protocol, software tools, and labeling process in collaboration with textile scientists. Additionally, we introduce ThermoFreq, a novel thermal frequency normalization method that reduces temperature noise effects in segmentation tasks. We evaluate our dataset alongside six existing RGBT datasets using state-of-the-art (SOTA) models. Experimental results demonstrate the superior performance of the SOTA models with ThermoFreq, highlighting its effectiveness in addressing noise challenges inherent in RGBT semantic segmentation across diverse environmental conditions. We make our dataset publicly accessible to foster further research and collaborations.

1. Introduction

The textile industry faces significant sustainability challenges [1], striving to reduce textile waste and comply with global Net Zero goals [2,3]. The repair of clothing materials to prolong the useful life of garments is indispensable in terms of reducing the environmental impact [4]. Advanced machine learning methodologies, particularly in semantic segmentation, are critical in the automation of damage detection and repair processes, thereby fostering a more sustainable fashion industry. Semantic segmentation of RGBT image frames is a highly pertinent task capable of identifying textile materials and damages in dynamic range scenes. This task encompasses the categorization of garment regions based on semantic content, alongside the identification of damage. However, there is a conspicuous scarcity of dedicated and publicly accessible datasets [5,6,7,8] for this specific application, which hinders the development and evaluation of frameworks aimed at addressing sustainability challenges.
In this paper, we present RGBT-Textile, a meticulously designed segmentation dataset for the tasks of textile analysis and damage detection. The inception of our dataset is influenced by contemporary RGBT segmentation methodologies [9,10,11,12,13,14] as well as existing benchmark datasets. Through rigorous data collection and annotation, we have compiled the RGBT-Textile dataset, which consists of 1441 RGBT images accompanied by ground-truth segmentation masks. In Figure 1, we provide several exemplars from the RGBT-Textile dataset, illustrating the alignment of RGB and thermal image frames alongside their corresponding segmentation labels. We evaluate the performance of state-of-the-art RGBT semantic segmentation models using the proposed RGBT-Textile dataset, thereby establishing a benchmark performance standard. Performance metrics extracted from this dataset underscore the inherent complexity of detecting damages in textiles and consequently contribute to sustainability goals related to the textile industry.
Integrating thermal infrared (TIR) images, which are insensitive to illumination variations, with visible light data is preferred in several challenging scenarios and high dynamic scenes, such as low illumination, fog, and rain [15,16,17]. In RGBT segmentation, foreground–background differentiation and complementary information extraction are primary challenges. Previous methods have used simple fusion techniques [18,19], while attention-based approaches such as FEANet [20], CCFFNet [21], AFNet [22], and CMX [23] have addressed these more effectively. However, issues such as sub-optimal differentiation, computational overhead, and insufficient complementary information mining persist.
Normalization techniques are crucial for aligning the statistical properties of modalities to ensure robust feature extraction and fusion. Traditional methods like Min-Max, Mean-STD, and histogram-based normalization are widely used but face limitations: Min-Max is outlier-sensitive, Mean-STD amplifies noise, and histogram approaches require downstream task adaptation for thermal images [24]. However, when applied to thermal images, such approaches require careful consideration of the downstream task [25]. To address this challenge, this paper further proposes ThermoFreq, a frequency-domain thermal normalization technique using Fourier transform to selectively adjust components based on statistical distribution, which transforms an image into the frequency domain using Fourier transform. The normalized images in the frequency domain are then transformed back to the image space and used to train the models. ThermoFreq enhances thermal characteristics, leading to the superior segmentation performance of the state-of-the-art (SOTA) RGBT models in challenging scenarios.
In the following sections, we report our detailed data collection process and compare SOTA RGBT segmentation models on existing datasets with and without ThermoFreq. Our dataset and findings serve as a catalyst for future research initiatives, fostering collaboration and innovation within the computer vision and textile garment repair community. Our contributions are as follows:
  • The RGBT-Textile dataset, a novel RGB-Thermal image dataset of textile materials with damages, primarily for the segmentation of textile materials and damages along with benchmarks and experimental results. The dataset is available at this download link https://drive.google.com/drive/folders/1HLri3SDPHSY0AsAmaEloh_0Q13E91u2v (accessed on 28 March 2025).
  • The frequency-based thermal image normalization technique ThermoFreq, to selectively adjust frequency components based on their statistical distribution, resulting in improved segmentation performance of the RGBT models.
  • Experiments with the SOTA RGBT segmentation models and benchmark datasets to highlight the effectiveness of ThermoFreq in addressing temperature noise challenges.

2. Related Works

2.1. RGBT Datasets

Researchers have contributed several RGBT datasets that are highly valuable for developing and evaluating RGBT semantic segmentation algorithms. The KAIST Multispectral Pedestrian Dataset [26] offers RGBT images captured under various ambient conditions, supporting both object detection and semantic segmentation tasks. Similarly, the LLVIP Dataset [27] provides RGBT images in different lighting conditions and supports object detection, semantic segmentation, and recognition tasks. The MFNet Dataset [15] includes multispectral RGBT images captured under various environmental conditions, as well as supporting object detection and semantic segmentation. Additionally, the OSU Thermal Pedestrian Database [28] contains RGBT images under various environmental conditions, which aids in object detection and semantic segmentation. The PST900 Dataset [17], derived from the DARPA Subterranean Challenge, includes thermal image pairs across four classes and supports semantic segmentation tasks. These datasets collectively offer a comprehensive range of RGBT data, facilitating advances in semantic segmentation and other vision-related tasks.

2.2. RGBT Segmetation Models

RGBT fusion methods have become prominent in computer vision for tasks such as pedestrian detection, vehicle detection, semantic segmentation, detection of salient objects, crowd counting, and fusion tracking. Notable advances include the Causal Mode Multiplexer [29], Triple Flow Network [30], ICAFusion [31], Feature-guided pretraining [32], and Fusion Architecture [33], which enhance detection performance through iterative cross-attention and multimodal data fusion. In semantic segmentation, the best performing models such as CRM_RGBT [34] and MMSFormer [11] have set new standards with their unique architectures and masking mechanisms, leading to more accurate results [14]. Salient object detection and crowd counting have also seen improvements through better fusion architectures and loss functions [9,35,36]. Fusion-based tracking methods leverage RGBT data for robust performance in challenging lighting conditions [37]. These advances underscore the importance of multimodal data for enhanced computer vision performance. Inspired by these developments and to address challenges in garment damage detection, we developed the proposed RGBT-Textile dataset as an RGBT dataset, detailed in Section 3.

3. RGBT-Textile Dataset

We present RGBT-Textile, an RGBT dataset designed for the detection of textile damage through the semantic segmentation task. This dataset consists of 1441 pairs of RGB and thermal images, each pair meticulously annotated with segmentation labels. These labels categorize the data into three distinct classes: background, garment, and damage. In Figure 1, we provide samples from the dataset, where the damaged area of a garment is indicated in red, the intact garment is marked in blue, and the background is represented in black.

3.1. Data Collection Protocol and Setups

To acquire the dataset, garments were first screened to ensure variations in types of material and damages. Data were then acquired according to the simple protocol as depicted in Figure 2A, which consisted of: (i) background acquisition to capture thermal ambience, (ii) RGBT acquisition, and (iii) RGBT acquisition after momentarily heating the garment using a hair dryer. As depicted in Figure 2B, RGB and thermal cameras were arranged so that the garments could be placed underneath the cameras at a distance of approximately 0.5 m. A Logitech BRIO 4K UHD webcam (Logitech International, Lausanne, Switzerland) was used to capture RGB video frames with 640 × 480 resolution, while thermal infrared frames were captured using a thermal camera (A65SC, FLIR systems Inc., Wilsonville, OR, USA) with 640 × 512 resolution. The spatial alignment between the RGB and thermal frames was achieved using a custom 3D printed mount that enabled placing the RGB camera on the thermal camera as shown in Figure 2B.
To ensure synchronization between RGB and thermal frames, we developed a multithreaded application with a user interface using Python v3.10 and Qt v6.0 framework [38], as shown in Figure 2D. The source code [39] for this application was adapted from a multithreaded application developed for PhysioKit [40] to acquire the iBVP dataset [41], comprising synchronized RGB and thermal facial video frames along with physiological signals. Each acquired image was annotated for damaged regions, textile material shape, and nontextile parts by two annotators, and rigorously validated by three other researchers, including textile experts. A Computer Vision Annotation Tool (CVAT) [42] was used to mark polygons as shown in Figure 2C, which were further transformed into segmentation masks.

3.2. Diversity of Textile Damage and Materials

The RGBT-Textile dataset encompasses a broad spectrum of textile damage and material types to reflect real-world challenges in garment inspection. Damages were categorized into three primary classes: structural damage (for example, holes, tears, frayed edges), surface anomalies (for example, stains, discolorations, burns), and manufacturing flaws (for example, stitching errors, uneven dyeing) [4,32]. These damages vary in size (5–150 mm2) and spatial distribution, with irregular shapes mimicking the natural patterns of wear and tear observed in post-consumer textiles [1]. The dataset includes garments composed of cotton (42%), polyester (28%), blended fabrics (20%), and synthetic fibers (10%), selected for their prevalence in fast fashion and their susceptibility to thermal variations during imaging [43].
As shown in Table 1, the characteristics of the damages are further quantified by the variability in the thermal response ( δ T = 0.5–3.0 °C), which complicates segmentation due to the overlapping temperature profiles between the damaged and intact regions [25]. For example, polyester exhibits a lower thermal emissivity than cotton, which requires the extraction of modality-specific characteristics [44]. This diversity underscores the need for specialized approaches, as generic RGBT models often fail to distinguish subtle thermal gradients or irregular damage boundaries, as discussed in Section 6.3.

4. Thermal Frequency Normalization: ThermoFreq

In RGBT segmentation tasks, normalization assumes paramount importance due to the varied distributions of RGB and thermal pixels. The features present in thermal images are frequently less prominent and are subject to significant levels of ambient-temperature noise, thereby complicating the extraction of meaningful features. Earlier research with thermal images predominantly relied on the utilization of manufacturer-implied pseudo-color maps, which often linearly mapped raw temperature values onto an 8-bit image frame [45]. However, subsequent studies have emphasized the necessity of application-specific optimal mapping, such as optimal quantization for the robust tracking of regions of interest [25]. Most contemporary thermal cameras provide access to a raw temperature image matrix, as opposed to pseudo-color maps or quantized images. This particularity highlights the necessity for tailored normalization techniques to ensure the effective fusion and precise interpretation of multispectral data, a requirement that researchers often overlook.
The significance of enhancing the saliency of features in thermal imaging has been highlighted by previous research. For example, optimal quantization increased the reliability of tracking respiratory signals and vasomotor activities in low-resolution mobile thermal imaging [25,46,47]. In 3D reconstruction, applying a cosine transformation to intensity values [48] proved beneficial, although this technique increases contrast for certain intensity levels while diminishing it for others. In the realm of deep learning models, improving the saliency of input thermal images can serve as a promising preprocessing step with considerable potential to enhance downstream performance, yet this area has received insufficient attention in the current literature. Drawing upon empirical observations that highlight the occurrence of noise predominantly at higher frequencies relative to informative pixels, we propose ThermoFreq as an essential preprocessing step for thermal imaging segmentation, aimed at mitigating the effects of temperature noise in segmentation tasks. This technique involves the manipulation of the frequency components of the image within the frequency domain. The process described in ThermoFreq includes transforming the image from the spatial to the frequency domain, evaluating and adjusting the temperature frequency values across the dataset, and then reverting the image back to the spatial domain. In practice, this involves the removal or nullification of high-frequency values, followed by scaling the residual pixel values to a range of 0 to 255.
The first step in frequency-based normalization is to transform an image from the spatial domain to the frequency domain using the Fourier transform. For a given image f ( x , y ) , the 2D Fourier Transform isdefined as:
F ( u , v ) = x = 0 M 1 y = 0 N 1 f ( x , y ) e 2 π i u x M + v y N
where M and N are the dimensions of the image, F ( u , v ) represents the frequency domain representation of the image, and ( u , v ) are the frequency coordinates. After an image is transformed into the frequency domain, the magnitude of the high-frequency components is set to zero to remove noise. For instance, to zero out a central region of the frequency components, we define a mask:
H ( u , v ) = 1 if ( u u 0 ) 2 + ( v v 0 ) 2 > D 0 0 otherwise
where ( u 0 , v 0 ) is the center of the frequency domain and D 0 is the cutoff distance as denoted t. The threshold t determines the radius of the region around the center that is zeroed out, effectively filtering out those frequency components. The modified frequency domain image G ( u , v ) is then obtained by element-wise multiplication of F ( u , v ) with the mask H ( u , v ) :
G ( u , v ) = F ( u , v ) · H ( u , v )
After manipulating the frequency components, the inverse Fourier transform is applied to transform the image back to the spatial domain. The 2D inverse Fourier transform is defined as:
g ( x , y ) = 1 M N u = 0 M 1 v = 0 N 1 G ( u , v ) e 2 π i u x M + v y N
where g ( x , y ) is the resulting image in the spatial domain. Finally, the image is normalized to the range [ 0 , 255 ] . This is achieved by:
g norm ( x , y ) = g ( x , y ) min ( g ) max ( g ) min ( g ) × 255
where min ( g ) and max ( g ) are the minimum and maximum pixel values in the image g ( x , y ) , respectively. This process is depicted in Figure 3.
This process retains the mean temperature value of an original image while increasing the standard deviation, indicating a greater spread of temperature values and enhanced image features.
Figure 4 demonstrates how ThermoFreq enhances thermal images by manipulating specific frequencies in the MF dataset [49], by plotting the mean and standard deviation of each image in the dataset, before and after applying ThermoFreq. The effect observed in Figure 4 is with a cut-off distance of D 0 = 1 .
Figure 5 provides a qualitative depiction of the effect of modifying the cut-off distance D 0 on thermal images sourced from two distinct datasets: MF (left) and PST (right). The images illustrate how the alteration of specific frequencies by ThermoFreq enhances features and contrast within the thermal images. Through adjustment of D 0 , the figure reveals visual alterations in thermal images, underscoring the ability of ThermoFreq to enhance fine details while preserving the overall mean temperature (see also Figure 4).
The observed color variations in the pseudo-color maps, such as increased redness or yellowness in the MF data set, reflect temperature deviations relative to ambient conditions, where the shades of pseudo-color correlate with the thermal contrast of the object against its surroundings. In contrast, the PST dataset, comprising underwater thermal scenes, exhibits minimal visual changes due to inherently constrained temperature gradients in aquatic environments. Although all processing was performed on the raw thermal data, we show the pseudo-color maps for visual illustration and further contextualize these observations; the processed images demonstrate ThermoFreq’s adaptive enhancement capabilities sufficiently. This aligns with the technique’s design to amplify discriminative features while mitigating noise, even when ambient thermal profiles limit observable variability.
Figure 6 represents temperature data in the 3rd dimension to elucidate the effects of varying cut-off distances D 0 on discrete thermal images sourced from two datasets: MF (upper) and PST (lower). These images demonstrate the manner in which ThermoFreq adjusts specific frequencies to enhance distinct features and contrast. Through modulation of D 0 , the qualitative influence on thermal images is reflected, underscoring the technique’s ability to augment features and enhance the overall saliency of the images.

5. Experiments

We first evaluate the effectiveness of ThermoFreq using existing RGBT segmentation models. Next, we evaluate the impact of the tolerance threshold t on segmentation performance across different datasets. Lastly, we evaluate the impact of the tolerance threshold t on individual classes of the data set RGBT-Textile. For evaluation, we used two transformer-based RGBT models, viz. CRM_RGBTSeg [34] and MMSFormer [11], which have demonstrated state-of-the-art performance for the RGBT segmentation task. These models are variations of the Swin Transformer models [50] adapted to handle an additional channel of thermal images. In addition, we also report the performance using U-Net [51] models with ResNet18 and ResNet50 [52] backbones. We adopt three commonly used metrics for evaluation: mean Intersection over Union (mIoU), mean precision, and mean recall [18,20,20]. Our model is implemented using PyTorch v1.9.0 framework [53], PyTorch lighting [54], and MMCV libraries [55], and trained on an NVIDIA GeForce RTX 3090 GPU. We follow the training and testing settings from [12,34] and resize the training and testing images to 224 × 224 .

6. Results and Discussion

We first evaluate the effectiveness of ThermoFreq and the impact of the tolerance threshold t on the performance of semantic segmentation models, specifically CRM_RGBTSeg, MMSFormer, and CNN-based models (UNet with ResNet18 and ResNet50 backbones), across RGBT datasets. For this study, we limit our models to CNN [52,56,57,58] and transformer-based architecture [11,59] due to their high generalization capability. The results in Table 2 and Table 3 show performance gains when models are trained using ThermoFreq, with transformer-based models generally outperforming CNN-based models by 1–2%.

6.1. Evaluation of ThermoFreq with Transformer and CNN-Based Models

Table 2 shows the Mean Intersection over Union (mIoU) for CRM_RGBTSeg using Swin Base, Swin Small, Swin Tiny, and UNet with ResNet18 and ResNet50 backbones, both with and without normalization. Transformer-based models consistently outperform CNN-based models across different datasets by approximately 1–2%. For example, on the KAIST Multispectral Pedestrian Dataset, CRM_RGBTSeg with Swin Base achieves an mIoU of 0.567 with normalization, compared to 0.555 with ResNet50 and 0.553 with ResNet18. This suggests that transformer-based models are more robust in learning complex features from RGBT data.
ThermoFreq normalization improves mIoU for all versions of CRM_RGBTSeg and CNN-based models. In particular, Swin Tiny sees a significant improvement, with mIoU increasing from 0.512 to 0.554. In contrast, ResNet-based models show more modest gains, such as an increase from 0.542 to 0.550 for ResNet18. The improved performance of transformer models is attributed to their better ability to model global dependencies and integrate information across multiple modalities, such as RGB and thermal images, while CNN-based models are more reliant on local features.
The performance of MMSFormer with and without normalization is also presented in Table 2. The results show that MMSFormer, although improved with normalization, still lags behind CRM_RGBTSeg by a small margin. For example, on the KAIST Multispectral Pedestrian Dataset, MMSFormer achieves an mIoU of 0.561 with normalization, while UNet with ResNet50 achieves 0.552, highlighting the superior performance of transformer-based architectures even when compared to CNN models. The results indicate that the complementary random masking strategy in CRM_RGBTSeg is more effective for RGBT segmentation than the residue-based approach in MMSFormer.

6.2. Impact of Different Tolerance Thresholds

Table 3 shows the mIoU for CRM_RGBTSeg and MMSFormer with different tolerance thresholds t (0, 0.5, 1, 5, 10). We observe that t values affects model performance across datasets. For example, the KAIST multispectral pedestrian data set benefits from higher t values, achieving the highest mIoU of 0.567 at t = 10 for CRM_RGBTSeg. Conversely, the PST900 dataset, which has a lower temperature range, achieves its best mIoU of 0.885 at a lower t value of 0.5 for CRM_RGBTSeg. This indicates that datasets with a higher thermal temperature range benefit more from higher t values, as more aggressive denoising is necessary to handle the increased noise. On the other hand, datasets with a lower temperature range require lower t values to avoid losing important information during denoising.
The findings suggest that a small amount of normalization (tolerance level of 0.5) has little effect, while a moderate level (tolerance level of 1) generally improves performance by removing noise. However, high tolerance levels (5 or 10) can be detrimental, especially for datasets with small temperature ranges, as they discard critical information. This effect is more pronounced for smaller models, whereas larger models inherently ignore noise during training. In summary, moderate normalization enhances segmentation performance, particularly for smaller models and datasets with large temperature ranges.

6.3. Class Specific Performance on RGBT-Textile Dataset

Table 4 highlights the segmentation performance of CRM_RGBTSeg, MMSFormer, and UNet-ResNet18/ResNet50 on the RGBT-Textile dataset with background, garment, and damaged area as semantic classes, across different thresholds t ( t = 0 , 0.5 , 1 , 5 , 10 ). The results provide insights into the characteristics and challenges related to the detection of garment damage. The mIoU values for both models indicate that the dataset is moderately challenging, with the highest performance metrics peaking at t = 0.5 and t = 1 . This suggests that complexity of the dataset allows models to achieve good segmentation accuracy with relatively short processing times, but prolonged processing does not necessarily enhance performance and may introduce noise or overfitting. The precision and recall metrics follow a similar trend as mIoU, with optimal values at t = 0.5 and 1, and a gradual decline at t = 5 and t = 10 .
The mIoU values for CRM_RGBTSeg, MMSFormer, UNet-ResNet50, and UNet-ResNet18 show that the dataset is moderately challenging, with the best results generally achieved at t = 0.5 and t = 1 . UNet-ResNet50 shows competitive performance, especially for the background class, while CRM_RGBTSeg performs best overall for the damage area class. Precision and recall metrics follow similar trends, with peak performance at t = 0.5 and t = 1 , and a decline for t = 5 and t = 10 , suggesting that higher tolerance levels can introduce noise.

6.3.1. Background

The background class shows the highest segmentation performance across all models, with CRM_RGBTSeg achieving an mIoU between 0.850 and 0.870 and MMSFormer between 0.836 and 0.854. UNet-ResNet50 outperforms the other models with an mIoU of up to 0.883 at t = 1 , indicating that it can capture the relatively distinct background features of the dataset well.

6.3.2. Garment

The performance for the garment class is slightly lower than for the background, with CRM_RGBTSeg achieving an mIoU up to 0.830, and UNet-ResNet50 performing best with an mIoU of 0.849. This reflects the more variable nature of garments, including different textures and colors, increasing the difficulty of segmentation.

6.3.3. Damaged Area

The most challenging class is the damaged area, with mIoU values reaching 0.645 for CRM_RGBTSeg and 0.675 for UNet-ResNet50 at t = 1 . The lower performance suggests that the damaged areas are harder to distinguish due to their irregular shapes and smaller presence in the dataset.
The varying mIoU values between classes indicate that the RGBT-Textile data set contains a range of variability and complexity. Although the background and garment are segmented more reliably by the models, damaged areas present more challenges. This variability necessitates robust segmentation models that can generalize well across different classes and advance the state of damage detection. The lower performance for the damage area class highlights the need to further refine the normalization technique that can specifically enhance the corresponding features. Enhancing the quality and quantity of damaged area annotations, augmenting the dataset with more diverse examples of damaged areas, could further help improve segmentation performance.
These findings indicate that the UNet-ResNet50 model offers competitive performance across all semantic classes, particularly excelling in the background and garment segmentation tasks. The CRM_RGBTSeg model remains strong for damage area detection. Both models benefit from normalization at moderate tolerance thresholds ( t = 0.5 and t = 1 ). Further improvements could be achieved by increasing the diversity of garment and damage area samples in the dataset and enhancing the models’ ability to capture subtle features in challenging classes like the damaged area.
In the context of textile semantic segmentation tasks, among Swin-based Transformers such as CRM_RGBTseg, CNN-based models, and other Transformer models such as MMSFormer and UNet with ResNet backbones, each offers distinct advantages and challenges. Transformer models excel at modeling long-range dependencies and capturing global context, which translates into superior performance for tasks involving complex data modalities, such as RGB-Thermal (RGBT) segmentation. They consistently outperform CNNs by 1–2% in mean Intersection over Union (mIoU) across various datasets, as shown in Table 4. However, this performance comes at the cost of increased training time and a significantly higher number of parameters, making Transformer models computationally more expensive and difficult to deploy in resource-constrained environments. On the other hand, CNN-based models such as UNet with ResNet18 and ResNet50 are more efficient in terms of training speed and memory usage, making them easier to deploy and maintain in production systems. Despite their efficiency, CNNs are more reliant on local feature extraction, which can limit their ability to capture global context as effectively as Transformers. In summary, Transformer models offer better performance for complex segmentation tasks, but at the cost of higher computational overhead, while CNNs provide a more lightweight and deployable option with slightly lower accuracy.

6.4. Robustness to Synthetic Thermal Noise

To ascertain the efficacy of ThermoFreq under various ambient conditions, we enhanced all datasets with thermal-specific perturbations derived from prior work on thermal robustness [60]. These modifications consisted of: (i) Thermal Noise: The sensitivity of a thermal infrared imaging camera is determined by a specification known as the noise equivalent temperature difference (NETD), which specifies the minimum discernible temperature difference that the camera can detect with reliability. To simulate this form of noise, a random temperature value ( κ ) was introduced to each pixel such that 0 < κ < T h N E T D m a x , with the maximum NETD value, T h N E T D m a x , established as 0.1 °C, reflecting the worst-case specification criteria for a low-cost thermal imaging camera [60]. (ii) Random Thermal Occlusion: Masking 10–20% of regions with synthesized patches exhibiting extreme low or high temperatures. Subsequently, the models were retrained using these augmented data, while maintaining the integrity of the test sets.
Table 5 shows that ThermoFreq-enhanced models maintain superior performance under synthetic noise compared to baselines. For example, CRM_RGBTSeg + ThermoFreq achieves 0.843 mIoU on augmented RGBT-Textile vs. 0.821 for vanilla CRM_RGBTSeg, an improvement of 2.2%. CNN-based models exhibit larger performance drops (up to 4.8% for UNet-ResNet18), indicating greater sensitivity to thermal perturbations without explicit normalization.
This demonstrates ThermoFreq’s practical value for real-world deployment where ambient-temperature variations induce thermal noise. By explicitly normalizing thermal responses during preprocessing, our method reduces the dependency on data augmentation for robustness.

6.5. Comparative Analysis

Existing RGBT models, such as CRM_RGBTSeg and MMSFormer, demonstrate reduced efficacy on RGBT-Textile compared to pedestrian-centric datasets (e.g., KAIST, LLVIP). For example, irregular damage geometries (e.g., jagged tears) yield 12–18% lower precision in transformer-based models due to their dependence on global context, which struggles with localized anomalies [21]. Similarly, CNN architectures such as UNet-ResNet50 exhibit limited sensitivity to subtle thermal gradients ( δ T < 1.0 °C), critical for detecting stains or discolorations in synthetic fabrics [15]. These limitations highlight the need for domain-specific adaptations, such as ThermoFreq’s frequency-sensitive normalization, to address textile-specific noise and variability.

7. Conclusions

In this study, we present the dataset RGBT-Textile, crafted specifically for the semantic segmentation of textile damages. Through a comprehensive series of experiments, we assessed the performance of this dataset by employing benchmark methodologies such as the CRM_RGBTSeg and MMSFormer models. These models were evaluated with and without the implementation of our novel frequency-based normalization technique, ThermoFreq. The empirical findings illustrate that the ThermoFreq normalization technique considerably enhances the performance metrics of the RGBT models. Specifically, we noticed performance improvements ranging from 1% to 1.5% mIoU in different datasets. This improvement trend is consistent, albeit slightly diminished (approximately 0.5 to 0.7) for models trained in the absence of ThermoFreq. The ThermoFreq technique effectively mitigates the inherent noise present in thermal images, thus ensuring robust feature extraction and facilitating effective data fusion. This results in a better and more efficient fusion of RGB and thermal data, culminating in superior model performance in challenging segmentation scenarios. In summary, our RGBT-Textile dataset, in conjunction with the ThermoFreq normalization technique, establishes a strong foundation for advancements in the field of textile damage detection. These contributions underscore the potential for the integration of RGB and thermal data to improve segmentation accuracy, paving the way for more sustainable and automated textile repair solutions. Future research efforts will aim to expand the dataset with additional samples and labels, further investigate the advantages of our normalization technique, and promote collaboration within the computer vision and textile repair communities.

Author Contributions

Conceptualization, F.R., J.J., G.R. and Y.C.; methodology, F.R., J.J., G.R. and Y.C.; software, F.R., J.J. and G.R.; validation, F.R., J.J., G.R. and Y.C.; formal analysis, F.R., J.J. and G.R.; investigation, F.R., J.J., G.R., L.H., B.P. and Y.C.; resources, S.B., N.B. and Y.C.; data curation, F.R., J.J. and G.R.; writing—original draft preparation, F.R., J.J. and G.R.; writing—review and revision, F.R., J.J. and Y.C.; visualization, F.R. and J.J.; overall supervision, Y.C.; project administration, B.P., S.B., N.B. and Y.C.; funding acquisition, B.P., S.B., N.B. and Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Engineering and Physical Sciences Research Council (EPSRC), UKRI Digital Economy Programme Sustainable Digital Society (Award EP/V042289/1), Consumer Experience (CX) Digital Tools for Dematerialisation for the Circular Economy. Additionally, this work was funded in part by the EPSRC (EP/V011766/1) for the UK Research and Innovation (UKRI) Interdisciplinary Circular Economy Centre for Textiles: Circular Bioeconomy for Textile Materials. The APC was exempted for this publication as an invited paper.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset is available at this download link https://drive.google.com/drive/folders/1HLri3SDPHSY0AsAmaEloh_0Q13E91u2v (accessed on 28 March 2025).

Acknowledgments

Authors immensely thank Two Lives (Niklas Vila Karpe, Tina Lilienthal) for providing access to damaged garments, and consenting to capture RGB-thermal images of the same.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
3DThree dimensional
CNNConvolution Neural Network
mIoUMean Intersection over Union
RGBTRGB (color imaging) and thermal infrared imaging
SOTAState-of-the-art

References

  1. Millward-Hopkins, J.; Purnell, P.; Baurley, S. Scenarios for reducing the environmental impacts of the UK clothing economy. J. Clean. Prod. 2023, 420, 138352. [Google Scholar]
  2. Jacometti, V. Fashion and a Circular Economy; Ellen MacArthur Foundation: Cowes, UK, 2019. [Google Scholar]
  3. Jacometti, V. Circular economy and waste in the fashion industry. Laws 2019, 8, 27. [Google Scholar] [CrossRef]
  4. Boiten, V.J.; Han, S.L.C.; Tyler, D. Circular Economy Stakeholder Perspectives: Textile Collection Strategies to Support Material Circularity; European Union’s: Brussels, Belgium, 2017. [Google Scholar]
  5. Papers with Code—Thermal Image Segmentation—paperswithcode.com. Available online: https://paperswithcode.com/task/thermal-image-segmentation/latest (accessed on 27 March 2025).
  6. Wang, Q.; Yin, C.; Song, H.; Shen, T.; Gu, Y. UTFNet: Uncertainty-guided trustworthy fusion network for RGB-thermal semantic segmentation. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar]
  7. Pang, Y.; Huang, Y.; Weng, C.; Lyu, J.; Bai, C.; Yu, X. Enhanced RGB-T saliency detection via thermal-guided multi-stage attention network. Vis. Comput. 2025, 1–19. [Google Scholar] [CrossRef]
  8. Zhao, S.; Liu, Y.; Jiao, Q.; Zhang, Q.; Han, J. Mitigating modality discrepancies for RGB-T semantic segmentation. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 9380–9394. [Google Scholar] [PubMed]
  9. Bao, W.; Huang, M.; Hu, J.; Xiang, X. Attention-guided multi-modal and multi-scale fusion for multispectral pedestrian detection. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Shenzhen, China, 4–7 November 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 382–393. [Google Scholar]
  10. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  11. Reza, M.K.; Prater-Bennette, A.; Asif, M.S. MMSFormer: Multimodal Transformer for Material and Semantic Segmentation. IEEE Open J. Signal Process. 2024, 5, 599–610. [Google Scholar]
  12. Cheng, B.; Misra, I.; Schwing, A.G.; Kirillov, A.; Girdhar, R. Masked-attention mask transformer for universal image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1290–1299. [Google Scholar]
  13. Rayhan, F.; Ahmed, S.; Mahbub, A.; Jani, R.; Shatabda, S.; Farid, D.M. Cusboost: Cluster-based under-sampling with boosting for imbalanced classification. In Proceedings of the 2017 2nd International Conference on Computational Systems and Information Technology for Sustainable Solution (CSITSS), Bengaluru, India, 21–23 December 2017; pp. 1–5. [Google Scholar]
  14. Chen, Y.T.; Shi, J.; Ye, Z.; Mertz, C.; Ramanan, D.; Kong, S. Multimodal object detection via probabilistic ensembling. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 139–158. [Google Scholar]
  15. Ha, Q.; Watanabe, K.; Karasawa, T.; Ushiku, Y.; Harada, T. MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 5108–5115. [Google Scholar]
  16. Cho, Y.; Bianchi-Berthouze, N.; Marquardt, N.; Julier, S.J. Deep thermal imaging: Proximate material type recognition in the wild through deep learning of spatial surface temperature patterns. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, Montreal, QC, Canada, 21–26 April 2018; pp. 1–13. [Google Scholar]
  17. Shivakumar, S.S.; Rodrigues, N.; Zhou, A.; Miller, I.D.; Kumar, V.; Taylor, C.J. Pst900: Rgb-thermal calibration, dataset and segmentation network. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 9441–9447. [Google Scholar]
  18. Sun, Y.; Zuo, W.; Liu, M. Rtfnet: Rgb-thermal fusion network for semantic segmentation of urban scenes. IEEE Robot. Autom. Lett. 2019, 4, 2576–2583. [Google Scholar]
  19. Sun, Y.; Zuo, W.; Yun, P.; Wang, H.; Liu, M. FuseSeg: Semantic segmentation of urban scenes based on RGB and thermal data fusion. IEEE Trans. Autom. Sci. Eng. 2020, 18, 1000–1011. [Google Scholar]
  20. Deng, F.; Feng, H.; Liang, M.; Wang, H.; Yang, Y.; Gao, Y.; Chen, J.; Hu, J.; Guo, X.; Lam, T.L. FEANet: Feature-enhanced attention network for RGB-thermal real-time semantic segmentation. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 4467–4473. [Google Scholar]
  21. Wu, W.; Chu, T.; Liu, Q. Complementarity-aware cross-modal feature fusion network for RGB-T semantic segmentation. Pattern Recognit. 2022, 131, 108881. [Google Scholar]
  22. Xu, J.; Lu, K.; Wang, H. Attention fusion network for multi-spectral semantic segmentation. Pattern Recognit. Lett. 2021, 146, 179–184. [Google Scholar]
  23. Zhang, J.; Liu, H.; Yang, K.; Hu, X.; Liu, R.; Stiefelhagen, R. CMX: Cross-modal fusion for RGB-X semantic segmentation with transformers. IEEE Trans. Intell. Transp. Syst. 2023, 24, 14679–14694. [Google Scholar] [CrossRef]
  24. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1597–1604. [Google Scholar] [CrossRef]
  25. Cho, Y.; Julier, S.J.; Marquardt, N.; Bianchi-Berthouze, N. Robust tracking of respiratory rate in high-dynamic range scenes using mobile thermal imaging. Biomed. Opt. Express 2017, 8, 4480–4503. [Google Scholar] [CrossRef]
  26. Hwang, S.; Park, J.; Kim, N.; Choi, Y.; Kweon, I.S. Multispectral pedestrian detection: Benchmark dataset and baseline. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1037–1045. [Google Scholar]
  27. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A Visible-Infrared Paired Dataset for Low-Light Vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Virtual, 11–17 October 2021. [Google Scholar]
  28. Davis, L.S.; Keck, M. The relationship of visible and thermal infrared imagery. In Proceedings of theIEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05)-Workshops, San Diego, CA, USA, 20–26 June 2005; p. 133. [Google Scholar]
  29. Kim, T.; Shin, S.; Yu, Y.; Kim, H.G.; Ro, Y.M. Causal Mode Multiplexer: A Novel Framework for Unbiased Multispectral Pedestrian Detection. arXiv 2024, arXiv:2403.01300. [Google Scholar]
  30. Tang, H.; Li, Z.; Zhang, D.; He, S.; Tang, J. Divide-and-Conquer: Confluent Triple-Flow Network for RGB-T Salient Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 1958–1974. [Google Scholar] [CrossRef]
  31. Shen, J.; Chen, Y.; Liu, Y.; Zuo, X.; Fan, H.; Yang, W. ICAFusion: Iterative cross-attention guided feature fusion for multispectral object detection. Pattern Recognit. 2024, 145, 109913. [Google Scholar] [CrossRef]
  32. Ouyang, J.; Jin, P.; Wang, Q. Multimodal feature-guided pre-training for RGB-T perception. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 16041–16050. [Google Scholar] [CrossRef]
  33. Yuan, M.; Shi, X.; Wang, N.; Wang, Y.; Wei, X. Improving RGB-infrared object detection with cascade alignment-guided transformer. Inf. Fusion 2024, 105, 102246. [Google Scholar] [CrossRef]
  34. Shin, U.; Lee, K.; Kweon, I.S.; Oh, J. Complementary random masking for RGB-thermal semantic segmentation. arXiv 2023, arXiv:2303.17386. [Google Scholar]
  35. Zhang, J.; Cao, M.; Yang, X.; Xie, W.; Lei, J.; Li, D.; Yang, G.; Huang, W.; Li, Y. Efficientmfd: Towards more efficient multimodal synchronous fusion detection. arXiv 2024, arXiv:2403.09323. [Google Scholar]
  36. Cao, B.; Sun, Y.; Zhu, P.; Hu, Q. Multi-modal Gated Mixture of Local-to-Global Experts for Dynamic Image Fusion. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–3 October 2023. [Google Scholar]
  37. Li, Q.; Zhang, C.; Hu, Q.; Fu, H.; Zhu, P. Confidence-aware fusion using dempster-shafer theory for multispectral pedestrian detection. IEEE Trans. Multimed. 2022, 25, 3420–3431. [Google Scholar] [CrossRef]
  38. Embedded Software Development Tools & Cross Platform IDE|Qt Creator—qt.io. Available online: https://www.qt.io/product/development-tools (accessed on 30 May 2024).
  39. Joshi, J.; Cho, Y. TCC RGB+T Fusion Data Collection & Visualisation Software. 2025. Available online: https://github.com/PhysiologicAILab/rgbt_capture (accessed on 31 March 2025).
  40. Joshi, J.; Wang, K.; Cho, Y. Physiokit: An open-source, low-cost physiological computing toolkit for single-and multi-user studies. Sensors 2023, 23, 8244. [Google Scholar] [CrossRef]
  41. Joshi, J.; Cho, Y. IBVP dataset: RGB-thermal RPPG dataset with high resolution signal quality labels. Electronics 2024, 13, 1334. [Google Scholar] [CrossRef]
  42. CVAT.ai Corporation. Computer Vision Annotation Tool (CVAT); CVAT.ai Corporation: Palo Alto, CA, USA, 2023. [Google Scholar]
  43. Thermal Testing—IDFL—idfl.com. 2023. Available online: https://idfl.com/2022/08/02/thermal-testing/ (accessed on 27 March 2023).
  44. Vollmer, M. Infrared thermal imaging. In Computer Vision: A Reference Guide; Springer: Berlin/Heidelberg, Germany, 2021; pp. 666–670. [Google Scholar]
  45. Alpar, O.; Krejcar, O. Quantization and equalization of pseudocolor images in hand thermography. In Proceedings of the Bioinformatics and Biomedical Engineering: 5th International Work-Conference, IWBBIO 2017, Granada, Spain, 26–28 April 2017; Proceedings, Part I 5. Springer: Berlin/Heidelberg, Germany, 2017; pp. 397–407. [Google Scholar]
  46. Cho, Y.; Bianchi-Berthouze, N.; Julier, S.J. DeepBreath: Deep learning of breathing patterns for automatic stress recognition using low-cost thermal imaging in unconstrained settings. In Proceedings of the 2017 Seventh International Conference on Affective Computing and Intelligent Interaction (ACII), San Antonio, TX, USA, 23–26 October 2017; pp. 456–463. [Google Scholar]
  47. Cho, Y.; Julier, S.J.; Bianchi-Berthouze, N. Instant stress: Detection of perceived mental stress through smartphone photoplethysmography and thermal imaging. JMIR Ment. Health 2019, 6, e10140. [Google Scholar]
  48. Krefer, A.G.; Lie, M.M.I.; Borba, G.B.; Gamba, H.R.; Lavarda, M.D.; de Souza, M.A. A method for generating 3D thermal models with decoupled acquisition. Comput. Methods Programs Biomed. 2017, 151, 79–90. [Google Scholar] [CrossRef] [PubMed]
  49. Takumi, K.; Watanabe, K.; Ha, Q.; Tejero-De-Pablos, A.; Ushiku, Y.; Harada, T. Multispectral object detection for autonomous vehicles. In Proceedings of the Theon Thematic Workshops of ACM Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 35–43. [Google Scholar]
  50. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  51. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  52. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  53. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 8026–8037. [Google Scholar]
  54. Falcon, W.A. Pytorch lightning. GitHub 2019, 3. Available online: https://github.com/Lightning-AI/pytorch-lightning (accessed on 27 March 2023).
  55. MMCV Contributors. OpenMMLab Computer Vision Foundation. 2018. Available online: https://github.com/open-mmlab/mmcv (accessed on 27 March 2023).
  56. Rayhan, F. FR-MRInet: A deep convolutional encoder-decoder for brain tumor segmentation with relu-RGB and sliding-window. Int. J. Comput. Appl. 2018, 975, 8887. [Google Scholar]
  57. Rayhan, F.; Galata, A.; Cootes, T.F. ChoiceNet: CNN learning through choice of multiple feature map representations. Pattern Anal. Appl. 2021, 24, 1757–1767. [Google Scholar] [CrossRef]
  58. Rayhan, F.; Galata, A.; Cootes, T.F. Not all points are created equal-an anisotropic cost function for facial feature landmark location. In Proceedings of the the 31st British Machine Vision Virtual Conference, BMVC, Virtual, 7–10 September 2020. [Google Scholar]
  59. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 7262–7272. [Google Scholar]
  60. Joshi, J.N.; Berthouze, N.; Cho, Y. Self-adversarial Multi-scale Contrastive Learning for Semantic Segmentation of Thermal Facial Images. In Proceedings of the 33rd British Machine Vision Conference 2022, BMVC, London, UK, 21–24 November 2022. [Google Scholar]
Figure 1. Sample RGB and thermal images with the corresponding segmentation masks from the RGBT-Textile.
Figure 1. Sample RGB and thermal images with the corresponding segmentation masks from the RGBT-Textile.
Sensors 25 02306 g001
Figure 2. Data acquisition process. (A) Acquisition protocol; (B) RGBT camera setup; (C) Damaged garment and corresponding labeling with segmentation mask; and (D) Python-QT based application interface for synchronized RGBT acquisition.
Figure 2. Data acquisition process. (A) Acquisition protocol; (B) RGBT camera setup; (C) Damaged garment and corresponding labeling with segmentation mask; and (D) Python-QT based application interface for synchronized RGBT acquisition.
Sensors 25 02306 g002
Figure 3. Overview of ThermoFreq Normalization Process.
Figure 3. Overview of ThermoFreq Normalization Process.
Sensors 25 02306 g003
Figure 4. Mean and Standard deviation of MF dataset. Plot on the left shows the mean and std before applying ThermoFreq, and the right shows the mean and std after applying ThermoFreq.
Figure 4. Mean and Standard deviation of MF dataset. Plot on the left shows the mean and std before applying ThermoFreq, and the right shows the mean and std after applying ThermoFreq.
Sensors 25 02306 g004
Figure 5. Thermal images from the MF (left) and PST (right) datasets, illustrating the visual effects of different cutoff distances D 0 .
Figure 5. Thermal images from the MF (left) and PST (right) datasets, illustrating the visual effects of different cutoff distances D 0 .
Sensors 25 02306 g005
Figure 6. Three dimensional plots for thermal images from the MF (top) and PST (bottom) datasets, to visualize the differences in the distribution of temperature values across the spatial plane, as a result of applying different cutoff distances D 0 . The brighter yellow shades indicate high temperature values, while the darker blue shades indicate low temperature values.
Figure 6. Three dimensional plots for thermal images from the MF (top) and PST (bottom) datasets, to visualize the differences in the distribution of temperature values across the spatial plane, as a result of applying different cutoff distances D 0 . The brighter yellow shades indicate high temperature values, while the darker blue shades indicate low temperature values.
Sensors 25 02306 g006
Table 1. Damage characteristics in the RGBT-Textile dataset, including size ranges, material types, thermal variations, and occurrence frequencies. Thermal response variability ( δ T = 0.5–3.0 °C) complicates segmentation due to overlapping temperature profiles [16].
Table 1. Damage characteristics in the RGBT-Textile dataset, including size ranges, material types, thermal variations, and occurrence frequencies. Thermal response variability ( δ T = 0.5–3.0 °C) complicates segmentation due to overlapping temperature profiles [16].
Damage TypeSize Range (mm2)Material TypeThermal Variation (°C)Occurrence Frequency (%)
Holes5–50Cotton, Blended1.2–2.530
Stains10–120Polyester, Synthetic0.5–1.825
Tears15–150All materials1.0–3.020
Stitching Errors5–30Cotton, Synthetic0.7–1.515
Discolorations20–100Blended, Polyester0.6–1.210
Table 2. Comparison of mean Intersection over Union (mIoU) for RGBT datasets using CRM_RGBTSeg and MMSFormer, both with and without normalization. Results highlight that normalization consistently improves the performance. The performance of CNN-based UNet models are additionally shown to compare against transformer-based models.
Table 2. Comparison of mean Intersection over Union (mIoU) for RGBT datasets using CRM_RGBTSeg and MMSFormer, both with and without normalization. Results highlight that normalization consistently improves the performance. The performance of CNN-based UNet models are additionally shown to compare against transformer-based models.
ModelCRM_RGBTSegMMS-
Former-84
UNet
Backbone Swin Base Swin Small Swin Tiny ResNet18 ResNet50
Normalization
Datasets
KAIST Multispectral Pedestrian0.5520.5670.5440.5600.5120.5540.5520.5650.5500.555
LLVIP Dataset0.6820.7010.6800.6920.6640.6790.6630.6750.6700.673
MF Dataset0.5780.5860.5720.5820.5560.5700.5780.5890.5770.580
OSU Thermal Pedestrian0.9800.9940.9800.9940.9700.9940.9800.9920.9900.993
McubeS0.5230.5450.5190.5420.4970.5300.5040.5220.5120.520
PST9000.8770.8890.8710.8850.8490.8800.8740.8550.8720.875
RGBT-Textile (Ours)0.8420.8460.8400.8430.8220.8350.8420.8520.8400.845
Table 3. Comparison of mean Intersection over Union (mIoU) for RGBT datasets using CRM_RGBTSeg, MMSFormer, and UNet-ResNet18/ResNet50, with different t thresholds (t = 0, 0.5, 1, 5, 10). The results highlight the performance improvements with normalization for both models across different datasets, indicating their robustness and accuracy in RGBT semantic segmentation tasks.
Table 3. Comparison of mean Intersection over Union (mIoU) for RGBT datasets using CRM_RGBTSeg, MMSFormer, and UNet-ResNet18/ResNet50, with different t thresholds (t = 0, 0.5, 1, 5, 10). The results highlight the performance improvements with normalization for both models across different datasets, indicating their robustness and accuracy in RGBT semantic segmentation tasks.
Dataset\ModelCRM_RGBTSegMMSFormerUnet ResNet18Unet ResNet50
Tolerance t = 0 t = 0.5 t = 1 t = 5 t = 10 t = 0 t = 0.5 t = 1 t = 5 t = 10 t = 0 t = 0.5 t = 1 t = 5 t = 10 t = 0 t = 0.5 t = 1 t = 5 t = 10
KAIST Multispectral Pedestrian0.5520.5630.5640.4630.3670.5520.5610.5620.4640.3650.5500.5600.5610.4600.3650.5550.5650.5660.4620.368
LLVIP Dataset0.6820.6970.6980.6000.5010.6630.6710.6720.5740.4750.6700.6840.6850.5890.4820.6730.6880.6890.5900.485
MFNet Dataset0.5560.5780.5790.4800.3810.5780.5850.5860.4870.3890.5770.5950.5960.4890.3870.5800.5980.5990.4910.390
OSU Thermal Pedestrian0.9800.9900.9910.7930.6940.9800.9880.9890.7910.6920.9900.9930.9940.7900.6930.9930.9960.9960.7920.695
McubeS0.5230.5410.5420.4440.3450.5040.5180.5190.4210.3220.5120.5290.5310.4220.3270.5200.5350.5370.4250.330
PST9000.8770.8850.8680.7480.6390.8740.8810.8720.7290.6250.8720.8830.8700.7350.6280.8750.8860.8730.7370.630
Dataset (Ours)0.8220.8420.8430.7450.6460.8420.8480.8490.7510.6520.8400.8560.8570.7490.6500.8450.8610.8620.7540.655
Table 4. Performance comparison of CRM_RGBTSeg, MMSFormer and U-Net models on RGBT-Textile dataset for different t thresholds ( t = 0 , 0.5 , 1 , 5 , 10 ). Metrics include mean Intersection over Union (mIoU), precision, and recall for the entire dataset and individual classes (background, garment, damaged area).
Table 4. Performance comparison of CRM_RGBTSeg, MMSFormer and U-Net models on RGBT-Textile dataset for different t thresholds ( t = 0 , 0.5 , 1 , 5 , 10 ). Metrics include mean Intersection over Union (mIoU), precision, and recall for the entire dataset and individual classes (background, garment, damaged area).
Performance Metrics\ModelCRM_RGBTSegMMSFormerUNet-ResNet18UNet-ResNet50
Tolerance t = 0 t = 0.5 t = 1 t = 5 t = 10 t = 0 t = 0.5 t = 1 t = 5 t = 10 t = 0 t = 0.5 t = 1 t = 5 t = 10 t = 0 t = 0.5 t = 1 t = 5 t = 10
mIoU/Whole dataset0.8220.8420.8410.8300.8200.8100.8280.8270.8170.8080.8400.8560.8570.7490.6500.8450.8610.8620.7540.655
Precision/Whole dataset0.8300.8500.8490.8380.8280.8180.8350.8340.8240.8150.8480.8640.8650.7570.6560.8530.8700.8710.7590.661
Recall/Whole dataset0.8150.8350.8340.8230.8150.8030.8200.8190.8110.8030.8320.8490.8500.7380.6460.8370.8540.8550.7420.650
mIoU/Background0.8500.8700.8690.8580.8500.8360.8540.8530.8450.8380.8600.8780.8790.7690.6610.8650.8820.8830.7730.665
mIoU/Garment0.8100.8300.8290.8200.8100.7960.8130.8120.8080.7980.8250.8420.8430.7400.6330.8300.8480.8490.7450.640
mIoU/Damaged area0.5800.6450.6400.6300.6200.5650.6290.6230.6170.6080.5900.6550.6600.5500.4480.6050.6700.6750.5550.452
Table 5. Performance comparison (mIoU) on original vs. thermally-augmented datasets. Models trained with ThermoFreq show greater robustness to synthetic noise/occlusions compared to baselines.
Table 5. Performance comparison (mIoU) on original vs. thermally-augmented datasets. Models trained with ThermoFreq show greater robustness to synthetic noise/occlusions compared to baselines.
DatasetCRM_RGBTSegMMSFormerUNet-ResNet50
Orig Aug Orig Aug Orig Aug
KAIST Multispectral Pedestrian0.5670.5410.5650.5380.5550.512
LLVIP Dataset0.7010.6830.6750.6510.6730.634
MF Dataset0.5860.5690.5890.5630.5800.547
RGBT-Textile (Ours)0.8460.8430.8520.8310.8450.809
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rayhan, F.; Joshi, J.; Ren, G.; Hernandez, L.; Petreca, B.; Baurley, S.; Berthouze, N.; Cho, Y. Advancing Textile Damage Segmentation: A Novel RGBT Dataset and Thermal Frequency Normalization. Sensors 2025, 25, 2306. https://doi.org/10.3390/s25072306

AMA Style

Rayhan F, Joshi J, Ren G, Hernandez L, Petreca B, Baurley S, Berthouze N, Cho Y. Advancing Textile Damage Segmentation: A Novel RGBT Dataset and Thermal Frequency Normalization. Sensors. 2025; 25(7):2306. https://doi.org/10.3390/s25072306

Chicago/Turabian Style

Rayhan, Farshid, Jitesh Joshi, Guangyu Ren, Lucie Hernandez, Bruna Petreca, Sharon Baurley, Nadia Berthouze, and Youngjun Cho. 2025. "Advancing Textile Damage Segmentation: A Novel RGBT Dataset and Thermal Frequency Normalization" Sensors 25, no. 7: 2306. https://doi.org/10.3390/s25072306

APA Style

Rayhan, F., Joshi, J., Ren, G., Hernandez, L., Petreca, B., Baurley, S., Berthouze, N., & Cho, Y. (2025). Advancing Textile Damage Segmentation: A Novel RGBT Dataset and Thermal Frequency Normalization. Sensors, 25(7), 2306. https://doi.org/10.3390/s25072306

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop