Next Article in Journal
SAM for Road Object Segmentation: Promising but Challenging
Previous Article in Journal
Biomechanics of Spiral Fractures: Investigating Periosteal Effects Using Digital Image Correlation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CSANet: Context–Spatial Awareness Network for RGB-T Urban Scene Understanding

1
School of Electronic Information, Xijing University, Xijing Road, Chang’an District, Xi’an 710123, China
2
School of Computer Science, Northwestern Polytechnical University, Dongxiang Road, Chang’an District, Xi’an 710129, China
3
School of Artificial Intelligence, Tianjin University of Science and Technology, Dagu South Road, Hexi District, Tianjin 300457, China
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(6), 188; https://doi.org/10.3390/jimaging11060188
Submission received: 9 April 2025 / Revised: 22 May 2025 / Accepted: 4 June 2025 / Published: 9 June 2025

Abstract

:
Semantic segmentation plays a critical role in understanding complex urban environments, particularly for autonomous driving applications. However, existing approaches face significant challenges under low-light and adverse weather conditions. To address these limitations, we propose CSANet (Context Spatial Awareness Network), a novel framework that effectively integrates RGB and thermal infrared (TIR) modalities. CSANet employs an efficient encoder to extract complementary local and global features, while a hierarchical fusion strategy is adopted to selectively integrate visual and semantic information. Notably, the Channel–Spatial Cross-Fusion Module (CSCFM) enhances local details by fusing multi-modal features, and the Multi-Head Fusion Module (MHFM) captures global dependencies and calibrates multi-modal information. Furthermore, the Spatial Coordinate Attention Mechanism (SCAM) improves object localization accuracy in complex urban scenes. Evaluations on benchmark datasets (MFNet and PST900) demonstrate that CSANet achieves state-of-the-art performance, significantly advancing RGB-T semantic segmentation.

1. Introduction

Semantic segmentation, a vital branch of computer vision (CV), aims to classify each pixel in an image into a specific category [1,2]. It has been widely applied in fields such as autonomous driving [3], urban planning [4], and intelligent transportation [5]. However, existing optical image-based semantic segmentation methods face significant challenges in autonomous driving scenarios: (1) segmentation performance often degrades in complex street scenes, particularly for small objects [6,7,8]; and (2) segmentation accuracy is severely impacted under adverse conditions such as low light, fog, rain, and snow [9,10,11]. Advancements in sensor technology have improved access to infrared waveband information, making thermal imaging a robust solution for perception and understanding in driving environments. Thermal infrared (TIR) imaging provides unique advantages for semantic segmentation in urban environments, such as robustness to variable lighting conditions and the ability to detect heat-emitting objects regardless of ambient illumination. However, TIR sensors also have intrinsic limitations. Their sensitivity to surface emissivity can cause material-dependent variations in thermal signatures, while environmental temperature changes and sensor-inherent thermal noise can further affect image quality and reliability. Moreover, TIR images often lack fine texture details present in RGB images, potentially reducing segmentation accuracy for small or visually similar objects. These factors highlight the necessity for carefully designed fusion strategies to fully leverage the complementary properties of RGB and TIR modalities in challenging real-world scenarios [12,13,14]. The primary challenge in multi-modal image processing lies in the effective extraction and fusion of features from different modalities.
Currently, multi-modal feature fusion paradigms can be categorized into three types, as illustrated in Figure 1: (a) direct feature fusion in the decoder module [15,16]; (b) parallel feature extraction and fusion during processing [17,18,19]; and (c) separate stages for feature extraction, fusion, and decoding [20,21]. While these paradigms are widely applied in RGB-T semantic segmentation, the first paradigm struggles to ensure feature complementarity, and the second risks feature confusion. To address these limitations, we adopt the third paradigm, which carefully integrates the unique characteristics of multi-modal features. This approach not only facilitates effective feature complementarity in subsequent fusion stages but also exploits the synergy of features at different levels, thereby improving overall model performance. Furthermore, as shown in Figure 1d, multi-modal feature extraction strategies can be classified into two categories: Siamese structures with shared parameters [22,23] and non-Siamese structures without shared parameters [16,24]. Given that RGB images provide color information of natural scenes while TIR images capture infrared waveband data, shared parameter structures often fail to effectively exploit their complementary properties. Existing studies demonstrate that convolutional neural networks (CNNs) excel at extracting local fine-grained features [25], whereas Transformer models leverage self-attention mechanisms to capture long-range dependencies [26]. Based on these observations, we propose a non-Siamese hybrid structure combining CNNs and Transformers to extract and integrate multi-modal features from RGB-T data effectively.
To fuse feature maps of different modalities and scales, existing feature fusion strategies include element-wise addition, element-wise multiplication, feature concatenation, and attention mechanisms. Element-wise addition [27] directly merges features, which enhances model robustness, while element-wise multiplication [28] emphasizes feature selection and shared information. Feature concatenation [29] preserves the integrity of original features, enabling the model to effectively understand the semantic information of multi-modal data. However, as shown in Figure 2, these three fusion methods are prone to feature redundancy or information loss. As a novel feature fusion approach, the attention mechanism [30] extracts important regional features and achieves fine-grained fusion. Therefore, in this study, we design attention mechanisms to achieve efficient fusion of multi-modal features. Additionally, feature extraction in the model can be categorized into shallow local features and deep semantic features. As the network depth increases, the size of feature maps gradually decreases [31]. To enhance model performance, shallow and deep features must be selectively fused to mitigate redundant feature interference. Based on these insights, CSANet incorporates the Channel-Space Cross-Fusion Module (CSCFM) and the Multi-Head Fusion Module (MHFM) to address the enhancement and fusion of shallow local features and deep semantic information. Meanwhile, the Spatial Coordinate Attention Mechanism (SCAM) is introduced to improve the localization precision of moving objects in complex urban scenes. The contributions of this study are summarized as follows:
  • By fully exploring the feature differences between multi-modal data, we propose a novel Context Spatial Awareness Network (CSANet) for RGB-T semantic segmentation, achieving state-of-the-art performance on the MFNet and PST900 datasets.
  • We design the CSCFM to extract shallow fine-grained local features from multi-modal data and the MHFM to capture and enhance deep semantic feature information. A cross-stage, multi-scale hierarchical framework is employed to efficiently fuse features across modalities.
  • The SCAM is introduced into the encoder structure of CSANet to improve the localization accuracy of moving objects in complex scenes. Additionally, a lightweight decoder framework is used to restore multi-modal feature resolution while reducing computational complexity.
The remainder of this article is organized as follows: Section 2 reviews the related work. Section 3 presents the proposed CSANet framework. Section 4 discusses experimental analyses, and Section 5 concludes the study.

2. Related Works

2.1. RGB Semantic Segmentation

With the development of convolutional neural networks (CNNs), Long et al. [32] first introduced fully convolutional networks (FCNs) for semantic segmentation, proposing an end-to-end, pixel-to-pixel training approach. To enhance the extraction of image details and edge information, Noh et al. [1] proposed a deconvolution network for semantic segmentation tasks. However, early semantic segmentation methods faced challenges such as high memory consumption, low computational efficiency, and limited accuracy. To address these issues, Badrinarayanan et al. [33] introduced SegNet, an encoder–decoder architecture designed to progressively restore pixel details. Early models often relied on relatively shallow networks, such as VGGNet [31], but deeper networks introduced issues like gradient vanishing or explosion. To overcome this, He et al. [26] proposed ResNet, utilizing residual learning to enable deeper networks and improve learning capacity. For complex scene segmentation, the DeepLab series [34,35,36,37] employed dilated convolutions with varying rates to expand the receptive field, enabling the capture of multi-scale visual features. The UNet series [38,39,40], on the other hand, utilized contraction and symmetrical extension paths to integrate contextual information. For example, UNet++ [40] improved upon UNet by introducing nested and dense skip connections, further enhancing segmentation performance. The Transformer model [41] revolutionized natural language processing tasks, and Dosovitskiy et al. [42] extended its application to computer vision, demonstrating its versatility. Strudel et al. [43] applied Transformers to semantic segmentation by integrating the Vision Transformer (ViT) as the core component of a segmentation pipeline. ViT [34] processes input images as patch sequences, capturing global dependencies more effectively than traditional CNN-based convolutions, making it advantageous for global feature extraction and enhancement. Despite its strengths, ViT’s effectiveness relies heavily on access to large-scale training datasets, limiting its ability to extract local features as efficiently as CNNs. For RGB-T semantic segmentation, which often involves smaller datasets and requires detailed local visual information in low-illumination conditions, constructing a ViT model suitable for RGB-T remains a challenge. Additionally, the computational resource requirements of ViT pose constraints for deployment in scenarios with limited hardware capabilities. To address these challenges, hybrid models integrating CNNs and Transformers have been proposed. By combining CNNs’ strength in extracting local features with Transformers’ ability to capture global dependencies, these models achieve a balance between computational efficiency and segmentation performance. Furthermore, the introduction of the multi-head attention mechanism significantly reduces the computational complexity of ViT, making it more scalable for resource-limited scenarios while maintaining high performance.

2.2. RGB-T Semantic Segmentation

Thermal infrared (TIR) images capture the thermal information of scene objects, which is perceived by the human brain as spatial information. In depth maps, however, different segmentation regions may share the same depth values without distinct features. Ha et al. [44] first introduced MFNet, which employs a dual-stream architecture to extract RGB-T features, followed by cascaded fusion to accomplish semantic segmentation tasks. RTFNet [17] and FuseSeg [45] adopt architectures similar to MFNet but rely on element-wise addition for feature fusion. Shivakumar et al. [46] proposed a cross-level fusion structure that integrates feature maps from different levels of RGB and TIR images. However, these methods predominantly use simple fusion strategies, such as element-wise addition and concatenation, to capture cross-modal features. Such approaches often result in information redundancy by neglecting the inherent differences between RGB and TIR modalities. More sophisticated RGB-T semantic segmentation methods, such as those in [12,16], emphasize structural classification of features across different stages of the model. By designing hierarchical fusion strategies and integrating multi-level supervision, these methods achieve a refined understanding of spatial and structural information. These approaches not only enhance recognition accuracy for object boundaries and positions but also highlight the significance of layer-specific fusion strategies and multi-task supervision in RGB-T semantic segmentation. Recently, Zhang et al. [14,20,47] combined CNNs with Transformers for feature extraction and fusion, significantly improving multi-modal semantic segmentation accuracy. The research community has increasingly adopted models like Swin Transformer [48] and SegFormer [49] as encoders or for feature fusion, further enhancing RGB-T semantic segmentation performance. These methods integrate traditional CNN-based techniques with Transformer-based attention mechanisms, providing more powerful and nuanced feature representations for complex segmentation tasks. The Swin Transformer [48], with its hierarchical Transformer structure, efficiently processes multi-scale features. By employing a sliding window mechanism, it reduces the computational complexity of self-attention, enabling the model to handle images of varying sizes. SegFormer [49], on the other hand, combines the self-attention mechanism with multi-scale feature extraction. Its lightweight design and strong performance enable it to handle high-resolution inputs while maintaining a low computational cost. These developments represent a significant step forward in RGB-T semantic segmentation, balancing computational efficiency with accuracy.

3. Methodology

To efficiently achieve RGB-T semantic segmentation in urban scenes, we propose a Context Spatial Awareness Network (CSANet), as illustrated in Figure 3. CSANet is designed to effectively learn local and global features from RGB and TIR images while performing fine-grained fusion of multi-modal features. Specifically, CSANet incorporates three key modules: the Channel–Spatial Cross-Fusion Module (CSCFM), which enhances the representation of visual information by fusing multi-modal features; the Multi-Head Fusion Module (MHFM), which performs global feature modeling and multi-modal feature calibration; and the Spatial Coordinate Attention Mechanism (SCAM), which improves the representation of positional information within object regions. Additionally, the final-stage feature fusion promotes interaction between visual and positional information, enhancing the model’s generalization capability.
For the encoder of CSANet, we adopt PVTv2 [50,51] as the backbone network, referred to as PVE. RGB and TIR image pairs (denoted as P RGB and P TIR , respectively) from the same scene are input into the encoder network to extract multi-modal features across four stages without parameter sharing. In the first stage, the feature map size is scaled to 1 / 4 of the original image size. In the subsequent three stages, the feature map size is halved successively. This hierarchical scaling of feature sizes across stages progressively refines features, enabling the model to focus on abstract and semantic representations as the network depth increases. This strategy balances computational efficiency and enhances the extraction of multi-scale feature information. The extracted RGB and TIR features at each stage are represented as follows:
f rgb _ i = PVE P RGB , i = 1 , 2 , 3 , 4 f tir _ i = PVE P TIR , i = 1 , 2 , 3 , 4
where f rgb _ i and f tir _ i represent the RGB and TIR features extracted at the i-th stage of the encoder, respectively. The PVE backbone combines the strengths of CNNs for local feature extraction and Transformers for capturing global context dependencies, as described in [51]. By integrating linear-time attention mechanisms, PVE significantly enhances model performance while balancing computational efficiency. To obtain enhanced fine-grained visual features C i , the RGB and TIR features from the same stage are input into the CSCFM. Spatial attention mechanisms are applied to enrich the TIR features with visual information, effectively addressing segmentation challenges for multi-scale objects in low-illumination conditions. The overall feature extraction process is as follows:
C i = F CSCFM f rgb _ i , f tir _ i , i = 1 , 2 , 3 M i = F MHFM f rgb _ i , f tir _ i , i = 4 S i = F SCAM f tir _ i , i = 1 , 2 , 3 , 4
In the final stage of feature extraction, MHFM is used to fuse RGB and TIR features, enabling refined global feature modeling. Additionally, SCAM is integrated into different stages of the feature extraction process to improve localization accuracy for objects of varying scales, leveraging TIR-based positional information. This multi-stage integration ensures accurate and robust segmentation performance in complex urban scenes. The process of the proposed method is shown in Algorithm 1.
Algorithm 1 CSANet: RGB-T Semantic Segmentation
Require: RGB image P RGB R 3 × H × W , TIR image P TIR R 3 × H × W .
Ensure: Segmentation map M R n _ c l s × H × W .
  1:
Extract multi-stage features for RGB and TIR:
  2:
for each stage i { 1 , 2 , 3 , 4 }  do
  3:
   Extract RGB features: f rgb _ i = PVE ( P RGB , i ) .
  4:
   Extract TIR features: f tir _ i = PVE ( P TIR , i ) .
  5:
end for
  6:
for each stage i { 1 , 2 , 3 }  do
  7:
   Fuse features using CSCFM: C i = F CSCFM ( f rgb _ i , f tir _ i ) .
  8:
   Enhance positional information using SCAM: S i = F SCAM ( f tir _ i ) .
  9:
end for
10:
Fuse features at final stage using MHFM:
11:
M 4 = F MHFM ( f rgb _ 4 , f tir _ 4 ) .
12:
Enhance positional information at final stage:
13:
S 4 = F SCAM ( f tir _ 4 ) .
14:
Combine multi-stage features:
15:
F fuse = { C i , S i } i = 1 3 { M 4 , S 4 } .
16:
Decode fused features:
17:
M = Decoder ( F fuse ) .
18:
return Segmentation map M .

3.1. Channel–Spatial Cross-Fusion Module

The visual features play a crucial role in pixel restoration and model cognition, as the specific category of a pixel is predominantly determined by the visual information. Therefore, the fusion and enhancement of shallow visual features are critical for improving model prediction performance. To fully capture the visual features of RGB-T data, we propose the Channel–Spatial Cross-Fusion Module (CSCFM), whose specific structure is shown in Figure 4.
Specifically, we first perform addition and multiplication operations on the input features f rgb _ i and f tir _ i to obtain f sum _ i and f mul _ i , respectively. The f mul _ i feature highlights shared salient features and compensates for the deficiencies in each modality. Next, f mul _ i is processed using the spatial attention mechanism to extract local details and enhance the connections between different regions within the image. Subsequently, the convolution-processed f sum _ i and f mul _ i are multiplied and input into the channel attention mechanism to produce the fused feature F i . This process adjusts the contribution of feature channels, enhancing the generalization ability and robustness of the model.
F i = CA Conv f sum _ i × SA f mul _ i , i = 1 , 2 , 3
where CA ( · ) and SA ( · ) represent the channel attention and spatial attention functions, respectively. The spatial attention mechanism aims to enhance visual detail features by emphasizing the most significant values in each feature map. Specifically, CSCFM extracts the maximum value of each feature map along the one-dimensional channel to construct the attention map and employs an activation function to prevent gradient explosion, ensuring stable and efficient feature enhancement. The calculation is given by
SA = σ Conv Max f mul _ i
where Conv ( · ) and Max ( · ) denote convolution and maximum pooling operations, respectively. The channel attention mechanism distinguishes the saliency between feature maps to enhance feature extraction and channel modeling capabilities. First, the input feature f R B × C i × H i × W i is reshaped to f ˜ R B × C i × d , where d = H i × W i . The matrix product of f ˜ and its transpose is computed to obtain the relation matrix w ˜ :
w ˜ = f ˜ Transpose f ˜ , w ˜ R B × C i × C i
where w ˜ represents the relation matrix. Next, the maximum relationship value v ˜ of each channel in w ˜ relative to other channels is computed, corresponding to the autocorrelation coefficient of the channel. The maximum relationship value is extended to match the dimension of w ˜ , and an element-wise subtraction is performed to obtain the reverse maximum relationship coefficient matrix E. The enhanced attention-weighted feature f ˜ w is then computed using the softmax function and matrix multiplication:
v ˜ = Max w ˜ ; E = v ˜ w ˜ f ˜ w = E f ˜ ; F i = α f ˜ w + f
where f ˜ w R B × C i × d is restored to the original dimension of f R B × C i × H i × W i , and α is a learnable parameter for the residual connection. The resulting fused feature F i represents the cross-fused output of CSCFM, integrating salient spatial and channel relationships.

3.2. Multi-Head Fusion Module

The thermal features of object regions in TIR images can be perceived as positional information, while the semantic features in RGB images not only represent positional information but also play a critical role in pixel-level feature reconstruction. To fully extract the semantic features of multi-modal data, we construct the Multi-Head Fusion Module (MHFM), which fuses multi-modal visual features to enhance the model’s global understanding and refine semantic segmentation results. The structure of MHFM is shown in Figure 5.
Specifically, MHFM takes the input features { f rgb _ 4 , f tir _ 4 , F 3 } R B × C × H × W , where f rgb _ 4 and f tir _ 4 are the deep features extracted from the fourth stage of RGB and TIR encoders, and F 3 is the fused feature from the previous stage. The input features are first flattened and linearly transformed into { f rgb _ 4 p , f tir _ 4 p , F 3 p } R B × C × H × W , similar to the Q, K, and V matrix transformations in Transformer models. The multi-head attention mechanism is then employed to capture and enhance global contextual features.
W 1 = D P Soft max f mul _ 1 p F 3 p C 4 f mul _ 1 p W 2 = D P ( W 1 ) + f mul _ 1 p
where D P ( · ) denotes the dropout function. The resulting matrix W 2 is processed by a feedforward neural network to obtain the fused feature F 3 fuse , which combines visual and semantic features and enhances positional information for object regions:
F 3 fuse = W 2 + D P L N Linear ( W 2 )
To further integrate visual and positional information, F 3 fuse is combined with F 3 sum and F 3 mul through a weighted multiplication operation:
F 4 = F 3 sum × F 3 fuse + F 3 mul × F 3 fuse
where L N ( · ) represents the layer normalization function. This process ensures that the model can effectively perceive global positional context and spatial correlations, enabling accurate classification of object-region pixels. To restore the segmentation results to the original resolution, F 4 R B × 2 × H × W is upsampled 32 times using linear interpolation. Additionally, a binary image supervision mechanism is introduced to supervise and learn the positional and visual features of the same object region, improving the model’s generalization capability.

3.3. Spatial Coordinate Attention Mechanism

To fully extract and enhance the positional feature information of RGB-T object regions, inspired by the coordinate attention mechanism [52], we propose the Spatial Coordinate Attention Mechanism (SCAM). SCAM embeds spatial position information into the channel attention feature map by leveraging the interdependence between spatial and channel dimensions, resulting in a three-dimensional attention weight. This attention weight enhances the feature representation of different object categories and extracts local detail information, facilitating accurate object region localization. As shown in Figure 6, let F R C × H × W denote the input feature map, where C, H, and W represent the number of channels, height, and width, respectively.
To aggregate the feature map F along the horizontal direction, we apply a 1 × 1 convolution to the X coordinate, which extracts long-distance dependencies in the horizontal dimension while preserving vertical positional information:
z x = Conv 1 × 1 X ( F )
where z x R C × H × 1 represents the X coordinate feature map. Next, z x is processed through a 1 × 1 convolution combined with batch normalization and a nonlinear activation function to produce the feature map f X :
f X = Swish BN Conv 1 × 1 ( z x )
where Swish ( · ) is the nonlinear activation function, BN ( · ) represents batch normalization, and f X R C / r × H × 1 encodes spatial information in the vertical direction, with r being the channel compression ratio. The channel dimension of f X is then restored to match the number of channels in F using another 1 × 1 convolution:
g X = Conv 1 × 1 ( f X )
where g X is the feature map after channel transformation. Similarly, for the Y coordinate, 1 × 1 convolution is applied to aggregate the feature map F along the vertical direction, retaining horizontal positional information while capturing long-distance dependencies in the vertical dimension:
z y = Conv 1 × 1 Y ( F ) f Y = Swish BN Conv 1 × 1 ( z y ) g Y = Conv 1 × 1 ( f Y )
After obtaining g X and g Y , the broadcast addition operation is performed to combine the two features, followed by the Sigmoid function to compute the spatial coordinate weight β :
β = Sigmoid g X g Y
where β R C × H × W represents the spatial coordinate weight. To generate the spatially enhanced feature map, the matrix multiplication of β and F is performed:
F w = β F
where F w is the weighted feature map with spatial information embedding. By embedding the spatial coordinate weights, SCAM effectively captures the object position information in RGB-T data under complex scenes, allowing the model to accurately filter out background pixel interference and enhance object region localization.

3.4. Decoder and Optimization Function

To recover the local details and global context information of RGB-T images, we construct the decoder framework based on UNet, which gradually upsamples the features and fuses the restored features from different stages. Instead of feature concatenation, we adopt feature addition as the fusion strategy to achieve a more complete integration of multi-scale information. This strategy enhances the decoder’s ability to reconstruct visual details and context features in RGB-T data. As shown in Figure 3, the decoder framework consists of four decoding units, each including a dropout layer to alleviate overfitting, two convolutional layers for transforming feature channels, and a bilinear interpolation layer for resizing feature maps. The decoding process of each unit is formally described as
F i = U P Conv Conv D P F i + F i + 1
where U P ( · ) denotes the upsampling function and D P ( · ) represents the dropout operation. For pixel restoration, we use addition to fuse deep semantic features with shallow visual features, improving the model’s prediction accuracy.
For model optimization, we employ a weighted cross-entropy loss function [53] to supervise each segmentation class. As shown in Figure 3, CSANet performs parallel supervised training on both binary images and labeled images, resulting in a total loss function L total composed of L binary and L ann . Considering the singularity of binary image categories, we use weight coefficients to adjust the proportions of different loss functions. For labeled images, the pixel values are proportionally distributed in the range 1 , n , where n is the number of semantic segmentation categories. The weight W i reflects the proportional significance of each object category. In contrast, the pixel value range of binary images is 0 , 1 , which captures the positional information of object regions. The constructed loss functions are defined as follows:
J θ = 1 N i = 1 N P i log P ^ i + 1 P i log 1 P ^ i L total = W 1 × L binary + W 2 × L ann
where J ( θ ) = L binary = L ann , P i 0 , 1 represents the ground truth of the i-th pixel, P ^ i 0 , 1 is the predicted value of the i-th pixel, and N = H × W is the total number of pixels in the image. To better balance the supervision tasks, we set the ratio of W 1 to W 2 as 2:1 during the optimization process. By reasonably assigning these weight coefficients, the optimal model parameters can be effectively obtained. To further clarify, the semantic label loss is used to supervise the network in distinguishing between multiple semantic categories at the pixel level, while the binary mask loss guides the network to distinguish object regions from the background by supervising a binary segmentation map. The motivation for introducing binary supervision is to enhance the network’s ability to localize object boundaries and reinforce the presence of objects, especially in challenging or ambiguous regions. This complementary supervision helps the model achieve more accurate and robust segmentation results.

4. Experiments and Results

4.1. Experimental Protocol

4.1.1. Dataset

To verify the effectiveness of our method, we conducted experimental analysis on two RGB-T datasets: MFNet [44] and PST900 [46]. The MFNet dataset consists of 1569 pairs of RGB and TIR images, with 820 daytime and 749 nighttime scenes. Each image pair has a resolution of 480 × 640 pixels and includes nine object categories: car, person, bike, curve, car stop, guardrail, color cone, and bump. For the experiments, the dataset was split into three subsets, a training set (50%), a testing set (25%), and a validation set (25%), with an equal distribution of daytime and nighttime images in each subset. The PST900 dataset contains 894 pairs of RGB and TIR images with a resolution of 1280 × 720 pixels. It includes five object categories: hand drill, backpack, fire-extinguisher, survivor, and background. The dataset was divided into two subsets, a training set (66.6%) and a validation set (33.3%), ensuring an even distribution of daytime and nighttime images in both subsets. Notably, no preprocessing operations were applied to the RGB and TIR images to ensure the applicability of our CSANet in real-world scenarios.

4.1.2. Evaluation Metrics

To quantitatively analyze the advantages of the proposed method, we adopt two commonly used metrics, mAcc and mIoU , to evaluate the segmentation performance of the model. The mAcc (mean Accuracy) measures the recognition accuracy of the model for each category, while the mIoU (mean Intersection over Union) comprehensively reflects the boundary positioning accuracy of the model across different object regions. The formulations are defined as follows:
mAcc = 1 N i = 1 N T P i T P i + F N i mIoU = 1 N i = 1 N T P i T P i + F P i + F N i
where N denotes the number of object categories. For each category i, T P i , F P i , and F N i represent the true positive, false positive, and false negative counts, respectively.

4.1.3. Implementation Details

The experiments were conducted using an NVIDIA A100 GPU (NVIDIA Corporation, Santa Clara, CA, USA) to implement CSANet within the PyTorch 1.2.0. framework. To improve the robustness of model training, data augmentation strategies, including random cropping and flipping, were applied to RGB and TIR image pairs. The model was optimized using the Ranger algorithm, with the initial learning rate set to 5 × 10 5 . The training process spanned 500 epochs with a batch size of 4, and the momentum parameter was set to 0.9 . For the feature extraction stage, the parameters were adaptively initialized using the pre-trained PVTv2-b5 model weights. During the verification and testing stages, data augmentation strategies were excluded to ensure the fairness and rationality of the experimental evaluation.

4.2. Comparison with State-of-the-Art Methods

4.2.1. Evaluation on the MFNet Dataset

To demonstrate the effectiveness and advantages of the proposed multimodal feature extraction and fusion schemes, we compare CSANet with several state-of-the-art methods on the MFNet dataset. These methods compared include MFNet [44], RTFNet [17], PSTNet [46], MLFNet [18], FuseSeg [45], ABMDRNet [15], FEANet [19], MFFENet [54], GMNet [55], MMNet [56], EGFNet [57], MTANet [16], CCFFNet [58], CCAFFMNet [59], DSGBINet [60], CMXSegF [20], FDCNet [61], ECGFNet [62], MMSMCNet [63], LASNet [12], SFAF-MA [64], DBCNet [65], CAINet [66], U-KAN [67], and U-Mamba [68]. For CSANet, results are reported as mean ± standard deviation (STD) over five independent runs with different random seeds. For other methods, results are taken from the original papers. As shown in the results presented in Table 1 and Table 2, CSANet achieves state-of-the-art performance on the MFNet dataset, significantly outperforming existing methods in both mAcc and mIoU . Specifically, CSANet achieves an mAcc of 79.1% and an mIoU of 62.5%, surpassing the second-best MMSMCNet by 3.9% and 4.4%, respectively. These improvements highlight the effectiveness of CSANet’s multimodal feature extraction and fusion mechanism. For large-scale object categories such as Car and Person, CSANet achieves IoU scores of 85.6% and 91.0%, demonstrating its ability to accurately capture global semantic information and maintain high segmentation accuracy for dominant object regions. Moreover, CSANet excels in segmenting fine-grained and challenging objects. For instance, in the car stop category, CSANet achieves an accuracy of 64.7% and an IoU of 48.5%, which are 1.9% and 5.1% higher than the second-best results, respectively. Similarly, for the bike and guardrail categories, CSANet achieves IoU scores of 71.7% and 44.2%, outperforming other methods by substantial margins. These results demonstrate CSANet’s robustness in handling both large-scale objects and small, complex structures, ensuring accurate segmentation across diverse scene elements. Both U-KAN and U-Mamba exhibit inferior performance on the RGB-T semantic segmentation task, mainly because they lack specialized cross-modal fusion mechanisms and are not tailored to leverage the complementary characteristics of RGB and thermal data.
To further evaluate CSANet’s robustness under varying illumination conditions, Table 3 compares its performance in daytime and nighttime scenes. For daytime scenes, CSANet achieves an mAcc of 76.2% and an mIoU of 61.1%, outperforming the second-best LASNet by 7.3% and 4.8%, respectively. This demonstrates CSANet’s ability to leverage multimodal data effectively in well-illuminated environments, ensuring precise segmentation of both dominant and fine-grained objects. In nighttime scenes, where RGB images suffer from significant visual degradation, CSANet achieves an mAcc of 71.9% and an mIoU of 58.3%, outperforming LASNet by 2.7% and 3.2%, respectively. These results underscore CSANet’s ability to utilize thermal (TIR) data to compensate for the deficiencies in RGB imagery, enabling robust segmentation performance under low-light conditions. Compared to other methods such as MMSMCNet and MTANet, which exhibit noticeable performance degradation in nighttime scenes, CSANet maintains high accuracy and segmentation quality, further validating its robustness in challenging real-world scenarios. For the qualitative analysis, Figure 7 provides qualitative comparisons of CSANet with the top five methods (LASNet, MMSMCNet, MTANet, DBCNet, and MFFENet) on both daytime and nighttime images. The visual results demonstrate CSANet’s superior segmentation capabilities across diverse scenarios. In daytime scenes, CSANet effectively segments both large-scale and fine-grained objects with precise boundary delineation and accurate semantic labeling. For instance, in the second row of Figure 7, CSANet accurately captures the intricate details of pedestrians and bicycles, while other methods either fail to detect these objects or produce incomplete segmentations. The advantages of CSANet are even more pronounced in nighttime scenes. In the last two rows, CSANet successfully segments challenging categories such as guardrail and bike with high precision, while competing methods struggle with incomplete or inaccurate segmentations due to the limitations of RGB imagery under low-light conditions. The integration of TIR data enables CSANet to detect and segment objects that are poorly visible in RGB images, ensuring consistent and robust performance across varying lighting conditions.
The results presented in Table 1, Table 2 and Table 3, along with the qualitative analysis in Figure 7, collectively highlight the strengths of CSANet in both quantitative and qualitative aspects. CSANet’s superior performance can be attributed to its effective multimodal feature extraction and fusion strategy, which ensures the accurate localization and segmentation of objects of varying scales, even under challenging conditions such as low illumination. Its ability to consistently outperform state-of-the-art methods in both daytime and nighttime scenes underscores its robustness and practical applicability to real-world RGB-T semantic segmentation tasks. Furthermore, the significant improvements in fine-grained segmentation tasks, such as bike, guardrail, and car stop, demonstrate CSANet’s ability to handle intricate scene elements, a critical requirement for applications such as autonomous driving and surveillance. The consistent, competitive results achieved by CSANet validate the effectiveness of its multimodal learning strategy and its ability to balance global and local feature representation, making it a promising solution for fine-grained scene understanding.

4.2.2. Evaluation on PST900 Dataset

To evaluate the generalization ability of the proposed method across diverse scenes, we conducted experiments on the PST900 dataset and compare CSANet with state-of-the-art methods, including MFNet [44], PSTNet [17], MFFENet [54], GMNet [55], EGFNet [57], MTANet [16], CCFFNet [58], DSGBINet [60], FDCNet [61], MMSMCNet [63], LASNet [12], DBCNet [65], and CAINet [66]. The quantitative results are summarized in Table 4 and Table 5, while qualitative comparisons of the top three performing methods are shown in Figure 8. As seen in Table 4 and Table 5, CSANet achieves the best overall performance, with an mAcc of 95.57% and an mIoU of 86.01%, outperforming the second-best CAINet by 1.3% in mAcc and 1.28% in mIoU . These results highlight the effectiveness of CSANet in leveraging multimodal RGB-T data to improve segmentation accuracy and robustness. CSANet demonstrates consistent superiority across most categories. Notably, for the small-scale and challenging category of fire extinguisher, CSANet achieves an IoU of 98.41%, which is 6.45% higher than the second-best CCFFNet. This significant improvement underscores the ability of CSANet to accurately segment objects with fine-grained details and complex boundaries, further enhancing overall segmentation performance. In addition to fine-grained categories, CSANet also achieves competitive results for large-scale objects such as backpacks, with an IoU of 89.88%, and for the survivor category, with an IoU of 76.51%, which is 1.02% higher than the third-best LASNet. These results further demonstrate CSANet’s robustness in capturing both global semantic contexts and detailed spatial information, making it suitable for diverse scene understanding tasks. Figure 8 presents qualitative comparisons of CSANet with CAINet, LASNet, and MFFENet. The visual results provide a more intuitive understanding of CSANet’s advantages. In the fire extinguisher category (first row), CSANet effectively delineates the object boundary and produces a more complete segmentation mask compared to other methods, which either over-segment the object or fail to capture its fine-grained details. Similarly, in the survivor category (fourth row), CSANet accurately segments the person in the scene, with minimal boundary errors, whereas CAINet and LASNet exhibit noticeable inaccuracies in the segmentation mask. Furthermore, in complex scenes with multiple objects, such as the backpack category (third row), CSANet demonstrates superior segmentation quality, correctly identifying and separating multiple objects, while other methods produce fragmented or incomplete masks. These qualitative and quantitative analyses collectively demonstrate the robustness and effectiveness of CSANet in RGB-T semantic segmentation tasks, particularly in scenarios involving small-scale objects and intricate boundaries. The ability to consistently outperform existing state-of-the-art methods highlights CSANet’s practical applicability for real-world applications, such as emergency response and autonomous systems, where accurate segmentation of diverse and challenging objects is critical.

4.2.3. Computational Complexity

To evaluate the feasibility of the proposed CSANet for practical applications, we compare its computational complexity with state-of-the-art methods on the MFNet dataset.
Table 6 summarizes the floating-point operations (FLOPs) and model parameters, along with segmentation accuracy metrics ( mAcc and mIoU ) for different methods. CSANet achieves competitive computational efficiency with the 84.32 G FLOPs and 38.36 M parameters, which is significantly lower than methods such as RTFNet (245.71 G FLOPs, 185.24 M parameters) and PSTNet (337.04 G FLOPs, 254.51 M parameters). Notably, while CAINet has the smallest computational cost (12.16 M parameters, 67.49 G FLOPs), its segmentation performance ( mAcc = 73.3%, mIoU = 58.6%) is inferior to that of CSANet ( mAcc = 79.1%, mIoU = 62.5%), indicating that a lower computational cost does not necessarily result in superior segmentation accuracy. Compared to other high-performing models such as LASNet and MMSMCNet, CSANet achieves a balance between computational efficiency and segmentation accuracy. For instance, LASNet has higher FLOPs (233.81 G) and parameters (154.62 M) while achieving lower mAcc (76.8%) and mIoU (60.6%) relative to CSANet. Similarly, MMSMCNet, with 181.82 G FLOPs and 98.58 M parameters, achieves an mIoU of 58.1%, which is 4.4% lower than CSANet. These results emphasize the efficiency of CSANet in balancing model complexity and performance. While CSANet does not achieve the lowest FLOPs or parameter count among all methods, it demonstrates strong competitiveness by achieving state-of-the-art segmentation accuracy. This balance of computational cost and performance makes CSANet a practical choice for real-world applications, particularly in scenarios where hardware resources are constrained but high segmentation accuracy is required. Furthermore, the reduction in computational cost compared to other high-performing methods highlights the scalability of CSANet for deployment in tasks such as autonomous driving and surveillance. To further enhance the deployment of CSANet in resource-constrained environments such as embedded systems, mobile platforms, or real-time applications (e.g., autonomous vehicles and drones), several model optimization strategies can be considered. Techniques such as model pruning, quantization, and knowledge distillation could be employed to further reduce the computational cost and memory footprint of CSANet, enabling efficient inference on hardware with limited resources. Additionally, replacing the current backbone with a lighter variant (e.g., a smaller version of PVT or other efficient architectures) can further improve suitability for embedded deployment.
Given the current model complexity (84.32 G FLOPs, 38.36 M parameters), CSANet is expected to achieve real-time inference on high-end GPUs, with the potential for further acceleration through the aforementioned techniques. On embedded devices, with appropriate model compression and optimization, CSANet could potentially approach or meet the typical real-time requirements (e.g., 30–60 FPS) for practical applications. We consider this an important direction for future work and plan to investigate and report the detailed deployment performance of CSANet on embedded platforms.

4.3. Ablation Study

To evaluate the contributions of each component in CSANet, including the Cross-Stage Cross-Modal Feature Module (CSCFM), the Multi-Head Feature Module (MHFM), and the Spatial Context Attention Module (SCAM), we conducted ablation studies on the MFNet dataset. The results, presented in Table 7, demonstrate the significant improvements each component brings to the overall segmentation performance. The baseline model, achieving an mIoU of 53.2% and an mAcc of 64.8%, struggles to effectively handle small and complex objects, as reflected by the low IoU scores for car stops (25.6%) and poles (28.2%), as well as limited performance in categories with fine-grained boundaries, such as bikes (59.3%). Incorporating CSCFM into the baseline significantly improves the mIoU to 58.7%, with notable gains in challenging categories such as car stops (31.7%, +6.1%) and poles (34.8%, +6.6%). These results highlight CSCFM’s ability to integrate multi-scale and multi-modal features across different stages, addressing the challenges posed by scale variation and object deformation. Additionally, by adding MHFM to the baseline, the mIoU increases to 57.3%, reflecting a 4.1% improvement. This enhancement is particularly evident in categories requiring accurate boundary delineation, such as bikes (61.7%, +2.4%) and persons (89.4%, +1.2%), demonstrating MHFM’s capability to refine fine-grained boundary features. The inclusion of SCAM further boosts the mIoU to 56.8%, with consistent improvements in small-scale object categories like guardrails (35.9%, +3.2%) and car stops (29.7%, +4.1%), showcasing SCAM’s ability to extract spatial position information and enhance feature representations for small or sparsely distributed objects in complex scenes. When all three modules are fully integrated, CSANet achieves the best performance, with an mIoU of 62.5% and an mAcc of 79.1%, representing improvements of 9.3% and 14.3%, respectively, over the baseline. Across all categories, CSANet demonstrates consistent gains, particularly in small and challenging objects. For example, the IoU for car stops improves from 25.6% to 44.2%, while the IoU for poles increases from 28.2% to 39.6%. These results validate the synergistic effects of CSCFM, MHFM, and SCAM in enhancing segmentation performance for both large-scale objects (e.g., car, person) and small-scale or fine-grained objects (e.g., bike, car stop, pole), effectively balancing global semantic understanding and local detail refinement. The visualized results in Figure 9 further corroborate the quantitative findings. Compared to the baseline and ablated variants, CSANet consistently produces more accurate and complete segmentation masks, particularly for objects with complex boundaries or small scales. These results demonstrate that the combination of CSCFM, MHFM, and SCAM allows CSANet to achieve state-of-the-art performance on the MFNet dataset, addressing the challenges of multi-scale object segmentation and small-scale object recognition in RGB-T semantic segmentation tasks. To analyze the contribution of each modality, we further conducted ablation experiments using only the RGB or only the TIR modality as input. As shown in Table 8, using only RGB images, CSANet achieves an mIoU of 56.7% and mAcc of 72.4%. Using only TIR images, the model achieves an mIoU of 54.2% and mAcc of 69.8%. In contrast, the full model utilizing both modalities achieves an mIoU of 62.5% and mAcc of 79.1%. For specific categories such as cars, persons, bikes, guardrails, cars stops, and poles, the multi-modal model consistently outperforms the single-modality variants. These results clearly demonstrate that fusing RGB and thermal modalities significantly improves segmentation accuracy across all object classes, confirming the effectiveness of our multi-modal design. To further validate the effectiveness of our proposed Spatial Coordinate Attention Module (SCAM), we compare it with two widely used alternatives: Coordinate Convolution (CoordConv) and Coordinate Attention (CA). For a fair comparison, we replaced the proposed SCAM with CoordConv and CA modules in our framework and evaluated their performance on the MFNet dataset. As shown in Table 9, CSANet with CoordConv achieves an mIoU of 59.3%, and with CA, it achieves an mIoU of 60.1%, both lower than our SCAM-based model (62.5%). Moreover, our SCAM-based model consistently outperforms the others in terms of both overall (mAcc, mIoU) and per-class IoU metrics, especially for challenging categories such as guardrails, car stops, and poles. This demonstrates the superiority of SCAM in capturing spatial dependencies and enhancing segmentation quality in complex urban scenes.

5. Conclusions

In this study, we present the Context Spatial Awareness Network (CSANet) for RGB-T semantic segmentation tasks. CSANet introduces a novel framework that categorizes multimodal features into visual and semantic features, designing specialized extraction and fusion strategies for each feature type. Three key modules, the Cross-Stage Cross-Modal Feature Module (CSCFM), the Multi-Head Feature Module (MHFM), and the Spatial Context Attention Module (SCAM), are proposed to handle multi-scale features, refine detailed features, and enhance spatial position features, respectively. These modules enable fine-grained feature fusion while addressing the interaction challenges between feature categories and achieving efficient cross-stage feature integration. The effectiveness of CSANet is demonstrated through extensive experiments on MFNet and PST900 datasets, where it achieves state-of-the-art segmentation performance. Ablation studies further validate the contributions of each module, highlighting their complementary roles in improving segmentation accuracy and robustness. Moreover, CSANet demonstrates competitive computational complexity, balancing efficiency and performance, making it suitable for real-world applications. In future work, we plan to extend the application of CSANet to more multimodal semantic segmentation tasks, such as RGB-D or LiDAR-based segmentation, to further evaluate its robustness and generalization across diverse data modalities.

Author Contributions

Conceptualization, R.L. and Z.W.; methodology, R.L. and C.Z.; software, R.L. and J.G.; validation, R.L. and C.Z.; formal analysis, R.L. and J.G.; investigation, Z.W.; resources, Z.W. and C.Z.; data curation, R.L.; original draft preparation, R.L.; review and editing, Z.W. and C.Z.; visualization, Z.W.; supervision, Z.W. and C.Z.; project administration, Z.W. and J.G.; funding acquisition, Z.W. and J.G. All authors have read and agreed on the published version of the manuscript.

Funding

This work was supported in part by the Youth Talent Support Program of Shaanxi Science and Technology Association under Grant 23JK0701, in part by the Xi’an Science and Technology Planning Projects under Grant 20240103, and in part by the China Postdoctoral Science Foundation under Grant 2024M754225.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in https://github.com/darkseid-arch/CSANet (accessed on 8 April 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Noh, H.; Hong, S.; Han, B. Learning deconvolution network for semantic segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1520–1528. [Google Scholar]
  2. Song, Z.; Zhao, L.; Zhou, J. Learning hybrid semantic affinity for point cloud segmentation. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4599–4612. [Google Scholar] [CrossRef]
  3. Feng, D.; Haase-Schütz, C.; Rosenbaum, L.; Hertlein, H.; Glaeser, C.; Timm, F.; Wiesbeck, W.; Dietmayer, K. Deep multi-modal object detection and semantic segmentation for autonomous driving: Datasets, methods, and challenges. IEEE Trans. Intell. Transp. Syst. 2020, 22, 1341–1360. [Google Scholar] [CrossRef]
  4. Neupane, B.; Horanont, T.; Aryal, J. Deep learning-based semantic segmentation of urban features in satellite images: A review and meta-analysis. Remote Sens. 2021, 13, 808. [Google Scholar] [CrossRef]
  5. Li, Y.; Cai, J.; Zhou, Q.; Lu, H. Joint semantic-instance segmentation method for intelligent transportation system. IEEE Trans. Intell. Transp. Syst. 2022, 24, 15540–15547. [Google Scholar] [CrossRef]
  6. Kampffmeyer, M.; Salberg, A.B.; Jenssen, R. Semantic segmentation of small objects and modeling of uncertainty in urban remote sensing images using deep convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1–9. [Google Scholar]
  7. Dong, R.; Pan, X.; Li, F. DenseU-net-based semantic segmentation of small objects in urban remote sensing images. IEEE Access 2019, 7, 65347–65356. [Google Scholar] [CrossRef]
  8. Ma, A.; Wang, J.; Zhong, Y.; Zheng, Z. FactSeg: Foreground activation-driven small object semantic segmentation in large-scale remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5606216. [Google Scholar] [CrossRef]
  9. Pfeuffer, A.; Dietmayer, K. Robust semantic segmentation in adverse weather conditions by means of sensor data fusion. In Proceedings of the 2019 22th International Conference on Information Fusion (FUSION), Ottawa, ON, Canada, 2–5 July 2019; pp. 1–8. [Google Scholar]
  10. Valada, A.; Vertens, J.; Dhall, A.; Burgard, W. Adapnet: Adaptive semantic segmentation in adverse environmental conditions. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 4644–4651. [Google Scholar]
  11. Ravishankar, H.; Venkataramani, R.; Thiruvenkadam, S.; Sudhakar, P.; Vaidya, V. Learning and incorporating shape models for semantic segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 11–13 September 2017; pp. 203–211. [Google Scholar]
  12. Li, G.; Wang, Y.; Liu, Z.; Zhang, X.; Zeng, D. RGB-T semantic segmentation with location, activation, and sharpening. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 1223–1235. [Google Scholar] [CrossRef]
  13. Shin, U.; Lee, K.; Kweon, I.S.; Oh, J. Complementary random masking for rgb-thermal semantic segmentation. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 11110–11117. [Google Scholar]
  14. Zhang, J.; Liu, R.; Shi, H.; Yang, K.; Reiß, S.; Peng, K.; Fu, H.; Wang, K.; Stiefelhagen, R. Delivering arbitrary-modal semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 1136–1147. [Google Scholar]
  15. Zhang, Q.; Zhao, S.; Luo, Y.; Zhang, D.; Huang, N.; Han, J. ABMDRNet: Adaptive-weighted bi-directional modality difference reduction network for RGB-T semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2633–2642. [Google Scholar]
  16. Zhou, W.; Dong, S.; Lei, J.; Yu, L. MTANet: Multitask-aware network with hierarchical multimodal fusion for RGB-T urban scene understanding. IEEE Trans. Intell. Veh. 2022, 8, 48–58. [Google Scholar] [CrossRef]
  17. Sun, Y.; Zuo, W.; Liu, M. RTFNet: RGB-thermal fusion network for semantic segmentation of urban scenes. IEEE Robot. Autom. Lett. 2019, 4, 2576–2583. [Google Scholar] [CrossRef]
  18. Guo, Z.; Li, X.; Xu, Q.; Sun, Z. Robust semantic segmentation based on RGB-thermal in variable lighting scenes. Measurement 2021, 186, 110176. [Google Scholar] [CrossRef]
  19. Deng, F.; Feng, H.; Liang, M.; Wang, H.; Yang, Y.; Gao, Y.; Chen, J.; Hu, J.; Guo, X.; Lam, T.L. FEANet: Feature-enhanced attention network for RGB-thermal real-time semantic segmentation. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 4467–4473. [Google Scholar]
  20. Zhang, J.; Liu, H.; Yang, K.; Hu, X.; Liu, R.; Stiefelhagen, R. CMX: Cross-modal fusion for RGB-X semantic segmentation with transformers. IEEE Trans. Intell. Transp. Syst. 2023, 24, 14679–14694. [Google Scholar] [CrossRef]
  21. Fu, Y.; Chen, Q.; Zhao, H. CGFNet: Cross-guided fusion network for RGB-thermal semantic segmentation. Vis. Comput. 2022, 38, 3243–3252. [Google Scholar] [CrossRef]
  22. Liu, J.; Zhou, W.; Cui, Y.; Yu, L.; Luo, T. GCNet: Grid-like context-aware network for RGB-thermal semantic segmentation. Neurocomputing 2022, 506, 60–67. [Google Scholar] [CrossRef]
  23. Fu, K.; Fan, D.P.; Ji, G.P.; Zhao, Q.; Shen, J.; Zhu, C. Siamese network for RGB-D salient object detection and beyond. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5541–5559. [Google Scholar] [CrossRef]
  24. Fu, K.; Fan, D.P.; Ji, G.P.; Zhao, Q. JL-DCF: Joint learning and densely-cooperative fusion framework for RGB-D salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3052–3062. [Google Scholar]
  25. Mustafa, H.T.; Yang, J.; Zareapoor, M. Multi-scale convolutional neural network for multi-focus image fusion. Image Vis. Comput. 2019, 85, 26–35. [Google Scholar] [CrossRef]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  27. Zhang, Z.; Zhang, X.; Peng, C.; Xue, X.; Sun, J. Exfuse: Enhancing feature fusion for semantic segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 269–284. [Google Scholar]
  28. Sun, J.; Li, Y. Multi-feature fusion network for road scene semantic segmentation. Comput. Electr. Eng. 2021, 92, 107155. [Google Scholar] [CrossRef]
  29. Li, X.; Zhao, H.; Han, L.; Tong, Y.; Tan, S.; Yang, K. Gated fully fusion for semantic segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 11418–11425. [Google Scholar]
  30. Niu, Z.; Zhong, G.; Yu, H. A review on the attention mechanism of deep learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  31. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  32. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  33. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  34. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv 2014, arXiv:1412.7062. [Google Scholar]
  35. Yurtkulu, S.C.; Şahin, Y.H.; Unal, G. Semantic segmentation with extended DeepLabv3 architecture. In Proceedings of the 2019 27th Signal Processing and Communications Applications Conference (SIU), Sivas, Turkey, 24–26 April 2019; pp. 1–4. [Google Scholar]
  36. Liu, H.; Xu, Z.; Xu, B. Research on improved DeepLabv3+ image Semantic Segmentation algorithm. In Proceedings of the 7th International Conference on Control Engineering and Artificial Intelligence, Sanya, China, 28–30 January 2023; pp. 137–142. [Google Scholar]
  37. Liu, G.; Chai, Z. Image semantic segmentation based on improved DeepLabv3+ network and superpixel edge optimization. J. Electron. Imaging 2022, 31, 013011. [Google Scholar] [CrossRef]
  38. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015, Proceedings, Part III 18; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  39. Alom, M.Z.; Yakopcic, C.; Hasan, M.; Taha, T.M.; Asari, V.K. Recurrent residual U-Net for medical image segmentation. J. Med. Imaging 2019, 6, 014006. [Google Scholar] [CrossRef] [PubMed]
  40. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018, Proceedings 4; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  41. Vaswani, A. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  42. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  43. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 7262–7272. [Google Scholar]
  44. Ha, Q.; Watanabe, K.; Karasawa, T.; Ushiku, Y.; Harada, T. MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 5108–5115. [Google Scholar]
  45. Sun, Y.; Zuo, W.; Yun, P.; Wang, H.; Liu, M. FuseSeg: Semantic segmentation of urban scenes based on RGB and thermal data fusion. IEEE Trans. Autom. Sci. Eng. 2020, 18, 1000–1011. [Google Scholar] [CrossRef]
  46. Shivakumar, S.S.; Rodrigues, N.; Zhou, A.; Miller, I.D.; Kumar, V.; Taylor, C.J. Pst900: Rgb-thermal calibration, dataset and segmentation network. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 9441–9447. [Google Scholar]
  47. Li, H.; Sun, Y. IGFNet: Illumination-Guided Fusion Network for Semantic Scene Understanding using RGB-Thermal Images. In Proceedings of the 2023 IEEE International Conference on Robotics and Biomimetics (ROBIO), Koh Samui, Thailand, 4–9 December 2023; pp. 1–6. [Google Scholar]
  48. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  49. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  50. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 568–578. [Google Scholar]
  51. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pvt v2: Improved baselines with pyramid vision transformer. Comput. Vis. Media 2022, 8, 415–424. [Google Scholar] [CrossRef]
  52. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  53. Phan, T.H.; Yamamoto, K. Resolving class imbalance in object detection with weighted cross entropy losses. arXiv 2020, arXiv:2006.01413. [Google Scholar]
  54. Zhou, W.; Lin, X.; Lei, J.; Yu, L.; Hwang, J.N. MFFENet: Multiscale feature fusion and enhancement network for RGB–thermal urban road scene parsing. IEEE Trans. Multimed. 2021, 24, 2526–2538. [Google Scholar] [CrossRef]
  55. Zhou, W.; Liu, J.; Lei, J.; Yu, L.; Hwang, J.N. GMNet: Graded-feature multilabel-learning network for RGB-thermal urban scene semantic segmentation. IEEE Trans. Image Process. 2021, 30, 7790–7802. [Google Scholar] [CrossRef]
  56. Lan, X.; Gu, X.; Gu, X. MMNet: Multi-modal multi-stage network for RGB-T image semantic segmentation. Appl. Intell. 2022, 52, 5817–5829. [Google Scholar] [CrossRef]
  57. Zhou, W.; Dong, S.; Xu, C.; Qian, Y. Edge-aware guidance fusion network for rgb–thermal scene parsing. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; Volume 36, pp. 3571–3579. [Google Scholar]
  58. Wu, W.; Chu, T.; Liu, Q. Complementarity-aware cross-modal feature fusion network for RGB-T semantic segmentation. Pattern Recognit. 2022, 131, 108881. [Google Scholar] [CrossRef]
  59. Yi, S.; Li, J.; Liu, X.; Yuan, X. CCAFFMNet: Dual-spectral semantic segmentation network with channel-coordinate attention feature fusion module. Neurocomputing 2022, 482, 236–251. [Google Scholar] [CrossRef]
  60. Xu, C.; Li, Q.; Jiang, X.; Yu, D.; Zhou, Y. Dual-space graph-based interaction network for RGB-thermal semantic segmentation in electric power scene. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 1577–1592. [Google Scholar] [CrossRef]
  61. Zhao, S.; Zhang, Q. A feature divide-and-conquer network for RGB-T semantic segmentation. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 2892–2905. [Google Scholar] [CrossRef]
  62. Zhou, W.; Lv, Y.; Lei, J.; Yu, L. Embedded control gate fusion and attention residual learning for RGB–thermal urban scene parsing. IEEE Trans. Intell. Transp. Syst. 2023, 24, 4794–4803. [Google Scholar] [CrossRef]
  63. Zhou, W.; Zhang, H.; Yan, W.; Lin, W. MMSMCNet: Modal memory sharing and morphological complementary networks for RGB-T urban scene semantic segmentation. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 7096–7108. [Google Scholar] [CrossRef]
  64. He, X.; Wang, M.; Liu, T.; Zhao, L.; Yue, Y. SFAF-MA: Spatial feature aggregation and fusion with modality adaptation for RGB-thermal semantic segmentation. IEEE Trans. Instrum. Meas. 2023, 72, 5012810. [Google Scholar] [CrossRef]
  65. Zhou, W.; Gong, T.; Lei, J.; Yu, L. DBCNet: Dynamic bilateral cross-fusion network for RGB-T urban scene understanding in intelligent vehicles. IEEE Trans. Syst. Man Cybern. Syst. 2023, 53, 7631–7641. [Google Scholar] [CrossRef]
  66. Lv, Y.; Liu, Z.; Li, G. Context-aware interaction network for rgb-t semantic segmentation. IEEE Trans. Multimed. 2024, 26, 6348–6360. [Google Scholar] [CrossRef]
  67. Li, C.; Liu, X.; Li, W.; Wang, C.; Liu, H.; Liu, Y.; Chen, Z.; Yuan, Y. U-kan makes strong backbone for medical image segmentation and generation. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 4652–4660. [Google Scholar]
  68. Ma, J.; Li, F.; Wang, B. U-mamba: Enhancing long-range dependency for biomedical image segmentation. arXiv 2024, arXiv:2401.04722. [Google Scholar]
Figure 1. Different multi-modal feature extraction and fusion strategies. (a) Direct feature fusion in the decoder module; (b) Parallel feature extraction and fusion during processing; (c) Separate stages for feature extraction, fusion, and decoding; (d) Comparison between non-Siamese and Siamese structures for multi-modal feature extraction.
Figure 1. Different multi-modal feature extraction and fusion strategies. (a) Direct feature fusion in the decoder module; (b) Parallel feature extraction and fusion during processing; (c) Separate stages for feature extraction, fusion, and decoding; (d) Comparison between non-Siamese and Siamese structures for multi-modal feature extraction.
Jimaging 11 00188 g001
Figure 2. Feature visualization of different feature fusion methods, including (a) original images, (b) feature redundancy, (c) feature loss, (d) effective feature.
Figure 2. Feature visualization of different feature fusion methods, including (a) original images, (b) feature redundancy, (c) feature loss, (d) effective feature.
Jimaging 11 00188 g002
Figure 3. The architecture of our CSANet. The left side comprises the Encoder module, while the right side encompasses thedecoder module. The encoder module integrates different feature extraction and fusion modules to enhance visual and positional information. The decoder module employs multiple supervisions to improve the accuracy of model prediction. Upon feeding multi-modal images into the encoder to extract hierarchical features T / R 1 , T / R 2 , T / R 3 , T / R 4 , the encoder includes the modules CSCFM, MHFM, and SCAM, designed to amalgamate these hierarchical features into F 1 , F 2 , F 3 , and F 4 , respectively. Subsequently, the fused features F i are inputted into the decoder for progressive pixel restoration. The framework aims to enhance the effectiveness of the encoder–decoder structure in RGB-T semantic segmentation tasks by using multi-modal feature complementarity.
Figure 3. The architecture of our CSANet. The left side comprises the Encoder module, while the right side encompasses thedecoder module. The encoder module integrates different feature extraction and fusion modules to enhance visual and positional information. The decoder module employs multiple supervisions to improve the accuracy of model prediction. Upon feeding multi-modal images into the encoder to extract hierarchical features T / R 1 , T / R 2 , T / R 3 , T / R 4 , the encoder includes the modules CSCFM, MHFM, and SCAM, designed to amalgamate these hierarchical features into F 1 , F 2 , F 3 , and F 4 , respectively. Subsequently, the fused features F i are inputted into the decoder for progressive pixel restoration. The framework aims to enhance the effectiveness of the encoder–decoder structure in RGB-T semantic segmentation tasks by using multi-modal feature complementarity.
Jimaging 11 00188 g003
Figure 4. The detailed architecture of the Channel–Spatial Cross-Fusion Module (CSCFM).
Figure 4. The detailed architecture of the Channel–Spatial Cross-Fusion Module (CSCFM).
Jimaging 11 00188 g004
Figure 5. The detailed architecture of Multi-Head Fusion Module (MHFM).
Figure 5. The detailed architecture of Multi-Head Fusion Module (MHFM).
Jimaging 11 00188 g005
Figure 6. The architecture of the Spatial Coordinate Attention Mechanism.
Figure 6. The architecture of the Spatial Coordinate Attention Mechanism.
Jimaging 11 00188 g006
Figure 7. Visual comparisons of CSANet and top 5 methods in typical daytime and nighttime images on MFNet dataset.
Figure 7. Visual comparisons of CSANet and top 5 methods in typical daytime and nighttime images on MFNet dataset.
Jimaging 11 00188 g007
Figure 8. Visual comparisons of CSANet and top 3 methods on PST900 dataset.
Figure 8. Visual comparisons of CSANet and top 3 methods on PST900 dataset.
Jimaging 11 00188 g008
Figure 9. Visualized results (Baseline, +w/o CSCFM, +w/o MHFM, +w/o SCAM and CSANet) of ablation studies.
Figure 9. Visualized results (Baseline, +w/o CSCFM, +w/o MHFM, +w/o SCAM and CSANet) of ablation studies.
Jimaging 11 00188 g009
Table 1. Quantitative comparisons (%) on the test set of MFNet dataset (Part 1). The top 3 results in each column are highlighted in red, green and blue. “-” denotes that the authors do not provide the corresponding results.
Table 1. Quantitative comparisons (%) on the test set of MFNet dataset (Part 1). The top 3 results in each column are highlighted in red, green and blue. “-” denotes that the authors do not provide the corresponding results.
MethodsYearsCarPersonBikeCurve
AccIoUAccIoUAccIoUAccIoU
MFNet201777.265.967.058.953.942.936.229.9
RTFNet201991.386.378.267.871.558.269.843.7
PSTNet2020-76.8-52.6-55.3-29.6
MLFNet2021-82.3-68.1-67.3-27.3
FuseSeg202193.187.981.471.778.564.668.444.8
ABMDRNet202194.384.890.069.675.760.364.045.1
FEANet202193.387.882.771.176.761.165.546.5
MFFENet202193.188.283.274.177.162.967.246.2
GMNet202194.186.583.073.176.961.759.744.0
MMNet2022-83.9-69.3-59.0-43.2
EGFNet202295.887.689.069.880.658.871.542.8
MTANet202295.888.190.971.580.360.775.340.9
CCFFNet202294.589.683.674.273.263.167.250.5
CCAFFMNet202295.289.185.972.582.367.571.846.3
DSGBINet202295.287.489.269.585.264.766.046.3
CMXSegF2022-89.4-74.8-64.7-47.3
FDCNet202294.187.591.472.478.161.770.143.8
ECGFNet202389.483.585.272.172.961.662.840.5
MMSMCNet202396.289.293.269.183.463.574.446.4
LASNet202394.984.281.767.182.156.970.741.1
SFAF-MA202394.387.883.972.472.059.564.446.0
DBCNet202493.087.482.773.670.361.871.247.1
CAINet202493.088.574.666.385.268.765.955.4
U-Mamba202482.774.377.965.559.948.852.436.2
U-KAN202581.272.476.464.158.347.251.735.0
CSANet (Ours)-95.3 ± 0.285.6 ± 0.291.0 ± 0.265.7 ± 0.386.0 ± 0.271.7 ± 0.349.5 ± 0.337.2 ± 0.3
Table 2. Quantitative comparisons (%) on the test set of MFNet dataset (Part 2). The top 3 results in each column are highlighted in red, green and blue. “-” denotes that the authors do not provide the corresponding results.
Table 2. Quantitative comparisons (%) on the test set of MFNet dataset (Part 2). The top 3 results in each column are highlighted in red, green and blue. “-” denotes that the authors do not provide the corresponding results.
MethodsYearsCar StopGuardrailColor ConeBumpmAccmIoU
AccIoUAccIoUAccIoUAccIoU
MFNet201719.19.90.18.530.325.230.027.745.139.7
RTFNet201932.124.313.43.640.426.073.557.262.251.7
PSTNet2020-25.1-15.1-39.4-45.0-48.4
MLFNet2021-30.4-15.7-55.6-40.1-53.8
FuseSeg202129.122.763.76.455.846.966.447.970.654.5
ABMDRNet202144.133.131.05.161.747.466.250.069.554.8
FEANet202126.622.170.86.666.655.377.348.973.255.3
MFFENet202152.337.165.07.658.552.473.447.474.357.1
GMNet202155.042.371.214.554.748.773.147.474.157.3
MMNet2022-24.7-4.6-42.2-50.762.752.8
EGFNet202248.733.833.67.065.348.371.147.172.754.8
MTANet202262.838.938.713.763.845.970.847.275.256.1
CCFFNet202238.731.930.64.855.249.772.956.368.357.6
CCAFFMNet202232.525.256.817.358.350.676.658.372.958.2
DSGBINet202256.743.47.83.382.061.772.848.972.658.1
CMXSegF2022-30.1-8.1-52.4-59.4-58.2
FDCNet202234.427.261.57.364.052.074.556.674.156.3
ECGFNet202344.830.845.211.157.249.765.150.969.155.3
MMSMCNet202356.641.926.98.870.248.877.557.675.258.1
LASNet202356.839.659.518.958.148.877.240.175.454.9
SFAF-MA202334.024.735.64.355.839.167.952.667.553.8
DBCNet202446.233.878.262.957.750.974.645.474.856.2
CAINet202434.731.565.69.055.648.985.060.773.258.6
U-Mamba202429.518.141.624.362.742.848.838.363.548.2
U-KAN202528.216.640.122.761.441.247.637.062.146.7
CSANet (Ours)-64.7 ± 0.448.5 ± 0.378.2 ± 0.344.2 ± 0.387.0 ± 0.258.9 ± 0.362.1 ± 0.253.5 ± 0.279.1 ± 0.1562.5 ± 0.13
Table 3. Quantitative comparison (%) of daytime and nighttime scenes.
Table 3. Quantitative comparison (%) of daytime and nighttime scenes.
MethodsDaytimeNighttime
mAccmIoUmAccmIoU
DSGBINet62.847.265.348.9
FEANet64.752.361.544.2
CAINet66.453.960.241.3
GMNet65.152.763.446.5
FDCNet63.649.261.744.9
MFFENet66.552.364.947.6
DBCNet67.254.562.345.1
MTANet65.853.461.243.9
MMSMCNet66.954.762.044.6
LASNet68.956.369.255.1
CSANet (Ours)76.261.171.958.3
Table 4. Quantitative comparisons (%) on the test set of the PST900 dataset (Part 1). The top 3 results in each column are highlighted in red, green and blue. “-” denotes that the authors do not provide the corresponding results.
Table 4. Quantitative comparisons (%) on the test set of the PST900 dataset (Part 1). The top 3 results in each column are highlighted in red, green and blue. “-” denotes that the authors do not provide the corresponding results.
MethodsYearsBackgroundHand-DrillBackpack
AccIoUAccIoUAccIoU
MFNet2017-98.63-41.13-64.27
PSTNet2020-98.85-53.60-69.20
MFFENet2021-99.40-72.50-81.02
GMNet202199.8199.4490.2985.1789.0183.82
EGFNet202299.4899.2697.9964.6794.1783.05
MTANet2022-99.33-62.05-87.50
CCFFNet202299.999.489.782.877.575.8
DSGBINet202299.7399.3994.5374.9988.6585.11
FDCNet202299.7299.1582.5270.3677.4572.17
MMSMCNet202399.5599.3997.9662.3696.9489.22
LASNet202399.7799.4691.8182.8090.8086.48
DBCNet2024-99.40-77.19-82.67
CAINet202499.6699.5095.8780.3096.0988.02
CSANet (Ours)-99.7399.5595.6583.7697.9089.88
Table 5. Quantitative comparisons (%) on the test set of PST900 dataset (Part 2). The top 3 results in each column are highlighted in red, green and blue. “-” denotes that the authors do not provide the corresponding results.
Table 5. Quantitative comparisons (%) on the test set of PST900 dataset (Part 2). The top 3 results in each column are highlighted in red, green and blue. “-” denotes that the authors do not provide the corresponding results.
MethodsYearsFire-ExtinguisherSurvivormAccmIoU
AccIoUAccIoU
MFNet2017-60.35-20.70-57.02
PSTNet2020-70.12-50.03-68.36
MFFENet2021-66.38-75.60-78.98
GMNet202188.2873.7980.8678.3689.6184.12
EGFNet202295.1771.2983.3074.3094.0278.51
MTANet2022-64.95-79.14-78.60
CCFFNet202287.679.979.772.786.982.1
DSGBINet202294.7879.3181.3775.5691.8182.87
FDCNet202291.7771.5278.3672.3685.9677.11
MMSMCNet202397.3673.2984.2874.7095.2079.80
LASNet202392.3677.7583.4375.4991.6384.40
DBCNet2024-72.95-76.68-81.78
CAINet202488.3877.2191.3578.6994.2784.73
CSANet (Ours)-98.4186.3586.1876.5195.5786.01
Table 6. Computational complexity comparison of different methods.
Table 6. Computational complexity comparison of different methods.
MethodsInput SizeFLOPs/G ↓Params/M ↓mAccmIoU
RTFNet640 × 480245.71185.2462.251.7
PSTNet640 × 480337.04254.51-48.4
FuseSeg640 × 480129.3720.3870.654.6
ABMDRNet640 × 480194.3364.6069.554.8
EGFNet640 × 480201.2962.8272.754.8
MTANet640 × 480264.69121.5875.256.1
FDCNet640 × 480159.0552.9174.156.3
MMSMCNet640 × 480181.8298.5875.258.1
LASNet640 × 480233.8193.5875.454.9
DBCNet640 × 48067.4947.8774.856.2
CAINet640 × 480123.6212.1673.258.6
CSANet (Ours)640 × 48084.3238.3679.162.5
Table 7. Quantitative results (%) of ablation studies on the MFNet dataset. The results include overall metrics ( mAcc and mIoU ) and per-class IoU .
Table 7. Quantitative results (%) of ablation studies on the MFNet dataset. The results include overall metrics ( mAcc and mIoU ) and per-class IoU .
VariantsmAccmIoUCarPersonBikeGuardrailCar StopPole
Baseline64.853.285.188.259.332.725.628.2
Baseline + w/o CSCFM76.958.786.890.162.536.431.734.8
Baseline + w/o MHFM69.557.386.089.461.735.830.233.5
Baseline + w/o SCAM75.356.886.489.862.135.929.731.2
CSANet (Ours)79.162.587.691.065.737.244.239.6
Table 8. Quantitative results (%) of single-modality and multi-modality ablation studies on the MFNet dataset.
Table 8. Quantitative results (%) of single-modality and multi-modality ablation studies on the MFNet dataset.
VariantsmAccmIoUCarPersonBikeGuardrailCar StopPole
CSANet (RGB only)72.456.782.185.961.233.535.832.1
CSANet (TIR only)69.854.280.483.659.431.233.630.8
CSANet (RGB+TIR)79.162.587.691.065.737.244.239.6
Table 9. Quantitative results (%) of different attention modules on the MFNet dataset.
Table 9. Quantitative results (%) of different attention modules on the MFNet dataset.
VariantsmAccmIoUCarPersonBikeGuardrailCar StopPole
CSANet w/CoordConv76.859.381.288.362.138.538.735.8
CSANet w/CoordAtt77.560.182.389.163.039.639.836.5
CSANet w/SCAM79.162.587.691.065.737.244.239.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, R.; Wang, Z.; Guo, J.; Zhang, C. CSANet: Context–Spatial Awareness Network for RGB-T Urban Scene Understanding. J. Imaging 2025, 11, 188. https://doi.org/10.3390/jimaging11060188

AMA Style

Li R, Wang Z, Guo J, Zhang C. CSANet: Context–Spatial Awareness Network for RGB-T Urban Scene Understanding. Journal of Imaging. 2025; 11(6):188. https://doi.org/10.3390/jimaging11060188

Chicago/Turabian Style

Li, Ruixiang, Zhen Wang, Jianxin Guo, and Chuanlei Zhang. 2025. "CSANet: Context–Spatial Awareness Network for RGB-T Urban Scene Understanding" Journal of Imaging 11, no. 6: 188. https://doi.org/10.3390/jimaging11060188

APA Style

Li, R., Wang, Z., Guo, J., & Zhang, C. (2025). CSANet: Context–Spatial Awareness Network for RGB-T Urban Scene Understanding. Journal of Imaging, 11(6), 188. https://doi.org/10.3390/jimaging11060188

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop