Next Article in Journal
The Design and Implementation of a Dynamic Measurement System for a Large Gear Rotation Angle Based on an Extended Visual Field
Previous Article in Journal
TDA-L: Reducing Latency and Memory Consumption of Test-Time Adaptation for Real-Time Intelligent Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SC-CoSF: Self-Correcting Collaborative and Co-Training for Image Fusion and Semantic Segmentation

1
Key Laboratory of Big Data Intelligent Computing, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
2
School of Artificial Intelligence, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(12), 3575; https://doi.org/10.3390/s25123575
Submission received: 1 May 2025 / Revised: 28 May 2025 / Accepted: 3 June 2025 / Published: 6 June 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Multimodal image fusion and semantic segmentation play pivotal roles in autonomous driving and robotic systems, yet their inherent interdependence remains underexplored. To address this gap and overcome performance bottlenecks, we propose SC-CoSF, a novel coupled framework that jointly optimizes these tasks through synergistic learning. Our approach replaces traditional duplex encoders with a weight-sharing CNN encoder, implicitly aligning multimodal features while reducing parameter overhead. The core innovation lies in our Self-correction and Collaboration Fusion Module (Sc-CFM), which integrates (1) a Self-correction Long-Range Relationship Branch (Sc-LRB) to strengthen global semantic modeling, (2) a Self-correction Fine-Grained Branch (Sc-FGB) for enhanced visual detail retention through local feature aggregation, and (3) a Dual-branch Collaborative Recalibration (DCR) mechanism for cross-task feature refinement. This design preserves critical edge textures and color contrasts for segmentation while leveraging segmentation-derived spatial priors to guide fusion. We further introduce the Interactive Context Recovery Mamba Decoder (ICRM) to restore lost long-range dependencies during the upsampling process; meanwhile, we propose the Region Adaptive Weighted Reconstruction Decoder (ReAW), which is mainly used to reduce feature redundancy in image fusion tasks. End-to-end joint training enables gradient propagation across all task branches via shared parameters, exploiting inter-task consistency for superior performance. Experiments demonstrate significant improvements over independently optimized baselines in both fusion quality and segmentation accuracy.

1. Introduction

In the field of autonomous driving, precise and robust scene analysis is crucial. However, in complex environments (such as adverse weather or nighttime driving), relying solely on visible light sensors may make it difficult to accurately identify targets [1,2,3]. In contrast, infrared sensors are not affected by these factors and can provide clear target information in low light and harsh weather conditions; however, their spatial resolution is lower. Therefore, infrared and visible modality fusion (IVMF) has become an effective approach to enhance scene understanding and target detection performance [4,5].
From a practical application perspective, IVMF aims to generate high-quality, visually perceptive fused images, along with high-precision semantic segmentation results for scene analysis tasks. In recent years, advances in deep learning have significantly driven the development of the IVMF field [6,7,8,9]. For instance, CDDFuse [10] cleverly combines the global feature extraction capability of Transformers with the local feature-capturing advantages of CNNs, achieving efficient fusion of multimodal images. In the field of semantic segmentation, SegFormer [11] outputs multi-scale features by adopting a hierarchical structure Transformer encoder, without the need for positional encoding, thereby avoiding performance degradation when the test resolution differs from the training resolution. Although the above methods have achieved encouraging results in terms of visual effects and segmentation performance, there are still several challenges.
First, in semantic segmentation, although the Transformer-based global context modeling capability has shown remarkable results, its high computational complexity makes it difficult to fully leverage its advantages on low-performance devices, which can lead to insufficient feature aggregation. Additionally, during the upsampling process, some global context modeling ability may be lost, weakening the model’s ability to make semantic decisions for continuous regions. At the same time, the introduction of the Mamba architecture has brought innovation to the field [12], offering hope for solving the existing problems. Mamba excels at efficiently capturing global context information and is renowned for its computational efficiency based on the State Space Model (SSM) [13]. Vision Mamba [14] and VMamba [15] extended this architecture to computer vision tasks, further enhancing performance by strengthening the unidirectional scanning mechanism.
Secondly, in image fusion, the limited receptive field of CNNs, due to finite kernel sizes, causes local artifacts and hampers the preservation of fine-grained details, reducing visual quality. Additionally, the weight-sharing mechanism across regions makes CNNs prone to learning redundant feature representations, leading to feature redundancy. Finally, most existing methods treat image fusion and semantic segmentation as independent problems, overlooking the inherent correlation between the two. Although some studies [16,17] validate the effectiveness of fused images by using the results of image fusion as input for semantic segmentation or object detection, as shown in Figure 1, this cascading approach lacks effective strategies to fully utilize the inherent consistency between the two, failing to maximize their potential. At this point, multi-task learning (MTL) [18], which has been widely applied in the field of machine learning, inspired us. This approach aims to construct a unified model capable of simultaneously addressing multiple related tasks, with the goal of leveraging the underlying commonalities among tasks to enhance overall performance and reduce inefficiencies and redundancies during training [19,20]. In particular, MTL has demonstrated significant potential and promising prospects in dense prediction tasks such as semantic segmentation [21,22,23].
To address the aforementioned challenges and inspired by recent advances in state-of-the-art technologies, in this paper, we propose SC-CoSF, a novel coupled framework for image fusion and semantic segmentation. This model leverages the similarity between the two tasks, treating them as a joint task to promote mutual benefits rather than handling them independently. On this basis, we design the Self-correction and Collaboration Fusion Module (Sc-CFM), which comprises a Self-correction Long-Range Relationship Branch (Sc-LRB), a Self-correction Fine-Grained Branch (Sc-FGB), and a Dual-branch Collaborative Recalibration (DCR) mechanism. The Sc-LRB employs Mamba-based global attention and cross-attention to effectively capture cross-region dependencies and enhance semantic modeling, thereby improving edge sharpness and structural coherence; the Sc-FGB integrates multi-scale convolution operations to extract fine texture details and enrich visual information; and the DCR mechanism fuses the outputs of these two branches to dynamically reshape visual features and semantic information flows, ensuring that the fused results exhibit both high contrast and rich spatial semantics. Finally, we introduce two task-specific decoders. The Interactive Context Restoration Mamba Decoder (ICRM) for semantic segmentation recovers remote contextual information lost during upsampling. The Region-Adaptive Weighted Reconstruction Decoder (ReAW) for image fusion adaptively suppresses redundant features to enhance the visual quality of the fused image. By jointly optimizing the above tasks, the encoded features propagate gradients across all task branches through shared parameters, fully utilizing the inherent consistency between tasks and achieving better performance than independent optimization.
To summarize, our main contributions are as follows:
  • We design a Self-correction and Collaboration Fusion Module (Sc-CFM), which integrates a Self-correction Long-Range Relationship Branch (Sc-LRB) to enhance semantic modeling capability and a Self-correction Fine-Grained Branch (Sc-FGB) to comprehensively capture multi-scale visual information, and employs a Dual-branch Collaborative Recalibration (DCR) mechanism to simultaneously recalibrate semantic features and visual representations. The Sc-CFM is specifically engineered to autonomously optimize the embedded semantic and visual features within latent representations, while simultaneously facilitating synergistic fusion between enhanced semantic guidance and raw visual signals. This dual-phase refinement process not only suppresses feature-level noise but also reinforces cross-modal alignment, ultimately generating discriminative feature representations that substantially improve downstream task performance.
  • We propose an Interactive Context Restoration Mamba Decoder and a Region-adaptive Weighted Reconstruction Decoder, aiming to recover remote information lost during the upsampling process and reduce feature redundancy in the fused image.
  • By establishing a coupling module that integrates image fusion with semantic segmentation in a complementary manner, our approach effectively harnesses the strengths of both tasks, thereby delivering a performance that maximizes each modality’s advantages.

2. Related Work

2.1. Multi-Modality Fusion and Segmentation

Recent deep learning-based multimodal image fusion methods have made significant progress. Early attempts [24,25,26] focused on adjusting network structures or loss functions to achieve fusion results, but often overlooked downstream task performance. Some fused images with good metrics may not be suitable for practical applications. Zhang et al. [27] introduced the CMX framework to improve semantic segmentation by integrating RGB and other modalities using cross-modal calibration and fusion modules. Tang et al. [16] proposed SeAFusion, which incorporates semantic segmentation feedback to balance low-level visual details and high-level semantic information, generating fused images that retain both thermal and texture details. Semantic segmentation and image fusion are jointly trained in an integrated manner. Liu et al. [17] achieved collaborative fusion of textures and semantics from infrared and visible images, leveraging mutual promotion to improve detail restoration and semantic representation. Jiang et al. [28] introduced domain adaptation branches to capture semantic and geometric information, bridging the gap between modalities and improving fused image quality for downstream tasks. However, these fusion-segmentation methods often rely on cascaded architectures, resulting in complex training and lacking an end-to-end design. As a result, they struggle to produce reliable fused images and accurate segmentation outputs with a single unified network.

2.2. State Space Models

State Space Models (SSMs) [29], inspired by linear time-invariant systems, have emerged as highly efficient sequence-to-sequence architectures. In particular, the Structured State-Space Sequence model (S4) [13] has been recognized as a groundbreaking framework for capturing long-range dependencies. Building on this success, the incorporation of selective mechanisms into S4, as demonstrated by Mamba [12], has enabled it to outperform Transformers and other state-of-the-art architectures. Capitalizing on these impressive results, researchers have recently extended SSMs to computer vision applications. For example, Vision Mamba [14] combines SSMs with bidirectional scanning to establish robust inter-patch relationships, and VMamba [15] further enhances this approach by incorporating four-directional scanning, thereby more comprehensively capturing the interrelations among image patches. Inspired by the above methods, we incorporate State Space Models (SSMs) into our multimodal task and design a Mamba-based fusion strategy. We also propose a dual-branch Mamba decoder for information selection. This approach balances low computational complexity and high accuracy by integrating the SSM mechanism into the fusion process.

2.3. Multi-Task Learning

Multi-task learning (MTL) seeks to enhance the performance of several related tasks by sharing and utilizing relevant information across them. Various approaches have been proposed to achieve effective multi-task learning by modifying the model architecture. For instance, Felix et al. [30] designed the architecture by using a cross-task sharing approach at the lower layers of the model. Ishan Misra et al. [31] proposed a novel module that enhances multi-task learning performance by learning a linear combination of activation maps from different tasks, allowing the network to find an optimal balance between shared and task-specific representations. Meanwhile, many emerging techniques for multi-task learning have also been proposed. For example, Zhang et al. [32] achieve mutually beneficial closed-loop learning by adaptively enhancing complementary information between tasks through the use of a task attention module (TAM). Other methods approach multi-task learning from an optimization perspective, combining task objectives into weighted sums [33,34], gradient normalization [35,36], and dynamic adjustment of task importance [17]. These methods further enhance the collaboration between tasks. Inspired by the above methods, we have incorporated the concept of multi-task learning into our framework design. By leveraging low-level feature sharing and a weighted loss function, we effectively utilize the inherent consistency between different tasks, mitigating the performance trade-off problem in multi-task learning.

3. Methodology

3.1. Problem Formulation

For tasks such as multi-modal image fusion or segmentation, a widely adopted method is to design a neural network that fully utilizes its capacity to determine an optimal set of parameters The optimization model is formulated as follows:
min ω k f ( k , L ( x , y ; ω k ) ) ,
where k represents the output of the task-specific network L with learnable parameters ω k . f ( · ) is a constraint term used to optimize the network. Here, we suppose a pair of an RGB image x R H × W × C x and an infrared image y R H × W × C y , respectively.
The previous methods mostly designed image fusion and semantic segmentation frameworks in a cascading manner, making it difficult to find a balance between different tasks. Therefore, here, we design a coupling learning framework to integrate the objectives of different tasks, which can be written in the following form:
min ω f , ω s g f f , Φ ( x , y ; ω f ) + h s s , Ψ ( x , y ; ω s ) + S ( ω )
where ω = { ω f , ω s } denotes the parameters of the fusion network Φ and segmentation network Ψ , respectively, with some parameters shared between them. f and s denote the fused image and segmentation results, which are produced by the fusion network Φ and Ψ . S ( · ) is a constraint term for joint optimization, implemented via a bespoke distance-based loss.

3.2. Overvall Architecture

The proposed SC-CoSF is designed using a parallel strategy, consisting of image fusion and semantic segmentation subnetworks. The overall structure details are shown in Figure 2. Sc-CFM effectively extracts and couples semantic and visual information from multimodal data, while the segmentation network (ICRM) and fusion network (ReAW) extract key information from the coupled features, fully capturing detailed information and extracting semantic features.

3.2.1. Sc-CFM

After passing through the weight-shared backbone, the RGB image and infrared image will each generate multimodal feature maps F R = { F 1 R , , F n R } and F I = { F 1 I , , F n I } . To effectively couple the above features, we propose the Self-correction and Collaboration Fusion Module (Sc-CFM), as shown in Figure 3. The Sc-CFM refines visual and semantic representations across modalities and harmonizes them for joint optimization. This results in a fused feature map containing rich visual effects and segmentation information. The module comprises three components: Self-correction Long-Range Relationship Branch (Sc-LRB), Self-correction Fine-Grained Branch (Sc-FGB), and Dual-branch Collaborative Recalibration (DCR) for recalibrating the information from different branches.
Sc-LRB
We propose a Self-correction Long-Range Relationship Branch (Sc-LRB) based on Mamba to address the efficiency issue in establishing long-range relationships in segmentation tasks. Although ViT has demonstrated outstanding performance in various visual tasks [37], enhancing contextual understanding for semantic segmentation, its self-attention mechanisms are computationally intensive. To address this issue, we introduce the efficient S6 structure [12] and construct our self-correction module for long-range relationship modeling. Consequently, the cross-attention mechanism [38] can capture interrelationships between positions in the input, thereby greatly facilitating feature extraction and complementarity between different data sources. Therefore, our Sc-LRB incorporates two correction strategies: self-modal correction and cross-modal complementarity.
In the self-modal correction stage, the Sc-LRB first receives visible and infrared feature maps from the lower-level encoder and processes them separately through a linear layer and depthwise separable convolution for initial transformation. Subsequently, each of these two feature streams is fed into the S6 module, where linear projection layers are used to generate corresponding matrices (such as B, C, and Δ ) for systematically capturing global dependencies. The module’s output is normalized via layer normalization (LN), passed through a linear projection to restore dimensionality, and finally, added to the original input via residual connection, thereby actively correcting intra-modal feature deviations. The formula can be represented as follows:
O i r = F i R + L i n e a r ( L N ( S 6 ( D W C o n v ( L i n e a r ( F i R ) ) ) ) ) ,
O i i = F i I + L i n e a r ( L N ( S 6 ( D W C o n v ( L i n e a r ( F i I ) ) ) ) ) ,
Then, we obtain the self-corrected feature map O R = { O 1 r , , O n r } and O I = { O 1 i , , O n i } . For cross-modal complementarity, it takes the corrected dual-modality features as input, with most processing steps similar to those in the previous stage. The difference lies in our custom-designed Cross S6 block, which is capable of simultaneously processing both streams of information, thereby enabling cross-modal information complementation at this stage. The process can be formulated as follows:
W i r = O i r + L i n e a r ( L N ( c r o s s S 6 ( D W C o n v ( L i n e a r ( O i r , O i i ) ) ) ) )
W i i = O i i + L i n e a r ( L N ( c r o s s S 6 ( D W C o n v ( L i n e a r ( O i r , O i i ) ) ) ) )
The obtained completed features can be described as: W R = { W 1 r , , W n r } and W I = { W 1 i , , W n i } . The specific architecture of the module is shown in Figure 3a, and Figure 3b illustrates the working principle of the S6 module. For the Cross S6 block, inspired by [39], we implement the corresponding functionality by replacing the matrix C between different modalities, with the following formula:
A ¯ 1 , B ¯ 1 = exp ( Δ A 1 ) , Δ B 1 , A ¯ 2 , B ¯ 2 = exp ( Δ A 2 ) , Δ B 2 ,
h 1 t = A 1 ¯ h 1 t 1 + B 1 ¯ x 1 t , h 2 t = A 2 ¯ h 2 t 1 + B 2 ¯ x 2 t ,
y 1 t = C 2 h 1 t + D 1 x 1 t , y 2 t = C 1 h 2 t + D 2 x 2 t ,
where A 1 / 2 and B 1 / 2 are the state transition and input matrices, and C 1 / 2 and D 1 / 2 are the output projection matrices and the direct input-to-output projection matrices for each modality. x 1 / 2 t represents the input at time step t, and y 1 / 2 is the selective scan output. In the SSM, the matrix C is used to map the current state to the output space, determining how the state generates the model’s output. By exchanging the matrix C between different data, we achieve cross-enhancement of features. Subsequently, we obtain the initial fused features F s = { F 1 s , , F n s } by concatenating W R and W I and use a channel-wise attention mechanism to extract key information from the fused features, while suppressing low-density information, which can be formulated as follows:
α i , j = 1 H i W i h = 1 H i w = 1 W i F i s ( h , w , j ) ,
F i S = F i s σ ( Conv 1 × 1 ( α i ) ) + F i s ,
where α i = [ α i , 1 , , α i , C i ] R 1 × 1 × C i is the average pooling results of each feature map in F i s . Finally, we obtain the fused segmentation images F S = { F 1 S , , F n S } after long-range relationship correction.
Sc-FGB
In addition to semantic features, fine-grained and rich visual features also play a crucial role in the accurate analysis of details such as edges and steps. Compared to Transformers, convolutional neural networks (CNNs) are more efficient at mining local features and can significantly capture fine-grained details of images. Therefore, we have designed a multi-kernel convolution-based Self-correction Fine-Grained Branch (Sc-FGB) to enhance the extraction of fine-grained features, with specific details shown in Figure 3c. Similar to the Sc-LRB, it receives multimodal feature maps from the shared encoder and outputs a corrected and fused map enriched with abundant visual information and fine-grained features. Specifically, inspired by SegNext [40], we first apply multi-scale convolutions ( 3 × 3 , 5 × 5 , 7 × 7 ) to perform lateral correction of feature maps from different modalities and then sum these feature maps, which can be formulated as follows:
H i R / I ( 3 ) = Conv 3 × 3 Conv 1 × 1 ( F i R / I ) , H i R / I ( 5 ) = Conv 5 × 5 Conv 1 × 1 ( F i R / I ) , H i R / I ( 7 ) = Conv 7 × 7 Conv 1 × 1 ( F i R / I ) ,
H i ˜ R / I = H i R / I ( 3 ) + H i R / I ( 5 ) + H i R / I ( 7 ) ,
where H i ˜ R and H i ˜ I represent the combined characteristics of visible and infrared modalities, respectively. Next, we introduce a channel attention weighting mechanism to integrate information from different scales and obtain an initial visual fusion map F v = { F 1 v , , F n v } through concatenation. The formulas are as follows:
A i = σ DWConv ( GAP ( H ˜ i R / I ) ) , H ^ i R / I = A i H ˜ i R / I , F i v = Concat ( H ˜ i R , H ˜ i I ) ,
For the concatenated result, we further apply depthwise separable convolutions to correct the details and expand the number of channels. Then, we split it along the channel dimension into two parts, F α and F β , to capture detailed features from multiple perspectives. Finally, these two parts interact through Hadamard multiplication, achieving the final feature fusion. This process can be represented as follows:
F i V = Conv 1 × 1 ( F α σ ( F β ) ) ,
Then, we obtain the fused visual feature F V = { F 1 V , , F n V } after fine-grained correction.
DCR
To synergistically recalibrate the semantically corrected and visually corrected feature streams produced by the preceding branches, we introduce the Dual-branch Collaborative Recalibration (DCR) module (shown in Figure 3d). The two corrected feature sets, F S and F V , are first merged via element-wise summation to form a base feature map. To incorporate spatial structural cues, a coordinate attention-inspired mechanism is employed [41]: global average pooling is performed separately along the height and width axes, yielding ( H , 1 ) and ( 1 , W ) descriptors. Each descriptor is then projected through a 1 × 1 convolution into an intermediate channel space and passed through a nonlinear activation to produce spatial attention maps. These maps are broadcast across the fused feature map and multiplied elementwise to amplify structurally salient regions and suppress noise. The resulting recalibrated feature representation retains both semantic and visual coherence and serves as the final output of the fusion process. The formulas can be represented as follows:
F i = F i S + F i V , F = { F 1 , , F n } ,
α i , j , p h = 1 W 0 w < W F i ( p , w , j ) ,
β i , j , q w = 1 H 0 h < H F i ( h , q , j ) ,
where F i is the sum result between F i S and F i V , α i h R H × 1 × C i and β i w R 1 × W × C i store the average pooling results of each feature map. Then, through feature map concatenation and convolution operations, we obtain a feature map that contains rich spatial information, which can be written as follows:
W i = σ ( Conv 1 × 1 ( [ α i h , β i w ] ) ) ,
Next, through a split operation, we separately send the horizontal and vertical feature maps into independent convolutional layers, and use activation functions to generate spatial attention weights. Then, these attention weights are applied to the original input feature map, thereby enhancing the spatial dimension of the feature map, which can be represented as follows:
F i = F i α ^ i h β ^ i w ,
where α ^ i h and β ^ i w are the split results of W i . Finally, the fused images after collaborative recalibration can be defined as: F = { F 1 , , F n } .

3.2.2. ICRM

We employ an Interactive Context Restoration Mamba Decoder (ICRM) to decode the feature maps output by Sc-CFM, in order to generate the final prediction result. To fully extract rich local context from the encoder and to mitigate global-context loss during upsampling, we introduce the Mamba architecture as our decoder. The decoder progressively strengthens the relationships between different features in the image through an interactive information restoration process. In this process, each feature map is first subjected to an upsampling operation and then passed through multiple Interactive Context Restoration Visual State Space Blocks (ICRVSSBs) for feature extraction and upsampling; the architecture is shown in the lower part of Figure 4. Finally, the encoded features are passed into the MLP layer for classification, resulting in semantic segmentation output. Our ICRVSSB module is an improvement upon the VSSB architecture based on Vmamba [15], which can be formulated as follows:
Υ 1 = LN S S 2 D SiLU DWConv Linear ( Υ ) ,
where Υ is the feature map fed into the VSSB. The core mechanism of VSSB is the Selective Scan 2D (SS2D) module, as shown in Figure 4 upper part. In SS2D, the 2D data are first unfolded into sequences along four different traversal paths (top-left to bottom-right, bottom-right to top-left, top-right to bottom-left, and bottom-left to top-right). Then, each sequence is processed in parallel through an independent S6 module. The processed result sequences are reshaped and merged to ultimately obtain the output map. Through this process, SS2D is able to integrate pixel information from different directions within the image, effectively enhancing and restoring the global context in the 2D space.
While existing VSSBs excel at extracting global context, they still fall short in capturing channel and spatial information. Therefore, we replace the multiplicative branch in the original VSSB with a channel and spatial attention mechanism. The formula is shown as follows:
Υ 2 = σ Θ C Θ S ( Υ ) ,
where Θ C represents channel attention and Θ S represents spatial attention. Finally, we fuse the outputs of the two branches through Hadamard product interaction and use a linear layer to map the channels back to the size of the input channels, which can be written as follows:
Ω = Linear ( Υ 1 Υ 2 ) .
The above operations enhance the model’s information interaction and context restoration capabilities, thereby constructing a more robust decoder module.

3.2.3. ReAW

In order to suppress irrelevant information and address feature redundancy, we propose the Region-adaptive Weighted Reconstruction Decoder (ReAW), which further processes the feature maps output by Sc-CFM to produce the final fused image. This allows for the dynamic extraction of region-specific features from the input feature maps. Meanwhile, we believe that the original input image contains important information for image reconstruction, which can guide the more abstractly extracted feature maps. During the feature selection and exclusion process, this guidance helps the feature maps make more reasonable decisions, ultimately achieving better fused image reconstruction. Therefore, in the final image fusion, we incorporate the relevant information from the original image into the final results.
For feature maps with rich visual effects, we first dynamically assign the number of regions to focus on for each feature map. Then, the incoming feature maps are processed by convolution operations of different scales, denoted as e ( · ) , for feature extraction based on the number of regions. The function e ( · ) can be represented by the following formula:
e ( F i ) = RELU ( BN ( Conv ( F i ) ) ) ,
Then, two linear layers and a ReLU activation function are used to compute the weight coefficients for different convolutions, thereby dynamically determining which regions need to be focused on. A Softmax function is applied to normalize the coefficients. After obtaining the normalized coefficients, the extracted features corresponding to each region are multiplied by their respective feature weights, and the weighted features of all regions are summed to obtain the total weighted fusion feature, which can be formulated as follows:
b j ( F i ) = λ j e j ( F i ) ,
B i = [ b 1 , b 2 , , b l ] ,
where λ j is the weight of j-th region, l is the number of regions, and B i represents the result after processing the feature map F i . Subsequently, each feature B i is upsampled to the original image size for further processing. Finally, the multi-scale weighted feature fusion result is concatenated with the original image features and then fed into a reconstruction layer, which consists of two convolutional layers and activation functions, to obtain the final image fusion result.

3.3. Loss Function

In this chapter, we mainly discuss the configuration of the loss functions used during task training, including the segmentation loss for the semantic segmentation task and the fusion loss for image fusion.
Segmentation loss: The conventional cross-entropy loss is used to supervise semantic segmentation, which can be formulated as follows:
L = 1 H × W h = 1 H w = 1 W c = 1 C y h , w , c log ( p ^ h , w , c ) ,
Here, C is the number of classes, y h , w , c is the one-hot encoding of the true label for the pixel ( h , w ) and class c, and p ^ h , w , c is the predicted probability by the model that the pixel ( h , w ) belongs to class c.
Fusion loss: To generate images with rich visual effects, we design the image fusion loss following [42]. Specifically, we first generate positive samples of the input images through common data augmentation strategies (such as gamma transformation and contrast stretching). Then, we construct the fusion loss from multiple aspects by combining content loss, gradient loss, and color loss. In terms of content loss, the L1 loss between the fused image and the larger value between the corresponding infrared and visible light image positive samples is calculated to ensure that the fusion image retains the most significant parts of both input images. The formula is as follows:
L con = 1 N i = 1 N | F i M a x ( J i R , J i I ) | ,
where N is the number of pixels, J i R / I is the pixel value of the RGB image and infrared image, and F i is the pixel value of the fused image.
For the gradient loss, we compute the L1 distance between the gradients of the fused image and the maximum gradients of the corresponding augmented visible and infrared images in both horizontal and vertical directions. Specifically,
L grad = 1 N i = 1 N | F i J i | x + | F i J i | y ,
where F i and J i represent the gradient of fusion image and the source image.
Finally, the color loss is calculated by the L1 loss between the color channels (Cr and Cb channels) of the fused image and the positive sample in the YCbCr color space, which can be written as follows:
L color = 1 N i = 1 N | C r , F i C r , J i | + | C b , F i C b , J i | ,
where C r , F i and C b , F i , respectively, represent the C r and C b channel values of the fused image at pixel i; C r , J i and C b , J i represent the C r and C b channel values of the source image at pixel i. The overall fusion loss is achieved through the weighted sum of the three losses mentioned above. The formula is as follows:
L total = λ c o n L con + λ g r a d L grad + λ c o l o r L color .

4. Experiment

4.1. Datasets and Implementation Details

4.1.1. Datasets

The proposed method was validated on the MFNet [43] and FMB [17] datasets, which contain 1569 and 1500 pairs of visible and infrared images, respectively, with 393 and 280 pairs used for the test sets. The MFNet dataset includes semantic annotations for 9 classes, while the FMB dataset provides semantic annotations for 14 classes.

4.1.2. Implementation Details

All experiments are conducted on a single NVIDIA 3090 GPU. We jointly trained both tasks for 500 epochs. We employ the AdamW optimizer with an initial learning rate 6 × 10−5. In addition, some common data augmentation strategies (random flipping, random scaling, etc.) were also applied in our training. Additionally, infrared information is replicated across the channel dimension three times during data preprocessing to match the H × W × 3 input shape of RGB images. The shared encoders are configured as CNN-based structures following ConvNeXt [44], with the number of blocks set to 4 based on its original configuration.

4.2. Results of Semantic Segmentation

We provide comparative experiments for semantic segmentation, evaluating against seven state-of-the-art competitors: ConvNeXt [44], SeAFusion [16], SegMiF [17], SGFNet [45], LASNet [46], EAEFNet [47], and MDRNet+ [48]. The qualitative results are depicted in Figure 5 and Figure 6. The experimental results clearly demonstrate the significant advantages of our proposed method in semantic segmentation tasks. For example, as shown in Figure 5, our method preserves pedestrian contours more clearly, demonstrating a more accurate semantic decision-making capability, while the compared methods only achieve vague or approximate region segmentation. Among all the compared methods, except for SeAFusion and SegMiF, none are able to detect the background between persons. Moreover, in terms of contour segmentation accuracy, our method significantly outperforms existing approaches. Furthermore, in Figure 6, our method exhibits superior segmentation performance in terms of fine details (such as lamps). Specifically, compared with existing methods, our approach demonstrates significant advantages in the detection and segmentation of lamp objects. It consistently preserves salient features and achieves accurate detection in long-range scenarios, where other methods often suffer from missed detections or insufficient recognition. Moreover, in close-range scenes, our method captures fine contour details and achieves more precise segmentation, while the compared methods yield approximate contours with limited detail. These results clearly illustrate the superiority of our approach in multi-scale object preservation and fine-grained segmentation.
We also conducted a quantitative comparison using the mean Intersection over Union (mIoU) as the performance evaluation metric. To align with previous experiments, we selected eight categories from the FMB dataset as our evaluation targets. As shown in Table 1 and Table 2, the model we proposed achieved the highest mIoU scores on all test datasets. Overall, our method achieves an improvement of 0.9% in mIoU over the current state-of-the-art methods on the MFNet dataset, and a 1.4% gain in accuracy on the FMB dataset, further validating its effectiveness and generalizability. Specifically, our model achieved the highest IoU values in the Bump category in Table 1 and the Pole category in Table 2, indicating that our model has a significant advantage in segmenting small objects. These results fully demonstrate that our method has reached the current state-of-the-art level in semantic segmentation tasks.

4.3. Results of Image Fusion

The performance of image fusion is evaluated through comparative experiments involving six advanced fusion methods, namely U2Fusion [49], SeAFusion [16], SegMiF [17], CDDFuse [10], TGFuse [50], and DATFuse [51]. The qualitative comparison results are shown in Figure 7 and Figure 8, from which we can clearly see that our method demonstrates significant advantages. In Figure 7, our approach significantly improves environmental contrast and object visibility (e.g., leaves and distant buildings) in nighttime image fusion, and it more effectively highlights the detailed features of shadowed objects. In contrast, other methods often lose their ability to perceive the environment under conditions of low visibility. In addition, Figure 8 demonstrates another advantage of our method: Guided by semantic information, our approach can significantly enhance the contrast between the detected objects and their surrounding environment, thereby effectively improving the texture details in the scene and presenting a clearer and more realistic visual effect. In contrast, other methods fail to highlight the boundaries and details of objects as prominently when processing similar scenes, which affects the overall image quality and visualization.
Due to the lack of ground-truth data, we adopted no-reference metrics EN [52] and SD [53] to evaluate the amount of information and contrast. As shown in Figure 9, SC-CoSF achieved the highest scores on most metrics. In addition, the widest section of the violin plot is concentrated in the upper-middle part of the curve, further confirming that our method consistently achieves superior fusion performance across the majority of images. The above results demonstrate the advanced performance of our method in image fusion.

4.4. Complexity Analysis

To evaluate the computational complexity of various semantic segmentation and image fusion methods, we report the number of parameters and FLOPs based on input images with a resolution of 600 × 800 (see Table 3). Although SC-CoSF does not achieve the lowest values in these metrics, its complexity remains within a practically acceptable range and is lower than the complexity of several task-specific segmentation or fusion models. This analysis underscores the practicality of SC-CoSF and provides valuable insights for balancing accuracy and efficiency, especially in scenarios with strict real-time requirements such as autonomous driving and robotic perception systems.

4.5. Ablation Study

We conducted extensive ablation experiments on the MFNet dataset to verify the effectiveness of our proposed correction module. In addition, we designed corresponding ablation experiments for different task heads to validate the effectiveness of joint learning. Specifically, our model includes the following variants: Model A: Replacing Sc-CFM with tensor addition; Model B: Removing Sc-FGB; Model C: Removing Sc-LRB; Model D: Replacing DCR with tensor addition; Model E: Removing the semantic segmentation head; Model F: Removing the image fusion head. It is worth noting that for variants E and F, we only conducted the corresponding ablation experiments for the tasks they are associated with. The qualitative and quantitative results of semantic segmentation are shown in Table 4 and Figure 10. It can be seen that removing any module affects the model’s performance. Notably, the results of variants A and C show a significant decrease in contour accuracy for objects (such as pedestrians), highlighting the importance of Sc-LRB in the semantic segmentation task, as it provides strong global contextual capabilities. Additionally, we introduce attention heatmap visualizations in Figure 11 to intuitively compare how each module influences the segmentation process. The results clearly illustrate that the attention heatmaps of models A and C fail to encompass the complete outline of the vehicle, further underscoring the superior performance of Sc-LRB in semantic segmentation tasks.
Figure 12 and Figure 13 present the performance of different variants in image fusion, showing that each module plays a crucial role in the final results. For example, replacing either Sc-LRB or Sc-FGB would make it difficult for the model to recognize and handle the shadows at crossings. These results demonstrate the positive impact of our design on both semantic segmentation and image fusion.

5. Discussion

In this study, we propose the SC-CoSF architecture to address multimodal image fusion and semantic segmentation challenges, showing significant advantages for autonomous driving and robotics. Our experiments indicate that by employing a shared-weight backbone and the Self-correction and Collaboration Fusion Module (Sc-CFM), the model effectively integrates heterogeneous information, improving both fusion quality and segmentation accuracy. Unlike conventional methods that treat multimodal inputs independently, our approach leverages the Interactive Context Restoration Mamba Decoder (ICRM) and Region Adaptive Weighted Reconstruction Decoder (ReAW) to exploit the inherent relationships between visual and semantic cues. This end-to-end, multi-task learning mechanism optimizes individual tasks and enhances the overall robustness in dynamic environments.
Despite these promising outcomes, SC-CoSF has limitations that require further investigation. First, its generalization to rare or occluded scenarios needs validation on larger, more diverse datasets. Second, the depth and complexity of our two-branch architecture add computational overhead, potentially hindering real-time performance on embedded platforms. Finally, while the shared-weight strategy reduces parameter count, more lightweight or adaptive fusion schemes could be explored. Future work will (1) evaluate SC-CoSF on real-world driving and robotics benchmarks, (2) investigate model compression and hardware-aware optimization for faster inference, and (3) extend the framework to incorporate temporal coherence for video-based fusion and segmentation.
In summary, the SC-CoSF architecture offers a novel solution for multimodal image fusion and semantic segmentation, providing both theoretical insights and practical benefits. We anticipate that further improvements will advance the state-of-the-art in this rapidly evolving field.

6. Conclusions

In this work, we investigated the interdependence between multimodal image fusion and semantic segmentation in autonomous driving and robotics. To address performance limitations, we present SC-CoSF, an innovative framework that enables synergistic learning between these tasks. Our approach introduces three key contributions: First, we develop the Self-correction and Collaboration Fusion Module (Sc-CFM), which includes Sc-LRB for global context modeling, Sc-FGB for local detail preservation, and DCR for cross-modal feature calibration. Second, we design a dual-decoder system with ICRM for long-range dependency recovery and ReAW for adaptive feature reconstruction. Third, we employ an end-to-end joint optimization strategy, using shared parameters to synchronize gradient propagation across tasks, fully exploiting the complementarity between image fusion and semantic segmentation. Quantitative evaluations on multiple public datasets show that SC-CoSF outperforms traditional approaches that optimize tasks independently, improving both fused image quality and segmentation accuracy. Additionally, the collaborative training mechanism enhances inter-modal consistency, increasing model stability and robustness across diverse scenarios. These results demonstrate the potential of tightly coupled multimodal learning in advancing perception capabilities for autonomous driving and robotics.

Author Contributions

D.Y. conducted the experiment and wrote the paper. L.Q. conceived original ideas and structured the paper. Y.S. processed the raw data and analyzed the results. All authors have read and agreed to the published version of the manuscript.

Funding

This work received partial support from various funding sources, including the National Key Research and Development Project, China (Grant 2019YFE0110800), the National Natural Science Foundation of China (Grants 62276040, 62276041, 62221005 and 61976031), the National Key Research Instrument Development Program, China (Grant 62027827), Chongqing Education Commission Science and Technology Research Project, China (Grant KJQN202200624), and Chongqing Big Data Collaborative Innovation Center Funding (Grant CQBDCIC202303), Natural Science Foundation of Chongqing (Grant CSTB2022NSCQ-MSX0436), Chongqing Key Laboratory of Precision Diagnosis and Treatment for Kidney Diseases. (Corresponding author: Lihong Qiao).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, W.; Lin, X.; Lei, J.; Yu, L.; Hwang, J.N. MFFENet: Multiscale Feature Fusion and Enhancement Network For RGB–Thermal Urban Road Scene Parsing. IEEE Trans. Multimed. 2022, 24, 2526–2538. [Google Scholar] [CrossRef]
  2. Zhou, W.; Dong, S.; Xu, C.; Qian, Y. Edge-aware guidance fusion network for rgb–thermal scene parsing. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 3571–3579. [Google Scholar]
  3. Zhao, Z.; Bai, H.; Zhu, Y.; Zhang, J.; Xu, S.; Zhang, Y.; Zhang, K.; Meng, D.; Timofte, R.; Van Gool, L. DDFM: Denoising diffusion model for multi-modality image fusion. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 8082–8093. [Google Scholar]
  4. Luo, F.; Li, Y.; Zeng, G.; Peng, P.; Wang, G.; Li, Y. Thermal infrared image colorization for nighttime driving scenes with top-down guided attention. IEEE Trans. Intell. Transp. Syst. 2022, 23, 15808–15823. [Google Scholar] [CrossRef]
  5. Wang, Y.; Miao, L.; Zhou, Z.; Zhang, L.; Qiao, Y. Infrared and visible image fusion with language-driven loss in CLIP embedding space. arXiv 2024, arXiv:2402.16267. [Google Scholar]
  6. Tang, H.; Li, Z.; Zhang, D.; He, S.; Tang, J. Divide-and-Conquer: Confluent Triple-Flow Network for RGB-T Salient Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 1958–1974. [Google Scholar] [CrossRef]
  7. Li, H.; Yang, Z.; Zhang, Y.; Jia, W.; Yu, Z.; Liu, Y. MulFS-CAP: Multimodal Fusion-Supervised Cross-Modality Alignment Perception for Unregistered Infrared-Visible Image Fusion. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 3673–3690. [Google Scholar] [CrossRef]
  8. Zhou, M.; Zheng, N.; He, X.; Hong, D.; Chanussot, J. Probing Synergistic High-Order Interaction for Multi-Modal Image Fusion. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 840–857. [Google Scholar] [CrossRef]
  9. Liu, H.; Mao, Q.; Dong, M.; Zhan, Y. Infrared-Visible Image Fusion Using Dual-Branch Auto-Encoder With Invertible High-Frequency Encoding. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 2675–2688. [Google Scholar] [CrossRef]
  10. Zhao, Z.; Bai, H.; Zhang, J.; Zhang, Y.; Xu, S.; Lin, Z.; Timofte, R.; Van Gool, L. Cddfuse: Correlation-driven dual-branch feature decomposition for multi-modality image fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 5906–5916. [Google Scholar]
  11. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  12. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar]
  13. Gu, A.; Goel, K.; Ré, C. Efficiently modeling long sequences with structured state spaces. arXiv 2021, arXiv:2111.00396. [Google Scholar]
  14. Zhu, L.; Liao, B.; Zhang, Q.; Wang, X.; Liu, W.; Wang, X. Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model. arXiv 2024, arXiv:2401.09417. [Google Scholar]
  15. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Liu, Y. VMamba: Visual State Space Model. arXiv 2024, arXiv:2401.10166. [Google Scholar]
  16. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  17. Liu, J.; Liu, Z.; Wu, G.; Ma, L.; Liu, R.; Zhong, W.; Luo, Z.; Fan, X. Multi-interactive feature learning and a full-time multi-modality benchmark for image fusion and segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 8115–8124. [Google Scholar]
  18. Zhang, Y.; Yang, Q. An overview of multi-task learning. Natl. Sci. Rev. 2017, 5, 30–43. [Google Scholar] [CrossRef]
  19. Ruder, S. An overview of multi-task learning in deep neural networks. arXiv 2017, arXiv:1706.05098. [Google Scholar]
  20. Caruana, R. Multitask learning. Mach. Learn. 1997, 28, 41–75. [Google Scholar] [CrossRef]
  21. Vandenhende, S.; Georgoulis, S.; Van Gansbeke, W.; Proesmans, M.; Dai, D.; Van Gool, L. Multi-task learning for dense prediction tasks: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3614–3633. [Google Scholar] [CrossRef]
  22. Zhang, D.; Zheng, R. TriangleNet: Edge Prior Augmented Network for Semantic Segmentation through Cross-Task Consistency. Int. J. Intell. Syst. 2022, 2023, 1–16. [Google Scholar] [CrossRef]
  23. Gonçalves, D.N.; Junior, J.M.; Zamboni, P.; Pistori, H.; Li, J.; Nogueira, K.; Gonçalves, W.N. MTLSegFormer: Multi-task Learning with Transformers for Semantic Segmentation in Precision Agriculture. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Vancouver, BC, Canada, 17–24 June 2023; pp. 6290–6298. [Google Scholar]
  24. Zhao, Z.; Bai, H.; Zhang, J.; Zhang, Y.; Zhang, K.; Xu, S.; Chen, D.; Timofte, R.; Gool, L.V. Equivariant Multi-Modality Image Fusion. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 25912–25921. [Google Scholar]
  25. Li, X.; Li, X.; Ye, T.; Cheng, X.; Liu, W.; Tan, H. Bridging the Gap between Multi-focus and Multi-modal: A Focused Integration Framework for Multi-modal Image Fusion. In Proceedings of the 2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 1617–1626. [Google Scholar]
  26. Yuan, Y.; Wu, J.; Liang Jing, Z.; Leung, H.; Pan, H. Multimodal Image Fusion based on Hybrid CNN-Transformer and Non-local Cross-modal Attention. arXiv 2022, arXiv:2210.09847. [Google Scholar]
  27. Liu, H.; Zhang, J.; Yang, K.; Hu, X.; Stiefelhagen, R. CMX: Cross-Modal Fusion for RGB-X Semantic Segmentation with Transformers. IEEE Trans. Intell. Transp. Syst. 2022, 24, 14679–14694. [Google Scholar]
  28. Jiang, C.; Liu, X.; Zheng, B.; Bai, L.; Li, J. HSFusion: A high-level vision task-driven infrared and visible image fusion network via semantic and geometric domain transformation. arXiv 2024, arXiv:2407.10047. [Google Scholar]
  29. Alonso, C.A.; Sieber, J.; Zeilinger, M.N. State Space Models as Foundation Models: A Control Theoretic Overview. arXiv 2024, arXiv:2403.16899. [Google Scholar]
  30. Bragman, F.J.S.; Tanno, R.; Ourselin, S.; Alexander, D.C.; Cardoso, M.J. Stochastic Filter Groups for Multi-Task CNNs: Learning Specialist and Generalist Convolution Kernels. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1385–1394. [Google Scholar]
  31. Misra, I.; Shrivastava, A.; Gupta, A.; Hebert, M. Cross-stitch networks for multi-task learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3994–4003. [Google Scholar]
  32. Zhang, Z.; Cui, Z.; Xu, C.; Jie, Z.; Li, X.; Yang, J. Joint task-recursive learning for semantic segmentation and depth estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 235–251. [Google Scholar]
  33. Chen, Z.; Badrinarayanan, V.; Lee, C.Y.; Rabinovich, A. GradNorm: Gradient Normalization for Adaptive Loss Balancing in Deep Multitask Networks. arXiv 2017, arXiv:1711.02257. [Google Scholar]
  34. Jeong, J.; Lee, S.; Kim, J.; Kwak, N. Consistency-based Semi-supervised Learning for Object detection. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: San Francisco, CA, USA, 2019; Volume 32. [Google Scholar]
  35. Bu, Z.; Jin, X.; Vinzamuri, B.; Ramakrishna, A.; Chang, K.W.; Cevher, V.; Hong, M. Unlearning as multi-task optimization: A normalized gradient difference approach with an adaptive learning rate. arXiv 2024, arXiv:2410.22086. [Google Scholar]
  36. Mortaheb, M.; Vahapoglu, C.; Ulukus, S. FedGradNorm: Personalized Federated Gradient-Normalized Multi-Task Learning. In Proceedings of the 2022 IEEE 23rd International Workshop on Signal Processing Advances in Wireless Communication (SPAWC), Oulu, Finland, 4–6 July 2022; pp. 1–5. [Google Scholar]
  37. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2010, arXiv:2010.11929. [Google Scholar]
  38. Chen, C.F.; Fan, Q.; Panda, R. CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classification. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 347–356. [Google Scholar]
  39. Wu, D.; Wang, Y.; Wu, X.; Qu, T. Cross-attention Inspired Selective State Space Models for Target Sound Extraction. arXiv 2024, arXiv:2409.04803. [Google Scholar]
  40. Guo, M.H.; Lu, C.; Hou, Q.; Liu, Z.; Cheng, M.M.; Hu, S. SegNeXt: Rethinking Convolutional Attention Design for Semantic Segmentation. arXiv 2022, arXiv:2209.08575. [Google Scholar]
  41. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13708–13717. [Google Scholar]
  42. Zhang, H.; Zuo, X.; Jiang, J.; Guo, C.; Ma, J. Mrfs: Mutually reinforcing image fusion and segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 26974–26983. [Google Scholar]
  43. Ha, Q.; Watanabe, K.; Karasawa, T.; Ushiku, Y.; Harada, T. MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 1–24 September 2017; pp. 5108–5115. [Google Scholar]
  44. Liu, Z.; Mao, H.; Wu, C.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11966–11976. [Google Scholar]
  45. Wang, Y.; Li, G.; Liu, Z. SGFNet: Semantic-Guided Fusion Network for RGB-Thermal Semantic Segmentation. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 7737–7748. [Google Scholar] [CrossRef]
  46. Chen, Y.; Zhan, W.; Jiang, Y.; Zhu, D.; Guo, R.; Xu, X. LASNet: A light-weight asymmetric spatial feature network for real-time semantic segmentation. Electronics 2022, 11, 3238. [Google Scholar] [CrossRef]
  47. Liang, M.; Hu, J.; Bao, C.; Feng, H.; Deng, F.; Lam, T.L. Explicit Attention-Enhanced Fusion for RGB-Thermal Perception Tasks. IEEE Robot. Autom. Lett. 2023, 8, 4060–4067. [Google Scholar] [CrossRef]
  48. Zhao, S.; Liu, Y.; Jiao, Q.; Zhang, Q.; Han, J. Mitigating Modality Discrepancies for RGB-T Semantic Segmentation. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 9380–9394. [Google Scholar] [CrossRef] [PubMed]
  49. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef] [PubMed]
  50. Rao, D.; Xu, T.; Wu, X.J. TGFuse: An Infrared and Visible Image Fusion Approach Based on Transformer and Generative Adversarial Network. IEEE Trans. Image Process. 2023, 1. [Google Scholar] [CrossRef] [PubMed]
  51. Tang, W.; He, F.; Liu, Y.; Duan, Y.; Si, T. DATFuse: Infrared and visible image fusion via dual attention transformer. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3159–3172. [Google Scholar] [CrossRef]
  52. Roberts, J.W.; Van Aardt, J.A.; Ahmed, F.B. Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2008, 2, 023522. [Google Scholar]
  53. Aslantas, V.; Bendes, E. A new image quality metric for image fusion: The sum of the correlations of differences. Aeu-Int. J. Electron. Commun. 2015, 69, 1890–1896. [Google Scholar] [CrossRef]
Figure 1. Compared to previous fusion-segmentation frameworks, our method redefines the framework by integrating image fusion and semantic segmentation subnetworks. By introducing a joint training mechanism that optimizes both tasks simultaneously, our model is optimized using synchronous single-stage joint optimization. This parallel training structure effectively enhances the performance of both tasks during joint training, reduces training difficulty, and eliminates the performance bottleneck between tasks.
Figure 1. Compared to previous fusion-segmentation frameworks, our method redefines the framework by integrating image fusion and semantic segmentation subnetworks. By introducing a joint training mechanism that optimizes both tasks simultaneously, our model is optimized using synchronous single-stage joint optimization. This parallel training structure effectively enhances the performance of both tasks during joint training, reduces training difficulty, and eliminates the performance bottleneck between tasks.
Sensors 25 03575 g001
Figure 2. The overall framework of our proposed SC-CoSF.
Figure 2. The overall framework of our proposed SC-CoSF.
Sensors 25 03575 g002
Figure 3. The overall framework of Sc-CFM, consisting of (a) a Self-correction Long-Range Relationship Branch (Sc-LRB), (b) the principle of the S6 block, (c) a Self-correction Fine-Grained Branch (Sc-FGB), and (d) Dual-branch Collborative Recalibration (DCR).
Figure 3. The overall framework of Sc-CFM, consisting of (a) a Self-correction Long-Range Relationship Branch (Sc-LRB), (b) the principle of the S6 block, (c) a Self-correction Fine-Grained Branch (Sc-FGB), and (d) Dual-branch Collborative Recalibration (DCR).
Sensors 25 03575 g003
Figure 4. The Interactive Context Restoration Visual State Space Block (ICRVSSB) used in the Mamba decoder. The lower part of the image shows the architecture of ICRVSSB, while the upper part presents the structure of SS2D.
Figure 4. The Interactive Context Restoration Visual State Space Block (ICRVSSB) used in the Mamba decoder. The lower part of the image shows the architecture of ICRVSSB, while the upper part presents the structure of SS2D.
Sensors 25 03575 g004
Figure 5. Qualitative segmentation on the MFNet dataset.
Figure 5. Qualitative segmentation on the MFNet dataset.
Sensors 25 03575 g005
Figure 6. Qualitative segmentation on the FMB dataset.
Figure 6. Qualitative segmentation on the FMB dataset.
Sensors 25 03575 g006
Figure 7. Qualitative comparison results on the FMB dataset.
Figure 7. Qualitative comparison results on the FMB dataset.
Sensors 25 03575 g007
Figure 8. Qualitative comparison results on the FMB dataset.
Figure 8. Qualitative comparison results on the FMB dataset.
Sensors 25 03575 g008
Figure 9. Quantitative analysis results of image fusion.
Figure 9. Quantitative analysis results of image fusion.
Sensors 25 03575 g009
Figure 10. Qualitative segmentation of ablation studies.
Figure 10. Qualitative segmentation of ablation studies.
Sensors 25 03575 g010
Figure 11. Attention heatmap visualizations of ablation studies.
Figure 11. Attention heatmap visualizations of ablation studies.
Sensors 25 03575 g011
Figure 12. Qualitative fusion of ablation studies.
Figure 12. Qualitative fusion of ablation studies.
Sensors 25 03575 g012
Figure 13. Quantitative fusion of ablation studies.
Figure 13. Quantitative fusion of ablation studies.
Sensors 25 03575 g013
Table 1. Quantitative segmentation on the MFNet dataset.
Table 1. Quantitative segmentation on the MFNet dataset.
MFNetCarPersonBikeCurveCar StopGuar.ConeBumpmIoU
SeAFusion84.271.158.733.120.10.040.433.948.8
LASNet84.267.156.941.139.618.948.940.154.9
SegMiF87.871.463.247.531.10.048.950.556.1
MDRNet+87.169.860.947.837.86.257.156.056.8
SGFNet88.477.664.345.831.00.657.155.057.6
ConvNeXt89.171.962.344.343.00.051.752.657.0
EAEFNet87.672.663.848.635.014.252.458.358.9
Ours90.774.165.947.145.71.855.659.259.8
Table 2. Quantitative segmentation on the FMB dataset.
Table 2. Quantitative segmentation on the FMB dataset.
FMBCarPersonTruckT-LampT-SignBuil.Vege.PolemIoU
SeAFusion76.259.615.134.468.080.183.538.451.9
MDRNet+79.661.320.719.471.582.285.544.055.9
SegMiF79.031.125.949.174.780.184.649.556.0
LASNet79.454.732.223.270.682.186.145.356.1
ConvNeXt75.063.036.030.165.382.483.244.857.4
EAEFNet83.365.430.623.872.583.986.348.659.7
SGFNet78.168.945.045.072.183.185.845.460.4
Ours77.567.439.346.772.383.485.052.961.8
Table 3. Complexity and parameters of segmentation and fusion methods.
Table 3. Complexity and parameters of segmentation and fusion methods.
MethodFLOPs (G)Params (M)Task
MDRNet+891.82210.87Segmentation
LASNet371.0393.58Segmentation
ConvNeXt278.49114.36Segmentation
EAEFNet316.49147.21Segmentation
SGFNet225.63125.12Segmentation
CDDFuse863.221.19Image Fusion
DATFuse8.680.01Image Fusion
TGFuse137.3419.34Image Fusion
U2Fusion633.090.66Image Fusion
SeAFusion102.5313.06Segmentation & Image Fusion
SegMiF526.2045.60Segmentation & Image Fusion
Ours304.52139.62Segmentation & Image Fusion
Table 4. Quantitative segmentation of ablation studies.
Table 4. Quantitative segmentation of ablation studies.
ModelCarPersonBikeCurveCar StopGuar.ConeBumpmIoU
A88.071.465.542.240.30.250.448.156.0
B90.073.764.345.540.00.752.362.458.6
C88.861.064.639.142.41.751.958.556.2
D90.573.766.146.039.15.355.361.059.5
F90.573.866.043.840.61.154.863.959.2
Full Model90.774.165.947.145.71.855.659.259.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, D.; Qiao, L.; Shu, Y. SC-CoSF: Self-Correcting Collaborative and Co-Training for Image Fusion and Semantic Segmentation. Sensors 2025, 25, 3575. https://doi.org/10.3390/s25123575

AMA Style

Yang D, Qiao L, Shu Y. SC-CoSF: Self-Correcting Collaborative and Co-Training for Image Fusion and Semantic Segmentation. Sensors. 2025; 25(12):3575. https://doi.org/10.3390/s25123575

Chicago/Turabian Style

Yang, Dongrui, Lihong Qiao, and Yucheng Shu. 2025. "SC-CoSF: Self-Correcting Collaborative and Co-Training for Image Fusion and Semantic Segmentation" Sensors 25, no. 12: 3575. https://doi.org/10.3390/s25123575

APA Style

Yang, D., Qiao, L., & Shu, Y. (2025). SC-CoSF: Self-Correcting Collaborative and Co-Training for Image Fusion and Semantic Segmentation. Sensors, 25(12), 3575. https://doi.org/10.3390/s25123575

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop