Next Article in Journal
Research on Damage Identification in Transmission Tower Structures Based on Cross-Correlation Function Amplitude Vector
Previous Article in Journal
A Welding Defect Detection Model Based on Hybrid-Enhanced Multi-Granularity Spatiotemporal Representation Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rethinking Infrared and Visible Image Fusion from a Heterogeneous Content Synergistic Perception Perspective

1
School of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing 210000, China
2
College of Electronic and Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(15), 4658; https://doi.org/10.3390/s25154658
Submission received: 19 May 2025 / Revised: 20 July 2025 / Accepted: 21 July 2025 / Published: 27 July 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Infrared and visible image fusion (IVIF) endeavors to amalgamate the thermal radiation characteristics from infrared images with the fine-grained texture details from visible images, aiming to produce fused outputs that are more robust and information-rich. Among the existing methodologies, those based on generative adversarial networks (GANs) have demonstrated considerable promise. However, such approaches are frequently constrained by their reliance on homogeneous discriminators possessing identical architectures, a limitation that can precipitate the emergence of undesirable artifacts in the resultant fused images. To surmount this challenge, this paper introduces HCSPNet, a novel GAN-based framework. HCSPNet distinctively incorporates heterogeneous dual discriminators, meticulously engineered for the fusion of disparate source images inherent in the IVIF task. This architectural design ensures the steadfast preservation of critical information from the source inputs, even when faced with scenarios of image degradation. Specifically, the two structurally distinct discriminators within HCSPNet are augmented with adaptive salient information distillation (ASID) modules, each uniquely structured to align with the intrinsic properties of infrared and visible images. This mechanism impels the discriminators to concentrate on pivotal components during their assessment of whether the fused image has proficiently inherited significant information from the source modalities—namely, the salient thermal signatures from infrared imagery and the detailed textural content from visible imagery—thereby markedly diminishing the occurrence of unwanted artifacts. Comprehensive experimentation conducted across multiple publicly available datasets substantiates the preeminence and generalization capabilities of HCSPNet, underscoring its significant potential for practical deployment. Additionally, we also prove that our proposed heterogeneous dual discriminators can serve as a plug-and-play structure to improve the performance of existing GAN-based methods.

1. Introduction

The primary objective of infrared and visible image fusion (IVIF) is to synergistically combine the complementary strengths of two distinct imaging modalities, outlined as follows: infrared sensors, which capture thermal radiation to highlight objects, and visible sensors, which provide rich textural details of the scene [1,2,3,4,5]. Effective fusion yields a composite image that is not only more informative but also exhibits enhanced robustness in challenging or degraded environments, such as low-light conditions or scenes obscured by smoke. This improved representation is crucial for a variety of downstream visual tasks, including object detection, surveillance, and autonomous navigation. Indeed, emerging applications, for instance in smartphone photography, stand to benefit significantly from IVIF’s capacity to improve imaging quality and adaptability to diverse environmental conditions, improving real-time fusion on consumer devices equipped with multiple imaging sensors, as depicted in Figure 1.
Current IVIF methodologies can be broadly classified into traditional model-based techniques and contemporary deep learning-based approaches. Traditional methods often depend on handcrafted feature extraction operators (e.g., multi-scale transforms like wavelet or pyramid decompositions and sparse representations) and explicitly defined fusion rules [6]. While these methods can achieve satisfactory results under specific conditions, their inherent limitation lies in their restricted feature representation capabilities, and they often struggle to adapt to the vast diversity of real-world scenes. In contrast, deep learning approaches, particularly those leveraging convolutional neural networks (CNNs) and generative adversarial networks (GANs), have demonstrated superior performance due to their powerful non-linear modeling capabilities, learning intricate fusion strategies directly from data, and producing fused images that are often more visually natural and harmonious [7].
Despite these advancements, even state-of-the-art GAN-based IVIF methods can encounter difficulties when confronted with severely degraded input images, such as those where prominent objects are heavily obscured or the entire scene is affected by adverse conditions like dense fog or poor illumination. A common issue in such scenarios is the introduction of undesirable artifacts in the fused output, which can compromise the image quality and the efficacy of subsequent analyses (see Figure 2).
Upon a closer examination of the limitations within existing GAN frameworks for IVIF, a prevalent issue emerges as follows: the predominant use of homogeneous discriminators, i.e., discriminators sharing identical architectural structures. This design choice can inadvertently curtail the discriminative power of the network. If a discriminator cannot accurately differentiate and assess the unique contributions of each source modality within the fused image, the feedback provided to the generator may be suboptimal, potentially leading to the generation of these artifacts. Such a homogenized approach fails to adequately address the intrinsic differences in information content and statistical properties between infrared and visible spectra, thereby imposing a ceiling on achievable fusion performance.
To address these shortcomings, this paper introduces HCSPNet, a pioneering GAN-based framework that, for the first time in the IVIF domain, employs dual discriminators with heterogeneous structures. Our approach re-evaluates the IVIF problem through the lens of “Heterogeneous Content Synergistic Perception”. The central tenet of HCSPNet is to leverage purpose-built, structurally distinct discriminators to more accurately guide the generator in learning to preserve essential information from the source images while concurrently suppressing the formation of artifacts.
The innovative aspects of HCSPNet are manifested in the design of its unique dual-discriminator system. Firstly, each of the two discriminators incorporates a bespoke Adaptive Salient Information Distillation (ASID) module. The internal architecture of these ASID modules is specifically tailored to the distinct characteristics of infrared and visible images, respectively. Their function is to efficiently extract and emphasize the most salient features pertinent to each modality—typically thermal radiation patterns from infrared and intricate textures from visible light. This targeted distillation enables the discriminators to more effectively evaluate the fidelity with which the fused image preserves these critical source components, thereby providing more refined guidance to the generator, while this paper does not focus on changes to the generator’s structure, it highlights how the discriminator system can serve as a plug-and-play enhancement to existing GAN-based methods.
The main contributions of this work are:
  • We propose HCSPNet, the pioneering GAN-based framework for IVIF that utilizes heterogeneous dual discriminators. This design is specifically focused on maintaining the integrity of crucial information from source images, especially in challenging, degraded scenarios.
  • The two discriminators, with our proposed ASID module, are optimized for their respective sensor modalities, allowing for the simultaneous targeted learning of thermal radiation regions and local details. This framework can serve as a plug-and-play framework for existing GAN-based methods.
  • We have conducted extensive experiments on several public IVIF datasets, as well as on related tasks such as medical and biological image fusion. The results not only confirm the superior fusion quality of HCSPNet but also highlight its capability to improve performance in downstream high-level vision tasks and demonstrate the broad applicability of our proposed framework.
The remainder of this paper is organized as follows: Section 2 reviews related work in the field of IVIF. Section 3 details the proposed HCSPNet architecture and its core components. Section 4 presents and analyzes the comprehensive experimental results. Finally, Section 5 concludes the paper.

2. Related Works

2.1. Model-Based IVIF

Traditional IVIF methodologies focus on designing feature extraction operators and fusion rules to achieve optimal outcomes. These methods can be categorized into four main approaches. The first category is based on the principle that an image can be decomposed into distinct components due to the varying scales of different features. Effective fusion results are achieved by integrating these components using carefully designed rules, such as pyramid transforms [9,10], wavelet transforms [11,12], and curvelet transforms [13,14,15].
Sparse assumption, aiming to enhance the representation of valuable information, learns an over-complete dictionary and characterizes image features using sparse coefficients [16,17,18,19]. Subspace-based fusion methods, recognizing the redundancy inherent in most images, propose mapping images to a low-dimensional subspace to capture essential structural information while effectively reducing interference from redundant data [20,21]. Additionally, hybrid methods, which seek to improve fusion performance by combining the strengths of the aforementioned strategies, have been widely proposed [22,23].

2.2. Deep Learning-Based IVIF

Learning-based algorithms leverage the powerful fitting capabilities of neural networks, making them increasingly effective and prominent in the IVIF field. CNN-based frameworks are the most widely used. Liu et al. introduced a multi-focus fusion strategy [24], facilitating fusion tasks by learning the optimal information mapping. Building on this, Li et al. utilized a pre-trained ResNet-50 model to extract deep features from images [25], while Zhang et al. employed sequential convolutional layers to extract features and enhance the fusion results with additional information [26].
Another widely used strategy is based on autoencoders. Li et al. [27] incorporated dense connections in the encoder to enhance information utilization. Furthermore, Li et al. introduced a multi-scale approach that focused on features at various scales to achieve feature reuse [28]. Additionally, Wang et al. combined the strengths of previous models to enable better aggregation of useful features [29].
Despite the generally strong performance, the effectiveness of these methods is constrained by the lack of ground truth data. To address this limitation, Ma et al. introduced the GAN model [30], providing new insights into IVIF. This breakthrough led to the development of several enhanced methods [31,32,33], which will be further discussed in the following subsections.

2.3. GANs for IVIF

FusionGAN [30] was the first to introduce GANs into IVIF, providing a novel solution to the challenge of lacking ground truth images in the fusion process. However, FusionGAN primarily evaluates the generated images against visible images, which leads the generator to focus predominantly on the visible image distribution while neglecting the unique characteristics of infrared images.
To address this limitation, Ma et al. proposed DDcGAN [31], incorporating two homogeneous discriminators with identical structures to compel the generator to learn both the detailed features from visible images and the thermal radiation information from infrared images. This dual-discriminator GAN, specifically tailored for image fusion, has inspired the development of several subsequent models. Additionally, GAN-FM [32] enhanced feature learning through improved discriminator architectures, integrating a Markovian discriminator to emphasize local image block structures. Further advancements include TarDAL, presented by Liu et al. [8], which employs a joint training strategy for downstream detection tasks, achieving bi-level optimization between fusion and detection.
In contrast to existing methods, we propose HCSPNet, a pioneering GAN-based framework with heterogeneous discriminators. This novel framework is specifically designed to ensure that fused images effectively retain the essential components from the source images. The two heterogeneous discriminators in this framework serve as a plug-and-play solution for existing GAN-based methods. Extensive experiments demonstrate the superiority of our approach across IVIF, medical image fusion, and biological image fusion tasks, as well as its advancements in downstream high-level vision tasks. Moreover, we showcase the generalization capability of our heterogeneous dual-discriminator framework, which can be seamlessly integrated into existing GAN-based methods in a plug-and-play manner, leading to significant performance improvements.

3. Methodology

3.1. Overview

In this paper, we propose a novel GAN-based framework, termed HCSPNet, which rethinks the IVIF problem from a heterogeneous content synergistic perception perspective. As shown in Figure 3, HCSPNet employs two distinct discriminators, each specifically designed for heterogeneous source images, ensuring that the fused images effectively inherit key components from the source images—namely, the salient thermal radiation information from infrared images and the detailed texture information from visible images.

3.2. Generator

Given that this paper focuses on proposing a plug-and-play structure with heterogeneous dual discriminators, we follow the practice of existing methods [27,28,29] to employ a dense connection encoder–decoder structure for image fusion. As depicted in Figure 3, the encoder E and decoder D of our generator G both utilize a multi-scale skip-connected architecture to facilitate information extraction, fusion, and reconstruction, thereby enhancing the robustness of image fusion under degraded conditions. Given an infrared image I and a visible image V of size W × H , the encoder generates a set of feature maps { E i k } k = 1 4 and { E v k } k = 1 4 with spatial resolutions of H 2 k × W 2 k . These extracted features are subsequently passed to the decoder for feature aggregation and reconstruction. To accomplish this, we employ multiple Conv-ReLU-BatchNorm (CRB) modules in conjunction with skip connections. After passing through the final decoder layer, we obtain the fused image X, which is then evaluated by the dual heterogeneous discriminators.

3.3. Heterogeneous Discriminators

To effectively address the issue of artifact generation and to fully leverage the discriminative capacity, HCSPNet introduces a novel dual-discriminator framework. This framework comprises two structurally distinct discriminators: the salient discriminator D s , tailored for assessing infrared-related components, and the detailed discriminator D d , designed for visible-light-related components. This heterogeneous design is a cornerstone of the “Heterogeneous Content Synergistic Perception” approach. Specifically, our method incorporates adaptive salient information distillation (ASID) modules, with structures tailored to the unique characteristics of the source inputs. ASID facilitates the extraction of salient and significant information from the fused images, which encourages the discriminators to focus on key components when assessing whether the fused image effectively preserves critical information from the source images.

Adaptive Salient Information Distillation Module

To effectively extract salient information from the fused image, we employ the combined guidance of wavelet decomposition and adaptive attention. Given that the fused image must preserve thermal radiation information from the infrared image and detailed texture from the visible image, the low-frequency wavelet coefficients [16] are used to differentiate between the fused image and the infrared image. In contrast, high-frequency wavelets are used to distinguish between the fused and visible images. The adaptive attention mechanism combines spatial attention and channel attention, with their information integrated through a weighted gate mechanism.
As illustrated in Figure 3, we use the infrared image I as an example. The Adaptive Salient Information Distillation (ASID) generates the distilled content I s , which combines the guidance of the low-frequency information I w and the adaptive attention map I a . This approach helps the distilled content emphasize the most salient information. Initially, we generate the down-sampled distilled content I s d using a window-based linear model as follows:
I s d i = σ w down I i + μ w , i s w ,
where d o w n ( · ) , s w , and i represent the down-sampling operation, local window, and pixel point i, respectively. { σ w , μ w } are the linear aggregation coefficients for the pixels in window s w , which are optimized by the following objective function L h :
L h = i s w I a i 2 I s d i I w i 2 + ϵ σ w 2 ,
where ϵ is a constraint value for σ w . We define A H = I a i ( down ( I ) ) i , A L = I a i I w i , and p n as the number of the pixels in s w . Considering the partial derivatives of the optimization function L to σ w and μ w and locating the zero point, we calculate the optimized results for { σ w , μ w } as follows:
L h σ w = i s w [ 2 ϵ σ w + 2 down I i I a i 2 ( σ w down I i + μ w I w i ) ] = 0 , A H 2 ¯ p n × p k h i ¯ A H ¯ × A H ¯ + ϵ σ w = A H A L ¯ p n × p k h i ¯ A H ¯ × A L ¯ , σ w = A H A L ¯ p n × p k h i ¯ A H ¯ × A L ¯ A H 2 ¯ p n × p k h i ¯ A H ¯ × A H ¯ + ϵ ,
L h μ w = i s w [ I a i 2 ( σ w down I i + μ w I w i ) ] = 0 , p k h i ¯ μ w = A L ¯ σ w × A H ¯ , μ w = A L ¯ σ w × A H ¯ / p k h i ¯ ,
where · ¯ denote the average operation. For pixels covered by multiple windows, we average those window-wise coefficients { σ i , μ i } for pixel i. By assembling { σ i , μ i } into { σ i , μ i } , Equation (1) can be rewritten as follows:
I s d = σ i down ( I ) + μ i ,
where ⊙ represents the Hadamard product. We then up-sample { σ i , μ i } as { σ h , μ h } and acquire the high-resolution aggregated feature f k 1 h for enriching spatial details as follows:
I s = ASID s I , I w , I a = σ h I + μ h .
This process enables the extraction of salient information from visible, infrared, and fused images, encouraging the heterogeneous discriminators to focus on these salient components, thus facilitating the preservation of those important components from the original images and reducing undesired artifacts.

3.4. Optimization Strategy

Our HCSPNet is optimized by two parts of the loss functions, i.e., the generator loss ( L G ) and the discriminator loss ( L D ).
(1)
Generator Loss: Our generator loss consists of the adversarial loss ( L ad ) and the basic loss ( L b ), generating a high-quality fused image as follows:
L G = L ad + α L b ,
where α is the parameter used to balance the two terms. The adversarial loss is defined as
L ad = E X d X [ ( D s ( X ) 1 ) 2 ] + E X d X [ ( D d ( X ) 1 ) 2 ] ,
where E means expectation and d X is the distribution of X.
The basic loss L basic aims to constrain the preservation of important information from source images. Following [7], L basic is defined as
L b = L i + β L v ,
where
L i = X I F 2 , L v = X V 1 ,
where β is the trade-off parameter. · F 2 and · 1 denote the Frobenius norm and 1 norm.
(2)
Discriminator Loss: Discriminator loss L D is combined with the salient discriminator loss L D s and the detailed discriminator loss L D d .
Given that the two heterogeneous discriminators may aggravate the imbalance conflict when the generator learns different features, we jointly train the two discriminators instead of calculating them separately [34,35]. The weight parameter γ is introduced to allow the model to adaptively adjust the optimization to solve the training imbalance arising from the difference of the discriminative structure as follows:
L D = L D s + γ L D d ,
where
L D s = E I d I [ ( D s ( I s ) 1 ) 2 ] + E X d X [ ( D s ( X s ) ) 2 ] ,
L D d = E V d V [ ( D d ( V d ) 1 ) 2 ] + E X d X [ ( D d ( X d ) ) 2 ] ,
where X s = ASID s ( X ) , X d = ASID d ( X ) . d I and d V denote the distributions of I and V.

4. Experiment

4.1. Implementation Details

  • Datasets: Our empirical evaluation of HCSPNet was conducted on the following five publicly accessible IVIF datasets: TNO [36], INO [37], RoadScene [38], LLVIP [39], and M 3 F D [8]. For the TNO dataset (190 images), the latter 100 pairs were used for training and the rest for testing. From RoadScene (221 pairs), the initial 151 pairs formed the training set. For LLVIP (300 pairs selected as per DeRUN [7]) and M 3 F D (300 pairs), the first 100 pairs from each were allocated for training. To augment the training data, images were uniformly cropped into 128 × 128 pixel blocks, yielding a total of 14,503 training samples.
  • Comparative Methods: We compare our proposed method, HCSPNet, with nine state-of-the-art (SOTA) methods, including U2Fusion [38], FusionGAN [30], DDcGAN [31], GANMcC [40], GAN-FM [32], TarDAL [8], DeRUN [7], GAN-HA [41], DSFD [42], and ReFusion [43]. The selection emphasizes GAN-based techniques to provide a direct comparison for our novel discriminative framework, supplemented by U2Fusion [38] and DeRUN [7] for broader context. All comparisons utilized publicly available code and their originally parameter settings.
  • Evaluation Metric: Following the approaches outlined in [7,17], we utilize six commonly used metrics to evaluate the performance of IVIF methods from multiple perspectives. These metrics include entropy (EN) [44], average gradient (AG) [45], spatial frequency (SF) [46], feature mutual information (FMI) [47], visual information fidelity (VIF) [48], and the universal image quality index (UIQI). EN measures information richness; AG reflects texture detail via gradients; SF indicates the overall activity level and edge information; FMI quantifies shared information between fused and source images; VIF assesses perceptual similarity to source images; and UIQI evaluates structural similarity, luminance, and contrast. Higher values for all these metrics signify better fusion quality.
  • Training Details: The proposed HCSPNet is implemented by PyTorch on two RTX4090 GPUs and is optimized by an Adam optimizer with momentum terms (0.9, 0.999). Our epoch is set as 160 with a batch size of 8. The learning rates are initialized as 2 × 10 4 and 1 × 10 4 for the generator and the heterogeneous discriminators, respectively. Furthermore, the hyperparameters α , β , and γ in Equations (7), (9), and (11) are set to 10, 1, and 1, respectively.

4.2. Results on the IVIF Task

  • Quantitative Comparisons: The average scores for the six evaluation metrics across the five datasets are presented in Table 1. HCSPNet consistently achieved the highest scores for nearly all metrics across all datasets. This demonstrates its superior capability in preserving gradients, textures, enhancing visual quality, and maximizing information content. Notably, HCSPNet often showed substantial improvements over the cutting-edge methods, ReFusion an DSFD, underscoring the efficacy of the heterogeneous content synergistic perception framework.
  • Qualitative Comparisons: To demonstrate the superiority of our model, we qualitatively assess three representative image pairs from the test set and compare the results with the nine state-of-the-art (SOTA) methods. The fusion outcomes of HCSPNet and the comparative algorithms are illustrated in Figure 4 and Figure 5. The evaluation focuses on salient thermal information, represented by bright human body regions, and texture details, such as lane lines, license plates, and bushes.
From a visual perspective, HCSPNet demonstrates several distinct advantages. First, our method closely aligns with the requirements of IVIF by producing fused images that incorporate both infrared and visible information. Notably, HCSPNet preserves a pixel distribution similar to infrared images while maintaining significant texture information from visible images. Second, HCSPNet outperforms other SOTA algorithms in capturing more salient thermal region details and clearer texture information. This can be attributed to the dual discriminator structures designed in our model, which enhance its ability to learn from the source images.
In contrast, competing methods that employ homogeneous discriminators, such as TarDAL, struggle to reconcile information from these disparate sources, often leading to undesirable artifacts. For instance, as shown in Figure 4, TarDAL produces a blurred representation of the human subject under uneven lighting. Furthermore, in Figure 5, it generates unnatural textures on the surface of the moving vehicle, demonstrating a failure to correctly interpret the salient features from the source inputs.

4.3. Ablation Study

In this subsection, we conduct ablation studies to validate the efficacy of the HCSPNet structure on the RoadScene dataset, aiming to highlight the practice of our method.
  • Effect of our discriminators: We employ dual heterogeneous discriminators, namely, the salient discriminator ( D s ) and the detailed discriminator ( D d ), to approach the IVIF problem from a heterogeneous content synergistic perception perspective. As demonstrated in Table 2, our method achieves the highest performance when both D s and D d are utilized, underscoring the effectiveness and superiority of our dual-discriminator framework.
  • Generalization of our discriminators: Our dual heterogeneous discriminators, equipped with ASID modules, enhance the extraction of salient information from the fused images, directing the discriminators to concentrate on key components when determining whether the fused images effectively preserve crucial information from the source images. Consequently, our dual-discriminator framework ensures that the generator retains essential content from the source images. In this subsection, we further integrate our dual-discriminator structure into existing GAN-based methods. As shown in Table 3, we observe significant performance improvements when incorporating our framework into other methods, further validating the superiority of our approach.
  • Sensitivity analysis of hyperparameters: In our framework, the final loss function is balanced by three key hyperparameters: α , β , and γ . To analyze their impact on model performance, we conduct a sensitivity analysis by varying each parameter while keeping the others at their default values ( α = 10 ,   β = 1 ,   γ = 1 ). The quantitative results of this analysis on the RoadScene dataset are presented in Table 4. As shown, our chosen set of parameters consistently yields the best or nearly the best results across all six evaluation metrics (EN, AG, SF, FMI, VIF, and UIQI), while performance varies with different hyperparameter values, the model demonstrates robust performance across a reasonable range, indicating that it is not overly sensitive to the precise choice of these parameters. This analysis confirms the rationale behind our selected hyperparameters, which effectively balance the contributions of the adversarial loss, the content loss, and the dual-discriminator losses to achieve optimal fusion quality.

4.4. Subsequent Application Verification

In addition to its role as a low-level vision task, IVIF can also function as a pre-processing step that integrates thermal information from infrared images with texture details from visible images. This fusion enhances the performance of high-level visual tasks. To validate this, we apply the proposed model to both object detection and object tracking tasks.
  • Object detection: Object detection is a widely used machine vision technique [49] for identifying specific objects, such as people, vehicles, and other important entities, within images. Fused images with enhanced quality of salient information are expected to yield better downstream performance compared to individual images. Furthermore, algorithms that effectively preserve salient information from heterogeneous source images tend to outperform others in object detection tasks. Following the methodology in [7,50], we employ the general-purpose detector YOLOv5 for object detection experiments on the LLVIP and M 3 F D datasets. The results for these two datasets are presented in Table 5 and Figure 6. In this context, Precision indicates the probability of correctly detecting true positives among all detected objects, with higher precision implying more accurate identification of real samples. Recall measures the probability of correctly recognizing all positive samples, with higher recall indicating fewer missed targets. The mean average precision (mAP) serves as a comprehensive metric that balances Precision and Recall to assess model performance. The mAP score ranges from 0 to 1, with values closer to 1 representing better performance. The mAP@.5 and mAP@.95 correspond to the mAP values at confidence thresholds of 0.5 and 0.95, respectively. The results clearly demonstrate that HCSPNet achieves superior performance across these metrics.
  • Visual tracking: Visual tracking, which aims to localize the position of specific objects in video sequences, is a crucial task in computer vision. Following the methodology in [51], we apply our algorithm to the VOT-RGBT sub-challenge, which focuses on short-term tracking as part of VOT2020, to comparatively assess the impact of the fusion approach on tracking performance. The VOT-RGBT benchmark, as outlined in [33,52], consists of 60 video sequences with aligned infrared and visible image pairs. For evaluation, we employ the following two trackers: the learning adaptive discriminative correlation filter (LADCF) [53] and the group feature selection and discriminative filter (GFSDCF) [54]. LADCF demonstrated top performance in the VOT-ST2018 sub-challenge, while GFSDCF further refines LADCF by eliminating redundant information using a group feature selection strategy. Both trackers’ codes are publicly available. For quantitative analysis, we use the following three metrics to evaluate tracking performance: Accuracy, Failure, and Expected Average Overlap (EAO). Accuracy measures the average overlap between the predicted bounding boxes and the ground truth, while Failure assesses tracker robustness. EAO evaluates the expected average overlap between the tracker’s output and the ground truth over time. For both Accuracy and EAO, higher values indicate better performance, while lower Failure values signify better robustness. The results are shown in Table 6 and Figure 7. HCSPNet demonstrates superior performance in enhancing visual tracking compared to other fusion methods.

4.5. Extension of the Proposed HCSPNet

We further validate the effect of our methods in the medical image fusion and biological image fusion tasks, which are highly similar to the infrared and visible image fusion task for the heterogeneous sources. Following the common practice, we adopt EN, AG, VIF, and UIQI to evaluate the performance.
(1)
Medical Image Fusion: For the medical image fusion task, we conduct experiments on PET and MRI image fusion using the HWBA dataset [55], which consists of 144 image pairs for training. We utilize the same evaluation metrics and comparison methods as applied in the IVIF task. The results, as shown in Table 7, indicate that our method achieves superior performance across all metrics, significantly surpassing existing methods. This enhanced performance is primarily due to our proposed heterogeneous content synergistic perception framework and the specifically designed attention module. Visual results of the medical image fusion are presented in Figure 8. These figures illustrate that HCSPNet provides better fusion outcomes with improved salient information compared to other methods, which may suffer from issues related to detail distortion and color component preservation.
(2)
Biological Image Fusion: In the biological image fusion task, we use the ATC dataset [56] for network training, focusing on the green fluorescent protein (GFP) and phase contrast (PC) image fusion (GFP) task. The training set consists of 60 image pairs. Quantitative results are reported in Table 8. As depicted in Table 8, our method achieves the best performance across all metrics. Additionally, visual comparisons are provided in Figure 9. These figures demonstrate that our approach effectively preserves salient information from GFP and PC images while mitigating phase noise issues from PC images.

5. Discussion

Compared to GAN-HA [41], a concurrent work that also addresses the limitations of homogeneous dual discriminators, our method presents clear and significant differences. First, GAN-HA [41] takes a relatively straightforward approach by assuming that crucial information in infrared and visible images is primarily located at the channel and spatial levels. To extract this information, it employs channel attention and spatial attention mechanisms. In contrast, our method introduces an Adaptive Salient Information Distillation (ASID) module, which integrates wavelet decomposition and adaptive attention to guide the extraction of both high-frequency and low-frequency components. This approach enables the heterogeneous discriminators to focus more effectively on salient features, enhancing the preservation of critical details from the original images and further reducing unwanted artifacts. These differences clearly distinguish our approach from that of GAN-HA [41]. Additionally, quantitative analysis confirms that our method consistently outperforms GAN-HA. When we replace our discriminators with those used in GAN-HA, we observe noticeable performance gains. In summary, we propose a novel GAN-based structure with heterogeneous dual discriminators, which can serve as a plug-and-play framework for existing GAN-based methods.

6. Future Work

While HCSPNet has demonstrated superior performance on the task of infrared and visible image fusion, its core principle of “Heterogeneous Content Synergistic Perception” holds significant promise for other challenging image fusion tasks. A particularly promising application domain is remote sensing image fusion, specifically the fusion of panchromatic (Pan) and multispectral (MS) images. In this scenario, the panchromatic image provides high-resolution spatial details, while the multispectral image contains rich spectral information, creating a complementary relationship highly analogous to that of visible and infrared images.
Therefore, our proposed heterogeneous dual-discriminator framework could be adapted to specifically evaluate whether the fused image simultaneously preserves the spatial fidelity of the Pan image and the spectral fidelity of the MS image—a critical requirement for tasks such as pansharpening. Furthermore, addressing related challenges within this domain, such as noise, could be integrated into our framework. For instance, techniques explored in pan-denoising research (e.g., PWRCTV [57]) represent a relevant direction for integration. Consequently, extending and applying the HCSPNet framework to the field of remote sensing image fusion represents a compelling avenue for future investigation.

7. Conclusions

In this paper, we have introduced HCSPNet, a novel generative adversarial network framework designed to address critical limitations in existing infrared and visible image fusion techniques. Traditional GAN-based methods, often constrained by homogeneous discriminator architectures, are prone to generating artifacts and may struggle in scenarios with degraded image quality. HCSPNet overcomes these issues by pioneering a “Heterogeneous Content Synergistic Perception” approach, which materializes as a dual-discriminator system where each discriminator is structurally distinct and specifically tailored to the unique characteristics of either infrared or visible source imagery. These specialized discriminators are further enhanced by ASID modules, which direct their focus towards the most critical information within each modality—thermal signatures and textural details. The comprehensive experimental evaluations conducted on multiple public IVIF datasets, as well as on analogous medical and biological image fusion tasks, unequivocally demonstrate the superiority of HCSPNet. Furthermore, the application of HCSPNet-fused images has shown tangible benefits in downstream tasks such as object detection and visual tracking, highlighting its practical utility. The successful generalization of its core principles to other fusion domains further underscores the robustness and broad applicability of our proposed framework. HCSPNet, therefore, represents a significant advancement in the field of multi-sensor image fusion, offering a more effective and reliable solution for integrating information from heterogeneous sensor sources.

Author Contributions

Conceptualization, G.H. and M.J.; methodology, G.H.; software, M.S.; validation, M.S.; formal analysis, G.H.; investigation, G.H.; resources, K.-K.M.; data curation, K.-K.M.; writing—original draft preparation, G.H.; writing—review and editing, M.J. and K.-K.M.; visualization, K.-K.M.; supervision, K.-K.M.; project administration, K.-K.M.; funding acquisition, M.J. and K.-K.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by The National Natural Science Foundation of China (62471253), the General Project of Natural Science Research in Colleges and Universities of Jiangsu Province (23KJB520027), and the Research Fund of Nanjing University of Posts and Telecommunications (NY222160).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ju, M.; Yu, X. Semantic-Aware Low-Light Image Enhancement Network for Recognizing Semantics in Intelligent Transportation Systems. IEEE Trans. Intell. Transp. Syst. 2025; early access. [Google Scholar] [CrossRef]
  2. He, C.; Shen, Y.; Fang, C.; Xiao, F.; Tang, L.; Zhang, Y.; Zuo, W.; Guo, Z.; Li, X. Diffusion Models in Low-Level Vision: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 47, 4630–4651. [Google Scholar] [CrossRef]
  3. Zhu, H.; Wu, H.; He, D.; Lan, R.; Liu, Z.; Pan, X. AcFusion: Infrared and Visible Image Fusion Based on Self-Attention and Convolution With Enhanced Information Extraction. IEEE Trans. Consum. Electron. 2023, 70, 4155–4167. [Google Scholar] [CrossRef]
  4. Li, H.; Fu, Y. FCDFusion: A fast, low color deviation method for fusing visible and infrared image pairs. Comput. Vis. Media 2025, 11, 195–211. [Google Scholar] [CrossRef]
  5. Li, H.; Wu, Z.; Shao, R.; Zhang, T.; Fu, Y. Noise Calibration and Spatial-Frequency Interactive Network for STEM Image Enhancement. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 11–15 June 2025; pp. 21287–21296. [Google Scholar]
  6. Ju, M.; Xie, S.; Li, F. Improving skip connection in u-net through fusion perspective with mamba for image dehazing. IEEE Trans. Consum. Electron. 2024, 70, 7505–7514. [Google Scholar] [CrossRef]
  7. He, C.; Li, K.; Xu, G.; Zhang, Y.; Hu, R.; Guo, Z.; Li, X. Degradation-Resistant Unfolding Network for Heterogeneous Image Fusion. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 12611–12621. [Google Scholar]
  8. Liu, J.; Fan, X.; Huang, Z.; Wu, G.; Liu, R.; Zhong, W.; Luo, Z. Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5802–5811. [Google Scholar]
  9. Burt, P.J.; Adelson, E.H. The Laplacian pyramid as a compact image code. In Readings in Computer Vision; Elsevier: Amsterdam, The Netherlands, 1987; pp. 671–679. [Google Scholar]
  10. Ju, M.; Ding, C.; Guo, Y.J.; Zhang, D. IDGCP: Image dehazing based on gamma correction prior. IEEE Trans. Image Process. 2019, 29, 3104–3118. [Google Scholar] [CrossRef] [PubMed]
  11. Lewis, J.J.; O’Callaghan, R.J.; Nikolov, S.G.; Bull, D.R.; Canagarajah, N. Pixel-and region-based image fusion with complex wavelets. Inf. Fusion 2007, 8, 119–130. [Google Scholar] [CrossRef]
  12. Li, K.; Xu, G.; Yan, J.; Tang, L.; Zhang, Y.; Wang, Y.; Li, X. Hqg-net: Unpaired medical image enhancement with high-quality guidance. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 18404–18418. [Google Scholar]
  13. Nencini, F.; Garzelli, A.; Baronti, S.; Alparone, L. Remote sensing image fusion using the curvelet transform. Inf. Fusion 2007, 8, 143–156. [Google Scholar] [CrossRef]
  14. Ju, M.; Ding, C.; Guo, C.A.; Ren, W.; Tao, D. IDRLP: Image dehazing using region line prior. IEEE Trans. Image Process. 2021, 30, 9043–9057. [Google Scholar] [CrossRef]
  15. Ju, M.; Ding, C.; Ren, W.; Yang, Y. IDBP: Image dehazing using blended priors including non-local, local, and global priors. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4867–4871. [Google Scholar] [CrossRef]
  16. He, C.; Li, K.; Zhang, Y.; Tang, L.; Zhang, Y.; Guo, Z.; Li, X. Camouflaged object detection with feature decomposition and edge reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22046–22055. [Google Scholar]
  17. Ju, M.; Ding, C.; Ren, W.; Yang, Y.; Zhang, D.; Guo, Y.J. IDE: Image dehazing and exposure using an enhanced atmospheric scattering model. IEEE Trans. Image Process. 2021, 30, 2180–2192. [Google Scholar] [CrossRef]
  18. Xiong, Z.; Zhang, X.; Hu, Q.; Han, H. RepVGGFuse: An approach for infrared and visible image fusion network based on RepVGG architecture. In Proceedings of the 2023 4th International Conference on Computing, Networks and Internet of Things, Xiamen, China, 26–28 May 2023; pp. 375–379. [Google Scholar]
  19. He, C.; Li, K.; Zhang, Y.; Yang, Z.; Tang, L.; Zhang, Y.; Kong, L.; Farsiu, S. Segment concealed object with incomplete supervision. IEEE Trans. Pattern Anal. Mach. Intell. 2025; early access. [Google Scholar] [CrossRef]
  20. Zheng, Y.; Essock, E.A.; Hansen, B.C. An advanced image fusion algorithm based on wavelet transform: Incorporation with PCA and morphological processing. In Proceedings of the Image Processing: Algorithms and Systems III, San Jose, CA, USA, 18–22 January 2004; SPIE: Bellingham, WA, USA, 2004; Volume 5298, pp. 177–187. [Google Scholar]
  21. Cvejic, N.; Bull, D.; Canagarajah, N. Region-based multimodal image fusion using ICA bases. IEEE Sens. J. 2007, 7, 743–751. [Google Scholar] [CrossRef]
  22. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  23. He, C.; Li, K.; Zhang, Y.; Xu, G.; Tang, L.; Zhang, Y.; Guo, Z.; Li, X. Weakly-Supervised Concealed Object Segmentation with SAM-based Pseudo Labeling and Multi-scale Feature Grouping. Adv. Neural Inf. Process. Syst. 2023, 36, 30726–30737. [Google Scholar]
  24. Liu, Y.; Chen, X.; Peng, H.; Wang, Z. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  25. Li, H.; Wu, X.-j.; Durrani, T.S. Infrared and visible image fusion with ResNet and zero-phase component analysis. Infrared Phys. Technol. 2019, 102, 103039. [Google Scholar] [CrossRef]
  26. Zhang, Y.; Liu, Y.; Sun, P.; Yan, H.; Zhao, X.; Zhang, L. IFCNN: A general image fusion framework based on convolutional neural network. Inf. Fusion 2020, 54, 99–118. [Google Scholar] [CrossRef]
  27. Li, H.; Wu, X.J. DenseFuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 2018, 28, 2614–2623. [Google Scholar] [CrossRef] [PubMed]
  28. Li, H.; Wu, X.J.; Durrani, T. NestFuse: An infrared and visible image fusion architecture based on nest connection and spatial/channel attention models. IEEE Trans. Instrum. Meas. 2020, 69, 9645–9656. [Google Scholar] [CrossRef]
  29. Wang, Z.; Wang, J.; Wu, Y.; Xu, J.; Zhang, X. UNFusion: A unified multi-scale densely connected network for infrared and visible image fusion. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 3360–3374. [Google Scholar] [CrossRef]
  30. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  31. Ma, J.; Xu, H.; Jiang, J.; Mei, X.; Zhang, X.P. DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 2020, 29, 4980–4995. [Google Scholar] [CrossRef]
  32. Zhang, H.; Yuan, J.; Tian, X.; Ma, J. GAN-FM: Infrared and visible image fusion using GAN with full-scale skip connection and dual Markovian discriminators. IEEE Trans. Comput. Imaging 2021, 7, 1134–1147. [Google Scholar] [CrossRef]
  33. He, C.; Zhang, R.; Xiao, F.; Fang, C.; Tang, L.; Zhang, Y.; Kong, L.; Fan, D.P.; Li, K.; Farsiu, S. RUN: Reversible Unfolding Network for Concealed Object Segmentation. arXiv 2025, arXiv:2501.18783. [Google Scholar]
  34. Zheng, Y.; Zhong, B.; Liang, Q.; Mo, Z.; Zhang, S.; Li, X. Odtrack: Online dense temporal token learning for visual tracking. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 26–27 February 2024; Volume 38, pp. 7588–7596. [Google Scholar]
  35. Zheng, Y.; Zhong, B.; Liang, Q.; Li, G.; Ji, R.; Li, X. Toward unified token learning for vision-language tracking. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 2125–2135. [Google Scholar] [CrossRef]
  36. Toet, A. The TNO multiband image data collection. Data Brief 2017, 15, 249–251. [Google Scholar] [CrossRef] [PubMed]
  37. Ma, J.; Chen, C.; Li, C.; Huang, J. Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fusion 2016, 31, 100–109. [Google Scholar] [CrossRef]
  38. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 502–518. [Google Scholar] [CrossRef]
  39. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A visible-infrared paired dataset for low-light vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3496–3504. [Google Scholar]
  40. Ma, J.; Zhang, H.; Shao, Z.; Liang, P.; Xu, H. GANMcC: A generative adversarial network with multiclassification constraints for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 2020, 70, 5005014. [Google Scholar] [CrossRef]
  41. Lu, G.; Fang, Z.; Tian, J.; Huang, H.; Xu, Y.; Han, Z.; Kang, Y.; Feng, C.; Zhao, Z. GAN-HA: A generative adversarial network with a novel heterogeneous dual-discriminator network and a new attention-based fusion strategy for infrared and visible image fusion. Infrared Phys. Technol. 2024, 142, 105548. [Google Scholar] [CrossRef]
  42. Hu, T.; Nan, X.; Zhou, X.; Shen, Y.; Zhou, Q. A dual-stream feature decomposition network with weight transformation for multi-modality image fusion. Sci. Rep. 2025, 15, 7467. [Google Scholar] [CrossRef]
  43. Bai, H.; Zhao, Z.; Zhang, J.; Wu, Y.; Deng, L.; Cui, Y.; Jiang, B.; Xu, S. Refusion: Learning image fusion from reconstruction with learnable loss via meta-learning. Int. J. Comput. Vis. 2025, 133, 2547–2567. [Google Scholar] [CrossRef]
  44. Lu, S.; Liu, Y.; Kong, A.W.K. Tf-icon: Diffusion-based training-free cross-domain image composition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 2294–2305. [Google Scholar]
  45. Lu, S.; Wang, Z.; Li, L.; Liu, Y.; Kong, A.W.K. Mace: Mass concept erasure in diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 6430–6440. [Google Scholar]
  46. Lu, S.; Zhou, Z.; Lu, J.; Zhu, Y.; Kong, A.W.K. Robust watermarking using generative priors against image editing: From benchmarking to advances. arXiv 2024, arXiv:2410.18775. [Google Scholar]
  47. Zheng, Y.; Zhong, B.; Liang, Q.; Tang, Z.; Ji, R.; Li, X. Leveraging local and global cues for visual tracking via parallel interaction network. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 1671–1683. [Google Scholar] [CrossRef]
  48. Zheng, Y.; Zhong, B.; Liang, Q.; Li, N.; Song, S. Decoupled Spatio-Temporal Consistency Learning for Self-Supervised Tracking. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 10635–10643. [Google Scholar]
  49. Liu, L.; Ouyang, W.; Wang, X.; Fieguth, P.; Chen, J.; Liu, X.; Pietikäinen, M. Deep learning for generic object detection: A survey. Int. J. Comput. Vis. 2020, 128, 261–318. [Google Scholar] [CrossRef]
  50. Ju, M.; He, C.; Ding, C.; Ren, W.; Zhang, L.; Ma, K.K. All-inclusive image enhancement for degraded images exhibiting low-frequency corruption. IEEE Trans. Circuits Syst. Video Technol. 2024, 35, 838–856. [Google Scholar] [CrossRef]
  51. Li, H.; Wu, X.J.; Kittler, J. MDLatLRR: A novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 2020, 29, 4733–4746. [Google Scholar] [CrossRef]
  52. Kristan, M.; Matas, J.; Leonardis, A.; Felsberg, M.; Pflugfelder, R.; Kamarainen, J.K.; Cehovin Zajc, L.; Drbohlav, O.; Lukezic, A.; Berg, A.; et al. The seventh visual object tracking VOT2019 challenge results. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  53. Xu, T.; Feng, Z.H.; Wu, X.J.; Kittler, J. Learning adaptive discriminative correlation filters via temporal consistency preserving spatial feature selection for robust visual object tracking. IEEE Trans. Image Process. 2019, 28, 5596–5609. [Google Scholar] [CrossRef]
  54. Xu, T.; Feng, Z.H.; Wu, X.J.; Kittler, J. Joint group feature selection and discriminative filter learning for robust visual object tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 7950–7960. [Google Scholar]
  55. Johnson, K.A.; Becker, J.A. The Whole Brain Atlas. 1997. Available online: https://www.med.harvard.edu/aanlib/ (accessed on 19 May 2025).
  56. Koroleva, O.A.; Tomlinson, M.L.; Leader, D.; Shaw, P.; Doonan, J.H. High-throughput protein localization in Arabidopsis using Agrobacterium-mediated transient expression of GFP-ORF fusions. Plant J. 2005, 41, 162–174. [Google Scholar] [CrossRef] [PubMed]
  57. Xu, S.; Ke, Q.; Peng, J.; Cao, X.; Zhao, Z. Pan-Denoising: Guided Hyperspectral Image Denoising via Weighted Represent Coefficient Total Variation. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5528714. [Google Scholar] [CrossRef]
Figure 1. System architecture of the proposed method: (a) Integration of smartphone visible sensors with infrared sensors, and (b) is the fusion procedure of our method.
Figure 1. System architecture of the proposed method: (a) Integration of smartphone visible sensors with infrared sensors, and (b) is the fusion procedure of our method.
Sensors 25 04658 g001
Figure 2. IVIF results under degraded scenarios. The state-of-the-art GAN-based method, TarDAL [8], fails to fully integrate salient information from the source images and often produces undesired artifacts, such as blurring and unnatural texture. In contrast, HCSPNet produces more visually appealing results, owing to its dual heterogeneous discriminators.
Figure 2. IVIF results under degraded scenarios. The state-of-the-art GAN-based method, TarDAL [8], fails to fully integrate salient information from the source images and often produces undesired artifacts, such as blurring and unnatural texture. In contrast, HCSPNet produces more visually appealing results, owing to its dual heterogeneous discriminators.
Sensors 25 04658 g002
Figure 3. (a) Framework of our proposed HCSPNet. (b,c) Generator and adaptive salient information distillation (ASID) module. Note that we take I as the input for an example and the actual input can be I, V, and X.
Figure 3. (a) Framework of our proposed HCSPNet. (b,c) Generator and adaptive salient information distillation (ASID) module. Note that we take I as the input for an example and the actual input can be I, V, and X.
Sensors 25 04658 g003
Figure 4. Visual comparisons in low-light scenarios. (a,b) Infrared image and visible images. (cj) Fused images processed by comparison methods.
Figure 4. Visual comparisons in low-light scenarios. (a,b) Infrared image and visible images. (cj) Fused images processed by comparison methods.
Sensors 25 04658 g004
Figure 5. Visual comparisons in glare scenarios. (a,b) Infrared image and visible images. (cj) Fused images processed by comparison methods.
Figure 5. Visual comparisons in glare scenarios. (a,b) Infrared image and visible images. (cj) Fused images processed by comparison methods.
Sensors 25 04658 g005
Figure 6. Object detection on fused images, where we select yolov5 as the detector.
Figure 6. Object detection on fused images, where we select yolov5 as the detector.
Sensors 25 04658 g006
Figure 7. Object tracking on fused images, where we select GFSDCF as the tracker. For easy observation and comparison, we follow the procedure of MDLatLRR [51] and add all tracking notations from different methods into the visible image. In the resulting figures, the bounding boxes are color-coded as follows: the ground truth is denoted by a red box, while the tracking results for DeRUN, GAN-HA, and our proposed HCSPNet are indicated by purple, green, and blue boxes, respectively.
Figure 7. Object tracking on fused images, where we select GFSDCF as the tracker. For easy observation and comparison, we follow the procedure of MDLatLRR [51] and add all tracking notations from different methods into the visible image. In the resulting figures, the bounding boxes are color-coded as follows: the ground truth is denoted by a red box, while the tracking results for DeRUN, GAN-HA, and our proposed HCSPNet are indicated by purple, green, and blue boxes, respectively.
Sensors 25 04658 g007
Figure 8. Visual comparisons of medical image fusion. (a,b) PET and MRI images. (cj) Fused images from the compared methods.
Figure 8. Visual comparisons of medical image fusion. (a,b) PET and MRI images. (cj) Fused images from the compared methods.
Sensors 25 04658 g008
Figure 9. Visual comparisons of biological image fusion. (a,b) GFP and PC images. (cj) Fused images from the compared methods.
Figure 9. Visual comparisons of biological image fusion. (a,b) GFP and PC images. (cj) Fused images from the compared methods.
Sensors 25 04658 g009
Table 1. Average scores of the six metrics in INO, TNO, RoadScene, LLVIP, and M 3 F D datasets, with the best result in red. * Indicates the method does not release the official code.
Table 1. Average scores of the six metrics in INO, TNO, RoadScene, LLVIP, and M 3 F D datasets, with the best result in red. * Indicates the method does not release the official code.
DataMetricsU2FusionFusionGANDDcGANGANMcCGAN-FMTarDALDeRUNGAN-HADSFD *ReFusionHCSPNet
TNOEN ↑6.6126.8286.6196.7586.3846.5676.8416.2656.6756.974
AG ↑4.0163.7593.9532.6662.2962.4934.1514.0324.2084.369
SF ↑10.6778.88811.05110.67411.87412.16212.82511.76711.18613.720
FMI ↑0.8780.8680.8750.8680.8750.8820.8720.8530.8790.895
VIF ↑0.5710.6430.6450.4460.2860.4560.4950.4720.5770.646
UIQI ↑0.7710.8460.5820.4260.7120.7520.7130.7360.8630.868
INOEN ↑7.0967.1527.2637.0356.8056.9857.2037.0667.0387.429
AG ↑3.6963.0073.6372.5852.3772.3703.1553.7053.8163.882
SF ↑10.1517.4299.7346.4815.9705.9907.8698.85510.75811.931
FMI ↑0.8960.8980.8960.8650.8580.8920.8930.8960.8960.899
VIF ↑0.4650.5580.6550.4640.2680.4570.4610.5020.6700.738
UIQI ↑0.7660.8500.6230.5200.8410.7760.6540.7050.7190.843
RoadSceneEN ↑7.1927.4727.1117.0487.2097.4317.2937.2937.2837.499
AG ↑3.5295.1025.0023.5503.4093.8184.8015.1735.2325.263
SF ↑9.90412.88313.6468.6088.8339.12212.56014.0358.08914.20614.067
FMI ↑0.8580.8480.8390.8470.8500.8610.8540.8620.8660.868
VIF ↑0.2290.4420.4630.3370.2740.4270.4200.4530.8440.4590.465
UIQI ↑0.8440.8070.6990.4340.7200.7310.7690.8320.8630.868
LLVIPEN ↑7.1266.9245.2636.8616.3656.7087.3387.1087.5027.695
AG ↑3.7482.9032.3652.0412.0442.2763.0823.3524.3154.676
SF ↑13.7379.25910.8475.7287.3087.43210.34815.32515.02315.577
FMI ↑0.9040.9020.9030.8990.9010.9010.9030.9000.9060.914
VIF ↑0.4050.4310.2830.3620.2060.3640.4701.0010.9731.053
UIQI ↑0.7850.2750.4360.7690.8240.5330.6940.7030.8230.848
M 3 F D EN ↑6.4286.7286.8436.4286.6396.7807.0857.5367.7127.803
AG ↑4.2393.6404.2402.4112.7692.7193.6714.2375.0205.483
SF ↑12.82210.07212.8816.4588.3557.65410.76812.30814.76415.36216.226
FMI ↑0.8870.8810.8710.8660.8780.8740.8780.8830.8880.891
VIF ↑0.3490.4450.6030.3020.2110.3810.3780.8570.8440.8750.892
UIQI ↑0.7210.6090.4320.8060.8090.6260.8140.8230.8830.886
Table 2. Ablation studies of our HCSPNet.
Table 2. Ablation studies of our HCSPNet.
MethodsRoadScene
EN ↑AG ↑SF ↑FMI ↑VIF ↑UIQI ↑
w/o A S I D 4.5363.3618.0240.7790.3320.833
w/ D s 5.1484.90412.6040.8430.3680.850
w/ D d 5.1373.88310.5000.8470.2580.842
Ours7.4995.26314.0670.8680.4650.868
Table 3. Generalization of our dual-discriminators framework, where “+” denotes incorporating our framework into existing GAN-based methods.
Table 3. Generalization of our dual-discriminators framework, where “+” denotes incorporating our framework into existing GAN-based methods.
MethodsRoadScene
EN ↑AG ↑SF ↑FMI ↑VIF ↑UIQI ↑
GANMcC7.0483.5508.6080.8470.3370.434
GANMcC+7.1353.6768.6230.8500.3410.439
GAM-FM7.2093.4098.8330.8500.2740.720
GAM-FM+7.4043.88610.0170.8600.3600.735
TarDAL7.4313.8189.1220.8610.4270.731
TarDAL+7.6164.25311.7050.8700.4670.778
GAN-HA7.2935.17314.0350.8620.4530.832
GAN-HA+7.3065.18614.3360.8750.4630.837
Table 4. Sensitivity analysis of α , β , and γ .
Table 4. Sensitivity analysis of α , β , and γ .
Metrics α β γ
0.1110 (Ours)1000.11 (Ours)101000.11 (Ours)10100
EN ↑7.0697.1677.4997.2737.1967.4997.2036.9497.1857.4997.2076.975
AG ↑4.4155.0375.2634.3865.0765.2635.2175.0274.7785.2635.0154.615
SF ↑11.32613.30814.06712.21913.31814.06714.00312.89313.01414.06713.71613.805
FMI ↑0.8610.8600.8680.8550.8650.8680.8660.8660.8630.8680.8620.860
VIF ↑0.3660.4270.4650.4370.4530.4650.4500.4380.4470.4650.4520.426
UIQI ↑0.8490.8610.8680.8580.8560.8680.8530.8620.8570.8680.8600.852
Table 5. Object detection on LLVIP and M 3 F D datasets with yolov5, where the best result is in red. We only make comparisons with cutting-edge methods.
Table 5. Object detection on LLVIP and M 3 F D datasets with yolov5, where the best result is in red. We only make comparisons with cutting-edge methods.
DatasetsLLVIP M 3 FD
MetricsPrecision ↑Recall ↑mAP@.5 ↑mAP@.95 ↑Precision ↑Recall ↑mAP@.5 ↑mAP@.95 ↑
Infrared0.9290.8490.9050.4720.7930.5470.6030.359
Visible0.9310.8800.9330.5210.7960.5850.6450.385
GAN-FM0.9280.8470.9150.4670.7650.5280.5790.362
TarDAL0.9390.8820.9230.4890.7660.5700.6180.355
DeRUN0.9360.8850.9370.5200.7360.5850.6310.368
GAN-HA0.9540.8800.9350.5330.8130.5920.6630.373
HCSPNet0.9580.8850.9470.5660.8250.6030.6680.406
Table 6. Object tracking with different tracking methods, where the best result is shown in red. We only make comparisons with cutting-edge methods.
Table 6. Object tracking with different tracking methods, where the best result is shown in red. We only make comparisons with cutting-edge methods.
TrackerLADCFGFSDCF
MetricsAccuracy ↑Failure ↓EAO ↑Accuracy ↑Failure ↓EAO ↑
Infrared0.47273.480.2090.58545.710.239
Visible0.40768.230.1600.43860.050.183
GAN-FM0.51941.220.1950.58638.350.262
TarDAL0.46948.570.1780.49346.220.203
DeRUN0.41736.820.1730.45760.370.178
GAN-HA0.57238.470.2260.61831.450.228
HCSPNet0.59637.320.2310.63029.430.267
Table 7. Quantitative results of the medical image fusion task with PET-MRI data, where the best result is shown in red.
Table 7. Quantitative results of the medical image fusion task with PET-MRI data, where the best result is shown in red.
MetricsU2FusionFusionGANDDcGANGANMcCGAN-FMTarDALDeRUNGAN-HAHCSPNet
EN ↑4.0854.2234.2753.4434.2713.1093.9534.8134.935
AG ↑5.8785.9165.9354.9885.3672.9723.6555.5376.083
VIF ↑0.5060.4570.4730.2980.3830.3060.5170.5530.675
UIQI ↑0.9330.8570.7730.5180.5560.4830.6450.6820.933
Table 8. Quantitative results of the biological image fusion with the GFP-PC task, where the best result is shown in red.
Table 8. Quantitative results of the biological image fusion with the GFP-PC task, where the best result is shown in red.
MetricsU2FusionFusionGANDDcGANGANMcCGAN-FMTarDALDeRUNGAN-HAHCSPNet
EN ↑6.5726.5816.6724.8646.6824.8575.6876.6786.935
AG ↑3.7693.8534.7383.6524.6575.3843.7685.0835.169
VIF ↑0.3640.5130.6250.4380.7020.6560.3470.5670.772
UIQI ↑0.7690.8530.8620.4180.8360.3980.7160.8830.903
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shen, M.; Huang, G.; Ju, M.; Ma, K.-K. Rethinking Infrared and Visible Image Fusion from a Heterogeneous Content Synergistic Perception Perspective. Sensors 2025, 25, 4658. https://doi.org/10.3390/s25154658

AMA Style

Shen M, Huang G, Ju M, Ma K-K. Rethinking Infrared and Visible Image Fusion from a Heterogeneous Content Synergistic Perception Perspective. Sensors. 2025; 25(15):4658. https://doi.org/10.3390/s25154658

Chicago/Turabian Style

Shen, Minxian, Gongrui Huang, Mingye Ju, and Kai-Kuang Ma. 2025. "Rethinking Infrared and Visible Image Fusion from a Heterogeneous Content Synergistic Perception Perspective" Sensors 25, no. 15: 4658. https://doi.org/10.3390/s25154658

APA Style

Shen, M., Huang, G., Ju, M., & Ma, K.-K. (2025). Rethinking Infrared and Visible Image Fusion from a Heterogeneous Content Synergistic Perception Perspective. Sensors, 25(15), 4658. https://doi.org/10.3390/s25154658

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop