Next Article in Journal
Intelligent Vehicle Target Detection Algorithm Based on Multiscale Features
Previous Article in Journal
L1-Constrained Fractional-Order Gradient Descent for Axial Dimension Estimation of Conical Targets
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Language-Driven Cross-Attention for Visible–Infrared Image Fusion Using CLIP

Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(16), 5083; https://doi.org/10.3390/s25165083
Submission received: 12 June 2025 / Revised: 6 August 2025 / Accepted: 14 August 2025 / Published: 15 August 2025
(This article belongs to the Section Sensors and Robotics)

Abstract

Language-guided multimodal fusion, which integrates information from both visible and infrared images, has shown strong performance in image fusion tasks. In low-light or complex environments, a single modality often fails to fully capture scene features, whereas fused images enable robots to obtain multidimensional scene understanding for navigation, localization, and environmental perception. This capability is particularly important in applications such as autonomous driving, intelligent surveillance, and search-and-rescue operations, where accurate recognition and efficient decision-making are critical. To enhance the effectiveness of multimodal fusion, we propose a text-guided infrared and visible image fusion network. The framework consists of two key components: an image fusion branch, which employs a cross-domain attention mechanism to merge multimodal features, and a text-guided module, which leverages the CLIP model to extract semantic cues from image descriptions containing visible content. These semantic parameters are then used to guide the feature modulation process during fusion. By integrating visual and linguistic information, our framework is capable of generating high-quality color-fused images that not only enhance visual detail but also enrich semantic understanding. On benchmark datasets, our method achieves strong quantitative performance: SF = 2.1381, Qab/f = 0.6329, MI = 14.2305, SD = 0.8527, VIF = 45.1842 on LLVIP, and SF = 1.3149, Qab/f = 0.5863, MI = 13.9676, SD = 94.7203, VIF = 0.7746 on TNO. These results highlight the robustness and scalability of our model, making it a promising solution for real-world multimodal perception applications.

1. Introduction

The core of multimodal fusion lies in acquiring complementary information from multiple data sources, thereby compensating for the incomplete data provided by a single modality and achieving a more comprehensive and enriched scene representation [1]. Infrared and visible light image fusion is a highly representative task in this domain [2,3]. Infrared imaging relies on detecting thermal radiation, maintaining high robustness in poorly lit environments and highlighting thermal targets with strong contrast [4]; however, it often lacks sufficient texture details [5,6]. In contrast, visible light images provide abundant structural and textural information but can suffer significant detail loss under low-light conditions or occlusion [7]. By leveraging the complementarity of these two modalities, one can retain the salient thermal information while integrating the clear textural details from visible light images, thus providing a more accurate and comprehensive representation of the original scene. The fused images typically exhibit both prominent target contrast and rich textural detail, offering superior visual perception for human observers and supplying more reliable input for subsequent machine vision tasks [8,9,10].
Language-guided multimodal fusion, which integrates visible and infrared images with linguistic descriptions, has shown great potential in enhancing image fusion performance, especially under low-light and adverse weather conditions. Visible images often carry rich color and texture details but tend to suffer from severe information loss in poorly lit or harsh environments. In contrast, infrared images maintain robustness in such conditions but lack sufficient texture and geometric information. To bridge the gap between these modalities, we incorporate semantic cues extracted from textual descriptions. These cues help align and constrain multimodal features more accurately, while also compensating for missing information in either modality. As a result, the fused 2D images exhibit enhanced texture richness and clearer structural representation, leading to improved robustness and accuracy in downstream tasks.
In computer vision applications, tasks such as detection, navigation, localization, and environmental perception are being applied to increasingly diverse scenarios. For example, prior work has explored object detection from Bayer images [11] to extract low-level visual information directly from raw sensor data, as well as underground object detection using ground-penetrating radar (GPR) [12] to enable 3D mapping and target recognition in complex subterranean environments. These studies reflect the growing demand for detection technologies under extreme visual conditions and across unconventional modalities. Fused images can provide robots with a more comprehensive understanding of the scene, particularly in low-light or obstacle-dense environments. For example, in autonomous driving, the integration of visible and infrared features with semantic guidance enables easier detection of roads and obstacles even in extreme environments such as nighttime or tunnels, thereby enhancing driving safety. Similarly, in rescue missions, semantically enriched fusion results allow robots to maintain accurate perception and navigation in smoke-filled or low-visibility settings, enabling efficient search and rescue operations. Overall, language-guided fusion enhances a system’s adaptability and stability in complex, real-world environments.
To address these challenges, we propose a visible-infrared fusion framework based on vision-language models. This framework employs a cross-domain attention mechanism to extract complementary features across modalities and integrates a language-guided module that embeds semantic information into the fusion process. The resulting fused images not only preserve salient thermal targets from infrared input but also retain rich texture and structural details from visible images, offering more reliable inputs for subsequent computer vision tasks.
Due to the lack of explicit supervision from real fused images, we introduce a language-guided fusion process into our framework. By generating textual descriptions for visible-light images, we are able to capture more comprehensive scene details and structural information, while also injecting these high-level semantic cues into the subsequent fusion stages. In the language-driven fusion module, we adopt a network called visible captioning, which takes visible-light images as input and outputs corresponding textual descriptions of the scene. Under the guidance of these descriptions, the fusion process dynamically adjusts the weights and priorities of features according to the scene context, thereby producing fused images that are more semantically aligned with real-world environments. Overall, our contributions are summarized as follows: 1. We propose a language-driven, vision-language-based infrared and visible image fusion framework, significantly enhancing fusion performance and interpretability. 2. We introduce a cross-domain attention mechanism to effectively integrate multimodal features from visible and infrared images, leveraging the complementary advantages of both modalities. 3. We design a semantic interaction guidance module, which extracts textual semantic information from content descriptions of visible-light images to guide the generation of high-quality fused images from multimodal features. As illustrated in Figure 1, our framework outlines the entire processing flow and the interplay among its constituent modules. 4. We achieved the best overall performance across multiple metrics on both the TNO and LLVIP datasets.

2. Related Work

2.1. Conventional Image Fusion Methods:

Sparse representation theory models image signals as a linear combination of as few “atoms” or transformation primitives as possible, derived from an overcomplete dictionary [13]. This approach is especially powerful in image fusion because it captures data-driven representations by learning a complete dictionary that can adapt to diverse image features, allowing for efficient feature extraction and representation [14]. In practice, sparse representation methods excel at emphasizing intrinsic structures (e.g., salient edges, key textures) while suppressing irrelevant details. Advances in dictionary learning—such as structured sparsity, adaptive dictionary construction, and online dictionary learning—have further strengthened their ability to represent complex image patterns, widening the scope of possible applications in fusion tasks. Such methods have proven highly effective in scenarios that demand high-precision feature selection, including medical imaging diagnostics and remote sensing fusion, where minimal loss of critical detail is paramount. Multi-scale transformation techniques break down the original image into sub-bands or sub-images at multiple scales, mirroring the hierarchical processing observed in the human visual system. By decomposing images into various levels of detail, these methods help preserve fine-grained information and important structures during the fusion process [2,15]. Representative approaches include nonlinear methods, pixel-level weighted averaging, estimation-based procedures, and color composite fusion [16,17,18,19,20]. Through effectively merging complementary information from the source images, multi-scale methods enhance the clarity, sharpness, and informativeness of the final fused outputs. Recent innovations, such as wavelet and curvelet transforms, further boost the ability to capture multi-resolution features. This multi-level understanding of the scene makes multi-scale transformation techniques especially suitable for application fields like medical diagnostics, night-vision enhancement, and surveillance, where preserving both large-scale structures and subtle details is critical. Subspace representation approaches project high-dimensional image features onto a low-dimensional subspace, thereby efficiently capturing essential structures in the input images while reducing computational overhead and memory consumption [21]. Widely adopted techniques include Principal Component Analysis (PCA) [22,23], Independent Component Analysis (ICA) [24,25], and Non-Negative Matrix Factorization (NMF) [26,27]. These methods typically focus on extracting the most salient features from the source images and discarding redundant or noise-related information, thus facilitating an informative and efficient fusion process. Subspace-based approaches have demonstrated considerable success in diverse applications, such as medical imaging (where dimensionality reduction can reveal latent anatomical structures), remote sensing (where large-scale datasets necessitate efficient processing), and multi-modal data integration (where merging data from various sensors requires robust feature-level transformations). Saliency detection seeks to emulate human visual behavior by identifying the most conspicuous regions or objects in an image. In computer vision and pattern recognition, such models play a pivotal role in tasks like object detection, image segmentation, and scene understanding [25,27,28]. When applied to image fusion, saliency-based methods generally involve computing spatial weights and extracting salient objects so that essential visual information is retained in the fused image [29,30,31,32,33]. Recent strides in deep learning-based saliency detection—leveraging convolutional neural networks (CNNs) or Transformer-based architectures—have led to heightened accuracy and robustness, making these techniques more adaptable to complex and dynamic scenarios. Notably, in infrared-visible fusion, saliency models help maintain the most critical infrared targets while preserving the contextual detail from visible imagery, resulting in output images that effectively integrate both modalities. Collectively, these classical approaches—sparse representation, multi-scale transformation, subspace representation, and saliency detection—form an essential foundation for image fusion. They complement data-driven deep learning methods by offering strong theoretical grounding and well-defined transform domains. By leveraging the strengths of each technique and integrating them with emerging deep architectures, researchers continue to push the boundaries of multimodal fusion, leading to more accurate, robust, and application-specific solutions.

2.2. Deep Learning-Based Image Fusion Methods

Convolutional neural networks (CNNs) have been widely explored for image fusion tasks, yielding a series of high-performing methods. One representative approach is PMGI [34], which employs a lightweight architecture to preserve gradient and intensity information effectively, enabling quick and robust fusion in practical applications. To further enhance texture detail retention, SDNet [35] introduces an optimization objective grounded in texture richness. This method incorporates an adaptive decision module to balance gradient preservation and texture representation, resulting in more comprehensive fusion outputs even in complex scenarios. Additionally, STDFusionNet [3] addresses the challenge of identifying critical regions by incorporating a salient target mask as spatial guidance, thus reinforcing the fusion of key objects from multiple modalities. Refs. [36,37] extract textual descriptions from depth priors to assist depth estimation. Ref. [38] extracts visible and non-visible features using a diffusion model and guides image fusion based on the textual descriptions of depth. A number of studies have also examined the impact of illumination on fusion quality. For instance, refs. [39,40] systematically analyze how varying lighting conditions influence the performance of image fusion algorithms, thereby offering insights into potential optimizations. Their findings underscore the importance of designing fusion strategies that can adapt to diverse lighting situations for more consistent and reliable results. Ref. [41] investigates image fusion for images with different focal lengths. In recent years, the combined use of CNN and Transformer architectures for multimodal image fusion has demonstrated considerable promise. By leveraging the self-attention mechanism inherent in Transformers, methods such as SwinFusion [42] and SwinFuse [43] capitalize on long-range dependencies to exploit the complementary nature of features across different modalities. These hybrid approaches not only improve fusion quality in various multimodal tasks but also exhibit strong generalization capabilities, extending the applicability of image fusion techniques to a broader range of real-world scenarios.

2.3. Language–Vision Models

As deep network architectures continue to evolve at a rapid pace and large-scale datasets become increasingly available, language–vision models are gaining prominence in the field of generative modeling. Among these, CLIP [44] stands as a foundational work. It employs two neural network encoders to process text and images separately, achieving semantic alignment through contrastive learning. Leveraging unsupervised training and extensive data, CLIP exhibits robust zero-shot recognition capabilities and highly efficient feature extraction for both text and images, thereby laying a solid groundwork for subsequent text-driven image generation and processing methods. Building upon CLIP, Style-CLIP [45] has become a highly influential application. It integrates StyleGAN’s [46] powerful image generation capabilities with textual prompts, enabling users to edit images interactively via natural language—a flexible and user-friendly approach to personalized image editing. Beyond GAN-based models, text-conditioned diffusion models have also garnered attention for their superior generation quality and flexibility. For instance, DiffusionCLIP [47] leverages diffusion models in conjunction with CLIP’s semantic understanding to perform text-driven image processing, while Stable Diffusion [48] further incorporates text encoders and attention mechanisms to achieve outstanding text-guided image generation. These methods commonly support interactive multimodal fusion and fine-grained control, greatly expanding the possibilities for image creation and editing. Addressing the limitations of existing methods in handling degraded or complex scenes, TextIF [49] proposes a text-guided image fusion framework that allows users to generate customized fused images via interactive text prompts. Meanwhile, ref. [50] employs CLIP to map text into a multimodal embedding space, representing fusion targets and image features through relationships among embedding vectors and introducing a language-driven loss function to optimize fusion performance. Despite notable progress in related technologies, most existing methods rely on relatively simple textual information, making it difficult to capture the full complexity of multimodal fusion tasks. To address this issue, we introduce an enhanced strategy: extracting more comprehensive and diverse fusion features from both visible and infrared images, and generating object-level textual descriptions from visible images to enrich textual information at a finer granularity. Our experiments demonstrate that this approach not only significantly improves the accuracy and reliability of fused images but also better accommodates a wide range of complex application scenarios, thereby creating a broader space for continued progress in multimodal image synthesis.

3. Method

RGB images typically capture abundant texture, color, and shape information, providing a detailed representation of the scene for both human and computer vision systems. However, in extreme environments (e.g., nighttime, smoke, heavy fog) or in scenarios with severe occlusion, RGB images often suffer substantial loss of visible information. By contrast, non-visible light (e.g., infrared or thermal imaging) can effectively detect thermal radiation or other spectral signals under these conditions, thereby strongly complementing the limitations of traditional RGB images. By leveraging a cross-attention mechanism to fuse features from these two distinct modalities, the system can fully exploit their complementary advantages at the feature level, achieving more comprehensive and robust perception of the environment and targets.

3.1. Image Encoder

In this module, both the RGB image and the non-visible light image (e.g., infrared) are used as inputs to the encoder. To extract crucial information from both visible and non-visible modalities, we employ a Transformer/Restormer-based network structure [51] as the primary feature extractor. This process can be formalized as follows:
F r g b = F r I ( I i r ) , F n i v = F n i v I ( I n i v ) ,
where I i r R H × W × 3 is the three-channel RGB image, and I n i v R H × W × 1 denotes the single-channel non-visible light image. H and W represent the image height and width, respectively. F r I and F n i v I are the encoders for RGB and non-visible light data, respectively.
Within each encoder, the Transformer/Restormer architecture leverages self-attention, feed-forward networks, and multi-scale feature pyramids to perform feature extraction at multiple resolutions. This design not only captures global semantic context at larger scales but also preserves local texture details at finer scales, laying a solid foundation for subsequent cross-modal information interaction.

3.2. Cross-Modality Feature Fusion

Non-visible light provides extended coverage or captures target information in spectral bands beyond human vision, such as temperature distribution, shape contours, and localized heat sources. Meanwhile, RGB images offer color, texture, and high-resolution details. To fully exploit their complementary advantages, we adopt a cross-attention mechanism (CR-ATT) in the cross-modality fusion layer, aiming to strengthen the interaction and information flow between non-visible light and RGB features. The resulting fused features enable deeper collaboration across different modalities, yielding a more discriminative and robust multi-modal representation. First, we project the RGB features F r g b and non-visible light features F n i v into { Q , K , V } components:
{ Q r , K r , V r } = F q k v r ( F r g b )
{ Q d , K d , V d } = F q k v d ( F n i v )
where F q k v r and F q k v d are linear transformations that generate queries, keys, and values for RGB and non-visible light features, respectively. Next, to encourage further spatial interaction between the two modalities, we exchange their query vectors:
F f d = softmax Q r K d T d k V d
F f r = softmax Q d K r T d k V r
where d k is a scaling factor to maintain numerical stability. Here, F f d represents the features obtained by combining RGB queries with non-visible keys and values, while F f r is derived from combining non-visible queries with RGB keys and values. Through this cross-attention, each modality can “focus” on the most informative regions or channels in the other modality, thereby strengthening the representation of specific targets or regions of interest.
Finally, we concatenate F f d and F f r along the channel dimension:
F f 0 = Concat F f d , F f r ,
F f i = F tran F f 0 ,
F t r a n refers to a Transformer layer that performs further feature extraction on the directly concatenated multimodal features. Thus yielding the fused multi-modal feature representation. This representation integrates spatial and geometric information from non-visible light with the color and texture details from RGB images, proving beneficial for downstream tasks such as image fusion, object detection, or semantic segmentation.
By using cross-attention to fuse RGB and non-visible light features, the system can more effectively aggregate information from multiple modalities, demonstrating superior performance in complex and dynamically changing real-world environments. In low-light or smoke-filled settings, infrared or thermal images offset the limitations of RGB-based observation; under normal lighting conditions, RGB features provide more intuitive visual details. By leveraging these two types of information simultaneously, we not only gain higher accuracy in target recognition and localization but also substantially improve the system’s adaptability in applications such as autonomous driving, surveillance, medical image analysis, and nighttime navigation.

3.3. Language-Driven Image Fusion

In the image fusion process, deep learning techniques typically rely on multiple loss functions to constrain the model. However, the absence of real fused images for direct supervision makes it challenging to effectively regulate the fusion output via loss functions. Moreover, many problems cannot be explicitly modeled, thereby limiting the model’s overall performance. Motivated by the idea that text descriptions generated from images can provide valuable visual cues, we propose a language-guided fusion method. Specifically, we adopt ExpansionNet v2 from [52], which takes an RGB image as input and outputs a textual description of the image content. Given a pair of RGB images I i r and niv images I n i v , the RGB image is first processed by the image captioning network to generate the corresponding text description. These text descriptions are then concatenated and fed into the frozen text encoder of CLIP, resulting in text embedding features that encapsulate information about the RGB image. To utilize these text embedding features for guiding image fusion, we extract a set of semantic parameters from the text embeddings. These parameters contain high-level guidance information such as object attributes and spatial relationships. We then employ a Multilayer Perceptron (MLP) to learn these relationships and further map the textual semantic information ( σ ^ and v) for feature F f i scaling and bias control, respectively. The process is formulated as follows:
F ^ f i = ( 1 + σ ^ ) F f i + μ ^ ,
where ⊙ denotes the Hadamard product (element-wise multiplication).
In this formulation, the semantic parameters interact with the fused features through a feature modulation mechanism, specifically via scale and bias adjustments. This allows the model to flexibly refine multimodal representations from a semantic perspective, injecting richer contextual information into the fusion process, enhancing feature expressiveness, and achieving effective alignment of multimodal features within the semantic space. Moreover, we introduce residual connections within the language-driven feature modulation network to alleviate gradient vanishing issues during deep network training, reduce the difficulty of model fitting, and enhance the stability of the fusion process. Under semantic guidance, the model can adaptively assign appropriate weights and priorities to different features, making the fusion results more aligned with the semantic requirements of specific scenes. The final fused image exhibits clear semantic structures and strong hierarchical representation while preserving critical information from both target and background, thus laying a solid perceptual foundation for downstream tasks such as image understanding, detection, and recognition.

3.4. Loss Function

Loss Function for the Fusion Process: To ensure that the fused image retains both structural details and intensity characteristics from the source modalities, we introduce two complementary loss functions: a multi-channel gradient loss and a multi-channel intensity loss. To preserve fine texture and edge information, we employ the multi-channel gradient loss L M C G , which encourages gradient consistency between the fused image and the most salient gradients from either the visible or infrared input. It is defined as follows:
L M C G = 1 H W i = 1 3 I f i max | I i r i | , | I n i v i | 1
where ∇ denotes the gradient operator, and I f i represents the i-th channel (i.e., red, green, or blue) of the fused image I f . Similarly, I i r i denotes the i-th channel of the visible image I i r . To ensure that the fused image maintains overall brightness and structural coherence with the source inputs, we also apply the multi-channel intensity loss L M C I , formulated as follows:
L M C I = 1 H W i = 1 3 I f i max I i r i , I n i v i 1
The final loss function used to train the fusion network is a weighted combination of these two components:
L f = L M C G + L M C I
This joint loss encourages the fused image to retain both sharp edge information and meaningful intensity patterns from the source modalities, facilitating high-quality color fusion.

4. Experiment

4.1. Implementation Details and Datasets

Datasets: We conducted comprehensive experiments on three widely used public datasets for visible-infrared fusion: LLVIP [53], and TNO [54]. These datasets encompass a diverse range of scenes, object types, lighting conditions, and thermal distributions, providing a solid benchmark to evaluate the generalization ability and robustness of our proposed language-guided fusion method. Specifically, the LLVIP dataset contains paired RGB and infrared images collected in indoor and outdoor environments under different illumination settings (daytime, nighttime, and low-light), while the TNO dataset focuses on military and surveillance scenarios with strong thermal contrast. These diverse characteristics help validate the practical applicability of our method across multiple domains.
Metrics: Five statistical metrics are adopted for quantitative evaluation, namely Mutual Information (MI) [55], Visual Information Fidelity (VIF) [56], Spatial Frequency (SF) [57], Qabf [58], and Standard Deviation (SD). Collectively, they offer a comprehensive assessment of the fused image from multiple perspectives:
  • MI (Mutual Information): Evaluates the degree to which information from both source images is preserved and integrated in the fused result, indicating how well the fusion process combines complementary details.
  • VIF (Visual Information Fidelity): Measures the fidelity of the fused image relative to the source images, focusing on how accurately essential visual content is conveyed.
  • SF (Spatial Frequency): Examines the spatial frequency components within the fused data, reflecting the level of detail and sharpness retained.
  • Qabf: Quantifies the edge information contributed by each source image, offering insight into how effectively structural details are preserved in the final result.
  • SD (Standard Deviation): Assesses the overall contrast of the fused image, highlighting its dynamic range and distinguishing ability between different intensity levels.
These metrics jointly capture various dimensions of image quality—ranging from information integration and detail clarity to structural fidelity and contrast—thus providing a well-rounded evaluation of fusion performance.

4.2. Analysis of Fusion Results with Textual Guidance

As illustrated in Figure 2, subfigures (a) and (b) show the original infrared and visible images, respectively. Subfigures (c) and (d) present the fused results without and with textual guidance. A comparative analysis reveals several notable improvements brought by the textual guidance module:
Enhanced detail clarity: In subfigure (d), the license plate region (highlighted by the red box) exhibits significantly sharper edges and higher character recognizability (e.g., “B92”) than in subfigure (c). This suggests that textual guidance enhances the model’s semantic attention to critical targets such as license plates, leading to more precise feature preservation in the fusion output.
Improved texture retention: The architectural textures enclosed in the green box are better preserved in the text-guided result. Compared to (c), subfigure (d) demonstrates clearer structural boundaries and reduced blurring, indicating that textual semantics help maintain structural integrity and detail fidelity.
Salient semantic target enhancement: The traffic sign in the blue circle is more distinct and easier to recognize in (d) than in (c), with sharper edges and improved contrast. This highlights the effectiveness of language-guided fusion in reinforcing visually and semantically important elements.
These improvements stem from the textual guidance mechanism, which extracts high-level semantic cues (e.g., “license plate”, “building”, “traffic sign”) from image descriptions and encodes them into attention-aware parameters. These semantic priors are then used to guide the fusion process, enabling the network to focus on meaningful regions while suppressing redundant or ambiguous background information. Consequently, the final fused image achieves better semantic consistency, structural completeness, and perceptual quality.

4.3. Comparison of Visible and Infrared Image Fusion

In Figure 3, Figure 4 and Figure 5, we present three sets of fusion results of visible and infrared images for the scenarios “pedestrian crossing,” “nighttime vehicles and pedestrians,” and “multiple pedestrians,” respectively. We use red boxes to indicate the areas in the fused images where differences most likely occur among the various methods. Compared to the other models, our method demonstrates notable advantages in the following aspects:

4.3.1. Saliency and Detail Preservation of Pedestrians

Infrared images generally highlight pedestrians’ thermal radiation but appear somewhat blurry in terms of texture and contours. Visible images, on the other hand, contain abundant detail yet may lose key information under insufficient illumination. From the three sets of figures, it is evident that in processing pedestrian areas, other models either diminish the high-contrast features in the infrared image, causing pedestrians to be less prominent, or fail to adequately preserve the texture from the visible image, resulting in blurring or distortion in pedestrian clothing or outlines. In contrast, our method effectively combines the intense highlights from infrared with the fine textures from visible images, ensuring that pedestrians in the final fused image remain both clear and highly recognizable.

4.3.2. Clarity of Background and Noise Suppression

In the right-side red box of Figure 3, as well as in the background regions of Figure 4 and Figure 5, there is often substantial noise or variations in lighting. Many of the compared models still retain noticeable noise after fusion or cannot simultaneously present the structured information from the visible image and the overall brightness from the infrared image. In our method’s cross-modal feature interaction, local and global information are dynamically balanced, allowing the background to maintain appropriate brightness and texture details while effectively reducing noise, thereby avoiding the issue of the image being “too dark” or “too bright.”

4.3.3. Distant Targets and Scene Structure

In the top red box of Figure 5 (the “pedestrians”) and the distant background in Figure 4 (vehicles), some far-off targets or scene structures are visible. In the fusion results of other models, such distant objects often appear blurred or lost. Through deeper interactions of multi-modal features, our method performs more precise texture extraction and contrast enhancement for distant areas, thus more completely retaining the shapes and structural information of these targets.

4.3.4. Overall Visual Consistency and Layered Appearance

Infrared images may show overexposure in regions with high temperatures or reflective surfaces, while visible images could lose details at night or in shadowed areas. By adaptively balancing exposure and shadowing between these two modalities, our method achieves a better equilibrium between detail preservation and overall visual quality, resulting in a stronger sense of depth and layering. Whether it is pedestrians, vehicles, or background buildings, the final fused image achieves a relatively natural visual effect.

4.3.5. Overall Performance Assessment

Taken together, the results from these three sets of experiments indicate that our method not only preserves the salient targets from infrared images and the detailed information from visible images but also effectively controls noise and prevents exposure imbalance. The visual and quantitative evaluations both demonstrate strong stability and superior performance. As shown in Table 1, our algorithm achieves the highest score in 4 of the 5 evaluation metrics and ranks third in the remaining one; Table 2 similarly confirms that we maintain leading performance on a large-scale, high-resolution dataset. Through both visual inspection and numerical assessment, it can be concluded that our method exhibits significant overall advantages in visible-infrared image fusion.
Figure 6 presents visible light images, infrared images, and the fusion results under nighttime scenes involving multiple pedestrians, vehicles, and complex backgrounds. Based on the comparison, the advantages of our fusion method are evident: Visible images contain rich detail information such as road textures, traffic signs, and clothing of individuals. However, under low-light conditions, they suffer from poor illumination, making target recognition more difficult. Infrared images clearly highlight thermal features of pedestrians and vehicles, offering strong target detection capability. However, they lack structural and texture details, resulting in weak overall scene representation. Fusion results effectively combine the advantages of both modalities. They retain detailed information from the visible images—subtext descriptions containing RGB information serve as interpretable prior guidance for image fusion results such as road structures, object contours, and background textures—while also enhancing the prominent thermal targets captured by the infrared images, making key objects like pedestrians and vehicles more distinct and clear. Specifically, in the fusion images, pedestrian clothing and facial contours are not only thermally highlighted but also show enhanced texture details. The appearance of vehicles and surrounding objects is more natural and realistic. Notably, in the third and fourth columns, the fusion results significantly improve the visibility of multiple targets in nighttime scenes, demonstrating strong environmental adaptability and enhanced visual clarity. This indicates that our proposed method performs robustly and is practically valuable in low-light and complex environments.

4.4. Impact of Caption Variability on Fusion Performance

To evaluate the robustness of our method to different forms of textual input, we employed multiple image captioning models to generate diverse textual descriptions from the same visible-light images. These descriptions vary in wording and phrasing but all broadly reflect the semantic content of the input image. We then conducted fusion and evaluated the results using standard quantitative metrics. As shown in Table 3, the performance across different textual inputs remains highly consistent, indicating that as long as the generated descriptions are semantically aligned with the image content, their impact on fusion quality is minimal. This demonstrates the flexibility and generalizability of our language-guided fusion framework.

4.5. Dataset-Specific Differences in Structural Metrics

We observe a notable disparity in structural and perceptual metrics—namely Spatial Frequency (SF), Standard Deviation (SD), and Visual Information Fidelity (VIF)—between the LLVIP and TNO datasets. As shown in Table 1 and Table 2, all compared methods achieve consistently higher SF, SD, and VIF scores on LLVIP, whereas the corresponding values on TNO are significantly lower. This phenomenon can be attributed to several key differences between the two datasets:
(1) Resolution and texture density of input images. LLVIP is a high-resolution dataset primarily composed of near-field, human-centered scenes captured under various illumination conditions. It contains rich texture details, well-defined object boundaries, and fine structural edges—especially in the RGB modality. As SF and SD are sensitive to spatial variations and intensity contrast, the richer visual content in LLVIP naturally leads to higher metric values. In contrast, the TNO dataset consists mainly of medium-to-long-range surveillance imagery, where both infrared and visible images have relatively lower resolution and fewer fine details. The infrared images in particular exhibit smoother transitions and blurred contours, resulting in lower structural fidelity in the fused output and subsequently lower SF and SD values.
(2) Thermal contrast and dynamic range differences. TNO infrared images typically exhibit strong thermal contrast, with hot targets (e.g., humans or vehicles) appearing bright against dark backgrounds. This limited dynamic range may hinder the model’s ability to preserve fine visual information, thereby reducing the VIF score. In LLVIP, the infrared and visible modalities are better aligned in terms of brightness distribution, allowing for more balanced fusion and clearer preservation of both modalities’ information, which contributes to consistently higher VIF scores.
(3) Scene conditions and target applications. LLVIP is designed for pedestrian perception tasks in nighttime or low-light environments, where detailed semantic and structural information is critical. As such, the fusion results emphasize edge clarity and texture preservation, leading to elevated values in SF, SD, and VIF. TNO, on the other hand, is oriented toward long-range surveillance and defense-related scenarios, where the focus is on enhancing thermal target saliency rather than preserving high-frequency texture or visual fidelity. This results in a shift of emphasis in the fusion process, explaining the lower structural metric values observed.

4.6. Ablation Study

Feature Fusion Module: This module achieves efficient multimodal feature fusion by integrating Transformer-based feature extraction with a cross-attention mechanism. Specifically, the Transformer feature extraction component captures long-range dependencies and global contextual information within multimodal data, thereby extracting deeper and more discriminative features from RGB and infrared images. The cross-attention mechanism then dynamically computes correlation weights between the RGB and infrared features, effectively highlighting complementary key information and suppressing redundant or noisy signals. This fusion strategy fully exploits the complementary strengths of the two modalities in both structure and detail, resulting in fused images that are clearer and richer in visual content. To further validate the effectiveness of this module, we removed it while keeping other components intact and compared the results with the evaluation metrics listed in Rows 1 and 2 of Table 4. After removing the feature fusion module, all metrics declined to varying degrees, indicating that it plays a critical role in extracting multimodal features and enhancing the fused image quality.
Language-Driven Fusion Module: In this module, text descriptions containing image information serve as interpretable prior guidance for image fusion. We design an MLP to predict fusion parameters that dynamically adjust the weights and priorities of the RGB and infrared features during the fusion process. To test its effectiveness, we fixed the MLP-predicted parameters to 1 and 0, thereby removing the influence of language on the image fusion. Comparing Rows 2 and 3 of Table 4, we observe a significant performance drop without language guidance, indicating that the RGB priors embedded in the text descriptions effectively improve fusion quality. By integrating these text-based constraints, the fused images achieve higher fidelity and clarity, while also offering enhanced interpretability for multimodal fusion.

4.7. Limitations

Although our qualitative analysis demonstrates advantages in terms of target clarity, background reconstruction, and visual consistency, it still has certain limitations. First, qualitative evaluation is inherently subjective, and the results may be influenced by individual viewer preferences or perceptual differences, making it difficult to ensure complete objectivity. Second, while the selected image examples are representative, they may not fully cover all possible scene variations, such as extreme weather conditions or complex background occlusions. Third, qualitative analysis alone cannot comprehensively reflect the effectiveness of the fused images in downstream tasks such as object detection or depth estimation, which limits the depth of assessment regarding the method’s practical applicability. In future work, we plan to incorporate more task-driven evaluation methods, combining both subjective and objective metrics to further validate the robustness and generalizability of our model in real-world applications.

5. Conclusions

In this paper, we propose a language-guided framework for fusing infrared and visible light images, which integrates both visual features and semantic cues to enhance fusion quality. The overall architecture consists of two main components: a cross-modal fusion branch, which leverages cross-domain attention to effectively combine visible and infrared modalities, and a language-driven module, which introduces high-level semantic guidance into the fusion process. To obtain the semantic descriptions, we employ an off-the-shelf image captioning model to extract textual descriptions from visible light images. These descriptions are then encoded using the CLIP text encoder to generate semantic representations, which guide the feature modulation and enhance the fusion outcome. By aligning visual content with language-derived semantics, our method enables the generation of high-quality, color-enhanced fusion results. This approach not only improves interpretability but also demonstrates strong potential for practical applications in low-light surveillance, medical diagnostics, and remote sensing scenarios.

Author Contributions

Methodology, X.W.; Formal analysis, X.W.; Resources, J.W.; Writing—original draft, X.W.; Writing—review & editing, X.W.; Supervision, P.Z.; Project administration, Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

All the authors are from Aerospace Information Research Institute, Chinese Academy of Sciences. All authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Wang, Y.; Huang, W.; Sun, F.; Xu, T.; Rong, Y.; Huang, J. Deep multimodal fusion by channel exchanging. Adv. Neural Inf. Process. Syst. 2020, 33, 4835–4845. [Google Scholar]
  2. Ma, J.; Ma, Y.; Li, C. Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 2019, 45, 153–178. [Google Scholar] [CrossRef]
  3. Ma, J.; Tang, L.; Xu, M.; Zhang, H.; Xiao, G. STDFusionNet: An infrared and visible image fusion network based on salient target detection. IEEE Trans. Instrum. Meas. 2021, 70, 5009513. [Google Scholar] [CrossRef]
  4. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar] [CrossRef]
  5. Llordés, A.; Garcia, G.; Gazquez, J.; Milliron, D.J. Tunable near-infrared and visible-light transmittance in nanocrystal-in-glass composites. Nature 2013, 500, 323–326. [Google Scholar] [CrossRef]
  6. Kim, J.U.; Park, S.; Ro, Y.M. Uncertainty-guided cross-modal learning for robust multispectral pedestrian detection. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1510–1523. [Google Scholar] [CrossRef]
  7. Li, C.; Song, D.; Tong, R.; Tang, M. Illumination-aware faster R-CNN for robust multispectral pedestrian detection. Pattern Recognit. 2019, 85, 161–171. [Google Scholar] [CrossRef]
  8. Ma, J.; Zhang, H.; Shao, Z.; Liang, P.; Xu, H. GANMcC: A generative adversarial network with multiclassification constraints for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 2020, 70, 5005014. [Google Scholar] [CrossRef]
  9. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  10. Zhu, Z.; Yang, X.; Lu, R.; Shen, T.; Xie, X.; Zhang, T. CLF-Net: Contrastive learning for infrared and visible image fusion network. IEEE Trans. Instrum. Meas. 2022, 71, 5021015. [Google Scholar] [CrossRef]
  11. Lin, J.; Zhang, J.; Lu, G. Keypoint Detection and Description for Raw Bayer Images. arXiv 2025, arXiv:2503.08673. [Google Scholar]
  12. Zhang, J.; Lu, G. Underground mapping and localization based on ground-penetrating radar. In Proceedings of the Asian Conference on Computer Vision, Hanoi, Vietnam, 8–12 December 2024; pp. 2018–2033. [Google Scholar]
  13. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-D deep learning approach for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef]
  14. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  15. Zhu, P.; Ma, X.; Huang, Z. Fusion of infrared-visible images using improved multi-scale top-hat transform and suitable fusion rules. Infrared Phys. Technol. 2017, 81, 282–295. [Google Scholar] [CrossRef]
  16. Dogra, A.; Goyal, B.; Agrawal, S. Bone vessel image fusion via generalized reisz wavelet transform using averaging fusion rule. J. Comput. Sci. 2017, 21, 371–378. [Google Scholar] [CrossRef]
  17. Dogra, A.; Goyal, B.; Agrawal, S.; Ahuja, C.K. Efficient fusion of osseous and vascular details in wavelet domain. Pattern Recognit. Lett. 2017, 94, 189–193. [Google Scholar] [CrossRef]
  18. Shen, R.; Cheng, I.; Basu, A. QoE-based multi-exposure fusion in hierarchical multivariate Gaussian CRF. IEEE Trans. Image Process. 2012, 22, 2469–2478. [Google Scholar] [CrossRef]
  19. Therrien, C.W.; Scrofani, J.W.; Kreb, W.K. An adaptive technique for the enhanced fusion of low-light visible with uncooled thermal infrared imagery. In Proceedings of the International Conference on Image Processing, Santa Barbara, CA, USA, 14–17 July 1997; Volume 1, pp. 405–408. [Google Scholar]
  20. Xue, Z.; Blum, R.S. Concealed weapon detection using color image fusion. In Proceedings of the 6th International Conference on Information Fusion, Cairns, Australia, 8–11 July 2003; Volume 1, pp. 622–627. [Google Scholar]
  21. Tang, H.; Liu, G.; Tang, L.; Bavirisetti, D.P.; Wang, J. MdedFusion: A multi-level detail enhancement decomposition method for infrared and visible image fusion. Infrared Phys. Technol. 2022, 127, 104435. [Google Scholar] [CrossRef]
  22. Li, H.; Liu, L.; Huang, W.; Yue, C. An improved fusion algorithm for infrared and visible images based on multi-scale transform. Infrared Phys. Technol. 2016, 74, 28–37. [Google Scholar] [CrossRef]
  23. Zhang, X.; Dai, X.; Zhang, X.; Jin, G. Joint principal component analysis and total variation for infrared and visible image fusion. Infrared Phys. Technol. 2023, 128, 104523. [Google Scholar] [CrossRef]
  24. Cvejic, N.; Bull, D.; Canagarajah, N. Region-based multimodal image fusion using ICA bases. IEEE Sens. J. 2007, 7, 743–751. [Google Scholar] [CrossRef]
  25. Mitianoudis, N.; Stathaki, T. Pixel-based and region-based image fusion schemes using ICA bases. Inf. Fusion 2007, 8, 131–142. [Google Scholar] [CrossRef]
  26. Kong, W.; Lei, Y.; Zhao, H. Adaptive fusion method of visible light and infrared images based on non-subsampled shearlet transform and fast non-negative matrix factorization. Infrared Phys. Technol. 2014, 67, 161–172. [Google Scholar] [CrossRef]
  27. Wang, J.; Peng, J.; Feng, X.; He, G.; Fan, J. Fusion method for infrared and visible images by using non-negative sparse representation. Infrared Phys. Technol. 2014, 67, 477–489. [Google Scholar] [CrossRef]
  28. Wang, S.; Yue, J.; Liu, J.; Tian, Q.; Wang, M. Large-scale few-shot learning via multi-modal knowledge discovery. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020. Proceedings Part X 16. [Google Scholar]
  29. Bavirisetti, D.P.; Dhuli, R. Two-scale image fusion of visible and infrared images using saliency detection. Infrared Phys. Technol. 2016, 76, 52–64. [Google Scholar] [CrossRef]
  30. Cui, G.; Feng, H.; Xu, Z.; Li, Q.; Chen, Y. Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Opt. Commun. 2015, 341, 199–209. [Google Scholar] [CrossRef]
  31. Han, J.; Pauwels, E.J.; Zeeuw, P.D. Fast saliency-aware multi-modality image fusion. Neurocomputing 2013, 111, 70–80. [Google Scholar] [CrossRef]
  32. Liu, C.H.; Qi, Y.; Ding, W.R. Infrared and visible image fusion method based on saliency detection in sparse domain. Infrared Phys. Technol. 2017, 83, 94–102. [Google Scholar] [CrossRef]
  33. Zhang, B.; Lu, X.; Pei, H.; Zhao, Y. A fusion algorithm for infrared and visible images based on saliency analysis and non-subsampled Shearlet transform. Infrared Phys. Technol. 2015, 73, 286–297. [Google Scholar] [CrossRef]
  34. Zhang, H.; Xu, H.; Xiao, Y.; Guo, X.; Ma, J. Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12797–12804. [Google Scholar]
  35. Zhang, H.; Ma, J. SDNet: A versatile squeeze-and-decomposition network for real-time image fusion. Int. J. Comput. Vis. 2021, 129, 2761–2785. [Google Scholar] [CrossRef]
  36. Zhang, J.; Lu, G. Vision-language embodiment for monocular depth estimation. In Proceedings of the Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 10–17 June 2025; pp. 29479–29489. [Google Scholar]
  37. Zhang, J.; Reddy, P.K.; Wong, X.; Aloimonos, Y.; Lu, G. Embodiment: Self-supervised depth estimation based on camera models. In Proceedings of the 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Abu Dhabi, United Arab Emirates, 14–18 October 2024; pp. 7809–7816. [Google Scholar]
  38. Zhang, J.; Li, Z.; Lu, G. Language-depth navigated thermal and visible image fusion. arXiv 2025, arXiv:2503.08676. [Google Scholar]
  39. Tang, L.; Yuan, J.; Zhang, H.; Jiang, X.; Ma, J. PIAFusion: A progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion 2022, 83, 79–92. [Google Scholar] [CrossRef]
  40. Tang, L.; Xiang, X.; Zhang, H.; Gong, M.; Ma, J. DIVFusion: Darkness-free infrared and visible image fusion. Inf. Fusion 2023, 91, 477–493. [Google Scholar] [CrossRef]
  41. Zhang, J.; Xu, N.; Zhang, H.; Lu, G. Depth Estimation Based on 3D Gaussian Splatting Siamese Defocus. arXiv 2024, arXiv:2409.12323. [Google Scholar]
  42. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  43. Wang, Z.; Chen, Y.; Shao, W.; Li, H.; Zhang, L. SwinFuse: A residual swin transformer fusion network for infrared and visible images. IEEE Trans. Instrum. Meas. 2022, 71, 5016412. [Google Scholar] [CrossRef]
  44. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Virtually, 18–14 July 2021; pp. 8748–8763. [Google Scholar]
  45. Patashnik, O.; Wu, Z.; Shechtman, E.; Cohen-Or, D.; Lischinski, D. Styleclip: Text-driven manipulation of stylegan imagery. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 2085–2094. [Google Scholar]
  46. Karras, T.; Laine, S.; Aittala, M.; Hellsten, J.; Lehtinen, J.; Aila, T. Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8110–8119. [Google Scholar]
  47. Kim, G.; Kwon, T.; Ye, J.C. Diffusionclip: Text-guided diffusion models for robust image manipulation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 2426–2435. [Google Scholar]
  48. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10684–10695. [Google Scholar]
  49. Yi, X.; Xu, H.; Zhang, H.; Tang, L.; Ma, J. Text-IF: Leveraging Semantic Text Guidance for Degradation-Aware and Interactive Image Fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 27026–27035. [Google Scholar]
  50. Wang, Y.; Miao, L.; Zhou, Z.; Zhang, L.; Qiao, Y. Infrared and visible Image Fusion with Language-driven Loss in CLIP Embedding Space. arxiv 2024, arXiv:2402.16267. [Google Scholar]
  51. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M. Restormer: Efficient transformer for high-resolution image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5728–5739. [Google Scholar]
  52. Hu, J.C.; Cavicchioli, R.; Capotondi, A. Exploiting Multiple Sequence Lengths in Fast End to End Training for Image Captioning. In Proceedings of the 2023 IEEE International Conference on Big Data (BigData), Sorrento, Italy, 15–18 December 2023; pp. 2173–2182. [Google Scholar]
  53. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A visible-infrared paired dataset for low-light vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  54. Toet, A. The TNO Multiband Image Data Collection. Data Brief 2017, 15, 249–251. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  55. Qu, G.; Zhang, D.; Yan, P. Information measure for performance of image fusion. Electron. Lett. 2002, 38, 1–2. [Google Scholar] [CrossRef]
  56. Han, Y.; Cai, Y.; Cao, Y.; Xu, X. A new image fusion performance metric based on visual information fidelity. Inf. Fusion 2013, 14, 127–135. [Google Scholar] [CrossRef]
  57. Eskicioglu, A.M.; Fisher, P.S. Image quality measures and their performance. IEEE Trans. Commun. 1995, 43, 2959–2965. [Google Scholar] [CrossRef]
  58. Xydeas, C.S.; Petrovic, V. Objective image fusion performance measure. Electron. Lett. 2000, 36, 308–309. [Google Scholar] [CrossRef]
  59. Li, H.; Wu, X.-J.; Kittler, J. RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf. Fusion 2021, 73, 72–86. [Google Scholar] [CrossRef]
  60. Tang, W.; He, F.; Liu, Y. YDTR: Infrared and visible image fusion via Y-shape dynamic transformer. IEEE Trans. Multimed. 2022, 25, 5413–5428. [Google Scholar] [CrossRef]
  61. Zhang, U.; Liu, Y.; Sun, P.; Yan, H.; Zhao, X.; Zhang, L. IFCNN: A general image fusion framework based on convolutional neural network. Inf. Fusion 2020, 54, 99–118. [Google Scholar] [CrossRef]
  62. Tang, W.; He, F.; Liu, Y.; Duan, Y. MATR: Multimodal medical image fusion via multiscale adaptive transformer. IEEE Trans. Image Process. 2022, 31, 5134–5149. [Google Scholar] [CrossRef]
  63. Xu, H.; Ma, J.; Zhang, X.-P. MEF-GAN: Multi-exposure image fusion via generative adversarial networks. IEEE Trans. Image Process. 2020, 29, 7203–7216. [Google Scholar] [CrossRef]
  64. Wang, D.; Liu, J.; Fan, X.; Liu, R. Unsupervised misaligned infrared and visible image fusion via cross-modality image generation and registration. arXiv 2022, arXiv:2205.11876. [Google Scholar] [CrossRef]
  65. Prabhakar, K.R.; Srikar, V.S.; Babu, R.V. Deepfuse: A deep unsupervised approach for exposure fusion with extreme exposure image pairs. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4714–4722. [Google Scholar]
  66. Li, H.; Wu, X. DenseFuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 2018, 28, 2614–2623. [Google Scholar] [CrossRef]
  67. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 502–518. [Google Scholar] [CrossRef]
  68. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  69. Rasheed, H.; Maaz, M.; Shaji, S.; Shaker, A.; Khan, S.; Cholakkal, H.; Anwer, R.M.; Xing, E.; Yang, M.; Khan, F.S. GLAMM: Pixel Grounding Large Multimodal Model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 13009–13018. [Google Scholar]
  70. Lin, W.; Wei, X.; An, R.; Ren, T.; Chen, T.; Zhang, R.; Guo, Z.; Zhang, W.; Zhang, L.; Li, H. Perceive Anything: Recognize, Explain, Caption, and Segment Anything in Images and Videos. arXiv 2025, arXiv:2506.05302. [Google Scholar] [CrossRef]
Figure 1. Overview of the framework. For visible and infrared images, individual features are extracted, and cross-attention mechanisms are used to fuse the features from both modalities. Subsequently, a text description corresponding to the visible image is generated using an image-to-text description network. The textual description is encoded through the CLIP text encoder, and a multi-layer perceptron (MLP) is employed to predict semantic information and fusion parameters, effectively guiding the image fusion process.
Figure 1. Overview of the framework. For visible and infrared images, individual features are extracted, and cross-attention mechanisms are used to fuse the features from both modalities. Subsequently, a text description corresponding to the visible image is generated using an image-to-text description network. The textual description is encoded through the CLIP text encoder, and a multi-layer perceptron (MLP) is employed to predict semantic information and fusion parameters, effectively guiding the image fusion process.
Sensors 25 05083 g001
Figure 2. Comparison of fusion results with and without textual guidance. (a) Infrared image; (b) Visible image; (c) Fused result without textual guidance; (d) Fused result with textual guidance. The red, green, and blue boxes highlight key semantic regions including license plate details, building structures, and traffic signs, respectively.
Figure 2. Comparison of fusion results with and without textual guidance. (a) Infrared image; (b) Visible image; (c) Fused result without textual guidance; (d) Fused result with textual guidance. The red, green, and blue boxes highlight key semantic regions including license plate details, building structures, and traffic signs, respectively.
Sensors 25 05083 g002
Figure 3. Infrared and visible image fusion experiment on “pedestrian crossing” images from the TNO dataset. PMGI [59], SeAFusion [9], YDTR [60], IFCNN [61], MATR [62], MEFGAN [63], UMF-CMGR [64], ours.
Figure 3. Infrared and visible image fusion experiment on “pedestrian crossing” images from the TNO dataset. PMGI [59], SeAFusion [9], YDTR [60], IFCNN [61], MATR [62], MEFGAN [63], UMF-CMGR [64], ours.
Sensors 25 05083 g003
Figure 4. Infrared and visible image fusion experiment on “nighttime vehicles and pedestrians” images from the LLVIP dataset. PMGI [59], SeAFusion [9], YDTR [60], IFCNN [61], MATR [62], MEFGAN [63], UMF-CMGR [64], ours.
Figure 4. Infrared and visible image fusion experiment on “nighttime vehicles and pedestrians” images from the LLVIP dataset. PMGI [59], SeAFusion [9], YDTR [60], IFCNN [61], MATR [62], MEFGAN [63], UMF-CMGR [64], ours.
Sensors 25 05083 g004
Figure 5. Infrared and visible image fusion experiment on “multiple pedestrians” images from the LLVIP dataset. PMGI [59], SeAFusion [9], YDTR [60], IFCNN [61], MATR [62], MEFGAN [63], UMF-CMGR [64], ours.
Figure 5. Infrared and visible image fusion experiment on “multiple pedestrians” images from the LLVIP dataset. PMGI [59], SeAFusion [9], YDTR [60], IFCNN [61], MATR [62], MEFGAN [63], UMF-CMGR [64], ours.
Sensors 25 05083 g005
Figure 6. Infrared and visible image fusion experiment on LLVIP dataset.
Figure 6. Infrared and visible image fusion experiment on LLVIP dataset.
Sensors 25 05083 g006
Table 1. Quantitative evaluation results on the TNO dataset.
Table 1. Quantitative evaluation results on the TNO dataset.
MethodSFQab/fMISDVIF
DeepFuse [65]8.35000.384713.220566.88720.5752
DenseFuse [66]9.32380.473513.705381.72830.6875
RFN-Nest [59]5.84570.329213.454767.87650.5404
PMGI [59]8.71950.378713.737669.23640.6904
U2Fusion [67]11.03680.393413.445366.50350.7680
IFCNN [61]11.85900.496213.290973.70530.6090
FusionGAN [68]8.04760.268213.081761.63390.4928
MEFGAN [63]7.84810.207613.945443.73320.7330
SeAFusion [9]11.93550.490814.066393.38510.8919
YDTR [60]3.25670.141012.386556.06680.2792
MATR [62]5.36320.272313.070578.07200.3920
UMF-CMGR [64]8.23880.367112.630160.72360.3934
Ours11.31490.586313.967694.72030.7746
Table 2. Quantitative evaluation results on the LLVIP dataset.
Table 2. Quantitative evaluation results on the LLVIP dataset.
MethodSFQab/fMISDVIF
DeepFuse [65]12.41750.462014.04440.458638.3328
DenseFuse [66]12.59000.470014.07230.466938.7011
RFN-Nest [59]10.68250.384414.12840.465839.7194
PMGI [59]12.09970.395114.07370.448737.9572
U2Fusion [67]17.28890.498513.41410.491737.4284
IFCNN [61]21.76980.609214.48350.676244.0938
FusionGAN [68]9.20620.060012.89810.114126.9133
MEFGAN [63]15.19050.364413.95750.872059.7947
SeAFusion [9]20.91940.618114.90160.839251.8096
YDTR [60]7.07550.196113.38580.336533.1625
MATR [62]13.50660.428211.99890.457534.1515
UMF-CMGR [64]13.44810.370713.40370.384135.1731
Ours22.13810.632914.23050.852745.1842
Table 3. Quantitative results under different image captioning models on the TNO dataset.
Table 3. Quantitative results under different image captioning models on the TNO dataset.
MethodSFQab/fMISDVIF
GLaMM [69]8.35000.384713.220566.88720.5752
Perceive [70]9.32380.473513.705381.72830.6875
ExpansionNet v2 [52]11.31490.586313.967694.72030.7746
Table 4. Ablation study of our methods on the TNo: Cro: Cross Attention. Lan: Language-Driven Fusion Module.
Table 4. Ablation study of our methods on the TNo: Cro: Cross Attention. Lan: Language-Driven Fusion Module.
CrossLanSFQab/fMISDVIF
7.32400.44328.623251.42280.6238
9.46120.519612.414072.45430.7426
11.31490.586313.967694.72030.7746
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, X.; Wu, J.; Zhang, P.; Yu, Z. Language-Driven Cross-Attention for Visible–Infrared Image Fusion Using CLIP. Sensors 2025, 25, 5083. https://doi.org/10.3390/s25165083

AMA Style

Wang X, Wu J, Zhang P, Yu Z. Language-Driven Cross-Attention for Visible–Infrared Image Fusion Using CLIP. Sensors. 2025; 25(16):5083. https://doi.org/10.3390/s25165083

Chicago/Turabian Style

Wang, Xue, Jiatong Wu, Pengfei Zhang, and Zhongjun Yu. 2025. "Language-Driven Cross-Attention for Visible–Infrared Image Fusion Using CLIP" Sensors 25, no. 16: 5083. https://doi.org/10.3390/s25165083

APA Style

Wang, X., Wu, J., Zhang, P., & Yu, Z. (2025). Language-Driven Cross-Attention for Visible–Infrared Image Fusion Using CLIP. Sensors, 25(16), 5083. https://doi.org/10.3390/s25165083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop