Next Article in Journal
Enhancing Road Safety with AI-Powered System for Effective Detection and Localization of Emergency Vehicles by Sound
Previous Article in Journal
Spatio-Temporal Agnostic Sampling for Imbalanced Multivariate Seasonal Time Series Data: A Study on Forest Fires
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Infrared and Visible Image Fusion Network Based on Res2Net and Multiscale Transformer

College of Electrical Engineering, University of South China, Hengyang 421001, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(3), 791; https://doi.org/10.3390/s25030791
Submission received: 29 November 2024 / Revised: 22 January 2025 / Accepted: 25 January 2025 / Published: 28 January 2025
(This article belongs to the Section Optical Sensors)

Abstract

:
The aim of infrared and visible image fusion is to produce a composite image that can highlight the infrared targets and maintain plentiful detailed textures simultaneously. Despite the promising fusion performance of current deep-learning-based algorithms, most fusion algorithms highly depend on convolution operations, which limits their capability to represent long-range contextual information. To overcome this challenge, we design a novel infrared and visible image fusion network based on Res2Net and multiscale Transformer, called RMTFuse. Specifically, we devise a local feature extraction module based on Res2Net (LFE-RN) in which dense connections are adopted to reuse the information that might be lost in convolution operation and a global feature extraction module based on multiscale Transformer (GFE-MT) which is composed of a Transformer module and a global feature integration module (GFIM). The Transformer module extracts the coarse-to-fine semantic features of the source images, while GFIM is used to further aggregate the hierarchical features to strengthen contextual feature representations. Furthermore, we employ the pre-trained VGG-16 network to compute the loss of features with different depths. Massive experiments on mainstream datasets indicate that RMTFuse is superior to the state-of-the-art methods in both subjective and objective assessments.

1. Introduction

With images captured by one-modality sensors, it is difficult to comprehensively and efficiently characterize the imaging scene on account of theoretical and technological limitations [1]. The visible sensor describes the scene texture details by reflective visible light, but it is vulnerable to various environmental interference like scene brightness. By comparison, the infrared sensor highlights the radiative properties of salient targets through thermal radiation, but it ignores details and is also susceptible to noise. Visible and infrared image fusion can integrate complementary features from the two types of images, resulting in an informative fusion result and thereby promoting numerous visual tasks [2], such as semantic segmentation [3], object tracking as well as detection [4,5,6], and biometric recognition [7].
In the last few decades, researchers have put forward a large number of fusion algorithms for infrared and visible images. Generally, they are divided into two main categories: traditional image fusion methods and deep-learning-based fusion models. Traditional image fusion methods are concerned with extracting important features from the original images. The method on the basis of multiscale transformation [8,9] is a common traditional method for image fusion. It first extracts multiscale features from original images and subsequently merges these features using hand-crafted fusion rules. To reconstruct the fusion image, the inverse multiscale transformation is finally adopted. The main challenge of this type of fusion method lies in effectively extracting features from the original images and formulating appropriate rules to obtain an informative fusion image. Aside from multiscale transformation, other traditional image fusion algorithms include subspace-based [10], sparse-representation-based [11], optimization-based [12], hybrid methods [13], and others [14]. However, a great hindrance is present in the advancement of traditional image fusion techniques. For one thing, such handcrafted fusion strategies and rules cannot fully leverage the complementary features of two different modalities. For another, these traditional methods are not suitable for integration in complicated scenes as a result of the limitation of representation ability.
In the past decade, deep learning theories have been deployed in diverse computer vision challenges successfully, thus promoting the implementation of image fusion approaches based on deep learning theories. Existing deep-learning-based methods are roughly categorized into three types, namely convolutional neural network (CNN)-based methods [15,16], autoencoder (AE)-based methods [17,18], and generative adversarial network (GAN)-based methods [19,20]. Owing to their powerful capability in extracting features, CNN-based fusion methods effectively address the manual design issues presented in traditional methods by performing integration strategies and activity level measurements holistically through convolution operations. They concentrate on the way to devise excellent network frameworks and formulate appropriate loss functions to perform feature extraction, fusion, and reconstruction, ultimately achieving distinctive fused results. Unlike CNN-based algorithms, the fusion methods based on AE train both the encoder and decoder to achieve the extraction and reconstruction of complementary features, whereas the process of merging these obtained features is accomplished through carefully designed fusion rules. In contrast, the methods based on GAN introduce the generative adversarial mechanism into the fusion network, thereby constantly optimizing the generated results and forcing them to approximate the ideal probability distribution with non-supervision.
With powerful feature extraction and generalization abilities, deep-learning-based fusion methods have made considerable progress in the image fusion domain, far surpassing traditional methods. Unfortunately, these deep learning methods above tend to lose global features as they rely on the CNN to execute particular operations during the phase of feature extraction [21]. The CNN indeed has a good local feature extraction ability and generalization ability, but it struggles to effectively capture long-range dependencies in images owing to the limited receptive field. In contrast, Transformer performs well in modeling the long-range dependencies of images. With this motivation, we develop an effective fusion network for infrared and visible images that successfully integrates the CNN and vision Transformer. The fusion network can maintain the merit of the CNN as well as improve the long-range dependencies of input images, thus enhancing fusion quality. Specifically, a local feature extraction module based on Res2Net (LFE-RN) is devised to fully utilize the multiscale features of a local neighborhood. In order to avoid weakening some beneficial information for fusion, we apply dense connections in LFE-RN. Moreover, we devise a global feature extraction module based on multiscale Transformer (GFE-MT) for further preserving global context semantic information.
The primary contributions in this paper can be characterized as follows:
  • We put forward a novel and efficient network that combines CNN and Transformer for the fusion of infrared and visible images. The proposed network performs well in integrating complementary information by effectively utilizing both local and global features of source images.
  • We devise a densely shaped LFE-RN to effectively exploit local features and reuse information that could be lost during the feature extraction operation. A Transformer module and global feature integration module (GFIM) are devised in GFE-MT to further preserve global contextual information.
  • A targeted perceptual loss function is devised to retain the high similarity between source images and the fused result.
  • Extensive experiments on two dominant datasets, i.e., TNO and RoadScene, illustrate that our method surpasses other state-of-the-art image fusion methods in terms of both subjective effects and objective evaluations.
The remainder of this paper is organized in the following manner: In Section 2, a brief overview of relevant research on image fusion is provided. Section 3 introduces our RMTFuse in detail. In Section 4, we carry out an ablation study to validate the impacts of each module, as well as illustrate the outstanding performance of RMTFuse compared to other methods. Finally, some concluding remarks on our work are presented in Section 5.

2. Related Work

In this section, we first review the recent advances in deep-learning-based image fusion methods, followed by a detailed description of Res2Net and vision Transformer.

2.1. Deep-Learning-Based Image Fusion Methods

2.1.1. CNN-Based Fusion Methods

Liu et al. [22] were pioneers in utilizing CNNs for image fusion tasks. In their research, a CNN was employed to calculate a binary map to obtain fused multi-focus images. Subsequently, Liu et al. [23] introduced a Siamese CNN into the network to acquire weight maps from input images to settle the problem of integrating infrared and visible images. They broke input images into pyramids and conducted fusion in a multiscale manner. In [16], Zhang et al. converted the problem of unified image fusion to maintain the radio of gradient and intensity in the source images. They also proposed a generic loss function applicable to different fusion tasks. Xu et al. [24] developed a unified fusion model that estimated the significance of different source images in an adaptive way. In order to adequately exploit valuable information from source images, Guo et al. [25] evaluated the weight score matrix of input images to identify the respective contributions of each image and introduced a masking strategy into their loss function. To better facilitate the downstream tasks during image fusion, a semantic-aware fusion framework was designed in [26], namely SeAFusion.

2.1.2. AE-Based Fusion Methods

Benefiting from the development of deep learning, some researchers have put forward the AE-based image fusion algorithms. In order to accomplish feature extraction and image reconstruction, most of them pre-train an autoencoder on large-scale datasets. Subsequently, an appropriate fusion rule is devised to perform the feature fusion task. A famous and pioneering AE-based method for the integration of infrared and visible images was DenseFuse [27]. It adopted a dense block for extracting image features and selected an additive strategy or l1-norm strategy to achieve feature fusion. Furthermore, Li et al. [28] brought nest connections into the network to extract multiscale features. Nevertheless, they all applied manual fusion strategies to integrate features, weakening the performance of pre-trained fusion model. To tackle this problem, they later presented an end-to-end fusion framework [17], namely RFN-Nest, where feature fusion is performed by a learnable residual fusion framework. Considering that multiscale characteristics greatly affect the performance of fused images, Wang et al. [29] developed a multiscale encoder–decoder fusion network based on Res2Net and double nonlocal attention models to retain as much significant information as possible. More recently, a multilevel dual-branch attention network was proposed in MDAN [18] to effectively reduce the information loss in the feature extraction process, so as to obtain an information-rich fused image.

2.1.3. GAN-Based Fusion Methods

A GAN can effectively evaluate probability distributions in an unsupervised way, which makes it a good choice for image fusion tasks. In 2019, Ma et al. [20] firstly introduced a GAN into the image fusion field, casting the fusion task as an antagonistic game between a generator and a discriminator. Unfortunately, the obtained result closely resembled a sharpened infrared image as there is only one discriminator. Subsequently, they further introduced dual discriminators into the GAN-based image fusion to achieve fusion balance [30]. In GANMcC [31], the image fusion issue is converted into the multi-classification constraint, and the obtained fused result is more balanced. In 2022, a target-aware adversarial model was devised in TarDAL [32] that effectively combined image fusion and follow-up detection tasks. More recently, in MAGAN [33], multi-attention mechanisms were introduced into a generator and two discriminators to fuse images selectively. However, GANs are hard to optimize, and the generated images may be biased towards either of the input images.
Although the above deep-learning-based fusion methods have achieved satisfactory results, there are still some shortcomings that need to be addressed. On the one hand, most of these fusion methods only employ the feature maps of the final layer and are unable to preserve features across different scales, which may weaken the useful information for the image fusion task. On the other hand, these methods usually rely on convolution operations, which fail to fully capture global significant features because of the limited receptive fields. As a result, the Res2Net block and multiscale vision Transformer are introduced into the proposed model to extract the multiscale local and global features, improving the quality of fused images.

2.2. Res2Net

In [34], a new multiscale backbone module called Res2Net was proposed to strengthen the multiscale expression ability of a CNN. It constructs hierarchical residual connections to represent multiscale features at a fine-grained level. Figure 1 shows the detailed architecture of Res2Net. A 1 × 1 convolutional layer is first adopted for adjusting the channel numbers of input feature maps to facilitate subsequent processing. The gained feature maps are then equally divided into s feature subsets that have the same spatial size, where s is the scale control parameter. Afterward, with the exception of the first one, every subset undergoes the operation with a corresponding convolution of 3 × 3 and is integrated into the following feature subset. Then, all of these acquired features are fed into another 1 × 1 convolution and connected with primitive feature information, thereby obtaining an output result that has a larger receptive field.
In this paper, the Res2Net is viewed as a convolution block and introduced into the fusion framework for fine-grained multiscale feature extraction. In addition, dense connections are employed to achieve feature reuse and strengthen information propagation. The scale control parameter s is set to 4 in subsequent experiments.

2.3. Vision Transformer

In 2017, the Transformer architecture was first proposed [35] and employed in natural language processing (NLP) tasks, such as word prediction and sentiment categorization. Motivated by the success of Transformer in NLP, the vision Transformer (ViT) structure was put forward by Dosovitskiy et al. [36] to perform image classification tasks. They split the input images into 16 × 16 patches, which were then fed into a standard Transformer. Given its powerful ability for long-range modeling, Transformer has excelled in various visual tasks, which include target detection, semantic segmentation, image restoration, etc. This contributed to the application of Transformer-based networks in the domain of image fusion. In 2021, VS et al. [37] designed a novel fusion strategy based on spatial CNN and Transformer branches to fuse local and global semantic information. Ma et al. [38] devised a unified fusion network on the basis of Swin Transformer. Specifically, the Transformer was utilized to represent the long dependencies within the same domain as well as across domains for better complementary information integration. In addition, Tang et al. [39] proposed DATFuse, in which a dual attention residual structure and Transformer were combined to achieve better fusion performance. In [40], Transformer was introduced into a CNN-based fusion framework to improve the global dependencies of images. In their network, the CNN module composed of a structure branch and detail branch was utilized for shallow feature extraction, while the Transformer module was designed for the exploration of long-range dependencies.
However, most of these Transformer-based fusion methods simply combine the CNN and Transformer in a cascade manner to successively extract the local and global features. Moreover, they ignore the representation of multiscale features, which inevitably weakens significant information for fusion. Unlike these methods, we design a local–global parallel fusion network to adequately extract significant features from source images. In the proposed network, we introduce the Res2Net as a multiscale convolutional block into the local feature extraction module to exploit features at different scales. In addition, since the original vision Transformer can only output single-scale features, the multiscale Transformer is devised in our network to acquire coarse-to-fine semantic features.

3. Methods

In this section, a thorough introduction to the proposed fusion network RMTFuse is provided. Firstly, an overall framework of RMTFuse is briefly introduced. Then, we emphatically describe the fusion network. Finally, we present the hybrid loss function and its mathematical expression.

3.1. Overall Framework

The proposed network is end-to-end and can be classified into two modules: the feature extraction module and the image reconstruction module. Firstly, the visible image I v i s and the infrared image I i r are concatenated in the channel dimension. Then, they are sent to the shallow feature extraction module (SFEM) to obtain shallow features. Next, the mixed shallow features are input to a dual-stream structure, which comprises LFE-RN and GFE-MT, to acquire local and global complementary features simultaneously. Subsequently, the obtained local and global features are concatenated before being fed into the image reconstruction module to achieve feature reconstruction. Finally, we are able to obtain the fused result I f . The overview of the proposed RMTFuse is shown in Figure 2. For convenience, Table 1 lists some abbreviations utilized in this section.

3.2. Network Architecture

As exhibited in Figure 2, the two-channel map { I v i s , I i r } obtained by concatenating visible and infrared images is first sent to the SFEM which comprises two successive convolutional layers to extract shallow feature F S O . Each convolutional layer has a kernel size of 3 × 3, and LReLU is adopted as the activation function. Subsequently, local and global features are obtained by using parallel modules, LFE-RN and GFE-MT, respectively. In the end, the obtained feature maps are concatenated and transmitted to the image reconstructor to generate the final result. We will describe the LFE-RN and GFE-MT together with the image reconstructor in detail.

3.2.1. Local Feature Extraction Module Based on Res2Net

To enable the network to learn texture details at different scales, we introduce the Res2Net blocks into the local feature extraction module. In Figure 2, it can be observed that LFE-RN deploys two Res2Net blocks and an ordinary convolutional layer. The Res2Net blocks represent the fine-grained multiscale features by aggregating multiple receptive fields hierarchically. Meanwhile, dense connections are adopted to achieve feature reuse, effectively retain multiscale detailed information, and improve fusion performance. Concretely, F S O is sent to the first Res2Net block, which is expressed as
F R 1 O = R B 1 ( F S O )
where R B 1 ( · ) and F R 1 O denote the first Res2Net block and the output intermediate features, respectively. Then,   F R 1 O and F S O are concatenated and fed into the second block, which can be represented as
F R 2 O = R B 2 ( C ( F S O , F R 1 O ) )
where R B 2 ( · ) indicates the second Res2Net block, C ( · ) indicates the concatenation conducted on the channel dimension, and F R 2 O represents the obtained features extracted from the second Res2Net block. Subsequently, F R 2 O , F R 1 O , and F S O are concatenated and sent to the last convolutional layer, which can be formulated as
F L F E R N O = C o n v 1 × 1 ( C ( F S O , F R 1 O , F R 2 O ) )
where C o n v 1 × 1 ( · ) is an ordinary convolutional layer with kernel size 1 × 1 and LReLU activation function. It is deployed to eliminate channel dimension differences. F L F E R N O is the final result of LFE-RN. Table 2 shows the parameter settings of LFE-RN. H and W represent the height and width of input images.

3.2.2. Global Feature Extraction Module Based on Multiscale Transformer

Due to the fact that the fusion performance will degrade without global context information, GFE-MT is devised for constructing long-range relationships. The design of GFE-MT is based on [41], as depicted in Figure 2. It includes two parts: the Transformer module and GFIM.
(1)
Transformer Module
The fusion network can generate favorable fused results by introducing a primitive ViT module only. However, its fusion performance may be affected in the absence of global multiscale information. Therefore, we design a multiscale vision Transformer for global feature preservation across different scales. The Transformer module is made up of three different Transformer blocks with varied scales. To be exact, the overlapped patch merging module in [42] is employed within the original ViT model to acquire a hierarchical feature map, which has a resolution of H 2 i + 1 × W 2 i + 1 . H × W means the size of input feature map, and i 1,2 , 3 . Moreover, the operation of overlapped patch merging can make the previously uncorrelated image block sequences correlative by setting K = 7, S = 4, P = 3, and K = 3, S = 2, P = 1, where K, S, and P mean convolutional kernel size, convolutional stride, and padding size. Table 3 illustrates the parameter settings of the Transformer module. T B i (i = 1, 2, 3) represents the Transformer blocks. O-C symbolizes the number of output channels. E-D stands for the embedding dimensions. N-B denotes the number of Transformer encoders used within a Transformer block.
The detailed architecture of the Transformer block adopted in our module is depicted in Figure 3. The obtained features will be fed into the subsequent Transformer encoder after the overlapped patch merging operation. The employed encoder comprises two LayerNorm (LN) units, multi-head self-attention (MSA), and a multi-layer perceptron (MLP). In addition, it has two residual connection operations, and the exact process can be formulated as follows:
F R C 1 = M S A ( L N ( F T I ) ) + F T I
F R C 2 = M L P ( L N ( F R C 1 ) ) + F R C 1
where F T I denotes the input of Transformer encoder. F R C 1 and F R C 2 represent the corresponding output results of the first and second residual connection operations.
(2)
Global Feature Integration Module
We can obtain hierarchical features through the multiscale Transformer module, from coarse to fine-grained semantic features. Therefore, our goal is to determine how to effectively integrate these features and strengthen the feature representations. To this end, we propose the GFIM which adopts convolution operation, upsampling, and channel connection to gradually recover the spatial size of input feature maps and merge channels. According to Figure 4, the core idea of GFIM is to cascade high-level scale features to low-level scale ones. In GFIM, F T i represents the input tensor with a resolution of H 2 i + 1 × W 2 i + 1 × C , where i 1,2 , 3 , C 32,64,128 , and F T 0 is the input of the Transformer module. We define C o n v 3 × 3 ( · ) as a convolutional block with 3 × 3 kernel size, followed by the LReLU activation function. Firstly, the fine-grained features F T 3 derived from the third Transformer block are fed into convolutional blocks C o n v 3 × 3 3 ( · ) , C o n v 3 × 3 2 ( · ) , and C o n v 3 × 3 1 ( · ) via element-level projection mapping to acquire features F 33 , F 32 , and F 31 . Obviously, although those obtained feature maps are consistent with the lower-level feature information F T 2 , F T 1 , and F T 0 in the channel dimension, their spatial sizes do not coincide. Therefore, bicubic interpolation is employed to upsample feature maps F 33 , F 32 , and F 31 to ensure they have the same spatial size as F T 2 , F T 1 , and F T 0 , respectively. The specific calculated process is denoted as
F 3 j = U P ( C o n v 3 × 3 j F T 3 )
where j 1,2 , 3 , and F 3 j is the feature map acquired by the upsampling operation. U P ( · ) denotes the upsampling operation.
The obtained F 33 and feature maps F T 2 are then concatenated for the purpose of merging global information of F T 3 and F T 2 . The concatenated features are first projected and mapped by a convolution block C o n v 1 × 1 2 ( · ) with kernel size 1 × 1 and LReLU activation function. To interact with lower-scale feature F T 1 and F T 0 , the feature maps F T 2 ~ obtained by the above projection mapping are then sent to two convolutional blocks C o n v 3 × 3 5 ( · ) and C o n v 3 × 3 4 ( · ) and subsequent upsampling blocks to obtain feature maps F 22 and F 21 which are consistent with F T 1 and F T 0 in channel dimension and spatial size, respectively. The specific process is formulated as
F 2 j = U P ( C o n v 3 × 3 j + 3 C o n v 1 × 1 2 ( C ( F T 2 , F 33 ) ) )
where j 1,2 and C ( · ) implies the operation of concatenation.
To effectively integrate shallow and deep feature information, the obtained feature maps F 32 and F 22 are concatenated with input tensor F T 1 . Subsequently, the concatenated features F T 1 ~ pass through convolutional blocks C o n v 1 × 1 1 ( · ) and C o n v 3 × 3 6 ( · ) and an upsampling block to obtain feature maps F 11 that could interact with input F T 0 . Then, we concatenate F 31 , F 21 , F 11 , and F T 0 and send them to convolutional blocks C o n v 1 × 1 0 ( · ) and C o n v 3 × 3 7 ( · ) successively to obtain the final output feature maps F G F E M T O . The whole calculation process is denoted as follows:
F 11 = U P ( C o n v 3 × 3 6 ( C o n v 1 × 1 1 ( C ( F 32 , F 22 , F T 1 ) ) ) )
F G F E M T O = C o n v 3 × 3 7 ( C o n v 1 × 1 0 ( C ( F 31 , F 21 , F 11 , F T 0 ) ) )

3.2.3. Image Reconstructor

In our network, the obtained local features F L F E R N O and global features F G F E M T O are then integrated via a concatenation strategy, and these aggregated feature maps are sent to the image reconstructor to generate the fused result. As shown in Figure 2, the image reconstructor consists of four successive convolutional layers with 3 × 3 kernel size and stride = 1.

3.3. Loss Function

As the proposed RMTFuse adopts unsupervised learning, the design of the loss function in the model is essential to the quality of fused results. In order to constrain the generated fusion images to capture additional complementary features from inputs, for example, the abundant texture details and significant infrared targets, we devise a hybrid loss function. It comprises per-pixel loss L p i x e l and perceptual loss L p e r , and the expression is as follows:
L t o t a l = L p i x e l + α L p e r
where α is a weight-off factor for balancing the total loss function.

3.3.1. Per-Pixel Loss

This part of the loss includes L s s i m and L T V , which can be formulated as
L p i x e l = L s s i m + β L T V
where L s s i m and L T V represent SSIM loss and total variation loss, respectively. β is a weighting factor used to control the contributions of each term.
The structural similarity (SSIM) index can reflect distortion from different perspectives, such as brightness, contrast, and structure [43]. With the aim of maintaining more structure information from input images, we introduce SSIM loss to constrain the structural similarity of output and input images. It is computed as
L s s i m = 1 S S I M ( I f , I s )
where I f is the output and I s denotes the source images. S S I M ( · ) represents the structural similarity measurement.
To better retain the gradient information in original images and eliminate the noise of generated results, we employed the total variation (TV) loss introduced in VIF-Net [44]. It is formulated as
T p , q = I f p , q I s ( p , q )
L T V = p , q ( T p , q + 1 T p , q 2 + T p + 1 , q T p , q 2 )
where T ( · )   means the difference between the source images and the fused image. · 2 represents the l 2 norm, while p and q denote the horizontal and vertical coordinates of the image pixels, respectively.

3.3.2. Perceptual Loss

It challenging to acquire the perceptual difference between source images and generated images in the image fusion field by only per-pixel loss [45]. For instance, two identical images differing from each other by just a few pixels may be quite different when measured in terms of per-pixel loss, despite their perceptual similarity. To this end, we introduce perceptual loss into hybrid loss for retaining high similarity between merged images and original images, so as to further improve fusion performance. It is denoted as
L p e r = k = 2,4 , 6 φ k ( I f ) φ k ( I s ) 1
where φ k indicates the k-th layer feature map obtained from the pre-trained VGG-16 model [46]. · 1 denotes the l 1 norm.

4. Experiments and Discussion

In this section, we first present the training details of our RMTFuse. Then, we provide the comparison algorithms and objective indices. Next, ablation experiments are performed to validate the effectiveness of our network structure. Afterward, the comparative and generalization experiments are presented to illustrate the superiority of RMTFuse. We also provide the computational complexity analysis. In the end, object detection experiments are conducted to assess different fusion models in the context of computer visual challenges.

4.1. Training Details

In this work, we choose 36 pairs of corresponding infrared and visible images from the TNO [47] benchmark as the training set. In order to create enough samples for training our model, original image pairs are cropped into patches of size 256 × 256 by setting 12 for the overlapping cropping step. Thus, approximately 20,000 image patches are collected and converted into grayscale images. We employ the Adam optimizer to update weights and set the learning rate to 0.0001 during the process of training. Meanwhile, the batch size and epoch are established at 8 and 30, respectively. The weight parameters, α and β , are set to 0.001. Moreover, our model is built on the PyTorch framework, and all experiments are implemented on the NVIDIA GeForce RTX 4090 GPU (NVIDIA Corporation, Santa Clara, CA, USA).

4.2. Comparison Methods and Objective Indices

In our experiments, two publicly accessible datasets, i.e., TNO [47] and RoadScene [24], are used to comprehensively estimate the fusion quality of our RMTFuse. Nine state-of-the-art image fusion methods, including BF [14], FusionGAN [20], DenseFuse [27], GANMcC [31], LRRNet [48], YDTR [49], SwinFusion [38], DATFuse [39], and CMTFusion [50], are picked for performance comparison. All source codes of the comparison methods above are accessible to the public, and we configure the parameters following the recommendations of the corresponding papers.
For quantitative evaluation, eight widely used metrics are deployed for assessing the fusion results objectively, which include standard deviation (SD) [51], visual information fidelity (VIF) [52], entropy (EN) [53], gradient-based fusion performance (Qabf) [54], spatial frequency (SF) [55], modified fusion measure (Nabf) [56], multiscale SSIM (MS-SSIM) [57], and feature mutual information with pixel (FMIpixel) [58]. SD reflects the distribution and contrast of fusion results in a statistical way. VIF computes the distortion of the image to measure the information fidelity in fusion results in the view of the human visual system. EN indicates the amount of information that an image contains. Qabf aims to measure the level of edge information transmitted from original images to the final generated result. SF evaluates the change rate of the image gray level and reveals the textural characteristics of an image. Nabf aims to compute noise and artifacts added to the generated results. MS-SSIM measures the similarity between the original images and the fusion image on different levels of scale. FMIpixel calculates mutual information of features from pixel features. Moreover, except for Nabf, the higher the value of the evaluation index, the more effective the fusion network is.

4.3. Ablation Experiments

Owing to the introduction of LFE-RN and GFE-MT in the proposed RMTFuse, it is equipped with the ability to extract local and global complementary features from source images adequately, thereby generating informative fused results. To verify the effects of LFE-RN and GFE-MT, 25 image pairs from the TNO dataset are selected to perform ablation studies. To validate the necessity of LFE-RN, we replace the Res2Net blocks with normal convolution layers, named w/o RN. To illustrate the significance of GFE-MT, we remove the GFE-MT from the whole architecture, called w/o GFE-MT. Similarly, the original ViT module is employed to replace the multiscale Transformer module to explore the impacts of the proposed multiscale vision Transformer structure in GFE-MT, termed w/o MT.
The ablation experiment results on the TNO dataset of different structures are depicted in Figure 5. Two local areas of each fusion result are marked, and one of them is zoomed in as a close-up for better observation. Clearly, the fusion result of w/o GFE-MT suffers from insignificant infrared targets as a result of lacking the global feature extraction ability. Moreover, some important detailed textures are lost, leading to unclear fusion results. With the addition of the original ViT module, more prominent features of the source images are maintained, leading to a brighter target person in the fused result of w/o MT. However, the overall scene of w/o MT is drowned in darkness, which is unable to bring a pleasant visual experience. In contrast, the fused result of w/o RN alleviates the darkened result to some extent, but the thermal target is still contaminated by the visible image compared with our fused image. With the introduction of Res2Net and multiscale Transformer module, our result achieves the best fusion performance, which not only maintains the prominent information in infrared images, but also exhibits excellent ability in preserving the texture details, such as the clear door frame and floor. From the ablation experiment results above, it can be seen that adding the multiscale Transformer module can enhance the long-range dependency modeling, which enables the proposed model to make the best of local and global information. In addition, the Res2Net block in the proposed network also facilitates the preservation of useful complementary information in source images.
The objective comparisons of our RMTFuse with other network structures are depicted in Table 4. The top-performing result for each evaluation metric is marked in bold, while the second-best result is underlined. It is clear that the complete network structure displays optimal performance over all metrics, which is consistent with the subjective evaluation results. These experimental results all show that our proposed network structure is effective and reasonable.

4.4. Comparative Experiments

To intuitively illustrate the strength of our model, we pick out three representative scenes from the TNO dataset: (1) Kaptien_1123; (2) Sandpath_18; (3) Marne_04. The three sets of infrared and visible images along with their corresponding fusion images obtained by different methods under the same environment are shown in Figure 6, Figure 7 and Figure 8. And we use green boxes and red boxes to mark salient objects and texture details, respectively, to display the fusion difference clearly. Some marked areas are then magnified and positioned in the corners. Despite the favorable fused images achieved by nine comparison algorithms, there are still some limitations compared with RMTFuse. Specifically, BF, DenseFuse, YDTR, and DATFuse can preserve part of the texture details, but the intensity information of prominent objects is weakened to different degrees, such as the targets in Figure 6 and Figure 7 and the tires and engine of the jeep in Figure 8. Moreover, the merged images generated by these algorithms have lower contrast. Similarly, the fused results of LRRNet exhibit low brightness of infrared targets, which is unfavorable for object tracking and detection tasks. Although FusionGAN and GANMcC are able to successfully highlight the thermal targets, the obtained fused results are quite fuzzy, which leads to the loss of substantial scene information, such as the floor and sky in Figure 6. Similar phenomena can be observed in Figure 7 and Figure 8. In contrast, the fusion results obtained by CMTFusion have good contrast and successfully preserve the meaningful information of original images. Unfortunately, the infrared target is still weakened to a certain extent. It is notable that the fusion performance of SwinFusion outperforms the aforementioned approaches; however, the ability of our RMTFuse to preserve texture details is better. As presented in Figure 6 and Figure 8, the proposed RMTFuse effectively maintains the high-contrast areas in infrared images and clearly shows the door frame, the jeep windows, and the edge of clouds. Similarly, in Figure 7, the tree branches are apparent in our fusion result, which is unmatched by other algorithms. Overall, our RMTFuse provides the best visual experience. Moreover, the fusion results generated by our network contain both salient infrared targets and rich texture information.
In order to further demonstrate the excellent performance of RMTFuse, 25 image pairs from the TNO dataset are selected to perform the objective assessment. The mean values of eight different assessment metrics calculated from these images generated by the aforementioned fusion approaches are listed in Table 5. The optimal metric value is highlighted in bold, while the second-best metric value is underlined. It is clear that our RMTFuse performs the best in all metrics. The optimal VIF and SD illustrate that our fused results provide satisfying visual effects and high contrast. The highest EN means scene information is effectively maintained in the fused images. Moreover, RMTFuse obtains the highest values in SF, MS-SSIM, and Qabf, indicating that we can preserve more source image details, structural information, and edge information. The best FMIpixel metric means the proposed RMTFuse conserves a great deal of feature information transmitted from the original image. Furthermore, the lower Nabf reveals that our results contain less noise and artifacts. Overall, our method delivers the optimal fusion performance. Undoubtedly, this result is in line with the subjective assessment.

4.5. Generalization Experiments

The quality of generalization experiments is a significant factor in evaluating data-driven algorithms. Hence, we carry out comparison experiments on the RoadScene dataset for the purpose of verifying the generalization capability of the proposed RMTFuse. Notably, the proposed model of RMTFuse is trained on the TNO dataset and directly tested on the RoadScene dataset. For subjective evaluation, three representative scenes are selected to present the fused results, as shown in Figure 9, Figure 10 and Figure 11. It can be observed that the traditional method BF loses some local details and thermal information, such as the pedestrians in every scene, the tree branches in Figure 9, and the manhole covers in Figure 10. Similarly, the thermal radiation information is unremarkable in the generated results of DenseFuse and CMTFusion. Although these two methods can successfully preserve some texture details, they cannot learn salient information in infrared images. The fused images of FusionGAN are relatively blurred and lose lots of scene textures. Despite the fact that GANMcC and YDTR have more texture information, compared with our method, the obtained texture details are still limited, such as the tire details in the car rear in Figure 10. As for the LRRNet method, it fails to retain infrared thermal information sufficiently, for example, the significant targets in each scene. Obviously, our RMTFuse performs well in retaining salient objects and texture information simultaneously. As shown in Figure 11, the pedestrian is salient and the street line is clear in the fused result of RMTFuse. Unfortunately, there is a bias in the integration of the brightness information in both SwinFusion and DATFuse, resulting in an overexposure scene and missing the boundary information and detailed textures in the car rear. The same result can also be observed in other scenes.
We randomly pick 40 image pairs from the RoadScene dataset to evaluate objectively. Table 6 displays the average values of RMTFuse with the other nine comparative fusion approaches on different assessment metrics. It is evident that our RMTFuse performs best on six metrics, which include SD, VIF, Qabf, SF, Nabf, and MS-SSIM. This reveals that our fused results not only preserve texture details and abundant information, but also have the best visual perception and the least amount of noise. Furthermore, RMTFuse ranks second in EN and FMIpixel metrics, which means that the fusion results generated by RMTFuse contain abundant scene information and feature information.

4.6. Computational Complexity Analysis

To evaluate the computation complexity of the proposed RMTFuse, we present the average running time of different methods on TNO and RoadScene datasets in Table 7. The traditional method BF is implemented on a computer platform with an i9-13900KS CPU, and deep-learning-based methods are conducted on an NVIDIA GeForce RTX 4090 GPU. As shown in Table 7, our RMTFuse ranks moderate among all the fusion methods owing to the introduction of the multiscale Transformer module. However, it achieves better fusion performance compared with other methods. In general, the proposed RMTFuse remains competitive.

4.7. Detection Performance

To validate the effectiveness of our RMTFuse in facilitating computer vision tasks, we evaluate the fused results generated by different methods using a prevailing object detection method. Specifically, the YOLOX detector [59] is utilized for detection. And we randomly select 52 image pairs from the MFNet [3] dataset for evaluation, including both daytime and nighttime images.
Figure 12 and Figure 13 show some typical objection detection results of the source images and different fused results. In the 00008N scene, BF is unable to maintain the thermal target, while FusionGAN cannot keep the sharpened edges of the pedestrian target. Consequently, the detector cannot detect the pedestrian from the fusion results. It can be observed that only LRRNet, YDTR, DATFuse, CMTFusion, and our RMTFuse detect both people and two cars. However, due to the interference of negative information, LRRNet and CMTFusion weaken the significant targets, resulting in low confidence in pedestrian targets. In addition, the YDTR and DATFuse fail to preserve the detailed textures in source images, which leads to lower confidence in detecting cars than our RMTFuse. A similar phenomenon can be seen in the 00315D scenario. In the 00315D scene, the visible image has lower confidence for pedestrian detection than the infrared image as a result of illumination factors. SwinFusion and the proposed RMTFuse accurately detect all targets in the scene, while the remaining methods fail to detect the car in the distance. Overall, our method fully integrates thermal targets and texture details from source images and has a confidence level closer to the source images over all detected objects. This indicates that RMTFuse is more conducive to computer vision tasks.
To further measure the detection performance of different approaches, we utilize the mean average precision (mAP) to conduct a quantitative assessment. Table 8 shows the assessment results. mAP@0.5 and mAP@0.9 imply the mAP values when the IoU threshold is 0.5 and 0.9, respectively. The closer the mAP value is to 1, the higher the quality of object detection. Obviously, the proposed RMTFuse achieved the highest average detection accuracy at both thresholds, which reveals the optimal detection performance. And we can also observe that, in comparison to other fusion algorithms, our RMTFuse has a higher accuracy of detection on the car. This indicates that our method is able to better maintain the valuable information in infrared and visible images. In brief, our RMTFuse excels in the image fusion task, which in turn helps to improve object detection accuracy.

5. Conclusions

In this study, we develop a novel end-to-end image fusion model for infrared and visible images, named RMTFuse. The proposed network utilizes the strengths of the Res2Net block and multiscale Transformer structure to achieve the extraction of local and global complementary features. In our network, we devise a local feature extraction module based on Res2Net (LFE-RN) to capture multiscale local features and adopt dense connections in the module to achieve feature reuse. To realize long-range dependency modeling and further expand the receptive field, we develop a global feature extraction module based on multiscale Transformer (GFE-MT), in which a global feature integration module (GFIM) is devised for aggregating feature information at different scales. It is worth remarking that LFE-RN and GFE-MT are ordered in parallel, allowing more efficient utilization of local detail information and global contextual features. Moreover, we utilize the pre-trained VGG-16 network for extracting features at different levels to compute the loss. Massive experiments on the TNO and RoadScene datasets illustrate that our proposed RMTFuse surpasses current state-of-the-art approaches in terms of both subjective effects and objective evaluations. Furthermore, ablation experiments are performed to verify the effects of different module designs employed in the proposed network. And the extended experiments on object detection also reveal the advantages of RMTFuse for computer vision tasks. In the future, we expect to further study how to integrate the image alignment task with the image fusion task organically to realize a unified model for unaligned images. We will also generalize the algorithm to different image fusion fields, for example, multi-focus image fusion, medical image fusion, and so on.

Author Contributions

Conceptualization, B.T. and B.Y.; methodology, B.T.; software, B.T.; validation, B.T. and B.Y.; writing—original draft preparation, B.T.; writing—review and editing, B.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 61871210.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Source of dataset in experimental analysis: https://figshare.com/articles/dataset/TNO_Image_Fusion_Dataset/1008029 (accessed on 10 June 2024), https://github.com/hanna-xu/RoadScene (accessed on 10 June 2024), and https://github.com/haqishen/MFNet-pytorch (accessed on 12 June 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, H.; Xu, H.; Tian, X.; Jiang, J.; Ma, J. Image fusion meets deep learning: A survey and perspective. Inf. Fusion 2021, 76, 323–336. [Google Scholar] [CrossRef]
  2. Zhang, X.; Demiris, Y. Visible and infrared image fusion using deep learning. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10535–10554. [Google Scholar] [CrossRef] [PubMed]
  3. Ha, Q.; Watanabe, K.; Karasawa, T.; Ushiku, Y.; Harada, T. MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 5108–5115. [Google Scholar] [CrossRef]
  4. Chandrakanth, V.; Murthy, V.S.N.; Channappayya, S. Siamese cross domain tracker design for seamless tracking of targets in RGB and thermal videos. IEEE Trans. Artif. Intell. 2023, 4, 161–172. [Google Scholar] [CrossRef]
  5. Li, K.; Cai, L.; He, G.; Gong, X. MATI: Multimodal Adaptive Tracking Integrator for Robust Visual Object Tracking. Sensors 2024, 24, 4911. [Google Scholar] [CrossRef] [PubMed]
  6. Tian, D.; Yan, X.; Wang, C.; Zhang, W. IV-YOLO: A Lightweight Dual-Branch Object Detection Network. Sensors 2024, 24, 6181. [Google Scholar] [CrossRef] [PubMed]
  7. Ariffin, S.; Jamil, N.; Rahman, P. Can thermal and visible image fusion improves ear recognition? In Proceedings of the 2017 8th International Conference on Information Technology (ICIT), Amman, Jordan, 17–18 May 2017; pp. 780–784. [Google Scholar]
  8. Liu, Y.; Wu, Z.; Han, X.; Sun, Q.; Zhao, J.; Liu, J. Infrared and Visible Image Fusion Based on Visual Saliency Map and Image Contrast Enhancement. Sensors 2022, 22, 6390. [Google Scholar] [CrossRef]
  9. Chen, J.; Li, X.; Luo, L.; Mei, X.; Ma, J. Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 2020, 508, 64–78. [Google Scholar] [CrossRef]
  10. Cvejic, N.; Bull, D.; Canagarajah, N. Region-based multimodal image fusion using ICA bases. IEEE Sens. J. 2007, 7, 743–751. [Google Scholar] [CrossRef]
  11. Li, Y.; Liu, G.; Bavirisetti, D.; Gu, X.; Zhou, X. Infrared-visible image fusion method based on sparse and prior joint saliency detection and LatLRR-FPDE. Digit. Signal Process. 2023, 134, 103910. [Google Scholar] [CrossRef]
  12. Ma, J.; Chen, C.; Li, C.; Huang, J. Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fusion 2016, 31, 100–109. [Google Scholar] [CrossRef]
  13. Ma, J.; Zhou, Z.; Wang, B.; Zong, H. Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 2017, 82, 8–17. [Google Scholar] [CrossRef]
  14. Zhao, Z.; Xu, S.; Zhang, C.; Liu, J.; Zhang, J. Bayesian fusion for infrared and visible images. Signal Process. 2020, 177, 107734. [Google Scholar] [CrossRef]
  15. Yin, R.; Yang, B.; Huang, Z.; Zhang, X. DSA-Net: Infrared and Visible Image Fusion via Dual-Stream Asymmetric Network. Sensors 2023, 23, 7079. [Google Scholar] [CrossRef] [PubMed]
  16. Zhang, H.; Xu, H.; Xiao, Y.; Guo, X.; Ma, J. Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12797–12804. [Google Scholar] [CrossRef]
  17. Li, H.; Wu, X.-J.; Kittler, J. RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf. Fusion 2021, 73, 72–86. [Google Scholar] [CrossRef]
  18. Wang, J.; Jiang, M.; Kong, J. MDAN: Multilevel dual-branch attention network for infrared and visible image fusion. Opt. Lasers Eng. 2024, 176, 108042. [Google Scholar] [CrossRef]
  19. Li, K.; Liu, G.; Gu, X.; Tang, H.; Xiong, J.; Qian, Y. DANT-GAN: A dual attention-based of nested training network for infrared and visible image fusion. Digit. Signal Process. 2024, 145, 104316. [Google Scholar] [CrossRef]
  20. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  21. Yan, C.; Meng, L.; Li, L.; Zhang, J.; Wang, Z.; Yin, J.; Zhang, J.; Sun, Y.; Zheng, B. Age-invariant face recognition by multi-feature fusion and decomposition with self-attention. ACM Trans. Multimed. Comput. Commun. Appl. 2022, 18, 1–18. [Google Scholar] [CrossRef]
  22. Liu, Y.; Chen, X.; Peng, H.; Wang, Z. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  23. Liu, Y.; Chen, X.; Cheng, J.; Peng, H.; Wang, Z. Infrared and visible image fusion with convolutional neural networks. Int. J. Wavelets Multiresolut. Inf. Process. 2018, 16, 1850018. [Google Scholar] [CrossRef]
  24. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef] [PubMed]
  25. Guo, C.; Fan, D.; Jiang, Z.; Zhang, D. MDFN: Mask deep fusion network for visible and infrared image fusion without reference ground-truth. Expert Syst. Appl. 2023, 211, 118631. [Google Scholar] [CrossRef]
  26. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  27. Li, H.; Wu, X.-J. DenseFuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 2019, 28, 2614–2623. [Google Scholar] [CrossRef] [PubMed]
  28. Li, H.; Wu, X.-J.; Durrani, T. NestFuse: An infrared and visible image fusion architecture based on nest connection and spatial/channel attention models. IEEE Trans. Instrum. Meas. 2020, 69, 9645–9656. [Google Scholar] [CrossRef]
  29. Wang, Z.; Wu, Y.; Wang, J.; Xu, J.; Shao, W. Res2Fusion: Infrared and visible image fusion based on dense Res2net and double nonlocal attention models. IEEE Trans. Instrum. Meas. 2022, 71, 5005012. [Google Scholar] [CrossRef]
  30. Ma, J.; Xu, H.; Jiang, J.; Mei, X.; Zhang, X. DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 2020, 29, 4980–4995. [Google Scholar] [CrossRef] [PubMed]
  31. Ma, J.; Zhang, H.; Shao, Z.; Liang, P.; Xu, H. GANMcC: A generative adversarial network with multiclassification constraints for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 2021, 70, 5005014. [Google Scholar] [CrossRef]
  32. Liu, J.; Fan, X.; Huang, Z.; Wu, G.; Liu, R.; Zhong, W.; Luo, Z. Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5792–5801. [Google Scholar] [CrossRef]
  33. Huang, S.; Song, Z.; Yang, Y.; Wan, W.; Kong, X. MAGAN: Multiattention generative adversarial network for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 2023, 72, 5016614. [Google Scholar] [CrossRef]
  34. Gao, S.; Cheng, M.; Zhao, K.; Zhang, X.; Yang, M.; Torr, P. Res2Net: A new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 652–662. [Google Scholar] [CrossRef]
  35. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  36. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  37. Vs, V.; Valanarasu, J.M.J.; Oza, P.; Patel, V.M. Image fusion transformer. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 3566–3570. [Google Scholar]
  38. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  39. Tang, W.; He, F.; Liu, Y.; Duan, Y.; Si, T. DATFuse: Infrared and visible image fusion via dual attention transformer. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3159–3172. [Google Scholar] [CrossRef]
  40. Chen, J.; Ding, J.; Yu, Y.; Gong, W. THFuse: An infrared and visible image fusion network using transformer and hybrid feature extractor. Neurocomputing 2023, 527, 71–82. [Google Scholar] [CrossRef]
  41. Li, W.; Zhang, Y.; Wang, G.; Huang, Y.; Li, R. DFENet: A dual-branch feature enhanced network integrating transformers and convolutional feature learning for multimodal medical image fusion. Biomed. Signal Process. Control 2023, 80, 104402. [Google Scholar] [CrossRef]
  42. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. arXiv 2021, arXiv:2105.15203. [Google Scholar]
  43. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  44. Hou, R.; Zhou, D.; Nie, R.; Liu, D.; Xiong, L.; Guo, Y. VIF-Net: An Unsupervised Framework for Infrared and Visible Image Fusion. IEEE Trans. Comput. Imaging 2020, 6, 640–651. [Google Scholar] [CrossRef]
  45. Johnson, J.; Alahi, A.; Li, F.-F. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 694–711. [Google Scholar] [CrossRef]
  46. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  47. Toet, A. TNO Image Fusion Dataset. 2014. Available online: https://figshare.com/articles/dataset/TNO_Image_Fusion_Dataset/1008029 (accessed on 10 June 2024).
  48. Li, H.; Xu, T.; Wu, X.-J.; Lu, J.; Kittler, J. LRRNet: A novel representation learning guided fusion network for infrared and visible images. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 11040–11052. [Google Scholar] [CrossRef]
  49. Tang, W.; He, F.; Liu, Y. YDTR: Infrared and visible image fusion via Y-shape dynamic transformer. IEEE Trans. Multimed. 2023, 25, 5413–5428. [Google Scholar] [CrossRef]
  50. Park, S.; Vien, A.G.; Lee, C. Cross-modal transformers for infrared and visible image fusion. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 770–785. [Google Scholar] [CrossRef]
  51. Rao, Y.-J. In-fibre Bragg grating sensors. Meas. Sci. Technol. 1997, 8, 355. [Google Scholar] [CrossRef]
  52. Han, Y.; Cai, Y.; Cao, Y.; Xu, X. A new image fusion performance metric based on visual information fidelity. Inf. Fusion 2013, 14, 127–135. [Google Scholar] [CrossRef]
  53. Roberts, J.W.; Van Aardt, J.A.; Ahmed, F.B. Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2008, 2, 023522. [Google Scholar] [CrossRef]
  54. Xydeas, C.S.; Petrović, V. Objective image fusion performance measure. Electron. Lett. 2000, 36, 308–309. [Google Scholar] [CrossRef]
  55. Eskicioglu, A.M.; Fisher, P.S. Image quality measures and their performance. IEEE Trans. Commun. 1995, 43, 2959–2965. [Google Scholar] [CrossRef]
  56. Kumar, B.K.S. Multifocus and multispectral image fusion based on pixel significance using discrete cosine harmonic wavelet transform. Signal, Image Video Process. 2013, 7, 1125–1143. [Google Scholar] [CrossRef]
  57. Ma, K.; Zeng, K.; Wang, Z. Perceptual quality assessment for multi-exposure image fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef] [PubMed]
  58. Haghighat, M.; Razian, M.A. Fast-FMI: Non-reference image fusion metric. In Proceedings of the 2014 IEEE 8th International Conference on Application of Information and Communication Technologies (AICT), Astana, Kazakhstan, 15–17 October 2014; pp. 1–3. [Google Scholar] [CrossRef]
  59. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
Figure 1. Architecture of Res2Net.The feature information of different depths is obtained by a multiscale residual connection module.
Figure 1. Architecture of Res2Net.The feature information of different depths is obtained by a multiscale residual connection module.
Sensors 25 00791 g001
Figure 2. Overview of the proposed RMTFuse.
Figure 2. Overview of the proposed RMTFuse.
Sensors 25 00791 g002
Figure 3. Detailed architecture of Transformer block.
Figure 3. Detailed architecture of Transformer block.
Sensors 25 00791 g003
Figure 4. Detailed architecture of GFIM.
Figure 4. Detailed architecture of GFIM.
Sensors 25 00791 g004
Figure 5. Representative results of ablation experiment on different structures.
Figure 5. Representative results of ablation experiment on different structures.
Sensors 25 00791 g005
Figure 6. Subjective results of RMTFuse with 9 different methods on Kaptien_1123.
Figure 6. Subjective results of RMTFuse with 9 different methods on Kaptien_1123.
Sensors 25 00791 g006
Figure 7. Subjective results of RMTFuse with 9 different methods on Sandpath_18.
Figure 7. Subjective results of RMTFuse with 9 different methods on Sandpath_18.
Sensors 25 00791 g007
Figure 8. Subjective results of RMTFuse with 9 different methods on Marne_04.
Figure 8. Subjective results of RMTFuse with 9 different methods on Marne_04.
Sensors 25 00791 g008
Figure 9. Subjective results of RMTFuse with 9 different methods on FLIR_08999.
Figure 9. Subjective results of RMTFuse with 9 different methods on FLIR_08999.
Sensors 25 00791 g009
Figure 10. Subjective results of RMTFuse with 9 different methods on FLIR_07732.
Figure 10. Subjective results of RMTFuse with 9 different methods on FLIR_07732.
Sensors 25 00791 g010
Figure 11. Subjective results of RMTFuse with 9 different methods on FLIR_08202.
Figure 11. Subjective results of RMTFuse with 9 different methods on FLIR_08202.
Sensors 25 00791 g011
Figure 12. Object detection results on scene 00008N.
Figure 12. Object detection results on scene 00008N.
Sensors 25 00791 g012
Figure 13. Object detection results on scene 00315D.
Figure 13. Object detection results on scene 00315D.
Sensors 25 00791 g013
Table 1. List of some abbreviations.
Table 1. List of some abbreviations.
AbbreviationDescription
SFEMShallow feature extraction module
LFE-RNLocal feature extraction module based on Res2Net
GFE-MTGlobal feature extraction module based on multiscale Transformer
GFIMGlobal feature integration module
MSAMulti-head self-attention
MLPMulti-layer perceptron
Table 2. Detailed parameter design of LFE-RN.
Table 2. Detailed parameter design of LFE-RN.
LayerOutput SizeInput ChannelOutput ChannelKerner SizeStrideActivation
RB1 H × W 1616---------
RB2 H × W 3216---------
C o n v 1 H × W 483211LReLU
Table 3. Detailed parameter design of Transformer module.
Table 3. Detailed parameter design of Transformer module.
Output SizeLayerO-CKSPE-DN-B
TB1 H 4 × W 4 Overlapped Patch
Merging
32743-------
Transformer Encoder------------323
TB2 H 8 × W 8 Overlapped Patch Merging64321------
Transformer Encoder------------644
TB3 H 16 × W 16 Overlapped Patch Merging128321------
Transformer Encoder------------1286
Table 4. Objective comparison results of different structures on TNO dataset.
Table 4. Objective comparison results of different structures on TNO dataset.
MethodSDVIFENQabfSFNabfMS-SSIMFMIpixel
w/o RN39.31400.75107.07830.548611.38970.02770.93400.9045
w/o GFE-MT34.39670.64186.83030.494811.22540.03110.92910.9029
w/o MT38.36600.70937.00150.533811.23530.02590.92930.9040
RMTFuse40.51530.77137.09490.562111.92820.02550.93420.9085
Table 5. Objective comparison results on TNO dataset.
Table 5. Objective comparison results on TNO dataset.
MethodSDVIFENQabfSFNabfMS-SSIMFMIpixel
BF29.39300.67046.58790.44228.24450.09930.86770.9080
FusionGAN29.10100.41836.48500.22456.26540.07720.72880.8858
DenseFuse25.29510.58696.44950.35216.88190.08300.87390.9048
GANMcC31.66950.52466.69220.27706.30300.06870.85570.8964
LRRNet39.26910.56046.97190.36269.63240.05620.85140.8921
YDTR26.91680.61306.46260.39377.92140.05860.85050.8988
SwinFusion39.43480.75926.93210.532111.38320.03590.89310.9059
DATFuse28.02750.69296.51870.49669.88250.04350.80540.8740
CMTFusion36.08100.67556.97690.489010.61600.05270.92330.9037
RMTFuse40.51530.77137.09490.562111.92820.02550.93420.9085
Table 6. Objective comparison results on RoadScene dataset.
Table 6. Objective comparison results on RoadScene dataset.
MethodSDVIFENQabfSFNabfMS-SSIMFMIpixel
BF30.42170.59096.7010 0.33487.79280.18410.77950.8680
FusionGAN38.02790.38067.04120.25628.04540.14580.75620.8533
DenseFuse31.72080.57916.81330.39288.30050.1470 0.85880.8641
GANMcC41.74630.51357.17430.34738.62120.12910.84950.8560
LRRNet41.90760.48557.1070 0.340311.87980.10540.79760.8511
YDTR35.81280.58336.88620.448310.27540.09920.86220.8616
SwinFusion44.67160.62886.98860.467611.77870.07420.84990.8593
DATFuse32.34010.5970 6.72390.487111.46610.08340.76110.8546
CMTFusion45.39940.62267.33550.443311.69160.09830.89190.8584
RMTFuse47.29990.65497.30020.561013.85690.03590.93390.8654
Table 7. The average running time of different fusion methods on TNO and RoadScene datasets (unit: seconds).
Table 7. The average running time of different fusion methods on TNO and RoadScene datasets (unit: seconds).
MethodTNORoadScene
BF1.57340.2803
FusionGAN0.26860.2924
DenseFuse0.10250.0521
GANMcC0.46000.5215
LRRNet0.17220.0918
YDTR0.39610.1085
SwinFusion1.81240.8726
DATFuse0.06090.0377
CMTFusion0.22080.1294
RMTFuse0.28020.1072
Table 8. Object detection performance for infrared, visible, and different fused images.
Table 8. Object detection performance for infrared, visible, and different fused images.
MethodmAP@0.5mAP@0.9
PersonCarAvg.PersonCarAvg.
IR0.73150.25470.49310.26570.24980.2578
VIS0.53620.73300.63460.21030.50490.3576
BF0.62270.67420.64850.26160.43810.3499
FusionGAN0.51730.53350.52540.11660.34260.2296
DenseFuse0.72500.64370.68440.27120.42790.3496
GANMcC0.71140.68090.69620.21700.41420.3156
LRRNet0.68840.73950.71400.22420.45810.3412
YDTR0.73500.70700.72100.26680.45930.3631
SwinFusion0.71950.73880.72920.26110.42670.3439
DATFuse0.76550.72530.74540.27770.38530.3315
CMTFusion0.69620.70120.69870.26650.40860.3376
RMTFuse0.73650.75490.74570.26850.46980.3692
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tan, B.; Yang, B. An Infrared and Visible Image Fusion Network Based on Res2Net and Multiscale Transformer. Sensors 2025, 25, 791. https://doi.org/10.3390/s25030791

AMA Style

Tan B, Yang B. An Infrared and Visible Image Fusion Network Based on Res2Net and Multiscale Transformer. Sensors. 2025; 25(3):791. https://doi.org/10.3390/s25030791

Chicago/Turabian Style

Tan, Binxi, and Bin Yang. 2025. "An Infrared and Visible Image Fusion Network Based on Res2Net and Multiscale Transformer" Sensors 25, no. 3: 791. https://doi.org/10.3390/s25030791

APA Style

Tan, B., & Yang, B. (2025). An Infrared and Visible Image Fusion Network Based on Res2Net and Multiscale Transformer. Sensors, 25(3), 791. https://doi.org/10.3390/s25030791

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop