Next Article in Journal
Object-Oriented Crop Classification Using Time Series Sentinel Images from Google Earth Engine
Next Article in Special Issue
Low-Cost Object Detection Models for Traffic Control Devices through Domain Adaption of Geographical Regions
Previous Article in Journal
Automatic Road Inventory Using a Low-Cost Mobile Mapping System and Based on a Semantic Segmentation Deep Learning Model
Previous Article in Special Issue
Infrared and Visible Image Fusion Method Based on a Principal Component Analysis Network and Image Pyramid
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IFormerFusion: Cross-Domain Frequency Information Learning for Infrared and Visible Image Fusion Based on the Inception Transformer

Department of Weapon Engineering, Naval University of Engineering, Wuhan 430030, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(5), 1352; https://doi.org/10.3390/rs15051352
Submission received: 1 January 2023 / Revised: 20 February 2023 / Accepted: 22 February 2023 / Published: 28 February 2023

Abstract

:
The current deep learning-based image fusion methods can not sufficiently learn the features of images in a wide frequency range. Therefore, we proposed IFormerFusion, which is based on the Inception Transformer and cross-domain frequency fusion. To learn features from high- and low-frequency information, we designed the IFormer mixer, which splits the input features through the channel dimension and feeds them into parallel paths for high- and low-frequency mixers to achieve linear computational complexity. The high-frequency mixer adopts a convolution and a max-pooling path, while the low-frequency mixer adopts a criss-cross attention path. Considering that the high-frequency information relates to the texture detail, we designed a cross-domain frequency fusion strategy, which trades high-frequency information between the source images. This structure can sufficiently integrate complementary features and strengthen the capability of texture retaining. Experiments on the TNO, OSU, and Road Scene datasets demonstrate that IFormerFusion outperforms other methods in object and subject evaluations.

1. Introduction

The image fusion technique aims to fuse images captured by different sensors to generate a fused image with better human visual effects and scene representation. As a result of thermal-based imaging, the infrared sensor can work in low-illuminance environments and all weather conditions and generate infrared images to emphasize prominent targets. The infrared image contains rich, low-frequency information, but it lacks texture detail. The visible image contains a rich texture and detailed spatial features, but it is susceptible to influence by illuminance and weather conditions. Due to the good complementarity of infrared and visible images, the fused image can provide high-quality results for person re-identification [1], object tracking [2], remote sensing [3,4], and salient object detection [5].
The image fusion methods can be categorized into traditional methods and deep learning-based methods. Traditional methods usually consist of three parts, a manual representation model to extract features, fusion strategies to the feature maps or weight maps, and an inverse feature extractor to reconstruct images. The traditional methods can be categorized into spatial domain [6] and frequency domain methods [7,8,9]. The spatial domain methods are usually simple to infer, but have poor effects on edge preserving. The frequency domain methods usually adopt domain transform operation and fuse procedures, which are more complex [10]. Although traditional fusion methods can achieve good fusion results, the manual design of transformation algorithms and fusion strategies can result in higher computational complexity, limiting the fusion performance. Conversely, deep learning-based models can extract features and generate high-performance fusion images without the need for a complex manual design. Deep learning-based methods can be categorized into four types: convolution neural network(CNN)-based fusion methods [11,12], auto-encoder-based fusion methods [13,14], generative adversarial network(GAN)-based fusion methods [15,16], and transformer-based fusion methods [17,18]. Although the above methods can generate fused images with a good performance, some issues still exist in order for them to be improved. CNN-based, auto-encoder-based, and GAN-based fusion methods generally use convolutional layers to extract the features. Thus, these methods cannot establish long-range dependencies with the limitation of the perceptive field of convolutional layers. The vision transformer develops attention mechanisms to build a long-range relationship between the image patches. However, there are still some drawbacks. Firstly, the existing methods fail to learn information in a wide frequency range, which is important for infrared and visible image fusion tasks. Secondly, the existing methods a lack consideration for the relationship between the frequency information of the source images.
Above all, we proposed IFormerFusion based on the Inception Transformer and cross-domain frequency fusion. On the one hand, we designed the IFormer mixer based on the Inception Transformer, which adopts convolution/max-pooling paths to process high-frequency information and a criss-cross attention path to process low-frequency information. The IFormer mixer was designed to capture high-frequency information to Transformers structure and learn information in a wide frequency range. On the other hand, cross-domain frequency fusion can trade high-frequency information between the source images to guide the model to retain more high-frequency information. IFormerFusion has four parts: feature extraction, cross-domain frequency fusion, feature reconstruction, and fused image reconstruction. The feature extractor can effectively extract comprehensive features in a wide frequency range. The cross-domain frequency fusion can learn features in a wide frequency range and trade high-frequency information between the source images to strengthen the texture-retaining capability of the model. The fusion results are concatenated and fed to the Inception Transformer-based feature reconstruction part to reconstruct deep features. Finally, a CNN-based fused image reconstruction part is utilized to reconstruct the images. Above all, the main contributions of this work can be summarized as follows:
  • We propose an infrared and visible image fusion method, IFormerFusion, which can efficiently learn features from source images in a wide frequency range. IFormerfusion can sufficiently retain texture details and maintain the structure of the source images.
  • We designed the IFormer mixer, which consists of convolution/max-pooling paths and a criss-cross attention path. The convolution/max-pooling can learn high-frequency information, while the criss-cross attention path can learn low-frequency information.
  • The cross-domain frequency fusion can trade high-frequency information between the source images to sufficiently learn comprehensive features and strengthen the capability to retain texture.
  • Experiments conducted using the TNO, OSU, and Road Scene datasets show that IFormerFusion obtains better results in both visual quality evaluation and quantitatively evaluation.

2. Related Works

2.1. Vision Transformer

The transformer, which was designed for natural language processing (NLP), has achieved notable success in a broad range of computer vision tasks, such as image classification [19,20], semantic segmentation [21,22], and object detection [23,24,25,26,27,28]. In 2020, Dosovitskiy et al. proposed the vision transformer, which splits an image into a sequence of flattened 16 × 16 patches and regards them as words in pieces of text [29]. Then, the vision transformer will embed the patches linearly and adopt a self-attention mechanism, which also brings an increased computational complexity quadratic to the image size, which is the bottleneck of the application in pixel-level tasks (such as image fusion) that require dense prediction. In 2021, Liu et al. proposed the Swin Transformer, in which self-attention computation is limited in the local window and shifted window partitioning in successive blocks [21]. Thus, the Swin Transformer has a linear complexity corresponding to the image size [21]. In 2022, Si et al. noticed that it is incompetent at learning information in a wide frequency range and proposed the Inception Transformer, which can sufficiently learn comprehensive features with both high- and low-frequency information [30]. The input images are split into three partitions and fed into the Inception mixer to learn features in a wide frequency range. Through the Inception mixer, the Inception Transformer has greater efficiency through a channel splitting mechanism to adopt parallel convolution/max-pooling paths and self-attention path to learn information within a wide frequency range [30].

2.2. Deep Learning-Based Image Fusion Methods

In the beginning, deep learning models were only utilized to extract features and generate weight maps for fusion [11]. The majority of fusion methods retain the structure of the traditional fusion framework. With more researchers designing networks and loss functions, CNN-based methods gradually differ from traditional frameworks. In 2021, Long et al. designed a parallel, aggregated, residual dense block consisting of a dense block path and a residual dense block path and the proposed RXDNFuse [12]. In 2019, Li et al. introduced an auto-encoder-based fusion network, DenseFuse [13]. The auto-encoder-based fusion methods can benefit from the high interpretability of the traditional fusion methods and the feature extraction ability of the CNN. Thus, many researchers focus on improving each part of the auto-encoder structure. In 2020, to fuse the extracted feature more effectively and reconstruct the feature maps, Li et al. proposed NestFuse, which adopts a nest-connected decoder network and the attention-based fusion strategy [31]. In 2021, Xu et al. proposed a learnable fusion strategy for the first time, which could quantitatively measure the classification significance of feature maps by using the back-propagation integral gradient of the classification results [32]. In 2022, Wang et al. proposed Res2Fusion, which adopts a multi-scale feature extraction strategy without down-sampling, and established long-distance feature dependency through nonlocal attention mechanisms [33].
Some researchers utilized the generative adversarial network(GAN) to fuse images in an implicit manner. In 2019, Ma et al. first introduced GAN to infrared and visible image fusion and proposed FusionGAN [15]. Subsequently, they proposed other GAN-based fusion methods (DDcGAN [34] in 2020 and GANMcC [16] in 2021). DDcGAN is utilized to fuse multi-resolution images. GANMcC is utilized to generate fused images with a balance between the gradient and intensity of the source images.
To further improve the fusion effects, some researchers replaced the CNN layers with transformer structures. In 2022, Liu et al. proposed MFST, which adopts a self-adaptive transformer fusion strategy [14]. In 2022, Rao et al. proposed TGFuse, which adopts a transformer-based generator [35]. In 2022, Wang et al. proposed SwinFuse, which adopts residual Swin Transformer blocks as the encoder network. Based on the Swin Transformer, in 2022, Ma et al. proposed a cross-domain long-range fusion method, which includes inter- and Swin Transformer-based cross-domain modules to extract and fuse deep features [18].
Different from the mentioned methods, we propose IFormerFusion, which sufficiently learns the features of the images in a wide frequency range through the IFormer mixer that consists of parallel convolution, max-pooling, and attention paths in a wide frequency range. Moreover, we designed a cross-domain frequency fusion strategy to sufficiently integrate complementary features and strengthen the ability to retain texture.

3. Methodology

3.1. Overall Framework

Let I1 and I2   H × W × C represent the input images of IFormerFusion, while C is the channel number, and H and W are the image sizes. As shown in Figure 1, IFormerFusion is constructed with four parts: feature extraction, cross-domain frequency fusion, feature reconstruction, and fused image reconstruction.
Feature Extraction: Firstly, the input images or features are first embedded by convolutional embedding layers, which extend the channel dimension of the input to the required number. In this experiment, the required channel number is 60. The embedding result can be expressed as:
ϕ = E m b e d ( I ) ,
We then design feature extraction blocks called base blocks. In each base block, the input feature will split into three partitions including two high-frequency partitions (H1 and H2) and a low-frequency partition (L) through the channel dimension and feed them into the IFormer mixer. Next, a feed-forward network (FFN) is deployed to refine the result of the IFormer mixer. Layer normalization (LN) is adopted before both the FFN and the split operation in the IFormer mixer, which is inferred specifically in the next section. Moreover, the residual connection is adopted in the mixer. The inference can be expressed as:
ϕ h 1 , ϕ h 2 ,   ϕ l = S p l i t ( N o r m ( ϕ ) ) ,
ϕ M = ϕ + M i x e r ( ϕ h 1 , ϕ h 2 ,   ϕ l ) ,
ϕ F = ϕ M + F F N ( N o r m ( ϕ M ) ) ,
where ϕ h 1 and ϕ h 2 represent the high-frequency information, ϕ l represents the low-frequency information, ϕ M represents the residual mixed results, and ϕ F represents the extract features. Feature extraction consists of N1 base blocks. In this experiment, N1 is 3.
Cross-Domain Frequency Fusion: The cross-domain frequency fusion part is designed to learn features in a wide frequency range and fuse cross-domain frequency information. The cross-domain frequency fusion part consists of N2 cross blocks. The base block is the same as the base block in feature extraction. However, in the cross block, the high-frequency information of two input features is exchanged. The embedding results of input features I 1 and I 2 can be expressed as:
ϕ 1 = E m b e d ( I 1 ) ,
ϕ 2 = E m b e d ( I 2 ) ,
The split partitions of ϕ i are ϕ h 1 i , ϕ h 2 i , and ϕ l i (i = 1, 2), which can be expressed as:
ϕ h 1 1 , ϕ h 2 1 , ϕ l 1 = S p l i t ( N o r m ( ϕ 1 ) ) ,
ϕ h 1 2 , ϕ h 2 2 , ϕ l 2 = S p l i t ( N o r m ( ϕ 2 ) ) ,
The mixed results ϕ M 1 and ϕ M 2 can be expressed as:
ϕ M 1 = ϕ 1 + M i x e r ( ϕ h 1 2 , ϕ h 2 2 , ϕ l 1 ) ,
ϕ M 2 = ϕ 2 + M i x e r ( ϕ h 1 1 , ϕ h 2 1 , ϕ l 2 ) ,
The extract features ϕ F 1 and ϕ F 2 can be expressed as:
ϕ F 1 = ϕ M 1 + F F N ( N o r m ( ϕ M 1 ) ) ,
ϕ F 2 = ϕ M 2 + F F N ( N o r m ( ϕ M 2 ) ) ,
In this experiment, N2 is 6. The results ϕ F 1 and ϕ F 2 are concatenated and fed to the feature reconstruction part. The concatenated result ϕ F can be expressed as:
ϕ F = C o n c a t ( ϕ F 1 , ϕ F 2 ) ,
Feature Reconstruction: The results of the Cross-Domain Frequency Fusion part are fed into the Feature Reconstruction part, which consists of N3 reconstruction blocks. The reconstruction block is half of the base block, which has a single path to process the concatenated result. In this experiment, N3 is 3.
Fused Image Reconstruction: Finally, a simple CNN-based part to reconstruct fused images is utilized to reconstruct fused images. The fused image reconstruction part consists of two convolutional layers.

3.2. IFormer Mixer

The architecture of the IFormer mixer is shown in Figure 2. After splitting the input feature into three partitions through the channel dimension, high-frequency and low-frequency mixers are adopted to learn the features in a wide frequency range. The high-frequency mixer has a max-pooling (MaxPool) path, which consists of a max pooling subsequently a linear layer [36], and a parallel convolution path, which consists of a linear subsequently a depthwise convolution (DwConv) layer [37]. The low-frequency mixer has a criss-cross attention path, which consists of average pooling, criss-cross attention (CC-Atten) [38], and up-sampling (UpSample). The detailed inference is as follows.
With the input X  H × W × C , X can be divided into two parts through channel dimension, the high-frequency part, X h H × W × C h , and the low-frequency part, X l   H × W × C l , where Ch + Cl = C. Then, X h and X l are assigned to the high-frequency mixer and low-frequency mixer, respectively. The details of the high- and low-frequency mixers follow.
High-frequency mixer: Considering the sharpness sensitivity of the maximum filter and the detail perception of the convolution operation, we propose two high-frequency paths to take advantage of the sharpness sensitivity of max-pooling and the detail perception capability of the convolution layers to learn high-frequency information. Firstly, the input X h is divided into X h 1 H × W × C h / 2 and X h 2   H × W × C h / 2 . X h 1 is fed into the max pooling path. X h 2 is fed into the parallel convolution path. The outputs of the high-frequency mixer Y h 1 and Y h 2 can be expressed as:
Y h 1 = L i n e a r ( M a x P o o l ( X h 1 ) ) ,
Y h 2 = D w C o n v ( L i n e a r ( X h 2 ) ) ,
Low-frequency mixer: Considering the strong capability of the attention mechanism for learning global representation, we use criss-cross attention [38] to establish long-range dependency to learn low-frequency information. However, in image fusion tasks, dense prediction can bring great computation complexity with the large resolution of feature maps in the low-frequency mixer. Therefore, we utilize an average-pooling operation to reduce the scale of X l before the criss-cross attention operation and an up-sample layer to restore the original scale. In this experiment, the kernel size and stride for the average pooling are 4, and the size of up-sample layers is also 4. This branch can be defined as
Y l = U p s a m p l e ( C C ( A v g P o o l ( X l ) ) ) ,
where Y l is the output of the low-frequency mixer, and CC represents criss-cross attention.
Finally, the outputs Y h 1 ,   Y h 2 ,   a n d   Y l are concatenated through the channel dimension and fused to obtain Y c :
Y c = F u s i o n ( C o n c a t ( Y h 1 ,   Y h 2 ,   Y l ) ) ,

3.3. Loss Function

Three loss functions are adopted for the IFormerFusion in the training phase, which is explained as follows.
The structural similarity (SSIM) loss L S S I M can be expressed as:
L S S I M = α ( 1 S S I M ( I F , I 1 ) + β ( 1 S S I M ( I F , I 2 ) ,
where α = β = 0.5 in this experiment and SSIM(⋅) represents the philosophy of structural similarity [39].
Inspired by IFCNN [40] and SwinFusion [18], we deploy the intensity loss to supervise the model to capture potential intensity information. The intensity loss L I n t can be expressed as:
L I n t = 1 H W I F m a x ( I 1 , I 2 ) 1 ,
where ||⋅||1 represents the l1-norm, and max(⋅) represents the chosen max value. H and W represent the image sizes.
We deploy texture loss [18] to evaluate the texture details, which can be extracted by the maximum function. Thus, the texture loss L T e x t can be expressed as:
L T e x t = 1 H W I F m a x ( | I 1 | , | I 2 | ) 1 ,
where ∇ represents the Sobel gradient operator, |⋅| represents for the absolute operation, m a x ( ) represents to choose max value, and ||⋅||1 represents the l1-norm.
Finally, the total loss function L t o t a l is a weighted sum of all loss functions, which can be expressed as:
L t o t a l = λ 1 L S S I M + λ 2 L I n t + λ 3 L T e x t ,
where λ 1 , λ 2 , and λ 3 are weighted to balance each loss.

4. Experimental Results and Analysis

In this section, IFormerFusion is compared with eight advanced fusion methods with images selected from three public datasets. Subsequently, we conducted quantitative and qualitative comparisons with nine advanced methods: DenseFuse [13], GANMcC [16], IFCNN [40], NestFuse [31], Res2Fusion [33], RFN-Nest [41], SwinFuse [42], SwinFusion [18], and U2Fusion [43]. Finally, a computational complexity analysis is conducted.

4.1. Experiment Setup

The MSRS dataset [44] is selected for training. A total of nine deep learning methods are selected to compare the methods, i.e., DenseFuse, GANMcC IFCNN, NestFuse, Res2Fusion, RFN-Nest, SwinFuse, SwinFusion, and U2Fusion. The TNO image fusion dataset [45], OSU Color-Thermal dataset [46], and RoadScene dataset [43] are selected to evaluate the above methods.
All the experiments in this paper are conducted on Intel(R) Xeon(R) Silver 4210 CPU and NVIDIA GeForce RTX 3090 GPU. The PyTorch program is used. The batch size is eight. The images are randomly cropped to 128 × 128 patches and normalized to [0, 1]. The Adam optimizer is used. The model is trained for 200 epochs. The initial learning rate is 0.001 and it decays to half this in epochs 20, 40, 80, 120, and 180.

4.2. Evaluation Metrics

A total of six metrics are selected for evaluation, including mutual information (MI) [47], fast mutual information (FMI) [48], the peak signal-to-noise ratio (PSNR), the structural similarity index measurement (SSIM) [39], visual information fidelity (VIF) [49], and Qabf [50]. These metrics measure the performance of the fusion method from different aspects. Suppose the infrared image, the visible image, and the fused image are I, V, and F, respectively. Their detailed definitions are described as follows:
The mutual information metric is a quality index that measures the amount of information transferred from the source images to the fused image. mutual information is a fundamental concept in information theory and measures the dependence of two random variables. The definition of the mutual information metric can be expressed as:
M I = M I ( I , F ) + M I ( V , F ) ,
where M I ( I , F ) and M I ( V , F ) are the amounts of information transferred from the infrared images and visible images to the fused image, respectively. The MI between two random variables can be calculated by the Kullback–Leibler approach, which can be expressed as:
M I ( X , F ) = x , f P X , F ( x , f ) l o g P X , F ( x , f ) P X ( x ) P F ( f )
where P X , F ( x , f ) is the joint histogram of the source image x and the fused image F ; P X ( x ) and P F ( f ) are the marginal histograms of the source image X and the fused image F, respectively.
The fast mutual information metric calculates the regional mutual information between the corresponding windows in the fused image and the source images [48]. The mutual information I n f o ( I ,   F ) and I n f o ( V ,   F ) can be expressed as:
I n f o ( X ,   F ) = 2 n i = 1 n I n f o i ( X , F ) H i ( X ) + H i ( F )  
where Hi(X) and Hi(F) are the entropies of X and F, and the mutual information I n f o i ( X , F ) can be expressed as:
I n f o i ( X ,   F ) = H i ( X ) + H i ( F ) 2
The fast mutual information metric can be expressed as:
F M I ( I , V , F ) = 1 2 ( I n f o ( I , F ) + I n f o ( V , F ) )
The peak signal-to-noise ratio metric is the ratio of the maximum possible power of a signal to the destructive noise power, which affects its representation accuracy. The PSNR of the fused image and source images can be expressed as:
P S N R = 10 l o g ( m a x ( F ) M S E ( I , F ) ) + 10 l o g ( m a x ( F ) M S E ( V , F ) )
where max(F) is the maximum value in the fused image, and the M S E ( I , F ) and M S E ( V , F ) are the mean-squared errors, which can be expressed as:
M S E ( X , F ) = 1 M N   i = 0 M 1 j = 0 M 1 [ X ( i , j ) F ( i , j ) ] 2
where X ( i , j ) and F ( i , j ) are the value of X and F in row i and column j, and M and N are the weight and height of the images, respectively.
The structural similarity index measurement metric calculates the similarity of the fused images and source images in terms of luminance, contrast, and structure [39]. The SSIM metric can be expressed as:
S S I M ( X , F ) = ( 2 μ x μ f + C 1 ) ( 2 σ x f + C 2 ) ( μ x 2 + μ f 2 + C 1 ) ( σ x 2 + σ f 2 + C 2 )
where μ x and μ f are the average gradients of X and F; σ x and σ f are the standard deviation of X and F; σ x f is the correlation coefficient of X and F; C 1 ,   C 2 , and C 3 are ( 0.01 L ) 2 , ( 0.03 L ) 2 , and 1 2 ( 0.01 L ) 2 , respectively; L the dynamic range of the pixel values, which is 255.
The visual information fidelity metric calculates image distortions including additive noise, blurs, and changes [49]. VIF is derived from the quantification of two types of mutual information: the mutual information between the input and the output of the HVS channel (described via a stationary white Gaussian noise model) when no distortion channel is presented (i.e., reference mutual information) and the mutual information between the input of the distortion channel and the output of the HVS channel for the test image [51]. The visual information fidelity can be expressed as:
V I F ( X , F ) = k s u b b a n d s I ( C s , k , F s , k | R S , k ) k s u b b a n d s I ( C s , k , X s , k | R S , k )
where the subbands are a collection of specific sub-bands; I ( C s , k , F s , k | R S , k ) and I ( C s , k , F s , k | R S , k ) are the information extracted from specific sub-bands of the source image and fused image; C s , k is the random field from source images; F s , k and X s , k are the output of the HVS channel for the fused image and source images; R S , k is the model parameter of the sub-bands.
Thus, the visual information fidelity for fusion (VIFF) can be expressed as:
V I F F = 1 2 ( V I F ( V , F ) + V I F ( I , F ) )
The Qabf metric measures the similarity of the edge transferred from the source images to the fused image. Qabf can be expressed as:
Q a b f = i = 1 M j = 1 N ( Q A , F ( i , j ) w A ( i , j ) + Q B , F ( i , j ) w B ( i , j ) ) i = 1 M j = 1 N ( w A ( i , j ) + w B ( i , j ) )
where w X ( i , j ) is the weight matrix of source images and Q X , F ( i , j ) is the edge information transferred from the source image to the fused image. The Q X , F ( i , j ) can be expressed as:
Q X , F ( i , j ) = Q g X , F ( i , j ) Q a X , F ( i , j )
where Q g X , F ( i , j ) and Q a X , F ( i , j ) are the retaining edge intensity and direction in the pixel ( i , j ) , respectively.

4.3. Results on the TNO Dataset

Visual Quality Evaluation: Four pairs of images are selected to evaluate the visual effects of IFormerFusion and nine comparable methods, as shown in Figure 3. Some targets and details are highlighted with boxes to display the information that is worthy of attention. DenseFuse, NestFuse, RFN-Nest, and SwinFuse extract features through a well-designed encoder and adopt an additional attention-based or learnable fusion strategy. Their fused images have lower brightness and contrast values than those of the other methods do, which indicates a weaker ability to retain detailed texture information. In Figure 3a, the fused images of DenseFuse, NestFuse, and SwinFuse cannot display the text information of the pedestrian and the store billboard. In Figure 3b, the target on the building is hard to distinguish from the background. In Figure 3c,d, the tree branches lack details. The fused images of GANMcC retain the sharpness of the target in the infrared images. However, the background information, such as the plants and buildings, is fuzzy and lacks details. Res2Fusion offers better visual effects than the above methods do. In Figure 3a, the store billboard and pedestrian details are retained from the source images. Nevertheless, in Figure 3b, the target over the building is still not clear. In Figure 3c, the details of the tree branches are retained, but in Figure 3d, more specific details of the tree are lost. The fused images of IFCNN, SwinFusion, U2Fusion, and IFormerFusion generate fused images that can balance the gradient and intensity information. The targets are displayed with rich texture details and sharp edges. The details of background information, including the plants and buildings, are retained. However, IFormerFusion obtains the best visual effects and retains rich texture details and sharp edges.
Quantitative Evaluation: To quantitatively evaluate IFormerFusion and the comparable methods, six metrics are calculated using 21 pairs of images. The result is shown in Figure 4. The average value of each metric is shown in Table 1. IFormerFusion obtains the best values in MI, FMI, PSNR, VIFF, and Qabf, and the second-best value in SSIM. More specifically, the highest MI and FMI values show that the proposed method can transfer more feature and edge information from the source images to the fused image. The highest PSNR value shows that the proposed method causes the least information distortion during fusion. The highest VIFF value shows that the proposed method has more effective visual information. The highest Qabf value shows that the proposed method can obtain more visual information from the source images. The DenseFuse has the highest SSIM value, which indicates the advantage of structural information maintenance. However, DenseFuse is weaker in terms of the other metrics. Above all, the result indicates that IFormerFusion produces the best fusion results.

4.4. Results on the OSU Dataset

Visual Quality Evaluation: A pair of images are selected to evaluate the visual effects of IFormerFusion and nine comparable methods. The results are shown in Figure 5. In the visible image, the pedestrian in the shadow of the building is hard to distinguish, while they are clearly observable in the infrared image. In the fused images of DenseFuse, GANMcC, NestFuse, and RFN-Nest, the gradient of the pedestrian is close to the building. Moreover, the edge of the pedestrian is blurred by the background building in GANMcC and RFN-Nest. In the fused images of IFCNN, Res2Fusion, SwinFuse, SwinFusion, and IFormerFusion, the brightness of the pedestrian is similar to that of the infrared images, which indicates that the information in the infrared images is well retained. In the fused images of GANMcC and RFN-Nest, the sculpture on the lawn is blurred. The constrain of the lawn in the fused images of IFCNN, NestFuse, SwinFuse, and U2Fusion is discordant, which is more influenced by the infrared images. Thus, the texture detail of the lawn is lost. The lawn in the fused images of Res2Fusion, SwinFusion, and IFormerFusion has better visual effects. Above all, the proposed method, IFormerFusion, can retain texture detail and fuse with a balance of the gradient and intensity.
Quantitative Evaluation: To quantitatively evaluate IFormerFusion and compare the methods, six metrics are calculated using 40 pairs of images. The result is shown in Figure 6. The average value of each metric is shown in Table 2. IFormerFusion obtains the best results in terms of PSNR, VIFF, and Qabf, which indicates that the proposed method has the least information distortion and obtains the best visual effects during fusion. The proposed method lags behind the best method by a narrow margin in terms of MI, FMI, and SSIM. Thus, the result indicates that IFormerFusion can transfer a lot of information from the source images and produces the best results.

4.5. Results on the Road Scene Dataset

Visual Quality Evaluation: A pair of images is selected to evaluate the visual effects of IFormerFusion and nine comparable methods. The result is shown in Figure 7. In the fused images of DenseFuse, GANMcC, Res2Fusion, and RFN-Nest, the tree on the left is hard to distinguish from the background. On the right of the images, the texture of the tire rim is not rich, and the people standing by the area are fuzzy. In the fused image of NestFuse, though the tree branches are clear, artifacts exist in the car light, which indicates that NestFuse cannot balance the light information. In the fused images of IFCNN, SwinFuse, SwinFusion, U2Fusion, and IFormerFusion, the details of tree branches can be distinguished on the left. On the right, the tire rim is sharp, and the people in front of the car can be distinguished.
Quantitative Evaluation: To quantitatively evaluate IFormerFusion and compare methods, six metrics are calculated on 221 pairs of images from Road Scene datasets. The average value of each metric is shown in Table 3. The NestFuse obtains the best results in terms of MI and FMI. However, as shown in Figure 7, artifacts exist on the car, which will introduce extra information to the fused images to increase the MI and FMI values. IFormerFusion obtains the best values in terms of PSNR, VIFF, and Qabf. IFormerFusion obtains the second-best value for SSIM, behind that of DenseFuse. DenseFuse also obtains second-best result in terms of PNSR. However, DenseFuse produces unsatisfactory results in other metrics; in other words, DenseFuse can retain the structure and information from the source images, but has worse visual effects and less information. Above all, the result demonstrates that IFormerFusion produces the best fusion results.

4.6. Computation Efficiency

Moreover, we provide the computation efficiency of IFormerFusion and comparable methods, as shown in Table 4. DenseFuse, IFCNN, and NestFuse have a high computation efficiency because these methods consist of convolutional layers and simple fusion strategies. The RFN-Nest and SwinFuse one have a low computational efficiency because these methods utilize attention-based fusion strategies. Though GANMcC and U2Fusion also consist of convolutional layers, the computation is more complex. Thus, these methods have a lower computational efficiency. Res2Fusion and SwinFusion have the lowest computational efficiency because more complex attention mechanisms are used in these methods. However, the proposed IFormerFusion has a competitive computational efficiency and retains the linear computational complexity of the image size. All the methods are tested using a public code.

5. Conclusions

We propose IFormerFusion, a cross-domain frequency information learning infrared and visible image fusion network based on the Inception Transformer. The IFormer mixer based on the Inception Transformer consists of the high-frequency mixer, which consists of max pooling and convolution paths, and the low-frequency mixer, which contains a criss-cross attention path. The high-frequency mixer can take the advantage of convolution and max-pooling for capturing high-frequency information. The low-frequency mixer can establish long-range dependency, and the criss-cross attention can reduce the computational complexity. Thus, the IFormer mixer can learn information in a wide frequency range. Moreover, the high-frequency information is traded in the cross-domain frequency fusion part to achieve the sufficient integration of complementary features and strengthen the capability to retain texture. The proposed IFormerFusion can comprehensively learn features from high- and low-frequency information to retain texture details and maintain the structure.
We conducted experiments on TNO, OSU, and Road Scene datasets and compared them with nine advanced deep-learning methods using six metrics. The results demonstrate that IFormerFusion performs well at preserving the structure and retaining texture details. In addition, IFormerFusion presents the balanced intensity of the targets and background in the fused image. IFormerFusion also has a competitive computational efficiency.

Author Contributions

Conceptualization, Z.X.; methodology, Z.X.; software, Z.X.; validation, Z.X.; formal analysis, Z.X.; investigation, Q.H.; resources, H.H., Q.H. and X.Z.; data curation, H.H.; writing—original draft preparation, Z.X.; writing—review and editing, X.Z.; visualization, Z.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Test images from TNO and OSU datasets are available at https://github.com/xiongzhangzzz/IFormerFusion (accessed on 21 February 2023). The Road Scene dataset can download from https://github.com/StaRainJ/road-scene-infrared-visible-images (accessed on 3 December 2022). The MSRS dataset can download from https://github.com/Linfeng-Tang/MSRS (accessed on 3 December 2022). The code and model are publicly accessible at https://github.com/xiongzhangzzz/IFormerFusion (accessed on 21 February 2023).

Acknowledgments

Thanks to the author of the compared methods for providing public codes and the author of the MSRS, TNO, OSU and Road Scene datasets for providing public datasets.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Feng, Z.; Lai, J.; Xie, X. Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification. IEEE Trans. Image Process. 2020, 29, 579–590. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, X.; Ye, P.; Qiao, D.; Zhao, J.; Peng, S.; Xiao, G. Object Fusion Tracking Based on Visible and Infrared Images Using Fully Convolutional Siamese Networks. In Proceedings of the 2019 22th International Conference on Information Fusion (FUSION), Ottawa, ON, Canada, 2–5 July 2019; pp. 1–8. [Google Scholar]
  3. Ma, J.; Yu, W.; Chen, C.; Liang, P.; Guo, X.; Jiang, J. Pan-GAN: An Unsupervised Pan-Sharpening Method for Remote Sensing Image Fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  4. Rajah, P.; Odindi, J.; Mutanga, O. Feature Level Image Fusion of Optical Imagery and Synthetic Aperture Radar (SAR) for Invasive Alien Plant Species Detection and Mapping. Remote Sens. Appl. Soc. Environ. 2018, 10, 198–208. [Google Scholar] [CrossRef]
  5. Zhou, W.; Zhu, Y.; Lei, J.; Wan, J.; Yu, L. CCAFNet: Crossflow and Cross-Scale Adaptive Fusion Network for Detecting Salient Objects in RGB-D Images. IEEE Trans. Multimed. 2021, 24, 2192–2204. [Google Scholar] [CrossRef]
  6. Li, S.; Kang, X.; Hu, J. Image Fusion With Guided Filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar] [CrossRef]
  7. Budhiraja, S.; Sharma, R.; Agrawal, S.; Sohi, B.S. Infrared and Visible Image Fusion Using Modified Spatial Frequency-Based Clustered Dictionary. Pattern Anal. Appl. 2020, 24, 575–589. [Google Scholar] [CrossRef]
  8. Yin, M.; Duan, P.; Liu, W.; Liang, X. A Novel Infrared and Visible Image Fusion Algorithm Based on Shift-Invariant Dual-Tree Complex Shearlet Transform and Sparse Representation. Neurocomputing 2017, 226, 182–191. [Google Scholar] [CrossRef]
  9. Yin, M.; Liu, W.; Zhao, X.; Yin, Y.; Guo, Y. A Novel Image Fusion Algorithm Based on Nonsubsampled Shearlet Transform. Optik 2014, 125, 2274–2282. [Google Scholar] [CrossRef]
  10. Zhang, P.; Yuan, Y.; Fei, C.; Pu, T.; Wang, S. Infrared and Visible Image Fusion Using Co-Occurrence Filter. Infrared Phys. Technol. 2018, 93, 223–231. [Google Scholar] [CrossRef]
  11. Li, H.; Wu, X.-J.; Kittler, J. Infrared and Visible Image Fusion Using a Deep Learning Framework. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; IEEE: Beijing, China, 2018; pp. 2705–2710. [Google Scholar]
  12. Long, Y.; Jia, H.; Zhong, Y.; Jiang, Y.; Jia, Y. RXDNFuse: A Aggregated Residual Dense Network for Infrared and Visible Image Fusion. Inf. Fusion 2021, 69, 128–141. [Google Scholar] [CrossRef]
  13. Li, H.; Wu, X.-J. DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans. Image Process. 2019, 28, 10. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Liu, X.; Gao, H.; Miao, Q.; Xi, Y.; Ai, Y.; Gao, D. MFST: Multi-Modal Feature Self-Adaptive Transformer for Infrared and Visible Image Fusion. Remote Sens. 2022, 14, 3233. [Google Scholar] [CrossRef]
  15. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A Generative Adversarial Network for Infrared and Visible Image Fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  16. Ma, J.; Zhang, H.; Shao, Z.; Liang, P.; Xu, H. GANMcC: A Generative Adversarial Network With Multiclassification Constraints for Infrared and Visible Image Fusion. IEEE Trans. Instrum. Meas. 2021, 70, 5005014. [Google Scholar] [CrossRef]
  17. Li, J.; Zhu, J.; Li, C.; Chen, X.; Yang, B.-H. CGTF: Convolution-Guided Transformer for Infrared and Visible Image Fusion. IEEE Trans. Instrum. Meas. 2022, 71, 5012314. [Google Scholar] [CrossRef]
  18. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-Domain Long-Range Learning for General Image Fusion via Swin Transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  19. Zhai, X.; Kolesnikov, A.; Houlsby, N.; Beyer, L. Scaling Vision Transformers. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 1204–1213. [Google Scholar]
  20. Ding, M.; Xiao, B.; Codella, N.C.F.; Luo, P.; Wang, J.; Yuan, L. DaViT: Dual Attention Vision Transformers. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  21. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 9992–10002. [Google Scholar]
  22. Chen, Z.; Duan, Y.; Wang, W.; He, J.; Lu, T.; Dai, J.; Qiao, Y. Vision Transformer Adapter for Dense Predictions. arXiv 2022, arXiv:2201.10147. [Google Scholar]
  23. Liu, Z.; Hu, H.; Lin, Y.; Yao, Z.; Xie, Z.; Wei, Y.; Ning, J.; Cao, Y.; Zhang, Z.; Dong, L.; et al. Swin Transformer V2: Scaling Up Capacity and Resolution 2022. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  24. Sun, Z.; Cao, S.; Yang, Y.; Kitani, K. Rethinking Transformer-Based Set Prediction for Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2020; pp. 3591–3600. [Google Scholar]
  25. Jebens, M.; White, R. Remote Sensing and Artificial Intelligence in the Mine Action Sector. J. Conv. Weapons Destr. 2021, 25, 28. [Google Scholar]
  26. Jebens, M.; Sawada, H. To What Extent Could the Development of an Airborne Thermal Imaging Detection System Contribute to Enhance Detection? J. Conv. Weapons Destr. 2020, 24, 14. [Google Scholar]
  27. Fardoulis, J.; Depreytere, X.; Gallien, P.; Djouhri, K.; Abdourhmane, B.; Sauvage, E. Proof: How Small Drones Can Find Buried Landmines in the Desert Using Airborne IR Thermography. J. Conv. Weapons Destr. 2020, 24, 15. [Google Scholar]
  28. Baur, J.; Steinberg, G.; Nikulin, A.; Chiu, K.; de Smet, T.S. Applying Deep Learning to Automate UAV-Based Detection of Scatterable Landmines. Remote Sens. 2020, 12, 859. [Google Scholar] [CrossRef] [Green Version]
  29. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations (ICLR), Online, 12–14 May 2021. [Google Scholar]
  30. Si, C.; Yu, W.; Zhou, P.; Zhou, Y.; Wang, X.; Yan, S. Inception Transformer. arXiv 2022, arXiv:2205.12956. [Google Scholar]
  31. Li, H.; Wu, X.-J.; Durrani, T. NestFuse: An Infrared and Visible Image Fusion Architecture Based on Nest Connection and Spatial/Channel Attention Models. IEEE Trans. Instrum. Meas. 2020, 69, 9645–9656. [Google Scholar] [CrossRef]
  32. Xu, H.; Zhang, H.; Ma, J. Classification Saliency-Based Rule for Visible and Infrared Image Fusion. IEEE Trans. Comput. Imaging 2021, 7, 824–836. [Google Scholar] [CrossRef]
  33. Wang, Z.; Wu, Y.; Wang, J.; Xu, J.; Shao, W. Res2Fusion: Infrared and Visible Image Fusion Based on Dense Res2net and Double Nonlocal Attention Models. IEEE Trans. Instrum. Meas. 2022, 71, 5011410. [Google Scholar] [CrossRef]
  34. Ma, J.; Xu, H.; Jiang, J.; Mei, X.; Zhang, X.-P. DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion. IEEE Trans. Image Process. 2020, 29, 4980–4995. [Google Scholar] [CrossRef]
  35. Rao, D.; Wu, X.; Xu, T. TGFuse: An Infrared and Visible Image Fusion Approach Based on Transformer and Generative Adversarial Network. arXiv 2022, arXiv:2201.10147. [Google Scholar]
  36. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; IEEE: Boston, MA, USA, 2015; pp. 1–9. [Google Scholar]
  37. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: Honolulu, HI, USA, 2017; pp. 1800–1807. [Google Scholar]
  38. Huang, Z.; Wang, X.; Huang, L.; Huang, C.; Wei, Y.; Shi, H.; Liu, W. CCNet: Criss-Cross Attention for Semantic Segmentation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 2 November 2019. [Google Scholar]
  39. Wang, Z.; Bovik, A.; Sheikh, H.R.; Simoncelli, E. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2014, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  40. Zhang, Y.; Liu, Y.; Sun, P.; Yan, H.; Zhao, X.; Zhang, L. IFCNN: A General Image Fusion Framework Based on Convolutional Neural Network. Inf. Fusion 2020, 54, 99–118. [Google Scholar] [CrossRef]
  41. Li, H.; Wu, X.-J.; Kittler, J. RFN-Nest: An End-to-End Residual Fusion Network for Infrared and Visible Images. Inf. Fusion 2021, 73, 72–86. [Google Scholar] [CrossRef]
  42. Wang, Z.; Chen, Y.; Shao, W.; Li, H.; Zhang, L. SwinFuse: A Residual Swin Transformer Fusion Network for Infrared and Visible Images. IEEE Trans. Instrum. Meas. 2022, 71, 5016412. [Google Scholar] [CrossRef]
  43. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef]
  44. Tang, L.; Yuan, J.; Zhang, H.; Jiang, X.; Ma, J. PIAFusion: A Progressive Infrared and Visible Image Fusion Network Based on Illumination Aware. Inf. Fusion 2022, 83–84, 79–92. [Google Scholar] [CrossRef]
  45. Toet, A. TNO Image Fusion Dataset. Figshare. Dataset. 2014. Available online: https://figshare.com/articles/dataset/TNO_Image_Fusion_Dataset/1008029/1 (accessed on 21 February 2023).
  46. Davis, J.W.; Sharma, V. Background-Subtraction Using Contour-Based Fusion of Thermal and Visible Imagery. Comput. Vis. Image Underst. 2007, 106, 162–182. [Google Scholar] [CrossRef]
  47. Peng, H.; Long, F.; Ding, C. Feature Selection Based on Mutual Information Criteria of Max-Dependency, Max-Relevance, and Min-Redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  48. Haghighat, M.; Razian, M.A. Fast-FMI: Non-Reference Image Fusion Metric. In Proceedings of the 2014 IEEE 8th International Conference on Application of Information and Communication Technologies (AICT), Astana, Kazakhstan, 15–17 October 2014; pp. 1–3. [Google Scholar]
  49. Han, Y.; Cai, Y.; Cao, Y.; Xu, X. A New Image Fusion Performance Metric Based on Visual Information Fidelity. Inf. Fusion 2013, 14, 127–135. [Google Scholar] [CrossRef]
  50. Aslantas, V.; Bendes, E. A New Image Quality Metric for Image Fusion: The Sum of the Correlations of Differences. AEU-Int. J. Electron. Commun. 2015, 69, 1890–1896. [Google Scholar] [CrossRef]
  51. Zhai, G.; Min, X. Perceptual Image Quality Assessment: A Survey. Sci. China Inf. Sci. 2020, 63, 211301. [Google Scholar] [CrossRef]
Figure 1. The framework of IFormerFusion.
Figure 1. The framework of IFormerFusion.
Remotesensing 15 01352 g001
Figure 2. The architecture of the IFormer mixer.
Figure 2. The architecture of the IFormer mixer.
Remotesensing 15 01352 g002
Figure 3. Subject comparisons of images selected from the TNO dataset: (a) street; (b) Nato camp; (c) Kaptein 1123; (d) Kaptein 1654. Some targets and details are highlighted with red boxes.
Figure 3. Subject comparisons of images selected from the TNO dataset: (a) street; (b) Nato camp; (c) Kaptein 1123; (d) Kaptein 1654. Some targets and details are highlighted with red boxes.
Remotesensing 15 01352 g003
Figure 4. Object comparisons of 21 pairs of images selected from the TNO dataset.
Figure 4. Object comparisons of 21 pairs of images selected from the TNO dataset.
Remotesensing 15 01352 g004
Figure 5. Subject comparisons of images selected from the OSU dataset. Some targets and details are highlighted with red boxes.
Figure 5. Subject comparisons of images selected from the OSU dataset. Some targets and details are highlighted with red boxes.
Remotesensing 15 01352 g005
Figure 6. Object comparisons of 40 pairs of images selected from the OSU dataset.
Figure 6. Object comparisons of 40 pairs of images selected from the OSU dataset.
Remotesensing 15 01352 g006
Figure 7. Subject comparisons of images selected from the Road Scene dataset. Some targets and details are highlighted with red boxes.
Figure 7. Subject comparisons of images selected from the Road Scene dataset. Some targets and details are highlighted with red boxes.
Remotesensing 15 01352 g007
Table 1. Average results of 21 pairs of images from the TNO dataset. The best values are in bold, and the second-best values are in italic.
Table 1. Average results of 21 pairs of images from the TNO dataset. The best values are in bold, and the second-best values are in italic.
MethodMIFMIPSNRSSIMVIFFQabf
DenseFuse1.45170.226116.51740.74590.45920.3383
GANMcC1.47960.224414.23520.64240.40120.2380
IFCNN1.65180.251016.17840.71260.51700.5041
NestFuse1.69230.252413.99480.59670.49330.3667
Res2Fusion2.25480.338015.65560.69980.57070.4764
RFN-Nest1.40500.208415.00330.67760.48810.3598
SwinFuse1.65240.246014.91310.64920.49600.4453
SwinFusion2.22370.338915.18550.71020.59410.5216
U2Fusion1.24510.186516.06280.67370.48090.4249
Proposed2.33410.353816.77120.72580.60850.5765
Table 2. Average results of 40 pairs of images from the OSU dataset. The best values are in bold, and the second-best values are in italic.
Table 2. Average results of 40 pairs of images from the OSU dataset. The best values are in bold, and the second-best values are in italic.
MethodMIFMIPSNRSSIMVIFFQabf
DenseFuse1.89420.263414.74070.59880.32710.3906
GANMcC1.80140.254814.60340.53770.29430.2120
IFCNN1.94840.260214.75240.60740.33870.5322
NestFuse2.34740.328012.43010.47390.31130.3910
Res2Fusion2.63720.355715.07580.62680.36140.5013
RFN-Nest1.82550.249614.69950.56990.31630.2588
SwinFuse2.30020.312113.06340.55600.32940.4224
SwinFusion2.77630.369313.92790.61250.37720.5296
U2Fusion1.78180.243014.76070.60740.33000.4627
Proposed2.71550.363615.23450.61370.40140.5865
Table 3. Average results of 221 pairs of images from the Road Scene dataset. The best values are in bold, and the second-best values are in italic.
Table 3. Average results of 221 pairs of images from the Road Scene dataset. The best values are in bold, and the second-best values are in italic.
MethodMIFMIPSNRSSIMVIFFQabf
DenseFuse2.02840.287416.42690.72300.42630.3827
GANMcC1.91180.264513.37210.62200.37370.3360
IFCNN2.03230.282716.32970.68600.45700.5449
NestFuse2.67350.368511.99090.57840.38800.3754
Res2Fusion2.42730.333814.15260.66340.47820.5135
RFN-Nest1.92420.264013.98580.63440.40480.3129
SwinFuse2.20210.297714.30880.65660.44640.5066
SwinFusion2.33990.333914.31910.69320.46600.4611
U2Fusion1.93640.265415.77280.67010.44090.5221
Proposed2.23080.315817.01100.70520.48180.5490
Table 4. Computation efficiency of IFomrerFusion and compared methods (unit: second). The best values are in bold, and the second-best values are in italic.
Table 4. Computation efficiency of IFomrerFusion and compared methods (unit: second). The best values are in bold, and the second-best values are in italic.
MethodTNOOSURoadScene
DenseFuse0.00600.00790.0060
GANMcC1.85201.00680.5242
IFCNN0.01600.01320.0119
NestFuse0.01670.02130.0144
Res2Fusion10.2080.9192.937
RFN-Nest0.26870.05100.1752
SwinFuse0.12290.12480.1929
SwinFusion8.62112.20205.1209
U2Fusion2.31071.26390.6492
Proposed0.34790.26780.2488
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiong, Z.; Zhang, X.; Hu, Q.; Han, H. IFormerFusion: Cross-Domain Frequency Information Learning for Infrared and Visible Image Fusion Based on the Inception Transformer. Remote Sens. 2023, 15, 1352. https://doi.org/10.3390/rs15051352

AMA Style

Xiong Z, Zhang X, Hu Q, Han H. IFormerFusion: Cross-Domain Frequency Information Learning for Infrared and Visible Image Fusion Based on the Inception Transformer. Remote Sensing. 2023; 15(5):1352. https://doi.org/10.3390/rs15051352

Chicago/Turabian Style

Xiong, Zhang, Xiaohui Zhang, Qingping Hu, and Hongwei Han. 2023. "IFormerFusion: Cross-Domain Frequency Information Learning for Infrared and Visible Image Fusion Based on the Inception Transformer" Remote Sensing 15, no. 5: 1352. https://doi.org/10.3390/rs15051352

APA Style

Xiong, Z., Zhang, X., Hu, Q., & Han, H. (2023). IFormerFusion: Cross-Domain Frequency Information Learning for Infrared and Visible Image Fusion Based on the Inception Transformer. Remote Sensing, 15(5), 1352. https://doi.org/10.3390/rs15051352

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop