Next Article in Journal
Wideband Interference Cancellation System Based on a Fast and Robust LMS Algorithm
Next Article in Special Issue
Dataset Condensation via Expert Subspace Projection
Previous Article in Journal
MAE-Based Self-Supervised Pretraining Algorithm for Heart Rate Estimation of Radar Signals
Previous Article in Special Issue
A Zero-Shot Low Light Image Enhancement Method Integrating Gating Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SFPFusion: An Improved Vision Transformer Combining Super Feature Attention and Wavelet-Guided Pooling for Infrared and Visible Images Fusion

International Joint Laboratory on Artificial Intelligence of Jiangsu Province, School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi 214122, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work and should be considered co-first authors.
Sensors 2023, 23(18), 7870; https://doi.org/10.3390/s23187870
Submission received: 10 August 2023 / Revised: 8 September 2023 / Accepted: 11 September 2023 / Published: 13 September 2023
(This article belongs to the Special Issue AI-Driven Sensing for Image Processing and Recognition)

Abstract

:
The infrared and visible image fusion task aims to generate a single image that preserves complementary features and reduces redundant information from different modalities. Although convolutional neural networks (CNNs) can effectively extract local features and obtain better fusion performance, the size of the receptive field limits its feature extraction ability. Thus, the Transformer architecture has gradually become mainstream to extract global features. However, current Transformer-based fusion methods ignore the enhancement of details, which is important to image fusion tasks and other downstream vision tasks. To this end, a new super feature attention mechanism and the wavelet-guided pooling operation are applied to the fusion network to form a novel fusion network, termed SFPFusion. Specifically, super feature attention is able to establish long-range dependencies of images and to fully extract global features. The extracted global features are processed by wavelet-guided pooling to fully extract multi-scale base information and to enhance the detail features. With the powerful representation ability, only simple fusion strategies are utilized to achieve better fusion performance. The superiority of our method compared with other state-of-the-art methods is demonstrated in qualitative and quantitative experiments on multiple image fusion benchmarks.

1. Introduction

Image fusion is the combination of multiple source images carrying complementary information to generate a high-quality fused image, which is used to make up for the limitation of incomplete image information obtained by a single type of sensor. As a task in the field of image processing, image fusion is applicable to various downstream tasks, including target detection [1,2], object tracking [3,4], and semantic segmentation [5].
Infrared image and visible image are two typical data sources studied in image fusion. Infrared image contains salient thermal infrared features, which are not affected by external conditions such as bad weather and poor illumination, but usually lack texture details [6]. Visible image contains more texture details and is suitable for human visual perception but very sensitive to illumination and weather conditions [7]. Therefore, infrared and visible image fusion methods are utilized to generate one image that contains both salient features and rich texture details.
In general, the key to infrared and visible image fusion is how to extract and fuse complementary features while reducing redundant information. In recent years, a large number of fusion algorithms are proposed, which can be roughly divided into non-deep learning and deep learning-based methods. Non-deep learning methods are mainly based on signal processing operations [8,9]. For example, multi-scale transform (MST) [10,11] can decompose the original image into components of different scales and extract multi-scale features. Then, an effective fusion strategy is designed to fuse the features. Finally, the image is reconstructed using the inverse transformation of feature extraction. In addition, the methods of non-deep learning also include sparse representation (SR) [12,13,14,15]-based methods, subspace [16,17]-based methods, and low-rank representation (LRR) [18,19,20,21]-based methods. Although the non-deep learning methods can synthesize satisfactory results, they still have some drawbacks: (1) manually designed fusion strategies cannot adapt to complex image fusion conditions and have poor generalization ability; (2) manual feature extraction has limitations in comprehensively capturing multi-modal images, which introduces noise and causes image distortion.
With the rise of deep learning, a lot of fusion methods based on deep learning are proposed to solve the above shortcomings [22]. Existing deep learning-based fusion methods can be divided into four categories: (1) convolutional neural network (CNN) [7,23]-based methods; (2) auto-encoder (AE) [24,25]-based methods; (3) generative adversarial network (GAN) [26,27]-based methods; and (4) Transformer [28,29]-based methods.
The CNN-based methods have excellent feature extraction capabilities, but the simple use of CNN cannot extract the deep and global features of source images. Therefore, in order to extract more deep features, the AE-based training strategy was introduced to image fusion task, which has strong generalization ability and can obtain richer image features. However, this kind of fusion method still needs complex fusion strategies to be manually designed. Moreover, given the powerful ability of GAN to estimate probability distributions in an unsupervised manner, they are well suited for unsupervised tasks such as image fusion. The GAN-based methods establish an adversarial game between the generator and the discriminator and continuously improve the performance of the fused image. Unfortunately, these methods are extremely unstable and prone to image distortion. In addition, with the development of Transformer and the mining of its powerful global feature extraction ability, scholars have begun to study whether Transformer-based methods are suitable for image fusion. However, for methods fully adapted to infrared and visible image fusion tasks, it is necessary to enhance multi-scale base information and detail textures while extracting global features, which the vast majority of algorithms combining Vision Transformer and CNN cannot do. In addition, the convolution kernel of CNN also has certain limitations for extracting image detail features and salient features.
Through the above description of the problem, we can find that most of the existing deep learning-based methods except Transformer-based one rely on the convolution operation. Although CNN has good local feature extraction ability, its small receptive field makes it difficult to model the long-range dependence, and thus, it is difficult to obtain the global features of the source images. In addition, most fusion networks introducing a pooling operation can extract rich image features using scale transformation, but it also brings a lot of information loss and artifacts.
In order to solve the above problems, the wavelet-guided pooling is applied into super feature attention (SFA) to form a novel Transformer architecture, which will be introduced in next section. The SFA mechanism can learn an efficient global representation, especially for shallow features. The number of image primitives for subsequent processing is reduced in a token space. The multi-scale features and detail information can be simultaneously extracted using wavelet-guided pooling. In the feature extraction stage, three convolutional layers are utilized to expand the channel number of the features and to enlarge the receptive field.
To be specific, the SFA attention aggregates tokens into local features by learning sparse relationships between tokens and local regions, and then, we use a self-attention machine to fully mine the local and global dependencies of the source images to capture complementary features. After that, the local features are mapped to the original space to extract global features. Finally, the deep information are extracted using embedded wavelet-guided pooling, where the multi-scale base component is transmitted to the next encoding layer and the detail components are skipped to the corresponding decoding layer after using the l 1 -norm and softmax-like fusion strategy for inverse wavelet-guided pooling. As shown in Figure 1, compared with SwinFuse [29] (base on the Swin Transformer [30]), our method can extract global features while enhancing the detail textures, as shown in the green box; the reflection of the tree is clearly displayed. SwinFuse is less capable of processing multimodal images than we are, resulting in generated images darker and affecting visual senses. Our method combines the efficient global representation ability of Transformer with the detail enhancement of wavelet-guided pooling, which can extract more fine-grained features and reconstruct better-quality fused images.
The main contributions of this paper are summarized as follows:
  • An improved Transformer architecture combined with wavelet-guided pooling and attention mechanism (super feature) is proposed, which preserves global features while enhancing detail information.
  • A simple yet efficient fusion network is proposed, which only employs a simple l 1 -norm and the softmax-like fusion strategy to achieve better fusion performance.
  • The qualitative and quantitative experiments on several public image fusion datasets demonstrate the superiority of the proposed fusion method.
The rest of this paper is organized as follows: Section 2 discusses related work on image fusion. In Section 3, we present the details of the proposed fusion method. Section 4 shows the experimental results and compares it with some state-of-the-art methods. Finally, we present the conclusions in Section 5.

2. Related Work

2.1. Deep Learning Model for Image Fusion

Due to the high efficiency and feature representation ability of deep learning, it is widely used in the field of image fusion. In IFCNN [31], convolutional layers are first used to extract salient image features from multiple input images. Then, the convolutional features of multiple input images are fused using appropriate fusion rules. Finally, the fusion features were reconstructed using two convolutional layers to generate an information-rich fused image. The representative auto-encoder-based method is DenseFuse [24]. The encoder network composed of cascaded networks is used to extract the features of the source images; then, a manually designed fusion strategy is conducted on these features; and finally, image features are reconstructed through the decoder network. In addition, NestFuse [25] similarly adopt the above three basic steps to fuse features. In the end-to-end model termed LRRNet [32], Li et al. propose a learnable low-rank representation (LRR) approach to the fusion task, which can preserve the image details and enhance the salient features of the source images. With the rise of generative adversarial network, Ma et al. [26] first applied the GAN [33] algorithm to image fusion. However, GAN-related networks are extremely unstable, so Ma et al. proposed DDcGAN [27], which contains two discriminators to force the generator to fuse useful information from original images.
However, these methods ignore the importance of global dependencies and cannot be well enhanced for details. Therefore, our proposed method combining super feature attention and wavelet-guided pooling extracts the global information of the image while enhancing the detail texture, so as to improve the global dependence of the images.

2.2. Transformer for Image Fusion

The Transformer structure was proposed by Vaswani et al. [34], which was applied to machine translation. Subsequently, Dosovitskiy et al. [35] proposed the Vision Transformer (ViT) to solve the image classification problem. In recent years, many Transformer-based infrared and visible image fusion models have been proposed. The Transformer architecture also was introduced into the image fusion task [36]; the original authors proposed a two-stage training approach to extract features and developed a Transformer-based multi-scale fusion strategy that attends to both local and long-range information.
With the proposal of Swin Transformer [30], some scholars have applied it to the field of image fusion. The Swin Transformer employs window-based local attention to restrict attention to non-overlapping local windows. Despite being able to reduce the redundancy in local regions, the features of local windows still exists in shallow layers and global features cannot be fully obtained. SwinFusion [28] adopts a hybrid structure of CNN and Swin Transformer. The intra-domain fusion module extracts features of the source images through the attention mechanism and then exchanges features via the inter-domain fusion modules. The small kernel convolution layer is used to fuse the global features. Finally, the image is reconstructed with the reconstruction unit. However, it lacks long-term dependence on feature fusion, and it is not sufficient for deep global detail feature extraction. SwinFuse [29] designs a fusion network based on the residual Swin Transformer, where the fusion result is obtained through global feature extraction, fusion layer, and feature reconstruction. DATFuse [6] designs a dual-attention residual module for feature extraction and a Swin Transformer module for global complementary information preservation. However, they do not change the Swin Transformer feature extraction module, so they are also equally limited when integrating complementary features and global interactions.
Recently, THFuse [37] introduced the Transformer into the CNN-based fusion network to focus on both global and local information. In this work, they extracted the shallow features using the dual-branch CNN module, and the vision Transformer module was used to obtain the global channel and spatial relationship in the features. In TCCFusion [38], they designed the local feature extraction branch to retain local complementary information and the global feature extraction branch composed of three Transformer blocks to maintain long-range dependencies. These two modules are sufficient to capture local and global useful information. Yi et al. [39] proposed the fusion network (TCPMFNet), which is based on the auto-encoder structure and designed the Transformer–convolutional parallel mixed fusion strategy to achieve outstanding feature fusion performance. DGLT-Fusion [40] decouples global–local information learning into Transformer and CNN modules, which enables the network to extract better global–local information. TGFuse [41] was proposed as a fusion algorithm that combines Transformer and GAN. The Transformer module is simply used to learn the global fusion relationship. In contrast, we combine the Super Feature Transformer module and wavelet-guided pooling to preserve more detail texture information while learning global features. Wave-ViT [42] was proposed as a framework using wavelet transforms and self-attention learning, which is suitable for image recognition, object detection, and instance segmentation. However, that Transformer module is less capable of extracting global features than our super feature Transformer, and our method can achieve excellent results using a simple Haar wavelet pooling compared to their complex DWT-Convolution-IDWT. Our pooling operation is able to capture multi-scale features while extracting vertical, horizontal, and diagonal detail textures. Compared with our method, they only use 3 × 3 convolution to extract local features and do not fully extract multi-scale information and detail textures. In addition, image fusion requires more texture details of visible images, and our method can better meet the requirements of the field of image fusion, but Wave-ViT does not.
Although the above Transformer-based methods can achieve good results, they rely too much on convolutional layers when extracting local features and fail to enhance certain meaningful features. In general, the improved Vision Transformer proposed by our SFPFusion well combines the attention mechanism and Haar wavelet-guided pooling operation to extract global features while preserving texture details, which are ignored by most Vision Transformers applied in the field of image fusion.

3. Proposed Fusion Method

In this section, the proposed fusion method will be presented. Firstly, we give the details of the framework and model architecture. Then, the details of training phase will be given.

3.1. Overall Framework

As shown in Figure 2, our network framework mainly contains three main parts: feature extraction, feature fusion, and feature reconstruction. Our proposed method aims to generate a fused image by fusing local and global complementary features from source images. The following will be the details of the framework.

3.1.1. Feature Extraction

The Convolutional Block: Given a pair of already registered testing infrared I i r R C × H × W and visible image I v i R C × H × W (where C, H, and W represent the channel number of input images, height, and width, respectively). The first stage consists of two modules, C B 1 and W a P 1 . The shallow features of the images F i r and F v i are extracted with a convolutional block ( C B 1 ), which can be expressed as follows:
F i r , F v i = H C B 1 ( I i r ) , H C B 1 ( I v i )
where H C B 1 ( · ) represents the first convolutional block.
The convolutional block consists of multiple convolutional layers, which provides an effective way to extract local information. The  C B 1 contains three convolutional layers using ReLU activation function, whose convolution kernel is 3 and stride is 1. The other main function of C B 1 is to transform the input channel number to 64, which will facilitate the subsequent extraction of deeper features by expanding the number of channel.
Wavelet-guided Pooling: After that, we extract multi-scale base information ( F L L ) and detail features ( F V D , F H D , F D D ) using the first wavelet-guided pooling operation ( W a P 1 ), which can be formulated as follows:
F L L ω , F V D ω , F H D ω , F D D ω = W a P 1 ( F ω ) , ω i r , v i
Compared with the ordinary pooling operation, wavelet-guided pooling has four kernels:
L L = 1 1 1 1 V D = 1 1 1 1
H D = 1 1 1 1 D D = 1 1 1 1
The multi-scale component L L obtains the global multi-scale base information, while the detail components V D , H D , and D D extract the vertical, horizontal, and diagonal detail textures.
The traditional pooling operation has some limitations. For max-pooling, significant features are lost in the pooling operation and images cannot be reconstructed accurately. Average pooling, on the other hand, dilutes meaningful features. Inspired by [43], we use Haar wavelet-guided pooling in the Super Feature Pooling Transformer module. Haar wavelet is one of the classic wavelet transforms. Compared with the traditional pooling operation, Haar wavelet-guided pooling decomposes the original image into different component channels, so as to extract multi-scale information and texture details. Moreover, it has minimal information loss and can reconstruct images completely without any post-processing step.
For our framework, only the multi-scale base information F L L passes to the encoding layer and the detail features ( F V D , F H D , F D D ) are skipped to the corresponding unpooling layer after feature fusion. Our unpooling operation is based on the inverse transform of wavelet-guided pooling, which can reconstruct the image accurately and minimize the information loss.
Super Feature Pooling Transformer: However, in [43], they only use convolutional layers to extract image features, which has certain limitations. Given that the inherent small receptive field of CNN cannot effectively extract global information of source image, the Super Token Transformer [44] is introduced. It can obtain efficient global representation and enhance detail information simultaneously. Specifically, we improve it and embed wavelet-guided pooling into the Transformer architecture, termed as Super Feature Pooling Transformer, so that the global features can be augmented to obtain richer features (multi-scale features and detail information).
To be specific, our Super Feature Pooling Transformer is mainly composed of four modules: position encoding (PE), super feature attention (SFA), embedded pooling layer (EPL), and patch merging module (PM).
As shown in Figure 2, the multi-scale features F L L 1 extracted by W a P 1 are used as the input of the Super Feature Pooling Transformer. For convenience of expression, we replace F L L 1 with F i n .
Given the input features F i n R C × H × W , we firstly add position information to all the tokens by using position encoding (PE). PE containing a 3 × 3 depth-wise convolution can learn the local representation better than absolute position encoding (APE) [34] and relative position encoding (RPE) [30,45]. The formula is defined as follows:
T = H p o s ( F i n )
where H p o s ( · ) represents the positional encoding, which is used to add position information into all the tokens. T R C × H × W are the visual tokens we obtain.
After extracting the position information, we next employ super feature attention to extract global features:
F S F = S F A ( L N ( T ) ) + T
where S F A ( · ) represents the super feature attention mechanism and L N ( · ) denotes LayerNorm operation. In the following, we will describe the process of super feature attention in detail.
Given the visual tokens T R C × H × W , we first compute the initial super features S R C × m × n using average pooling in regular grid regions, where the grid size is h × w , m = H h , and n = W w . The number of super features is m × n . In order to better extract complementary features and global features, it is necessary to calculate the mapping associations Q R m n × h w × 9 between T and S. The details of SFA operation, please refer to our Supplementary Materials.
Following the super feature attention module, an embedded pooling layer is deployed to augment multi-scale features and detail information, which can be defined as follows:
F L L , F V D , F H D , F D D = Pooling ( F S F )
where Pooling ( · ) represents the embedded wavelet-guided pooling operation.
Our embedded pooling layer has the same structure as the above wavelet-guided pooling, as shown in Figure 2. The extracted multi-scale base information is passed to the next layer, and the detail features are given in the unpooling operation for image reconstruction.
As shown in Figure 3a, we can observe that the super feature attention mechanism well builds long-range dependencies and extracts the global salient features. In addition, the multi-scale base information (b) and detail information (c)–(e) obtained by the embedded pooling layer are well enhanced, which helps to reconstruct the image with fine features.
Finally, the extracted multi-scale base features F L L go through the patch merging module to increase the number of output channels. With super feature attention, we can make full use of long-range dependencies to extract enhanced global features F G F .
F G F = H P M ( F L L )
where H P M ( · ) represents the patch merging module.
In summary, with the combination of PE, SFA, and EPL, the fusion network is able to fully implement local information extraction and global information integration.
After extracting enhanced global information and detail textures from S F P 1 in the second stage, we put the output features into the third stage S F P 2 with the same structure to extract deeper features again, as shown in Figure 2. In the fourth stage, when the global features and complementary features of the three stages are extracted, they are passed to the convolution block ( C B 2 ) to obtain local semantic information and are mapped into the high-order feature space. The network architecture of the specific feature extraction stages is shown in Table 1.

3.1.2. Fusion Strategy

Global Features Fusion Strategy: It is well known that choosing an appropriate fusion strategy is very important for image fusion. In this work, we use the fusion strategy based on l 1 -norm and softmax-like. The  l 1 -norm represents the sum of the absolute values of the elements in the vector.
Given the extracted infrared F o u t i r R C × H × W and visible F o u t v i R C × H × W features, we first sum the vectors in each channel using the l 1 -norm operation and then the weighting maps are generated using the softmax-like operation, which can be calculated with the following:
W ω ( x , y ) = F o u t ω ( x , y ) 1 i i r , v i F o u t i ( x , y ) 1 , ω i r , v i
where · 1 denotes the l 1 -norm. ( x , y ) indicates the corresponding position in deep features ( F o u t i r and F o u t v i ) and weighting maps ( W i r and W v i ).
Finally, the extracted features in each channel are multiplied by the weighting maps to obtain the fused feature maps F f . The specific process of the fusion strategy is shown in Algorithm 1.
Detail Features Fusion Strategy: For the pooling layer, the detail information preserved by wavelet-guided pooling are sparse but also very significant. Therefore, it can achieve good results without complex operations. Moreover, it is very important to preserve as much texture details as possible in the field of image fusion. In this work, we sum the saved infrared and visible detail information corresponding to the scale, which can be formulated as follows:
F d e t f = F d e t i r + F d e t v i , d e t V D , H D , D D
where F d e t i r and F d e t v i represent the texture details of infrared and visible images in the vertical, horizontal, and diagonal directions, respectively. F d e t f represents the fused features of the corresponding scale detail parts.
Algorithm 1 Procedure for fusion strategy.
  Require: 
 
 
   the extracted infrared F o u t i r R C × H × W and visible F o u t v i R C × H × W features
  Ensure: 
 
 
   fused feature maps F f ;
   1:
Take the absolute value of each pixel value;
   2:
Make the sum of pixel values within each channel;
   3:
Calculate the weighting maps by softmax-like operation according to Equation (9);
   4:
for each i [ 1 , C ] do
   5:
   Multiplying features within i-th channel and weighting maps, which can be expressed as follows:
F f i = ω i r , v i W ω ( x , y ) × F o u t i ω ( x , y )
   6:
end for
   7:
Concatenate the channel of F f i , and the number of channel is restored to C;
   8:
return  F f ;

3.1.3. Feature Reconstruction

In this phase, the fused features are used to reconstruct the image through four convolutional blocks and three wavelet-guided unpoolings. Specifically, our wavelet-guided unpooling is the inverse operation of wavelet-guided pooling. The detail information extracted in the feature extractor adopts an addition strategy to the corresponding scale and is used to reconstruct the image together with the fused global feature, as shown in Figure 2.
In addition, our image reconstruction stage can fully recover the image features without any post-processing operations, so as to achieve the effect of minimum information loss. The specific details of the image reconstructor are shown in Table 2.

3.2. Training Phase

Our fusion network is based on an auto-encoder and the fusion strategy needs to be discarded during the training phase. In the training phase, the trained feature extractor can better extract global features and detail information about source images, and the trained feature reconstructor can recover image features well and reduce information loss, as shown in Figure 4.
In order to reconstruct the input image more accurately, the loss function is composed of a pixel loss function L p i x e l and a structural similarity loss function L s s i m , which can be formulated as follows:
L t o a l = L p i x e l + λ L s s i m
where λ denotes the tradeoff value between L p i x e l and L s s i m .
Specifically, L p i x e l is mean square error, which is used to make sure that the reconstructed image is more similar to the input image at the pixel level, as calculated in the following equation:
L p i x e l = O I F 2
where O and I indicate the output and input images, respectively. · F 2 is the l 2 -norm.
In addition, the SSIM loss L s s i m is formulated as follows:
L s s i m = 1 S S I M O , I
where S S I M ( · ) denotes the structural similarity between the output image and input image [46].

4. Experimental Results and Analysis

4.1. Experimental Settings

In the training phase, we randomly selected 40,000 images from MS-COCO [47] to train our auto-encoder network, where all images are resized to 224 × 224 . The batch size, epochs, and learning rate are set to 4, 4, and 1 × 10 4 , respectively. The hyperparameter λ in Equation (12) is set as 10, and we use the same tradeoff value in all experiments. Moreover, our iterations are i = 1 . All the involved experiments are conducted on an NVIDIA RTX 3090Ti GPU (NVIDIA, Santa Clara, CA, USA) and Intel Core i7-10700 CPU (Intel, Santa Clara, CA, USA). In the test phase, for RGB images, we first convert the visible image to YCbCr color space, and then, take the Y channel of the visible images into our proposed network together with the infrared images for fusion. Finally, the fused image is converted back to the RGB color space by concatenating the Cb and Cr channels of the visible images.
To comprehensively evaluate the proposed method, we perform qualitative and quantitative experiments on the MSRS dataset [7] with 361 image pairs, the LLVIP dataset [48] with randomly selected 389 image pairs, and the TNO dataset [49] with randomly selected 16 image pairs. We compare our method with eight state-of-the-art (SOTA) approaches, including DenseFuse [24], FusionGAN [26], SwinFuse [29], U2Fusion [50], AUIF [51], CUFD [23], MUFusion [52], and AEFusion [53]. The implementations of these approaches are publicly available and the initial parameters of the compared methods remain the same.
Five quality metrics are selected for the fair and quantitative comparison between our fusion method and other SOTA methods, including standard deviation (SD) [54], visual information fidelity (VIF) [55], average gradient (AG) [56], entropy (EN) [57], and Q a b f [58]. SD reflects the distribution and contrast of the fused image, which is consistent with the visual perception. VIF is the metric based on natural scene statistics and the conception of image information extracted by the human visual system. AG quantifies the gradient information of the fused image and represents its detail textures. EN is used to represent the amount of information contained in image. Q a b f measures the amount of edge information. Moreover, a fusion algorithm with larger SD, VIF, AG, EN, and Q a b f indicates better fusion performance.

4.2. Comparative Experiments

In this section, our method will be compared with eight SOTA methods both qualitatively and quantitatively to show the superiority of our algorithm on MSRS, LLVIP, and TNO datasets. For more experiments, please refer to our Supplementary Materials.

4.2.1. Fusion Results on MSRS Dataset

One set of source image pairs of the MSRS dataset and their corresponding fused images obtained using different methods are demonstrated in Figure 5. In daytime scenarios, thermal radiation information from infrared images should be used as complementary information to visible images. As a whole, although FusionGAN enhances the thermal radiation information, it causes serious spectral contamination and affects the overall visual sense. SwinFuse and AUIF extract too much background features of infrared images, making the whole scene submerged in darkness, which are completely not suitable for daytime scenes. Moreover, as shown in the green and red boxes, DenseFuse and U2Fusion weaken infrared targets and also fail to retain detail information of visible images. Although AEFusion preserves more texture details, it similarly weakens salient features. In addtion, CUFD and MUFusion are able to synthesize the texture details of visible images and the salient features of infrared images. Unfortunately, a large number of artifacts are introduced into fused results, resulting in poor visual perception. Only our method enhances the thermal radiation information while preserving the texture information and provides a pleasing visual effect.
In order to better evaluate the quality of the generated images, we conduct qualitative comparisons on 361 image pairs from MSRS datasets to verify the effectiveness of our algorithm, as shown in Table 3. It is worth noting that our method exhibits excellent superiority in all five metrics. The best SD metric indicates that our fused images have richer contrast information, which achieves a good visual effect. The best result in VIF metric indicates that our results have better visual perceptual performance, which is consistent with human visual perceptions. For the AG and Q a b f metrics, they achieve the best results, meaning that our results have richer texture details and more edge information, benefiting from our proposed wavelet-guided pooling. Moreover, the best EN metric shows that our fused images contains more scene information.

4.2.2. Fusion Results on LLVIP Dataset

We further randomly select 389 image pairs of nighttime scenes on the LLVIP dataset to demonstrate the effectiveness of our method. In the dark scenes, the infrared images contain a lot of thermal radiation information and detail textures, which will complement the visible images containing limited detail information. As shown in Figure 6, FusionGAN fails to preserve texture details and the background suffers from severe contamination, causing visual conflicts. For SwinFuse and AUIF, the infrared features are weakened. These too dark images lead to the retention of little detail information, making people obtain less useful information. As shown in the red box, U2Fusion and DenseFuse do not retain more texture details, mainly due to inadequate feature extraction of infrared images. Although CUFD and AEFusion retain more thermal radiation information, they blur edge information and produce extensive artifacts at the same time, resulting in overall image blur. In addition, MUFusion introduces additional information into the fused images in some cases, which can be seen from the ground in a green box. It is worth noting that, as shown in the green box, our algorithm can very clearly show the reflection of branches in the dark, which has a great advantage compared with the algorithms that blur the targets.
The quantitative results of five metrics on image pairs from the LLVIP dataset are presented in Table 4. It can be observed that our method ranks first in three metrics. Our method achieves the highest AG metric, indicating that our fused images possess more detailed textures. Moreover, the best metric Q a b f implies that more edge information is preserved in fused results. For the best EN metric, it demonstrates that our method has good superiority in containing valid information. In addition, our method still has a pleasing performance on the VIF metric, only following FusionGAN. It shows that our fused images can still achieve good visual effects. On the SD metric, the proposed method follows FusionGAN and AEFusion by a narrow margin. And, with the help of the visualization results, we can see that FusionGAN and AEFusion blur a lot of detail information.

4.2.3. Fusion Results on TNO Dataset

For grayscale images of infrared and visible image fusion, we select 16 image pairs on the TNO dataset for comparison. As shown in Figure 7, DenseFuse, U2Fusion, and AEFusion weaken the thermal radiation information, and the part of sky introduces plenty of artifacts. Moreover, DenseFuse and AEFusion also blur the edge information, which can be seen from red box. For MUFusion, it introduces too much extra information, causing conflicts in human perception. FusionGAN retains too much of the infrared features, causing loss of details and background contamination. In contrast, SwinFuse, AUIF, and CUFD preserve the salient features well and can achieve good results. However, it also causes the loss of details. As shown in the red box, they all do not preserve the detail information of branches and fences well. In general, our method preserves the salient features and texture details well, which is attributed to our proposed method of wavelet-guided pooling combined with super feature attention.
The comparative subjective results of different methods on the TNO dataset are shown in Table 5. Our method achieves the best results on AG and Q a b f metrics, which indicates that our fused results contain richer gradient textures and more edge information. On the VIF metric, our method has a small gap compared with CUFD, indicating that our method can still achieve good visual results. However, in terms of SD and EN metrics, our method does not achieve as good results as the other two datasets. This is justified. Compared with the previous two datasets (all RGB images), TNO as the dataset of grayscale images mainly contains salient features. The visible images in TNO dataset do not contain more obvious texture details as visible images in RGB. From our fusion network, it can be seen that our method combines wavelet-guided pooling and super feature attention, which prefers to preserve the intensity information and global features of the source images. Therefore, it is understandable that our method performs worse on the TNO dataset than the first two datasets. Even so, the results of the two metrics can still be kept within the top three.
In general, the visualization results show that our method has obvious advantages in preserving global features and maintaining detail information, while achieving satisfied visual effects. The objective evaluation metrics also confirm that our method can achieve better fusion performance.

4.3. Comparison of the Efficiency

We also conduct comparison experiments about the average running time of different fusion methods. We use 361 pairs of infrared and visible images with the same size ( 640 × 480 ) to test the operational efficiency of these methods. As shown in Table 6, U2Fusion achieves relatively better performance on efficiency comparison. For our SFPFusion, due to the improvement in the Transformer and combined with wavelet pooling operation guidance, this is somewhat slower than methods with fewer parameters. However, our method is able to achieve better results than SwinFuse, which is based on Swin Transformer. In addition, our SFPFusion still obtains a comparable result on the efficiency, with an average time of less than 0.1 s per image.

4.4. Ablation Studies

4.4.1. Analysis of Transformer

Since our network introduces the super feature attention mechanism combined with wavelet-guided pooling, making it better adapted to infrared and visible image fusion, we use the original Super Token Transformer [44] in ablation experiments to demonstrate the advantages of our embedded wavelet-guided pooling. As shown in Figure 8d, although the Super Token Transformer without pooling preserves the features of visible images better, it weakens the salient features from infrared images and the detail part is not enhanced compared with our method. Moreover, the sky also forms an over-exposed scene. In addition, it can also be seen from the evaluation metrics (Table 7) that our fusion results are better than the original framework without pooling as a whole.
In addition, we conduct Vision Transformer ablation experiments to confirm that our improved Transformer is better than the original ViT. We introduce the original ViT [35] structure, which does not have embedded pooling, and thus, the unpooling operations of W a U P 2 and W a U P 3 are not included in the framework of the ablation experiments. As shown in Figure 8e, compared with it, our fusion result can preserve more sufficient texture details and global features and can accurately reconstruct the image without introducing noise. The quantitative evaluation in Table 7 also shows that the quality of our generated images are all better than using ViT.

4.4.2. Analysis of CNN

As we stated in Section 1, the size of convolution kernel of CNN determines how many features it can extract from its receptive field, which is its limitation. Therefore, we conduct ablation experiments combining CNN with pooling to confirm that our super feature attention mechanism can better capture long-range dependencies, perform effectively global representation, and obtain global features of the source images. When training the auto-encoder and the hyperparameter λ in Equation (12) is set as 10, the auto-encoder cannot reconstruct the image. In the ablation experiment, the hyperparameter λ is set to 100 to reconstruct the image well. As shown in Figure 8f, the fusion results using CNN weaken the salient features and fail to model the long-range dependence of the image, resulting in a very poor visual perception. Moreover, it also introduces a lot of artifacts in the sky. In contrast, our method is superior to the fusion results using CNN in both visual results and quantitative metrics (Table 7).

4.4.3. Analysis of Structure

In our framework, after the first convolutional block ( C B 1 ), we utilize wavelet-guided pooling to extract the first multi-scale features and detail information. In the ablation experiments, we discarded the first layer pooling ( W a P 1 ). As shown in Figure 8g, artifacts still appear in the sky despite sufficient features extracted from the generated image. And, as shown in Table 7, the evaluation metrics are all significantly lower than our method. This also indicates that W a P 1 is also important in extracting the information of the first layer.
In addition, we perform ablation experiments on the number of Super Feature Pooling Transformers. We use a S F P module and the visualization result is shown in Figure 8h. Although it preserves good salient features, the enhancement of the detail part is not as good as our two S F P modules, and the sky also forms overexposure. The superiority of our use of two S F P modules can also be seen in Table 7. Moreover, we try to use three S F P modules to extract features but unfortunately fail when training the auto-encoder. The reason for the failure is that the last layer of the reconstructed network uses a Sigmoid activation function, which causes the gradient to disappear in the network with deep superposition.

4.4.4. Analysis of Global Features Fusion Strategy

For the global feature fusion strategy, we use the simple l 1 -norm and softmax-like function-based strategy, which can also achieve good fusion results. Thus, in the ablation experiments, the addition strategy and the average strategy are performed. As shown in Figure 8i,j, the addition fusion strategy excessively enhances the salient features, resulting in over-exposure of the image. However, using the average strategy weakens the infrared intensity, resulting in the overall dark image, and also introduces a lot of noise in the sky. In addition, from the evaluation of metrics (Table 7), it can also be found that our method is generally better than the other two strategies, which confirms that our fusion strategy is simple and efficient.

4.4.5. Analysis of Detail Features Fusion Strategy

As mentioned in Section 3.1.2, the extracted detail information is sparse and clear, which can lead to better fusion performance without complicated operations. As shown in Figure 3 and Figure 9, the generated feature maps ( F V D 2 , F H D 2 , and F D D 2 ) clearly show the detail textures. The detail features extracted with our method are very prominent, so they can be fused well using only a simple addition strategy. In the ablation experiments, we use the average strategy for the fusion strategy of detailed features. As shown in Figure 8k, using the average strategy weakens the detail intensity, which results in the overall blurring of the images. Moreover, it can be seen from the quantitative results (Table 7) that all the metrics are significantly lower than those of our method. Therefore, the addition fusion strategy is very effective for detail feature fusion.

4.4.6. Analysis of Different λ

In Equation (12), our hyperparameter λ is set to 10 and the same value is used in all experiments. Therefore, it is necessary to compare the metrics of fused images with different λ . As shown in Table 8, the best experimental results can be achieved when λ = 10.

5. Conclusions

In this work, we propose a novel Vision Transformer-based fusion network (SFPFusion) which combines wavelet-guided pooling and the super feature attention mechanism to obtain effective global features while enhancing detail information. Specifically, the super feature attention mechanism decomposes ordinary global attention into the product of sparse correlation maps and low-dimensional attention, which can capture more long-range dependencies. Moreover, combined with the embedded wavelet-guided pooling layer, multi-scale base information and detail textures are well enhanced after extracting global features. We pass the preserved detail features (vertical, horizontal, and diagonal detail textures) to the unpooling layer, which can accurately recover the structure and texture information of the images. In addition, the proposed fusion network is also simple yet efficient by adopting the simple l 1 -norm and softmax-like fusion strategy. The experiments on three public available datasets demonstrate the effectiveness of our proposed method in terms of qualitative results and quantitative evaluation. Ablation experiments also confirm the role of different components in our fusion network. In the future, texture and color information will play an important role in image content recognition. As a result, infrared and visible image fusion can be widely used in the pretreatment of the image retrieval system [59], etc.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/s23187870/s1, Section S1. Super Feature Attention; Section S2. Comparative Experiments. Figure S1: Qualitative comparison of our method with 8 state-of-the-art methods on the MSRS dataset; Figure S2: Qualitative comparison of our method with 8 state-of-the-art methods on the TNO dataset.

Author Contributions

Conceptualization, C.C. and X.S.; Methodology, H.L., Y.X. and X.S.; Software, Y.X.; Validation, H.L.; Investigation, X.S.; Writing—original draft, Y.X.; Writing—review and editing, H.L. and C.C.; Visualization, C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The implementation of this work can be available at https://github.com/draymondbiao/SFPFusion.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (62202205); the National Social Science Foundation of China (21&ZD166); the Natural Science Foundation of Jiangsu Province, China (BK20221535); and the Fundamental Research Funds for the Central Universities (JUSRP123030).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guo, R.; Li, D.; Han, Y. Deep multi-scale and multi-modal fusion for 3D object detection. Pattern Recognit. Lett. 2021, 151, 236–242. [Google Scholar] [CrossRef]
  2. Gao, W.; Liao, G.; Ma, S.; Li, G.; Liang, Y.; Lin, W. Unified information fusion network for multi-modal RGB-D and RGB-T salient object detection. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 2091–2106. [Google Scholar] [CrossRef]
  3. Zhang, L.; Danelljan, M.; Gonzalez-Garcia, A.; Van De Weijer, J.; Shahbaz Khan, F. Multi-modal fusion for end-to-end rgb-t tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  4. Zhu, J.; Lai, S.; Chen, X.; Wang, D.; Lu, H. Visual prompt multi-modal tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 9516–9526. [Google Scholar]
  5. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  6. Tang, W.; He, F.; Liu, Y.; Duan, Y.; Si, T. DATFuse: Infrared and Visible Image Fusion via Dual Attention Transformer. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3159–3172. [Google Scholar] [CrossRef]
  7. Tang, L.; Yuan, J.; Zhang, H.; Jiang, X.; Ma, J. PIAFusion: A progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion 2022, 83, 79–92. [Google Scholar] [CrossRef]
  8. Kong, W.; Lei, Y.; Zhao, H. Adaptive fusion method of visible light and infrared images based on non-subsampled shearlet transform and fast non-negative matrix factorization. Infrared Phys. Technol. 2014, 67, 161–172. [Google Scholar] [CrossRef]
  9. Ma, J.; Zhou, Z.; Wang, B.; Zong, H. Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 2017, 82, 8–17. [Google Scholar] [CrossRef]
  10. Nencini, F.; Garzelli, A.; Baronti, S.; Alparone, L. Remote sensing image fusion using the curvelet transform. Inf. Fusion 2007, 8, 143–156. [Google Scholar] [CrossRef]
  11. Lewis, J.J.; O’Callaghan, R.J.; Nikolov, S.G.; Bull, D.R.; Canagarajah, N. Pixel-and region-based image fusion with complex wavelets. Inf. Fusion 2007, 8, 119–130. [Google Scholar] [CrossRef]
  12. Pati, Y.C.; Rezaiifar, R.; Krishnaprasad, P.S. Orthogonal matching pursuit: Recursive function approximation with applications to wavelet decomposition. In Proceedings of the 27th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 1–3 November 1993; pp. 40–44. [Google Scholar]
  13. Li, S.; Yin, H.; Fang, L. Group-sparse representation with dictionary learning for medical image denoising and fusion. IEEE Trans. Biomed. Eng. 2012, 59, 3450–3459. [Google Scholar] [CrossRef]
  14. Chen, C.; Li, Y.; Liu, W.; Huang, J. Image fusion with local spectral consistency and dynamic gradient sparsity. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2760–2765. [Google Scholar]
  15. Yang, B.; Li, S. Pixel-level image fusion with simultaneous orthogonal matching pursuit. Inf. Fusion 2012, 13, 10–19. [Google Scholar] [CrossRef]
  16. Zhou, Y.; Mayyas, A.; Omar, M.A. Principal component analysis-based image fusion routine with application to automotive stamping split detection. Res. Nondestruct. Eval. 2011, 22, 76–91. [Google Scholar] [CrossRef]
  17. Bavirisetti, D.P.; Xiao, G.; Liu, G. Multi-sensor image fusion based on fourth order partial differential equations. In Proceedings of the 2017 20th International Conference on Information Fusion (Fusion), Xi’an, China, 10–13 July 2017; pp. 1–9. [Google Scholar]
  18. Li, H.; Wu, X.J.; Kittler, J. Infrared and visible image fusion using a deep learning framework. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2705–2710. [Google Scholar]
  19. Gao, C.; Song, C.; Zhang, Y.; Qi, D.; Yu, Y. Improving the performance of infrared and visible image fusion based on latent low-rank representation nested with rolling guided image filtering. IEEE Access 2021, 9, 91462–91475. [Google Scholar] [CrossRef]
  20. Bhavana, D.; Kishore Kumar, K.; Ravi Tej, D. Infrared and visible image fusion using latent low rank technique for surveillance applications. Int. J. Speech Technol. 2022, 25, 551–560. [Google Scholar] [CrossRef]
  21. Li, H.; Wu, X.J.; Kittler, J. MDLatLRR: A novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 2020, 29, 4733–4746. [Google Scholar] [CrossRef]
  22. Li, H.; Wu, X.j.; Durrani, T.S. Infrared and visible image fusion with ResNet and zero-phase component analysis. Infrared Phys. Technol. 2019, 102, 103039. [Google Scholar] [CrossRef]
  23. Xu, H.; Gong, M.; Tian, X.; Huang, J.; Ma, J. CUFD: An encoder–decoder network for visible and infrared image fusion based on common and unique feature decomposition. Comput. Vis. Image Underst. 2022, 218, 103407. [Google Scholar] [CrossRef]
  24. Li, H.; Wu, X.J. DenseFuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 2018, 28, 2614–2623. [Google Scholar] [CrossRef]
  25. Li, H.; Wu, X.J.; Durrani, T. NestFuse: An infrared and visible image fusion architecture based on nest connection and spatial/channel attention models. IEEE Trans. Instrum. Meas. 2020, 69, 9645–9656. [Google Scholar] [CrossRef]
  26. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  27. Ma, J.; Xu, H.; Jiang, J.; Mei, X.; Zhang, X.P. DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 2020, 29, 4980–4995. [Google Scholar] [CrossRef] [PubMed]
  28. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  29. Wang, Z.; Chen, Y.; Shao, W.; Li, H.; Zhang, L. SwinFuse: A residual swin transformer fusion network for infrared and visible images. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  30. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  31. Zhang, Y.; Liu, Y.; Sun, P.; Yan, H.; Zhao, X.; Zhang, L. IFCNN: A general image fusion framework based on convolutional neural network. Inf. Fusion 2020, 54, 99–118. [Google Scholar] [CrossRef]
  32. Li, H.; Xu, T.; Wu, X.J.; Lu, J.; Kittler, J. LRRNet: A Novel Representation Learning Guided Fusion Network for Infrared and Visible Images. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 11040–11052. [Google Scholar] [CrossRef]
  33. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the Neural Information Processing Systems, Montreal, BC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  34. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  35. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  36. Vs, V.; Valanarasu, J.M.J.; Oza, P.; Patel, V.M. Image fusion transformer. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 3566–3570. [Google Scholar]
  37. Chen, J.; Ding, J.; Yu, Y.; Gong, W. THFuse: An infrared and visible image fusion network using transformer and hybrid feature extractor. Neurocomputing 2023, 527, 71–82. [Google Scholar] [CrossRef]
  38. Tang, W.; He, F.; Liu, Y. TCCFusion: An infrared and visible image fusion method based on transformer and cross correlation. Pattern Recognit. 2023, 137, 109295. [Google Scholar] [CrossRef]
  39. Yi, S.; Jiang, G.; Liu, X.; Li, J.; Chen, L. TCPMFNet: An infrared and visible image fusion network with composite auto encoder and transformer–convolutional parallel mixed fusion strategy. Infrared Phys. Technol. 2022, 127, 104405. [Google Scholar] [CrossRef]
  40. Yang, X.; Huo, H.; Wang, R.; Li, C.; Liu, X.; Li, J. DGLT-Fusion: A decoupled global–local infrared and visible image fusion transformer. Infrared Phys. Technol. 2023, 128, 104522. [Google Scholar] [CrossRef]
  41. Rao, D.; Xu, T.; Wu, X.J. Tgfuse: An infrared and visible image fusion approach based on transformer and generative adversarial network. arXiv 2023, arXiv:2201.10147. [Google Scholar] [CrossRef] [PubMed]
  42. Yao, T.; Pan, Y.; Li, Y.; Ngo, C.W.; Mei, T. Wave-vit: Unifying wavelet and transformers for visual representation learning. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–24 October 2022; pp. 328–345. [Google Scholar]
  43. Yoo, J.; Uh, Y.; Chun, S.; Kang, B.; Ha, J.W. Photorealistic style transfer via wavelet transforms. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9036–9045. [Google Scholar]
  44. Huang, H.; Zhou, X.; Cao, J.; He, R.; Tan, T. Vision Transformer with Super Token Sampling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22690–22699. [Google Scholar]
  45. Shaw, P.; Uszkoreit, J.; Vaswani, A. Self-attention with relative position representations. arXiv 2018, arXiv:1803.02155. [Google Scholar]
  46. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  47. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  48. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A visible-infrared paired dataset for low-light vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 3496–3504. [Google Scholar]
  49. Toet, A. The TNO multiband image data collection. Data Brief 2017, 15, 249–251. [Google Scholar] [CrossRef] [PubMed]
  50. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 502–518. [Google Scholar] [CrossRef]
  51. Zhao, Z.; Xu, S.; Zhang, J.; Liang, C.; Zhang, C.; Liu, J. Efficient and model-based infrared and visible image fusion via algorithm unrolling. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1186–1196. [Google Scholar] [CrossRef]
  52. Cheng, C.; Xu, T.; Wu, X.J. MUFusion: A general unsupervised image fusion network based on memory unit. Inf. Fusion 2023, 92, 80–92. [Google Scholar] [CrossRef]
  53. Li, B.; Lu, J.; Liu, Z.; Shao, Z.; Li, C.; Du, Y.; Huang, J. AEFusion: A multi-scale fusion network combining Axial attention and Entropy feature Aggregation for infrared and visible images. Appl. Soft Comput. 2023, 132, 109857. [Google Scholar] [CrossRef]
  54. Rao, Y.J. In-fibre Bragg grating sensors. Meas. Sci. Technol. 1997, 8, 355. [Google Scholar] [CrossRef]
  55. Han, Y.; Cai, Y.; Cao, Y.; Xu, X. A new image fusion performance metric based on visual information fidelity. Inf. Fusion 2013, 14, 127–135. [Google Scholar] [CrossRef]
  56. Cui, G.; Feng, H.; Xu, Z.; Li, Q.; Chen, Y. Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Opt. Commun. 2015, 341, 199–209. [Google Scholar] [CrossRef]
  57. Roberts, J.W.; Van Aardt, J.A.; Ahmed, F.B. Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2008, 2, 023522. [Google Scholar]
  58. Xydeas, C.S.; Petrovic, V. Objective image fusion performance measure. Electron. Lett. 2000, 36, 308–309. [Google Scholar] [CrossRef]
  59. Kelishadrokhi, M.K.; Ghattaei, M.; Fekri-Ershad, S. Innovative local texture descriptor in joint of human-based color features for content-based image retrieval. Signal Image Video Process. 2023, 17, 4009–4017. [Google Scholar] [CrossRef]
Figure 1. Visualization results of different Transformer-based methods. From left to right: infrared image, visible image, Swin Transformer-based fusion model (SwinFuse), and the proposed SFPFusion.
Figure 1. Visualization results of different Transformer-based methods. From left to right: infrared image, visible image, Swin Transformer-based fusion model (SwinFuse), and the proposed SFPFusion.
Sensors 23 07870 g001
Figure 2. The proposed module combining super feature attention and wavelet-guided pooling. L represents the number of times attention extracts features. The features F S F extracted with super feature attention are passed to the embedded pooling layer to extract deep multi-scale features ( F L L ) and detail textures ( F V D , F H D , and F D D ). In the reconstruction, the components are aggregated by the wavelet-guided unpooling.
Figure 2. The proposed module combining super feature attention and wavelet-guided pooling. L represents the number of times attention extracts features. The features F S F extracted with super feature attention are passed to the embedded pooling layer to extract deep multi-scale features ( F L L ) and detail textures ( F V D , F H D , and F D D ). In the reconstruction, the components are aggregated by the wavelet-guided unpooling.
Sensors 23 07870 g002
Figure 3. Visualization results of feature maps within the S F P 1 . F S F represents the global features extracted by the attention mechanism. Then, the global multi-scale information is obtained by the multi-scale kernels L L , V D , H D , and D D to capture vertical, horizontal, and diagonal texture details.
Figure 3. Visualization results of feature maps within the S F P 1 . F S F represents the global features extracted by the attention mechanism. Then, the global multi-scale information is obtained by the multi-scale kernels L L , V D , H D , and D D to capture vertical, horizontal, and diagonal texture details.
Sensors 23 07870 g003
Figure 4. The framework of training process. In this process, we train an auto-encoder network without fusion strategy.
Figure 4. The framework of training process. In this process, we train an auto-encoder network without fusion strategy.
Sensors 23 07870 g004
Figure 5. Qualitative comparison of our method with 8 state-of-the-art methods on the MSRS dataset.
Figure 5. Qualitative comparison of our method with 8 state-of-the-art methods on the MSRS dataset.
Sensors 23 07870 g005
Figure 6. Qualitative comparison of our method with 8 state-of-the-art methods on the LLVIP dataset.
Figure 6. Qualitative comparison of our method with 8 state-of-the-art methods on the LLVIP dataset.
Sensors 23 07870 g006
Figure 7. Qualitative comparison of our method with 8 state-of-the-art methods on the TNO dataset.
Figure 7. Qualitative comparison of our method with 8 state-of-the-art methods on the TNO dataset.
Sensors 23 07870 g007
Figure 8. Visualized results of ablation studies. From (c) to (k): fused results of our method, fused results of using original Super Token Transformer, fused results of using original Vision Transformer, fused results of using CNN with pooling, without W a P 1 , global features using average strategy, global features using addition strategy and detail features with different directions using average strategy.
Figure 8. Visualized results of ablation studies. From (c) to (k): fused results of our method, fused results of using original Super Token Transformer, fused results of using original Vision Transformer, fused results of using CNN with pooling, without W a P 1 , global features using average strategy, global features using addition strategy and detail features with different directions using average strategy.
Sensors 23 07870 g008
Figure 9. Visualization results of feature maps within the S F P 1 .
Figure 9. Visualization results of feature maps within the S F P 1 .
Sensors 23 07870 g009
Table 1. Network architecture of feature extraction. Input and Output denote the number of channels in the corresponding feature maps.
Table 1. Network architecture of feature extraction. Input and Output denote the number of channels in the corresponding feature maps.
StageBlockLayer NameKernelInputOutputSizeActivation
ExtractionStage1 CB 1 Layer1 1 × 1 13 224 × 224 ReLU
Layer2 3 × 3 364 224 × 224 ReLU
Layer3 3 × 3 6464 224 × 224 ReLU
WaP 1 Pooling 2 × 2 6464 112 × 112
Stage2 SFP 1 CPE 3 × 3 , 64 , 64 grid 8 , heads 1 2 × 2 , 64 , 64 3 × 3 , 64 , 128 × 3 112 × 112
SFA 112 × 112
Pooling 56 × 56
PM 56 × 56
Stage3 SFP 2 CPE 3 × 3 , 128 , 128 grid 4 , heads 2 2 × 2 , 128 , 128 3 × 3 , 128 , 320 × 5 56 × 56
SFA 56 × 56
Pooling 28 × 28
PM 28 × 28
Stage4 CB 2 Layer1 3 × 3 320512 28 × 28 ReLU
Table 2. Network architecture of feature reconstruction. Input and Output denote the number of channels in the corresponding feature maps.
Table 2. Network architecture of feature reconstruction. Input and Output denote the number of channels in the corresponding feature maps.
BlocksLayersKernelInputOutputActivation
Reconstruction CB 3 Layer1 3 × 3 512256ReLU
CB 4 Layer1 3 × 3 256256ReLU
Layer2 3 × 3 256256ReLU
Layer3 3 × 3 256256ReLU
Layer4 3 × 3 256128ReLU
CB 5 Layer1 3 × 3 128128ReLU
Layer2 3 × 3 12864ReLU
CB 6 Layer1 3 × 3 6464ReLU
Layer2 3 × 3 641Sigmoid
Table 3. Quantitative results on 361 image pairs from the MSRS dataset. (Bold: best, red: second best, blue: third best).
Table 3. Quantitative results on 361 image pairs from the MSRS dataset. (Bold: best, red: second best, blue: third best).
MethodsSDVIFAG Q abf EN
DenseFuse [24]7.42370.69992.08730.35725.9340
FusionGAN [26]7.17580.86923.11930.21105.9937
SwinFuse [29]4.92460.41021.96730.17204.4521
U2Fusion [50]6.82170.58632.06940.29725.5515
AUIF [51]5.26220.39811.82380.19704.6460
CUFD [23]7.63840.64882.90030.43976.0652
MUFusion [52]6.92330.60863.14740.41105.9682
AEFusion [53]8.21040.85482.69680.42396.5374
Ours8.26500.92143.89500.53616.5636
Table 4. Quantitative results on 389 image pairs from the LLVIP dataset. (Bold: best, red: second best, blue: third best).
Table 4. Quantitative results on 389 image pairs from the LLVIP dataset. (Bold: best, red: second best, blue: third best).
MethodsSDVIFAG Q abf EN
DenseFuse9.29630.75032.67140.34586.8287
FusionGAN10.08231.02632.17060.27657.1741
SwinFuse7.54690.62902.95800.33926.0825
U2Fusion9.42560.72122.36850.34256.7588
AUIF7.54330.58772.87900.33486.1555
CUFD9.17010.71872.51980.32646.8448
MUFusion8.74520.78753.54120.41286.9242
AEFusion9.84000.63022.03970.14777.2764
Ours9.81030.96714.73530.60887.3006
Table 5. Quantitative results on 16 image pairs from the TNO dataset. (Bold: best, red: second best, blue: third best).
Table 5. Quantitative results on 16 image pairs from the TNO dataset. (Bold: best, red: second best, blue: third best).
MethodsSDVIFAG Q abf EN
DenseFuse9.22030.73493.88040.44656.8256
FusionGAN8.12340.61972.81200.22606.4629
SwinFuse9.26330.79825.59860.45426.9484
U2Fusion9.38690.72005.44560.46536.9395
AUIF9.28050.74825.28200.45137.0402
CUFD9.41360.87814.51780.37827.0743
MUFusion9.53790.78515.57560.38187.3032
AEFusion9.46550.78033.41140.30747.0716
Ours9.44000.80216.23000.48027.1578
Table 6. The average inference time (unit: second) on 361 pairs of images from MSRS dataset. (Bold: best).
Table 6. The average inference time (unit: second) on 361 pairs of images from MSRS dataset. (Bold: best).
MethodDenseFuseFusionGANSwinFuseU2FusionAUIFCUFDMUFusionAEFusionOurs
Inference time0.18130.06770.25920.03420.115872.61570.70450.22440.0739
Table 7. The average values of the five objective metrics obtained with different ablation studies on TNO dataset. (Bold: best, red: second best).
Table 7. The average values of the five objective metrics obtained with different ablation studies on TNO dataset. (Bold: best, red: second best).
StrategiesSDVIFAG Q abf EN
Ours9.44000.80216.23000.48027.1578
Original-SFT9.41010.84516.15720.46967.0922
Original-ViT8.90650.69715.66910.40406.6331
CNN-Pooling8.90910.71005.83720.46526.7445
W/O W a P 1 9.30700.77135.96070.47126.9916
One-SFP9.31450.79356.00230.46977.0172
GF-Add10.11520.71164.41280.37146.7826
GF-Avg9.19150.78846.51820.47727.0576
Detail-Avg9.15560.68123.33290.38096.9184
Table 8. The average values of the five objective metrics obtained with different λ on TNO dataset. (Bold: best, red: second best).
Table 8. The average values of the five objective metrics obtained with different λ on TNO dataset. (Bold: best, red: second best).
λ SDVIFAG Q abf EN
Ours ( λ = 10)9.44000.80216.23000.48027.1578
λ = 19.59010.79646.22200.47687.2029
λ = 1009.40010.79996.17620.48207.1582
λ = 10009.46440.79076.06820.48007.1623
λ = 10,0009.55370.80106.10700.47637.1716
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, H.; Xiao, Y.; Cheng, C.; Song, X. SFPFusion: An Improved Vision Transformer Combining Super Feature Attention and Wavelet-Guided Pooling for Infrared and Visible Images Fusion. Sensors 2023, 23, 7870. https://doi.org/10.3390/s23187870

AMA Style

Li H, Xiao Y, Cheng C, Song X. SFPFusion: An Improved Vision Transformer Combining Super Feature Attention and Wavelet-Guided Pooling for Infrared and Visible Images Fusion. Sensors. 2023; 23(18):7870. https://doi.org/10.3390/s23187870

Chicago/Turabian Style

Li, Hui, Yongbiao Xiao, Chunyang Cheng, and Xiaoning Song. 2023. "SFPFusion: An Improved Vision Transformer Combining Super Feature Attention and Wavelet-Guided Pooling for Infrared and Visible Images Fusion" Sensors 23, no. 18: 7870. https://doi.org/10.3390/s23187870

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop