Next Article in Journal
Geometric Configuration Design and Fast Imaging for Multistatic Forward-Looking SAR Based on Wavenumber Spectrum Formation Approach
Next Article in Special Issue
Guided Local Feature Matching with Transformer
Previous Article in Journal
Current Status and Challenges of BDS Satellite Precise Orbit Products: From a View of Independent SLR Validation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An ISAR and Visible Image Fusion Algorithm Based on Adaptive Guided Multi-Layer Side Window Box Filter Decomposition

1
School of Physics, Xidian University, Xi’an 710071, China
2
School of Electronics and Information Engineering, Wuxi University, Wuxi 214105, China
3
Computer Science and Engineering Group, Indian Institute of Information Technology, Sri City 441108, India
4
Science and Technology on Complex System Control and Intelligent Agent Cooperation Laboratory, Beijing Mechano-Electronic Engineering Institute, Beijing 100074, China
5
Department of Basic, Air Force Engineering University, Xi’an 710100, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(11), 2784; https://doi.org/10.3390/rs15112784
Submission received: 5 April 2023 / Revised: 12 May 2023 / Accepted: 24 May 2023 / Published: 26 May 2023

Abstract

:
Traditional image fusion techniques generally use symmetrical methods to extract features from different sources of images. However, these conventional approaches do not resolve the information domain discrepancy from multiple sources, resulting in the incompleteness of fusion. To solve the problem, we propose an asymmetric decomposition method. Firstly, an information abundance discrimination method is used to sort images into detailed and coarse categories. Then, different decomposition methods are proposed to extract features at different scales. Next, different fusion strategies are adopted for different scale features, including sum fusion, variance-based transformation, integrated fusion, and energy-based fusion. Finally, the fusion result is obtained through summation, retaining vital features from both images. Eight fusion metrics and two datasets containing registered visible, ISAR, and infrared images were adopted to evaluate the performance of the proposed method. The experimental results demonstrate that the proposed asymmetric decomposition method could preserve more details than the symmetric one, and performed better in both objective and subjective evaluations compared with the fifteen state-of-the-art fusion methods. These findings can inspire researchers to consider a new asymmetric fusion framework that can adapt to the differences in information richness of the images, and promote the development of fusion technology.

1. Introduction

Advancements in technology and economics have increased the popularity of sensors with different characteristic features [1]. The trade-offs in the characteristic features of different sensors have necessitated the use of fusion techniques to generate a robust and informative image from multi-sensor source images [2]. The fused images have all the desired characteristics of the source images [3]. Therefore, fusion techniques have been used in various applications such as medical diagnosis, monitoring, security, geological mapping, and agriculture monitoring [4,5,6].
Image fusion has been successfully applied in various fields [7]. However, the fusion sources have mostly been static images such as infrared and visible images, which do not provide dynamic information about the scene. This limitation can be overcome by fusing inverse synthetic aperture radar (ISAR) images with visible images. The fusion of ISAR and visible images yields informative images with detailed and dynamic information from the visible and ISAR images, respectively. The advantages of the fusion of the ISAR and visible images facilitate its application in security surveillance, air traffic control, wildlife conservation, and intelligent assistance [8,9,10,11].
In recent decades, numerous image fusion methods have emerged. The conventional image fusion methods can generally be grouped into three categories [12], namely, spatial domain methods [13,14,15], transform domain methods [16,17,18] and deep-learning-based methods [19]. The spatial domain methods directly operate on the pixel values of the input images and process the image on a pixel-by-pixel basis [20]. These methods have a simple processing flow and strong interpretability, and are widely used in pixel-level image fusion [21]. For example, Wan et al. [22] merged the images using the most significant features of a sparse matrix, which were obtained through the robust principal component analysis. Mitianoudis et al. [14] transformed the images using independent component analysis (ICA) and topographic independent component analysis. Fusion results were obtained in the transform domain using novel pixel-based or region-based rules. Yin et al. [23] proposed a dual-kernel side window box filter decomposition method and a saliency-based fusion rule for image fusion. Smadi et al. [24] adopted side window filtering employed on several filters to decompose and fuse the images. Yan et al. [25] decomposed the images using an edge-preserving filter and refined the final fused image with a guided filter. Zou [26] designed a multiscale decomposition method based on a guided filter and a side window box filter. The spatial-domain methods usually use a symmetric structure to design fusion strategies. However, in cases where there is a large difference in information intensity between the input images, weak signals are often overwhelmed by strong signals, resulting in the loss of some critical details.
The transform domain methods transform the images into another subspace where the features are classified and considered globally [27]. These methods are usually established on some dictionaries or models and can retain more meaningful information [17]. For instance, Liu et al. [27] represented the different components in the original images using sparse coefficients. Zhang et al. [28] represented the images through a multi-task robust sparse representation (MRSR) model, and were able to process the unregistered images. Liu et al. [29] decomposed the images using nonsubsampled shearlet transform (NSST) for pre-fusion. Qu et al. [30] represented the images in a nonsubsampled contour-let transform domain. Furthermore, a dual-pulse coupled neural network (PCNN) model was adopted to fuse the images. Panigrahy et al. [31] transformed images into the NSST domain and fused the images with adaptive dual channel PCNN. Indhumathi et al. [32] transformed the images with empirical mode decomposition (EMD) and then fused them with PCNN. Similarly to spatial-domain methods, transform-domain methods often use the same dictionary or model to extract feature information from images. These transformations use non-local or global information. However, this will lead to the loss of some important but globally less significant information, such as dim targets in infrared images.
Deep-learning-based methods have recently been introduced into the field of image fusion. These methods employ deep features to guide the fusion of two images [7,33]. Most of the deep-learning-based fusion strategies and generation methods report strong scene adaptability [34,35,36]. However, due to down-sampling in the network, the fusion results of deep-learning-based methods are usually blurred.
Although most of the conventional approaches have achieved satisfactory results, these methods fuse the two different source images in the same way. In the ISAR and visible image fusion, objects in the captured scene are mostly stationary or move slowly relative to the radar. As a result, ISAR images have large black backgrounds and many weak signals, while visible images are information-rich. This significant information difference makes conventional symmetric approaches inadequate for processing these images. These approaches will result in the mismatching of feature layers and affect the final fusion performance. For example, in [5], the same decomposition method is applied to both source images. Hence, significant information from these images may be lost during the fusion process. Similarly, in the sparse representation methods [17,27], the same over-complete dictionary is applied to different images, resulting in some features being ignored.
Hence, most conventional approaches cannot deal suitably with the fusion of different images containing quite different information content. Further research is needed on how to balance the differences in information between different source images to preserve as many details as possible. To the best of our knowledge, there are few studies considering the inequality information between two input images. This study aims to balance the information difference between different source images and preserve image details as much as possible in the fusion results. To achieve this objective, we analyzed heterogeneous image data, including ISAR, visible, and infrared data. We found that although the data acquisition methods are different, they describe different information about the same scene. There are some underlying connections between these pieces of information. Therefore, we hypothesized that these pieces of information can be guided and obtained through other source images. Under this assumption and motivated by the work of [5,25], we innovatively designed an asymmetric decomposition method, aiming to use strong signals to guide weak signals. Through the operation, we can obtain potential components from weak signals and enhance them, ensuring that weak signals can be fully preserved in the fusion result.
Based on the idea above, under the assumption that the images have been registered at the pixel level, the current research proposes a novel fusion method named the adaptive guided multi-layer side window box filter decomposition (AMLSW-GS) method. Firstly, an information abundance discrimination method is proposed to categorize the input images as detailed and coarse images. Secondly, a novel asymmetric decomposition framework is proposed to decompose the detailed and coarse images, respectively. The detailed image is decomposed into three layers constituting small details, long edges, and base energy. Similarly, the coarse image is used to yield an additional layer containing the strengthening information. Thirdly, a pre-fusion step is adopted to fuse the layers containing strengthening information and the layers containing the small details from the coarse image. Furthermore, three different fusion strategies are adopted to fuse the different layers, and the sub-fusion results are obtained. Finally, the fusion result is obtained as a summation of different sub-fusion results.
The main contributions of the current work are as follows:
(1)
We innovatively propose an asymmetric decomposition method to avoid losing vital information and balance the features of input images.
(2)
We design an information abundance measurement method to justify the amount of information in input images. This measurement estimates the information variance and edge details in the input images.
(3)
We propose three different fusion strategies to use different feature information for different layers. These strategies are designed according to the characteristics of different layers.
(4)
We publicize a pixel-to-pixel co-registered dataset containing visible and ISAR images. To the best of our knowledge, this is the first dataset related to the fusion of visible and ISAR images.
In summary, to address the problem of weak signal loss due to uneven fusion information intensity, under the assumption that the images are already registered at the pixel level and information can be guided, this study proposes an asymmetric decomposition method to enhance the weak signal and ensure the completeness of information in the fusion result. This article is organized as follows. In Section 2, the proposed method is presented in detail. In Section 3, the dataset is introduced and the experimental results on ISAR and VIS images are presented. In Section 4, some characteristics of the proposed method are discussed. In Section 5, we conclude this paper.

2. Proposed Method

In this paper, an ISAR and VIS fusion method based on AMLSW-GS decomposition is proposed. As shown in Figure 1, the ISAR and VIS images are decomposed using the AMLSW-GS method into layers at different scales. For the multi-scale layers, different fusion strategies are adopted to obtain the fusion results at different scales. Finally, the fusion results at different scales are summed together to yield the final fusion result.

2.1. Information Abundance Discrimination

To estimate the information complexity of different input images, a method based on spatial frequency and variance (SF-Var) is proposed. The SF-Var is a global evaluation parameter, which is computed as the weighted average of spatial frequency (SF) and global variance. S F is used to describe the total amount of edges in the images and is computed as
R F = 1 M × N x = 1 M y = 2 N I x , y I x , y 1 2
C F = 1 M × N x = 1 M y = 2 N I x , y I x 1 , y 2
S F = C F 2 + R F 2
where M and N denote the dimensions of image I, and R F and S F are the row and column frequencies, respectively.
The variance (Var) is a statistical indicator to evaluate the strength of the edges of an image and is computed as
M e a n = 1 M × N x = 1 M y = 1 N I x , y
V a r = 1 M × N x = 1 M y = 1 N I x , y M e a n 2
The complexity of the image content is judged based on two types of edge information: edge quantity information derived from SF, and edge intensity information derived from variance. The complexity is then calculated using the weighted average method as
S V = S F × V a r S F + V a r
where S V represents the evaluation index of the SF-Var method. The larger the S V , the richer the information.

2.2. Decomposition Method

In the conventional decomposition methods, the images are often decomposed in the same manner irrespective of their differences. As the information content of different images is quite different, the conventional methods may lose the image information. In this section, the AMLSW-GS decomposition method will be introduced. Images with different levels of content complexity are decomposed into layers of different scales to avoid information loss. The ISAR and VIS images are decomposed into multi-scale layers using the AMLSW-GS approach. The AMLSW-GS decomposition method comprises two steps, namely, multi-layer Gaussian side window filtering (MLSW) and multi-layer Gaussian side window filtering with guided strengthening (MLSW-GS).

2.2.1. Multi-Layer Gaussian Side Window Filtering

After the image information redundancy is judged, we can distinguish the complexity of the two input images. For the high-complexity image, we employ the multi-layer Gaussian side window filtering (MLSW) decomposition approach, as detailed in Figure 2.
It may be noted that there are two information extraction streams in the MLSW, a Gaussian high-frequency feature extraction stream and a side window high-frequency feature extraction stream. Both of these streams differ in their ability to extract detailed features.
The extraction process of the Gaussian high-frequency features is summarized as
I g i = G F ( I g ( i 1 ) )
where I g i is the i-th low frequency information of the Gaussian filter, i = 1, 2, …, n, I g 0 = I . n is the number of decomposition layers. GF() represents the Gaussian filter. In the experiments adopted in this study, the size of the filter window and the standard deviation are set to 3.
The process of side window high-frequency feature extraction is based on the side window filter [37]. The side window filter is an edge-preserving filter, which assumes the edge pixels to have a stronger correlation with the side pixels than the entire neighborhood. Hence, the reconstructed edges will not be too blurred as compared to the other traditional filters. In addition, the side window structure can also be applied to the other traditional filters such as the mean filter, median filter, Gaussian filter, etc. In this research, we adopted the box side window filter as the side window structure.
The process of side window high-frequency feature extraction, using the box side window filter, can be described as
I d i = S W F ( I d ( i 1 ) , r , i t e r )
where I d i is the i-th low-frequency information of the side window filter, i = 1, 2, …, n. I d 0 = I. n is the number of decomposition layers, S W F ( ) represents the side window filter [37], r represents the radius of the filter window, and i t e r represents the number of iterations. In our experiments, we adopted the recommended values given in [37]: r = 1 and i t e r = 7 .
Two of the above feature extraction streams yield two different background information images, namely, I g i and I d i . Both of these contain decreasing detail. The detail preservation layers D P i and the edge preservation layers E P i are then estimated as
D P i = I g ( i 1 ) I d i , i = 1 , 2 , . . . , n
E P i = I d i I g i , i = 1 , 2 , . . . , n
The base layer in this framework is the final layer of the Gaussian feature extraction stream I g n . It may be noted that in this study, n = 3.
Just as in the work in [5,25], edge-preserving filters are often used for image feature decomposition. They take advantage of the blurring characteristics of edge-preserving filters to remove some of the detailed information from the original image. Detailed information can be obtained by difference operation. The decomposition framework in this section is also based on such filter characteristics. However, two filters with different blur capabilities are used in this section to extract details, so that more detailed information can be extracted from images with different levels of blurring. However, as we know, when the same filter is applied to data with different intensities, the absolute strength of the difference values can also vary, which can cause some weak signals to be overwhelmed by strong signals. This is why we designed another enhanced decomposition framework.

2.2.2. Multi-Layer Gaussian Side Window Filter with Guided Strengthening

To make full use of the two images containing different information, and enhance the characterization of weak signals, we design a decomposition framework called the multi-layer Gaussian side window filter with guided strengthening (MLSW-GS). We apply the MLSW-GS to the low-complexity image which needs to be enhanced. The steps involved in MLSW-GS are summarized in Figure 3.
Based on the MLSW framework, we introduce a guided stream where an image with high-complexity information is used to guide the decomposition of an image with low-complexity information. The guiding process can be described as
S i = G u i d ( D P h i , D P l i )
where S i represents the i-th strengthening layer. D P h i represents the i-th detailed layer of MLSW, D P l d i represents the i-th detailed layer of the MLSW-GS, and i = 1, 2, 3, …. G u i d ( ) represents the guided filter [38]. D P h i contains complex information, and D P l d i contains little information. Through this guidance operation, we can relatively enhance the weak details in the low-complexity image layers.
The MLSW-GS yields a new layer called the strengthening layer, in addition to the three type layers obtained using MLSW.

2.3. Fusion Strategies

The AMLSW-GS framework yields different detailed layers for different images. For the image with high-complexity information, the MLSW decomposition method results in detailed preservation layers (DP), edge preservation layers (EP), and a base energy layer (B). For the image with low-complexity information, the MLSW-GS decomposition method yields four layers, namely, the strengthening layers (S), DP layers, EP layers, and base layer.

2.3.1. Fusion Strategy for Strengthening Layers

As the number of layers of the two images is different, pre-fusion processing of the strengthening layers is necessary. The strengthening layers have some enhanced information and the main image structure is similar to that of DP layers. Hence, we carry out a pre-fusion for these two layers; the fusion strategy can be described as
S F L i = D P l i + S i
where S F L i represents the i-th strengthening fusion layer, D P l i represents the i-th detail preservation layer of low complexity, and S i represents the i-th strengthening layer.

2.3.2. Fusion Strategy for Detail Preservation Layers

Based on Equation (12), the strengthening fusion layers (SFL) were obtained. In detailed preservation layers, small structures and edges are mainly preserved. Therefore, the more small structures that are preserved, the better the fusion result. Based on this principle, the spatial frequency and local variance (SF-LV) fusion strategy is adopted to fuse the SFL layers and the DP layers of the highly complex source. The fusion strategy can be described as
F D P i = D P h i , i f S F ( S F L i ) < S F ( D P h i ) a n d L V ( S F L i ) < L V ( D P h i ) S F L i o t h e r w i s e
where D P h i represents the i-th detail preservation layer of a high-complexity source, S F L i represents the i-th strengthening fusion layer, and S F ( · ) represents the local spatial frequency, which can be calculated using Equation (1) to Equation (3). The M and N in Equations (1) and (2) represent the size of the local area in the image. L V ( · ) represents local variance and is calculated as
L V ( x , y ) = r = P 2 P 2 s = Q 2 Q 2 I ( x + r , y + s ) μ 2 P × Q
where I represents the image, x and y represent the center position of the local image patch, r and s represent the pixel index in the image patch, P and Q represent the size of the image patch, and μ represents the mean of the pixels in the image patch.
The SF can describe the number of edges in the local area, while the LV can describe the distribution of pixels in the local area. This fusion strategy can describe the local attributes, which is important for locating small structures.

2.3.3. Fusion Strategy for Edge Preservation Layers

In edge preservation layers, long edges with strong intensity are mainly preserved. Therefore, the more accurate the edge location is, the better the fusion result. Based on this principle, a fusion method, named the morphological symmetric phase consistency motivated pulse coupled neural network (SPhCM-PCNN), is proposed to fuse the preservation layers.

Symmetric Phase Consistency

Classic phase consistency (PhC) [39] is an effective method for feature extraction, focused on edge features. Compared with other edge extraction methods, PhC methods have advantages in terms of edge localization and weak edge extraction. However, the approach is not effective when the frequency component of the signal is not rich. In addition, for some sharp edges, such as step edges, the aftershock phenomenon [40] will occur in the boundaries.
Therefore, to avoid the limitations of the PhC method, symmetry phase consistency (SPhC) is adopted. Based on the work of [41,42], symmetrical phase consistency makes use of the symmetry of the signal. Symmetry features can describe the signal features where all of the frequency components are aligned in symmetry and in phase. In the image, the position near the edge often shows strong symmetry or anti-symmetry. By filtering the symmetrical phase in the image, the edge position of the image can be located more quickly and accurately. As only the symmetrical phase is considered, the SPhC method can suppress the aftershock effect to a certain extent.
Classic phase consistency [40] is described by the Fourier series expansion and is calculated as
I ( x ) = n A n cos ( ϕ n ( x ) ) P h C ( x ) = max ϕ ¯ ( x ) [ 0 , 2 π ] n A n cos ( ϕ n ( x ) ϕ ¯ ( x ) ) n A n
where the I ( x ) is the image signal, and A n and ϕ n ( x ) represent the coefficients and local phase of the n-th Fourier component at location x, respectively. ϕ ¯ (x) is the weighted mean local phase of all the Fourier components at point x.
Equation (15) is hard to calculate. To compute the PhC features, local energy is considered to simplify the process. The local energy is defined as
E ( x ) = I ( x ) 2 + H ( x ) 2
where H ( x ) is the Hilbert transform of the I(x). H(x) can generate the 90-degree phase shift of the I(x).
For the one-dimension signal f ( x ) , supposing that M n e and N n o are even symmetric filters and odd symmetric filters in the scale n, the odd symmetric filters are 90-degree phase shift filters of the even symmetric filters. We can calculate the strength coefficients of the odd and even filters at position x as
[ e n ( x ) , o n ( x ) ] = [ f ( x ) * M n e , f ( x ) * N n o ]
where e n ( x ) and o n ( x ) represent the strength coefficients of the even components and odd components, respectively. * is a convolution operation.
Furthermore, we compute the local amplitude A n ( x ) at the scale n as
A n ( x ) = e n ( x ) 2 + o n ( x ) 2
The local energy E ( x ) at position x is calculated as
E ( x ) = e n ( x ) 2 + o n ( x ) 2
The phase consistency of the signal f ( x ) is computed as
P h C ( x ) = E ( x ) n A n ( x ) + ϵ
where ϵ is a small positive constant to avoid a zero denominator.
For the two-dimensional images, the PhC value is calculated in different directions, and results in all directions being summarized to obtain the PhC value. The phase consistency of the image can be described as
P h C 2 D ( x ) = j E θ j ( x ) n j A n θ j ( x ) + ϵ
where θ j is the j-th direction, and θ j = j π / m , j = 0 , 1 , 2 . . . , m 1 ; here, m = 6.
The phase consistency can locate the edges accurately, but an aftershock phenomenon exists in the phase consistency result [43]. In order to solve this problem, the symmetry of the phase is researched.
Symmetry is a basic element of image features. The edges are often at the most symmetric or asymmetric points in all the frequencies. Hence, we can use these special phases to obtain the image features. In the symmetry phase consistency, we only consider four special phases, π / 2 , 0, π / 2 , and π . These four phases are symmetric and asymmetric at all scales. The π / 2 and π / 2 are the symmetric phase and the 0 and π are the asymmetric phase. Hence, the symmetric phase consistency can be described as
S P h C ( x ) = S u m E ( x ) S u m A ( x ) + ϵ = k , q e n ( k ) , θ ( q ) ( x ) 2 + k , q o n ( k ) , θ ( q ) ( x ) 2 k , q e n ( k ) , θ ( q ) ( x ) 2 + o n ( k ) , θ ( q ) ( x ) 2 + ϵ ϕ n ( k ) , θ ( q ) π / 2 , 0 , π / 2 , π
where SPhC(x) represents the symmetric phase consistency at point x. S u m E ( x ) and S u m A ( x ) represent the sum of local energy and the sum of local amplitude in ϕ n ( k ) , θ ( q ) . ϕ n ( k ) , θ ( q ) is the symmetric phase, n ( k ) ) and θ ( q ) are the scale and the orientation, and k , q = 0 , 1 , 2 . . . . e n ( k ) , θ ( q ) ( x ) and o n ( k ) , θ ( q ) ( x ) represent the strength coefficients of the even components and odd components at scale n ( k ) and orientation θ ( q ) , respectively.
There are some noises that may affect the estimation of the fusion weights. Therefore, we add a morphological operation before the symmetric phase consistency. Furthermore, we obtain the morphological symmetric phase consistency result as
S P h C M ( x ) = t = 1 N ω t · ( S P h C ( x ) S E t S P h C ( x ) S E t )
where S E t donates a basic structure element at t scale, t = 1, 2, 3, …, N. Here, N = 3. ⊕ and ⊖ represent the morphological dilation and erosion operations. ω t represents the t-scale weight. It can be calculated by
ω t = 1 2 t + 1

Morphological Symmetric Phase Consistency Motivated Pulse Coupled Neural Network

In order to locate the edge accurately, the SPhCM-PCNN transforms the classical PCNN [44] model to the phase consistency domain. The SPhCM-PCNN framework is summarized in Figure 4
The main fusion framework of SPhCM-PCNN is a two-channel PCNN model. It can be calculated as
F i j 1 ( k ) = S i j 1 ( k )
F i j 2 ( k ) = S i j 2 ( k )
L i j ( k ) = 1 , i f r , t S Y r t ( k 1 ) > 0 0 , o t h e r w i s e
U i j ( k ) = m a x F i , j 1 ( k ) 1 + β i , j 1 L i j ( k ) , F i j 2 ( k ) 1 + β i j 2 L i j ( k )
Y i j ( k ) = 1 , i f U i j ( k ) > = θ i j ( k 1 ) 0 , o t h e r w i s e
θ i j ( k ) = θ i j ( k 1 ) Δ + V θ Y i j ( k )
T i j ( k ) = k , i f U i j ( k ) > = θ i j ( k 1 ) T i j ( k 1 ) , o t h e r w i s e
where S i j 1 and S i j 2 represent the pixels of the two images at point (i, j). The source images are edge preservation layers from the ISAR and visible images. L i j represents the link parameter. β i j 1 and β i j 2 represent the link strength. F i j 1 and F i j 2 represent the input feedback. U i j is the output of the dual channels. θ i j is the threshold of the step function. Y i j ( k ) is the the k-th iteration output of PCNN. Δ is the value of the threshold decline, and V θ represents the learning rate of the threshold. T i j represents the criterion that determines the number of iterations. S represents the neighborhood of the pixel at position (i, j).
Based on the dual-channel PCNN above, we replace the β i j 1 and β i j 2 with the results of SPhCM from the two sources. Furthermore, we obtain the fusion map of the edge preservation layers as
F E P i = W E P h i × E P h i + ( 1 W E P h i ) × E P l i
where F E P i represents the i-th fusion result of the edge preservation layers. W E P h i is the output weight of the SPhCM-PCNN, which has been normalized. E P h i and E P l i are the edge preservation layers of the high-complexity source and the low-complexity source.

2.3.4. Fusion Strategy for Base Layers

The base layers of both the images contain significant energy but lack high-frequency details. Hence, we adopted energy-based methods to fuse the energy layers.
The energy-based methods are summarized as
F B ( x , y ) = B h ( x , y ) 2 + B l ( x , y ) 2 B h ( x , y ) + B l ( x , y )
where the F B ( x , y ) represents the fusion result of base layers at point (x,y). B h and B l are the base layer of the high and low complexity sources.
Finally, all the sub-fusion results are added to obtain the final fusion result:
F = i = 1 n ( F D P i + F E P i ) + F B
where n represents the number of decomposition layers. Here, n = 3.

3. Experimental Results

The details of our dataset are discussed in Section 3.1. The qualitative and quantitative comparisons of the proposed method with the other state-of-the-art methods are presented in Section 3.2. The ISAR and visible dataset used in this study were obtained from different sources. The visible images were collected from the Internet [45]. The ISAR images were generated through simulations from the visible images, the details of which are explained in Section 3.1. All experiments were performed using Matlab 2022b by MathWorks, the United States. The computer used for experiments was equipped with an Intel Core i9-12900H CPU, 16 GB of memory produced by Samsung South Korea, and a 512 GB solid-state drive produced by Toshiba Japan.

3.1. ISAR and Visible Dataset

The dataset that we published as a part of this research constitutes co-registered visible and simulated ISAR images. A few examples of the images from the dataset are shown in Figure 5.
All the visible images in the dataset were collected from the Internet [45]. The ISAR images were simulated based on the visible images. The simulation process for ISAR image generation is summarized below:
1.
The targets in the visible images are labeled pixel by pixel; therefore, we can accurately locate the target area.
2.
We set the background pixel values to 0 and only keep the pixels in the target area.
3.
The 3D models of the targets are established to estimate the distance.
4.
The Range Doppler algorithm [46] is adopted to synthesize the ISAR data. To reduce the amount of computation, we resize the images to 151 × 151 .
For the pixel-to-pixel registration, we aligned the visible images to the center of the simulated ISAR images by resizing them to 151 × 151 .
In our dataset, the scenes of visible images are mainly aerial scenes. The targets are moving in the air and near space, including planes, fighters, bombers, and airships. The dataset contains 100 different registered image pairs and 500 more visible images for future target recognition applications. In addition, all of the images containing targets are labeled in the PASCAL Visual Object Classes (PASCAL VOC) format. This provides the possibility to use the dataset for fusion detection tasks.

3.2. Comparison with State-of-the-Art Methods

To demonstrate the superiority of the proposed method, we compared it with fifteen state-of-the-art methods: adaptive sparse representation fusion (ASR) [17], multi-layer decomposition based on latent low-rank representation (MDLatLRR) [47], anisotropic diffusion, and the Karhunen–Loeve transform fusion (ADF) [48,49]; the fourth order partial differential equation fusion (FPDE) [50]; fast multi-scale structural patch decomposition-based fusion (FMMEF) [51]; night-vision context enhancement fusion (GFCE) [52]; visual saliency map and weighted least square optimization fusion (VSMWLS) [53]; Gradientlet [54]; latent low-rank representation fusion (LatLRR) [16]; pulse-coupled neural networks in non-subsampled shearlet transform domain fusion (NSCT-PAPCNN) [18]; spatial frequency-motivated pulse-coupled neural networks in non-subsampled contourlet transform domain fusion (NSCT-SF-PCNN) [30]; multi-scale transform and sparse representation fusion (MST-SR) [27]; guided-bilateral filter-based fusion in wavelet domain (DWT-BGF) [55]; VGG19-L1-Fusion [19]; and Image Fusion with ResNet and zero-phase component analysis (ResNet50-ZPCA) [56]. Eight metrics have been adopted to evaluate the quality of the fusion results, namely, the Mutual information (MI) [57], Nonlinear Correlation Information Entropy (NCIE) [58], Objective Evaluation of Fusion Performance ( Q a b / f ) [59], Fusion Quality Index ( Q S ) [60], Disimilarity-based quality metric ( Q Y ) [61], Average, Structural Similarity (SSIM), and Feature Mutual Information (FMI) [62,63].
For the subjective evaluation, four examples of image pairs with their fusion results, using different methods, are shown in Figure 6, Figure 7, Figure 8 and Figure 9.
As shown in Figure 6, the fusion source images contain planes with clouds in the background. There is some scene information in the visible source and some speed information of the plane in the ISAR source. The fusion result should contain both of these sources of information as much as possible. For the results of the MDLatLRR, FPDE, Gradientlet, LatLRR, DWT-BGF, VGG19-L1-Fusion, and ResNet50-ZPCA methods, the details of the ISAR information are blurred, which is marked by the red box. This means that these methods lose some high-frequency information. This phenomenon occurs because these methods use the same decomposition framework, and decompose the visible image and ISAR image at the same level. However, in this scene, the average brightness of the visible image is significantly higher than that of the ISAR image. This causes the details of the ISAR image to be treated as noise and submerged in the visible information. As a result, the ISAR information is lost. For the methods of the ASR, ADF, FMMEF, GFCE, VLMWLS, and the MST-SR methods, the area marked by the red box shows more ISAR information. However, these approaches ignore the information from the VIS source. This is because the scene brightness of the ISAR is significantly lower than that of the visible image. Therefore, when performing fusion operations, the brightness of the visible image changes significantly due to the influence of the ISAR brightness, causing signal strength loss. Benefiting from the asymmetric decomposition framework, our proposed method can preserve the ISAR details while maintaining visible information. Most of the fusion results suppress the background intensity of the visible scene, which is marked by the green box. Some of the results, such as those for the NSCT-SF-PCNN and MST-SR methods, have uneven background and block effects. Our proposed method and the NSCT-PAPCNN method preserve the visible source information. In summary, the proposed method yields a better performance as compared to the other methods in terms of fusion quality.
In Figure 7, the fusion details of ISAR and visible sources are marked by red boxes. The visible background area is marked by green boxes. The red-marked areas have been up-sampled by a factor of two and placed in the upper right corner of the images to show the fusion details. The green marked areas aim to show the fused scene brightness. Due to the dark background and lower detail of the ISAR image, the brightness of the green areas indicates the amount of energy retained in the visible scene. The clearer the red areas and the brighter the green areas, the better the fusion result. We can see that the red areas are blurry in the results of the ASR, MDLatLRR, ADF, FPDE, Gradientlet, MST-SR, DWT-BGF, VGG19-L1-Fusion, and ResNet50-ZPCA methods. Only the results of the NSCT-PAPCNN and MST-SR methods and the proposed methods have a well-lit scene in the green area. However, there are artifacts around the plane in the results of the NSCT-PAPCNN method, and the background is not uniform in the results of the MST-SR method. Therefore, compared with the other methods, our proposed method obtained satisfactory fusion results.
In Figure 8, the subjective evaluations are the same as those discussed in the previous case. We can find that the fusion details, marked by the red bounding box, are not clear in the results of the ASR, MDLATLRR, ADF, FPDE, Gradientlet, LatLRR, DWT-BGF, VGG19-L1-Fusion, and ResNet50-ZPCA methods. The details in the NSCT-SF-PCNN and MST-SR methods are clear, but the fusion background is not uniform. The fusion scene marked by green is brighter in the results of the FMMEF, GFCE, NSCT-PAPCNN and proposed methods. Hence, the proposed method and FMMEF, GFCE, and NSCT-PAPCNN methods yield good performance for this scene.
In Figure 9, the edges in the red area are not obvious in the ASR, MDLATLRR, ADF, FPDE, FMMEF, Gradientlet, LatLRR, DWT-BGF, VGG19-L1-Fusion, and ResNet50-ZPCA methods. This means that the information on the ISAR is not fully preserved. The background marked by green is not clear in almost all the contrast-based methods, except for the GFCE and MST-SR methods. The results for the GFCE, LATLRR, NSCT-PAPCNN, NSCT-SF-PCNN, and MST-SR methods have relatively more artifacts around the plane due to the low resolution and noise. Compared with other methods, our proposed method has more details and high contrast. It preserves the information of the two sources as much as possible in this scene. As is evident from Figure 6, Figure 7, Figure 8 and Figure 9, our proposed method has a better performance in most of the scenes in comparison with the state-of-the-art methods.
For the objective evaluation, eight metrics of fusion results are shown in Table 1, Table 2, Table 3 and Table 4, where the best values, the second-best values, and the third-best values are marked in red, blue, and green, respectively.
As is evident from Table 1, Table 2, Table 3 and Table 4, the proposed method achieves the best values for all the evaluation metrics. In some cases, even if not the best, the proposed approach achieves the second- or third-best position, as summarized in Table 3 and Table 4. Hence, in view of objective metrics, our proposed method achieves the best performance compared to the other state-of-the-art methods.
To demonstrate the adaptability of the proposed method to different scenes, we conducted experiments on the datasets containing dozens of representative scenes and evaluated them with the eight metrics above. The results for the different methods are shown in Figure 10.
In all the evaluation metrics, a higher value indicates better performance. As shown in Figure 10, our proposed method, which is denoted using a red line, consistently achieves a top-ranked performance across most fusion scenes. These results demonstrate that the proposed method has strong adaptability to different scenarios.

4. Discussion

To further analyze the characteristics and applicability of the proposed method, some experiments were carried out in this section.

4.1. Parameter Analysis of the Proposed Method

In this section, we discuss the impact of parameters on the performance of the method. In the proposed method, many parameters need to be set: the number of decompositions s, the size of the guided filter k, the number of iterations in PCNN L, the number of scales in SPhC l, and the number of scales in morphological operation t. After conducting experiments, we found that the parameters that have a significant impact on the performance of the method are the number of decompositions s and the number of scales in morphological operation t. Hence, we mainly discuss the impact of these two parameters on the performance of the method.
In our experiments, we used the method of controlling the variables to explore. Except for the two parameters being explored, the values of the other parameters were set as follows: k = 3 , L = 200 , and l = 4 . For convenience of analysis, we chose MI and Q a b / f as the metrics for evaluating the performance.
As shown in Figure 11, as the number of decompositions increases, both MI and Q a b / f show an increase followed by a decrease. The best number of decompositions is 3. This is mainly due to the fact that at lower decomposition levels, details can be better distinguished to improve the performance of the fusion result with the increase in the number of decompositions. However, when the number of decompositions is too large, weak details are overly enhanced, resulting in a decrease in the structural similarity of the final fusion result with the input images. The enhancement operations can lead to information imbalance.
As shown in Figure 12, while the number of scales increases, both MI and Q a b / f also show an increase followed by a decrease. The best number on the scale is 3. This is due to the fact that when the number of scales is small, the multi-scale morphological operation can well locate the position of the object compared to the edge, which improves the final fusion performance. However, when the number of scales is large, due to the characteristics of the morphological operators, there will be more artifacts in the fusion weights, leading to incorrect weight allocation in the areas around the edges, resulting in poorer fusion results.
In summary, the adaptability and the stability of the parameters in the proposed method are weak. Our method performs better with fewer decomposition layers and morphological scales. This is because an increase in the number of layers results in the excessive enhancement of weak signals, and increasing the number of morphological scales results in the inaccurate location of the edges. Both factors will lead to decreased performance. Hence, to further improve the stability of the parameters, we need to further study the internal mechanism of the enhancement, and design a more effective enhancement method to achieve better performance.
As analyzed above, we chose s = 3 and t = 3 under consideration of the effectiveness and efficiency of the method.

4.2. Ablation Analysis of the Structure

To verify the effectiveness of the proposed asymmetric method, we conducted some experimental studies in this section.
Based on the analysis of the proposed method, we believe that the factors that mainly affect the fusion performance are concentrated in two aspects: the asymmetric decomposition framework and the selection of fusion strategies. Hence, we analyzed these aspects in this section. The fusion strategy for the energy layer is a classic weighted fusion, so we do not intend to discuss this fusion strategy in this section.
In the experiments, three different structures are discussed and four metrics including MI, NCIE, Q a b / f , and SSIM are adopted. The proposed method is denoted as AMLSW + SP + SL, where AMLSW represents the asymmetric decomposition framework we proposed, SP represents the SPhCM-PCNN fusion strategy, and SL represents the SF-LV fusion strategy. Meanwhile, we use MLSW to represent the symmetric decomposition strategy, which replaces MLSW-GS in AMLSW with MLSW. AMLSW + SP represents the decomposition method being AMLSW, and the fusion strategies for detail preservation layers and edge preservation layers are SPhCM-PCNN. AMLSW + SL represents the decomposition method being AMLSW, and the fusion strategies for detail preservation layers and edge preservation layers are SF-LV. The experimental results are shown in Figure 13.
Through Figure 13, we can find that by replacing the decomposition method with a symmetric one, which is shown in Figure 13d,h, some potential information about detail is lost, causing signal distortion in certain areas. This indicates that our asymmetric decomposition structure has the ability to preserve weak signals compared to the symmetric structure. By replacing the detail preservation fusion strategy with SPhCM-PCNN, which is shown in Figure 13e,i, we can find that some small structures and details are lost. This indicates that SF-LV has a better fusion performance on small structures and details, and can better preserve the details of small structures compared to SPhCM-PCNN. By replacing the edge preservation fusion strategy with SF-LV, which is shown in Figure 13f,j, we can find that some long edges are blurred. This indicates that SPhCM-PCNN has a better fusion performance for long edge preservation due to its high precision in edge localization, and that SPhCM-PCNN can better preserve the details of long edges compared to SF-LV.
To further demonstrate the effectiveness of each part of the algorithm, we selected four metrics to assess the final fusion results. These metrics are MI, NCIE, Q a b / f , and SSIM. The experimental results are shown in Table 5, where the best metric values have been bolded.
In Table 5, we can draw the same conclusion as above: (1) The asymmetric decomposition structure has better fusion performance compared with the symmetric decomposition structure by comparing the metrics of MLSW + SP + SL and AMLSW + SP + SL. (2) The SF-LV fusion strategy has a better performance in detail preservation by comparing the metrics of AMLSW + SP and AMLSW + SP + SL. (3) The SPhCM-PCNN fusion strategy has a better performance in edge preservation by comparing the metrics of AMLSW + SL and AMLSW + SP + SL. We can find that all of the structures are necessary for good fusion performance.
As analyzed above, both in the objective and subjective evaluations, the asymmetric structure proposed in this method performs better than the symmetric one. This is because the asymmetric structure extracts more weak signals and effectively enhances them. The fusion strategies can also effectively preserve the details according to the characteristics of the corresponding layers. The main structures proposed in this study are both necessary and effective. The proposed method can inspire researchers to consider a new asymmetric fusion framework that can adapt to the differences in information richness of the images, and promote the development of fusion technology.

4.3. Generalization Analysis of the Method

To validate the adaptability and applicability of our proposed method on real-world data, we discuss the applicability of the proposed method in this section. We use publicly available infrared and visible fusion datasets to validate the proposed method. The infrared and visible fusion datasets that are widely used by many researchers include TNO [64], RoadScene [33], Multi-Spectral Road Scenarios(MSRS) [65], etc.
For the evaluation carried out in this study, we used the TNO datasets for infrared and visible image fusion. The TNO dataset contains images captured in various bands, including visual (390–700 nm), near-infrared (700–1000 nm), and longwave-infrared (8–12 μ m). The scenarios were captured at nighttime so that the advantage of infrared could be exploited. All of the images are registered at the pixel level with different multi-band camera systems.
To illustrate the performance of our method, we compared it with three other infrared and visible image fusion methods: MDLatLRR, NSCT-SF-PCNN and ResNet50-ZPCA. The fusion results are shown in Figure 14.
Through the results shown in Figure 14, we find that our proposed method can also be applied to infrared and visible image fusion. Due to the fact that infrared and visible light images are obtained from different types of signal sources, there is a relatively large information difference. The proposed method can deal with two different images with large information differences. Therefore, our method is also suitable for fusing infrared and visible images. The image data used for fusion are all from real scenes, which indicates that our method is capable of processing the image information of real scenes with information differences.
To further demonstrate the applicability of our method, we used four fusion metrics to evaluate the fusion performance: MI, NCIE, Q a b / f , and Q Y . The metric values of different fusion methods in different scenes are shown in Table 6, where the best values have been bolded.
As is evident in Table 6, the proposed method achieves the best values for these four metrics. Hence, in view of objective metrics, the proposed method is applicable to the task of infrared and visible image fusion. Our proposed method exhibits good adaptability.

4.4. Analysis of Time Complexity

Because the proposed fusion method has a relatively complex structure, it is necessary to conduct a time complexity analysis of the method. In this part, we analyze the time consumption for the proposed method.
The time complexity analysis can be performed by analyzing the number of operations required to complete the method as a function of the input size. The time complexity of the method is typically denoted as O ( f ( n ) ) , where n represents the input size.
In our study, the entire processing procedure can be divided into four stages: image complexity assessment, image decomposition, multi-level image fusion, and the aggregation process. Suppose that the size of the input image is n = M × N . For the image complexity assessment stage, the SF and the variance are calculated. Both the SF and the variance have an O ( n ) complexity. Thus, the time complexity of this stage is O ( M N ) . For the image decomposition stage, the MLSW and MLSW-GS are adopted. MLSW-GS has a higher complexity than MLSW. The complexity of MLSW-GS is determined by the number of decomposition layers s, filter operation, and guided operation. The complexity of the filter operation is determined by the kernel size k. It has an O ( n k 2 ) complexity, and so is the guided operation. So in this stage, the time complexity is O ( 2 s M N k 2 ) . For the multi-level image fusion stage, SPhCM-PCNN, SF-LV and Energy based fusion strategies are adopted. Both SF-LV and Energy-based method have O ( n ) complexity. The time complexity of SPhCM-PCNN is O ( M N ( 4 l l o g ( M N ) + C c o f + L ) ) , where the l is the scale parameter in SPhC method. The 4 represents the four orientations in SPhC. C c o f = 1 3 ( t + 1 ) ( 4 ( t + 1 ) 2 1 ) 1 . t is the number of scales in SPhCM. L is the iteration number in PCNN. So in this stage, the time complexity is O ( M N ( 4 l l o g ( M N ) + C c o f + L ) ) . For the aggregating process stage, a sum operation is adopted. Sum operation has an O ( n ) complexity. So in this stage, the complexity is determined by the number of decomposition layers s. The time complexity of this stage is O ( s M N ) . Hence, the time complexity of the entire proposed method is O ( M N ( s + 1 + 2 s k 2 + 4 l l o g ( M N ) + C c o f + L ) ) . As we analyzed, the second and third stages account for the main part of the time consumption of the proposed method.
To demonstrate the run-time of the proposed method more specifically, we separately timed each stage of the method. The results are shown in Table 7. It should be noted that the input image size is 151 × 151 .
The results in Table 7 indicate that the time consumption of the proposed method is mainly concentrated in the image decomposition and multi-level image fusion stages, which is consistent with our previous analysis of time complexity.
To compare the time consumption with other state-of-the-art methods, we selected four of the methods mentioned above as references, timed them, and compared the results. The experimental results are shown in Table 8.
Compared with other methods, our method has a higher time complexity and longer processing time. It can not meet the real-time processing requirement. This needs further improvement to help our proposed method enter engineering practice. One feasible solution is to design a parallel structure for the method because the main structures of the method process the images in the local area, which is more amenable to parallel computation. Furthermore, simplification methods for some structures can also be used to improve the time performance.

5. Conclusions

In this work, a novel fusion method named adaptive guided multi-layer side window box filter decomposition (AMLSW-GS) is proposed to mitigate detail loss in the fusion process of ISAR and visible images. Firstly, a new information abundance discrimination method based on spatial frequency and variance is proposed to sort the input images into detailed and coarse images. Secondly, the MLSW decomposition and MLSW-GS decomposition methods are proposed to decompose the detailed image and the coarse image into different scales, respectively. Thirdly, four fusion strategies—sum fusion, SF-LV, SPhCM-PCNN, and energy-based methods—are adopted to fuse different scale features to retain more detail. Finally, the fusion result is obtained by accumulating different sub-fusion results. In this way, the proposed method can make full use of the latent features from different types of images and retain more detail in the fusion result. On the synthetic ISAR-VIS dataset and the real-world IR-VIS dataset, compared with other state-of-the-art fusion methods, the experimental results show that the proposed method has the best performance in both subjective and objective fusion quality evaluations. This demonstrates the superiority and effectiveness of the proposed asymmetric decomposition methods.
Additionally, the unstable parameters limit the overall performance of the method, and further optimization is needed for the weak signal enhancement method. Simultaneously, the fusion time is in the order of seconds, which cannot meet the real-time requirements of engineering. In the future, we will mainly research more effective methods of enhancing weak signals to better preserve information. Furthermore, we will consider the execution efficiency of the methods while designing fusion strategies to make them applicable in engineering practices.

Author Contributions

Conceptualization, J.Z. and W.T.; methodology, J.Z.; software, J.Z.; validation, J.Z., P.X. and D.Z.; formal analysis, D.Z.; investigation, J.Z.; resources, W.T.; data curation, H.L.; writing—original draft preparation, J.Z.; writing—review and editing, P.X., D.Z. and P.V.A.; visualization, J.H. and J.D.; supervision, P.X., D.Z. and H.Z.; project administration, D.Z.; funding acquisition, H.Z., D.Z. and J.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the 111 Project (B17035), Aeronautical Science Foundation of China (201901081002), Youth Foundation of Shaanxi Province (2021JQ-182), the Natural Science Foundation of Jiangsu Province (BK20210064), the Wuxi Innovation and Entrepreneurship Fund “Taihu Light” Science and Technology (Fundamental Research) Project (K20221046), the Start-up Fund for Introducing Talent of Wuxi University (2021r007), National Natural Science Foundation of China (62001443,62105258), Natural Science Foundation of ShanDong province (ZR2020QE294), the Fundamental Research Funds for the Central Universities (QTZX23059, QTZX23009), the Basic Research Plan of Natural Science in Shaanxi Province (2023-JC-YB-062, 021JQ-182).

Data Availability Statement

Data are available from the corresponding author upon reasonable request.

Acknowledgments

Thanks are due to Weiming Yan for valuable discussion.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yi, S.; Jiang, G.; Liu, X.; Li, J.; Chen, L. TCPMFNet: An infrared and visible image fusion network with composite auto encoder and transformer–convolutional parallel mixed fusion strategy. Infrared Phys. Technol. 2022, 127, 104405. [Google Scholar] [CrossRef]
  2. Huang, Z.; Yang, B.; Liu, C. RDCa-Net: Residual dense channel attention symmetric network for infrared and visible image fusion. Infrared Phys. Technol. 2023, 130, 104589. [Google Scholar] [CrossRef]
  3. Zhang, X.; Liu, G.; Huang, L.; Ren, Q.; Bavirisetti, D.P. IVOMFuse: An image fusion method based on infrared-to-visible object mapping. Digit. Signal Process. 2023, 137, 104032. [Google Scholar] [CrossRef]
  4. Liu, X.; Wang, R.; Huo, H.; Yang, X.; Li, J. An attention-guided and wavelet-constrained generative adversarial network for infrared and visible image fusion. Infrared Phys. Technol. 2023, 129, 104570. [Google Scholar] [CrossRef]
  5. Tan, W.; Zhou, H.; Rong, S.; Qian, K.; Yu, Y. Fusion of multi-focus images via a Gaussian curvature filter and synthetic focusing degree criterion. Appl. Opt. 2018, 57, 10092–10101. [Google Scholar] [CrossRef]
  6. Su, W.; Huang, Y.; Li, Q.; Zuo, F. GeFuNet: A knowledge-guided deep network for the infrared and visible image fusion. Infrared Phys. Technol. 2022, 127, 104417. [Google Scholar] [CrossRef]
  7. Liu, Y.; Chen, X.; Cheng, J.; Peng, H.; Wang, Z. Infrared and visible image fusion with convolutional neural networks. Int. J. Wavelets Multiresolut. Inf. Process. 2018, 16, 1850018. [Google Scholar] [CrossRef]
  8. Zhuang, D.; Zhang, L.; Zou, B. An Interferogram Re-Flattening Method for InSAR Based on Local Residual Fringe Removal and Adaptively Adjusted Windows. Remote Sens. 2023, 15, 2214. [Google Scholar] [CrossRef]
  9. Shao, Z.; Zhang, T.; Ke, X. A Dual-Polarization Information-Guided Network for SAR Ship Classification. Remote Sens. 2023, 15, 2138. [Google Scholar] [CrossRef]
  10. Byun, Y.; Choi, J.; Han, Y. An area-based image fusion scheme for the integration of SAR and optical satellite imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 6, 2212–2220. [Google Scholar] [CrossRef]
  11. Iervolino, P.; Guida, R.; Riccio, D.; Rea, R. A novel multispectral, panchromatic and SAR data fusion for land classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 3966–3979. [Google Scholar] [CrossRef]
  12. Li, B.; Xian, Y.; Zhang, D.; Su, J.; Hu, X.; Guo, W. Multi-sensor image fusion: A survey of the state of the art. J. Comput. Commun. 2021, 9, 73–108. [Google Scholar] [CrossRef]
  13. Li, S.; Kwok, J.T.; Wang, Y. Combination of images with diverse focuses using the spatial frequency. Inf. Fusion 2001, 2, 169–176. [Google Scholar] [CrossRef]
  14. Mitianoudis, N.; Stathaki, T. Pixel-based and region-based image fusion schemes using ICA bases. Inf. Fusion 2007, 8, 131–142. [Google Scholar] [CrossRef]
  15. Li, X.; Wang, L.; Wang, J.; Zhang, X. Multi-focus image fusion algorithm based on multilevel morphological component analysis and support vector machine. IET Image Process. 2017, 11, 919–926. [Google Scholar] [CrossRef]
  16. Li, H.; Wu, X.J. Infrared and visible image fusion using latent low-rank representation. arXiv 2018, arXiv:1804.08992. [Google Scholar]
  17. Liu, Y.; Wang, Z. Simultaneous image fusion and denoising with adaptive sparse representation. IET Image Process. 2015, 9, 347–357. [Google Scholar] [CrossRef]
  18. Yin, M.; Liu, X.; Liu, Y.; Chen, X. Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Trans. Instrum. Meas. 2018, 68, 49–64. [Google Scholar] [CrossRef]
  19. Li, H.; Wu, X.J.; Kittler, J. Infrared and visible image fusion using a deep learning framework. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2705–2710. [Google Scholar]
  20. Tan, W.; Tiwari, P.; Pandey, H.M.; Moreira, C.; Jaiswal, A.K. Multimodal medical image fusion algorithm in the era of big data. Neural Comput. Appl. 2020, 2020, 1–21. [Google Scholar] [CrossRef]
  21. Zhang, S.; Li, X.; Zhang, X.; Zhang, S. Infrared and visible image fusion based on saliency detection and two-scale transform decomposition. Infrared Phys. Technol. 2021, 114, 103626. [Google Scholar] [CrossRef]
  22. Wan, T.; Zhu, C.; Qin, Z. Multifocus image fusion based on robust principal component analysis. Pattern Recognit. Lett. 2013, 34, 1001–1008. [Google Scholar] [CrossRef]
  23. Yang, Y.; Zhang, W.; Huang, S.; Wan, W.; Liu, J.; Kong, X. Infrared and visible image fusion based on dual-kernel side window filtering and S-shaped curve transformation. IEEE Trans. Instrum. Meas. 2021, 71, 1–15. [Google Scholar] [CrossRef]
  24. Smadi, A.A.; Abugabah, A.; Mehmood, A.; Yang, S. Brain image fusion approach based on side window filtering. Procedia Comput. Sci. 2022, 198, 295–300. [Google Scholar] [CrossRef]
  25. Yan, H.; Zhang, J.X.; Zhang, X. Injected Infrared and Visible Image Fusion via L_{1} Decomposition Model and Guided Filtering. IEEE Trans. Comput. Imaging 2022, 8, 162–173. [Google Scholar] [CrossRef]
  26. Zou, D.; Yang, B. Infrared and low-light visible image fusion based on hybrid multiscale decomposition and adaptive light adjustment. Opt. Lasers Eng. 2023, 160, 107268. [Google Scholar] [CrossRef]
  27. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  28. Zhang, Q.; Levine, M.D. Robust multi-focus image fusion using multi-task sparse representation and spatial context. IEEE Trans. Image Process. 2016, 25, 2045–2058. [Google Scholar] [CrossRef]
  29. Liu, X.; Mei, W.; Du, H. Structure tensor and nonsubsampled shearlet transform based algorithm for CT and MRI image fusion. Neurocomputing 2017, 235, 131–139. [Google Scholar] [CrossRef]
  30. Qu, X.B.; Yan, J.W.; Hong-Zhi, X.; Zi-Qian, Z. Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain. Acta Autom. Sin. 2008, 34, 1508–1514. [Google Scholar] [CrossRef]
  31. Panigrahy, C.; Seal, A.; Mahato, N.K. MRI and SPECT image fusion using a weighted parameter adaptive dual channel PCNN. IEEE Signal Process. Lett. 2020, 27, 690–694. [Google Scholar] [CrossRef]
  32. Indhumathi, R.; Narmadha, T. Hybrid pixel based method for multimodal image fusion based on Integration of Pulse Coupled Neural Network (PCNN) and Genetic Algorithm (GA) using Empirical Mode Decomposition (EMD). Microprocess. Microsyst. 2022, 94, 104665. [Google Scholar] [CrossRef]
  33. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef] [PubMed]
  34. Li, H.; Wu, X.J. DenseFuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 2018, 28, 2614–2623. [Google Scholar] [CrossRef] [PubMed]
  35. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  36. Xu, H.; Ma, J.; Zhang, X.P. MEF-GAN: Multi-exposure image fusion via generative adversarial networks. IEEE Trans. Image Process. 2020, 29, 7203–7216. [Google Scholar] [CrossRef]
  37. Yin, H.; Gong, Y.; Qiu, G. Side window filtering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8758–8766. [Google Scholar]
  38. He, K.; Sun, J.; Tang, X. Guided Image Filtering. In Proceedings of the European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010. [Google Scholar]
  39. Kovesi, P. Image features from phase congruency. Videre: J. Comput. Vis. Res. 1999, 1, 1–26. [Google Scholar]
  40. Morrone, M.C.; Owens, R.A. Feature detection from local energy. Pattern Recognit. Lett. 1987, 6, 303–313. [Google Scholar] [CrossRef]
  41. Kovesi, P. Symmetry and asymmetry from local phase. In Proceedings of the Tenth Australian Joint Conference on Artificial Intelligence, Perth, Australia, 30 November–4 December 1997; Volume 190, pp. 2–4. [Google Scholar]
  42. Xiao, Z.; Hou, Z.; Guo, C. Image feature detection technique based on phase information: Symmetry phase congruency. J. Tianjin Univ. 2004, 37, 695–699. [Google Scholar]
  43. Zhang, F.; Zhang, B.; Zhang, R.; Zhang, X. SPCM: Image quality assessment based on symmetry phase congruency. Appl. Soft Comput. 2020, 87, 105987. [Google Scholar] [CrossRef]
  44. Johnson, J.L.; Padgett, M.L. PCNN Models and Applications. IEEE Trans. Neural Netw. 1999, 10, 480–498. [Google Scholar] [CrossRef]
  45. Teng, X. VisableAerial-Target. Available online: https://github.com/TengXiang164/VisableAerial-Target (accessed on 4 April 2023).
  46. Richards, M.A. Fundamentals of Radar Signal Processing; McGraw-Hill Education: New York, NY, USA, 2014. [Google Scholar]
  47. Li, H.; Wu, X.J.; Kittler, J. MDLatLRR: A novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 2020, 29, 4733–4746. [Google Scholar] [CrossRef] [PubMed]
  48. Zhang, X.; Ye, P.; Xiao, G. VIFB: A visible and infrared image fusion benchmark. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 104–105. [Google Scholar]
  49. Bavirisetti, D.P.; Dhuli, R. Fusion of infrared and visible sensor images based on anisotropic diffusion and Karhunen-Loeve transform. IEEE Sens. J. 2015, 16, 203–209. [Google Scholar] [CrossRef]
  50. Bavirisetti, D.P.; Xiao, G.; Liu, G. Multi-sensor image fusion based on fourth order partial differential equations. In Proceedings of the IEEE 2017 20th International Conference on Information Fusion (Fusion), Xi’an, China, 10–13 July 2017; pp. 1–9. [Google Scholar]
  51. Li, H.; Ma, K.; Yong, H.; Zhang, L. Fast multi-scale structural patch decomposition for multi-exposure image fusion. IEEE Trans. Image Process. 2020, 29, 5805–5816. [Google Scholar] [CrossRef] [PubMed]
  52. Zhou, Z.; Dong, M.; Xie, X.; Gao, Z. Fusion of infrared and visible images for night-vision context enhancement. Appl. Opt. 2016, 55, 6480–6490. [Google Scholar] [CrossRef] [PubMed]
  53. Ma, J.; Zhou, Z.; Wang, B.; Zong, H. Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 2017, 82, 8–17. [Google Scholar] [CrossRef]
  54. Ma, J.; Zhou, Y. Infrared and visible image fusion via gradientlet filter. Comput. Vis. Image Underst. 2020, 197, 103016. [Google Scholar] [CrossRef]
  55. Biswas, B. Guided-Bilateral Filter-Based Medical Image Fusion Using Visual Saliency Map in the Wavelet Domain. Available online: https://github.com/biswajitcsecu/Guided-Bilateral-Filter-based-Medical-Image-Fusion-Using-Visual-Saliency-Map-in-the-Wavelet-Domain (accessed on 4 April 2023).
  56. Li, H.; Wu, X.J.; Durrani, T.S. Infrared and Visible Image Fusion with ResNet and zero-phase component analysis. Infrared Phys. Technol. 2019, 102, 103039. [Google Scholar] [CrossRef]
  57. Cvejic, N.; Canagarajah, C.; Bull, D. Image fusion metric based on mutual information and Tsallis entropy. Electron. Lett. 2006, 42, 1. [Google Scholar] [CrossRef]
  58. Wang, Q.; Shen, Y.; Jin, J. Performance evaluation of image fusion techniques. Image Fusion Algorithms Appl. 2008, 19, 469–492. [Google Scholar]
  59. Xydeas, C.S.; Petrovic, V.S. Objective pixel-level image fusion performance measure. In Proceedings of the Sensor Fusion: Architectures, Algorithms, and Applications IV, SPIE, Orlando, FL, USA, 6–7 April 2000; Volume 4051, pp. 89–98. [Google Scholar]
  60. Piella, G.; Heijmans, H. A new quality metric for image fusion. In Proceedings of the IEEE 2003 International Conference on Image Processing (Cat. No. 03CH37429), Barcelona, Spain, 14–17 September 2003; Volume 3, pp. 3–173. [Google Scholar]
  61. Li, S.; Hong, R.; Wu, X. A novel similarity based quality metric for image fusion. In Proceedings of the IEEE 2008 International Conference on Audio, Language and Image Processing, Shanghai, China, 7–9 July 2008; pp. 167–172. [Google Scholar]
  62. Haghighat, M.B.A.; Aghagolzadeh, A.; Seyedarabi, H. A non-reference image fusion metric based on mutual information of image features. Comput. Electr. Eng. 2011, 37, 744–756. [Google Scholar] [CrossRef]
  63. Haghighat, M.; Razian, M.A. Fast-FMI: Non-reference image fusion metric. In Proceedings of the 2014 IEEE 8th International Conference on Application of Information and Communication Technologies (AICT), Astana, Kazakhstan, 15–17 October 2014; pp. 1–3. [Google Scholar]
  64. Toet, A. The TNO multiband image data collection. Data Brief 2017, 15, 249–251. [Google Scholar] [CrossRef] [PubMed]
  65. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain Long-range Learning for General Image Fusion via Swin Transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
Figure 1. Framework of the proposed method.
Figure 1. Framework of the proposed method.
Remotesensing 15 02784 g001
Figure 2. Framework of MLSW.
Figure 2. Framework of MLSW.
Remotesensing 15 02784 g002
Figure 3. Framework of the MLSW-GS.
Figure 3. Framework of the MLSW-GS.
Remotesensing 15 02784 g003
Figure 4. Framework of the SPhCM-PCNN.
Figure 4. Framework of the SPhCM-PCNN.
Remotesensing 15 02784 g004
Figure 5. Some samples in the public dataset.
Figure 5. Some samples in the public dataset.
Remotesensing 15 02784 g005
Figure 6. Results of different fusion methods in scene 1. The red box shows the detail preservation of ISAR, magnified in the upper right corner, while the green box demonstrates the information preservation of the visible image, mainly in terms of signal intensity. They are listed as follows: (a) Source of VIS. (b) Source of ISAR. (c) Result of ASR. (d) Result of MDLatLRR. (e) Result of ADF. (f) Result of FPDE. (g) Result of FMMEF. (h) Result of GFCE. (i) Result of VSMWLS. (j) Result of Gradientlet. (k) Result of LatLRR. (l) Result of NSCT-PAPCNN. (m) Result of NSCT-SF-PCNN. (n) Result of MST-SR. (o) Result of DWT-BGF. (p) Result of VGG19-L1-Fusion. (q) Result of ResNet50-ZPCA. (r) Result of the proposed method.
Figure 6. Results of different fusion methods in scene 1. The red box shows the detail preservation of ISAR, magnified in the upper right corner, while the green box demonstrates the information preservation of the visible image, mainly in terms of signal intensity. They are listed as follows: (a) Source of VIS. (b) Source of ISAR. (c) Result of ASR. (d) Result of MDLatLRR. (e) Result of ADF. (f) Result of FPDE. (g) Result of FMMEF. (h) Result of GFCE. (i) Result of VSMWLS. (j) Result of Gradientlet. (k) Result of LatLRR. (l) Result of NSCT-PAPCNN. (m) Result of NSCT-SF-PCNN. (n) Result of MST-SR. (o) Result of DWT-BGF. (p) Result of VGG19-L1-Fusion. (q) Result of ResNet50-ZPCA. (r) Result of the proposed method.
Remotesensing 15 02784 g006
Figure 7. Results of different fusion methods in scene 2. The red box shows the detail preservation of ISAR, magnified in the upper right corner, while the green box demonstrates the information preservation of the visible image, mainly in terms of signal intensity. They are listed as follows: (a) Source of VIS. (b) Source of ISAR. (c) Result of ASR. (d) Result of MDLatLRR. (e) Result of ADF. (f) Result of FPDE. (g) Result of FMMEF. (h) Result of GFCE. (i) Result of VSMWLS. (j) Result of Gradientlet. (k) Result of LatLRR. (l) Result of NSCT-PAPCNN. (m) Result of NSCT-SF-PCNN. (n) Result of MST-SR. (o) Result of DWT-BGF. (p) Result of VGG19-L1-Fusion. (q) Result of ResNet50-ZPCA. (r) Result of the proposed method.
Figure 7. Results of different fusion methods in scene 2. The red box shows the detail preservation of ISAR, magnified in the upper right corner, while the green box demonstrates the information preservation of the visible image, mainly in terms of signal intensity. They are listed as follows: (a) Source of VIS. (b) Source of ISAR. (c) Result of ASR. (d) Result of MDLatLRR. (e) Result of ADF. (f) Result of FPDE. (g) Result of FMMEF. (h) Result of GFCE. (i) Result of VSMWLS. (j) Result of Gradientlet. (k) Result of LatLRR. (l) Result of NSCT-PAPCNN. (m) Result of NSCT-SF-PCNN. (n) Result of MST-SR. (o) Result of DWT-BGF. (p) Result of VGG19-L1-Fusion. (q) Result of ResNet50-ZPCA. (r) Result of the proposed method.
Remotesensing 15 02784 g007
Figure 8. Different results of different fusion methods in scene 3. The red box shows the detail preservation of ISAR, magnified in the upper right corner, while the green box demonstrates the information preservation of the visible image, mainly in terms of signal intensity. They are listed as follows: (a) Source of VIS. (b) Source of ISAR. (c) Result of ASR. (d) Result of MDLatLRR. (e) Result of ADF. (f) Result of FPDE. (g) Result of FMMEF. (h) Result of GFCE. (i) Result of VSMWLS. (j) Result of Gradientlet. (k) Result of LatLRR. (l) Result of NSCT-PAPCNN. (m) Result of NSCT-SF-PCNN. (n) Result of MST-SR. (o) Result of DWT-BGF. (p) Result of VGG19-L1-Fusion. (q) Result of ResNet50-ZPCA. (r) Result of the proposed method.
Figure 8. Different results of different fusion methods in scene 3. The red box shows the detail preservation of ISAR, magnified in the upper right corner, while the green box demonstrates the information preservation of the visible image, mainly in terms of signal intensity. They are listed as follows: (a) Source of VIS. (b) Source of ISAR. (c) Result of ASR. (d) Result of MDLatLRR. (e) Result of ADF. (f) Result of FPDE. (g) Result of FMMEF. (h) Result of GFCE. (i) Result of VSMWLS. (j) Result of Gradientlet. (k) Result of LatLRR. (l) Result of NSCT-PAPCNN. (m) Result of NSCT-SF-PCNN. (n) Result of MST-SR. (o) Result of DWT-BGF. (p) Result of VGG19-L1-Fusion. (q) Result of ResNet50-ZPCA. (r) Result of the proposed method.
Remotesensing 15 02784 g008
Figure 9. Results of different fusion methods in scene 4. The red box shows the detail preservation of ISAR, magnified in the upper right corner, while the green box demonstrates the information preservation of the visible image, mainly in terms of signal intensity. They are listed as follows: (a) Source of VIS. (b) Source of ISAR. (c) Result of ASR. (d) Result of MDLatLRR. (e) Result of ADF. (f) Result of FPDE. (g) Result of FMMEF. (h) Result of GFCE. (i) Result of VSMWLS. (j) Result of Gradientlet. (k) Result of LatLRR. (l) Result of NSCT-PAPCNN. (m) Result of NSCT-SF-PCNN. (n) Result of MST-SR. (o) Result of DWT-BGF. (p) Result of VGG19-L1-Fusion. (q) Result of ResNet50-ZPCA. (r) Result of the proposed method.
Figure 9. Results of different fusion methods in scene 4. The red box shows the detail preservation of ISAR, magnified in the upper right corner, while the green box demonstrates the information preservation of the visible image, mainly in terms of signal intensity. They are listed as follows: (a) Source of VIS. (b) Source of ISAR. (c) Result of ASR. (d) Result of MDLatLRR. (e) Result of ADF. (f) Result of FPDE. (g) Result of FMMEF. (h) Result of GFCE. (i) Result of VSMWLS. (j) Result of Gradientlet. (k) Result of LatLRR. (l) Result of NSCT-PAPCNN. (m) Result of NSCT-SF-PCNN. (n) Result of MST-SR. (o) Result of DWT-BGF. (p) Result of VGG19-L1-Fusion. (q) Result of ResNet50-ZPCA. (r) Result of the proposed method.
Remotesensing 15 02784 g009
Figure 10. Comparison of different methods used on representative scenes. (a) MI values of Different Methods (b) NICE values of Different Methods (c) Q a b / f values of Different Methods (d) Q S values of Different Methods (e) Q Y values of Different Methods (f) Average values of Different Methods (g) SSIM values of Different Methods (h) FMI values of Different Methods.
Figure 10. Comparison of different methods used on representative scenes. (a) MI values of Different Methods (b) NICE values of Different Methods (c) Q a b / f values of Different Methods (d) Q S values of Different Methods (e) Q Y values of Different Methods (f) Average values of Different Methods (g) SSIM values of Different Methods (h) FMI values of Different Methods.
Remotesensing 15 02784 g010
Figure 11. Performance under different numbers of decompositions. (a) is the result of MI, and (b) is the result of Q a b / f . In this experiment, the number of scales in morphological operation t is 3.
Figure 11. Performance under different numbers of decompositions. (a) is the result of MI, and (b) is the result of Q a b / f . In this experiment, the number of scales in morphological operation t is 3.
Remotesensing 15 02784 g011
Figure 12. Performance under different numbers of scales. (a) is the result of MI, and (b) is the result of Q a b / f . In this experiment, the number of decompositions s is 3.
Figure 12. Performance under different numbers of scales. (a) is the result of MI, and (b) is the result of Q a b / f . In this experiment, the number of decompositions s is 3.
Remotesensing 15 02784 g012
Figure 13. Comparison of results of different model structures. (a) is the Source of VIS, (b) is the Source of ISAR, (c) AMLSW + SP + SL, (d) MLSW + SP + SL, (e) AMLSW + SP, (f) AMLSW + SL, (g) Difference between AMLSW + SP + SL and AMLSW + SP + SL, (h) Difference between MLSW + SP + SL and AMLSW + SP + SL, (i) Difference between AMLSW + SP and AMLSW + SP + SL, (j) Difference between AMLSW + SL and AMLSW + SP + SL.
Figure 13. Comparison of results of different model structures. (a) is the Source of VIS, (b) is the Source of ISAR, (c) AMLSW + SP + SL, (d) MLSW + SP + SL, (e) AMLSW + SP, (f) AMLSW + SL, (g) Difference between AMLSW + SP + SL and AMLSW + SP + SL, (h) Difference between MLSW + SP + SL and AMLSW + SP + SL, (i) Difference between AMLSW + SP and AMLSW + SP + SL, (j) Difference between AMLSW + SL and AMLSW + SP + SL.
Remotesensing 15 02784 g013
Figure 14. Results of different methods for infrared and visible image fusion. Different scenes are shown in different columns. The first row contains the visible images, the second row contains the infrared images, the third row contains the fusion results of MDLatLRR, the fourth row contains the fusion results of NSCT-SF-PCNN, the fifth row contains the fusion results of ResNet50-ZPCA, and the last row contains the fusion results of the proposed method.
Figure 14. Results of different methods for infrared and visible image fusion. Different scenes are shown in different columns. The first row contains the visible images, the second row contains the infrared images, the third row contains the fusion results of MDLatLRR, the fourth row contains the fusion results of NSCT-SF-PCNN, the fifth row contains the fusion results of ResNet50-ZPCA, and the last row contains the fusion results of the proposed method.
Remotesensing 15 02784 g014
Table 1. Objective values of different fusion methods in scene 1. The best values, the second-best values, and the third-best values are marked in red, blue, and green, respectively.
Table 1. Objective values of different fusion methods in scene 1. The best values, the second-best values, and the third-best values are marked in red, blue, and green, respectively.
MetricsMINCIE Q ab / f Q S Q Y AverageSSIMFMI
ASR0.55780.80840.57690.64640.6187104.25980.80410.4333
MDLatLRR0.58540.80930.53530.65040.6027104.82930.82970.4230
ADF0.59170.80860.35600.60500.3951104.25600.75820.4062
FPDE0.59040.80880.34750.61680.4046104.61140.78090.4011
FMMEF0.38830.80570.39300.74110.4927151.50030.90980.2929
GFCE0.39450.80550.44460.68810.5477127.97030.83810.3334
VSMWLS0.55420.80860.52020.65590.5670106.58520.81690.4269
Gradientlet0.60130.81130.61050.63920.6227106.74990.81150.4492
LatLRR0.48500.80700.51880.67270.5955117.10270.84060.3852
NSCT-PAPCNN0.56390.80940.51250.71600.6102194.67750.86170.3737
NSCT-SF-PCNN0.33350.80490.21280.36420.293558.50170.53150.2337
MST-SR0.37180.80610.34250.66300.4895149.49480.81280.2777
DWT-BGF0.59170.80940.33950.59660.4214104.25740.75160.3527
VGG19-L1-Fusion0.61390.80990.39370.62190.4810104.36850.79490.3896
ResNet50-ZPCA0.61440.80990.45290.63590.5367104.52810.81590.3980
Proposed0.89580.82380.73120.76870.8584194.96410.95830.4929
Table 2. Objective values of different fusion methods in scene 2. The best values, the second-best values, and the third-best values are marked in red, blue, and green, respectively.
Table 2. Objective values of different fusion methods in scene 2. The best values, the second-best values, and the third-best values are marked in red, blue, and green, respectively.
MetricsMINCIE Q ab / f Q S Q Y AverageSSIMFMI
ASR0.68300.80820.57800.58890.5707110.09050.79030.5250
MDLatLRR0.69050.80800.58590.59880.5294110.72540.80680.5460
ADF0.69880.80770.48310.57080.3093110.04390.75440.5403
FPDE0.62410.80730.46080.53400.4894110.49540.72880.4680
FMMEF0.47610.80640.28170.67970.3985151.61050.88420.2470
GFCE0.46520.80630.31660.56930.3774111.72590.73190.2891
VSMWLS0.68800.80820.55520.59550.5136109.97620.78980.5218
Gradientlet0.74620.80990.62170.58900.5791114.37280.79790.5601
LatLRR0.51490.80610.43080.60520.4206127.36590.79870.3742
NSCT-PAPCNN0.49370.80650.27900.60850.3434199.66880.76690.2563
NSCT-SF-PAPCNN0.32810.80480.17780.33510.267862.24040.50880.1991
MST-SR0.41390.80600.24860.58330.3345145.83040.75770.2367
DWT-BGF0.68130.80780.47210.55150.3389110.04050.73660.5045
VGG19-L1-Fusion0.71950.80810.50320.57780.4063110.29470.77770.5455
ResNet50-ZPCA0.71020.80810.54300.58840.4677110.38830.79840.5468
Proposed0.94300.81530.72100.70480.8719201.62020.93900.6030
Table 3. Objective values of different fusion methods in scene 3. The best values, the second-best values, and the third-best values are marked in red, blue, and green, respectively.
Table 3. Objective values of different fusion methods in scene 3. The best values, the second-best values, and the third-best values are marked in red, blue, and green, respectively.
MetricsMINCIE Q ab / f Q S Q Y AverageSSIMFMI
ASR0.51990.80850.68870.70180.769664.45220.85930.4409
MDLatLRR0.55930.80940.58370.69680.714164.72190.86400.3938
ADF0.52770.80860.63310.68650.755164.47050.86920.4146
FPDE0.53230.80860.59040.69100.732764.81310.86840.4004
FMMEF0.51090.80840.76340.82060.871393.72310.96960.4511
GFCE0.40160.80580.60820.77670.7982131.06070.91130.4097
VSMWLS0.53190.80850.63050.71900.742272.12180.87820.4065
Gradientlet0.52590.80820.63450.72800.785076.49900.90290.4077
LatLRR0.58370.81010.62530.73360.745278.24870.88760.3974
NSCT-PAPCNN0.60730.81230.68670.81460.8360116.89930.95530.4181
NSCT-SF-PAPCNN0.38360.80520.37980.48100.477365.11870.70160.3191
MST-SR0.66630.81520.73000.81990.8650112.08200.95700.4472
DWT-BGF0.48270.80690.24740.52870.410064.48230.67540.1920
VGG19-L1-Fusion0.56090.80900.28140.58820.482264.68620.75040.2689
ResNet50-ZPCA0.56080.80910.36750.63890.593164.63970.80520.3074
Proposed0.83040.82590.79110.83740.9289115.15810.99300.4723
Table 4. Objective values of different fusion methods in scene 4. The best values, the second-best values, and the third-best values are marked in red, blue, and green, respectively.
Table 4. Objective values of different fusion methods in scene 4. The best values, the second-best values, and the third-best values are marked in red, blue, and green, respectively.
MetricsMINCIE Q ab / f Q S Q Y AverageSSIMFMI
ASR0.41850.80730.61670.71600.776474.30900.94110.4506
MDLatLRR0.43290.80780.50320.69470.709374.75270.95450.4069
ADF0.42480.80770.49780.64910.711374.31170.95160.4100
FPDE0.42000.80760.48270.65870.707874.79380.95200.4061
FMMEF0.47990.80940.64920.77940.8293101.85660.99760.4448
GFCE0.41790.80750.55180.71320.7505146.77150.90150.4188
VSMWLS0.48200.80930.54150.73660.776690.69220.96570.4238
Gradientlet0.48630.80970.54520.73080.766293.19080.99160.4127
LatLRR0.45240.80820.54710.72280.742294.61530.95770.4114
NSCT-PAPCNN0.51420.81070.62150.77760.8176124.66200.98400.4386
NSCT-SF-PAPCNN0.39830.80680.49550.67450.719699.26780.88550.3881
MST-SR0.57210.81370.67420.79380.8555121.45130.99700.4671
DWT-BGF0.38350.80650.25870.52730.472574.35250.76720.2277
VGG19-L1-Fusion0.44220.80790.31820.61000.566674.79530.87710.3304
ResNet50-ZPCA0.44070.80800.36710.64360.625874.44570.91470.3525
Proposed0.65520.81790.66670.79590.8718124.49611.02090.4516
Table 5. Objective values of different structure methods. The best metric values are in bold.
Table 5. Objective values of different structure methods. The best metric values are in bold.
MethodsMINCIE Q ab / f SSIM
MLSW + SP + SL0.77110.82060.64110.8875
AMLSW + SP0.79460.82220.68270.9489
AMLSW + SL0.77090.82060.64130.8874
AMLSW + SP + SL0.80610.82300.67940.9523
Table 6. Objective value comparison of different methods in different scenes. The best metric values are in bold.
Table 6. Objective value comparison of different methods in different scenes. The best metric values are in bold.
MethodsMINCIE Q ab / f Q Y
Scene 1MDLatLRR0.30620.80420.39880.7311
NSCT-SF-PCNN0.24400.80350.39190.7521
ResNet50-ZPCA0.30230.80430.35900.6963
Proposed0.35460.80470.40770.7398
Scene 2MDLatLRR0.44230.80440.41330.7804
NSCT-SF-PCNN0.28720.80340.27770.6541
ResNet50-ZPCA0.43380.80420.34420.7124
Proposed0.50980.80620.43730.8115
Scene 3MDLatLRR0.31990.80440.38860.7135
NSCT-SF-PCNN0.34000.80590.54880.8798
ResNet50-ZPCA0.31220.80440.34440.6785
Proposed0.40610.80600.38720.7229
Table 7. Time consumption of different stages in the proposed method. Stage 1 is image complexity assessment. Stage 2 is image decomposition. Stage 3 is multi-level image fusion. Stage 4 is the aggregation process.
Table 7. Time consumption of different stages in the proposed method. Stage 1 is image complexity assessment. Stage 2 is image decomposition. Stage 3 is multi-level image fusion. Stage 4 is the aggregation process.
Stage 1Stage 2Stage 3Stage 4Total
0.00370.10521.51140.00081.6211
Table 8. Time consumption of different methods. The best metric values are in bold.
Table 8. Time consumption of different methods. The best metric values are in bold.
MethodsASRLatLRRADFFMMEFProposed
Time14.98683.35990.04400.08701.6211
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Li, H.; Zhao, D.; Arun, P.V.; Tan, W.; Xiang, P.; Zhou, H.; Hu, J.; Du, J. An ISAR and Visible Image Fusion Algorithm Based on Adaptive Guided Multi-Layer Side Window Box Filter Decomposition. Remote Sens. 2023, 15, 2784. https://doi.org/10.3390/rs15112784

AMA Style

Zhang J, Li H, Zhao D, Arun PV, Tan W, Xiang P, Zhou H, Hu J, Du J. An ISAR and Visible Image Fusion Algorithm Based on Adaptive Guided Multi-Layer Side Window Box Filter Decomposition. Remote Sensing. 2023; 15(11):2784. https://doi.org/10.3390/rs15112784

Chicago/Turabian Style

Zhang, Jiajia, Huan Li, Dong Zhao, Pattathal V. Arun, Wei Tan, Pei Xiang, Huixin Zhou, Jianling Hu, and Juan Du. 2023. "An ISAR and Visible Image Fusion Algorithm Based on Adaptive Guided Multi-Layer Side Window Box Filter Decomposition" Remote Sensing 15, no. 11: 2784. https://doi.org/10.3390/rs15112784

APA Style

Zhang, J., Li, H., Zhao, D., Arun, P. V., Tan, W., Xiang, P., Zhou, H., Hu, J., & Du, J. (2023). An ISAR and Visible Image Fusion Algorithm Based on Adaptive Guided Multi-Layer Side Window Box Filter Decomposition. Remote Sensing, 15(11), 2784. https://doi.org/10.3390/rs15112784

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop