Next Article in Journal
A Framework to Secure the Development and Auditing of SSL Pinning in Mobile Applications: The Case of Android Devices
Next Article in Special Issue
Identification of Functional Bioprocess Model for Recombinant E. Coli Cultivation Process
Previous Article in Journal
Detecting Metachanges in Data Streams from the Viewpoint of the MDL Principle
Previous Article in Special Issue
Recognition of a Single Dynamic Gesture with the Segmentation Technique HS-ab and Principle Components Analysis (PCA)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Infrared and Visible Image Information Fusion Method Based on Phase Congruency and Image Entropy

1
Key Laboratory of Complex System Safety and Control, Ministry of Education, Chongqing University, Chongqing 400044, China
2
Computer Information Systems Department, Buffalo State College, Buffalo, NY 14222, USA
3
College of Automation, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
4
Department of Mathematics and Computer Information Science, Mansfield University of Pennsylvania, Mansfield, PA 16933, USA
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(12), 1135; https://doi.org/10.3390/e21121135
Submission received: 17 October 2019 / Revised: 18 November 2019 / Accepted: 19 November 2019 / Published: 21 November 2019
(This article belongs to the Special Issue Entropy-Based Algorithms for Signal Processing)

Abstract

:
In multi-modality image fusion, source image decomposition, such as multi-scale transform (MST), is a necessary step and also widely used. However, when MST is directly used to decompose source images into high- and low-frequency components, the corresponding decomposed components are not precise enough for the following infrared-visible fusion operations. This paper proposes a non-subsampled contourlet transform (NSCT) based decomposition method for image fusion, by which source images are decomposed to obtain corresponding high- and low-frequency sub-bands. Unlike MST, the obtained high-frequency sub-bands have different decomposition layers, and each layer contains different information. In order to obtain a more informative fused high-frequency component, maximum absolute value and pulse coupled neural network (PCNN) fusion rules are applied to different sub-bands of high-frequency components. Activity measures, such as phase congruency (PC), local measure of sharpness change (LSCM), and local signal strength (LSS), are designed to enhance the detailed features of fused low-frequency components. The fused high- and low-frequency components are integrated to form a fused image. The experiment results show that the fused images obtained by the proposed method achieve good performance in clarity, contrast, and image information entropy.

1. Introduction

Both infrared and visible images are widely used in daily life. Due to the difference in wavelength, infrared and visible light contain different image information. Infrared images can reflect all the objects that emit infrared radiation. Visible-light images can provide the scene details. No matter whether an infrared or visible-light image, it is difficult for an image captured by a single shot to contain all-in-focus images in one scene. Infrared-visible fusion techniques can effectively combine the complementary information, which are the indicative features and detailed information extracted from infrared and visible images, respectively [1]. In the fused infrared-visible image, the targeted item can be highlighted and the corresponding indicative features as well as detailed information are retained. At present, image fusion techniques as a type of image pre-processing methods, especially for infrared-visible image fusion, have been widely applied to the target recognition in different environments, such as smart city, battlefield, remote sensing, and so on [2,3].
In recent years, transform domain based methods have become the mainstream in infrared-visible image fusion, which include pyramid, wavelet transform, multi-scale geometric transform, sparse representation [4,5], and so on. Pyramid, wavelet transform, and multi-scale geometric transform can be categorized as MST-based methods. MST-based methods have three main steps. First, MST is employed to decompose each source image into high-frequency sub-bands at different scales and directions as well as one low-frequency sub-band. Then, the obtained high- and low-frequency sub-bands are fused separately following different fusion rules. Finally, the fused image is obtained by performing the inverse MST on both fused high- and low-frequency sub-bands. Double-tree complex wavelet transform as a kind of wavelet transform can only capture a limited amount of edge information, but cannot correctly and effectively represent the discontinuity of lines and curves [6]. As a true two-dimensional (2D) multi-scale geometric analysis method, contourlet transform (CT) possesses localization, multi-resolution, multi-scale, multi-direction, and anisotropy.
As a shift-invariant version of CT, NSCT performs well in transform domain, and has been widely used in image fusion. NSCT has multi-scale and multi-direction features, which can solve the limitations of traditional wavelet methods in the representation of image curves and edges [7]. Compared with traditional MST-based image fusion methods, NSCT has shift invariance and also suppresses pseudo-Gibbs phenomenon [8,9]. Based on the above advantages, Liu proposed a general image fusion framework based on MST and sparse representation (SR) [10]. It overcomes the shortcomings of MST- and SR-based fusion methods at the same time. However, redundancy and loss of residue exist in this method [11]. Li proposed an infrared-visible image fusion method based on PC information, which fuses PC information into the coefficients of frequency bands [12]. However, the computational complexity of this fusion method is high.
In traditional infrared-visible images, the target information cannot be extracted effectively. Ding proposed an infrared-visible image fusion method based on non-downsampling shear transform (NSST) and sparse structure features [13]. First, source images are decomposed into high- and low-frequency sub-band coefficients. According to the advantages of principal component analysis (PCA) in principle information extraction, the PCA-based method is then used to fuse the low-frequency sub-band coefficients. At the same time, a new sparse-feature extraction method of high-frequency sub-band coefficients is proposed, which fuses the high-frequency components of source images. Finally, a fused image is obtained by inverse NSST. An infrared-visible image fusion method that integrates NSST and spiking cortical model was proposed by Kong [14]. This method uses NSST to reconstruct the decomposed components, which not only makes the fused image have good performance in human-eye visualization, but also effectively reduces the computational complexity. On the other hand, the fusion of different-scales and -direction sub-images can be realized by using a spiking cortical model. Xiang proposed an infrared-visible image fusion algorithm based on an adaptive dual-channel unit-linking PCNN in an NSCT domain [15]. This algorithm uses NSCT to decompose source images in multiple scales and directions. In order to make the adaptive dual-channel PCNN, the average gradient of each pixel is taken as the connection strength, and the time matrix adaptively determines the number of iterations. In the fusion process, a low-frequency sub-band and the modified spatial-frequency Laplacian of a high-frequency sub-band are used as input to excite the adaptive dual-channel unit-linking PCNN. Zhang proposed an NSCT-based infrared image fusion method, which used an adaptive Gaussian (AG) fuzzy membership method, compressed sensing (CS) technique, and total variation (TV) based gradient descent reconstruction algorithm to do the fusion calculation of infrared-visible image [8]. Wang proposed an infrared-visible image fusion method that integrates data compression based on sparse representation and compressed sensing [16]. This method first performs random projection compression on the remote sensing data, and then obtains the sparse coefficients of the compressed sample by sparse representation. Finally, fusion coefficients are combined with fusion influence factors, and the fused image is reconstructed by fusing sparse coefficients.
This paper proposes a novel precise decomposition framework for infrared-visible image fusion, in which image energy and details can be preserved well. First, NSCT is used to decompose source images to obtain corresponding high- and low-frequency components. The high-frequency sub-bands of each decomposed layer contain different information. For the top decomposed layer, the activity level of high-frequency coefficients is measured by a PCNN model [17]. For other decomposed layers, the absolute value of each high-frequency coefficient is taken as the activity level value following the absolute (ABS) maximum rule [10]. For low-frequency bands, PC is used as the image feature, whose value is not affected by image brightness, contrast, and illumination intensity. According to the information of PC, LSCM, and LSS, the low-frequency fusion rule is formulated. This rule enhances the detailed features of each source image. Finally, the fused image is reconstructed by performing inverse NSCT on the fused high- and low-frequency images. The main contributions of this paper can be summarized as follows:
  • The high- and low-frequency components of source images are processed separately based on their own features.
  • It applies PCNN and ABS to high-frequency sub-bands of different layers, which achieves a more precise decomposition of high-frequency components.
  • The proposed image fusion algorithm can capture the details of source images well by integrating the advantages of NSCT, PCNN, and PC.
The rest of the sections of this paper are structured as follows: Section 2 proposes an infrared-visible image fusion framework based on an NSCT domain and specifies the corresponding technical details; Section 3 analyzes the results of comparative experiments; and Section 4 concludes this paper.

2. The Proposed Algorithm

The proposed infrared-visible image fusion framework is shown in Figure 1, which has four main steps: image decomposition, the fusion of both high- and low-frequency sub-bands, and image reconstruction. It decomposes source images into 5-layer high- and low-frequency sub-bands first. Then, it applies different methods to the fusion of high- and low-frequency sub-bands, respectively. The decomposed high-frequency sub-bands are further categorized into two parts, H A , l < 5 l , k , H B , l < 5 l , k , and  H A , l = 5 l , k , H B , l = 5 l , k . H A , l < 5 l , k and H B , l < 5 l , k are fused by the method of maximum absolute value. The fused high-frequency sub-bands contain the overall image structure information. PCNN is used to fuse H A , l = 5 l , k and H B , l = 5 l , k . (The related details are explained in the following paragraph.) The fused low-frequency sub-bands retain the detailed information and the residual image information. Finally, it combines the fused high- and low-frequency sub-bands, which can make the fused image more informative.

2.1. NSCT

NSCT can overcome the frequency aliasing phenomenon caused by upsampling and downsampling on CT [18,19]. NSCT is a discrete image calculation framework that achieves shift-invariant, multi-scale, and multi-direction by using non-subsampled pyramid filter banks (NSPFBs) and non-subsampled directional filter banks (NSDFBs). Thus, the proposed solution uses NSCT to decompose source images into high- and low-frequency components.
Two source images are decomposed into high-frequency H A l , k , H B l , k and low-frequency L A , L B bands by performing L-level NSCT decomposition. H A l , k and H B l , k represent the high-frequency components at the decomposition level l and direction k of source image A and B, respectively, while L A and L B are the corresponding low-frequency components of source image A and B, respectively.

2.2. Fusion of High-Frequency Sub-Bands

The high-frequency sub-bands of different decomposed layers contain different information, which retains the overall image structure information. The maximum absolute value and PCNN fusion rules are applied to the fusion of different high-frequency sub-bands, which ensures that the structure information of source images is retained.
As shown in Equation (1), for the high-frequency sub-bands of the decomposed layer l = 5 , the activity level of high-frequency coefficients is measured by PCNN fusion rule. In our previous experiments, we used the different number of decomposition layers to test the performance of the proposed solution many times. According to the objective evaluation metrics, the corresponding results were compared. There is one trade-off between the performance and processing time. The performance of four decomposition layers is poor, and the processing time of six decomposition layers is long. Five decomposition layers can use a relatively short time to achieve a good performance in PCNN fusion. The proposed solution uses two different methods to fuse the high-frequency sub-bands from five decomposition layers. The method of maximum absolute value is used to fuse the high-frequency sub-bands from 1–4 layers. PCNN is applied to the fusion of the high-frequency sub-bands from the 5th layer. The fusion effects of the high-frequency sub-bands can be effectively improved, which is confirmed by the comparative experiments:
H F l , k i , j = H F l , k i , j l = 5 + H F l , k i , j l < 5 .
In Equation (1), H F l , k i , j represents the fused high-frequency coefficients. H F l , k i , j l = 5 represents the 5-level high-frequency fusion coefficients, which can be obtained by Equation (2). Equation (2) integrates the PCNN model, in which the entropy of the absolute value of high-frequency band is used as the network input. Then, the PCNN excitation times of high-frequency components M A , i j l , k N and M B , i j l , k N are calculated by Equation (3), where N denotes the number of iterations:
H F l , k i , j l = 5 = H A l , k i , j l = 5 , i f M A , i j l , k N M B , i j l , k N , H B l , k i , j l = 5 , o t h e r w i s e ,
M i j n = M i j n 1 + P i j n ,
where P i j n denotes the output model of PCNN [17].
Figure 2 shows the architecture of PCNN model used in the proposed image fusion method. In PCNN, F i j [ n ] and L i j [ n ] are the feeding input and the linking input of the neuron at position ( x , y ) in iteration n, respectively, which can be obtained by Equations (4) and (5).
F i j [ n ] = S i j ,
L i j [ n ] = V L k l W i j k l P k l [ n 1 ] ,
where F i j [ n ] is related to the intensity of input image S i j during the whole iteration process. L i j [ n ] is associated with the previous exciting status of eight surrounding neurons through the synaptic weights shown in Equation (6):
W i j k l = 0.5 1.0 0.5 1.0 0.0 1.0 0.5 1.0 0.5 .
The parameter V L represents the amplitude of linking input. U i j [ n ] is the internal activity that consists of two terms, which can be calculated by Equation (7):
U i j [ n ] = e a f U i j [ n 1 ] + F i j [ n ] ( 1 + β L i j [ n ] ) .
In the first term, e a f U i j [ n 1 ] is a decay of its previous value, where the parameter a f is an exponential decay coefficient. The second term F i j [ n ] ( 1 + β L i j [ n ] ) denotes the nonlinear modulation of L i j [ n ] and F i j [ n ] , where the parameter β is the linking strength. The output module P i j [ n ] of the PCNN has two statuses, including excited ( P i j [ n ] = 1 ) and unexcited ( P i j [ n ] = 0 ):
P i j [ n ] = 1 , if U i j [ n ] > E i j [ n 1 ] , 0 , o t h e r w i s e ,
E i j [ n ] = e a e E i j [ n 1 ] + V E P i j [ n ] .
The status depends on its two inputs, which are current internal activity U i j [ n ] and previous dynamic threshold E i j [ n 1 ] . According to Equation (9), the iteration is updating the dynamic threshold, where a e and V E are the exponential decay coefficient and the amplitude of E i j [ n ] , respectively.
Similarly, H F l , k i , j l < 5 represents the 1-to-4 level high-frequency fusion coefficients, which can be obtained by Equation (10):
H F l , k i , j l < 5 = H A l , k i , j l < 5 , E n t r o p y ( H A l , k l < 5 ) E n t r o p y ( H B l , k l < 5 ) , H B l , k i , j l < 5 , o t h e r w i s e .
In Equation (10), E n t r o p y ( H A l , k l < 5 ) and E n t r o p y ( H B l , k l < 5 ) represent the information entropy of high-frequency components H A l , k and H B l , k , respectively. The information entropy of high frequency component H x l , k can be calculated by Equation (11):
E n t r o p y ( H x l , k ) = 1 m × n j = 1 n i = 1 m log 2 H x l , k ( i , j ) ,
where m and n are the total column and row number of H x l , k , and  | H x l , k ( i , j ) | is the maximum entropy of the ABS. The maximum entropy of the ABS is used as the fusion measurement of high-frequency sub-bands.

2.3. Fusion Rule of Low-Frequency Sub-Bands

The low-frequency sub-bands of NSCT filtered images mainly describe the detailed information that corresponds to the texture and edge information of source images. In medical imaging, organ or cell lesions are often identified by the detailed information. Thus, the enhancement of detailed features from each source image is the key of low-frequency sub-bands fusion.
This paper uses PC to enhance image features that make low-frequency sub-bands more informative. PC as a dimensionless measure can evaluate the significance of each image feature. In low-frequency sub-bands, PC value reflects the sharpness of image object. Thus, PC is used as the phase of the coefficient with maximal local sharpness. Since an image can be regarded as 2D signals [9], PC of an image pixel at location (x,y) can be calculated by Equation (12).
P C ( x , y ) = k E θ k ( x , y ) ε + n k A n , θ k ( x , y )
where θ k is the orientation angle at k scale [9], A n , θ k denotes the amplitude of the n-th Fourier component, and angle θ k , ε is a positive constant to remove the PC components of image signals. E θ k ( x , y ) can be calculated by Equation (13):
E θ k ( x , y ) = F 2 θ k ( x , y ) + H 2 θ k ( x , y ) ,
where F θ k ( x , y ) = n b n , θ k ( x , y ) and H θ k ( x , y ) = n c n , θ k ( x , y ) . b n , θ k ( x , y ) and c n , θ k ( x , y ) are the convolution results of input image pixel at location (x,y), which can be evaluated by Equation (14):
[ b n , θ k ( x , y ) , c n , θ k ( x , y ) ] = [ I L ( x , y ) M n b , I ( x , y ) M n c ] ,
where I L ( x , y ) is the low-frequency image pixel value at location ( x , y ) . M n b and M n c are the even- and odd-symmetry filters of 2D log-Gabor at scale n. As a contrast invariant, PC has defects that do not reflect the local contrast changes. To compensate the lack of PC, a measure of sharpness change (SCM) shown in Equation (15) is developed:
S C M x , y = x 0 , y 0 Ω 0 I L x , y I L x 0 , y 0 2 ,
where Ω 0 represents a local area at location ( x , y ) . Meanwhile, LSCM shown in Equation (16) is introduced to calculate the contrast of location ( x , y ) neighborhood:
L S C M x , y = i = M M j = N N S C M x + i , y + j ,
where 2 M + 1 × 2 N + 1 denotes the neighborhood size. Since LSCM and PC cannot fully reflect the local signal strength, LSS shown in Equation (17) is introduced:
L S S x , y = M a x i ( M , M ) M a x j ( N , N ) x i j μ M N ,
where x i j is the pixel in location of this image patch, μ M N represents the mean value of this image patch.
As shown in Equation (18), a global measurement (GM) is proposed that integrates PC, LSCM, and LSS complements to measure different aspects of image information:
G M x , y = P C x , y α · L S C M x , y β · L S S x , y γ ,
where α , β , and  γ are the parameters used in GM to adjust PC, LSCM, and LSS, respectively. When GM is obtained, the fused image of low-frequency sub-bands can be calculated by the rule proposed in Equation (19):
L F ( x , y ) = L A x , y , i f L m a p A x , y = 1 , L B x , y , o t h e r w i s e ,
where L F ( x , y ) , L A ( x , y ) , L B ( x , y ) are low-frequency sub-bands of the fused image, source image I A and I B , respectively. L m a p i x , y denotes a decision map for the fusion of low-frequency sub-bands, which can be calculated by Equation (20):
L m a p i x , y = 1 , i f Φ i x , y > M ˜ × N ˜ 2 , 0 , o t h e r w i s e ,
where is the cardinality of a set, and  Φ i ( x , y ) can be calculated by Equation (21). The cardinality of a set is helpful to obtain the abundant image details and structure information:
Φ i x , y = x 0 , y 0 Ω 1 | G M i x 0 , y 0 max ( G M 1 x 0 , y 0 , . . . , G M i 1 x 0 , y 0 , G M i + 1 x 0 , y 0 , . . . , G M K x 0 , y 0 )
where Ω 1 represents a sliding window with a size of M ˜ × N ˜ centered at location (x,y), and K is the number of source images. GM defined in Equation (18) is expressed as a general term. In Equation (21), the subscript of GM is used to select the corresponding maximum value from source images.
For input source images A and B, the high-frequency components H A l , k , H B l , k and low-frequency components L A , L B are first obtained by NSCT decomposition. The activity level of high-frequency components H A l , k , H B l , k is then measured by using the absolute maximum rule and PCNN model. Meanwhile, it applies PC to the fusion of low-frequency components. Finally, the fused high- and low-frequency components H F and L F are inversely transformed by NSCT to obtain the fused image I F . Algorithm 1 shows the main steps of the proposed infrared-visible image fusion solution.
Algorithm 1 The proposed infrared-visible image fusion algorithm
Input:
  •  source image A and B
  •  Parameters: decomposition layer l, decomposition direction k
Input:
  •  fused image F
1:
for each source image A and B do
2:
Decompose source image A and B into corresponding high- and low-frequency sub-bands H A l , k , H B l , k and L A , L B by NSCT respectively
3:
end for
4:
for each decomposed layer do
5:
if the decomposed layer l = 5 of high-frequency sub-bands then
6:
  Measure the activity level of high-frequency coefficients by PCNN.
7:
  Obtain the 5 t h layer fusion coefficient of high-frequency sub-bands by PCNN as follows:
   H F l , k i , j l = 5 = H A l , k i , j l = 5 ,   i f M A , i j l , k N M B , i j l , k N H B l , k i , j l = 5 , o t h e r w i s e
8:
end if
9:
if the decomposed layer l < 5 of high-frequency sub-bands then
10:
  Use the maximum entropy of the ABS of coefficient as the actually measured value of activity level.
11:
  Obtain the first four-layer coefficients of high-frequency sub-bands as follows:
   H F l , k i , j l < 5 = H A l , k i , j l < 5 , E n t r o p y ( H A l , k l < 5 ) E n t r o p y ( H B l , k l < 5 ) H B l , k i , j l < 5 , o t h e r w i s e
12:
end if
13:
 The obtained image of high-frequency sub-bands is H F l , k i , j = H F l , k i , j l = 5 + H F l , k i , j l < 5
14:
end for
15:
for each source image A and B do
16:
It uses PC, LSCM, and LSS to design GM:
G M x , y = P C x , y α · L S C M x , y β · L S S x , y γ
17:
 Calculate the image of low-frequency sub-bands by the following rule:
L F ( x , y ) = L A x , y i f L m a p A x , y = 1 L B x , y o t h e r w i s e
18:
end for
19:
The inverse NSCT is applied to the obtained images of high- and low-frequency sub-bands H F , L F to get the fused image F.

3. Comparative Experiments

3.1. Experiment Preparation

In comparative experiments, 30 sets of infrared-visible images are used to test the fusion performance. The resolution of test images are 256 256 . The infrared wavelength is 700–2526 nm, and the visible wavelength is 390–700 nm. Infrared-visible image pairs were collected by Liu [10] and can be downloaded from quxiaobo.org. All the experiment’s program’s codes are programmed in Matlab 2014a (MathWorks, Natick, MA, USA) on an Intel(R) Core(TM)i7-4790CPU (Intel, Santa Clara, CA, USA) @ 3.60 GHz Desktop with 8.00 GB RAM.

3.2. Objective Evaluation Metrics

For the evaluation of fused image, a single evaluation metric cannot fully reflect the performance of fused image [20,21]. Therefore, it is necessary to use multiple metrics to do comprehensive performance analysis. This paper uses five objective metrics to evaluate the performances of different fusion methods, which include Q T E [22,23], Q A B / F [24,25], Q M I [23], Q C B [23,26], and Q V I F [25,27]. Q T E is used to evaluate the Tsallis entropy of the fused image. Q A B / F as a gradient-based quality index measures the edge information. Q M I is used to evaluate the similarity between the fused image and source images. Both Q C B and Q V I F measure the human visual performance of the fused image.

3.3. Experiment Results of Infrared-Visible Image Fusion

In this section, the proposed NSCT-based fusion framework is compared with seven popular fusion methods, such as the adaptive spare representation (ASR) based image fusion method proposed by Liu [28], the convolutional neural network (CNN) based image fusion method proposed by Liu [29], the multi-channel medical image fusion (CT) proposed by Zhu [25], the multi-modality image fusion method with joint patch clustering based dictionary learning (KIM) proposed by Kim [30], the image fusion based on multi-scale transform and sparse representation (MST-SR) proposed by Liu [10], a novel infrared and visual image fusion algorithm based on NSST and improved PCNN (NSST-PCNN) was proposed by Li [31], and an infrared and visible image fusion scheme based on NSCT and PC information (NSCT-PC) proposed by Li [9]. This section only picks the fused results of six comparative experiments from thirty attempts to analyze the fusion performance.

3.3.1. Comparative Experiments—1

Figure 3 shows the fused results of infrared-visible image fusion experiment—1. As shown in Figure 3c,f, the fused images obtained by ASR and KIM have low brightness. The light brightness in source image (a) is not well preserved in both (c) and (f), so images (c) and (f) have overall poor visual performance. The CNN method does not perform well in some local areas as shown in Figure 3d. According to the partially enlarged areas in Figure 3e, some local areas of the fused image obtained by CT have high brightness, and the image detailed information is not obvious. In Figure 3i, the saturation of the fused image is high, and the edge detailed information is not obvious. In addition, the fused image obtained by NSCT-PCNN has low contrast, and the global image features have poor performance. Compared the fused images (h) and (j) as well as the corresponding partially magnified images in Figure 3, NSCT-PC and the proposed method have the close visual performance of human eyes.

3.3.2. Comparative Experiments—2

Figure 4 shows the fused results of infrared-visible image fusion experiment—2. After the comparisons of fused images obtained by different methods, it gets the following conclusions. In Figure 4c,f, the fused images obtained by ASR and KIM have low brightness, and poor performance in global features. As shown in the magnified areas of Figure 4d,e, CNN does not obtain the clear details of fused image, the contrast of the partially enlarged image obtained by CT is high, and the corresponding edge information is not obvious. For the fused image (f) in Figure 4 obtained by KIM, the connection area of sky and forest has high edge brightness. As shown in Figure 4h, the fused image obtained by NSCT-PCNN has high brightness and poor visual effect. Compared with the experiment results of the other six fusion methods, the fused images obtained by NSCT-PC and the proposed method have better fusion performance.

3.3.3. Comparative Experiments—3

Figure 5 shows the fused results of infrared-visible image fusion experiment—3. In Figure 5c,f, both ASR and KIM obtain the fused images with high brightness, and do not preserve the detailed information of source image (b). Comparing with ASR, the fused image obtained by KIM is fuzzy and not conducive to human-eye observation. As shown in Figure 5d, the fused image obtained by CNN has high saturation. In Figure 5e,g, the detail texture information of fused images obtained by CT and MST-SR is not clear by observing the partially enlarged areas. Compared with the proposed method, the fused image obtained by NSCT-PCNN in Figure 5h has low saturation and poor performance in global features. As shown in Figure 5i,j, NSCT-PC and the proposed method have good performance in both global and local features.

3.3.4. Comparative Experiments—4

Figure 6 shows the fused results of infrared-visible image fusion experiment—4. In Figure 6c, the fused image obtained by ASR has a general visualization performance. As shown in Figure 6d, the car light has high brightness in the fused image obtained by CNN. In Figure 6e,g, the fused images obtained by CT and MST-SR are dark, and have poor overall visualization performance. As shown in Figure 6f,h, the fused images obtained by KIM and NSCT-PC have high brightness. After the analysis of detailed information, the detailed textures of fused images are not obvious, which are not conducive to human-eye observation. Comparing with NSCT-PC, the proposed method has better performance in both global and local features of source images.

3.3.5. Analysis of Comparative Experiment Results

As the analysis of 30 comparative experiments, Table 1 and Figure 7 show the average objective evaluation results of infrared-visible image fusion. In Table 1, all the best results are marked in bold. According to the results shown in Table 1 and Figure 7, the proposed method achieves the best performance in Q A B / F , Q M I , Q C B , and Q V I F , and the second best performance in Q T E . Q T E of the proposed method is a little bit lower than the best one obtained by NSST-PCNN. It means that both the proposed method and NSST-PCNN can retain more information of source images. Meanwhile, the similarities between the fused images obtained by these two methods and source images are also comparable. For the Q A B / F metric, the proposed method is slightly higher than other methods. Thus, the proposed method performs better in the preservation of image edge details. Additionally, the proposed method can also preserve the global and local features of source images well, and also achieve a good performance in human-eye visualization. As shown in Figure 7, the proposed method uses the shortest processing time in infrared-visible image fusion among all the eight fusion methods, which is much less than others as well as about 40% of the second shortest processing time. Thus, the results of comparative experiments confirm that the proposed infrared-visible image fusion solution has a low algorithm complexity and can effectively reduce the related costs.

4. Conclusions

In this paper, an NSCT-based precise high-frequency decomposition method for infrared-visible image fusion is proposed. The fusion method combines NSCT, PCNN model, and PC information to improve the visual quality of fused images. Specifically, the method uses NSCT to achieve the high- and low-frequency decomposition of source images. The fusion of high-frequency image coefficients is realized by introducing PCNN and ABS as the activity metrics of high-frequency coefficients. In the fusion of low-frequency components, it integrates the fusion rules of LSCM, LSS, and PC features to achieve the energy preservation and detail extraction of low-frequency components. Finally, the fused image is obtained by inverse NSCT over the fused high- and low-frequency components. Compared to other image fusion methods, the proposed method achieves good performance on the structural similarity and detail preservation in fused images. The experiment results confirm that the proposed method has good effectiveness and high speed in infrared-visible image fusion.
In the future, the proposed method will be optimized to increase the processing speed. A weighted fusion will be explored to improve the fusion performance. The statistical tests, such as Friedman’s test, will be introduced to compare the performance of the proposed method. The proposed image fusion method will also be extended to other multi-modality image fusion areas, such as medical image fusion, multi-focus image fusion, and so on as well as face recognition, especially in night scenes.

Author Contributions

Conceptualization, X.H. and G.Q.; Data curation, X.H. and H.W.; Formal analysis, G.Q. and H.W.; Funding acquisition, X.H. and H.W.; Investigation, G.Q.; Methodology, G.Q. and J.S.; Project administration, G.Q. and Y.C.; Software, G.Q.; Validation, H.W.; Writing—original draft, H.W.; Writing—review and editing, G.Q., Y.C., and J.S.

Funding

This work is jointly supported by the National Natural Science Foundation of China under Grant No. 61906026; the Common Key Technology Innovation Special of Key Industries of Chongqing science and Technology Commission under Grant No. cstc2017zdcy-zdyfX0067 and cstc2017zdcy-zdyfX0055; and the Artificial Intelligence Technology Innovation Significant Theme Special Project of Chongqing Science and Technology Commission under Grant No. cstc2017rgzn-zdyfX0014 and cstc2017rgzn-zdyfX0035.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qi, G.; Zhu, Z.; Chen, Y.; Wang, J.; Zhang, Q.; Zeng, F. Morphology-based visible-infrared image fusion framework for smart city. Int. J. Simul. Process Model. 2018, 13, 523–536. [Google Scholar] [CrossRef]
  2. Liu, C.H.; Qi, Y.; Ding, W.R. Infrared and visible image fusion method based on saliency detection in sparse domain. Infrared Phys. Technol. 2017, 83, 94–102. [Google Scholar] [CrossRef]
  3. Li, H.; Li, X.; Yu, Z.; Mao, C. Multifocus image fusion by combining with mixed-order structure tensors and multiscale neighborhood. Inf. Sci. 2016, 349, 25–49. [Google Scholar] [CrossRef]
  4. Li, Y.; Sun, Y.; Zheng, M.; Huang, X.; Qi, G.; Hu, H.; Zhu, Z. A Novel Multi-Exposure Image Fusion Method Based on Adaptive Patch Structure. Entropy 2018, 20, 935. [Google Scholar] [CrossRef]
  5. Yin, L.; Zheng, M.; Qi, G.; Zhu, Z.; Jin, F.; Sim, J. A Novel Image Fusion Framework Based on Sparse Representation and Pulse Coupled Neural Network. IEEE Access 2019, 7, 98290–98305. [Google Scholar] [CrossRef]
  6. Naik, V.V.; Gharge, S. Satellite image resolution enhancement using DTCWT and DTCWT based fusion. In Proceedings of the International Conference on Advances in Computing, Jaipur, India, 21–24 September 2016; pp. 1957–1962. [Google Scholar]
  7. Zhu, Z.; Zheng, M.; Qi, G.; Wang, D.; Xiang, Y. A Phase Congruency and Local Laplacian Energy Based Multi-Modality Medical Image Fusion Method in NSCT Domain. IEEE Access 2019, 7, 20811–20824. [Google Scholar] [CrossRef]
  8. Zhang, Q.; Maldague, X. An adaptive fusion approach for infrared and visible images based on NSCT and compressed sensing. Infrared Phys. Technol. 2016, 74, 11–20. [Google Scholar] [CrossRef]
  9. Li, H.; Qiu, H.; Yu, Z.; Zhang, Y. Infrared and visible image fusion scheme based on NSCT and low-level visual features. Infrared Phys. Technol. 2016, 76, 174–184. [Google Scholar] [CrossRef]
  10. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  11. Qi, G.; Zhu, Z.; Erqinhu, K.; Chen, Y.; Chai, Y.; Sun, J. Fault-diagnosis for reciprocating compressors using big data and machine learning. Simul. Model. Pract. Theory 2018, 80, 104–127. [Google Scholar] [CrossRef]
  12. Li, H.; Qiu, H.; Yu, Z.; Li, B. Multifocus image fusion via fixed window technique of multiscale images and non-local means filtering. Sign Process. 2017, 138, 71–85. [Google Scholar] [CrossRef]
  13. Ding, W.; Bi, D.; He, L.; Fan, Z. Infrared and visible image fusion method based on sparse features. Infrared Phys. Technol. 2018, 92, 372–380. [Google Scholar] [CrossRef]
  14. Kong, W.; Wang, B.; Yang, L. Technique for infrared and visible image fusion based on non-subsampled shearlet transform and spiking cortical model. Infrared Phys. Technol. 2015, 71, 87–98. [Google Scholar] [CrossRef]
  15. Xiang, T.; Li, Y.; Gao, R. A fusion algorithm for infrared and visible images based on adaptive dual-channel unit-linking PCNN in NSCT domain. Infrared Phys. Technol. 2015, 69, 53–61. [Google Scholar] [CrossRef]
  16. Liu, Y.; Liu, S.; Wang, Z. Infrared and visible image fusion based on random projection and sparse representation. Int. J. Remote Sens. 2014, 35, 1640–1652. [Google Scholar]
  17. Yin, M.; Liu, X.; Liu, Y.; Chen, X. Medical Image Fusion With Parameter-Adaptive Pulse Coupled Neural Network in Nonsubsampled Shearlet Transform Domain. IEEE Trans. Instrum. Meas. 2019, 68, 49–64. [Google Scholar] [CrossRef]
  18. Da Cunha, A.L.; Zhou, J.; Do, M.N. The nonsubsampled contourlet transform: Theory, design, and applications. IEEE Trans. Image Process. 2006, 15, 3089–3101. [Google Scholar] [CrossRef]
  19. Li, Y.; Sun, Y.; Huang, X.; Qi, G.; Zheng, M.; Zhu, Z. An Image Fusion Method Based on Sparse Representation and Sum Modified-Laplacian in NSCT Domain. Entropy 2018, 20, 522. [Google Scholar] [CrossRef]
  20. Li, H.; He, X.; Tao, D.; Tang, Y.; Wang, R. Joint medical image fusion, denoising and enhancement via discriminative low-rank sparse dictionaries learning. Pattern Recognit. 2018, 79, 130–146. [Google Scholar] [CrossRef]
  21. Li, H.; Wang, Y.; Yang, Z.; Wang, R.; Li, X.; Tao, D. Discriminative dictionary learning-based multiple component decomposition for detail-preserving noisy image fusion. IEEE Trans. Instrum. Meas. 2019. [Google Scholar] [CrossRef]
  22. Cvejic, N.; Canagarajah, C.N.; Bull, D.R. Image fusion metric based on mutual information and Tsallis entropy. Electron. Lett. 2006, 42, 626–627. [Google Scholar] [CrossRef]
  23. Liu, Z.; Blasch, E.; Xue, Z.; Zhao, J.; Laganière, R.; Wu, W. Objective Assessment of Multiresolution Image Fusion Algorithms for Context Enhancement in Night Vision: A Comparative Study. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 94–109. [Google Scholar] [CrossRef] [PubMed]
  24. Petrovi, V. Subjective tests for image fusion evaluation and objective metric validation. Inf. Fusion 2007, 8, 208–216. [Google Scholar] [CrossRef]
  25. Zhu, Z.; Yin, H.; Chai, Y.; Li, Y.; Qi, G. A Novel Multi-modality Image Fusion Method Based on Image Decomposition and Sparse Representation. Inf. Sci. 2018, 432, 516–529. [Google Scholar] [CrossRef]
  26. Chen, Y.; Blum, R.S. A new automated quality assessment algorithm for image fusion. Image Vis. Comput. 2009, 27, 1421–1432. [Google Scholar] [CrossRef]
  27. Sheikh, H.R.; Bovik, A.C. Image information and visual quality. IEEE Trans. Image Process. 2006, 15, 430–444. [Google Scholar] [CrossRef]
  28. Liu, Y.; Wang, Z. Simultaneous image fusion and denoising with adaptive sparse representation. Image Process. IET 2014, 9, 347–357. [Google Scholar] [CrossRef]
  29. Liu, Y.; Chen, X.; Cheng, J.; Peng, H. A medical image fusion method based on convolutional neural networks. In Proceedings of the 2017 20th International Conference on Information Fusion (Fusion), Xi’an, China, 10–13 July 2017; pp. 1–7. [Google Scholar]
  30. Kim, M.; Han, D.K.; Ko, H. Joint patch clustering-based dictionary learning for multimodal image fusion. Inf. Fusion 2016, 27, 198–214. [Google Scholar] [CrossRef]
  31. Li, M.; Yuan, X.; Luo, Z.; Qiu, X. Infrared and Visual Image Fusion Based on NSST and Improved PCNN. J. Phys. Conf. Ser. 2018, 1069, 012151. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The proposed infrared-visible image fusion framework.
Figure 1. The proposed infrared-visible image fusion framework.
Entropy 21 01135 g001
Figure 2. Architecture of the PCNN model used in the proposed method.
Figure 2. Architecture of the PCNN model used in the proposed method.
Entropy 21 01135 g002
Figure 3. Infrared-visible image fusion comparative experiments—1. (a,b) are source images, (cj) are the fused results of ASR, CNN, CT, KIM, MST-SR, NSCT-PCNN, NSCT-PC, and the proposed method, respectively. At the bottom of each image, two areas marked in green and red dashed line frames correspond to the magnified areas encompassed in green and red frames, respectively.
Figure 3. Infrared-visible image fusion comparative experiments—1. (a,b) are source images, (cj) are the fused results of ASR, CNN, CT, KIM, MST-SR, NSCT-PCNN, NSCT-PC, and the proposed method, respectively. At the bottom of each image, two areas marked in green and red dashed line frames correspond to the magnified areas encompassed in green and red frames, respectively.
Entropy 21 01135 g003
Figure 4. Infrared-visible image fusion comparative experiments—2. (a,b) are source images, (cj) are the fused results of ASR, CNN, CT, KIM, MST-SR, NSCT-PCNN, NSCT-PC, and the proposed method, respectively. At the bottom of each image, two areas marked in green and red dashed line frames correspond to the magnified areas encompassed in green and red frames, respectively.
Figure 4. Infrared-visible image fusion comparative experiments—2. (a,b) are source images, (cj) are the fused results of ASR, CNN, CT, KIM, MST-SR, NSCT-PCNN, NSCT-PC, and the proposed method, respectively. At the bottom of each image, two areas marked in green and red dashed line frames correspond to the magnified areas encompassed in green and red frames, respectively.
Entropy 21 01135 g004
Figure 5. Infrared-visible image fusion comparative experiments—3. (a,b) are source images, (cj) are the fused results of ASR, CNN, CT, KIM, MST-SR, NSCT-PCNN, NSCT-PCand the proposed method, respectively. At the bottom of each image, two areas marked in green and red dashed line frames correspond to the magnified areas encompassed in green and red frames, respectively.
Figure 5. Infrared-visible image fusion comparative experiments—3. (a,b) are source images, (cj) are the fused results of ASR, CNN, CT, KIM, MST-SR, NSCT-PCNN, NSCT-PCand the proposed method, respectively. At the bottom of each image, two areas marked in green and red dashed line frames correspond to the magnified areas encompassed in green and red frames, respectively.
Entropy 21 01135 g005
Figure 6. Infrared-visible Image Fusion Comparative Experiments—4. (a,b) are source images, (cj) are the fused results of ASR, CNN, CT, KIM, MST-SR, NSCT-PCNN, NSCT-PCand the proposed method, respectively. At the bottom of each image, two areas marked in green and red dashed line frames correspond to the magnified areas encompassed in green and red frames, respectively.
Figure 6. Infrared-visible Image Fusion Comparative Experiments—4. (a,b) are source images, (cj) are the fused results of ASR, CNN, CT, KIM, MST-SR, NSCT-PCNN, NSCT-PCand the proposed method, respectively. At the bottom of each image, two areas marked in green and red dashed line frames correspond to the magnified areas encompassed in green and red frames, respectively.
Entropy 21 01135 g006
Figure 7. Average objective evaluations of thirty infrared-visible image fusion comparative experiments.
Figure 7. Average objective evaluations of thirty infrared-visible image fusion comparative experiments.
Entropy 21 01135 g007
Table 1. Average objective evaluations of thirty infrared-visible image fusion comparative experiments.
Table 1. Average objective evaluations of thirty infrared-visible image fusion comparative experiments.
Q T E Q A B / F Q M I Q C B Q V I F
ASR0.41230.64701.93540.53340.4250
CNN0.44020.45282.09520.52880.4582
CT0.39310.56391.85110.49650.3848
KIM0.38960.60111.84080.49660.4062
MSR-SR0.41950.68882.01320.55240.4563
NSST-PCNN0.46970.60822.15460.53080.4299
NSCT-PC0.42620.66392.03370.56080.4352
Proposed0.45410.71222.18130.56220.4811

Share and Cite

MDPI and ACS Style

Huang, X.; Qi, G.; Wei, H.; Chai, Y.; Sim, J. A Novel Infrared and Visible Image Information Fusion Method Based on Phase Congruency and Image Entropy. Entropy 2019, 21, 1135. https://doi.org/10.3390/e21121135

AMA Style

Huang X, Qi G, Wei H, Chai Y, Sim J. A Novel Infrared and Visible Image Information Fusion Method Based on Phase Congruency and Image Entropy. Entropy. 2019; 21(12):1135. https://doi.org/10.3390/e21121135

Chicago/Turabian Style

Huang, Xinghua, Guanqiu Qi, Hongyan Wei, Yi Chai, and Jaesung Sim. 2019. "A Novel Infrared and Visible Image Information Fusion Method Based on Phase Congruency and Image Entropy" Entropy 21, no. 12: 1135. https://doi.org/10.3390/e21121135

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop