Next Article in Journal
Drone Mission Definition and Implementation for Automated Infrastructure Inspection Using Airborne Sensors
Next Article in Special Issue
Development of Noninvasive Classification Methods for Different Roasting Degrees of Coffee Beans Using Hyperspectral Imaging
Previous Article in Journal
Phase-Shifted Eccentric Core Fiber Bragg Grating Fabricated by Electric Arc Discharge for Directional Bending Measurement
Previous Article in Special Issue
Edge-Aware Unidirectional Total Variation Model for Stripe Non-Uniformity Correction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Infrared and Visible Image Fusion Based on Different Constraints in the Non-Subsampled Shearlet Transform Domain

1
Aeronautics and Astronautics Engineering College, Air Force Engineering University, Xi’an 710038, Shaanxi, China
2
School of Management Engineering, Xi’an University of Finance and Economics, Xi’an 710100, Shaanxi, China
3
The 93575 Unit of PLA, Chengde 067000, Hebei, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(4), 1169; https://doi.org/10.3390/s18041169
Submission received: 5 March 2018 / Revised: 7 April 2018 / Accepted: 8 April 2018 / Published: 11 April 2018
(This article belongs to the Special Issue Advances in Infrared Imaging: Sensing, Exploitation and Applications)

Abstract

:
There are many artificial parameters when fuse infrared and visible images, to overcome the lack of detail in the fusion image because of the artifacts, a novel fusion algorithm for infrared and visible images that is based on different constraints in non-subsampled shearlet transform (NSST) domain is proposed. There are high bands and low bands of images that are decomposed by the NSST. After analyzing the characters of the bands, fusing the high level bands by the gradient constraint, the fused image can obtain more details; fusing the low bands by the constraint of saliency in the images, the targets are more salient. Before the inverse NSST, the Nash equilibrium is used to update the coefficient. The fused images and the quantitative results demonstrate that our method is more effective in reserving details and highlighting the targets when compared with other state-of-the-art methods.

1. Introduction

Image fusion plays a very important role in many fields. In remote sensing [1], multispectral images and panchromatic images are integrated into an image; the fused image can obtain quality spatial resolution and the quality spectral resolution. In medical imaging [2], the MRI images and PET images are fused to help diagnose the condition. In modern military technologies [3,4], the fusion of thermal infrared (IR) images and visible images is becoming more important for night action. In this paper, we concentrate on the fusion of IR images and visible images, which can lead to better performance for target recognition and human visual perception.
The goal of IR and visible image fusion is to integrate them into an image, which can contain most information in the source images without distortion or loss. The visible spectrum is better in spatial resolution, but it is easily affected by illumination and disguise. In contrast, IR images are not affected by illumination and disguise, but do suffer form low resolution. Therefore, integrating the two types of images can be beneficial for target detection and recognition at night.
The image fusion algorithm can be divided into two categories. One type is fusion algorithms, which achieve in the spatial domain, such as the weight method, pyramid transform method [5], principal component analysis (PCA) [6], saliency method [7], and so on. Those algorithms are simple to implement, and they can obtain some useful images, but the fused images lose information on occasion sometimes. Another type of fusion algorithms fuse images in the transform domain. Chao [8] fused images based on wavelets in 2004, in the high- and low-level bands of the fused the images, respectively. The result is effective, but loses some orientation information in the image. Due to the insufficient orientations in wavelets, Do [9] proposed the contourlet transform based on wavelets. Wang [10] fused IR images and visible images based on the contourlet transform, and the fused images perform well with respect to the image details, but with sampling, there are Gibbs phenomena in the final images. Thus, the non-subsampled contourlet transform is used in the fusion algorithm [11], overcoming the shortcoming of the Gibbs phenomenon. Guo [12,13,14] proposed the shearlet transform in 2008, and it not only possessed the above properties, but it is equipped with a rich mathematical structure that is associated to multi-resolution analyses, and it forms a tight frame at various levels and directions, reducing the complexity and improving the efficiency of the operation. Wang [15] proposed a novel fusion algorithm that is based on improved shearlet transform to enhance the quality of the fused image. Wan [16] proposed a method based on the guided filter and improved sum-modified-Laplacian (SML) in the non-subsampled shearlet transform (NSST) domain. However, most of the above algorithms set a large number of artificial parameters, and there are many artifacts in the final results that seriously affect the target edge and the background information.
Ma [17] proposed an infrared and visible image fusion algorithm that is based on the total variation, the optimal image is calculated by iterations, and the details in the visible image can be retained as much as possible, while also maintaining the target characters in the IR image. However, the fusion results look like an enhanced IR image, which is not suitable for many applications. To improve the quality of the fused image, we proposed a novel fusion algorithm for IR images and visible images. On one hand, the details in the visible image can be reserved; on the other hand, the salient target and some important information are transformed in the final image. The algorithm fuses images that are based on the NSST, but uses the gradient and saliency constraints in the fusion rules.
The contributions of this paper are as follows: we take advantage of the total variation minimization to optimize the fusion image based on the NSST domain, and both details and target information are well reserved. In addition, we provide both qualitative and quantitative comparisons with several state-of-the-art algorithms on a publicly available dataset. When compared to those algorithms, our results are more suitable for human vision systems and processing, and thus, improve the reliability of automatic target detection and recognition systems.
The rest of the paper is organized as follows: Section 2 introduces the related NSST, and, at the same time, we establish the constraints based on the gradient and the saliency of the rules of different levels. In Section 3, we compare our approach with several of the most advanced methods. Section 4 summarizes the full text.

2. Methods

2.1. Preliminaries

When n = 2, the affine systems with composite dilations are defined as follows [10,11]:
{ ψ j , l , k ( x ) = | det B | j / 2 ψ ( A l B j x k ) : l , j Z , k Z 2 }
where ψ L 2 ( R 2 ) , A and B are 2 × 2 invertible matrices with det(B) = 1, A and B control the level and direction of the transform, respectively, and j, l and k are the parameters that control the levels, directions, and transform, respectively. The values of the two matrices play an important role in the transform, and the values of A and B for the shearlet transform are defined as follows:
A = ( a 0 0 a ) , B = ( 1 s 0 1 )
where a = 4 and s = 1 in the Equation (2).
When processing the digital image using the shearlet transform, we need the discrete shearlet transform. There are subsamples in the operation of discretization, so there is the impossibility of Gibbs phenomenon in the images. To overcome the shortage, the NSST is proposed [14].
The NSST can be more sparse and effective when simulating the images, which are two-dimensional, and it can be considered as the optimal tool to analyze two-dimensional data. There are also some methods of image fusion that are based on the NSST in recent years, which can reserve the characteristics of different frequencies, but the universal fusion rule relies on artificial parameters, which leads to the artifacts. Therefore, this paper proposes a fusion rule that is based on different constraints in the NSST domain to improve the quality of the fused image.

2.2. Proposed Method

2.2.1. Model of Image Fusion

The model of the traditional infrared and visible image fusion is mainly expressed as Equation (3):
f ˜ = α · I + β · V
In Equation (3), f ˜ represents the fused image, α and β represent the weight factors, respectively, I represents the infrared image, and V represents the visible image. In general, the setting of the weight factors needs to be man-made. Although this approach is convenient, the quality of the fused image is not optimal, and results in artifacts. Thus, in this paper, we assume the optimal fused image is f . The relationship between the optimal fused image and the original images can be expressed by Equation (4):
min f α · I β · V 2 2
Equation (4) is the fidelity constraint of the fused image, the result is similar to the fused image by Equation (3), rely on the Equation (4) only, and cannot avoid the existence of artifacts. Therefore, in different frequency bands, we build more constraints for different frequency bands, according to their characteristics.

2.2.2. Strategy for High Frequency Coefficients

The low-frequency coefficient fd and the high-frequency coefficient group {fg} can be obtained by an image f transform by the NSST. The high-frequency coefficients contain a great deal of detailed information [18] and the regularization term of the gradient constraint is used in our method to preserve the information better:
i f = ( i h f ) 2 + ( i v f ) 2
As Equation (5) shows, i f represent the gradient value of pixel i in the image f. i h f = f i f r ( i ) and i v f = f i f b ( i ) represent the neighbor to the right and below pixel i. When pixel i is located in the edges, the value of its gradient is large; on the other hand, when it is located in the smooth area, the value is small. The f(i) represents the value of the image in the pixel i, fr(i) represents the value of infrared image in the pixel i, and fr(i) represents the value of the visible image in the pixel i.
After the fusion, the fused image should preserve the advantage of the gradient in the infrared image or visible image as much as possible. If the gradients of some area in the different images are small, then the area should be inhibited because of the little information that is contained in this area.
As Ma [19] mentioned in the paper, they constrained the fused image to have a similar gradient with the visible image, as the equation below:
min f I 0
However, in infrared images, there is still some useful information for the gradient, so we constrain the gradient of the fused image f as:
min f g W I · I g W V · V g 1
where the f g is the gradient of the high-level frequency that should be solved, I g and V g are the gradients of the original images respectively, and WI and WV are the weights for the infrared image and the visible image, respectively. It is widely known that the l0 norm is difficult to solve, and the l1 norm encourages sparsity, thus, we consider minimizing the constraints that use the l1 norm as superior. We compare the l0 norm and the l1 norm, and the result is shown in Figure 1.
Figure 1 shows the high-frequency coefficients in the first iteration of the algorithm implementation in this paper, analyzing the data in the blue frame area. The l0 norm constraint should theoretically be 35, but the actual operation needs to set the threshold ε in order to determine its value. If ε = 0.5, then the value in the green frame area should be 0, and the value in the blue area will be smaller when ε is reduced. The l1 norm is constrained directly by the absolute value, sparseness is lacking, but more precise, and the calculation is simple. Thus, we choose the two coefficients of the l1 norm constraint.
The weights of the two source images are weighted, and the specific weights are set as follows:
W I i = I g i i N ( I g i + V g i )
Equation (8) represents the weight value at pixel i, N represents the number of pixels adjacent to i, and we select a smaller 3 × 3 pixel block for the weight calculation. By the weighting of Equation (8), the gradient difference is increased and the calculation is more favorable.
The fusion rules of the high frequency part can be set by the minimization cost function of Equation (9):
E g = f g α · I g β · V g 2 2 + f g W I · I g W V · V g 1
In Equation (9), the unknown parameters have the optimal fusion fg and the weight parameters α and β, which is an unsolvable problem. To simplify the solution, we set the relationship between the parameters α and β and update Equation (9):
E g = f g α · I g ( 1 α ) · V g 2 2 + f g W I · I g W V · V g 1
Equation (10) can be transformed into a classical total variation equation for solving. We assume that Z = f g W I · I g W V · V g and Equation (10) can be transformed, as follows:
E g = Z + W I · I g + W V · V g α · I g ( 1 α ) · V g 2 2 + Z 1
Discretizing Equation (11), we obtain the final formula [20]:
E g = Ω ( Z + W I · I g + W V · V g α · I g ( 1 α ) · V g ) 2 + Ω | Z |
where Ω represents the entire image field. Calculating the value of Z, the optimal high-frequency transform coefficients are obtained by transformation.

2.3. Strategy for Low-Frequency Coefficients

In the low-frequency coefficient, the main energy of the image is concentrated, and the subject of the image is in the low-frequency coefficient. Figure 2 shows the low frequency coefficients of multiple images, and different salient targets are embodied in different sensor images.
Figure 2 shows the two groups of low-frequency coefficients of the image, and the salient objects in Figure 2a,b are the thermal objects. As indicated by the red box in Figure 2a, the salient object of the main target in Figure 2b is the roof. The significance of the target in the infrared image is more obvious, the significance of the latter set of images includes heat source objects and light reflection text, and, in Figure 2c,d is the performance of the green box content. Thus, the key to the fusion of low-frequency coefficients is to optimize the performance of the significant targets.
To improve the salience of the fused image, we constrain the low-frequency coefficients after fusing. Firstly, the salient extraction of low-frequency coefficients is analyzed below.
The extraction of salient targets has been extensively studied. In 1998, Itti first proposed a salient detection method that was based on human vision [21], identified salient areas based on a variety of a priori knowledge of the visual mechanism, but the results were not accurate enough. Additionally, many scholars have proposed new detection methods. In 2007, Tie [22] proposed the method of using a random field model based on Itti’s model, where the global information, local information, and scale information were taken into account. In 2011, Xie [23] proposed a salient detection method based on the Bayesian model, the result is suitable for simple backgrounds, but there is noise with the complex background. In 2012, Stas [24] determined the salient areas based on neighborhood relationships, obtaining better results with the context information. In 2014, Yang [25] used the propagation map to obtain salient graphs, and was able to extract the main target. In 2016, Sun [26] used multi-regional mergers to obtain salient areas, thus improving the accuracy of the test. These algorithms have achieved good results in detecting, but these algorithms are computationally complex. In our method, we use the gray image as the experimental images, so we need to concisely calculate the salience of the low-frequency coefficients. A salient area is determined by using the prior image contrast:
P X i = e | X ( i ) | / j ω | X ( j ) |
In Equation (13), P X i represents the value of pixel i in the image X, X(i) represents the gray value of the pixel i, and ω represents the adjacent area of pixel i. As Equation (13) shows, the value of salience is higher in the pixels whose contrast is higher. However, we can obtain the single value of the pixel in the image; for the target, we need to detect the salient area:
S X i = P X m L d i s ( X ( i ) , X ( m ) ) + L v a l ( X ( i ) , X ( m ) )
P X m = max ( P X ω , i ω )
where Ldis and Lval represent the geometric distance and gray value distance between the two pixels. S X i represents the salient value of pixel i in image X. As Equation (15) shows, P X m represents the extreme values in area ω. The closer to the extremes of the salient point, the value of salience is higher. The salience of the low-frequency portion can be extracted by Equations (13)–(15).
We fuse the low-frequency coefficient with the salient constraints, as follows:
min S f S c 1
where Sf and Sc represent the salient values of the fused low-frequency coefficient and the constraint value, respectively, and Sc is calculated by the original images:
S c = H ( S I , S V ) = ( S I ) 2 + ( S V ) 2
H represents the synthesis of the two-source image’s salient maps; we need to select the suitable area to constrain the low-frequency coefficients. S(I) represents the salient area in the infrared image and the S(B) represents the salient area in the visible image. As in Equation (17), the difference between salient and non-salient areas were more pronounced, so the salient area can be emphasized.
We calculate the fused low frequency with the constraints, as follows:
E d = f d α · I d β · V d 2 2 + S f d S c 1
Discretizing this equation as:
E d = Ω ( f d α · I d β · V d ) 2 + Ω | S f d S c |

2.4. Nash Equilibrium to Update the Coefficients

After calculating the low and high coefficients, the obtained coefficients need to inverse the NSST. Before this step, we should confirm that the coefficients of the different frequencies are the best, so we use the Nash equilibrium for this purpose.
The idea of the Nash equilibrium can be shown by the following example:
The K represents the final set of the strategies:
K = { E d * , E g * }
We need to ensure that the final choice E d { E d * } is the best option for the every option of { E g * } and the final choice E g { E g * } is the best option for every option of { E d * } . We set the two parameters to update the different frequencies’ coefficients. Using σ d to update the low-frequency coefficients, and σ g to update the high-frequency coefficients, when the values of Ed and Eg are smaller than the parameters, we can ensure that the result of the fused image will be better.
Thus, we obtain the fused low-frequency coefficient, and we can obtain the fused image by the inverse NSST. The overall Algorithm 1 is described below:
Algorithm 1. The proposed method
Input: infrared image I, visible image V
Output: fused image f
Step 1.Initialize parameters, α = 0.5 , σ d = 0.04 , σ g = 0.01
Step 2.Decompose the I and V by NSST, get the Id , {Ig} and Vd, {Vg}
Step 3.Compute the Eg and Ed by Equation (12) and (18)
Where Eg < σg, and Ed < σd, compute the fd and fg
Else back to the step 3, update the α to minimize the Eg and Ed
Step 4.Use the fd and fg to inverse NSST to obtain the output.
Figure 3 shows the procedure of proposed method in this paper. After the NSST transformation of the visible and infrared images, the low and high frequency coefficients were obtained. Fused the coefficients based on the Nash equilibrium principle, and the final fused image was obtained with INSST transformation.

3. Results and Analysis

In order to verify the effectiveness of the proposed algorithm, we experiment the infrared and visible images in MATLAB 2016a (Core i3, clocked at 3.5 GHz, memory for 8 GB). The levels of NSST is four, the directions of the first level is four, and the next level is double than the last level. There are three groups of images, which include the salient target and rich details. When compared with other algorithms, such as contrast pyramid (CP) [5], PCA [6], Guided Filter fusion (GFF) [27], saliency fusion (SAL) [7], gradient transfer and total variation (GTF) [17], non-subsampled Contourlet transform (NSCT) [11], and the NSST [15]. The validity of the algorithms is verified by subjective and objective evaluation.
In Figure 4, four sets of experimental images are registered as infrared and visible images; the first group has a more obvious salience of the target, the second set of images has rich details, and the third set of images has a salient target and a blurred background; the fourth set of images has rich details and a blurred background.

3.1. Performance Evaluation

As the Figure 5 shows, analyze the red box area in the images, the result of PCA algorithm is poor, and the target is not obvious. In the Figure 5c,d,f, the target is brighter, but to enhance the target area brightness, a greater impact on the environmental area is produced at the same time. There is more serious blur in the Figure 5c, and the railing in the background is too smooth to detect in Figure 5d,f. In the images of Figure 5a,e,g, the target brightness is similar to the proposed algorithm, but the edge in Figure 5c is blurry, and the information of the background is missing. In the proposed algorithm, we reflect the target and retain the background information as much as possible. The green box area mainly reflects the details of the trees, and we can find that there is a shadow phenomenon in the result of PCA, GFF, and SAL. The results of CP, GTF, and NSCT are fuzzier. The details of the NSST has been promoted, but when compared with the proposed algorithm, our result is better in terms of the details.
There are more details in Figure 6, comparing the proposed algorithm with other algorithms. The red box is the text message in the roof, low brightness around the text, and their performance is not obvious in the results of CP, PCA, GTF, and the NSST. When compared with results of GFF, SAL, and NSCT, there are no artificial artifacts in the proposed method. The proposed algorithm preserves the detailed information on the basis of maximizing the brightness of the text background, and it is more suitable for the human eye system. The green box area in the figure is a more pronounced salient target, the contrast in the CP and the NSST is insufficient, and the information of the target can be preserved in the results of PCA, SAL, and NSCT, but the seat information on the left side of the target is basically lost. The result of the GFF algorithm has black block noise near the target. The proposed algorithm is better with respect to details and salient constraints, and it is superior.
Figure 7 shows the results of the third set of experimental images, processed by eight algorithms, and the overall comparison shows that proposed algorithm has the best effect in terms of detail reservation. The green box area in the figure shows some of the details of the trees, there is more retention of infrared image information in the results of CP, PCA, GTF, and the NSST, but also a lack of information is preserved in the visible images. The proposed algorithm is constrained by the gradient for the high-frequency part, and the details are more reserved in the result. There are also some algorithms that preserve the details, such as GFF, SAL, and NSCT, but when compared with the proposed algorithm, they are not superior enough. The red box section is the target area, the target is vague, and the edges are affected by the artificial artifacts in the GTF’s result. Additionally, there is a loss of background information around the target in the other algorithms. The proposed algorithm can reflect the target, and it has rich information of the background at the same time. It is more conducive to further processing.
Figure 8 shows that the proposed algorithm has obvious advantages in detail handling when compared with existing algorithms. In the proposed algorithm, the high frequency coefficients are constrained by the gradient constraint, and the details are more reserved in the result. Although there are also some other algorithms that can preserve the details, such as GFF, SAL, and PCA, they are not superior enough when compared with the proposed algorithm. Observing the red box tree area, the tree target is vague and the edges are affected by the artificial artifacts in the results of NSST, GTF, and CP algorithms. The target is reflected more definitely and more details are reserved with the proposed algorithms.

3.2. Further Analysis

After visual comparison and the analysis of the experimental results, it is obvious that the proposed algorithm has advantages in salient target enhancement and detail reservation. The advantages and disadvantages of objective indicators can be more effective in confirming the effectiveness of the proposed algorithm. Thus, eight algorithms were evaluated by using the commonly used fusion image quality evaluation index; they are the quality of image (Q) [28], quality of edge (Qe) [29], information entropy (H), average gradient (AVG), and overall cross entropy (OCE). The results are as follows:
Table 1, Table 2, Table 3 and Table 4 list the four groups of experimental results of the quality evaluation indicators.
Through the objective data of Table 1, Table 2, Table 3 and Table 4, the numerical values of the proposed algorithm are advantageous with respect to the image quality, edge quality, and the average gradient. When comparing the information entropy with these algorithms, the proposed algorithm in the first two groups of experiments were lower than the PCA algorithm but, compared with other algorithms, the value is higher. Analyzing the PCA algorithm, the reason for the higher value is that the targets are dark in the first two experimental images, so the contrasts are poor, and so the values are higher. When comparing the evaluation of the overall cross-entropy index, the results of the proposed method are better than the other methods, and it is shown that the information obtained from the two source images is the largest. Based on the experimental results, the proposed algorithm is at the optimal level.

4. Conclusions

The traditional fusion algorithms are summarized in this paper, and a new fusion algorithm is proposed for the problem of large artificial parameters and large errors. The high- and low-frequency coefficients’ characteristics after the NSST are analyzed, fusing the high-frequency coefficients based on the gradient constraints, and fusing the low-frequency coefficients based on the salient constraint. Before the inverse the NSST, the Nash equilibrium is used to update the coefficient, so that the fused image can preserve the details and enhance the target at the same time. After experiments and objective analysis, the validity of the algorithm is verified.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (61372167 and 61701524). This work was also supproted by the Scientific Research Program Funded by Shaanxi Provincial Education Department (Program No. 17JK0300).

Author Contributions

Yan Huang conceived the proposed algorithm and wrote the paper; Yan Huang and Dongpeng Wu designed and performed the experiments; Duyan Bi revisited the paper and provided technical guidance.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Libao, Z.; Jue, Z. A new adaptive fusion method based on saliency analysis for remote sensing images. Chin. J. Lasers 2015, 42, 0114001. [Google Scholar] [CrossRef]
  2. Wang, L.; Li, B.; Tian, L.-F. Multi-model medical image fusion using the inter-scale and intra-scale dependencies between image shift-invariant shearlet coefficients. Inf. Fusion 2014, 19, 20–28. [Google Scholar] [CrossRef]
  3. Wenshan, D.; Duyan, B.; Linyuan, H.; Dongpeng, W. Fusion of Infrared and Visible Image Based on Shearlet Transform and Neighborhood Structure Features. Acta Opt. Sinica 2017, 37, 1010002. [Google Scholar] [CrossRef]
  4. Musheng, C.; Zhishan, C. Study on fusion of visual and infrared images based on NSCT. Laser Optoelectron. Prog. 2015, 52, 061002. [Google Scholar] [CrossRef]
  5. Jin, H.Y.; Wang, Y. A fusion method for visible and infrared images based on contrast pyramid with teaching learning based optimization. Infrared Phys. Technol. 2014, 64, 134–142. [Google Scholar] [CrossRef]
  6. Shahdoosti, H.D.; Ghassemian, H. Combining the spectral PCA and spatial PCA fusion methods by an optimal filter. Inf. Fusion 2016, 27, 150–160. [Google Scholar] [CrossRef]
  7. Durga, P.B.; Ravindra, D. Two-scale image fusion of visible and infrared images using saliency detection. Infrared Phys. Technol. 2016, 76, 52–64. [Google Scholar]
  8. Rui, C.; Ke, Z.; Li, J. An image fusion algorithm using wavelet transform. Acta Electron. Sin. 2004, 32, 750–752. [Google Scholar]
  9. Do, M.; Vetterli, M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 14, 2091–2106. [Google Scholar] [CrossRef]
  10. Wang, X.H.; Wei, T.T.; Zhou, Z.G. Remote sensing image fusion method based on contourlet coefficients’ correlativity of directional region. J. Remote Sens. 2010, 14, 905–916. [Google Scholar]
  11. Wang, N.Y. Multimodal medical image fusion framework based on simplified PCNN in Nonsubsampled contourlet transform domain. J. Multimed. 2013, 8, 270–276. [Google Scholar] [CrossRef]
  12. Guo, K.H.; Labate, D.; Lim, W.; Guido, W.; Edward, W. Wavelets with composite dilations and their MRA properties. Appl. Comput. Harmon. Anal. 2006, 20, 202–236. [Google Scholar] [CrossRef]
  13. Glenn, E.; Labate, D.; Lim, W. Optimally sparse image representations using Shearlets. In Proceedings of the Fortieth Asilomar Conference on the Signals, Systems and Computers (ACSSC’06), Pacific Grove, CA, USA, 29 October–1 November 2006. [Google Scholar]
  14. Glenn, E.; Labate, D.; Lim, W. Sparse directional image representations using the discrete shearlet transform. Appl. Comput. Harmon. Anal. 2008, 25, 25–46. [Google Scholar]
  15. Luo, X.Q.; Zhang, C.Z.; Wu, X. A novel algorithm of remote sensing image fusion based on shift-invariant shearlet transform and regional selection. Int. J. Electron Commun. 2016, 70, 186–197. [Google Scholar] [CrossRef]
  16. Wan, W.; Yang, Y.; Lee, H.J. Practical remote sensing image fusion method based on guided filter and improved SML in the NSST domain. In Signal Image and Video Processing; Springer: London, UK, 2018; pp. 1–8. [Google Scholar]
  17. Ma, J.Y.; Chen, C.; Li, C.; Huang, J. Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fusion 2016, 31, 100–109. [Google Scholar] [CrossRef]
  18. Lei, W. Study for the Key Algorithms in Multi-Modal Medical Image Registration and Fusion; South China University of Technology: Guang Zhou, China, 2013. [Google Scholar]
  19. Ma, Y.; Chen, J.; Chen, C.; Fan, F.; Ma, J.Y. Infrared and visible image fusion using total variation model. Neurocomputing 2016, 202, 12–19. [Google Scholar] [CrossRef]
  20. Chan, T.F.; Osher, S.; Shen, J.H. The digital TV filter and nonlinear denoising. IEEE Trans. Image Process. 2001, 10, 231–241. [Google Scholar] [CrossRef] [PubMed]
  21. Itti, L.; Koch, C.; Niebur, E. A Model of Saliency-Based Visual Attention for Rapid Scene Analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef]
  22. Liu, T.; Sun, J.; Zheng, N.N. Learning to detect a salient object. In Proceedings of the IEEE Conference on Computer Vision & Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; Volume 33, pp. 1–8. [Google Scholar]
  23. Xie, Y.L.; Lu, H.C. Visual saliency detection based on Bayesian model. In Proceedings of the IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; Volume 263, pp. 645–648. [Google Scholar]
  24. Goferman, S.; Manor, L.; Tal, A. Context-Aware Saliency Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1915–1926. [Google Scholar] [CrossRef] [PubMed]
  25. Yang, Z.; Li, D.H.; Yu, J.G. Saliency detection based on diffusion maps. Optik 2014, 125, 5202–5206. [Google Scholar] [CrossRef]
  26. Feng, S.; Kai-yang, Q.; Wei, S.; Hong, G. Image saliency detection based on region merging. J. Comput. Aided Des. Comput. Graph. 2016, 28, 1679–1687. [Google Scholar]
  27. Li, S.T.; Kang, X.D.; Hu, J.W. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar] [PubMed]
  28. Wang, Z.; Bovik, A. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  29. Piella, G.; Heijmans, H. A new quality metric for image fusion. In Proceedings of the Tenth International Conference on Image Processing, Barcelona, Spain, 14–17 September 2003. [Google Scholar]
Figure 1. The illustration of two type of norm.
Figure 1. The illustration of two type of norm.
Sensors 18 01169 g001
Figure 2. The salient target in the low level coefficient. (a) The low-frequency band of the infrared image for the first group; (b) the low-frequency band of the visible image for the first group; (c) the low-frequency band of the infrared image for the second group; and (d) the low-frequency band of the visible image for the second group.
Figure 2. The salient target in the low level coefficient. (a) The low-frequency band of the infrared image for the first group; (b) the low-frequency band of the visible image for the first group; (c) the low-frequency band of the infrared image for the second group; and (d) the low-frequency band of the visible image for the second group.
Sensors 18 01169 g002
Figure 3. The illustration of the proposed method.
Figure 3. The illustration of the proposed method.
Sensors 18 01169 g003
Figure 4. Images for experiment: (a,b) the first set of experimental images; (c,d) the second set of experimental images; (e,f) the third set of experimental images; and (g,h) the fourth set of experimental images.
Figure 4. Images for experiment: (a,b) the first set of experimental images; (c,d) the second set of experimental images; (e,f) the third set of experimental images; and (g,h) the fourth set of experimental images.
Sensors 18 01169 g004
Figure 5. The results for first set of images. (a) Contrast pyramid (CP); (b) principal component analysis (PCA); (c) Guided Filter fusion (GFF); (d) saliency fusion (SAL); (e) gradient transfer and total variation (GTF); (f) non-subsampled Contourlet transform NSCT; (g) non-subsampled shearlet transform (NSST); and (h) proposed.
Figure 5. The results for first set of images. (a) Contrast pyramid (CP); (b) principal component analysis (PCA); (c) Guided Filter fusion (GFF); (d) saliency fusion (SAL); (e) gradient transfer and total variation (GTF); (f) non-subsampled Contourlet transform NSCT; (g) non-subsampled shearlet transform (NSST); and (h) proposed.
Sensors 18 01169 g005
Figure 6. The results for second set of images. (a) CP; (b) PCA; (c) GFF; (d) SAL; (e) GTF; (f) NSCT; (g) NSST; and (h) proposed.
Figure 6. The results for second set of images. (a) CP; (b) PCA; (c) GFF; (d) SAL; (e) GTF; (f) NSCT; (g) NSST; and (h) proposed.
Sensors 18 01169 g006
Figure 7. The results for third set of images. (a) CP; (b) PCA; (c) GFF; (d) SAL; (e) GTF; (f) NSCT; (g) NSST; and (h) proposed.
Figure 7. The results for third set of images. (a) CP; (b) PCA; (c) GFF; (d) SAL; (e) GTF; (f) NSCT; (g) NSST; and (h) proposed.
Sensors 18 01169 g007
Figure 8. The results for fourth set of images. (a) CP; (b) PCA; (c) GFF; (d) SAL; (e) GTF; (f) NSCT; (g) NSST; and (h) proposed.
Figure 8. The results for fourth set of images. (a) CP; (b) PCA; (c) GFF; (d) SAL; (e) GTF; (f) NSCT; (g) NSST; and (h) proposed.
Sensors 18 01169 g008
Table 1. Comparison of evaluation results of first experiment.
Table 1. Comparison of evaluation results of first experiment.
AlgorithmQQeHAVGOCE
CP6.040155.78006.75645.37596.7892
PCA5.749352.42117.46644.94506.2556
GFF5.101745.37096.78294.29216.1763
SAL6.043556.32537.18725.47976.4765
GTF4.117135.67406.67783.41046.8306
NSCT6.018650.64606.97404.82886.6516
NSST3.427929.11926.27502.78375.9401
Proposed6.197759.98856.60245.91976.9581
Table 2. Comparison of evaluation results of second experiment.
Table 2. Comparison of evaluation results of second experiment.
AlgorithmQQeHAVGOCE
CP4.669943.58126.17634.43446.8412
PCA4.111435.83046.98923.32267.1751
GFF4.746241.77746.83653.90037.1223
SAL4.905541.02446.83063.88677.3618
GTF2.875225.92166.62042.44837.4233
NSCT4.535338.15206.65163.59187.4169
NSST2.564321..41945.94012.02807.3417
Proposed5.067545.56686.84494.52257.4521
Table 3. Comparison of evaluation results of third experiment.
Table 3. Comparison of evaluation results of third experiment.
AlgorithmQQeHAVGOCE
CP6.590036.43376.23114.25656.6155
PCA2.867717.88035.92581.96026.4579
GFF4.949228.16466.27643.24476.5692
SAL5.327033.36926.27693.65196.7221
GTF2.924121.20036.02052.17276.0142
NSCT4.941729.59206.16133.29306.8353
NSST3.118518.47705.92272.07066.9215
Proposed5.724640.33326.64744.30087.2371
Table 4. Comparison of evaluation results of fourth experiment.
Table 4. Comparison of evaluation results of fourth experiment.
AlgorithmQQeHAVGOCE
CP2.431262.45406.37644.55403.6240
PCA3.647357.65535.94833.93692.7300
GFF4.577253.5435.49223.57702.0700
SAL4.234642.47314.52901.33301.0230
GTF1.854155.58726.51583.51092.3092
NSCT3.929859.83576.12753.55523.0523
NSST3.118552.26275.44913.88731.8936
Proposed5.903265.41936.43274.56923.9829

Share and Cite

MDPI and ACS Style

Huang, Y.; Bi, D.; Wu, D. Infrared and Visible Image Fusion Based on Different Constraints in the Non-Subsampled Shearlet Transform Domain. Sensors 2018, 18, 1169. https://doi.org/10.3390/s18041169

AMA Style

Huang Y, Bi D, Wu D. Infrared and Visible Image Fusion Based on Different Constraints in the Non-Subsampled Shearlet Transform Domain. Sensors. 2018; 18(4):1169. https://doi.org/10.3390/s18041169

Chicago/Turabian Style

Huang, Yan, Duyan Bi, and Dongpeng Wu. 2018. "Infrared and Visible Image Fusion Based on Different Constraints in the Non-Subsampled Shearlet Transform Domain" Sensors 18, no. 4: 1169. https://doi.org/10.3390/s18041169

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop