You are currently viewing a new version of our website. To view the old version click .
Entropy
  • Article
  • Open Access

28 June 2021

The Multi-Focus-Image-Fusion Method Based on Convolutional Neural Network and Sparse Representation

,
,
and
School of Mathematics and Statistics, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.

Abstract

Multi-focus-image-fusion is a crucial embranchment of image processing. Many methods have been developed from different perspectives to solve this problem. Among them, the sparse representation (SR)-based and convolutional neural network (CNN)-based fusion methods have been widely used. Fusing the source image patches, the SR-based model is essentially a local method with a nonlinear fusion rule. On the other hand, the direct mapping between the source images follows the decision map which is learned via CNN. The fusion is a global one with a linear fusion rule. Combining the advantages of the above two methods, a novel fusion method that applies CNN to assist SR is proposed for the purpose of gaining a fused image with more precise and abundant information. In the proposed method, source image patches were fused based on SR and the new weight obtained by CNN. Experimental results demonstrate that the proposed method clearly outperforms existing state-of-the-art methods in addition to SR and CNN in terms of both visual perception and objective evaluation metrics, and the computational complexity is greatly reduced. Experimental results demonstrate that the proposed method not only clearly outperforms the SR and CNN methods in terms of visual perception and objective evaluation indicators, but is also significantly better than other state-of-the-art methods since our computational complexity is greatly reduced.

1. Introduction

In the image processing field, multi-focus-image-fusion is a significant branch [,,]. Multi-focus-image-fusion is the process of combining two or more images with different focal points of the same scene into a composite image with all-focus, which is of service to humans and machine perception [,]. The multi-focus-image-fusion holds true for multifarious applications such as remote sensing and computer vision [].
In the past decade, sparse representation (SR)-based methods have been extensively applied to multi-focus-image-fusion []. SR has been proven as an extraordinarily powerful signal modeling method and has a good reputation in both theoretical research and practical application []. Yang and Li first applied SR in the field of image fusion []. After that, a large number of fusion methods based on SR emerged [,]. Liu and Wang proposed a adaptive sparse representation (ASR) model for simultaneous image fusion and denoising []. In the ASR model, a set of compact sub-dictionaries were learned from plentiful image patches which were pre-classified into several categories according to their gradient information. One of the sub-dictionaries can be adaptively selected by a given set of source image patches. In [], a convolutional sparse representation (CSR)-based image fusion framework was presented, in which each source image is decomposed into a base layer and a detailed layer. The SR-based method was, in nature, the local one with the nonlinear fusion regulation, which was used to merge the source image patches.
In contrast to the relatively complex fusion method based on SR, in [], Liu et al. proposed a multi-focus-image-fusion method based on the convolutional neural network (CNN). In this method, the decision map was received in line with the CNN model, which represented the accurate measurement of activity level. At last, a pixel-by-pixel weighted average strategy was employed to gain the fused image. Compared to SR methods, CNN was the global method with a linear fusion rule.
Uniting the merits of the two aforementioned methods, a novel fusion method was presented and more ample fused images were gained. In this method, source image patches are fused on the strength of the new weights obtained by CNN and SR.
The rest of this paper is organized as follows. In Section 2, some related work is discussed. In Section 3, the basic idea of the proposed fusion method is presented in detail. Experimental results and discussions are demonstrated in Section 4. Finally, Section 5 concludes the paper.

3. Proposed Fusion Algorithm

The proposed method based on CNN and SR includes three principal parts: (1) CNN-based weight map generation; (2) fusion of image patches based on the new SR; and (3) Fast image fusion based on patches. The following subsections describe the aforementioned steps at length. The algorithm flow is shown in Figure 2.
Figure 2. Flowchart of the proposed fusion algorithm: (a): Generation of weight map; (b): The process of the algorithm.

3.1. CNN-Based Weight Map Generation

We suppose that I 1 , I 2 are the two source images and the size is X × Y . I 1 is taken as a reference. These two images are fed to a pre-trained CNN to acquire the score map, the size of the which is [ ( X 16 + 2 ) / 2 ] × [ ( Y 16 + 2 ) / 2 ] (where [ · ] denotes the ceil operation). Every value of the score map that represents the focus level of a set of 16 × 16 patches of I 1 is between 0 and 1. The closer the pixel value is to 1, the more focused the image patches from I 1 are. After that, each pixel of the score map is extended to a 16 × 16 matrix that has the same element, and the focus map with the size X × Y is obtained through the pixel-wise overlap-averaging tactics. With this, the initial segment and small region removal were performed on the focus map to obtain the decision map. Later, the slider process is executed on the decision map. The patch size is 8 × 8 and the step size is 1. Each patch is averaged to obtain the pixel value of the corresponding position of the weight map E, i.e., the weight of the patches. The size of E is ( X 8 + 1 ) × ( Y 8 + 1 ) . The flow chart for generating the weight map E is shown in Figure 2a.

3.2. Fusion of Image Patches Based on the New SR

Given the image patches P q , q = 1 , 2 , which are represented as vectors V q , the normalization is performed via v q = V q V q ¯ · 1 , where V q ¯ is the mean value of V q .
Then, the normalized vectors v q are represented in the dictionary by the following formula:
v q = D α q , min α q 0
where D is the pre-trained dictionary via the K-SVD algorithm, as shown in Figure 1a. The α q that are earned by the OMP algorithm are the SR vectors of P q .
The fusion coefficients and the fusion means are, respectively, obtained as follows:
α F = α 1 , i f M 1 > M 2 α 2 , o t h e r w i s e
V F ¯ = V 1 ¯ , i f M 1 > M 2 V 2 ¯ , o t h e r w i s e
where M 1 = ω · α 1 1 , M 2 = ( 1 ω ) · α 2 1 and ω is the weight of P 1 obtained from E. The weight l 1 -norm M q reflects the actual activity level of the image patches, which can avoid the wrong selection of patch with the small value of the norm.
The fused result of V F is calculated by
V F = D α F + V F ¯
V F is reshaped into the 8 × 8 patch P F and P F is the fused image patch. In the end, each pixel’s value of the fused image I F is obtained by its average over its superposition.

3.3. Fast Image Fusion Based on Patches

By sliding window technology, I 1 , I 2 is divided into n × n patches I 1 t , I 2 t , t = 1 , , T . The number of patches from each image is T, T = ( X n + 1 ) ( Y n + 1 ) . In fact, the procedure proposed in Section 3.2 is not needed for each patch. In the very beginning, the weight map E is expressed in vector form E t that is used to choose the patch that does not need sparse coding.
When E t = 1 , i.e., the image patch of I 1 is in-focused, as can be shown, for example, when these in-focused patches are at the position of the red diamonds in Figure 3a, then the fusion result is I F t = I 1 t .
Figure 3. (a): The weight map; and (b): The number comparison of positions of image patches.
If E t = 0 , i.e., the image patch of I 1 is out-focused, as shown by the green squares in Figure 3a, the fusion result I F t is I 2 t .
In the case of 0 < E t < 1 , the image patch is located somewhere in between the in-focused and out-focused regions. For instance, these patches are the blue blocks in Figure 3a. Only in this case is the new SR fusion method adopted, where ω = E t and the fusion patch is gained.
It can be known that the above classification can greatly reduce the computational complexity.

4. Experiments

This section successively gives the experimental settings that include the source images to be processed, image fusion quality metrics, parameters setting, computational complexity analysis, compared methods and image fusion results to be visually and quantitatively analyzed.

4.1. Source Images

In order to illustrate the experimental results, different types of source images are applied. There are 12 pairs of source images, including five pairs of multi-focus grayscale images in Figure 4 and seven pairs of multi-focus color images in Figure 5. These images are obtained from the Lytro Multi-Focus Dataset that contains 20 pairs of color multi-focus images and four series of multi-focus images with three sources, and the Multi-Focus-Image-Fusion-Dataset that includes 150 different images used in multi-focus-image-fusion algorithms [,].
Figure 4. The multi-focus grayscale source images.
Figure 5. The multi-focus color source images.

4.2. Evaluation Metrics

For the sake of verifying the performance of image fusion methods, subjective and objective evaluation metrics are usually applied. Between them, subjective evaluation means that people explain the relative merits of the methods through the visual effects of the fusion results, and it is affected by uncertain factors such as the observer’s own conditions, professional knowledge, observation angle, application occasions and objective environment []. The subjective evaluation is thus less reliable and objective. The objective evaluation is required to assist subjective evaluation. Therefore, the objective evaluation is especially important. The objective evaluation method conducts the quantitative analysis of fused images through certain mathematical models, which can overcome the limitations of subjective evaluation and the evaluation results are stable and reliable []. Generally speaking, it is difficult to objectively evaluate the merit and fallacies of the fusion method by relying on only one evaluation index. Therefore, many researchers ecumenically adopt comprehensive evaluation with multiple evaluation indexes.
In this paper, five metrics were employed to evaluate the fusion quality. The larger the values of the metrics, the higher the fusion performance. The five metrics are introduced as follows:
1.
Mutual information mainly reflects how much information the fused image contains from the source images []. The greater the mutual information is, the more information of the source images the fused image contains, and the better the fusion effect is. Mutual information is defined as follows:
M I = M I ( I 1 , I F ) + M I ( I 2 , I F )
M I ( I q , I F ) = x , y h I q , I F ( x , y ) log 2 h I q , I F ( x , y ) h I q ( x ) h I F ( y )
Here, h I q ( x ) , h I F ( y ) are, respectively, the edge histogram of I q , I F . h I q , I F ( x , y ) are normalized joint histograms of I F and source images I q , respectively.
2.
The Chen–Blum metric Q C B is a human perception-inspired fusion metric. Q C B is calculated by the following steps.
At the very start, the masked contrast map for the input image I q ( x , y ) can be computed in:
C I q = l ( C I q ) n k ( C A ) p + m
where C is Peli’s contrast, k , l , m , n are real scalar parameters, and more details on the parameter settings can be found in [].
The information preservation value Q I q , I F ( x , y ) and the saliency map μ I q ( x , y ) can be calculated by the two following expressions:
Q I q , I F ( x , y ) = C I q ( x , y ) / C I F ( x , y ) , i f C I q ( x , y ) < C I F ( x , y ) C I F ( x , y ) / C I q ( x , y ) , o t h e r w i s e
μ I q ( x , y ) = C I q ( x , y ) C I 1 ( x , y ) + C I 2 ( x , y )
Then, the value of the global quality map can be calculated:
Q G Q M = x , y μ I 1 ( x , y ) Q I 1 , I F ( x , y ) + μ I 2 ( x , y ) Q I 2 , I F ( x , y )
where Q C B is the average of Q G Q M .
3.
The fusion metric Q G based on the gradient is a popular fusion metric which computes the amount of gradient information of the source images injected into the fused image []. It is calculated by
Q G = x = 1 X y = 1 Y ( Q e I 1 , I F ( x , y ) Q o I 1 , I F ( x , y ) τ I 1 ( x , y ) + Q e I 2 , I F ( x , y ) Q o I 2 , I F ( x , y ) τ I 2 ( x , y ) ) x = 1 X y = 1 Y ( τ I 1 ( x , y ) + τ I 2 ( x , y ) )
where Q e I q , I F ( x , y ) and Q o I q , I F ( x , y ) are the edge strength and orientation reservation values, respectively. The weight factor τ I q ( x , y ) shows the significance of Q I q , I F ( x , y ) .
4.
The fusion metric based on phase congruency Q P measures the image-salient features of the source images, such as the edges and corners in the fused image []. The definition of Q P is:
Q P = ( R r ) θ ( R H ) υ ( R h ) σ
where r , H , h refer to phase congruency, maximum and minimum moments, respectively. The exponential parameters θ , υ , σ are all set to 1. More details about Q P can be seen in [].
5.
Q Y was proposed by Yang et al., which is a structural similarity-based method of fusion assessment []. The definition of Q Y is shown as follows:
Q Y = max S S I M ( I 1 , I F ω ) , S S I M ( I 1 , I F ω ) , i f S S I M ( I 1 , I 2 ω ) < 0.75 μ ( ω ) S S I M ( I 1 , I F ω ) + ( 1 μ ( ω ) ) S S I M ( I 2 , I F ω ) , o t h e r w i s e
The details of local weight μ ( ω ) and the structural similarity of images S S I M ( I 1 , I 2 ) can be found in [,].

4.3. Parameters Setting

In this section, our training parameters are set. For image processing applications based on SR, the size of image patches is 8 × 8 and the step length of sliding window technology is 1 pixel, which has been proven to be an appropriate setting []. The dictionary is obtained according to the K-SVD methodology, and the 68,000 image patches are randomly selected from the natural images. According to the paper [], NSCT is selected for the multi-focus-image-fusion-based MST and MST-SR methods. The implementation of the compared method in this article was based on the exposed code, and we set the parameters according to their original reports. All experiments were performed on MATLAB R2017a. The computer processor is Intel(R) Xeon(R) Silver 4110CPU.

4.4. The Compared Methods

The effectiveness of the proposed algorithm was evaluated against state-of-the-art research methods. The first one was the NSCT-based method that uses the weighted average for low-pass sub-bands and ’max-absolute’ for high-pass sub-bands. The second compared algorithm was based on SR []. The third method was ASR []. The fourth method was NSCT-SR-1 []. Each of the pre-registered source images was decomposed by NSCT 1 level decomposition, and the low-pass and high-pass coefficients were obtained. The low-pass coefficients were merged with an SR-based fusion method, while the high-pass coefficients were fused using the absolute values of coefficients for activity level measurement. The fifth approach was CSR []. The sixth compared algorithm was based on CNN [].

4.5. Computational Complexity Analysis

In order to verify that the proposed algorithm reduced the computational complexity, Figure 3b is given. As shown in Figure 3b, the ordinate is the number of positions, and the abscissas 1–5 indicate five pairs of grayscale source images. The red and blue rectangles indicate the corresponding positions of the patches fused by SR in the traditional SR and the algorithm proposed in this paper, respectively. It can be seen from the histogram that the height of the yellow rectangle is much higher than that of the purple. Therefore, the fusion algorithm proposed in this paper greatly reduces the number of patches that need to be fused by SR, thereby enormously reducing the computational complexity.
Referring to Table 1, it can be seen that the running time of CNN-SR is less than that of the SR and CNN. The bold font in Table 1 indicates that the shortened running time of CNN-SR is greater than one minute. To sum up, the method proposed in this paper improves the computational efficiency.
Table 1. Running time with the units of the five grayscale images under different fusion methods.

4.6. Validity of the Proposed Fusion Method

In this section, the comparative methods and the proposed method are applied to the commonly used multi-focus grayscale images of Figure 4. Figure 6a,b, Figure 7a,b, Figure 8a,b, Figure 9a,b and Figure 10a,b are source images from the same sensor, that are focused at different locations. The objects in other positions are out-focused and blurred. Figure 6c–i,Figure 7c–i, Figure 8c–i, Figure 9c–i and Figure 10c–i are the fusion results of different methods that are, respectively, MST(NSCT), SR, ASR, MST-SR(NSCT-SR-1), CSR, CNN and the fusion method CNN-SR proposed in this paper. The evaluation indexes are shown in Table 2, Table 3, Table 4, Table 5 and Table 6, and the values for the best fusion performance are bolded.
Figure 6. The ‘flowerpot’ source images and fusion results obtained by different fusion methods: (a) source image flowerpot A; (b) source image flowerpot B; (c) NSCT; (d) SR; (e) ASR; (f) NSCT-SR-1; (g) CSR; (h) CNN; and (i) CNN-SR.
Figure 7. The ‘aircraft’ source images and fusion results obtained by different fusion methods: (a) source image aircraft A; (b) source image aircraft B; (c) NSCT; (d) SR; (e) ASR; (f) NSCT-SR-1; (g) CSR; (h) CNN; and (i) CNN-SR.
Figure 8. The ‘leaf’ source images and fusion results obtained by different fusion methods: (a) source image leaf A; (b) source image leaf B; (c) NSCT; (d) SR; (e) ASR; (f) NSCT-SR-1; (g) CSR; (h) CNN; and (i) CNN-SR.
Figure 9. The ‘newspaper’ source images and fusion results obtained by different fusion methods: (a) source image newspaper A; (b) source image newspaper B; (c) NSCT; (d) SR; (e) ASR; (f) NSCT-SR-1; (g) CSR; (h) CNN; and (i) CNN-SR.
Figure 10. The ‘temple’ source images and fusion results obtained by different fusion methods: (a) source image temple A; (b) source image temple B; (c) NSCT; (d) SR; (e) ASR; (f) NSCT-SR-1; (g) CSR; (h) CNN; and (i) CNN-SR.
Table 2. Objective evaluations of image pair ‘flowerpot’ fusion experimentations.
Table 3. Objective evaluations of image pair ‘aircraft’ fusion experimentations.
Table 4. Objective evaluations of image pair ‘leaf’ fusion experimentations.
Table 5. Objective evaluations of image pair ‘newspaper’ fusion experimentations.
Table 6. Objective evaluations of image pair ‘temple’ fusion experimentations.
An example of ’flowerpot’ fusion is shown in Figure 6. The left-hand magnification of the clock is shown at the lower left corner of each image in Figure 6. From the magnified details, it can be seen that the fusion results obtained by MST, SR, ASR and MST-SR are uneven in varying degrees. For the remaining three fusion results, the human visual system struggles to tell the difference. Hence, objective evaluation is needed.
As shown in Table 2, among the five indicators, CNN-SR leads four, including M I , Q G , Q Y , Q P . It follows that our proposed method can extract the most information from the source images. Although the Q C B of our proposed method is a little small, our proposed fusion method best preserves the structure and detailed information of the source images, and improves the clarity of the fused image.
Figure 7 shows the fusion results of the ‘aircraft’ images. In megascopic details, in the top left corner, the artificial information increases in Figure 7c,e. The fusion result based on SR Figure 7d has a slight artificial edge; the details of Figure 7f are inconspicuous; the details of Figure 7g,h are missing. In the enlarged area of the bottom left corner, there are different degrees of unevenness in Figure 7c,d,f. The fusion result of the algorithm in this paper shown in Figure 7i retains the best and restores the information at the bottom left.
Table 3 exhibits the objective evaluation of Figure 7. The objective results confirm that our approach is the best among the seven methods. The performance shows that the CNN-SR can extract the edge and structure information of the source image well.
As can be seen from the magnified details of Figure 8, the edge artifacts exist in Figure 8c–f, and the fusion results have low contrast, resulting in the loss of some useful details. Figure 8g with artificial edges is derived by the CSR-based approach. Figure 8h,i effectively preserve the detail of the source images without producing specific visual artifacts and brightness distortion. By comparison, using our method can achieve better image appearance.
The fusion performance measured by the objective metrics is shown in Table 4. The fusion method proposed in this paper is superior to other methods in terms of evaluation criteria M I , Q C B , Q G , Q Y . The performance of these indexes shows that the fused image obtained by the CNN-SR method not only contains more detailed information, however, it is also more suitable for human visual perception. While the Q P is inferior to the CNN-based approach, our approach obtains comparable performance. Therefore, the proposed fusion method is superior to the SR-based method.
The fused results of ‘newspaper’ are shown in Figure 9. The fusion details are shown in the lower left corner of all the images in Figure 9. By comparing the image details fused by different methods, it can be determined that Figure 9c–g are relatively fuzzy, with poor contrast brightness. The fused images Figure 9h,i have better performance in information recovery and contrast and have better fusion performance.
The indicators of the proposed method and the contrastive methods are shown in Table 5. Table 5 shows the best performance of our proposed fusion method on M I , Q C B , Q G , Q Y . It can be inferred that the proposed method has better performance in visual fidelity, image clarity and structure information level. For Q P , the image integrated by CSR algorithm shows the best result. However, the image integrated by CSR extracts less information than the image fused by our proposed method. Therefore, the fusion method proposed in this paper is superior to the other contradistinctive methods.
The image pairs ‘temple’ and fusion results are shown in Figure 10. The details are shown in the lower left corner of Figure 10. For the details in Figure 10, the integrated images of MST, SR, ASR, MST-SR and CSR have different degrees of artifacts. The fused image of CNN and our proposed method show better performance in terms of detailed information than other integrated images.
The objective evaluation indexes are listed in Table 6. It can be clearly seen in Table 6 that our method obviously obtains all the largest quality indicators. A conclusion can be drawn from the experiment. Through visual comparison and objective evaluations, the proposed method shows emulatory fusion performance compared with the previous methods.
Therefore, these experimental consequences show that the proposed method fully extracts the information of the multi-focus source images. After CNN-SR fusion, the fused images with clear edges and no artificial artifacts preserve the detailed information well and have high contrast. Meanwhile, uneven fusion does not occur. Both the subjective and objective evaluation of CNN-SR are better than that of other algorithms.

4.7. Fusion of Multi-Focus Color Images

The proposed method can be extended to multi-focus color image fusion. In order to prove the effectiveness of CNN-SR in the color images, the color source images, as shown in Figure 5, are adopted. Figure 11 is the fusion results of the different methods. Table 7 provides the average scores of the seven pairs of input images under different fusion methods. The visual fusion results and the quantitative estimate in Table 7 show that the CNN-SR method can gain the best fusion results.
Figure 11. From left to right are the results of multi-focus color source images of MST, SR, ASR, MST-SR, CSR, CNN and CNN-SR.
Table 7. Quantitative assessments of Figure 11 and values for the seven pairs of input images in Figure 11 are averaged.

5. Conclusions

We proposed a multi-focus-image-fusion method based on CNN and SR. In the method, the weight map was acquired according to the CNN model, where each pixel of the weight map represents the focus level of each source image patch. If the pixel value of the weight map is 0 or 1, this means that the image patch is in-focused or out-focused, which can be directly obtained from the source images. When the pixel value is greater than 0 and less than 1, it indicates that the image patch is between clear and blurred. The new SR method was adopted. In the novel SR method, the image patches are represented by the dictionary to gain the sparse vectors, and the weight of the patch is multiplied by the l 1 -norm of its sparse vector to obtain its actual activity level. The fused sparse vectors were received by the max weight l 1 -norm. The fused image can be gained by aggregating all the reconstructed patches with the pixel-wise overlap-averaging tactics. The classified disposal of the image patches makes the proposed fusion method have great computational efficiency and it retains as much information of source images in the fused image as possible. The qualitative and quantitative comparisons show that the proposed method achieves better fusion performance in visual and objective evaluation.

Author Contributions

Conceptualization, B.W. and X.F.; methodology, B.W. and X.F.; software, B.W. and K.W.; validation, B.W. and B.G.; formal analysis, B.W.; writing—original draft preparation, B.W.; writing—review and editing, B.W. and X.F.; visualization, K.W. and B.G.; supervision, X.F.; funding acquisition, B.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China grant number 61772389 and 61472303.

Data Availability Statement

Publicly available datasets were analyzed in this study: Lytro Multi-Focus Dataset: Mansour Nejati. 2015. https://mansournejati.ece.iut.ac.ir/content/lytro-multi-focus-dataset, accessed on 7 December 2020; Multi-Focus-Image-Fusion-Dataset: 2018. https://github.com/sametaymaz/Multi-focus-Image-Fusion-Dataset, accessed on 7 December 2020.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Meher, B.; Agrawal, S.; Panda, R.; Abraham, A. A survey on region based image fusion methods. Inf. Fusion 2019, 48, 119–132. [Google Scholar] [CrossRef]
  2. Guo, L.L.; Woźniak, M. An image super-resolution reconstruction method with single frame character based on wavelet neural network in internet of things. Mob. Netw. Appl. 2021, 26, 390–403. [Google Scholar] [CrossRef]
  3. Woźniak, M.; Polap, D. Soft trees with neural components as image-processing technique for archeological excavations. Pers. Ubiquitous Comput. 2020, 24, 363–375. [Google Scholar] [CrossRef] [Green Version]
  4. Liu, Y.; Wang, L.; Cheng, J.; Li, C.; Chen, X. Multi-focus image fusion: A survey of the state of the art. Inf. Fusion 2020, 64, 71–91. [Google Scholar] [CrossRef]
  5. Farid, M.S.; Mahmood, A.; Al-Maadeed, S.A. Multi-focus image fusion using content adaptive blurring. Inf. Fusion 2019, 45, 96–112. [Google Scholar] [CrossRef]
  6. Panigrahy, C.; Seal, A.; Mahato, N.K. Fractal dimension based parameter adaptive dual channel PCNN for multi-focus image fusion. Opt. Lasers Eng. 2020, 133, 106–141. [Google Scholar] [CrossRef]
  7. Nejati, M.; Samavi, S.; Shirani, S. Multi-focus image fusion using dictionary-based sparse representation. Inf. Fusion 2015, 25, 72–84. [Google Scholar] [CrossRef]
  8. Wang, K.P.; Qi, G.Q.; Zhu, Z.Q.; Chai, Y. A novel geometric dictionary construction approach for sparse representation based image fusion. Entropy 2017, 19, 306. [Google Scholar] [CrossRef] [Green Version]
  9. Yang, B.; Li, S.T. Multifocus image fusion and restoration with sparse representation. IEEE Trans. Instrum. Meas. 2009, 59, 884–892. [Google Scholar] [CrossRef]
  10. Li, Y.Y.; Sun, Y.J.; Huang, X.H.; Qi, G.Q.; Zheng, M.Y.; Zhu, Z.Q. An image fusion method based on sparse representation and sum modified-Laplacian in NSCT domain. Entropy 2018, 20, 522. [Google Scholar] [CrossRef] [Green Version]
  11. Li, S.T.; Yin, H.T.; Fang, L.Y. Group-sparse representation with dictionary learning for medical image denoising and fusion. IEEE Trans. Biomed. Eng. 2012, 59, 3450–3459. [Google Scholar] [CrossRef] [PubMed]
  12. Liu, Y.; Wang, Z.F. Simultaneous image fusion and denoising with adaptive sparse representation. IET Image Process. 2014, 9, 347–357. [Google Scholar] [CrossRef] [Green Version]
  13. Liu, Y.; Chen, X.; Ward, R.K.; Wang, Z.J. Image Fusion with Convolutional Sparse Representation. IEEE Signal Process. Lett. 2016, 23, 1882–1886. [Google Scholar] [CrossRef]
  14. Liu, Y.; Chen, X.; Peng, H.; Wang, Z.F. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  15. Wei, Q.; Bioucas-Dias, J.; Dobigeon, N.; Tourneret, J.Y. Hyperspectral and multispectral image fusion based on a sparse representation. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3658–3668. [Google Scholar] [CrossRef] [Green Version]
  16. Chen, L.; Li, J.B.; Chen, C.L.P. Regional multifocus image fusion using sparse representation. Opt. Express 2013, 21, 5182–5197. [Google Scholar] [CrossRef]
  17. Tang, H.; Xiao, B.; Li, W.S.; Wang, G.Y. Pixel convolutional neural network for multi-focus image fusion. Inf. Sci. 2018, 433, 125–141. [Google Scholar] [CrossRef]
  18. Dian, R.W.; Li, S.T.; Fang, L.Y.; Wei, Q. Multispectral and hyperspectral image fusion with spatial-spectral sparse representation. Inf. Fusion 2019, 49, 262–270. [Google Scholar] [CrossRef]
  19. Yin, H.P.; Li, Y.X.; Chai, Y.; Liu, Z.D.; Zhu, Z.Q. A novel sparse-representation-based multi-focus image fusion approach. Neurocomputing 2016, 216, 216–229. [Google Scholar] [CrossRef]
  20. Xu, M.; Hu, D.L.; Luo, F.L.; Liu, F.L.; Wang, S.Y.; Wu, W.W. Limited angle X ray CT reconstruction using image gradient l0 norm with dictionary learning. IEEE Trans. Radiat. Plasma Med. Sci. 2020, 5, 78–87. [Google Scholar] [CrossRef]
  21. Zhang, J.; Zhao, C.; Zhao, D.B.; Gao, W. Image compressive sensing recovery using adaptively learned sparsifying basis via l0 minimization. Signal Process. 2014, 103, 114–126. [Google Scholar] [CrossRef] [Green Version]
  22. Cai, T.T.; Wang, L. Orthogonal matching pursuit for sparse signal recovery with noise. IEEE Trans. Inf. Theory 2011, 57, 4680–4688. [Google Scholar] [CrossRef]
  23. Wang, J.; Kwon, S.; Shim, B. Generalized orthogonal matching pursuit. IEEE Trans. Signal Process. 2012, 60, 6202–6216. [Google Scholar] [CrossRef] [Green Version]
  24. Jeon, M.J.; Jeong, Y.S. Compact and Accurate Scene Text Detector. Appl. Sci. 2020, 10, 2096. [Google Scholar] [CrossRef] [Green Version]
  25. Woźniak, M.; Silka, J.; Wieczorek, M. Deep neural network correlation learning mechanism for CT brain tumor detection. Neural Comput. Appl. 2021, 6, 1–16. [Google Scholar]
  26. Vu, T.; Nguyen, C.V.; Pham, T.X.; Luu, T.M.; Yoo, C.D. Fast and Efficient Image Quality Enhancement via Desubpixel Convolutional Neural Networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  27. Lytro Multi-Focus Dataset. Available online: https://mansournejati.ece.iut.ac.ir/content/lytro-multi-focus-dataset (accessed on 7 December 2020).
  28. Multi-Focus-Image-Fusion-Dataset. Available online: https://github.com/sametaymaz/Multi-focus-Image-Fusion-Dataset (accessed on 7 December 2020).
  29. Tsagaris, V. Objective evaluation of color image fusion methods. Opt. Eng. 2009, 46, 066201. [Google Scholar] [CrossRef]
  30. Petrović, V. Subjective tests for image fusion evaluation and objective metric validation. Inf. Fusion 2007, 8, 208–216. [Google Scholar]
  31. Zhu, Z.Q.; Yin, H.P.; Chai, Y.; Li, Y.X.; Qi, G.Q. A novel multi-modality image fusion method based on image decomposition and sparse representation. Inf. Sci. 2018, 432, 516–529. [Google Scholar] [CrossRef]
  32. Chen, Y.; Blum, R.S. A new automated quality assessment algorithm for image fusion. Image Vis. Comput. 2009, 27, 1421–1432. [Google Scholar] [CrossRef]
  33. Xydeas, C.S.; Petrovic, V.S. Objective image fusion performance measure. Electron. Lett. 2004, 36, 308–309. [Google Scholar] [CrossRef] [Green Version]
  34. Zhao, J.; Laganiere, R.; Liu, Z. Performance assessment of combinative pixel-level image fusion based on an absolute feature measurement. Int. J. Innov. Comput. 2007, 3, 1433–1447. [Google Scholar]
  35. Yang, C.; Zhang, J.Q.; Wang, X.R.; Liu, X. A novel similarity based quality metric for image fusion. Inf. Fusion 2008, 9, 156–160. [Google Scholar] [CrossRef]
  36. Liu, Z.; Blasch, E.; Xue, Z.; Zhao, J.; Laganiere, R.; Wu, W. Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: A comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 94–109. [Google Scholar] [CrossRef] [PubMed]
  37. Zong, J.J.; Qiu, T.S. Medical image fusion based on sparse representation of classified image patches. Biomed. Signal Process. Control 2017, 34, 195–205. [Google Scholar] [CrossRef]
  38. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2017, 24, 147–164. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.