Next Article in Journal
A Structure Identification Method for Urban Agglomeration Based on Nighttime Light Data and Railway Data
Previous Article in Journal
Quantifying the Reliability and Uncertainty of Satellite, Reanalysis, and Merged Precipitation Products in Hydrological Simulations over the Topographically Diverse Basin in Southwest China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Multispectral Image Fusion via Fast Matrix Truncated Singular Value Decomposition

1
State Key Laboratory of High-Performance Computing, College of Computer Science, National University of Defense Technology, Changsha 410073, China
2
Beijing Institute for Advanced Study, College of Advanced Interdisciplinary Studies, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(1), 207; https://doi.org/10.3390/rs15010207
Submission received: 17 November 2022 / Revised: 27 December 2022 / Accepted: 28 December 2022 / Published: 30 December 2022

Abstract

:
Recently, methods for obtaining a high spatial resolution hyperspectral image (HR-HSI) by fusing a low spatial resolution hyperspectral image (LR-HSI) and high spatial resolution multispectral image (HR-MSI) have become increasingly popular. However, most fusion methods require knowing the point spread function (PSF) or the spectral response function (SRF) in advance, which are uncertain and thus limit the practicability of these fusion methods. To solve this problem, we propose a fast fusion method based on the matrix truncated singular value decomposition (FTMSVD) without using the SRF, in which our first finding about the similarity between the HR-HSI and HR-MSI is utilized after matrix truncated singular value decomposition (TMSVD). We tested the FTMSVD method on two simulated data sets, Pavia University and CAVE, and a real data set wherein the remote sensing images are generated by two different spectral cameras, Sentinel 2 and Hyperion. The advantages of FTMSVD method are demonstrated by the experimental results for all data sets. Compared with the state-of-the-art non-blind methods, our proposed method can achieve more effective fusion results while reducing the fusing time to less than 1% of such methods; moreover, our proposed method can improve the PSNR value by up to 16 dB compared with the state-of-the-art blind methods.

1. Introduction

High spatial resolution hyperspectral images are of great significance in agriculture [1], military [2,3,4], image processing [5,6], and remote sensing [7,8,9,10] because of their ability to possess rich spectral and spatial information at the same time. However, due to limitations of the physical components, the LR-HSI can only be obtained with low spatial information but high spectral information, and the HR-MSI of the same scene has high spatial information but low spectral information. The most common economical way to obtain HR-HSI is, therefore, to fuse the LR-HSI and HR-MSI.
As we know, HR-MSI, LR-HSI, and HR-HSI are images of the same scene with different degrees of spatial and spectral information. When applying most fusion methods, HR-MSI is considered to contain most of the spatial information of HR-HSI, while LR-HSI contains most of the spectral information of HR-HSI. For example, Dian et al. [11] assumed that LR-HSI contains a large amount of spectral information of HR-HSI, and the spectral basis was obtained from LR-HSI by TMSVD. Long et al. [12] found a significant correlation between the singular values of HR-HSI and LR-HSI. For both methods, a TMSVD operation is carried out on LR-HSI, and the different factor matrices from LR-HSI are used as prior information in these proposed methods. We assume the HR-MSI should have properties similar to LR-HSI and HR-HSI.
As mentioned, these works [11,12] have shown that the first two TMSVD factor matrices of HR-HSI and LR-HSI have strong similarity. A feature of singular value decomposition (SVD) is that most of the image information can be saved by only keeping the first few terms. Inspired by this, we can reasonably assume that HR-HSI can be represented as the product of three TMSVD factor matrices. The last TMSVD factor matrix of HR-HSI contains a lot of spatial information, and it obviously comes from HR-MSI. The relationship between the second TMSVD factor matrix of LR-HSI and HR-HSI has been clearly experimentally demonstrated in [12], but the relationship between first TMSVD factor matrix of LR-HSI and HR-HSI was not clearly shown. We will verify the relationship of the first matrix as shown in the experimental results that follow.
As shown in Figure 1, we found a strong correlation between HR-HSI and HR-MSI. The values of the third TMSVD factor matrix from the HR-MSI and HR-HSI are similar or opposite in the same position. As we can see from Figure 2, the first factor matrix value from HR-HSI and LR-HSI by TMSVD are also approximately the same at the same position, which can explain why the spectral basis comes from the LR-HSI based on matrix similarity, where there was previously no tangible evidence to explain this. We estimate the three SVD factor matrices of HR-HSI from LR-HSI and HR-MSI using TMSVD, and the three factor matrices are then used to construct the HR-HSI, which is different from estimating the spectral basis and the corresponding spectral coefficient matrix.
However, compared with SVD, TMSVD is limited by a truncated value. It only saves part of the original matrix information, which will cause the loss of some critical information. To further improve the fusion effect, we assume that the first TMSVD factor matrix of HR-MSI also contains some information about HR-HSI. Therefore, we introduce the first TMSVD factor matrix of HR-MSI into the fusion process.
We integrate the prior information of TMSVD on HR-MSI and LR-HSI to reconstruct HR-HSI, obtaining the factor matrices from LR-HSI and HR-MSI by TMSVD. We then recombine the obtained factor matrices to obtain a rough HR-HSI and, through a multiplicative iterative process, can finally solve the optimization question.
The main contributions of this paper are as follows:
  • We explain that the reason for the spectral basis comes from LR-HSI from the perspective of the matrix similarity after truncated singular value decomposition. No such quantitative analysis has been conducted prior to our work.
  • We found a strong correlation between HR-MSI and HR-HSI. All prior information about TMSVD was integrated into a new proposed fusion model without using the SRF and based only on the TMSVDfactor matrices from LR-HSI and HR-HSI.
  • We propose a new idea for the hyperspectral fusion method. We reconstruct the HR-HSI by estimating the three SVD factor matrices of HR-HSI from LR-HSI and HR-MSI.
  • We test our proposed method on two simulated data sets, Pavia University and CAVE, and a real data set in which the remote sensing images are generated by two different spectral cameras, Sentinel 2 and Hyperion. Compared with the non-blind methods, our proposed method achieves a more effective fusio result while reducing fusing time to less than 1% of such methods. Compared with the blind methods on the simulated data sets, our proposed method can improve the PSNR value by up to 16 dB. Moreover, our proposed method demonstrates a better performance on the real data set, which validates its practicality.
The rest of this paper is organized as follows. In Section 2, some background knowledge and the representative hyperspectral image super-resolution literature are presented. The FTMSVD method is presented in Section 3. The experimental details are outlined in Section 4. Presented in Section 5 are our experimental results and analysis, and the conclusion is found in Section 6.

2. Related Work

In recent years, the exploration of hyperspectral image (HSI) super-resolution methods has been gradually increasing. These methods can be broadly divided into three categories: matrix factorization-based methods, tensor factorization-based methods, and other methods.
In the matrix factorization-based methods, it is assumed that HR-HSI is a matrix formed by the multiplication of the spectral basis and the corresponding spectral coefficient matrix. The fusion problem is transformed into the problem of how to estimate the spectral basis and coefficient matrix. There are many ways to estimate the spectral basis and coefficient matrix, such as non-negative dictionary learning [13], K-SVD [14], and online dictionary learning [15]. Wycoff et al. [16] made full use of the prior information of non-negativity and sparsity of HR-HSI and conducted non-negative sparse matrix factorization for HR-MSI and LR-HSI to obtain the approximation of a non-negative spectral basis and sparse coefficient matrix, and the alternative direction multiplier method (ADMM) was then used in their optimization to obtain the HR-HSI. Akhtar et al. [17] estimated a non-negative dictionary based on the principle of local similarity of images. In [18], the spectral basis and coefficient matrix were learned from LR-HSI and MR-HSI under some prior information. Huang et al. [19] proposed obtaining the learning spectral basis by using K-SVD on LR-HSI. Han et al. [20] clustered image blocks and assumed that similar image blocks can linearly represent the given block, learning of the local similarity of HR-HSI was based on image segmentation. The method we propose in this paper is also a matrix factorization-based method. However, our FTMSVD method differs in two aspects. Firstly, in our FTMSVD method, information is directly obtained from LR-HSI and HR-MSI by TMSVD without the need for a complex operation such as dictionary learning. Secondly, in our proposed method, it is assumed that HR-HSI is composed of three SVD factor matrices, and we estimate the three SVD factor matrices of HR-HSI from LR-HSI and HR-MSI, but not the spectral basis and coefficient matrix.
Tensor factorization-based methods allow for preserving the 3D characteristics of hyperspectral images to a great extent. That is, tensors represent hyperspectral images, and the structure information of the image is better preserved. The question for fusion then concerns how to estimate the tensors. Dian et al. [21] used Tucker decomposition to decompose HR-HSI into three dictionaries of three dimensions and used a kernel tensor to describe the relationship between the dictionaries. Li et al. [22] proposed a method based on coupled sparse tensor representation (CSTF) in which HR-HSI is regarded as a three-dimensional tensor. The tensor could be approximated as a core tensor multiplied by three subtensors. In [11], subspace representation and low-rank tensor representation are combined. The spectral subspace is approximated by the singular value decomposition of LR-HSI, and the coefficients are estimated by the low-rank tensor. Prvost et al. [23] obtained the dictionary of the third modes by the operation of TMSVD, and they solved the generalized Sylvester equation to obtain the kernel tensor.
Other methods mainly include the Bayesian-based and deep convolutional neural network (CNN)-based methods. In the Bayesian-based methods, the prior distribution is used to solve the fusion problem. A representative example is the proposal in [24]. The method based on CNN also plays an important role in the realization of hyperspectral image fusion. Dong et al. [25,26] constructed a deep CNN for solving single image super-resolution and achieved excellent performance. Liu et al. [27] proposed a deep CNN named SSAU-Net, by introducing the spectral-spatial attention module to extract the shallow and deep features information from LR-HSI and HR-HSI. Work [28] cleverly combined subspace representation with CNN denoisier, which only need to train on the gray images. This method solves the difficult problem about the training for CNN model.
These methods are also roughly divided into two categories, depending on whether PSF and SRF are known during the fusion process, as blind fusion methods and non-blind fusion methods. Methods that are non-blind indicate that both PSF and SRF are known. On the contrary, PSF and SRF are unknown in blind methods. Hysure (blind version) [29] and CNMF [18] are blind methods in which the fusion process is achieved by estimating both SRF and PSF from HR-MSI and LR-HSI. The method in [30] is also a blind method because it only estimates the SRF and does not use the PSF during the fusion process. In either case, our proposed method demonstrates good performance without using SRF. When the PSF is known, we use it directly. If the PSF cannot be directly obtained, we can use the PSF that we set. Our experiments demonstrate that our method achieves good results in both cases.

3. Fusion Model

We denote the target HR-HSI as a three-dimensional tensor Z R W × H × L , which has W × H pixels and L bands. The LR-HSI is regarded as X R w × h × L , which has w × h pixels and L bands. Y R W × H × l represents the observed HR-MSI, with W × H pixels and l bands. It is observed that w < W , h < H , and l < L . X can be seen as a spatially downsampled version of Z , i.e.,
X = Z BS
where X R L × w h and Z R L × W H represent mode-n matricization matrix of X and Z , respectively. The matrices B R W H × W H and S R W H × w h represent the convolutional blur and downsampling matrix, respectively. In a practical situation, the PSF represents B , which is uncertain, and S can be obtained by the proportion of spatial dimensions of LR-HSI and HR-MSI.
The Y R W × H × l can also be seen as the spectral downsampling version of Z , i.e.,
Y = R Z
where Y R l × W H and R R l × L represent the mode-n matricization matrix of Y and the spectral response matrix, respectively. R represents the SRF, which is uncertain in most cases.
In [31], SVD was used for image compression, because it can denoise and represent most information of an image with a few elements. Therefore, if we can obtain the three SVD factor matrices of Z , we can recover the original HR-HSI, i.e.,
Z = U Σ V T
where U z R L × d , Σ z R d × d and V z T R d × W H are the corresponding factor matrices of Z by SVD. However, it is difficult to directly obtain the SVD factor matrices of Z . To approximate the SVD factor matrices of Z , we estimate the SVD factor matrices of Z from X and Y by using TMSVD, i.e.,
Z = U z Σ z V z T
where U z R L × q , Σ z R q × q , and V z T R q × W H are the corresponding estimation factor matrices obtained by TMSVD, and q is the truncated value. Therefore, Equations (1) and (2) can be converted to the following equation.
X = U z Σ z V z T B S
Y = R U z Σ z V z T
The fusion problem turns into the problem of estimating three factor matrices U z , Σ z , and V z T from HR-MSI and LR-HSI. To distinguish it from the target matrix Z , we denote Z s as the fusion matrix, i.e.,
Z s = U s Σ s V s T
where U s R L × q , Σ s R q × q and V s T R q × W H are the corresponding factor matrices of fusion result Z s by TMSVD. Since l < L , the truncated singular value q is constrained below l, and we set l as the truncated singular value for the consistency of TMSVD factor matrix dimensions. Through performing the TMSVD operation on Y and X with truncated value q , we can obtain the TMSVD factor matrices of Y and X , respectively, i.e.,
Y = U y Σ y V y T
X = U x Σ x V x T
where U y R l × q , Σ y R q × q and V y T R q × W H , and U x R L × q , Σ x R q × q and V x T R q × w h . They are the corresponding TMSVD factor matrices of X and Y.
As shown in Figure 1 and Figure 2, | V y T | | V z T | and U x U z , and it was found in [12] that Σ z s f × Σ x , where s f is the downsampling factor. Hence, we can obtain the following equations.
U s = U y
Σ s = s f × Σ x
V s T = V y T
However, constrained by the spectral dimension of HR-MSI, the q value is limited in l. Therefore, if we use U s , Σ s and V s T to reconstruct HR-HSI, much important information about HR-HSI will be lost, and the fusion effect is unsatisfactory. Due to the features of SVD, U y also obtains information of Y , and Y contains most of the spatial information of Z , so we introduce it in the reconstruction of V s T to improve the fusion performance, i.e.,
V s T = U y × V y T
The detailed process for estimating the SVD factor matrices of HR-HSI is presented in Algorithm 1.
Based on Equations (1) and (9), we can further optimize the quality of Z s as the following equation, i.e.,
min U s Σ s V s T X U s Σ s V s T B S F 2
where A F 2 means the Frobenius norm of A in this paper.
Algorithm 1 Estimate three rough SVD factor matrices of HR-HSI
Require: X Y
  • obtain the downsampling factor s f and the truncated value q from X and Y
  • [ n r , , q ] = size ( Y )
  • [ n r 1 , , ] = size ( X )
  • s f = n r / n r 1
  • Matrix X and Y to X and Y
  • The truncated singular value decomposition on X and Y with truncated value q
  • U x , Σ x , = SVD ( X , q )
  • U y , , V y T = SVD ( Y , q )
  • U s = U x
  • Σ s = s f × Σ x
  • V s T = U y V y T
  • return U s , Σ s , V s T
Because of the correlation in Figure 1 and [12], we assume that Σ s and V s T are approximately equal to Σ z and V z T , which means Σ s and V s T are fixed during the fusion process. We therefore simplify Equation (14), i.e.,
min U s X U s C F 2
We set C = Σ s V s T B S as constant throughout the fusion process. The meaning of C here is similar to the coefficient matrix in the work [11], except that we get it directly by the connection between HSI and MSI, without any additional operations. Note that there are two cases of B . If PSF is known in practice, we use it directly. If PSF is unknown, we use the one we set, which is a 5 × 5 Gaussian blur (standard deviation 1) in this paper, and solve the uncertainty of B in this way. Finally, a multiplicative iterative process as in [18] is introduced to optimize U s in Equation (15), i.e.,
U s = U s . × ( X C T ) . / ( U s C C T )
After several iterative rounds of optimization, we obtain a more accurate U s . We reconstruct it using Σ s and V s T to obtain a satisfactory Z . The whole process of our method is shown in Algorithm 2.
Algorithm 2 FTMSVD Algorithm
Require: X Y
  • Obtain the U s Σ s V s T via TMSVD with a truncated singular value q in Algorithm 1
  • if PSF is known then
  •    Use it directly
  • else [PSF is unknown]
  •    Set PSF as a 5 × 5 Gaussian blur (standard deviation 1)
  • end if
  • Set C = Σ s V s T B S
  • Z f = U s C
  • Use a multiplicative iterative processe to optimize the U s
  • for k = 1:K do
  •    Update U s by Equation (16)
  • end for
  • Z f = U s Σ s V s T
  • Z = Z f
  • return Z

4. Experiments

4.1. Data Set

To demonstrate the universality and effectiveness of our proposed FTMSVD method, we conduct experiments on two simulated data sets, the CAVE [32] and Pavia University [33] data sets. In addition, we test our proposed method on a real data set of remote sensing images generated by Sentinel 2 and Hyperion spectral cameras.
In the CAVE data set, each HSI has 512 × 512 pixels and 31 spectral bands and consists of 32 indoor scenes with high quality and a spectral range from 400 to 700 nm. We removed the final two fuzzy bands and kept the remaining 29 bands. The Pavia University data set was acquired by the ROSIS sensor, which has 610 × 340 space pixels with 115 spectral bands (0.43–0.86 μ m) from around Pavia University, northern Italy. We removed the final fuzzy bands and kept the remaining 93 bands to use the upper left corner ( 256 × 256 × 93 ). Compared with non-blind methods, we filter each HR-HSI band with the 7 × 7 Gaussian blur (standard deviation 2) and conduct downsampling every 32 pixels in two spatial modes to simulate the LR-HSI in the simulated data sets. The HR-MSIs of the Pavia University data set are obtained using the IKONOS class reflection spectral response filter [34]. The HR-MSIs (RGB image) from the CAVE data set are generated by a Gf-1-16m multispectral camera response function. Specifically, for the Pavia University data set, each LR-HSI is 8 × 8 × 93 and HR-MSI is 256 × 256 × 4 . For the CAVE data set, the LR-HSI is 16 × 16 × 29 and the HR-MSI is 512 × 512 × 3 . For comparison, we set the downsampling as 8 for blind methods and kept other settings the same as for non-blind in generating the LR-HSI ( 32 × 32 × 93 ) and the LR-HSI ( 64 × 64 × 29 ) from the Pavia University and CAVE data sets, respectively. The HR-MSIs are the same as in non-blind fusion processes.
For the real remote sensing data set, the LR-HSI was obtained by the Hyperion sensor on the Earth Observation 1 satellite, which has 220 spectral bands. Our experiments use one part of it as the LR-HSI ( 100 × 100 × 89 ). The corresponding HR-MSI is 300 × 300 × 4 , generated by the Sentinel-2A satellite.
All experiments were implemented in MATLAB R2014a on localhost with 2.90 GHz Intel i5-9400F CPU and 8.0 GB DDR3.

4.2. Experimental Settings

To demonstrate the superiority of our proposed method in different situations (non-blind and blind), we organize the experiments in the following order.
First of all, through two simulated data sets, we compare our proposed method with five state-of-the-art non-blind fusion methods, including Hysure(NB) (non-blind version) [29], CSTF [22], STEREO(NB) (non-blind version) [35], NLSTF [21], and LTTR [36]. We chose the parameters to obtain the best performance for each method in different data sets when the downsampling factor is 32. The experiment results demonstrate that our method results in a better performance in less execution time. The specific parameters of the five state-of-the-art methods are given below in Table 1 and Table 2.
Secondly, we verify the validity of V s T extracted from our finding in Figure 1 by using a randomly selected factor matrix to replace V s T and intuitively display the influence of V s T .
We then compare our proposed method with five state-of-the-art blind methods on two simulated data sets, in which the LR-HSI and HR-MSI are simulated based on different SRF and PSF values. The compared blind methods we used include the Hysure (https://github.com/alfaiate/HySure/ (blind version) accessed on 1 October 2022) [29], CNMF [18], GSA [37], SFIMHS [38], and MAPSMM [39]. Our proposed FTMSVD method, GSA, and SFIMHS do not require additional setting parameters. In fact, it is very difficult to debug the optimal parameters in the real blind fusion scenario, so we keep the parameter settings of Hysure, CNMF, and MAPSMM in the original literature for a realistic and reliable comparison in blind fusion experiments.
After that, we will use the different PSF to replace the one we set in our proposed method to show the impact of the fixed PSF and to further illustrate the generalization of our proposed method.
Finally, we take the experiment on the real data set and show the fusion visual fidelity.

4.3. Quantitative Metrics

To evaluate the effectiveness of our proposed method in all aspects, we use subjective and objective evaluation metrics. For the subjective metrics, we show the effectiveness of our proposed method through the visual presentation of the fused images. We used six objective indexes to simultaneously reflect the fusion results. Regarding the peak signal-to-noise ratio (PSNR) [11], the larger the value of PSNR, the better the spatial reconstruction quality of each band; regarding the Erreur Relative Globale Adimensionnelle de Synthese (ERGAS) [40], the smaller the ERGAS, the better the overall quality of the fused image; regarding spectral angle mapper (SAM) [34], the smaller the SAM, the better the quality of fusion; regarding the structural similarity (SSIM) [41], which is used to evaluate the similarity of two images by brightness, contrast, and structure, the closer the metric is to 1, the better the effect; regarding the universal Image Quality Index (UIQI) [42], which is used to address simultaneously spectral and radiometric quality, the closer the UIQI is to 1 the better result and regarding fusion time, which is used to evaluate the speed of the fusion method, the smaller the value, the faster the fusion.

5. Results and Analysis

5.1. Compared with Non-Blind Methods

The results of the non-blind situation are shown in Table 3 and Table 4, and the values in bold are the best results in the metric. All the data values in the table are the average of the five experiments. For visual comparison, we introduce the image of the experimental results for comparison. Figure 3 and Figure 4 are the corresponding images. Table 3 shows the average objective metrics on the Pavia University data set. Our FTMSVD method has the best performance compared with other non-blind methods on all objective metrics, which means that HR-HSI reconstructed by our FTMSVD method has the best spatial and spectral qualities. More specifically, the PSNR of our FTMSVD method is higher than that of CSTF method by 0.091 dB, but we reduce the execution time by about 258.485 s. From Table 4, it can be seen that our proposed method achieves the best result on the CAVE data set. The main improvement reason is that our method extracts the reconstruction information from X and Y based directly on TMSVD and without complex operations. The optimization we used is multiplicative iterative processes, which only optimize the U s in a short time. The reason why the optimization is so fast is because the optimize process only works on smaller matrices U s . Most matrix-based fusion methods take a large amount of time because they need to optimize both the spectral subspace and the corresponding coefficient matrix. In [11], the approximate estimation of subspace through LR-HSI can reduce the operation, but the coefficient matrix is still very large, and the fusion time is long. The data volume of the coefficient matrix is much larger than the spectral base. Due to the discovery of Figure 1, we can make an approximate substitution of the last two SVD factor matrices of HR-HSI by combining the work [12] with our findings. Therefore, we propose the optimization method in Equation (15) only need to optimize a smaller matrix.
As we can see from Figure 3 and Figure 4, the first row shows the fused images of each method at band 30, and the second row is the enlarged one of the marked area in the first row. The third row shows the error images at band 30, and the more black pixels there are, the closer the fused image is to the ground true image. Take a comprehensive view from Figure 3, when the downsampling factor is 32, the Hysure(NB), CSTF and our proposed method FTMSVD have the better visual effect than STEREO(NB), NLSTF and LTTR. Although the error images of the Hysure(NB) appears to perform better than FTMSVD on the CAVE data set, the FTMSVD still has a better performance than others. In combination with Table 3 and Table 4, FTMSVD method has better performance on both data sets.

5.2. Impact of the V s T to Fusion Effect

Here, we use a randomly selected factor matrix V R T to replace V s T in Algorithm 1 to verify the impact of our finding in Figure 1 on the fusion result. Figure 5 shows the PSNR values in two simulated data sets. One uses V s T and the other uses V R T . We can see that in using V R T , the value of PSNR is very low in the two data sets, which indicates poor fusion effect. The PSNR is much higher when using V s T in all data sets, which verifies that V s T can be effectively used to improve the fusion performance, and the improvement effect is pronounced. The main reason for this result is that V s T is directly obtained from the truncated singular value decomposition on HR-MSI, which contains large spatial information of HR-MSI. Due to the HR-MSI being a spectral downsampling of HR-HSI, V s T is a better approximation of V z T . Therefore, we can obtain a better fusion result by using V s T to reconstruct HR-HSI.

5.3. Comparison with Blind Methods

To comprehensively evaluate the superiority of our proposed method in blind fusion cases, we compare it with five state-of-the-art blind methods in this section. To make a more direct comparison, we will compare the two simulated data sets, in which the HR-HSI (which we obtain) can be compared with the fusion results of blind methods in terms of objective metrics. We exchange the PSF of CAVE and Pavia University to generate six LR-HSI and HR-MSI pairs. Here, we refer to the IKONOS class reflection spectral response filter as SRF1 and the Gf-1-16m multispectral camera response as SRF2. The 7 × 7 Gaussian blur (standard deviation 2) is defined as PSF1, and the 3 × 3 average blur is PSF2. Different combinations of these are used to generate the LR-HSI and the HR-MSI. We keep the PSF1 for the Pavia University data set and only use different SRFs. We change both the SRF and the PSF for the CAVE data set. The downsampling factor is 8. We compared the blind methods on the simulated data sets according to the objective fusion metrics, and the image of the fusion result of some data sets is visually displayed.
Table 5 shows the fusion result on the Pavia University data set. As we can see, our proposed method achieved the best performance on most metrics in either case. The PSNR value of our proposed method is much better than other blind methods, where the maximum PSNR reached is almost 16 dB, and the minimum is almost 5 dB. The improvement in PSNR demonstrates the superiority of our proposed method in the blind fusion effect. Although the fusion time of our proposed method is not the best, it can still achieve the second best fusion time, which is only slightly longer than the time for the best performance. Table 6 shows the experimental results on the CAVE data set. Our proposed method achieves the best result on all metrics except fusion time. In the case of SRF1 and PSF2, our proposed method can achieve 43.274 dB for PSNR, while the second best is only 36.445 dB. The fusion time of our proposed method is still a little longer than that of SFIMHS, but the PSNR value of SFIMHS is only 22.503 dB. From Table 5 and Table 6, our proposed method is seen to be significantly better than other methods in terms of fusion quality. Due to page limitations, we only show the fusion images using the SRF1 and PSF1 on both data sets. Figure 6 are the fusion images of six blind methods on two data sets. In the upper left corner of each image is an enlargement of the portion marked in that image. From the performance on the two data sets in Figure 6, we can intuitively see from the visual fidelity that the image produced from our method is closer to the original image than those of other compared blind methods. From Figure 6, we can intuitively see that our fusion effect is closer to the original image than other compared blind methods. The main reason is that the information we extract from LR-HSI and HR-MSI can effectively simulate the corresponding three SVD factor matrices of HR-HSI, which saves the primary spectral information of LR-HSI and the main spatial information of HR-MSI. In addition, the method is independent of SRF, which allows avoiding the error and the time cost associated with SRF estimation. In addition, our proposed method does not depend on SRF, which can avoid the error of estimating the SRF and save a lot of fusion time.
To further contrast the superiority of our method, we compare the performance of the compared methods in each spectral band. Figure 7 shows the change of PSNR for each spectral band over two data sets. As we can see from Figure 7a, the FTMSVD has the best performance than the other methods in each band. Figure 7b shows that the FTMSVD performs better than the compared methods at the most of the spectral bands. A higher PSNR value means that the spectral reflectance of the fusion result is more similar to the ground truth, which further demonstrates the advantages of our method.

5.4. The Impact of the PSF Set in the FTMSVD Method

Table 7 and Table 8 show the impact of different typical blur types as PSF set in our method. The data sets we used are the same as in Table 5 and Table 6 with SRF1 and PSF1. PSFS1 is the 5 × 5 Gaussian blur (standard deviation 1) (which we used in this paper), PSFS2 is the 5 × 5 average blur, PSFS3 is the 7 × 7 average blur, PSFS4 is the 9 × 9 Gaussian blur (standard deviation 2), and PSFS5 is the motion blur (len is 9 and theta is 0).
As we can see from Table 7 and Table 8, the PSF set in our proposed method can influence the fusion effect, but no matter which PSF we used, the fusion result is better than for the five compared blind methods in Table 5 and Table 6. The main reason is that the PSF is only used to optimize the first SVD factor matrix U s , and U s mainly contains the spectral information of HR-HSI, and its size of is smaller than V s T . For different PSF, the optimization effect of U s is not significantly different. Moreover, most spatial information of HR-HSI is in Σ s and V s T and unaffected by PSF. In this way, our proposed method only needs to assume a fixed PSF to replace its estimation. A multiplicative iterative process is then used to optimize U s , which can take a concise time to converge, so our proposed method can achieve a good fusion result within a short time.

5.5. Practicality of the Proposed Method

To demonstrate the practicality of our proposed method, we test our method with the five blind methods on the real remote sensing data set. The LR-HSI ( 100 × 100 × 89 ) and HR-MSI ( 300 × 300 × 4 ) are extracted from the real remote sensing data set. Figure 8 shows the LR-HSI, HR-MSI, and the corresponding fusion images of the different blind methods with enlargement of the marked position. Since the HR-HSI of the real data set is unknown, we directly compared the fusion effect from the fusion images. As Figure 8 shows, our proposed method has a better visual fidelity than others, and it is closer to the HR-MSI, which demonstrates its practicality.

6. Conclusions

In this paper, we find a strong correlation between the HR-MSI and HR-HSI. We experimentally verified the reason why the basic spectral values can be estimated from LR-HSI. We propose a fast fusion method based on TMSVD without SRF to implement the process of HR-MSI and LR-HSI fusion, which is suitable for both non-blind and blind fusion situations. Since LR-HSI contains the most spectral information of HR-HSI, the spatial information is mainly in HR-MSI. We estimated the SVD factor matrices of HR-HSI from HR-MSI and LR-HSI by TMSVD. Our proposed method solves the problems associated with the uncertainty of PSF and SRF, and a satisfactory fusion result can be rapidly produced. Our experimental results on the simulated hyperspectral data sets and real remote sensing data set all indicate the superiority of our proposed method. However, our pre-set PSF still has a slight impact on the fusion results of our proposed methods. Therefore, it is necessary for us to estimate a more suitable PSF based on the LR-HSI in future studies. The findings in this paper can be used as prior information for future fusion methods. Moreover, estimating SVD factor matrices of HR-HSI also represents a novel concept for conducting hyperspectral image fusion.

Author Contributions

Conceptualization, H.L., J.L., Y.P. and T.Z.; methodology, H.L. and J.L.; software, H.L.; validation, H.L.; data curation, H.L.; writing—original draft preparation, H.L.; writing—review and editing, and J.L, Y.P. and T.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the Opening Foundation of State Key Laboratory of High-Performance Computing, National University of Defense Technology, under Grant No. 202201-05.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Sample Availability

The code in the articles is available from the authors.

References

  1. Samiappan, S. Spectral Band Selection for Ensemble Classification of Hyperspectral Images with Applications to Agriculture and Food Safety. Ph.D. Thesis, Mississippi State University, Starkville, MS, USA, 2014. [Google Scholar]
  2. Tiwari, K.; Arora, M.; Singh, D. An assessment of independent component analysis for detection of military targets from hyperspectral images. Int. J. Appl. Earth Obs. Geoinf. 2011, 13, 730–740. [Google Scholar] [CrossRef]
  3. Gu, Y.; Wang, C.; Wang, S.; Zhang, Y. Kernel-based regularized-angle spectral matching for target detection in hyperspectral imagery. Pattern Recognit. Lett. 2011, 32, 114–119. [Google Scholar] [CrossRef]
  4. Luft, L.; Neumann, C.; Freude, M.; Blaum, N.; Jeltsch, F. Hyperspectral modeling of ecological indicators—A new approach for monitoring former military training areas. Ecol. Indic. 2014, 46, 264–285. [Google Scholar] [CrossRef]
  5. Cui, Y.; Zhang, B.; Yang, W.; Wang, Z.; Li, Y.; Yi, X.; Tang, Y. End-to-End Visual Target Tracking in Multi-Robot Systems Based on Deep Convolutional Neural Network. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017. [Google Scholar]
  6. Cui, Y.; Zhang, B.; Yang, W.; Yi, X.; Tang, Y. Deep CNN-based Visual Target Tracking System Relying on Monocular Image Sensing. In Proceedings of the 2018 International Joint Conference on Neural Networks, IJCNN 2018, Rio de Janeiro, Brazil, 8–13 July 2018. [Google Scholar]
  7. van der Meer, F.D.; van der Werff, H.M.; van Ruitenbeek, F.J.; Hecker, C.A.; Bakker, W.H.; Noomen, M.F.; van der Meijde, M.; Carranza, E.J.M.; Smeth, J.B.d.; Woldai, T. Multi- and hyperspectral geologic remote sensing: A review. Int. J. Appl. Earth Obs. Geoinf. 2012, 14, 112–128. [Google Scholar] [CrossRef]
  8. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral Remote Sensing Data Analysis and Future Challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar]
  9. Zhong, P.; Wang, R. Learning Conditional Random Fields for Classification of Hyperspectral Images. IEEE Trans. Image Process. Publ. IEEE Signal Process. Soc. 2010, 19, 1890–1907. [Google Scholar]
  10. He, N.; Fang, L.; Li, S.; Plaza, A.; Plaza, J. Remote Sensing Scene Classification Using Multilayer Stacked Covariance Pooling. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6899–6910. [Google Scholar] [CrossRef]
  11. Dian, R.; Li, S. Hyperspectral Image Super-resolution via Subspace-Based Low Tensor Multi-Rank Regularization. IEEE Trans. Image Process. Publ. IEEE Signal Process. Soc. 2019, 28, 5135–5146. [Google Scholar] [CrossRef]
  12. Peng, Y.; Li, J.; Zhang, L.; Xu, Y.; Long, J. Hyperspectral image super-resolution via subspace-based fast low tensor multi-rank regularization. Infrared Phys. Technol. 2021, 116, 103631. [Google Scholar]
  13. Dong, W.; Fu, F.; Shi, G.; X, C.; JJ, W.; GY, L.; X, L. Hyperspectral Image Super-Resolution via Non-Negative Structured Sparse Representation. IEEE Trans. Image Process. Publ. IEEE Signal Process. Soc. 2016, 25, 2337–2352. [Google Scholar]
  14. Bruckstein, M.A.E. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar]
  15. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G. Online dictionary learning for sparse coding. In Proceedings of the ICML’09: 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009. [Google Scholar]
  16. Wycoff, E.; Chan, T.H.; Jia, K.; Ma, W.K.; Ma, Y. A non-negative sparse promoting algorithm for high resolution hyperspectral imaging. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013. [Google Scholar]
  17. Akhtar, N.; Shafait, F.; Mian, A. Sparse Spatio-spectral Representation for Hyperspectral Image Super-resolution. In Proceedings of the ECCV 2014, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  18. Yokoya, N.; Yairi, T.; Iwasaki, A. Coupled Nonnegative Matrix Factorization Unmixing for Hyperspectral and Multispectral Data Fusion. IEEE Trans. Geosci. Remote Sens. 2012, 50, 528–537. [Google Scholar] [CrossRef]
  19. Huang, B.; Song, H.; Cui, H.; Peng, J.; Xu, Z. Spatial and Spectral Image Fusion Using Sparse Matrix Factorization. IEEE Trans. Geosci. Remote Sens. 2014, 52, 1693–1704. [Google Scholar] [CrossRef]
  20. Han, X.; Shi, B.; Zheng, Y. Self-Similarity Constrained Sparse Representation for Hyperspectral Image Super-Resolution(Article). IEEE Trans. Image Process. 2018, 27, 5625–5637. [Google Scholar] [CrossRef]
  21. Dian, R.; Fang, L.; Li, S. Hyperspectral image super-resolution via non-local sparse tensor factorization. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  22. Li, S.; Dian, R.; Fang, L.; Bioucas-Dias, J. Fusing Hyperspectral and Multispectral Images via Coupled Sparse Tensor Factorization. IEEE Trans. Image Process. 2018, 27, 4118–4130. [Google Scholar] [CrossRef]
  23. Prévost, C.; Usevich, K.; Comon, P.; Brie, D. Hyperspectral super-resolution with coupled Tucker approximation: Recoverability and SVD-based algorithms. IEEE Trans. Signal Process. 2020, 68, 931–946. [Google Scholar] [CrossRef] [Green Version]
  24. Hardie, R.C.; Eismann, M.T.; Wilson, G.L. MAP estimation for hyperspectral image resolution enhancement using an auxiliary sensor. IEEE Trans. Image Process. Publ. IEEE Signal Process. Soc. 2004, 13, 1174–1184. [Google Scholar] [CrossRef]
  25. Dong, C.; Loy, C.; He, K.M.; Tang, X.O. Learning a Deep Convolutional Network for Image Super-Resolution. In Proceedings of the 2014 European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  26. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. arXiv 2016, arXiv:1501.00092. [Google Scholar] [CrossRef] [Green Version]
  27. Zhang, S.L.L.Z.L.H.D. SSAU-Net: A Spectral–Spatial Attention-Based U-Net for Hyperspectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5542116. [Google Scholar]
  28. Kang, R.D.L. Regularizing Hyperspectral and Multispectral Image Fusion by CNN Denoiser. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 1124–1135. [Google Scholar]
  29. Simões, M.; Bioucas-Dias, J.; Almeida, L.B.; Chanussot, J. A Convex Formulation for Hyperspectral Image Superresolution via Subspace-Based Regularization. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3373–3388. [Google Scholar] [CrossRef] [Green Version]
  30. Peng, Y.; Long, J. Blind Fusion of Hyperspectral Multispectral Images Based on Matrix Factorization. Remote Sens. 2021, 13, 4219. [Google Scholar]
  31. Swathi, H.; Sohini, S.; Surbhi; Gopichand, G. Image compression using singular value decomposition. IOP Conf. Ser. Mater. Sci. Eng. 2017, 263, 042082. [Google Scholar] [CrossRef]
  32. Yasuma, F.; Mitsunaga, T.; Iso, D. Generalized Assorted Pixel Camera: Postcapture Control of Resolution, Dynamic Range, and Spectrum. IEEE Trans. Image Process. 2010, 19, 2241–2253. [Google Scholar] [CrossRef] [Green Version]
  33. Dell’Acqua, F.; Gamba, P.; Ferrari, A.; Palmason, J.; Benediktsson, J.; Arnason, K. Exploiting spectral and spatial information in hyperspectral urban data with high resolution. Geosci. Remote Sens. Lett. IEEE 2004, 1, 322–326. [Google Scholar] [CrossRef]
  34. Wei, Q.; Bioucas-Dias, J.; Dobigeon, N.; Tourneret, J.Y. Hyperspectral and Multispectral Image Fusion Based on a Sparse Representation. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3658–3668. [Google Scholar] [CrossRef] [Green Version]
  35. Kanatsoulis, C.; Fu, X.; Sidiropoulos, N.; Ma, W. Hyperspectral Super-Resolution: A Coupled Tensor Factorization Approach. IEEE Trans. Signal Process. 2018, 66, 6503–6517. [Google Scholar] [CrossRef] [Green Version]
  36. Dian, R.; Li, S.; Fang, L. Learning a Low Tensor-Train Rank Representation for Hyperspectral Image Super-Resolution. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 2672–2683. [Google Scholar] [CrossRef]
  37. Aiazzi, B.; Baronti, S.; Selva, M. Improving Component Substitution Pansharpening Through Multivariate Regression of MS +Pan Data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  38. Liu, J.G. Smoothing Filter-based Intensity Modulation: A spectral preserve image fusion technique for improving spatial details. Int. J. Remote Sens. 2000, 21, 3461–3472. [Google Scholar] [CrossRef]
  39. Eismann, M.T. Resolution Enhancement of Hyperspectral Imagery Using Maximum a Posteriori Estimation with a Stochastic Mixing Model. Ph.D. Thesis, University of Dayton, Dayton, OH, USA, 2004. [Google Scholar]
  40. Wald, L. Quality of high resolution synthesised images: Is there a simple criterion ? In Proceedings of the Third Conference “Fusion of Earth Data: Merging Point Measurements, Raster Maps and Remotely Sensed Images”, Sophia Antipolis, France, 26–28 January 2000; pp. 99–103. [Google Scholar]
  41. Yokoya, N.; Grohnfeldt, C.; Chanussot, J. Hyperspectral and Multispectral Data Fusion A comparative review of the recent literature. IEEE Geosci. Remote Sens. Mag. 2017, 5, 29–56. [Google Scholar] [CrossRef]
  42. Wang, Z.; Bovik, A. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
Figure 1. The correlation of the third factor matrix (the first row of matrix) after TMSVD of HR-HSI and HR-MSI.
Figure 1. The correlation of the third factor matrix (the first row of matrix) after TMSVD of HR-HSI and HR-MSI.
Remotesensing 15 00207 g001
Figure 2. The correlation of the first factor matrix after TMSVD of HR-HSI and LR-HSI.
Figure 2. The correlation of the first factor matrix after TMSVD of HR-HSI and LR-HSI.
Remotesensing 15 00207 g002
Figure 3. The first row shows the fused images by the competing method for Pavia University at band 30. The second row shows the magnified version of the marked area. The third row shows the error images of the test methods for Pavia University at band 30.
Figure 3. The first row shows the fused images by the competing method for Pavia University at band 30. The second row shows the magnified version of the marked area. The third row shows the error images of the test methods for Pavia University at band 30.
Remotesensing 15 00207 g003
Figure 4. The first row shows the fused images by the competing methods for oil_painting_ms (CAVE data set) at band 16. The second row shows the magnified version of the marked area. The third row shows the error images of the test methods for oil_painting_ms at band 16.
Figure 4. The first row shows the fused images by the competing methods for oil_painting_ms (CAVE data set) at band 16. The second row shows the magnified version of the marked area. The third row shows the error images of the test methods for oil_painting_ms at band 16.
Remotesensing 15 00207 g004
Figure 5. Impact of V s T to the fusion effect.
Figure 5. Impact of V s T to the fusion effect.
Remotesensing 15 00207 g005
Figure 6. The first row shows LR-HSI, fusion images, and ground truth images’ false-color images in Pavia university data set (formed by bands 30th, 60th, and 90th). The second row shows LR-HSI, fusion images, and ground truth images’ false-color images in CAVE data set (formed by bands 10th, 19th, and 28th). All fusion images and LR-HSI are obtained using SRF1 and PSF1.
Figure 6. The first row shows LR-HSI, fusion images, and ground truth images’ false-color images in Pavia university data set (formed by bands 30th, 60th, and 90th). The second row shows LR-HSI, fusion images, and ground truth images’ false-color images in CAVE data set (formed by bands 10th, 19th, and 28th). All fusion images and LR-HSI are obtained using SRF1 and PSF1.
Remotesensing 15 00207 g006
Figure 7. The PSNR of each spectral band for the compared methods using SRF1 and PSF1.
Figure 7. The PSNR of each spectral band for the compared methods using SRF1 and PSF1.
Remotesensing 15 00207 g007
Figure 8. The fusion results of the blind methods on the real data set (the first band).
Figure 8. The fusion results of the blind methods on the real data set (the first band).
Remotesensing 15 00207 g008
Table 1. Parameters of non-blind methods on the Pavia University data set.
Table 1. Parameters of non-blind methods on the Pavia University data set.
MethodParameters
Hysure(NB)basis = ’VCA’, λ φ = 1 × 10 7 , λ m = 1 × 10 3 , λ B = λ R = 10 , mu = 0.05, iters = 200
CSTF n w = 400 , n h = 400 , n s = 8 , and λ = 1 × 10 6 , β = 0.001 , K = 20
STEREO(NB) λ = 1 × 10 3 , s _ i t e r = 2 , t _ r a n k = 40 , m a x i t = 5 , k e r n e l _ l e n g t h = 21
NLSTF W = H = 10 , S = 14 , K = 151 , λ = 1 × 10 6 , λ 1 = λ 2 = λ 3 = 1 × 10 5
LTTR K = 3900 , e t a = 0.002
Table 2. Parameters of non-blind methods on the CAVE data set.
Table 2. Parameters of non-blind methods on the CAVE data set.
MethodParameters
Hysure(NB)basis = ’VCA’, λ φ = 1 × 10 1 , λ m = 1 × 10 5 , λ B = λ R = 10 , mu = 0.05, iters = 200
CSTF n w = 400 , n h = 400 , n s = 10 , and λ = 1 × 10 5 , β = 0.001 , K = 20
STEREO(NB) λ = 1 × 10 4 , s _ i t e r = 1 , t _ r a n k = 40 , m a x i t = 100 , k e r n e l _ l e n g t h = 31
NLSTF W = H = 10 , S = 14 , K = 151 , λ = 1 × 10 6 , λ 1 = λ 2 = λ 3 = 1 × 10 5
LTTR K = 760 , e t a = 0.0009
Table 3. Average quantitative results of the non−blind methods on the Pavia University data set.
Table 3. Average quantitative results of the non−blind methods on the Pavia University data set.
MethodPSNRERGASSAMUIQISSIMTIME
Hysure(NB)42.4940.1542.2050.9920.98854.925
CSTF43.0770.1442.1400.9930.989258.772
STEREO(NB)27.2630.7928.7320.7720.6961.370
NLSTF29.5560.6727.3820.9240.93660.850
LTTR32.4380.6767.6750.9080.932236.399
FTMSVD43.1680.1402.0490.9930.9890.287
Table 4. Average quantitative results of the non−blind methods on the CAVE data set.
Table 4. Average quantitative results of the non−blind methods on the CAVE data set.
MethodPSNRERGASSAMUIQISSIMTIME
Hysure(NB)40.1380.42315.3500.9340.956220.819
CSTF40.6740.4638.0320.9140.957335.912
STEREO(NB)31.3480.69016.9270.7390.83716.549
NLSTF24.6481.51814.3770.7170.814185.460
LTTR33.9471.21913.1320.7900.906319.596
FTMSVD41.1150.38713.9160.9370.9590.835
Table 5. Average quantitative results of the blind methods on the Pavia University data set.
Table 5. Average quantitative results of the blind methods on the Pavia University data set.
SRF & PSFMethodPSNRERGASSAMUIQISSIMTIME
SRF1 PSF1Hysure38.1800.9583.0070.9860.98152.678
CNMF26.3233.5276.0990.8240.80514.823
GSA27.5993.1528.0290.9190.9100.882
MAPSMM23.1445.4128.0610.7010.71540.453
SFIMHS23.9744.9596.8310.7660.7780.301
FTMSVD40.1220.7402.3570.9900.9870.356
SRF1 PSF2Hysure36.4451.1483.2640.9800.97453.378
CNMF24.6384.2017.8420.7810.74115.305
GSA24.2624.59311.4610.8640.8360.790
MAPSMM21.6346.3499.3480.5860.61741.387
SFIMHS22.5035.8228.0380.6990.7270.197
FTMSVD43.2740.5482.0140.9940.9890.380
Table 6. Average quantitative results of the blind methods on the CAVE data set.
Table 6. Average quantitative results of the blind methods on the CAVE data set.
SRF & PSFMethodPSNRERGASSAMUIQISSIMTIME
SRF1 PSF1Hysure28.8013.61919.4930.7990.851230.795
CNMF30.7353.2517.4110.7920.87454.099
GSA33.6462.26210.2100.8750.9101.061
MAPSMM27.0694.5637.9490.7380.849173.955
SFIMHS26.3095.1366.8340.7970.8800.276
FTMSVD40.4381.0446.2700.9500.9720.840
SRF1 PSF2Hysure26.9774.46322.1770.7400.805243.398
CNMF27.9084.4228.2200.7300.82559.904
GSA32.7082.48211.9180.8500.8781.111
MAPSMM26.0935.0958.7840.6620.813166.409
SFIMHS25.5805.4847.4080.7660.8640.257
FTMSVD41.6140.9645.1470.9540.9790.961
SRF2 PSF1Hysure35.5631.7408.2740.9220.955256.527
CNMF31.8602.6307.2760.8170.89651.648
GSA34.9741.9679.3060.8980.9261.066
MAPSMM27.1864.4507.8750.7860.870140.832
SFIMHS26.7924.6646.5330.8130.8850.250
FTMSVD40.8061.54213.9950.9380.9590.852
SRF2 PSF2Hysure35.3441.7958.1260.9190.953249.050
CNMF29.9293.3958.1310.7720.86745.055
GSA33.6652.22110.8380.8760.8991.017
MAPSMM26.2354.9628.3880.7410.849138.072
SFIMHS25.7785.2407.1680.7730.8650.213
FTMSVD41.1231.53513.9770.9380.9590.733
Table 7. The results of the different PSF set in FTMSVD on the Pavia University data set.
Table 7. The results of the different PSF set in FTMSVD on the Pavia University data set.
MethodPSNRERGASSAMUIQISSIMTIME
PSFS140.1220.7402.3570.9900.9870.453
PSFS242.9760.5592.0440.9930.9900.228
PSFS342.4730.5912.0900.9930.9880.255
PSFS443.0000.5642.0400.9930.9890.211
PSFS542.5290.5912.0430.9930.9890.203
Table 8. The results of the different PSF set in FTMSVD on the CAVE data set.
Table 8. The results of the different PSF set in FTMSVD on the CAVE data set.
MethodPSNRERGASSAMUIQISSIMTIME
PSFS140.4381.0446.2700.9500.9720.994
PSFS241.4310.9775.4780.9550.9780.737
PSFS341.4480.9875.0890.9530.9770.818
PSFS441.5440.9765.0840.9540.9780.792
PSFS541.1710.9945.5070.9550.9780.804
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, H.; Long, J.; Peng, Y.; Zhou, T. Hyperspectral Multispectral Image Fusion via Fast Matrix Truncated Singular Value Decomposition. Remote Sens. 2023, 15, 207. https://doi.org/10.3390/rs15010207

AMA Style

Lin H, Long J, Peng Y, Zhou T. Hyperspectral Multispectral Image Fusion via Fast Matrix Truncated Singular Value Decomposition. Remote Sensing. 2023; 15(1):207. https://doi.org/10.3390/rs15010207

Chicago/Turabian Style

Lin, Hong, Jian Long, Yuanxi Peng, and Tong Zhou. 2023. "Hyperspectral Multispectral Image Fusion via Fast Matrix Truncated Singular Value Decomposition" Remote Sensing 15, no. 1: 207. https://doi.org/10.3390/rs15010207

APA Style

Lin, H., Long, J., Peng, Y., & Zhou, T. (2023). Hyperspectral Multispectral Image Fusion via Fast Matrix Truncated Singular Value Decomposition. Remote Sensing, 15(1), 207. https://doi.org/10.3390/rs15010207

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop