Next Article in Journal
UAV-Based Remote Sensing Applications for Bridge Condition Assessment
Previous Article in Journal
A New Method for Winter Wheat Mapping Based on Spectral Reconstruction Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Snapshot Compressive Imaging with Non-Local Spatial-Spectral Residual Network

1
Jiangsu Key Laboratory of Big Data Analysis Technology, Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology, Nanjing University of Information Science and Technology, Nanjing 210044, China
2
School of Geographical Sciences, Nanjing University of Information Science and Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(9), 1812; https://doi.org/10.3390/rs13091812
Submission received: 26 February 2021 / Revised: 27 April 2021 / Accepted: 27 April 2021 / Published: 6 May 2021

Abstract

:
Snapshot Compressive Imaging is an emerging technology that is based on compressive sensing theory to achieve high-efficiency hyperspectral data acquisition. The core problem of this technology is how to reconstruct 3D hyperspectral data from the 2D snapshot measurement in a fast and high-quality manner. In this paper, we propose a novel deep network, which consists of the symmetric residual module and the non-local spatial-spectral attention module, to learn the reconstruction mapping in a data-driven way. The symmetric residual module uses symmetric residual connections to improve the potential of interaction between convolution operations and further promotes the fusion of local features. The non-local spatial-spectral attention module is designed to capture the non-local spatial-spectral correlation in the hyperspectral image. Specifically, this module calculates the channel attention matrix to capture the global correlations between all of the spectral channels, and it fuses the channel attention attained feature maps and the spatial attention weighted features as the module output, thus both of the spatial-spectral correlations of hyperspectral images can be fully utilized for reconstruction. In addition, a compound loss, including the reconstruction loss, the measurement loss, and the cosine loss, is designed to guide the end-to-end network learning. We experimentally evaluate the proposed method on simulation and real datasets. The experimental results show that the proposed network outperforms the competing methods in terms of the reconstruction quality and running time.

1. Introduction

Hyperspectral images (HSIs) are three-dimensional data cubes, in which the first two dimensions represent spatial information, and the third dimension represents spectral information of scene objects [1]. By performing high-resolution spectral imaging, each pixel can contain dozens or hundreds of spectral bands. Therefore, HSIs not only reflect the spatial geometric distribution of the scene, but they also obtain the spectral signature for each pixel in the scene. The spectral signature can reflect the variation of reflectance of a material with respect to wavelengths, such that they can be used to identifying materials and detect the object in the scene [2]. HSI has been applied in many fields, such as remote sensing [3], precision agriculture [4], and military applications [5].
Although hyperspectral data are three-dimensional, hyperspectral imagers usually detect the spatial-spectral data through one-dimensional line sensors or two-dimensional sensors. In order to acquire the full hyperspectral cube, some representative hyperspectral imaging devices, including push broom [6] and whisk broom [7] and staring imagers [8], need to perform spatial scanning or spectral scanning to complete the acquisition of three-dimensional spatial-spectral information. Different from these scanning based spectral imagers, snapshot compressive imaging systems take advantage of the compressing sensing technology to sample the whole spatial-spectral data by snapshot measurement without scanning [9,10]. According to this mechanism, a Coded Aperture Snapshot Spectral Imaging (CASSI) system [9], as a representative type of hyperspectral snapshot imaging system, has been developed for more than ten years. Specifically, CASSI systems obtain a 2D snapshot measurement by a linear random encoding of the whole data cube according to the compressive sensing theory. The most significant benefits of these snapshot hyperspectral imaging systems over the scanning based imagers are the lower data sampling volume and shorter imaging time. Owing to these advantages, CASSI systems have the capacity to achieve high-speed hyperspectral imaging. However, the snapshot measurement is a projection transformation of the original data value, not the data value itself. The CASSI system needs to solve an optimization problem to obtain the final reconstruction.
The task of reconstructing HSI from the acquired snapshot measurement is a highly ill-posed problem due to the under-mined acquisition mode of CASSI systems [10,11]. To cope with this issue, many studies try to exploit pre-defined image priors to formulate the reconstruction as a regularized optimization problem. Some commonly used priors include the sparse representation, the total variation (TV) [12], non-local similarity [13], and so on [14]. However, solving these problems requires the use of time-consuming iterative optimization, which leads to high reconstruction complexity. This has become an important factor hindering the practical application of the CASSI system. At the same time, these predefined priors cannot describe the spatial-spectral correlation characteristics of hyperspectral data well, which reduces the reconstruction quality. With the excellent learning ability of deep networks [15], scholars are committed to using deep convolutional networks to supervisely learn the explicit mapping from snapshot measurement to the original HSI. This end-to-end learning method can significantly reduce the reconstruction time. However, these existing deep learning methods do not make full use of the coupled spatial-spectral structure of hyperspectral data in network design. In terms of spectral dimension, there are correlations between not only adjacent channels, but also global channels. Because each channel of HSI is the imaging of the same scene at different wavelengths, and these wavelengths are densely sampled at certain intervals within a certain range. In terms of spatial dimension, neighboring pixels usually have similar spectral characteristics. For this reason, these prior structures should be used in the design of the network architecture, which can further improve the quality of reconstruction.
In this paper, we propose a novel Non-local Spatial-Spectral Residual Network (dubbed as NSSR-Net) to learn the parametric reconstruction mapping. The proposed network exploits the symmetric residual module and the non-local spatial-spectral attention module to represent the underlying hyperspectral data, and learns network parameters in a supervised manner under the constraint of a well-defined compound loss function, as shown in Figure 1. Subsequently, we only need to feed the snapshot measurement of the test sample to the well-trained network to achieve efficient and fast reconstruction. Our main contributions can be summarized as follows:
  • we propose a non-local spatial-spectral attention module to represent the HSI data, and both the spatial structure and the global correlations between spectral channels are exploited to improve the reconstruction quality;
  • we design a compound loss, consisting of the reconstruction loss, the measurement loss and the cosine loss, to supervise the network learning. In particular, the cosine loss can further enhance the fidelity of the reconstructed spectral signatures;
  • and experimental results demonstrate that the proposed model achieves better performance on simulation and real datasets, which proves the effectiveness and superiority of the proposed network.

2. Related Work

Hyperspectral snapshot compressive imaging is an important manner for achieving efficient spatial-spectral data acquisition. Specifically, it follows that the computational imaging mechanism encodes the scene content into a snapshot measurement through the principle of compressive sensing and decodes it through a reconstruction algorithm. How to develop a fast and efficient reconstruction algorithm is a key problem for hyperspectral snapshot compressive imaging. Many methods have been proposed to cope with this problem. The prior-driven method is a classical reconstruction framework, which models the reconstruction as a convex optimization problem with prior regularization, and obtains an ideal hyperspectral image through iterative optimization. With the development of deep learning, recent attention has focused more on developing network-driven methods and exploits the deep network to learn the reconstruction mapping from training dataset. In the following, we briefly introduce some representative work in these two categories of methods.
Prior-Driven methods: because to the inherently underdetermined measurement, the prior-driven methods utilize the diverse priors to regularize the reconstruction problem. The objective function of the reconstruction model can be formulated as a weighted sum of the regularization term associated with HSI priors and a data fidelity term associated with the imaging observation equation. A primary concern of prior-driven methods is how to design proper priors to characterize the spatial-spectral correlations in HSIs. Ref. [10] used the wavelet transform to represent each sub-band image of the unknown HSI and formulate the reconstruction as a sparse optimization problem. The total variation (TV) prior is recognized to be effective in maintaining the sharp structures, and it was also used for hyperspectral snapshot compressive reconstruction to improve the reconstruction accuracy [16]. Ref. [17] proposed an adaptive non-local sparse representation model to improve the performance. Liu et al. [18] exploited the weighted nuclear norm to characterize the low rank prior of a group of matched patches. The reconstruction performance of prior-driven methods largely depends on the prior regularization used. However, these priors are hand-crafted and cannot match the characteristics of hyperspectral data well, thus affecting the reconstruction quality.
Given the established reconstruction model, we need to perform iterative optimization to find the final reconstruction. Many optimization algorithms, including iterative shrinkage thresholding, proximal gradient, and the alternating direction method of multipliers, are used to reduce the optimization complexity of each iteration through decomposing the original complex problems into simple sub-problems [19]. However, each iteration still involves huge matrix multiplication, which is time-consuming.
Network-Driven methods: deep networks have made gratifying progress in vision-related tasks [15,20,21]. With the help of the excellent representation ability of deep network, some scholars apply it to compressive sensing reconstruction, forming a network-driven reconstruction method. Different from the iterative optimization based methods, the network-driven methods can directly learn an explicit mapping from the compressive measurement to the HSI and reconstruct the new HSI by just performing a feed-forward computation over the learnt network.
Here, we introduce some representative deep networks for hyperspectral snapshot reconstruction. Xiong et al. proposed a convolution network, dubbed Hscnn [22], to learn the incremental residual to enhance HSI reconstruction. Chol et al. trained an autoencoder to learn the nonlinear spectral representation and employed it as a spectral prior of the variational model for reconstruction [23]. Wang et al. [24] unrolled the iterative optimization of HSI reconstruction into a deep network, and then learned the parameters simultaneously. Zheng et al. [25] exploited a deep-learning-based denoisers as regularization priors and embedded it into the optimization framework for spectral snapshot compressive reconstruction. Miao et al. [26] proposed a two-stage conditional generative model, named λ –net, to generate the reconstruction conditional on the CASSI measurement and masks. A discriminator is also employed by λ –net to discriminate whether the network output is a reconstructed HSI or ground-truth.
Because of the correlation between spatial pixels and spectral bands in HSIs, a lot of work began to introduce the attention mechanism [27] to capture the spatial-spectral correlation [28,29] for hyperspectral image analysis. Ref. [30] combined four 3-D octave convolution blocks and two attention models that were introduced from spatial and spectral dimensions to capture spatial-spectral features from HSIs. This work achieved efficient hyperspectral classification. Ref. [31] proposed an interpretable spatial-spectral reconstruction network, consisting of cross-mode message inserting, spatial reconstruction network, and spectral reconstruction network, to achieve the efficient fusion of hyperspectral and multispectral image. With respect to hyperspectral snapshot compressive reconstruction, Ref. [32] used the self-attention mechanism to process the feature information separately from the channel dimension and the spatial dimension, achieving high-quality reconstruction. Ref. [26] employed non-local spatial attention module to capture the long range dependencies in space. However, the calculating of spatial attention map will consume a lot of computing and memory resources due to huge size of HSIs. Inspired by the non-local network in [33], our work designs a non-local spatial-spectral attention module consisting of the spectral attention path and the spatial attention path. The spectral path calculates the channel attention matrix to capture the global correlations between all of the spectral channels, the spatial path captures the spatial correlation within hyperspectral images. Therefore, the spatial-spectral correlations of hyperspectral images can both be effectively utilized for reconstruction.

3. CASSI Forward Model

Before detailing the proposed reconstruction network, we first briefly introduce the CASSI system. It encodes a 3D spectral scene into a 2D snapshot measurement according to a specific compressive projection manner. Physically, the spectral scene is first collected by the objective lens and spatially encoded by a coded aperture. Subsequently, the encoded scene is dispersed through a disperser, for example, the dispersion degree of each band is linear with its index, and the final snapshot measurement is captured by a 2D detector. Mathematically, the snapshot compressive spectral imaging measurement process can be formulated as:
y = Φ h + ε ,
where h R H W B is the vectorized representation of original HSI x with H, W as the spatial size, and B as the number of spectral channels, Φ R H W × H W B is the forward matrix that describes the CASSI system imaging model, and ε represents the noise corruption that naturally exists in the imaging system. According to the CASSI imaging principle, Φ is actually a combination of diagonal matrices with a special form that can be further expressed as
Φ = [ D ( C 1 ) , D ( C 2 ) , , D ( C B ) ] ,
where coded apertures { C i } i = 1 B R H × W are generated by shifting the mask with a different degree, D ( ) is an operation that represents a diagonal matrix. In particular, the sensing matrix Φ depends on the coded apertures and the measurement y is can be simply computed as:
y = i = 1 B X i C i + ε ,
where ⊙ means the element-wise product and { X i } i = 1 B R H × W are spectral bands of the original HSI x.

4. The Proposed Method

The core problem of hyperspectral snapshot compressive imaging is to reconstruct unknown 3D data from the 2D measurement that is captured by the imaging system. Different from the single-channel panchromatic image, hyperspectral images have many spectral bands, and there are correlations within and between these spectral bands. The correlations within these spectral bands mainly refer to te hspatial correlation, which is, the gray levels of adjacent pixels also have a certain similarity. Regarding the correlations between these spectral bands, not only does this correlation exist between adjacent bands, but there is also a global correlation between spectral bands that are far apart, which is, this correlation is non-local. We design a Non-local Spatial-spectral Residual Network (NSSR-Net) to learn the parameterized reconstruction mapping in order to exploit both the spatial and non-local spectral correlation prior for reconstruction.

4.1. Non Local Spatial-Spectral Residual Reconstruction Network

Figure 1 shows the network architecture of the proposed NSSR-Net. NSSR-Net first employs a 3 × 3 convolution layer to process the input snapshot measurement and generates a feature map with 64 channels. What are subsequently configured are the core components of the network, namely the residual module and the non-local spatial-spectral attention module. The non-local spatial-spectral attention module is set in the middle of the multiple symmetric residual modules. Finally, a 3 × 3 convolutional layer with sigmoid activation function is used to make the output of the network the same channel as the original HSI and normalize the range of each item in the output to [0, 1]. In the following, we elaborate on the details of the two core components of the proposed network.

4.1.1. The Symmetric Residual Module

The Deep Residual Network (ResNet) [34] is a widely-used network architecture. It use the proprietary operation of skip connection to link the input to the output, so that the convolutional block only needs to learn incremental information, which is, the residual between input and output, which can further speed up the network convergence.
Being motivated by [35], we design a symmetric residual learning module with more skips, so that the flow of information between convolutional blocks can be further enhanced. We briefly explain the difference between symmetric residual and classical residual through an illustration. { F i } i = 1 7 denote convolutional blocks. Figure 2a shows the classical residual module, and the output Y is calculated as,
Y = [ F 2 ( F 1 ( x ) ) + x ] + F 4 ( F 3 ( [ F 2 ( F 1 ( x ) ) + x ] ) .
In the symmetric residual module (b), the output is expressed as,
Y ^ = F 6 [ F 2 ( F 1 ( x ) ) + F 5 ( x ) ] + F 4 [ F 3 ( [ F 2 ( F 1 ( x ) ) + F 5 ( x ) ] + F 7 ( F 1 ( x ) ) ) ] .
It can be seen that through further linking, the final output can effectively realize the repeated use of different convolutional layer features, which greatly enhances the performance of feature extraction.

4.1.2. The Non-Local Spatial-Spectral Attention Module

As mentioned above, HSI exhibits coupled spatial-spectral correlation. For sake of capturing this correlation, we propose a non-local spatial-spectral attention module with the spectral attention path and the spatial attention path. The spectral attention calculates the non-local correlation inter spectral channels, as shown in Figure 3. The spatial attention path focuses on the spatial correlation of hyperspectral images. The final output S is the sum of spectral attention attended feature maps S e and spatial attention attended feature maps S a .
We now present the detailed processing of spectral attention path. Let x R H × W × B denote the input of this module. After 1 × 1 convolution operation upon x and dimension reshaping, we can obtain two matrices with sizes ( B , H × W ) and ( H × W , B ) , respectively. Subsequently, a weighted correlation matrix C r R B × B can be calculated after multiplying these two matrices and performing the s o f t m a x operation. C r represents the global correlation between the feature maps of different channels in Equation (6). Different from the non-local processing in spatial-dimension in [33], our non-local processing occurs in the spectral dimension, and the spectral dimension B is usually much smaller than spatial dimension. Therefore, the entire non-local spectral correlation prediction does not take up a lot of calculation and memory. After this operation, we also add weight symmetrization [36] to obtain a symmetric correlation matrix C s . The weight symmetrization can be briefly expressed by a linear operator [36]. Subsequently, the feature map x is subjected to 1 × 1 convolution processing and then multiplication with C s . After the succeeding 1 × 1 convolution and reshaping operation, we can obtain the final output S e of the spectral attention path. At the same time, in the spatial attention path, we use spatial attention to extract the spatial correlation of each feature map. The processing operation of the non-local spatial-spectral attention module can be mathematically formulated as:
C r = s o f t max ( ϕ 1 ( x ) × ϕ 2 ( x ) ) C s = C r + C r T 2 S e = c o n v ( C s × g ( x ) ) S a = x s i g m o i d ( c o n v ( x ) ) S = S e + S a
where ϕ , g indicates the corresponding convolution operation, × is the matrix multiplication, ⊙ is the element-wise multiplication, and C r T represents the transposition operation of the weighted correlation matrix C r . The coupled spatial-spectral correlation can be effectively represented by the incorporation of spectral attention path and spatial attention path. The ablation studies shown in Section 5 verify the effectiveness of the proposed non-local spatial-spectral attention module.

4.1.3. Loss Function

We design a compound loss function consisting of the reconstruction loss, the measurement loss, and the cosine loss to better guide the network learning. The reconstruction loss L r e c o n s t r u c t i o n directly considers the geometric distance between hyperspectral images, and the measurement loss L m e a s u r e m e n t is the L 1 loss between the snapshot measurement y of the original HSI and the snapshot measurement y ^ of the network reconstructed image. The cosine loss L c o s i n e is more helpful in maintaining the characteristics of the spectral signature. It determines the average cosine distance between hyperspectral pixels, treating them as vectors with the same dimension as the number of spectral bands. The mathematical formulation of the cosine loss between two hyperspectral pixels is defined as
l c o s i n e ( i , j ) = 1 cos ( θ i , j ) = 1 b = 1 B x i , j , b x ^ i , j , b b = 1 B x i , j , b 2 b = 1 B x ^ i , j , b 2 ,
where x is the ground truth of HSI, x ^ is the reconstructed HSI, x i , j , b denotes the entry of x at spatial location ( i , j ) and spectral band b, and θ is the spectral angle formed between reference hyperspectral pixel and reconstructed hyperspectral pixel. Figure 4 shows a concise diagram of the spectral angle and geometric distance between pixel 1 and pixel 2. The spectral cosine distance and the geometric distance are measured by cosine loss and L 1 loss, respectively. With the joint constraints of the distance difference and the angle difference, the reconstructed HSI can be as close as possible to the original HSI in each spectral band.
Finally, the overall compound loss function is mathematically defined as:
L t o t a l ( Θ ) = L r e c o n s t r u c t i o n + γ 1 L m e a s u r e m e n t + γ 2 L c o s i n e L r e c o n s t r u c t i o n = x ^ x 1 , L m e a s u r e m e n t = y ^ y 1 , L c o s i n e = 1 H W i = 1 H j = 1 W l c o s i n e ( i , j ) ,
where H, W are the spatial sizes of x, γ 1 and γ 2 are the parameters that tweak the weights of each term.

5. Experiments

In this section, we conduct a series of experiments, including the comparative experiments and ablation experiments, to evaluate the performance of the proposed method. The methods to be compared with our method include several start-of-the-art methods, namely, TwIST [37], GAP-TV [38], DeSCI [18], and λ -net [26]. The first three are Prior-Driven methods, and the last is the Network-Driven method. For a comprehensive evaluation, we perform a series of comparisons on simulation and real data.

5.1. Experimental Setting

All of the experiments are performed on an NVIDIA GTX TITAN X GPU. We employ Pytorch to implement the proposed network. Our network is trained from scratch and initializes all of the convolutional layers using the default setting of the Pytorch. The Adam optimizer [39] is used to minimize the loss function and its hyper parameters are set as learning rate lr = 0.00025, betas = (0.9, 0.999), eps = 10 8 , weight decay = 0. The batch size is 10. All of the competing methods use the code published by their authors.
We used the same data set to train the proposed network, as in [26]. The training data set of [26] contains 150 hyperspectral images with a size of 1392 × 1300 × 31 randomly selected from the ICVL dataset, and then a spectral interpolation is used to transform their channels from the original 31 channels into 24 channels. The wavelength range of these 24 channels is from 400 nm to 700 nm, and the wavelength of each spectral band is: 398.62, 404.40, 410.57, 417.16, 424.19, 431.69, 439.70, 448.25, 457.38, 467.13, 477.54, 488.66, 500.54, 513.24, 526.8, 541.29, 556.78, 573.33, 591.02, 609.93, 630.13, 651.74, 674.83, 699.51 nm. In the process of network training, data cubes with a size of 256 × 256 × 24 are randomly cut out from these hyperspectral data for data augmentation. Following the experimental strategy that was used in [26], the same test set was composed of 10 hyperspectral images is also used in this paper. These 10 test hyperspectral images are also selected from the ICVL dataset and their size is 256 × 256 × 24 .

5.2. Evaluation Metrics

Three quantitative image quality metrics, including PSNR, SSIM, and SAM [40,41], are used to evaluate the performance of various methods. Peak Signal to Noise Ratio (PSNR) and Structural SIMilarity (SSIM) are the first two metrics, which are widely used in the image restoration field. For hyperspectral images, we calculate the spatial fidelity of each 2D spectral band and use the average of all spectral bands as the final output. The higher the values of PSNR and SSIM, the better the performance. The last metric is Spectral Angle Mapper (SAM) [40], which is a specified metric in the hyperspectral image field. It measures the spectral fidelity between the hyperspectral pixels. Smaller SAM values indicate better reconstruction.

5.3. Ablation Studies

We conduct two ablation experiments to investigate the effectiveness of cosine loss and non-local spatial-spectral attention module. First, we test the impact of the non-local spatial-spectral attention module by removing it from the original network and evaluating the performance changes brought about by it. Table 1 shows the results of ablation studies, which are obtained as the average of three runs. Table 1 also reports the standard deviations of three quantitative metrics. It can be seen from Table 1 that the non-local spatial-spectral attention module can improve all the three metrics, which fully affirms its effectiveness. Furthermore, keeping the network architecture unchanged, we test the influence of the cosine loss term on network learning. According to the experimental results shown in Table 1, the cosine loss is conducive to improving SAM metrics of the reconstruction. Overall, the non-local spatial-spectral attention module and the compound loss term can better constrain the network learning and enhance the reconstruction performance, thus demonstrating the rationality of the NSSR-Net design.
We visualize four-channel feature maps before and after the processing of this module in Figure 5 to further analyze the role of the non-local spatial-spectral attention module. It can be seen that the feature maps after the processing of the non-local spatial-spectral attention module can have significantly more informative structures, which demonstrates the advantage of taking the spatial-spectral joint correlation in the reconstruction network.

5.4. Simulation Data Results

In this Simulation case, we use the same coded masks as in [26], which are from the real CASSI system and used to generate snapshot measurements of the test HSIs. Table 2 shows the PSNR, SSIM, and SAM values of five methods on 10 test images. According to the quantitative metrics in Table 2, our method has the best average PSNR, SSIM, and SAM values. Figure 6 and Figure 7 show the plots of PSNR and SSIM values of two scenes as a function of the number of spectral bands. The PSNR and SSIM plots of our method lie basically at the top, and our method does not show a sudden drop in reconstruction quality at certain spectral bands. Figure 8 and Figure 9 provide the snapshot measurements that correspond to the two scenes, as well as the reconstructed spectral signatures in the patches that are indicated by the rectangles. The correlation coefficients that are presented in the legend demonstrate that our method can reconstruct more accurate spectral signatures when compared with the other methods. Figure 10 and Figure 11 visualize the reconstructed spectral bands of two scenes. We can see that our algorithm can reconstruct clear structures and fine details. We also compare the reconstructed spectral signatures of different methods. Our method can better maintain the fidelity of the spectral signatures than the competitive methods.

5.5. Real Data Results

The real data used in our experiments is the Bird data captured by the hyperspectral imaging camera (The Bird data is downloaded from [18]’s Github homepage https://github.com/liuyang12/DeSCI.2019, accessed on 1 February 2020). Because of the complexity of hardware in the real imaging system, the obtained snapshot measurement of Bird hyperspectral data is troubled by noise, which makes reconstruction more difficult. The spatial size of the original real Bird data is 1021 × 703 and contains 24 spectral bands. We cropped a 512 × 512 sub-image for performance evaluation and comparison due to the limitation of computational resource of hardware.
Figure 12 shows four reconstructed spectral bands of Bird data. The reconstruction of GAP-TV still contains a lot of noise when compared with the Ground-truth, as can be seen from Figure 12. Although DeSCI can smooth out the noise, its reconstructed images lack texture details. Regarding the last spectral band (699.51 nm), only λ -net and our method can reconstruct the main structures of this spectral band. λ -net and our method have the similar visual quality. In terms of quantitative metric, our method has the best PSNR and SSIM values. Figure 13 shows the spectral correlation between the reconstruction results of each method and the ground truth. It can be seen that our method has the superiority of maintaining the fidelity of spectral signatures over the other methods.

5.6. Time Complexity Analysis

In addition to the quantitative indicators of reconstruction quality, it is also necessary to analyze the time complexity of the reconstruction methods. Therefore, we also compare the running time (in seconds) that is consumed by each method in reconstructing 256 × 256 × 24 hyperspectral images. TwIST, GAP-TV, and DeSCI run on the CPU, while DeSCI and the proposed method run on the GPU. Table 3 shows the running time results of each algorithm. It can be seen that the reconstruction speed of the Network-Driven method is faster than that of the Prior-Driven methods, because the Network-Driven methods do not require iterative optimization. Because of the two stages of cascaded reconstruction in λ -Net, its reconstruction process consumes more time than our method.

6. Conclusions

In this paper, we propose a novel network for HSI snapshot reconstruction from a single measurement. First, we design the symmetric residual module to integrate the fusion of local features. We propose a non-local spatial-spectral attention module to fully utilize this prior structures to further consider the joint correlation of the spatial and spectral of the HSI. Besides, the compound loss is designed to guide the network focus more on detail reconstruction. The experiment results on both simulation and real data have verified that our method has good performance, while maintaining a fast reconstruction time.

Author Contributions

Conceptualization: Y.X. and Y.S.; Data curation: X.C.; Methodology: Y.Y.; Software: Y.Y. and X.C.; Supervision: Y.X.; Writing—original draft: Y.Y.; Writing—review and editing: Y.X. and Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant U2001211 and 61672292, in part by the Land Observation Satellite Supporting Platform of National Civil Space Infrastructure Project, and in part by the Key University Science Research Project of Jiangsu Province under Grant Number 18KJA520007.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study can be available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chakrabarti, A.; Zickler, T. Statistics of real-world hyperspectral images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011; pp. 193–200. [Google Scholar]
  2. Tu, B.; Zhou, C.; Liao, X.; Zhang, G.; Peng, Y. Spectral-Spatial Hyperspectral Classification via Structural-Kernel Collaborative Representation. IEEE Geosci. Remote Sens. Lett. 2020, 18, 861–865. [Google Scholar] [CrossRef]
  3. Ojha, L.; Wilhelm, M.B.; Murchie, S.L.; McEwen, A.S.; Wray, J.J.; Hanley, J.; Massé, M.; Chojnacki, M. Spectral evidence for hydrated salts in recurring slope lineae on Mars. Nat. Geosci. 2015, 8, 829–832. [Google Scholar]
  4. Cao, X.; Zhou, F.; Xu, L.; Meng, D.; Xu, Z.; Paisley, J. Hyperspectral image classification with markov random fields and a convolutional neural network. IEEE Trans. Image Process. 2018, 27, 2354–2367. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Fauvel, M. Advances in spectral-spatial classification of hyperspectral images. Proc. IEEE 2012, 101, 652–675. [Google Scholar] [CrossRef] [Green Version]
  6. Mouroulis, P.; Green, R.O.; Chrien, T.G. Design of pushbroom imaging spectrometers for optimum recovery of spectroscopic and spatial information. Appl. Opt. 2000, 39, 2210–2220. [Google Scholar] [CrossRef] [Green Version]
  7. Wehr, A.; Lohr, U. Airborne laser scanning-an introduction and overview. ISPRS J. Photogramm. Remote Sens. 1999, 54, 68–82. [Google Scholar] [CrossRef]
  8. Shogenji, R.; Kitamura, Y.; Yamada, K. Multispectral imaging using compact compound optics. Opt. Express 2004, 12, 1643–1655. [Google Scholar] [CrossRef]
  9. Arce, G.; Brady, D.; Carin, L.; Arguello, H.; Kittle, D. Compressive coded aperture spectral imaging: An introduction. IEEE Signal Process. Mag. 2014, 31, 105–115. [Google Scholar] [CrossRef]
  10. Wagadarikar, A.; John, R.; Willett, R.; Brady, D. Single disperser design for coded aperture snapshot spectral imaging. OSA Appl. Opt. 2008, 47, 44–51. [Google Scholar] [CrossRef] [Green Version]
  11. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  12. Wang, Y.; Yang, J.; Yin, W.; Zhang, Y. A new alternating minimization algorithm for total variation image reconstruction. SIAM J. Imaging Sci. 2008, 1, 248–272. [Google Scholar] [CrossRef]
  13. Buades, A.; Coll, B.; Morel, J.-M. A non-local algorithm for image denoising. IEEE Conf. Comput. Vis. Pattern Recognit. 2005, 2, 60–65. [Google Scholar]
  14. Elad, M.; Aharon, M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef]
  15. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  16. Kittle, D.; Choi, K.; Wagadarikar, A.; Brady, D.J. Multiframe image estimation for coded aperture snapshot spectral imagers. Appl. Opt. 2010, 49, 6824–6833. [Google Scholar] [CrossRef]
  17. Wang, L.; Xiong, Z.; Shi, G.; Wu, F.; Zeng, W. Adaptive nonlocal sparse representation for dual-camera compressive hyperspectral imaging. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2104–2111. [Google Scholar] [CrossRef]
  18. Yuan, X. Generalized alternating projection based total variation minimization for compressive sensing. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2990–3006. [Google Scholar]
  19. Blumensath, T.; Davies, M.E. Iterative hard thresholding for compressed sensing. Appl. Comput. Harmon. Anal. 2009, 27, 265–274. [Google Scholar] [CrossRef] [Green Version]
  20. Sun, Y.; Chen, J.; Liu, Q.; Liu, B.; Guo, G. Dual-Path Attention Network for Compressed Sensing Image Reconstruction. IEEE Trans. Image Process. 2020, 29, 9482–9495. [Google Scholar] [CrossRef]
  21. Sun, Y.; Yang, Y.; Liu, Q.; Chen, J.; Yuan, X.-T.; Guo, G. Learning Non-Locally Regularized Compressed Sensing Network With Half-Quadratic Splitting. IEEE Trans. Multimed. 2020, 22, 3236–3248. [Google Scholar] [CrossRef]
  22. Xiong, Z.; Shi, Z.; Li, H.; Wang, L.; Liu, D.; Wu, F. Hscnn: Cnn-based hyperspectral image recovery from spectrally undersampled projections. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 518–525. [Google Scholar]
  23. Choi, I.; Jeon, D.S.; Nam, G.; Gutierrez, D.; Kim, M.H. High-quality hyperspectral reconstruction using a spectral prior. ACM Trans. Graph. 2017, 36, 218. [Google Scholar] [CrossRef] [Green Version]
  24. Wang, L.; Sun, C.; Zhang, M.; Fu, Y.; Huang, H. DNU: Deep Non-Local Unrolling for Computational Spectral Imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1658–1668. [Google Scholar]
  25. Zheng, S.; Liu, Y.; Meng, Z.; Qiao, M.; Tong, Z.; Yang, X.; Han, S.; Yuan, X. Deep plug-and-play priors for spectral snapshot compressive imaging. Photonics Res. 2021, 9, B18–B29. [Google Scholar] [CrossRef]
  26. Miao, X.; Yuan, X.; Pu, Y. lambda-net: Reconstruct Hyperspectral Images from a Snapshot Measurement. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 4058–4068. [Google Scholar]
  27. Vaswani, A.; Shazeer, N.; Parmar, N. FAttention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  28. Sun, H.; Zheng, X.; Lu, X.; Wu, S. Spectral–Spatial Attention Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3232–3245. [Google Scholar] [CrossRef]
  29. Sellami, A.; Abbes, A.B.; Barra, V. Fused 3-D spectral-spatial deep neural networks and spectral clustering for hyperspectral image classification. Pattern Recognit. Lett. 2020, 138, 594–600. [Google Scholar] [CrossRef]
  30. Sellami, A.; Abbes, A.B.; Barra, V. Hyperspectral Image Classification Based on 3-D Octave Convolution With Spatial–Spectral Attention Network. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2430–2447. [Google Scholar]
  31. Zhang, X.; Huang, W.; Wang, Q.; Li, X. SSR-NET: Spatial-Spectral Reconstruction Network for Hyperspectral and Multispectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2020, 1–13. [Google Scholar] [CrossRef]
  32. Meng, Z.; Ma, J.; Yuan, X. End-to-end low cost compressive spectral imaging with spatial-spectral self-attention. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 187–204. [Google Scholar]
  33. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  34. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  35. Liu, X.; Suganuma, M.; Sun, Z.; Okatani, T. Dual Residual Networks Leveraging the Potential of Paired Operations for Image Restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7000–7009. [Google Scholar]
  36. Hu, X.S.; Zagoruyko, S.; Komodakis, N. Exploring Weight Symmetry in Deep Neural Networks. arXiv 2018, arXiv:1812.11027. [Google Scholar] [CrossRef] [Green Version]
  37. Bioucas-Dias, J.M.; Figueiredo, M.A. A new twist: Two-step iterative shrinkage/thresholding algorithms for image restoration. IEEE Trans. Image Process. 2007, 16, 2992–3004. [Google Scholar] [CrossRef] [Green Version]
  38. Liu, Y.; Yuan, X.; Suo, J.; Brady, D.; Dai, Q. Rank minimization for snapshot compressive imaging. IEEE Int. Conf. Image Process. 2016, 16, 2539–2543. [Google Scholar] [CrossRef] [Green Version]
  39. Diederik, P.K.; Jimmy, B. Adam: A method for stochastic optimization. In Proceedings of the 33rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
  40. Liu, X.; Yang, C. A Kernel Spectral Angle Mapper algorithm for remote sensing image classification. In Proceedings of the International Congress on Image and Signal Processing (CISP), Beijing, China, 14–16 July 2013; pp. 814–818. [Google Scholar]
  41. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The overall framework of the proposed method. It consists of the training stage and test stage. The reconstruction of test image can be obtained by just feeding the CASSI measurement into the well-trained network.
Figure 1. The overall framework of the proposed method. It consists of the training stage and test stage. The reconstruction of test image can be obtained by just feeding the CASSI measurement into the well-trained network.
Remotesensing 13 01812 g001
Figure 2. A simple explanation of the symmetric residual module. (a) is the original residual block, (b) is a symmetric residual module with seven convolutional blocks.
Figure 2. A simple explanation of the symmetric residual module. (a) is the original residual block, (b) is a symmetric residual module with seven convolutional blocks.
Remotesensing 13 01812 g002
Figure 3. The Non-local Spatial-spectral Attention Module, consisting of the spectral attention and spatial attention.
Figure 3. The Non-local Spatial-spectral Attention Module, consisting of the spectral attention and spatial attention.
Remotesensing 13 01812 g003
Figure 4. A diagram of the spectral angle and geometric distance between pixel 1 and pixel 2 of a HSI with three bands.
Figure 4. A diagram of the spectral angle and geometric distance between pixel 1 and pixel 2 of a HSI with three bands.
Remotesensing 13 01812 g004
Figure 5. Visualization of feature maps. The upper column is feature maps before the non-local spatial-spectral attention module, and the lower is feature maps after processing by this module.
Figure 5. Visualization of feature maps. The upper column is feature maps before the non-local spatial-spectral attention module, and the lower is feature maps after processing by this module.
Remotesensing 13 01812 g005
Figure 6. The plot of PSNR and SSIM values of scene 10 as a function of the number of spectral bands.
Figure 6. The plot of PSNR and SSIM values of scene 10 as a function of the number of spectral bands.
Remotesensing 13 01812 g006
Figure 7. The plot of PSNR and SSIM values of scene 6 as a function of the number of spectral bands.
Figure 7. The plot of PSNR and SSIM values of scene 6 as a function of the number of spectral bands.
Remotesensing 13 01812 g007
Figure 8. The snapshot measurement of scene 10 and the reconstructed spectral signatures. The first row shows one spectral bands of the scene 10 and its snapshot measurement. The second row show the reconstructed spectral signatures of the two patches indicated by the rectangles. The correlation coefficients are also shown in the legend to quantitatively compare the accuracy of the spectral signatures that were reconstructed by the five methods.
Figure 8. The snapshot measurement of scene 10 and the reconstructed spectral signatures. The first row shows one spectral bands of the scene 10 and its snapshot measurement. The second row show the reconstructed spectral signatures of the two patches indicated by the rectangles. The correlation coefficients are also shown in the legend to quantitatively compare the accuracy of the spectral signatures that were reconstructed by the five methods.
Remotesensing 13 01812 g008
Figure 9. The snapshot measurement of scene 6 and the reconstructed spectral signatures. The first row shows one spectral bands of the scene 6 and its snapshot measurement. The second row show the reconstructed spectral signatures of the two patches indicated by the rectangles. The correlation coefficients are also shown in the legend to quantitatively compare the accuracy of the spectral signatures reconstructed by the five methods.
Figure 9. The snapshot measurement of scene 6 and the reconstructed spectral signatures. The first row shows one spectral bands of the scene 6 and its snapshot measurement. The second row show the reconstructed spectral signatures of the two patches indicated by the rectangles. The correlation coefficients are also shown in the legend to quantitatively compare the accuracy of the spectral signatures reconstructed by the five methods.
Remotesensing 13 01812 g009
Figure 10. The visualization of four spectral bands of scene 10 (wavelength: 477.54 nm, 526.8 nm, 630.13 nm and 699.51 nm) reconstructed by five methods. From the left column to the rightmost column correspond to Ground truth, TwIST (PSNR 18.921/SSIM 0.747), GAP-TV (21.502/0.796), DeSCI (23.331/0.834), λ -net (28.204/0.924) and ours (28.924/0.937).
Figure 10. The visualization of four spectral bands of scene 10 (wavelength: 477.54 nm, 526.8 nm, 630.13 nm and 699.51 nm) reconstructed by five methods. From the left column to the rightmost column correspond to Ground truth, TwIST (PSNR 18.921/SSIM 0.747), GAP-TV (21.502/0.796), DeSCI (23.331/0.834), λ -net (28.204/0.924) and ours (28.924/0.937).
Remotesensing 13 01812 g010
Figure 11. The visualization of four spectral bands of scene 6 (wavelength: 477.54 nm, 526.8 nm, 630.13 nm and 699.51 nm) reconstructed by five methods. From the left column to the rightmost column correspond to Ground truth, TwIST (PSNR 20.579/SSIM 0.744), GAP-TV (23.090/0.802), DeSCI(25.814/0.906), λ -net (28.573/0.902), and ours (30.002/0.944).
Figure 11. The visualization of four spectral bands of scene 6 (wavelength: 477.54 nm, 526.8 nm, 630.13 nm and 699.51 nm) reconstructed by five methods. From the left column to the rightmost column correspond to Ground truth, TwIST (PSNR 20.579/SSIM 0.744), GAP-TV (23.090/0.802), DeSCI(25.814/0.906), λ -net (28.573/0.902), and ours (30.002/0.944).
Remotesensing 13 01812 g011
Figure 12. Real data results: Four reconstructed spectral bands (wavelengths from top to bottom are 477.54 nm, 541.29 nm, 591.02 nm, and 699.51 nm). On the far left is the ground truth of Bird real data, and to the right are the reconstruction results of GAP-TV (22.540/0.754), DeSCI (25.036/0.777), λ -net (24.882/0.832), and ours (25.389/0.839). The numbers in the brackets are the PSNR and SSIM values corresponding to each method.
Figure 12. Real data results: Four reconstructed spectral bands (wavelengths from top to bottom are 477.54 nm, 541.29 nm, 591.02 nm, and 699.51 nm). On the far left is the ground truth of Bird real data, and to the right are the reconstruction results of GAP-TV (22.540/0.754), DeSCI (25.036/0.777), λ -net (24.882/0.832), and ours (25.389/0.839). The numbers in the brackets are the PSNR and SSIM values corresponding to each method.
Remotesensing 13 01812 g012
Figure 13. Real data results: the reconstructed spectral signatures of the Bird hyperspectral data captured by the real CASSI system. The correlation coefficients of the reconstructed spectral signatures and the ground-truth are shown in the legends.
Figure 13. Real data results: the reconstructed spectral signatures of the Bird hyperspectral data captured by the real CASSI system. The correlation coefficients of the reconstructed spectral signatures and the ground-truth are shown in the legends.
Remotesensing 13 01812 g013
Table 1. Ablation study for the non-local spatial-spectral attention module and the cosine loss. The numbers after ± denote standard deviations.
Table 1. Ablation study for the non-local spatial-spectral attention module and the cosine loss. The numbers after ± denote standard deviations.
ConfigurationPSNR ↑SSIM ↑SAM↓
without Non-local Module34.019 (±0.141)0.967 (±0.0020)0.091 (±0.0005)
without Cosine Loss34.303 ( ± 0.138 )0.970 ( ± 0.0005 )0.092 ( ± 0.0005 )
NSSR-Net34.764 ( ± 0.114 )0.972 ( ± 0.0009 )0.089 ( ± 0.0005 )
Table 2. Quantitative result comparison on 10 test images from the ICVL dataset.
Table 2. Quantitative result comparison on 10 test images from the ICVL dataset.
MethodsMetricsScene 1Scene 2Scene 3Scene 4Scene 5Scene 6Scene 7Scene 8Scene 9Scene 10Average
TwISTPSNR25.62118.41321.75021.24023.78420.57924.23220.20227.01418.92122.140
SSIM0.8560.8260.8260.8280.7990.7440.8700.7840.8880.7470.817
SAM0.1600.1750.1920.1970.2350.3350.1730.1990.2260.2410.213
GAP-TVPSNR30.66622.41023.49922.27326.98523.09024.85922.91329.10521.50224.730
SSIM0.8920.8690.8630.8290.7920.8020.8770.8410.9120.7960.847
SAM0.2070.1850.3020.1840.3300.3340.1660.1860.2250.2130.233
DeSCIPSNR31.14726.44324.74129.25129.37225.81428.40124.42434.41123.33127.732
SSIM0.9370.9470.8980.9490.9070.9060.9210.8720.9710.8340.914
SAM0.1680.0810.2630.1050.2290.2620.1480.1860.1750.1900.181
λ -NetPSNR36.10932.05433.34129.59835.40328.57335.21932.35533.41828.20432.427
SSIM0.9490.9750.9740.9370.9420.9020.9690.9510.9160.9240.944
SAM0.0980.0430.0910.1000.1290.2050.0710.0990.2000.1130.115
NSSR-Net (ours)PSNR40.04436.55734.63229.22538.47430.00238.29833.88637.21728.92434.726
SSIM0.9880.9890.9720.9530.9840.9440.9850.9660.9870.9370.971
SAM0.0550.0270.0730.1140.0810.1980.0460.0750.1230.1070.090
Table 3. The average run time for the reconstruction of a 256 × 256 × 24 hyperspectral image.
Table 3. The average run time for the reconstruction of a 256 × 256 × 24 hyperspectral image.
MethodsTwISTGAP-TVDeSCI λ -NetNSSR-Net
Times (s)70.9625.83594.54.281.19
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, Y.; Xie, Y.; Chen, X.; Sun, Y. Hyperspectral Snapshot Compressive Imaging with Non-Local Spatial-Spectral Residual Network. Remote Sens. 2021, 13, 1812. https://doi.org/10.3390/rs13091812

AMA Style

Yang Y, Xie Y, Chen X, Sun Y. Hyperspectral Snapshot Compressive Imaging with Non-Local Spatial-Spectral Residual Network. Remote Sensing. 2021; 13(9):1812. https://doi.org/10.3390/rs13091812

Chicago/Turabian Style

Yang, Ying, Yong Xie, Xunhao Chen, and Yubao Sun. 2021. "Hyperspectral Snapshot Compressive Imaging with Non-Local Spatial-Spectral Residual Network" Remote Sensing 13, no. 9: 1812. https://doi.org/10.3390/rs13091812

APA Style

Yang, Y., Xie, Y., Chen, X., & Sun, Y. (2021). Hyperspectral Snapshot Compressive Imaging with Non-Local Spatial-Spectral Residual Network. Remote Sensing, 13(9), 1812. https://doi.org/10.3390/rs13091812

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop