Next Article in Journal
GPR Data Processing and Interpretation Based on Artificial Intelligence Approaches: Future Perspectives for Archaeological Prospection
Next Article in Special Issue
Reconstruction of Sentinel-2 Image Time Series Using Google Earth Engine
Previous Article in Journal
Validating Salinity from SMAP and HYCOM Data with Saildrone Data during EUREC4A-OA/ATOMIC
Previous Article in Special Issue
Self-Supervised Denoising for Real Satellite Hyperspectral Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Column-Spatial Correction Network for Remote Sensing Image Destriping

1
Key Laboratory of Specialty Fiber Optics and Optical Access Networks, Joint International Research Laboratory of Specialty Fiber Optics and Advanced Communication, Shanghai Institute of Advanced Communication and Data Science, Shanghai University, Shanghai 200444, China
2
Computer Science Department, Aberystwyth University, Aberystwyth SY23 3DB, UK
3
JD AI Research, Beijing 100176, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2022, 14(14), 3376; https://doi.org/10.3390/rs14143376
Submission received: 24 May 2022 / Revised: 27 June 2022 / Accepted: 8 July 2022 / Published: 13 July 2022
(This article belongs to the Special Issue Remote Sensing Image Denoising, Restoration and Reconstruction)

Abstract

:
The stripe noise in the multispectral remote sensing images, possibly resulting from the instrument instability, slit contamination, and light interference, significantly degrades the imaging quality and impairs high-level visual tasks. The local consistency of homogeneous region in striped images is damaged because of the different gains and offsets of adjacent sensors regarding the same ground object, which leads to the structural characteristics of stripe noise. This can be characterized by the increased differences between columns in the remote sensing image. Therefore, the destriping can be viewed as a process of improving the local consistency of homogeneous region and the global uniformity of whole image. In recent years, convolutional neural network (CNN)-based models have been introduced to destriping tasks, and have achieved advanced results, relying on their powerful representation ability. Therefore, to effectively leverage both CNNs and the structural characteristics of stripe noise, we propose a multi-scaled column-spatial correction network (CSCNet) for remote sensing image destriping, in which the local structural characteristic of stripe noise and the global contextual information of the image are both explored at multiple feature scales. More specifically, the column-based correction module (CCM) and spatial-based correction module (SCM) were designed to improve the local consistency and global uniformity from the perspectives of column correction and full image correction, respectively. Moreover, a feature fusion module based on the channel attention mechanism was created to obtain discriminative features derived from different modules and scales. We compared the proposed model against both traditional and deep learning methods on simulated and real remote sensing images. The promising results indicate that CSCNet effectively removes image stripes and outperforms state-of-the-art methods in terms of qualitative and quantitative assessments.

1. Introduction

Remote sensing technology aims to achieve the long-distance information detection in a non-contact way. The multispectral remote sensing images of different ground objects are received by sensors and further processed and analyzed to realize the detection and identification of earth resources and geographical environments. However, there often exist various kinds of stripe noise in different remote sensing imaging systems given their unique image acquisition modes. The stripe noise not only reduces the visibility of images [1,2], but also adversely affects their interpretations and applications, such as classification [3,4], endmember extraction [5,6,7], unmixing [8,9], segmentation [10], target detection [11,12], etc. Therefore, the destriping is a necessary step of remote sensing image processing.
In the remote sensing image, the digital numbers (DNs) of pixels on the same column often derive from the same detector pixel. Therefore, the DNs of homogeneous region generally come from several adjacent detectors, which tend to maintain the local consistency of these regions. However, different detector pixels sometimes output different DNs for the same object irradiance due to various factors, such as optical system, detector pixel response, readout circuit, and electronic systems, making images contain the vertically distributed stripe noises. Therefore, the destriping, which aims to reduce the differences between adjacent columns caused by the stripe noise, plays a crucial role in improving the uniformity of the whole image.
In the past few decades, various methods dedicated to remote sensing image destriping have been proposed to mitigate the impact of stripe noise on image quality. The existing methods can be grouped into four categories according to the principle of stripe removal, namely, statistical-based, filtering-based, optimization-based, and deep learning-based methods, each of which possesses unique advantages and limitations. The statistical-based methods assume that the signal acquired by each pixel in the detector shares the same statistical distribution, such as histogram matching [13,14] and moment matching (MM) [15,16]. The statistical information of each column is matched to the same reference, e.g., the information of the whole image. According to the noise characteristics in the spatial and transformed domains, filtering-based methods remove the stripe by designing the appropriate filter [17,18]. Differently from them, optimization-based methods estimate the clean image from its striped counterpart by optimizing the designed object function, which usually contains the fidelity and constraint terms. The fidelity term is adopted to retain the original image structure. The constraint term is formulated from the smoothness of clean image and the characteristics of image spectrum or noise components, such as the total variation (TV) [19,20], low-rank property [21,22], or sparsity [23,24]. For example, Chang et al. [19] introduced an anisotropic spectral-spatial total variation (ASSTV) model for multispectral remote sensing image destriping. Subsequently, a low-rank-based single-image decomposition model (LRSID) was proposed to separate the original image from the stripe component [25].
Among above traditional destriping methods, it is proven that effectively exploring the structural characteristics of the stripe play a crucial role in the process of destriping. For example, statistical-based methods assume the image scene is evenly distributed, and baseline objects within each column of pixels in the scanning direction have the similar radiation distribution, among which, the histogram matching [13,14] and moment matching [15,16] are representative methods. They remove the stripe noise by adjusting the statistical information of each column, including the histogram or mean and standard deviation. Alternatively, the TV model [19] is usually adopted in the optimization-based destriping methods. The conventional TV model enhances the smoothness of images from both horizontal and vertical directions for the general denoising task. Given that the stripe noise does not damage the vertical structure information of an image, optimization-based methods consider limiting the smoothness along the horizontal direction to adjust the TV model to be more suitable for destriping.
Recently, deep learning-based methods rely on the advanced capability of convolutional neural networks (CNNs) to remove the various noises of images [26,27]. Zhang et al. [28] presented the spatial-spectral gradient network (SSGN) for removing the mixed noise using the gradient feature in hyperspectral images. Xiao et al. [29] developed the infrared cloud image stripe removal network (ICSRN) with a local-global combination structure. These methods mainly utilize the global contextual information to improve the uniformity of the whole image, yet the significance of the structural characteristics is often neglected [29,30]. To effectively leverage both CNNs and the structural characteristics of stripe noise, we propose a multi-scaled column-spatial correction network (CSCNet) for remote sensing image destriping, in which the local structural characteristic of stripe noise and the global contextual information of image are both explored at multiple feature scales. More specifically, we designed a column-based correction module (CCM) to capture the local feature, which enables the network to pay more attention to eliminating the differences between columns caused by the stripe noise. For the global feature, the spatial attention mechanism is adopted in the spatial-based correction module (SCM) to further improve the image’s uniformity. In addition, we utilize a feature fusion module (FFM) based on the channel attention mechanism to fuse features obtained from different correction modules and scales. Overall, our main contributions are summarized as follows:
(1)
Based on the structural characteristics of stripe, we propose a multi-scaled column-spatial correction network (CSCNet), aiming at improving the local consistency of homogeneous region and the global uniformity of whole image. The proposed CSCNet can effectively remove different kinds of stripe noise, including non-periodic, periodic, and wide stripe.
(2)
A column-based correction module is proposed to reduce the differences between columns caused by stripe noise. To the best of our knowledge, this was one of the first attempts to explore the column-based correction strategy in deep neural network-based models for destriping according to the structural characteristics of the stripe.
(3)
The proposed method has been evaluated on both simulated and real remote sensing images with promising results. Compared to existing methods, our CSCNet has achieved superior qualitative and quantitative assessments.
The remainder of this paper is organized as follows. In Section 2, existing methods for remote sensing image destriping are introduced. The proposed CSCNet and the related details are described in Section 3. The simulated and experimental results with real data are presented in Section 4. Finally, our conclusions are summarized in Section 5.

2. Related Work

In recent decades, various methods dedicated to remote sensing image destriping have been proposed. The existing methods can be coarsely grouped into four categories, i.e., statistical-based, filtering-based, optimization-based, and deep learning-based methods.

2.1. Statistical-Based Methods

Statistical-based methods focus on the statistical properties of DNs for each detector and assume a linear relationship between the response values of detector pixels and the input radiance. They assume the scene is evenly distributed in the remote sensing image, and that each column of data in the scanning direction has a similar radiation distribution. The most common methods, including histogram matching [13,14] and moment matching (MM) [15,16], mainly consist of two steps: allocating the reference and statistical matching. The histogram matching adopts the histogram as a clean reference and adjusts the histogram of each detector by referring to the reference. Similarly, MM assumes that the mean and standard deviation of all detectors are approximately equal. These statistics of each detector are rectified to those reference values based on the entire image for stripe removal. The statistics-based methods manage to obtain satisfactory destriping results when scenes are simple and homogeneous. However, the distribution of ground objects is rather complex in practice, and the statistical information originated from each detector often dramatically changes. Therefore, the assumption of distribution similarity cannot always holds, resulting in poor destriping results in those complex scenes.

2.2. Filtering-Based Methods

Filtering-based methods aim at processing the stripe in the spatial and transformed domains. Among early works, Crippe et al. [31] proposed a simple spatial filtering routine to remove the stripe. Fourier transform [32] and Wavelet decomposition [33] are the typical domain transform filters. Following the assumption of stripe noise with strong periodicity characteristics, Fourier transform-based methods aim at suppressing the specific frequencies caused by stripe to separate useful signals from a striped image by designing an appropriate filter in the transformed domain. Nevertheless, since signals and the noise are mixed together, useful signals are often compromised during the stripe removal. In particular, when the input radiance changes abruptly, ringing artifacts appear in the destriping results [34]. To sum up, Fourier transform methods are efficient to implement, yet the assumption for the periodicity of stripe limits the effectiveness of stripe removal. Moreover, wavelet decomposition is also applied to remove the stripe noise under the consideration of scaling directional properties [35]. It is worth noting that the selection of wavelet transform function plays a crucial role in the destriping process.

2.3. Optimization-Based Methods

Optimization-based methods treat the destriping as an ill-posed inverse problem and mainly include two types of methods, i.e., variation-based and low-rank-based methods. The former one obtains the desired destriping results by optimizing the variation model with priors. A destrping framework based on maximum a posterior (MAP) was firstly presented by Shen et al. [36], in which the Huber–Markov-based variation model is used as the prior likelihood probability density function. They treat the stripe as the isotropic noise, which lacks the consideration of the anisotropic property of the stripe. Consequently, unidirectional variation models were proposed which focus on the directional characteristic of the stripe. They constrain the image smoothness in the cross-stripe direction and retain the information in the along-stripe direction [19,20,23]. The aforementioned models are dedicated to estimate the image prior. On the contrary, since the stripe noise possesses significant directional features compared to the clean component, alternative methods choose to focus on utilizing the stripe prior [17,20]. The low-rank-based methods consider that the data, such as images, abundance matrices, and stripe noises, possess the low-rank characteristic; thus, they adopt the low-rank restriction during the estimation of desired results. According to the form of data processing, they can be divided into matrix-based [1,25] and tensor-based [37,38] methods. The matrix-based methods take the advantage of spectral coherence by lexicographically ordering the 3D cubic into a 2D matrix [39]. Though promising results have been demonstrated, matricization techniques preliminarily vectorize all image bands, which inevitably causes the loss of the spectral–spatial structural correlation of the image cube. Therefore, the subsequent tensor-based methods were proposed to make up for this drawback. However, the tensor-based methods are relatively inefficient given the large data size and the high computation complexity.

2.4. Deep Learning-Based Methods

Various assumptions or handcrafted priors have been designed in previous methods for the image and stripe components to promote the improvement of destriping. However, these assumptions and priors are often inconsistent with realistic scenarios, leading to the weak generalization ability of such methods. Most recently, deep convolutional neural network-based methods have been proposed to remove the stripe noise for remote sensing images [29,30,39]. The plain neural network-based methods utilize the strong learning capacity of CNN to output a clean image from a striped one [29,30]. Subsequently, residual learning is introduced into networks to obtain the stripe noise [39,40]. Moreover, the additional information, such as the horizontal and vertical gradient features of the image, are used to assist the model in learning discriminative features [28]. Zhang et al. [28] presented the spatial-spectral gradient network (SSGN) to remove mixed noises, including Gaussian and stripe noise. A two-stream wavelet enhanced U-net (TSWEU) model was presented by Chang et al. [39] to learn the relationship between the clean image and stripe noise, and the wavelet is adopted to extract the multiscale information of global contextual feature with a larger receptive field. Compared to traditional methods, deep learning-based methods have achieved a superior destriping effect, relying on the strong ability of feature learning; however, these methods neglect the special structural characteristic of the stripe.

3. Methodology

3.1. Overall Framework

We propose a column-spatial correction network (CSCNet) for remote sensing image destriping, which leverages the global image uniformity and the local difference between columns. The flowchart of the proposed method is shown in Figure 1. CSCNet learns an end-to-end mapping between the striped image and its clean counterpart. Considering that the striped remote sensing image retains the structural information in the vertical direction, both the striped image and its vertical gradient are used as inputs in CSCNet. First of all, we feed the original striped image and its vertical gradient to a 3 × 3 convolutional layer. Then, the extracted feature passes through the middle part of the network composed of the multiple multi-scaled column-spatial correction module (MCSCM) based on a residual design. Finally, a 3 × 3 convolutional layer followed by a residual connection to the original striped image are used to generate the destriping result. Except in downsampling and upsampling, the padding operation is utilized to keep the spatial sizes of features in all correction modules the same. The Charbonnier loss is adopted to optimize the proposed network [41]:
L ( I ^ , I * ) = I ^     I * 2 + ε 2 ,
where Î and I* d-truth (clean image), respectively. ε is a constant.

3.2. Multi-Scaled Column-Spatial Correction Module

As the basic unit of the network, the structure of MCSCM is shown in Figure 1, which is a multi-scaled residual structure and is mainly composed of a column-spatial correction module (CSCM) and a feature fusion module (FFM). CSCM is responsible for enhancing uniformity from the global and local perspectives, and FFM is utilized to fuse features derived from different correction modules. Firstly, we generate multi-scaled features by downsampling. These features are then fed to CSCM for the correction at different scale branches; subsequently, the generated smaller scale features are upsampled and gradually fused with larger scale features by FFM. Finally, the fused feature passes through a 3 × 3 convolutional layer followed by a residual connection to the input feature as the final prediction. To improve the readability, we give the details of the CSCM and FFM first, and then introduce their multi-scale extensions.

3.3. Column-Spatial Correction Module

The structure of the CSCM is shown in Figure 2. Since the destriping can be viewed as reducing the differences between columns caused by stripe noise, we intuitively designed two correction modules from the perspectives of reducing local differences and improving global uniformity, denoted column-based and spatial-based correction modules (CCM and SCM), respectively. The two modules generate correction coefficient maps to correct input features, and the corrected features are further fused as the results of CSCM.

3.3.1. Column-Based Correction Module

The column-based correction module (CCM) is designed to reduce the local differences based on the column; i.e., features in the same convolutional column should be allocated with similar correction coefficients. As illustrated in the upper part of Figure 2, the CCM first performs the column average pooling on each channel of input feature M R H × W × C , which calculates the average values column-wise. The yielded feature d 1 R 1 × W × C is then copied along the column direction H times as d 2 R H × W × C with the same size as the input feature. The copied feature is fed to two 1 × 1 convolutional layers sequentially, which are followed by relu and sigmoid activation, respectively, to form the column correction maps d ^ R H × W × C . We reduce the local differences by employing an element-wise product between the correction coefficient maps d ^ and the input feature M .

3.3.2. Spatial-Based Correction Module

Differently from CCM, the spatial-based correction module (SCM) focuses on utilizing the global feature to improve the image uniformity. As illustrated in the lower part of Figure 2, the input feature M R H × W × C first passes through the global average pooling (GAP) layer along the channel dimension. Then we generate the spatial correction map f ^ R H × W × 1 by feeding the pooled f R H × W × 1 to one convolutional layer followed by the sigmoid gating. To extract the spatial information in a larger range, the 5 × 5 convolutional kernel is used to increase the receptive field in SCM. Similarly to CCM, the output feature with the high uniformity is obtained by rescaling M with the coefficient map f ^ .

3.4. Feature Fusion Module

Based on the proposed network structure, features derived from different modules and scales need to be aggregated for generating discriminative representations. The details of multi-scale extension will be elaborated in the next section. Therefore, we carry out a fusion process based on the channel attention mechanism to enhance the feature selectively and remove the possible redundancy, denoted the feature fusion module (FFM).
As shown in Figure 3, the FFM receives input features F 1 R H × W × C and F 2 R H × W × C from different scales or correction modules, i.e., SCMs and CCMs. These features are firstly combined by the element-wise sum as: F = F 1 + F 2 . We apply the global average pooling on the spatial dimension of F R H × W × C to compute the channel-wise statistics w R 1 × 1 × C . The vector w is then passed through a 1 × 1 convolutional layer and a sigmoid activation layer to generate fusion weights w ^ R 1 × 1 × C . FFM strengthens the important features and suppresses the less significant ones based on learned weights. Therefore, we conduct a soft selection to fuse input features F 1 and F 2 , which are allocated with the weights w ^ and (1 w ^ ), respectively. Finally, the sum of two weighted features formulates the output of FFM. The feature fusion procedure can be summarized as:
U = w ^ F 1 + ( 1     w ^ ) F 2 ,
where U is the fusion result and ⊗ represents the element-wise product.

3.5. Multi-Scale Extension

To collect multi-scaled spatial information, as illustrated in Figure 1, the proposed column-spatial correction modules (CSCMs) are constructed at three different scale branches; the convolution with a padding operation in the correction modules projects features with the same spatial size in different branches. First of all, the input is downsampled to three scales, including original, one-half, and one-quarter size, which are denoted B 1 , B 2 , and B 3 , respectively. We then apply the CSCM twice in each branch. Moreover, to enhance complementary advantages between high and low resolutions, the information is exchanged across parallel streams after the first CSCM. More specifically, by downsampling or upsampling, the outputs of the first CSCM (denoted f 11 , f 21 , and f 31 , respectively) are resized to the same scale as other two branches. We then add features derived from three branches and feed the result to the second CSCM in each branch. For example, the input of second CSCM is B 2 , which is the sum of the downsampling of f 11 and f 12 , and the upsampling of f 13 . Finally, corrected features derived from three branches (denoted f 12 , f 22 , and f 32 , respectively) are fused using the proposed fusion module FFM. We fuse two low-resolution representations f 22 and f 32 first; then, the fused feature and high-resolution feature f 12 are passed through FFM to obtain the final output with the same size as the input.
To align the spatial sizes of features from different scales, we employ the anti-aliasing downsampling [42] and a 1 × 1 convolutional layer to downsample the features, where the anti-aliasing downsampling can improve the shift-equivariance of proposed model. Moreover, the bilinear interpolation, followed by a 1 × 1 convolutional layer, is used to upsample the features. The 1 × 1 convolutional layers are utilized to adjust the number of feature channels. The size of each feature is reduced by half for each downsampling, and the number of channels is doubled; i.e., we set the channel number of original scale branches to 64, and the other two branches are 128 and 256, respectively. The change in feature size during upsampling is the opposite to that in downsampling.

3.6. Training Details

The number of multi-scaled column-spatial correction modules in the proposed model is four. The relative analysis and discussion are given in Section 4.4. To speed up the training, we cropped image patches of the size 80 × 80 from the University of Houston dataset. These training samples were then expanded to 10,000 through the rotation. The training data were normalized to [0, 1]; the learning rate and epoch were set to 0.0001 and 200, respectively. The learning rate was reduced by half every 50 epochs. We set the batch size to 120, and the ADAM optimizer was adopted to minimize the total loss.

4. Experimental Results and Analysis

In this section, we compare our proposed CSCNet against eight classic destriping methods, namely, MM [15], ASSTV [19], LRSID [25], that of reference [43], PADMM [44], ICSRN [29], SSGN [28], and TSWEU [39] on multiple datasets. Both simulated and real images with the size of 256 × 256 are tested, and we present the visualization analysis of the destriping effect of all methods. Moreover, we also give the quantitative analysis of the model’s performance. The quantitative evaluation includes the peak signal-to-noise ratio (PSNR), structural similarity (SSIM), mean relative deviation (MRD), and non-uniformity of images.

4.1. Simulated Image Destriping

4.1.1. Simulated Data Preparation

We collected four remote sensing images as simulated datasets for fair comparisons, including Washington DC Mall, HYDICE Urban, Pavia University, and Salinas, which are denoted DC, Urban, PaviaU, and Salinas in the following section. By referring to [39], two ways of adding stripes were investigated, namely, adding noise to the entire image and a part of the image. The former consists of non-periodicity and periodicity stripes, whereas the latter adds stripes to certain rows and columns, respectively. The additive noise was utilized to generate the periodical stripe, and the others were mixed noise. The additive and mixed noise can be, respectively, formulated as: V l , j ( φ ) = v l , j ( φ ) + C l , j and V l , j ( φ ) = A l , j × v l , j ( φ ) + C l , j , where v l , j ( φ ) stands for the pixel located in the l-th row and j-th column. To simulate the vertical characteristics of stripe noise, A l , j and C l , j are the multiplicative and additive noise values and A l , j and C l , j are random values, but the same columns share the same A l , j and C l , j . The University of Houston dataset was employed to generate the simulated data for training. For fair comparisons, by referring to their original implementations, data-driven models were retrained with the same set of training data and followed the same testing protocol as the proposed model. The details of simulated data are described as follows.
Entire Image: In order to generate test images with the non-periodic stripe, we added the multiplicative and additive noise to each column of the original DC image. For the periodic stripe, we added noise to columns with a certain number of intervals on the Urban image. The non-periodically and periodically degraded images are shown in Figure 4b and Figure 5b, respectively. Figure 4c–j and Figure 5c–j show the correction results of all compared algorithms.
A Part of Image: In practice, the stripe noise often exists in some rows and columns of the image due to the instability of the remote sensing imaging system. In order to simulate the stripe noise more realistically, we randomly selected a number of columns and rows to add stripes to the PaviaU and Salinas images, respectively, where the selected rows are successive. Two degraded images are shown in Figure 6b and Figure 7b, respectively. We provide the corrected images of eight methods in Figure 6c–j and Figure 7c–j.

4.1.2. Evaluation

As shown in Figure 4, Figure 5, Figure 6 and Figure 7, the visualization of destriping results on simulated images indicates that CSCNet achieves the overall best performance for stripe noise removal. More specifically, for the comparisons against statistical-based methods, the complex ground target distribution led to the failure of MM, as shown in Figure 4c, Figure 5c, Figure 6c and Figure 7c, whereas in Figure 4c, the brightness of some areas on the original image was changed after correction. In particular, in Figure 7c, MM not only failed to remove the stripe noise existing in the part rows of the image, but also changed the structural information of the original image, causing additional noise. This is due to the assumption of statistic-based methods, i.e., the mean value and standard deviation of each column are approximately equal, not always being able to hold in real images.
From the destriping results of optimization-based methods, since the total variation (TV) model is employed in ASSTV to increase the smoothness between columns, the destriping performance was somewhat satisfactory. For example, we observe that ASSTV achieved relatively decent performance on the simulated images from Figure 4d, Figure 5d, Figure 6d and Figure 7d. However, as for LRSID, the destriping with the additive low-rank constraint was less satisfactory, including the residual stripes in Figure 4e, Figure 6e, and Figure 7e. Both the method of [43] and PADMM generated promising destriping results on the degraded image with periodical stripe noise (Figure 5f,g); the residual stripe still appeared in the restored images when dealing with the severe noise (Figure 4f,g). In addition, from Figure 7f,g), the method of [43] and PADMM cannot handle the situation of the stripe existing in the rows.
As for the deep learning-based methods, SSGN had relatively better results for images with lighter stripes compared to traditional methods, which can be observed in Figure 6i. However, as shown in Figure 4h, Figure 5h, Figure 6h and Figure 7h and Figure 4i, Figure 5i and Figure 7i, the severe stripes were not fully removed. Compared to these models, relying on the proposed CCM and SCM, our CSCNet is much more effective for various stripes by reducing the local difference between columns and improving the global uniformity of the whole image.
PSNR and SSIM: To quantitatively evaluate the destriping effect of the aforementioned methods, PSNR and SSIM were used to measure the corrected image. The corresponding results are shown in Table 1. The quantitative assessments listed in Table 1 show that the CSCNet achieved the highest PSNR values of all methods on the four simulated images. As for SSIM, our method was only 0.01 lower than ASSTV on the Urban data, which obtained the best results on other simulated images. The results of PSNR and SSIM validate that the proposed CSCNet obtained the overall best performance concerning stripe removal and structural information preservation.
To validate the effectiveness of methods impacted by different noise intensities, we added four types of additive noise to the Urban dataset and provide the relative evaluation indicators for comparisons, including PSNR and SSIM. The results are listed in Table 2. In Table 2, we can observe that CSCNet achieved the best results among PSNR and SSIM with different stripe noise intensities. As the noise intensity increased, the effectiveness of all methods experienced certain decreases. It can be observed that the traditional methods are more susceptible to the noise intensity, which can generate the relative advanced destriping results with low-intensity noise, though the effect is limited when handling the high-intensity noise. Compared to other methods, CSCNet possesses powerful generalization ability to deal with different stripe noise intensities.

4.2. Real Image Destriping

4.2.1. Real Data

Apart from the simulated images, we also evaluated the destriping effects of different methods on five real images to investigate their practicability. The two hyperspectral remote sensing images produced by the full spectrum airborne hyperspectral imager (FAHI), including the near-infrared (VNIR) and shortwave-infrared (SWIR) images, were utilized to test the destriping performances of nine methods. FAHI is the Chinese next-generation pushbroom hyperspectral image instrument [45], and the main parameters are listed in Table 3. Other test images were the public CHRIS images, including CHRIS_AM and CHRIS_UK, and Terra MODIS. The correction results are shown in Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13.

4.2.2. Evaluation

In particular, the destriping results on real images, including the correction results of continuous, discontinuous, thin, and wide stripes, demonstrate the practicality and generalization capacity of the proposed model. Consistently with the results on the simulated data, the classical MM is unsuitable when the ground target distribution between columns changes drastically, as shown in Figure 9b, Figure 10b and Figure 11b. The original image structure was often damaged to a certain degree after the correction.
The optimization-based methods, including ASSTV and LRSID, achieved relatively good correction results for thin stripes; see Figure 9c and Figure 11c, and Figure 9d, relying on the adopted TV model. However, the results in Figure 8c,d indicate that the wide stripe cannot be fully removed. Similarly, in Figure 10c,d, we can also observe that the destriping is rather unstable. In Figure 12c,d and Figure 13c,d, we can see that they affect the original image information after the destriping. For the method of [43], it can be observed from Figure 8e and Figure 11e that this method can remove simple stripe noise (Figure 9e and Figure 11e), and the destriping performance was poor when the noise distribution was complicated (Figure 8e and Figure 10e). PADMM possessed a stronger ability for stripe removal; however, it generally causes over-smoothness, losing the detailed information of original image after destriping (Figure 9f, Figure 12f and Figure 13f).
As shown in Figure 8h, Figure 10h, and Figure 11h, the destriping effects of deep learning-based methods on real images are similar to observations on simulated images. SSGN is more suitable for images with lighter stripes; however, from Figure 8g, Figure 9g, Figure 10g and Figure 11g, and Figure 9h and Figure 12h, it can be observed that the stripe noise remained after the destriping with ICSRN and SSGN. As we can see from the visualization results, TSWEU achieved relatively satisfactory performance except on the VNIR image with the residual stripes. In Figure 13, the detailed structures demonstrate that CSCNet achieved the balance between the stripe removal and original information preservation. Overall, the promising performance on various kinds of simulated and real images indicates the effectiveness of CSCNet on stripe removal.
MRD: The destriping methods should retain the original image information to the greatest extent while removing the stripe. Therefore, the MRD was adopted to measure the distortion of the original image after the different correction methods. It is defined as [34]:
M R D = 1 m n i = 1 m j = 1 n y ( i , j ) x ( i , j ) x ( i , j ) ,
where x ( i , j ) and y ( i , j ) stand for the pixel value in positions ( i , j ) of the original image and the corrected image, respectively. m and n are the numbers of rows and columns in the image, respectively. According to the definition, the average relative difference in images before and after destriping is calculated by MRD. The lower the MRD values are, the greater the ability of the method to preserve the image’s original information. The MRD results concerning nine methods on the five real remote sensing images are shown in Table 4. The lower MRD represents that the image distortion caused by the correction is relatively slighter. Since the destriping changes the information of original striped image, especially in the case of severe stripe noise, it is natural that the corrected image with residual stripe noise often has a lower MRD. Therefore, it is reasonable that the radiation quality of corrected image needs to be improved first before considering reducing the image distortion; i.e., the MRD comparison is meaningful only when compared methods can effectively remove the stripe noise. It was observed that stripes were not fully removed by ICSRN, resulting in the lowest MRD values on all images. A similar case also existed for the corrected result of SSGN on the SWIR image. In Terra MODIS image, our MRD was only about 0.01 larger than the method of [43] with the effective stripe removal. Therefore, CSCNet had a lower MRD on five real images, indicating that the distortion of useful information caused by our model is minimal, which verifies that the proposed model is more balanced between removing the stripe and preserving the detail information.

4.3. Image Uniformity

It is expected that the stripe removal will improve the uniformity of images. Therefore, the comparisons of image non-uniformity could be used to quantitatively validate the destriping ability of algorithms. The lower value of non-uniformity indicates the improvement of image uniformity, which can be calculated as [46]:
U = 1 S ¯ × i = 1 N s ( S i S ¯ ) 2 N s = σ i S ¯ .
where U represents the image non-uniformity, S ¯ is the mean value of DNs, N s is the sample number, S i is the DN of pixel i, and σ i stands for the standard deviation of pixel i. Two uniform regions with the size of 20 × 60 marked by red boxes in Figure 14 were selected from SWIR images. The calculated results of nine methods are shown in Table 5. Based on the evaluation results of non-uniformity improvement, it can be seen in Table 5 that corrected images obtained by CSCNet had lower non-uniformity. It is worth noting that even though the U value of CSCNet is comparable with those of ICSRN and SSGN in region 1, the clear residual stripes remained in the corrected images of ICSRN and SSGN. Therefore, compared to other methods, CSCNet is slightly larger than TSWEU in region 2. Overall, CSCNet improved the image uniformity and smoothness significantly, while achieving advanced destriping results simultaneously, compared to other methods.

4.4. Ablation Study

CCM and SCM: This section verifies the effectiveness of different modules in CSCNet, including CCM and SCM. The SCM was removed in CSCNet to evaluate the capacity of CCM, and vice visa. In Figure 15a, we can observe that the SCM can remove parts of noise by improving the global uniformity of image, but residual stripes remain in the image. Compared to the SCM, the destriping effect is significantly promoted by CCM, as shown in Figure 15b. In particular, in Figure 15b,c, it can be seen that both CCM and CSCNet effectively removed stripes on the SWIR image, which further demonstrates the effectiveness of CCM, that is, focusing on improving the local consistency of homogeneous regions. Relying on SCM and CCM, CSCNet achieved the best performance, as shown in Figure 15c.
The Number of MCSCM: CSCNet consists of multiple multi-scale column-spatial correction modules (MCSCMs). To evaluate the impact of the number of MCSCMs on the destriping performance, we investigated different numbers of MCSCM: 1, 2, 4, and 6. The corresponding PSNR comparisons are shown in Figure 16, and the corrected CHRIS_AM images for band 6 are given in Figure 17. In Figure 16, we can observe that the learning capacity of CSCNet on the training set was promoted with more MCSCMs: the PSNR values for 4 and 6 MCSCMs are both above 36. As shown in Figure 17, CSCNet with 4 and 6 MCSCMs demonstrated similar performance; thus, we selected four MCSCMs in the final CSCNet based on its correction performance, learning capacity, and network complexity.
Multi-scale extension: Moreover, we also evaluated the effectiveness of multi-scale extension. The proposed network without two downsampling branches was utilized for fair comparisons. PSNR during the training is shown in Figure 18, and the correction results of VNIR images for band 64 are shown in Figure 19. It can be observed from PSNR curves that the multi-scale structure significantly promotes the learning ability. Correspondingly, the proposed CSCNet achievd a better destriping effect on the VNIR image, demonstrating that multi-scale branches and the information exchanging among different resolutions effectively improved the stripe noise removal performance.

4.5. Model Complexity Analysis

We have conducted thorough analysis regarding the balance between the effectiveness and efficiency of the destriping models. Specifically, we list the network parameters and flops of four deep learning-based methods in Table 6. Since final CSCNet consists of multiple (four) multi-scaled column-spatial correction modules (MCSCM) in its final version and the parameters and flops naturally increase as the number of MCSCMs raises, we also constructed a light version of CSCNet (denoted Lite-CSCNet) to demonstrate the relationship between the network complexity and effectiveness. The Lite-CSCNet contains two MCSCM. The destriping results of all compared models on CHRIS_AM are shown in Figure 20.
In addition, we also report the inference time for every method on the CHRIS_AM image in Table 6 to evaluate the time complexity of our method. The traditional methods, including MM, ASSTV, LRSID, the method of [43], and PADMM, ran on the MATLAB 2016a platform with an Intel i5-10210U CPU at 1.6 GHz and with 8 GB memory; the deep learning methods, including ICSRN, SSGN, and CSCNet were conducted with pytorch 1.1.0 framework on the workstation with an Intel i7-8700k CPU at 3.8 GHz and with 128 GB memory and NVIDIA GTX1080Ti GPU. The results of TSWEU were obtained on the same workstation with the MATLAB 2016a, Matconvnet framework.
From Table 6, we can observe that ICSRN and SSGN are lightweight models given their simple network architectures, and the parameters and flops of Lite-CSCNet are comparable to those of TSWEU. The final CSCNet was superior compared to other models. In Table 7, it can be observed that the inference speed of MM was the fastest among all traditional methods relying on its simplest implementation, whereas it generated poor destriping performance with the complicated noise distribution, as expected. Compared to traditional methods, deep learning methods process images with much faster speeds. Among the four deep learning methods, although the running time of final CSCNet was not dominant due to its network design, its destriping performance was much better than those of other methods.
It is unreasonable to solely focus on the complexity of models without appropriate consideration of the destriping performance. Therefore, we display the destriping results of five deep learning-based models on the CHRIS_AM image for band 1 in Figure 20. As we can observe that the lightweight models cannot effectively remove the stripe noise. On the other hand, TSWEU and Lite-CSCNet manage to achieve certain improvements against these light models. Though the complexity of CSCNet is the highest, it effectively handles diverse noise distributions. Additionally, it is worth noting that the complexity of CSCNet matches that of the popular modern CNN-based vision models, and the inference speed is acceptable for real applications, as addressed before.
Moreover, since each part of CSCNet, including CCM, SCM, and FFM, is well motivated and was designed by referring to the characteristics of stripe noise, to further validate the improvements made by the proposed model do not come from the increase in the parameters, we conducted extensive ablation studies and parameter analysis to prove the effectiveness of each proposed module (Section 4.4). In summary, by leveraging both effectiveness and efficiency, we selected the current version of CSCNet as our final model.

5. Conclusions

In this paper, we proposed the multi-scaled column-spatial correction network (CSCNet) for remote sensing image destriping. We design the column-based and spatial-based correction modules (CCM and SCM) to focus on reducing the local differences between columns and improving the global uniformity of image, respectively. Moreover, we presented a channel attention-based feature fusion module to facilitate learning representative features from different modules and scales. Extensive experiments were conducted on several simulated and real remote sensing image datasets to verify the effectiveness of the proposed model. The advanced visual and qualitative assessment results indicate that the CSCNet is effective for various types of stripe noise and outperforms state-of-the-art methods. Moreover, CSCNet achieved an adequate balance between improving the uniformity and preserving the original structural information. The supervised deep learning methods often require paired training data, in which the simulated image is commonly generated by manually adding the stripe noise on the clean image according to the degradation model. Since the physical degradation procedure of the stripe is rather complex, the simplified additive or multiplicative model cannot well handle all kinds of stripes in practice. Therefore, the destriping performance of CSCNet is somewhat limited, when dealing with the extremely complex stripe noise distribution. To address the above issue, self-supervised or semi-supervised learning could be leveraged to further improve the generalization capability of the destriping model as our future work.

Author Contributions

Conceptualization, methodology, formal analysis and validation, J.L. and J.Z.; investigation, software, visualization and writing—original draft preparation, J.L.; data curation, D.Z.; resources and supervision, J.Z. and D.Z.; writing—review and editing, J.Z., J.L., J.H. and T.M.; project administration and funding acquisition, D.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Open Project of Key Laboratory of Intelligent Infrared Perception, Chinese Academy of Sciences (CAS-IIRP-2020-01).

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lu, X.; Wang, Y.; Yuan, Y. Graph-Regularized Low-Rank Representation for Destriping of Hyperspectral Images. IEEE Trans. Geoence Remote Sens. 2013, 51, 4009–4018. [Google Scholar] [CrossRef]
  2. Zhang, M.; Carder, K.; Muller-Karger, F.E.; Lee, Z.; Goldof, D.B. Noise Reduction and Atmospheric Correction for Coastal Applications of Landsat Thematic Mapper Imagery. Remote Sens. Environ. 1999, 70, 167–180. [Google Scholar] [CrossRef]
  3. Liu, G.; Wang, L.; Liu, D.; Fei, L.; Yang, J. Hyperspectral Image Classification Based on Non-Parallel Support Vector Machine. Remote Sens. 2022, 14, 2447. [Google Scholar] [CrossRef]
  4. Wang, W.; Han, Y.; Deng, C.; Li, Z. Hyperspectral Image Classification via Deep Structure Dictionary Learning. Remote Sens. 2022, 14, 2266. [Google Scholar] [CrossRef]
  5. Zare, A.; Gader, P. Hyperspectral Band Selection and Endmember Detection Using Sparsity Promoting Priors. IEEE Geoence Remote Sens. Lett. 2008, 5, 256–260. [Google Scholar] [CrossRef]
  6. Ayma Quirita, V.A.; da Costa, G.A.O.P.; Beltrán, C. A Distributed N-FINDR Cloud Computing-Based Solution for Endmembers Extraction on Large-Scale Hyperspectral Remote Sensing Data. Remote Sens. 2022, 14, 2153. [Google Scholar] [CrossRef]
  7. Song, M.; Li, Y.; Yang, T.; Xu, D. Spatial Potential Energy Weighted Maximum Simplex Algorithm for Hyperspectral Endmember Extraction. Remote Sens. 2022, 14, 1192. [Google Scholar] [CrossRef]
  8. Benhalouche, F.Z.; Benharrats, F.; Bouhlala, M.A.; Karoui, M.S. Spectral Unmixing Based Approach for Measuring Gas Flaring from VIIRS NTL Remote Sensing Data: Case of the Flare FIT-M8-101A-1U, Algeria. Remote Sens. 2022, 14, 2305. [Google Scholar] [CrossRef]
  9. Feng, X.; Han, L.; Dong, L. Weighted Group Sparsity-Constrained Tensor Factorization for Hyperspectral Unmixing. Remote Sens. 2022, 14, 383. [Google Scholar] [CrossRef]
  10. Decker, K.T.; Borghetti, B.J. Composite Style Pixel and Point Convolution-Based Deep Fusion Neural Network Architecture for the Semantic Segmentation of Hyperspectral and Lidar Data. Remote Sens. 2022, 14, 2113. [Google Scholar] [CrossRef]
  11. Guo, T.; Luo, F.; Fang, L.; Zhang, B. Meta-Pixel-Driven Embeddable Discriminative Target and Background Dictionary Pair Learning for Hyperspectral Target Detection. Remote Sens. 2022, 14, 481. [Google Scholar] [CrossRef]
  12. Hu, X.; Xie, C.; Fan, Z.; Duan, Q.; Zhang, D.; Jiang, L.; Wei, X.; Hong, D.; Li, G.; Zeng, X.; et al. Hyperspectral Anomaly Detection Using Deep Learning: A Review. Remote Sens. 2022, 14, 1973. [Google Scholar] [CrossRef]
  13. Wegener, M. Destriping multiple sensor imagery by improved histogram matching. Int. J. Remote Sens. 1990, 11, 859–875. [Google Scholar] [CrossRef]
  14. Horn, B.K.P.; Woodham, R.J. Destriping LANDSAT MSS images by histogram modification. Comput. Graph. Image Process. 1979, 10, 69–83. [Google Scholar] [CrossRef] [Green Version]
  15. Gadallah, F.L.; Csillag, F.; Smith, E.J.M. Destriping multisensor imagery with moment matching. Int. J. Remote Sens. 2000, 21, 2505–2511. [Google Scholar] [CrossRef]
  16. Liu, Z.J.; Wang, C.Y.; Wang, C. Destriping Imaging Spectrometer Data by an Improved Moment Matching Method. J. Remote Sens. 2002, 6, 279–284. [Google Scholar] [CrossRef]
  17. Carfantan, H.; Idier, J. Statistical Linear Destriping of Satellite-Based Pushbroom-Type Images. IEEE Trans. Geosci. Remote Sens. 2010, 48, 1860–1871. [Google Scholar] [CrossRef]
  18. Bouali, M.; Ladjal, S. Toward Optimal Destriping of MODIS Data Using a Unidirectional Variational Model. IEEE Trans. Geosci. Remote Sens. 2011, 49, 2924–2935. [Google Scholar] [CrossRef]
  19. Chang, Y.; Yan, L.; Fang, H.; Luo, C. Anisotropic Spectral-Spatial Total Variation Model for Multispectral Remote Sensing Image Destriping. IEEE Trans. Image Process. 2015, 24, 1852–1866. [Google Scholar] [CrossRef]
  20. Liu, X.; Lu, X.; Shen, H.; Yuan, Q.; Jiao, Y.; Zhang, L. Stripe Noise Separation and Removal in Remote Sensing Images by Consideration of the Global Sparsity and Local Variational Properties. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3049–3060. [Google Scholar] [CrossRef]
  21. Zhang, H.; He, W.; Zhang, L.; Shen, H.; Yuan, Q. Hyperspectral Image Restoration Using Low-Rank Matrix Recovery. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4729–4743. [Google Scholar] [CrossRef]
  22. Wang, M.; Yu, J.; Xue, J.H.; Sun, W. Denoising of Hyperspectral Images Using Group Low-Rank Representation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 4420–4427. [Google Scholar] [CrossRef] [Green Version]
  23. Chang, Y.; Yan, L.; Fang, H.; Liu, H. Simultaneous Destriping and Denoising for Remote Sensing Images With Unidirectional Total Variation and Sparse Representation. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1051–1055. [Google Scholar] [CrossRef]
  24. Zhao, Y.Q.; Yang, J. Hyperspectral Image Denoising via Sparse Representation and Low-Rank Constraint. IEEE Trans. Geosci. Remote Sens. 2015, 53, 296–308. [Google Scholar] [CrossRef]
  25. Chang, Y.; Yan, L.; Wu, T.; Zhong, S. Remote Sensing Image Stripe Noise Removal: From Image Decomposition Perspective. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7018–7031. [Google Scholar] [CrossRef]
  26. Sun, H.; Zheng, K.; Liu, M.; Li, C.; Yang, D.; Li, J. Hyperspectral Image Mixed Noise Removal Using a Subspace Projection Attention and Residual Channel Attention Network. Remote Sens. 2022, 14, 2071. [Google Scholar] [CrossRef]
  27. Zhang, J.; Cai, Z.; Chen, F.; Zeng, D. Hyperspectral Image Denoising via Adversarial Learning. Remote Sens. 2022, 14, 1790. [Google Scholar] [CrossRef]
  28. Zhang, Q.; Yuan, Q.; Li, J.; Liu, X.; Shen, H.; Zhang, L. Hybrid Noise Removal in Hyperspectral Imagery With a Spatial–Spectral Gradient Network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7317–7329. [Google Scholar] [CrossRef]
  29. Xiao, P.; Guo, Y.; Zhuang, P. Removing Stripe Noise From Infrared Cloud Images via Deep Convolutional Networks. IEEE Photon. J. 2018, 10, 7801114. [Google Scholar] [CrossRef]
  30. Kuang, X.; Sui, X.; Chen, Q.; Gu, G. Single Infrared Image Stripe Noise Removal Using Deep Convolutional Networks. IEEE Photon. J. 2017, 9, 7800615. [Google Scholar] [CrossRef]
  31. Crippen, R.E. A simple spatial filtering routine for the cosmetic removal of scan-line noise from Landsat TM P-tape imagery. Photogramm. Eng. Remote Sens. 1989, 55, 327–331. [Google Scholar]
  32. Jia, J.; Wang, Y.; Cheng, X.; Yuan, L.; Zhao, D.; Ye, Q.; Zhuang, X.; Shu, R.; Wang, J. Destriping Algorithms Based on Statistics and Spatial Filtering for Visible-to-Thermal Infrared Pushbroom Hyperspectral Imagery. IEEE Trans. Geoence Remote Sens. 2019, 57, 4077–4091. [Google Scholar] [CrossRef]
  33. Pande-Chhetri, R.; Abd-Elrahman, A. De-striping hyperspectral imagery using wavelet transform and adaptive frequency domain filtering. Isprs J. Photogramm. Remote Sens. 2011, 66, 620–636. [Google Scholar] [CrossRef]
  34. Acito, N.; Diani, M.; Corsini, G. Subspace-Based Striping Noise Reduction in Hyperspectral Images. IEEE Trans. Geoence Remote Sens. 2011, 49, 1325–1342. [Google Scholar] [CrossRef]
  35. Infante; Omar, S. Wavelet analysis for the elimination of striping noise in satellite images. Opt. Eng. 2001, 40, 1309–1314. [Google Scholar] [CrossRef]
  36. Shen, H.; Zhang, L. A MAP-Based Algorithm for Destriping and Inpainting of Remotely Sensed Images. IEEE Trans. Geosci. Remote Sens. 2009, 47, 1492–1502. [Google Scholar] [CrossRef]
  37. Xie, Q.; Zhao, Q.; Meng, D.; Xu, Z.; Gu, S.; Zuo, W.; Zhang, L. Multispectral Images Denoising by Intrinsic Tensor Sparsity Regularization. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1692–1700. [Google Scholar] [CrossRef]
  38. Chang, Y.; Yan, L.; Zhong, S. Hyper-Laplacian Regularized Unidirectional Low-Rank Tensor Recovery for Multispectral Image Denoising. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5901–5909. [Google Scholar] [CrossRef]
  39. Chang, Y.; Chen, M.; Yan, L.; Zhao, X.L.; Zhong, S. Toward Universal Stripe Removal via Wavelet-Based Deep Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2880–2897. [Google Scholar] [CrossRef]
  40. Chang, Y.; Yan, L.; Liu, L.; Fang, H.; Zhong, S. Infrared Aerothermal Nonuniform Correction via Deep Multiscale Residual Network. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1120–1124. [Google Scholar] [CrossRef]
  41. Charbonnier, P.; Blanc-Feraud, L.; Aubert, G.; Barlaud, M. Two deterministic half-quadratic regularization algorithms for computed imaging. In Proceedings of the IEEE 1st International Conference on Image Processing, Austin, TX, USA, 13–16 November 1994; Volume 2, pp. 168–172. [Google Scholar] [CrossRef]
  42. Zhang, R. Making convolutional networks shift-invariant again. In Proceedings of the International Conference on Machine Learning (PMLR), Long Beach, CA, USA, 9–15 June 2019; pp. 7324–7334. [Google Scholar]
  43. Chen, Y.; Huang, T.Z.; Zhao, X.L.; Deng, L.J.; Huang, J. Stripe noise removal of remote sensing images by total variation regularization and group sparsity constraint. Remote Sens. 2017, 9, 559. [Google Scholar] [CrossRef] [Green Version]
  44. Dou, H.X.; Huang, T.Z.; Deng, L.J.; Zhao, X.L.; Huang, J. Directional 0 Sparse Modeling for Image Stripe Noise Removal. Remote Sens. 2018, 10, 361. [Google Scholar] [CrossRef] [Green Version]
  45. Wang, Y.; Wei, L.; Yuan, L.; Li, C.; Lv, G.; Xie, F.; Han, G.; Shu, R.; Wang, J. New generation VNIR/SWIR/TIR airborne imaging spectrometer. In Proceedings of the International Symposium on Optoelectronic Technology and Application, Beijing, China, 9–11 May 2016. [Google Scholar] [CrossRef]
  46. Jia, J.; Wang, Y.; Zhuang, X.; Yao, Y.; Wang, S.; Zhao, D.; Shu, R.; Wang, J. High spatial resolution shortwave infrared imaging technology based on time delay and digital accumulation method. Infrared Phys. Technol. 2017, 81, 305–312. [Google Scholar] [CrossRef]
Figure 1. The framework of proposed multi-scaled column-spatial correction network (CSCNet). The main structure of CSCNet is the multiple multi-scaled column-spatial correction module (MCSCM), including the column-spatial correction and feature fusion sub-modules (CSCM and FFM). CSCM corrects the striped image from the perspectives of column difference and global uniformity. Moreover, we use its multi-scale extension to enhance the destriping performance. FFM is adopted to select significant features derived from different modules and scales.
Figure 1. The framework of proposed multi-scaled column-spatial correction network (CSCNet). The main structure of CSCNet is the multiple multi-scaled column-spatial correction module (MCSCM), including the column-spatial correction and feature fusion sub-modules (CSCM and FFM). CSCM corrects the striped image from the perspectives of column difference and global uniformity. Moreover, we use its multi-scale extension to enhance the destriping performance. FFM is adopted to select significant features derived from different modules and scales.
Remotesensing 14 03376 g001
Figure 2. The structure of the column-spatial correction module (CSCM). The CSCM consists of the column-based and spatial-based correction modules (CCM and SCM), which were designed from the perspectives of reducing local differences and improving global uniformity, respectively. FFM is used to fuse features generated by CCM and SCM.
Figure 2. The structure of the column-spatial correction module (CSCM). The CSCM consists of the column-based and spatial-based correction modules (CCM and SCM), which were designed from the perspectives of reducing local differences and improving global uniformity, respectively. FFM is used to fuse features generated by CCM and SCM.
Remotesensing 14 03376 g002
Figure 3. The structure for the feature fusion module (FFM).
Figure 3. The structure for the feature fusion module (FFM).
Remotesensing 14 03376 g003
Figure 4. Corrected Washington DC Mall images for band 42. (a) Original image. (b) Degraded image. Images corrected by (c) MM, (d) ASSTV, (e) LRSID, (f) that of reference [43], (g) PADMM, (h) ICSRN, (i) SSGN, (j) CSCNet.
Figure 4. Corrected Washington DC Mall images for band 42. (a) Original image. (b) Degraded image. Images corrected by (c) MM, (d) ASSTV, (e) LRSID, (f) that of reference [43], (g) PADMM, (h) ICSRN, (i) SSGN, (j) CSCNet.
Remotesensing 14 03376 g004
Figure 5. Corrected HYDICE Urban images for band 120. (a) Original image. (b) Degraded image. Images corrected by (c) MM, (d) ASSTV, (e) LRSID, (f) that of reference [43], (g) PADMM, (h) ICSRN, (i) SSGN, (j) CSCNet.
Figure 5. Corrected HYDICE Urban images for band 120. (a) Original image. (b) Degraded image. Images corrected by (c) MM, (d) ASSTV, (e) LRSID, (f) that of reference [43], (g) PADMM, (h) ICSRN, (i) SSGN, (j) CSCNet.
Remotesensing 14 03376 g005
Figure 6. Corrected Pavia University images for band 99. (a) Original image. (b) Degraded image. Images corrected by (c) MM, (d) ASSTV, (e) LRSID, (f) that of reference [43], (g) PADMM, (h) ICSRN, (i) SSGN, (j) CSCNet.
Figure 6. Corrected Pavia University images for band 99. (a) Original image. (b) Degraded image. Images corrected by (c) MM, (d) ASSTV, (e) LRSID, (f) that of reference [43], (g) PADMM, (h) ICSRN, (i) SSGN, (j) CSCNet.
Remotesensing 14 03376 g006
Figure 7. Corrected Salinas images for band 35. (a) Original image. (b) Degraded image. Images corrected by (c) MM, (d) ASSTV, (e) LRSID, (f) that of reference [43], (g) PADMM, (h) ICSRN, (i) SSGN, (j) CSCNet.
Figure 7. Corrected Salinas images for band 35. (a) Original image. (b) Degraded image. Images corrected by (c) MM, (d) ASSTV, (e) LRSID, (f) that of reference [43], (g) PADMM, (h) ICSRN, (i) SSGN, (j) CSCNet.
Remotesensing 14 03376 g007
Figure 8. Corrected VNIR images for band 64. (a) VNIR image. Images corrected by (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Figure 8. Corrected VNIR images for band 64. (a) VNIR image. Images corrected by (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Remotesensing 14 03376 g008
Figure 9. Corrected SWIR images for band 264. (a) SWIR image. Images corrected by (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Figure 9. Corrected SWIR images for band 264. (a) SWIR image. Images corrected by (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Remotesensing 14 03376 g009
Figure 10. Corrected CHRIS_AM images for band 1. (a) CHRIS_AM image. Images corrected by (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Figure 10. Corrected CHRIS_AM images for band 1. (a) CHRIS_AM image. Images corrected by (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Remotesensing 14 03376 g010
Figure 11. Corrected CHRIS_UK images for band 3. (a) CHRIS_UK image. Images corrected by (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Figure 11. Corrected CHRIS_UK images for band 3. (a) CHRIS_UK image. Images corrected by (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Remotesensing 14 03376 g011
Figure 12. Corrected Terra MODIS images for band 9. (a) Terra MODIS image. Images corrected by (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Figure 12. Corrected Terra MODIS images for band 9. (a) Terra MODIS image. Images corrected by (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Remotesensing 14 03376 g012
Figure 13. Zoomed results of Figure 12. (a) Original image, (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Figure 13. Zoomed results of Figure 12. (a) Original image, (b) MM, (c) ASSTV, (d) LRSID, (e) the method of [43], (f) PADMM, (g) ICSRN, (h) SSGN, (i) TSWEU, (j) CSCNet.
Remotesensing 14 03376 g013
Figure 14. Two ground target regions in the SWIR image.
Figure 14. Two ground target regions in the SWIR image.
Remotesensing 14 03376 g014
Figure 15. The images corrected with different models. Band 1 of the CHRIS_AM image and band 264 of the SWIR image are displayed sequentially from top to bottom. Corrected images obtained by (a) SCM, (b) CCM, (c) CSCNet.
Figure 15. The images corrected with different models. Band 1 of the CHRIS_AM image and band 264 of the SWIR image are displayed sequentially from top to bottom. Corrected images obtained by (a) SCM, (b) CCM, (c) CSCNet.
Remotesensing 14 03376 g015
Figure 16. The PSNR comparisons of the proposed CSCNet composed of different numbers of MCSCM during training. n stands for the number of MCSCM.
Figure 16. The PSNR comparisons of the proposed CSCNet composed of different numbers of MCSCM during training. n stands for the number of MCSCM.
Remotesensing 14 03376 g016
Figure 17. Corrected CHRIS_AM images for band 6. Images corrected by CSCNet composed of (a) 1, (b) 2, (c) 4, and (d) 6 MCSCMs.
Figure 17. Corrected CHRIS_AM images for band 6. Images corrected by CSCNet composed of (a) 1, (b) 2, (c) 4, and (d) 6 MCSCMs.
Remotesensing 14 03376 g017
Figure 18. The PSNR comparisons of the proposed CSCNet with and without the multi-scale extension during training.
Figure 18. The PSNR comparisons of the proposed CSCNet with and without the multi-scale extension during training.
Remotesensing 14 03376 g018
Figure 19. Corrected VNIR images for band 64. Images corrected by CSCNet (a) without and (b) with the multi-scale extension.
Figure 19. Corrected VNIR images for band 64. Images corrected by CSCNet (a) without and (b) with the multi-scale extension.
Remotesensing 14 03376 g019
Figure 20. Corrected CHRIS_AM images for band 1. (a) ICSRN, (b) SSGN, (c) TSWEU, (d) Lite-CSCNet, (e) CSCNet.
Figure 20. Corrected CHRIS_AM images for band 1. (a) ICSRN, (b) SSGN, (c) TSWEU, (d) Lite-CSCNet, (e) CSCNet.
Remotesensing 14 03376 g020
Table 1. PSNR and SSIM of the simulated image destriping results.
Table 1. PSNR and SSIM of the simulated image destriping results.
ImageIndexMMASSTVLRSIDRef. [43]PADMMICSRNSSGNCSCNet
DCPSNR26.3117.7523.3223.8223.8123.5924.2328.98
SSIM0.890.900.900.810.810.770.820.90
UrbanPSNR32.8623.7132.8729.4029.4929.2128.0534.43
SSIM0.960.970.940.950.960.940.940.96
PaviaUPSNR28.4829.4035.5528.9429.1428.9231.3436.56
SSIM0.950.980.980.970.980.960.980.99
SalinasPSNR22.0627.4326.8530.7330.7131.2434.5535.54
SSIM0.820.970.940.960.950.970.980.98
Table 2. PSNR and SSIM of the urban images with different stripe noise intensities.
Table 2. PSNR and SSIM of the urban images with different stripe noise intensities.
IntensityIndexMMASSTVLRSIDRef. [43]PADMMICSRNSSGNCSCNet
(−50, 50)PSNR27.8639.5346.2944.6847.4845.9239.1451.63
SSIM0.93910.98830.99400.99160.99770.99630.98090.9989
(−100, −50)
(50, 100)
PSNR27.8037.4841.9439.5739.9743.1638.9750.12
SSIM0.93650.98440.99220.98590.99250.99350.98160.9987
(−200, −100)
(100, 200)
PSNR27.2334.3837.0434.4834.3539.4536.5446.29
SSIM0.92430.97490.98350.97310.98060.98980.980.9973
(−300, −200)
(200, 300)
PSNR26.1232.1833.0330.7130.5440.0436.6246.56
SSIM0.90740.96280.97080.95680.96560.99000.97470.9972
Table 3. Main FAHI parameters (VNIR; SWIR; CCD; and MCT).
Table 3. Main FAHI parameters (VNIR; SWIR; CCD; and MCT).
ItemVNIRSWIR
Spectral range ( μ m)0.4–0.950.95–2.5
FOV (∘)14.714.7
Detector/array sizeCCD/1024 × 256MCT/512 × 512
Spectral resolution (nm)2.343
Band numbers64512
Table 4. MRD of real images’ destriping results.
Table 4. MRD of real images’ destriping results.
MethodVNIRSWIRCHRIS_AMCHRIS_UKTerra MODIS
MM0.07810.10420.01550.11450.1179
ASSTV0.04450.09810.04040.05280.0436
LRSID0.04640.07930.05670.01720.0418
Ref. [43]0.01820.11530.03960.05330.0093
PADMM0.15260.17580.06220.64240.0201
ICSRN0.01000.05860.00700.01310.0754
SSGN0.02350.06140.00850.01770.0077
TSWEU0.10490.21420.06300.14320.0758
CSCNet0.01120.07130.00820.01630.0196
Table 5. Image non-uniformity using nine methods.
Table 5. Image non-uniformity using nine methods.
MethodMMASSTVLRSIDRef. [43]PADMMICSRNSSGNTSWEUCSCNet
Region10.05560.05610.05560.07870.06110.05350.05140.09190.0548
Region20.04110.02320.02190.02710.03500.02210.02110.01810.0206
Table 6. Network parameters and flops of five deep learning-based models.
Table 6. Network parameters and flops of five deep learning-based models.
MethodICSRNSSGNTSWEULite-CSCNetCSCNet
Parameters (M)0.80.23.23.16.2
Flops (G)541110390175
Table 7. Inference time (in seconds) of every method on the CHRIS_AM image (for one band).
Table 7. Inference time (in seconds) of every method on the CHRIS_AM image (for one band).
MethodMMASSTVLRSIDRef. [43]PADMMICSRNSSGNTSWEULite-CSCNetCSCNet
Time0.019.9344.476.313.990.260.070.450.350.61
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, J.; Zeng, D.; Zhang, J.; Han, J.; Mei, T. Column-Spatial Correction Network for Remote Sensing Image Destriping. Remote Sens. 2022, 14, 3376. https://doi.org/10.3390/rs14143376

AMA Style

Li J, Zeng D, Zhang J, Han J, Mei T. Column-Spatial Correction Network for Remote Sensing Image Destriping. Remote Sensing. 2022; 14(14):3376. https://doi.org/10.3390/rs14143376

Chicago/Turabian Style

Li, Jia, Dan Zeng, Junjie Zhang, Jungong Han, and Tao Mei. 2022. "Column-Spatial Correction Network for Remote Sensing Image Destriping" Remote Sensing 14, no. 14: 3376. https://doi.org/10.3390/rs14143376

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop