Next Article in Journal
Tracking of Multiple Maneuvering Random Hypersurface Extended Objects Using High Resolution Sensors
Previous Article in Journal
Effect of Attention Mechanism in Deep Learning-Based Remote Sensing Image Processing: A Systematic Literature Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Remote Sensing Image Super-Resolution Based on Dense Channel Attention Network

1
School of Information Engineering, Ningxia University, Yinchuan 750021, China
2
State Key Laboratory of Information Engineering in Surveying, Mapping, and Remote Sensing, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(15), 2966; https://doi.org/10.3390/rs13152966
Submission received: 6 July 2021 / Revised: 22 July 2021 / Accepted: 23 July 2021 / Published: 28 July 2021

Abstract

:
In the recent years, convolutional neural networks (CNN)-based super resolution (SR) methods are widely used in the field of remote sensing. However, complicated remote sensing images contain abundant high-frequency details, which are difficult to capture and reconstruct effectively. To address this problem, we propose a dense channel attention network (DCAN) to reconstruct high-resolution (HR) remote sensing images. The proposed method learns multi-level feature information and pays more attention to the important and useful regions in order to better reconstruct the final image. Specifically, we construct a dense channel attention mechanism (DCAM), which densely uses the feature maps from the channel attention block via skip connection. This mechanism makes better use of multi-level feature maps which contain abundant high-frequency information. Further, we add a spatial attention block, which makes the network have more flexible discriminative ability. Experimental results demonstrate that the proposed DCAN method outperforms several state-of-the-art methods in both quantitative evaluation and visual quality.

1. Introduction

High-resolution (HR) remote sensing images provide detailed geometric information about land cover. Thus, HR remote sensing images are essential for many applications, such as object detection [1,2], urban planning [3], building extraction [4,5,6] and so on [7,8,9,10,11,12,13,14,15]. However, the spatial resolution of remote sensing images are influenced by the limitations of hardware and environmental factors [16,17,18]. Compared to current physical imaging technology, super resolution (SR) which recovers HR images from low-resolution (LR) images is more convenient and low cost. Thus, SR becomes an alternative method in the remote sensing field.
SR can be classified into single image super resolution (SISR) and multiple image super resolution (MISR) [19] by the number of input images. MISR methods utilize multiple LR images of the same area, to provide more information, to better reconstruct high spatial-frequency details and texture [20]. Nevertheless, it is difficult to obtain multiple remote sensing images of the same scene. Thus, this paper considers a study of SISR methods. The traditional SISR methods [21] are mainly grouped into interpolation-based methods and reconstruction-based methods [22]. The interpolation-based methods predict the unknown pixel using a simple linear or non-linear interpolation [23] operation. Although interpolation-based methods are convenient, they have limited performance for images which contain more details. The reconstruction-based methods utilize a certain kind of prior information to produce better results, such as local, nonlocal and sparse priors [23,24]. Although reconstruction-based methods [25] are flexible and allow consideration of different prior constraints [26,27,28,29,30,31], they also have difficulties with complicated remote sensing images. Recently, deep learning-based methods attracted a lot of attention in the remote sensing images super resolution task. In 2015, a super resolution convolutional neural network (SRCNN) [32] was first proposed by Dong et al. to achieve the natural images super resolution. As the poineer, SRCNN learned the mapping between the HR images and the corresponding LR images using a three layers network. While SRCNN had outperformed traditional-based methods, bicubic LR images made the network to operate in a high-dimensional space and largely increased the computational cost. To alleviate the problem, a fast super resolution convolutional neural netork (FSRCNN) [33] and an efficient subpixel convolutional network (ESPCN) [34] were proposed which used a deconvolutional layer and a subpixel convolutional layer to achieve direct reconstruction in a low-dimensional space and save the computational cost, respectively. These networks were shallow and their performance were limited by the network depth. However, increasing the network depth leads to a vanishing gradient and an exploding gradient. To handle the problem, the skip connection operation [35] was proposed by He et al., which combined the low-level features and high-level features to effectively alleviate the gradient vanishing. Thus, the skip connection operation was gradually used in the SR networks. Among these networks, the very deep super resolution convolutional network (VDSR) [36] which used a global residual connection was proposed by Kim et al. to spread the LR information to the network end. It was the first network which introduced the residual learning to SR and succeeded training a 20-layer network. Besides, enhanced deep super resolution (EDSR) [37] and SRResNet [38] also used global residual connection. In addition, EDSR and SRResNet employed residual blocks as basic network module, which introduced local residual connection to ease the deep network training difficulty. Later, Zhang et al. [39] constructed a residual in residual structure where residual blocks compose residual groups using short and long skip connections. Further, cascading residual network (CARN) [40], dense deep back-projection network (D-DBPN) [41], SRDenseNet [42], residual dense network (RDN) [43], employed the dense skip connections or multiple skip connections to increase the training effect.
Deep learning-based SR methods in the field of remote sensing also developed fast in recent years. In 2017, a local-global combined network (LGC) [19] was first proposed by Lei et al. to enhance the spatial resolution of remote sensing images. LGC learns multi-level information including local details and global priors using the skip connection operation. In 2018, a residual dense backprojection network (RDBPN) [22] was proposed by Pan et al., which consists of several residual dense backprojection blocks that contain the upprojection module and the downprojection module. In 2020, Zhang et al. proposed a scene adaptive method [44] via a multi-scale attention network to enhance the SR reconstruction details under the different remote sensing scenes. Recently, an approach named dense-sampling super resolution network (DSSR) [45] presented a dense sampling mechanism which reuses an upscaler to upsample and overcome the large-scale remote sensing images SR reconstruction problem. However, the complex spatial distribution of remote sensing images need more attention. In 2020, a second-order multi-scale super resolution network (SMSR) [46] was proposed by Dong et al. to reuse the learned multi-level information to the high-frequency regions of remote sensing images. The multi-perception attention network (MPSR) [47] and the multi-scale residual neural network (MRNN) [48] are also doing some related work about using multi-scale information. In addition, the generative adversarial network (GAN)-based SR method is used to generate visually pleasing remote sensing images. In 2019, Jiang et al. presented an edge-enhancement generative adversarial network (EEGAN) [49], which introduces an edge enhancement module to improve the remote sensing images SR performance. In 2020, Lei et al. proposed a coupled-discriminated generative adversarial network (CDGAN) [50] for solving the discrimination-ambiguity problem for the low-frequency regions in the remote sensing images.
Although the above-mentioned methods have good performance, their results can be further improved. First, the distributions of remote sensing images are very complex; therefore, we need more high-frequency details and texture to better reconstruct HR images. Secondly, redundancy feature information are not beneficial to recover details and increase computation cost. So, we propose a dense channel attention network (DCAN) which learns multi-level feature information and pays more attention to the important and useful regions in order to better reconstruct the final image. The major contributions are as follows:
(1)
We propose a DCAN for SR of the single remote sensing image, which makes full use of the features learned at different depths through densely using multi-level feature information and pay more attention to high-frequency regions. Both quantitative and qualitative evaluations demonstrate the superiority of DCAN over the state-of-the-art methods.
(2)
A dense channel attention mechanism (DCAM) is proposed to utilize the channel attention block through the dense skip connection manner. This mechanism can increase the flow of information through the network and improve the representation capacity of the network.
(3)
A spatial attention block (SAB) is added to the network. This helps the network have more flexible discriminative ability for different local regions. It contributes to reconstruct the final image. In addition, this helps the network have more flexible discriminative ability for global structure and focus on high-frequency information from the spatial dimension.

2. Method

2.1. Network Architecture

The architecture of our proposed DCAN is illustrated in Figure 1, which consists of three parts: shallow feature extraction, deep feature extraction, and reconstruction. The network first extracts the shallow features from the input LR image. Then, the second part extracts the deep features and increase the weights of important feature maps. Finally, the features which contain abundant useful information are sent to the third part of network for reconstruction. The network details will be introduced in the next part.
(1)
Shallow Feature Extraction: Let Conv(k, f, c) be a convolutional layer, where k, f, c represent the filter kernel size, the number of filters, and the number of filter channels, respectively. We use a 3 × 3 convolutional layer to extract the shallow feature F 0 from the input image I L R , which contributes to the next feature extraction. The shallow feature extraction f S F ( . ) operation can be formulated as follows:
F 0 = f S F ( I L R ) = C o n v ( k , f , c ) ( I L R )
(2)
Deep Feature Extraction: After the shallow feature extraction, the backbone part which contains a series of Dense Channel Attention blocks ( DCAB s ) and a Spatial Attention Block (SAB) is designed to extract deep features. The main block DCAB receives the feature maps from each single DCAB as input, the structure of which will be given in Section 2.2. Then, the feature F G which is the sum of output features of G DCAB s is sent to a convolutional layer:
F G = C o n v ( k , f , c ) ( F G )
where F G is generated by the convolution operation, G denotes the number of DCAB s . Then, the F G is sent to a SAB as follows:
F SAB = H SAB ( F G )
where H SAB denotes the operation of the SAB. The operational details will be described in Section 2.3.
(3)
Reconstruction: The reconstruction part contains two convolutional layers and a deconvolution layer. The SR image I S R is generated as follows:
I S R = C o n v ( k , f , c ) ( H U p ( C o n v ( k , f , c ) ( F SAB ) ) )
where H u p denotes the operation of the deconvolution layer.

2.2. Dense Channel Attention Mechanism

As discussed in Section 1, most existing super-resolution models do not make full use of the information from the input LR image. So, we propose a novel dense channel attention mechanism to solve the problem. It concentrates on the high-frequency information and weakens the useless information. Figure 2 shows DCAM, which can be described that the G t h DCAB computes the feature map F G from the outputs of the DCAB G 1 , DCAB G 2 , . . . , DCAB 1 as follows:
F G = H DCAB G ( F G 1 + F G 2 + + F 1 )
where H DCAB G denotes the operation of the G t h DCAB, F G 1 , F G 2 , , F 1 denote the outputs of the DCAB G 1 , DCAB G 2 , , DCAB 1 . The purpose of DCAM is to focus on high-frequency components and make better use of the information.
As shown in Figure 2, the DCAB is the basic block of the proposed DCAM. It mainly contains two convolutional layers and a CA (Channel Attention) block. To be specific, in Figure 3, the first convolutional layer before Relu consists of n f filters of size n c × 3 × 3 . The second convolutional layer after Relu contains n f filters of size n f × 3 × 3 . Let F i 1 be the input of DCAB i (the i-th DCAB in the DCAB s ), there is
X i = W ( 3 , n f , n f ) σ ( W ( 3 , n f , n c ) F i 1 )
where X i is an intermediate feature contaninting n f feature maps. W ( 3 , n f , n f ) and W ( 3 , n f , n c ) denote the weight matrix of n f filters of size n f × 3 × 3 and n f filters of size n c × 3 × 3 . σ ( . ) = m a x ( 0 , x ) indicates the Relu activation.
Then, CA is used to improve the discriminative learning ability. As shown in Figure 4, the mechanism of CA can be formulated as follows:
X i = X i S i = [ X 1 i , , X n f i ] [ S 1 i , , S n f i ]
where X i is the output of X i , X 1 i , , X n f i is the feature map of X i , S 1 i , , S n f i are the elements of S i , ⊗ indicates the elementwise product. S i is a n f dimensional channel statistical descriptor, which is used to update X i . S i is obtained by
S i = f s i g m o i d ( f u p ( σ ( f d o w n ( f g a p ( X i ) ) ) ) )
It contains several operations, including global average pooling f g a p ( ) , channel-down f d o w n = C o n v ( 1 , n f / r , n f ) , Relu activation σ ( ) , channel-up f u p = C o n v ( 1 , n f , n f / r ) , and sigmoid function f s i g m o i d , on X i (r is set to 16). The channel statistical descriptor can help express the different information among the feature maps of X i .
As shown in Figure 3, the input of the i t h DCAB is from the output of i 1 , , i t h 1 DCAB s . In general, the complete operation of DCAB can be formulated as follows:
F i = S i ( W ( 3 , n f , n f ) σ ( W ( 3 , n f , n c ) F i 1 ) ) + F i 1
This operation increases the flow of information through the network and the representation capacity of the network.

2.3. Spatial Attention Block

Considering the complicated spatial information and distribution of the remote sensing images, we add a SAB to increase the discriminative ability of the network. It helps the network have discriminative ability for different local regions and pay more attention to the regions which are more important and more difficult to reconstruct. As shown in Figure 5, the operation of SAB can be formulated as follows:
F SAB = F i n p u t f s i g m o i d ( f c o n v ( f c o n c a t ( f A v g p o o l i n g ( F i n p u t ) , f M a x p o o l i n g ( F i n p u t ) ) ) )
where F SAB is obtained by several operations, including average pooling f A v g p o o l i n g ( ) , f M a x p o o l i n g ( ) , concat f c o n c a t ( ) , channel-down f c o n v ( ) = C o n v ( 1 , 1 , 2 ) , sigmoid function f s i g m o i d ( ) , and element-wise product σ . The operation of SAB can focus on the local regions of F i n p u t which are useful to reconstruct.

2.4. Loss Function

We use the L 1 loss as the total loss because the L 1 loss can support better convergence. In addition, the L 1 loss can be described as follows:
L 1 ( θ ) = 1 n i = 1 n f DCAN ( I i L R ; θ ) I i H R
where θ represents the whole parameter of the DCAN network, n represents the number of training images. The purpose of the L 1 loss function is that the reconstructed HR image I i S R = f DCAN ( I i L R ; θ ) can be similar to its corresponding ground truth image I i H R .

3. Results

3.1. Experimental Settings

The relevant experimental settings on the experimental datasets, degradation method, and evaluation metrics are detailed in this section.
(1) Datasets: We use UC Merced dataset and RSSCN7 dataset for qualitative and quantitative analysis. The UC Merced dataset [51] is a classification dataset of remote sensing images containing 21 land use classes, and each class contains 100 images with size 256 × 256 and RGB channels. Figure 6 shows some images. The UC Merced dataset is mainly used as the experimental data and divided into two sections. The first section includes images in 21 classes of agricultural, baseball diamond, beach, building and so on. For each category, 90 images are taken to create the training set. Therefore, the second section includes the remaining ten images for each class which are taken to construct the test set. We can validate the performance of our model for each class. The results are discussed in Section 3.2.
In addition, the other dataset named RSSCN7 [52] is also used to train our method and verify the effectiveness of our method. The RSSCN7 dataset is a classification dataset of remote sensing images containing 7 land use classes, and each class contains 400 images with size 400 × 400 and RGB channels. Figure 7 shows some images. This dataset is divided into two sections. The first section contains images in 7 classes. For each class, 360 images are used to train the model. Thus, the second section includes the remaining 40 images for each class which are used to construct the test set. The test results are discussed in Section 3.2.
To validate the robustness of the proposed method, the real world data from GaoFen-2 in Ningxia, China, are used to test our model in Section 3.4. Experiments are designed by the × 4 scale factor and the × 8 scale factor, respectively. We obtain LR images by down sampling HR images using the bicubic operation.
All the parameter settings are same on the UC Merced dataset and RSSCN7 dataset. The training step is performed on the three channels of the RGB space. The channel number of the input image n c is 3, the filter size k is set as 3. The Adam [53] optimizer with β 1 = 0.9 and β 2 = 0.999 is used to train the proposed models under the batchsize of 16. The weights of the model are initialized using the method in [54]. We initialize the learning rate as 10 4 and halve it at every 2 × 10 5 batches updates. We implement the proposed algorithm with the PyTorch [55] framework, and we train the DCAN model using one NVIDIA Tesla V100 GPU.
(2) Evaluation Metrics: Two widely used image quality assessment metrics, the peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) are used as evaluation metrics. They are used as the criteria for evaluating the quality of reconstructed HR images. Given the ground truth HR image x and its corresponding super-resolved image y, the PSNR of the super-resolved image y is computed as follows:
MSE ( x , y ) = 1 n s i = 1 n s ( x i y i ) 2
PSNR ( x , y ) = 10 l o g 10 255 2 MSE ( x , y )
where x i and y i denote the ith pixel in x and y values, respectively. n s represents the image pixel number. The higher PSNR indicates that the quality of the SR image is better. In addition to using PSNR to measure image quality, we also use SSIM as the image quality assessment, the SSIM of the super-resolved image y is defined as follows:
SSIM ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 )
where μ x and μ y denote the mean values of x and y, and σ x and σ y denote the standard deviation values of x and y, and C 1 and C 2 are constants. The higher SSIM indicates that the quality of the SR image is better.

3.2. Comparisons with the Other Methods

The proposed DCAN method is compared with four state-of-the-art methods: Bicubic, SRCNN [32], VDSR [36], SRResNet [38]. In this experiment, we use the same training datasets and test datasets to compare these methods fairly. In addition, these models are trained well to test under the suitable conditions. Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14 and Figure 15 show some reconstructed images of the UC Merced test set using the corresponding algorithm and the ground truth HR images in the visual quality.
Figure 8 shows the super-resolved HR images obtained through different models for “airplane89.jpg” with a scale factor of × 4 . The figure demonstrates that the reconstructed result of our DCAN model is better than those of the other algorithms, such as the edge of the plane. The reconstructed images for “overpass18.jpg” with a scale factor of × 4 are shown in Figure 9. It is obvious that DCAN reconstructed more details and recovered sharper edges, such as the traffic index line of the road and the car. Comparison results for “harbor32.jpg” of a × 8 scale factor are shown in Figure 10. We can get the conclusion that our method successfully reconstructed the main part of the bridge. The reconstructed HR images of “storagetanks68.jpg” are compared in Figure 12. The results demonstrate that the DCAN method provided more enhanced textures and recovered more details in comparison with other algorithms. Figure 13 shows the reconstructed images of “mediumresidential34.jpg” of a × 4 scale factor. It can be seen that the result of our proposed method reconstructed the edges sharper than the other methods. Figure 14 compares the reconstructed images of “runway39.jpg” with other methods. It is obvious that the proposed methods reconstructed the details better than the other algorithms, such as the thin strip lines at the right of the image. As shown in Figure 11 and Figure 15, we enlarge some regions of super-resolved images to better compare the textures and details. The texture and details of our proposed method are superior than those of other methods. The above results show that our method achieves better results compared with other methods in terms of the visual quality.
Table 1 and Table 2 show the PSNR and SSIM of the reconstructed HR images of the UC Merced test set on the × 4 and × 8 enlargement, respectively.
Our proposed method is also better than other methods in the quantitative evaluation. Table 3 and Table 4 show the PSNR and SSIM of the reconstructed HR images of RSSCN7 test set on the × 4 enlargement and × 8 enlargement, respectively. Experimental results demonstrate the effectiveness of our method.

3.3. Model Analysis

In this section, we use three groups of experiments to analyze the proposed method using the UC Merced dataset, including the number of DCAB, the dense channel attention mechanism, and the SAB. In addition, we use PSNR as quantitative evaluation.
(1)
DCAB: A small DCAN which contains four DCABs was used to learn the effect of n f . The experimental results about the effect of n f are presented in Table 5. We use three widely used values, including 32, 64 and 128. When n f = 64 , the average PSNR values on the test dataset are 0.27 dB higher than n f = 32 , and it also are 0.16 dB higher than n f = 128 . So, we set the n f = 64 in the rest of the experiments.
(2)
Dense Channel Attention Mechanism: We use several DCAN which contain different DCAB s to study the effect of the number of DCABs. The experimental results about the effect of DCAB s are provided in Table 6. We set the number of DCABs to 4, 8 and 10, respectively. As shown in the table, we conclude that the PSNR of ten DCABs are higher than 4 and 8.
(3)
SAB: To study the effect of the SAB, we compare the DCAN-s which contain a SAB and the DCAN without SAB. The experimental results are provided in Table 7. It can be seen that the PSNR of DCAN-s is higher than DCAN. It demonstrates that SAB improves the network performance.
In addition, we explore the relationship between PSNR and epochs of network training. As shown in Figure 16, it can be seen that the PSNR of around 350 epochs begins to converge with a scale factor of ×4. As shown in Figure 17, at around 50 epochs, it begins to converge with a scale factor of ×8. We can draw the conclusion that, with our model, it is easy to achieve high performance.

3.4. Super-Resolving the Real-World Data

In this section, we use the real data to validate the robustness of our proposed method. The model is trained by UC Merced dataset and tested by some remote sensing images from GaoFen-2 in Ningxia, China, as the real-world data. The size and band number of real-word images are 400 × 400 and 3, respectively. Figure 18 and Figure 19 show one LR image on the × 4 and × 8 enlargement, respectively. The reconstructed results show that the DCAN method achieves good results in terms of the visual quality.

4. Discussion

The proposed DCAN method is proven to have a good performance with experimental results. In Section 3.2, our method outperforms several state-of-the-art methods in both quantitative evaluation and visual quality. In Section 3.3, the increase of PSNR after adding DCAB and SAB demonstrates the effectiveness of our approch. In Section 3.4, we deal with real satellite images from GaoFen-2 using the DCAN model and obtain satisfactory SR results. In addition, we discussed the experimental results in combination with theoretical analysis.
(1)
Effect of Dense Channel Attention Block: As shown in Table 6, the PSNR of super-resolved images increases after we add the DCAB. It can be seen that when the number of DCAB s increases from 4 to 10, it improves the performance of network. It proves that when we increase appropriately the number of DCAB s , the capacity of the network can be improved.
(2)
Effect of Spatial Attention Block: As shown in Table 7, it can be seen that the PSNR value ranges from 28.63 to 28.70 after we add the SAB. Thus, we conclude that SAB can improve the performance of network and help the network have more flexible discriminative ability for global structure and focus on high-frequency information from spatial dimension.
(3)
Effect of Scale Factor: As shown in Table 1 and Table 2, due to the increasing scale factor, the improvement of our proposed DCAN method reduces. It indicates that the large-scale super resolution is still a hard problem.

5. Conclusions

This article develops a DCAN network which achieves good performance in super-resolving the remote sensing images with complicated spatial distribution. Specifically, we design a network which densely uses the multi-level feature information and strengthens the effective information. In addition, we propose a dense channel attention mechanism which makes better use of multi-level feature maps which contain abundant high-frequency information. Further, we add a spatial attention block to pay more attention to the regions which are more important and more difficult to reconstruct. Results of the extensive experiments demonstrate the superiority of our method over the other compared algorithms.

Author Contributions

Funding acquisition, P.L. and X.S.; Investigation, Y.M.; Methodology, Y.M., P.L. and H.L.; Resources, P.L., H.L., X.S. and Y.Z.; Supervision, P.L.; Validation, Y.M.; Writing—original draft, Y.M. and P.L.; Writing—review & editing, P.L. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 42001307, and Grant 62061038, in part by the Ningxia Key RD Program under Grant 2020BFG02013, and in part by the Natural Science Foundation of Ningxia under Grant 2020AAC02006.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dong, Z.; Wang, M.; Wang, Y.; Zhu, Y.; Zhang, Z. Object detection in high resolution remote sensing imagery based on convolutional neural networks with suitable object scale features. IEEE Trans. Geosci. Remote Sens. 2019, 58, 2104–2114. [Google Scholar] [CrossRef]
  2. Amit, S.N.K.B.; Shiraishi, S.; Inoshita, T.; Aoki, Y. Analysis of satellite images for disaster detection. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 5189–5192. [Google Scholar]
  3. Mathieu, R.; Freeman, C.; Aryal, J. Mapping private gardens in urban areas using object-oriented techniques and very high-resolution satellite imagery. Landsc. Urban Plan. 2007, 81, 179–192. [Google Scholar] [CrossRef]
  4. Li, W.; He, C.; Fang, J.; Zheng, J.; Fu, H.; Yu, L. Semantic segmentation-based building footprint extraction using very high-resolution satellite images and multi-source GIS data. Remote Sens. 2019, 11, 403. [Google Scholar] [CrossRef] [Green Version]
  5. Zhao, K.; Kang, J.; Jung, J.; Sohn, G. Building extraction from satellite images using mask R-CNN with building boundary regularization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 247–251. [Google Scholar]
  6. Yuan, S.; Dong, R.; Zheng, J.; Wu, W.; Zhang, L.; Li, W.; Fu, H. Long time-series analysis of urban development based on effective building extraction. In Proceedings of the Geospatial Informatics X, International Society for Optics and Photonics, Online, 21 April 2020; Volume 11398, p. 113980M. [Google Scholar]
  7. Lu, T.; Ming, D.; Lin, X.; Hong, Z.; Bai, X.; Fang, J. Detecting building edges from high spatial resolution remote sensing imagery using richer convolution features network. Remote Sens. 2018, 10, 1496. [Google Scholar] [CrossRef] [Green Version]
  8. Zhao, P.; Liu, K.; Zou, H.; Zhen, X. Multi-stream convolutional neural network for SAR automatic target recognition. Remote Sens. 2018, 10, 1473. [Google Scholar] [CrossRef] [Green Version]
  9. Zhang, W.; Witharana, C.; Liljedahl, A.K.; Kanevskiy, M. Deep convolutional neural networks for automated characterization of arctic ice-wedge polygons in very high spatial resolution aerial imagery. Remote Sens. 2018, 10, 1487. [Google Scholar] [CrossRef] [Green Version]
  10. Xu, Y.; Zhu, M.; Li, S.; Feng, H.; Ma, S.; Che, J. End-to-end airport detection in remote sensing images combining cascade region proposal networks and multi-threshold detection networks. Remote Sens. 2018, 10, 1516. [Google Scholar] [CrossRef] [Green Version]
  11. Ma, J.; Zhao, J.; Jiang, J.; Zhou, H.; Guo, X. Locality preserving matching. Int. J. Comput. Vis. 2019, 127, 512–531. [Google Scholar] [CrossRef]
  12. Lv, P.; Zhong, Y.; Zhao, J.; Zhang, L. Unsupervised change detection based on hybrid conditional random field model for high spatial resolution remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4002–4015. [Google Scholar] [CrossRef]
  13. Kakareko, G.; Jung, S.; Ozguven, E.E. Estimation of tree failure consequences due to high winds using convolutional neural networks. Int. J. Remote Sens. 2020, 41, 9039–9063. [Google Scholar] [CrossRef]
  14. Song, Y.; Zhang, Z.; Baghbaderani, R.K.; Wang, F.; Qu, Y.; Stuttsy, C.; Qi, H. Land cover classification for satellite images through 1d cnn. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar]
  15. Kocatepe, A.; Ulak, M.B.; Kakareko, G.; Ozguven, E.E.; Jung, S.; Arghandeh, R. Measuring the accessibility of critical facilities in the presence of hurricane-related roadway closures and an approach for predicting future roadway disruptions. Nat. Hazards 2019, 95, 615–635. [Google Scholar] [CrossRef]
  16. Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  17. Shen, H.; Zhang, L.; Huang, B.; Li, P. A MAP approach for joint motion estimation, segmentation, and super resolution. IEEE Trans. Image Process. 2007, 16, 479–490. [Google Scholar] [CrossRef]
  18. Köhler, T.; Huang, X.; Schebesch, F.; Aichert, A.; Maier, A.; Hornegger, J. Robust multiframe super-resolution employing iteratively re-weighted minimization. IEEE Trans. Comput. Imaging 2016, 2, 42–58. [Google Scholar] [CrossRef]
  19. Lei, S.; Shi, Z.; Zou, Z. Super-resolution for remote sensing images via local–global combined network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1243–1247. [Google Scholar] [CrossRef]
  20. Li, F.; Jia, X.; Fraser, D.; Lambert, A. Super resolution for remote sensing images based on a universal hidden Markov tree model. IEEE Trans. Geosci. Remote Sens. 2009, 48, 1270–1278. [Google Scholar]
  21. Garzelli, A. A review of image fusion algorithms based on the super-resolution paradigm. Remote Sens. 2016, 8, 797. [Google Scholar] [CrossRef] [Green Version]
  22. Pan, Z.; Ma, W.; Guo, J.; Lei, B. Super-resolution of single remote sensing image based on residual dense backprojection networks. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7918–7933. [Google Scholar] [CrossRef]
  23. Chang, K.; Ding, P.L.K.; Li, B. Single image super-resolution using collaborative representation and non-local self-similarity. Signal Process. 2018, 149, 49–61. [Google Scholar] [CrossRef]
  24. Lu, X.; Yuan, H.; Yuan, Y.; Yan, P.; Li, L.; Li, X. Local learning-based image super-resolution. In Proceedings of the 2011 IEEE 13th International Workshop on Multimedia Signal Processing, Hangzhou, China, 17–19 October 2011; pp. 1–5. [Google Scholar]
  25. Zhong, Y.; Zhang, L. Remote sensing image subpixel mapping based on adaptive differential evolution. IEEE Trans. Syst. Man Cybern. Part B 2012, 42, 1306–1329. [Google Scholar] [CrossRef]
  26. Yang, J.; Wright, J.; Huang, T.S.; Ma, Y. Image super-resolution via sparse representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef] [PubMed]
  27. Lu, X.; Yan, P.; Yuan, Y.; Li, X.; Yuan, H. Utilizing homotopy for single image superresolution. In Proceedings of the First Asian Conference on Pattern Recognition, Beijing, China, 28 November 2011; pp. 316–320. [Google Scholar]
  28. Lu, X.; Yuan, H.; Yan, P.; Yuan, Y.; Li, X. Geometry constrained sparse coding for single image super-resolution. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 1648–1655. [Google Scholar]
  29. Lu, X.; Yuan, Y.; Yan, P. Image super-resolution via double sparsity regularized manifold learning. IEEE Trans. Circuits Syst. Video Technol. 2013, 23, 2022–2033. [Google Scholar] [CrossRef]
  30. Lu, X.; Yuan, Y.; Yan, P. Alternatively constrained dictionary learning for image superresolution. IEEE Trans. Cybern. 2013, 44, 366–377. [Google Scholar] [PubMed]
  31. Dong, W.; Zhang, L.; Shi, G.; Wu, X. Image deblurring and super-resolution by adaptive sparse domain selection and adaptive regularization. IEEE Trans. Image Process. 2011, 20, 1838–1857. [Google Scholar] [CrossRef] [Green Version]
  32. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef] [Green Version]
  33. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Lecture Notes in Computer Science, Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 391–407. [Google Scholar]
  34. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1874–1883. [Google Scholar]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  36. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1646–1654. [Google Scholar]
  37. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 June 2017; pp. 136–144. [Google Scholar]
  38. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 June 2017; pp. 4681–4690. [Google Scholar]
  39. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  40. Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate, and lightweight super-resolution with cascading residual network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 252–268. [Google Scholar]
  41. Haris, M.; Shakhnarovich, G.; Ukita, N. Deep back-projection networks for super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Like City, UT, USA, 18–22 June 2018; pp. 1664–1673. [Google Scholar]
  42. Tong, T.; Li, G.; Liu, X.; Gao, Q. Image super-resolution using dense skip connections. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4799–4807. [Google Scholar]
  43. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Like City, UT, USA, 18–22 June 2018; pp. 2472–2481. [Google Scholar]
  44. Zhang, S.; Yuan, Q.; Li, J.; Sun, J.; Zhang, X. Scene-adaptive remote sensing image super-resolution using a multiscale attention network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4764–4779. [Google Scholar] [CrossRef]
  45. Dong, X.; Sun, X.; Jia, X.; Xi, Z.; Gao, L.; Zhang, B. Remote sensing image super-resolution using novel dense-sampling networks. IEEE Trans. Geosci. Remote Sens. 2020, 59, 1618–1633. [Google Scholar] [CrossRef]
  46. Dong, X.; Wang, L.; Sun, X.; Jia, X.; Gao, L.; Zhang, B. Remote Sensing Image Super-Resolution Using Second-Order Multi-Scale Networks. IEEE Trans. Geosci. Remote Sens. 2020, 59, 3473–3485. [Google Scholar] [CrossRef]
  47. Dong, X.; Xi, Z.; Sun, X.; Gao, L. Transferred multi-perception attention networks for remote sensing image super-resolution. Remote Sens. 2019, 11, 2857. [Google Scholar] [CrossRef] [Green Version]
  48. Lu, T.; Wang, J.; Zhang, Y.; Wang, Z.; Jiang, J. Satellite image super-resolution via multi-scale residual deep neural network. Remote Sens. 2019, 11, 1588. [Google Scholar] [CrossRef] [Green Version]
  49. Jiang, K.; Wang, Z.; Yi, P.; Wang, G.; Lu, T.; Jiang, J. Edge-enhanced GAN for remote sensing image superresolution. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5799–5812. [Google Scholar] [CrossRef]
  50. Lei, S.; Shi, Z.; Zou, Z. Coupled adversarial training for remote sensing image super-resolution. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3633–3643. [Google Scholar] [CrossRef]
  51. Yang, Y.; Newsam, S. Bag-of-visual-words and spatial extensions for land-use classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in gEographic Information Systems, lSan Jose, CA, USA, 2–5 November 2010; pp. 270–279. [Google Scholar]
  52. Zou, Q.; Ni, L.; Zhang, T.; Wang, Q. Deep learning based feature selection for remote sensing scene classification. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2321–2325. [Google Scholar] [CrossRef]
  53. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  54. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 1026–1034. [Google Scholar]
  55. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic Differentiation in Pytorch. 2017. Available online: https://openreview.net/forum?id=BJJsrmfCZ (accessed on 28 October 2017).
Figure 1. The structure of the proposed dense channel attention network (DCAN).
Figure 1. The structure of the proposed dense channel attention network (DCAN).
Remotesensing 13 02966 g001
Figure 2. The structure of dense channel attention mechanism (DCAM).
Figure 2. The structure of dense channel attention mechanism (DCAM).
Remotesensing 13 02966 g002
Figure 3. The structure of dense channel attention block (DCAB).
Figure 3. The structure of dense channel attention block (DCAB).
Remotesensing 13 02966 g003
Figure 4. The structure of Channel Attention Block (CA).
Figure 4. The structure of Channel Attention Block (CA).
Remotesensing 13 02966 g004
Figure 5. The structure of Spatial Attention Block (SAB).
Figure 5. The structure of Spatial Attention Block (SAB).
Remotesensing 13 02966 g005
Figure 6. Some images in the UC Merced dataset: 21 land use classes, including buildings, agricultural, airplane, baseball diamond, beach, chaparral, dense residential, forest, freeway, golf course, harbor, intersection, medium residential, mobile home park, overpass, parking lot, river, runway, sparse residential, storage tanks, and tennis court. The above images correspond to the above categories respectively.
Figure 6. Some images in the UC Merced dataset: 21 land use classes, including buildings, agricultural, airplane, baseball diamond, beach, chaparral, dense residential, forest, freeway, golf course, harbor, intersection, medium residential, mobile home park, overpass, parking lot, river, runway, sparse residential, storage tanks, and tennis court. The above images correspond to the above categories respectively.
Remotesensing 13 02966 g006
Figure 7. Some images in the RSSCN7 dataset: 7 land use classes, including grass, filed, industry, river lake, forest, resident and parking. The above images correspond to the above categories respectively.
Figure 7. Some images in the RSSCN7 dataset: 7 land use classes, including grass, filed, industry, river lake, forest, resident and parking. The above images correspond to the above categories respectively.
Remotesensing 13 02966 g007
Figure 8. × 4 super resolved-images of “airplane89.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Figure 8. × 4 super resolved-images of “airplane89.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Remotesensing 13 02966 g008
Figure 9. × 4 super-resolved images of “overpass18.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Figure 9. × 4 super-resolved images of “overpass18.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Remotesensing 13 02966 g009
Figure 10. × 8 super-resolved images of “harbor32.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Figure 10. × 8 super-resolved images of “harbor32.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Remotesensing 13 02966 g010
Figure 11. The detail of super-resolved “harbor32.jpg” region with a scale factor of × 8 . (a) represent HR image, (be) represent the super-resolved image patch of SRCNN, VDSR, SRResNet, DCAN.
Figure 11. The detail of super-resolved “harbor32.jpg” region with a scale factor of × 8 . (a) represent HR image, (be) represent the super-resolved image patch of SRCNN, VDSR, SRResNet, DCAN.
Remotesensing 13 02966 g011
Figure 12. × 4 super-resolved images of “storagetanks68.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Figure 12. × 4 super-resolved images of “storagetanks68.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Remotesensing 13 02966 g012
Figure 13. × 8 super-resolved images of “mediumresidential34.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Figure 13. × 8 super-resolved images of “mediumresidential34.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Remotesensing 13 02966 g013
Figure 14. × 8 super-resolved images of “runway39.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Figure 14. × 8 super-resolved images of “runway39.jpg” on the UC Merced dataset via different algorithms, and the numbers under these images represent the PSNR and SSIM values. (a) represent HR image, (bf) represent the super-resolved image of Bicubic, SRCNN, VDSR, SRResNet, DCAN.
Remotesensing 13 02966 g014
Figure 15. The detail of super-resolved “runway39.jpg” region with a scale factor of × 8 . (a) represent HR image, (be) represent the super-resolved image patch of SRCNN, VDSR, SRResNet, DCAN.
Figure 15. The detail of super-resolved “runway39.jpg” region with a scale factor of × 8 . (a) represent HR image, (be) represent the super-resolved image patch of SRCNN, VDSR, SRResNet, DCAN.
Remotesensing 13 02966 g015
Figure 16. PSNR of DCAN with a scale factor of ×4 on the UC Merced dataset.
Figure 16. PSNR of DCAN with a scale factor of ×4 on the UC Merced dataset.
Remotesensing 13 02966 g016
Figure 17. PSNR of DCAN with a scale factor of ×8 on the UC Merced dataset.
Figure 17. PSNR of DCAN with a scale factor of ×8 on the UC Merced dataset.
Remotesensing 13 02966 g017
Figure 18. SR results of real data of × 4 and × 8 scale factor. (ad) represent the result of Bicubic ×   4 , DCAN ×   4 , Bicubic ×   8 , DCAN ×   8 , respectively.
Figure 18. SR results of real data of × 4 and × 8 scale factor. (ad) represent the result of Bicubic ×   4 , DCAN ×   4 , Bicubic ×   8 , DCAN ×   8 , respectively.
Remotesensing 13 02966 g018
Figure 19. SR results of real data of × 4 and × 8 scale factor. (ad) represent the result of Bicubic ×   4 , DCAN ×   4 , Bicubic ×   8 , DCAN ×   8 , respectively.
Figure 19. SR results of real data of × 4 and × 8 scale factor. (ad) represent the result of Bicubic ×   4 , DCAN ×   4 , Bicubic ×   8 , DCAN ×   8 , respectively.
Remotesensing 13 02966 g019
Table 1. The PSNR and SSIM of UC Merced test dataset with a scale factor of 4.
Table 1. The PSNR and SSIM of UC Merced test dataset with a scale factor of 4.
DataScaleBicubicSRCNNVDSRSRResNetDCAN
Agricultural × 4 25.24/0.452625.66/0.472825.87/0.0.475325.74/0.481126.20/0.5332
Airplane × 4 25.83/0.752426.40/0.765427.93/0.831127.95/0.819228.57/0.8627
Baseball diamond × 4 30.32/0.775430.73/0.800331.48/0.816631.78/0.819532.37/0.8314
Beach × 4 33.16/0.837533.47/0.850733.85/0.861134.05/0.866534.26/0.8678
Buildings × 4 21.44/0.660222.11/0.668722.55/0.744923.62/0.781424.12/0.8217
Chaparral × 4 24.15/0.651224.78/0.697824.86/0.714625.21/0.743125.54/0.7438
Dense residential × 4 23.34/0.672123.85/0.674224.27/0.721125.43/0.778525.92/0.8014
Forest × 4 26.28/0.601026.47/0.641326.63/0.657226.89/0.672327.11/0.6914
Freeway × 4 26.23/0.677326.69/0.691427.11/0.741127.91/0.770828.46/0.7928
Golf course × 4 30.76/0.771131.22/0.783431.24/0.791231.53/0.800332.31/0.8077
Harbor × 4 17.38/0.690617.90/0.699118.58/0.806419.79/0.845321.03/0.8603
Intersection × 4 24.37/0.695124.81/0.698324.89/0.743725.82/0.776326.11/0.7883
Medium residential × 4 23.47/0.653423.96/0.686424.51/0.728125.42/0.753325.63/0.7682
Mobile homepark × 4 21.83/0.663122.45/0.694323.19/0.751624.31/0.781524.93/0.8006
Overpass × 4 23.14/0.641723.51/0.646324.06/0.694425.41/0.746925.94/0.7664
Parking lot × 4 19.21/0.601119.83/0.621319.52/0.661720.04/0.714821.08/0.7760
River × 4 26.41/0.655126.82/0.690426.99/0.703827.07/0.710927.37/0.7162
Runway × 4 26.56/0.717727.24/0.740128.21/0.775329.64/0.796230.53/0.8093
Sparse residential × 4 26.15/0.670426.64/0.703827.08/0.726327.65/0.738427.81/0.7463
Storage tanks × 4 23.57/0.697024.01/0.699824.43/0.760725.26/0.787525.53/0.8211
Tennis court × 4 28.03/0.781228.67/0.781628.99/0.832430.02/0.851430.46/0.8601
Table 2. The PSNR and SSIM of UC Merced test dataset with a scale factor of 8.
Table 2. The PSNR and SSIM of UC Merced test dataset with a scale factor of 8.
DataScaleBicubicSRCNNVDSRSRResNetDCAN
Agricultural × 8 23.62/0.271123.74/0.281323.78/0.281723.61/0.280623.77/0.2874
Airplane × 8 22.98/0.615223.45/0.632123.89/0.656424.02/0.658324.33/0.6759
Baseball diamond × 8 27.15/0.651427.87/0.666127.98/0.682528.17/0.685628.41/0.6933
Beach × 8 30.31/0.738530.56/0.738330.94/0.752430.68/0.757431.02/0.7614
Buildings × 8 18.74/0.4497/19.14/0.462319.45/0.520919.92/0.549720.11/0.5687
Chaparral × 8 20.61/0.316520.53/0.339520.68/0.356720.93/0.374120.99/0.3977
Dense residential × 8 20.44/0.461720.92/0.477521.03/0.523621.55/0.549121.73/0.5702
Forest × 8 23.97/0.374124.14/0.401124.21/0.401824.13/0.401924.27/0.4147
Freeway × 8 23.84/0.494624.27/0.527124.59/0.549724.71/0.553324.95/0.5841
Golf course × 8 27.65/0.649128.15/0.660128.56/0.675328.63/0.681728.96/0.6893
Harbor × 8 14.84/0.508915.16/0.499715.29/0.584915.72/0.624315.91/0.6704
Intersection × 8 21.46/0.488021.89/0.511221.82/0.519422.36/0.543722.63/0.5722
Medium residential × 8 20.40/0.422720.77/0.441620.84/0.466321.36/0.499921.55/0.5208
Mobile homepark × 8 18.27/0.414318.59/0.436118.78/0.469819.42/0.506819.61/0.5301
Overpass × 8 20.72/0.418220.97/0.442421.25/0.465921.67/0.493222.86/0.5321
Parking lot × 8 16.98/0.377117.21/0.388916.95/0.405717.19/0.418717.29/0.4416
River × 8 24.41/0.502524.75/0.526924.74/0.531124.77/0.533524.94/0.5435
Runway × 8 23.25/0.556723.51/0.549824.03/0.584924.84/0.613625.63/0.6371
Sparse residential × 8 22.93/0.467823.28/0.491523.41/0.502223.72/0.516823.95/0.5298
Storage tanks × 8 21.22/0.535621.68/0.543921.87/0.583522.16/0.600822.79/0.6451
Tennis court × 8 24.43/0.596825.13/0.619425.41/0.645625.63/0.654226.13/0.8078
Table 3. The PSNR and SSIM of RSSCN7 test dataset with a scale factor of 4.
Table 3. The PSNR and SSIM of RSSCN7 test dataset with a scale factor of 4.
DataScaleBicubicSRCNNVDSRSRResNetDCAN
Grass × 4 32.69/0.807333.48/0.813234.04/0.823134.27/0.829434.39/0.8342
Industry × 4 24.42/0.661724.85/0.665125.76/0.718526.39/0.741426.61/0.7536
River lake × 4 29.89/0.801430.33/0.806230.91/0.823131.17/0.829431.35/0.8346
Filed × 4 31.83/0.722832.59/0.733233.10/0.752733.37/0.752833.51/0.7683
Forest × 4 27.79/0.599628.06/0.607228.27/0.627128.38/0.640328.53/0.6421
Resident × 4 23.82/0.643124.11/0.645324.87/0.693125.39/0.718225.59/0.7287
Parking × 4 23.76/0.618724.34/0.633225.12/0.676125.57/0.706425.82/0.7138
Table 4. The PSNR and SSIM of RSSCN7 test dataset with a scale factor of 8.
Table 4. The PSNR and SSIM of RSSCN7 test dataset with a scale factor of 8.
DataScaleBicubicSRCNNVDSRSRResNetDCAN
Grass × 8 29.67/0.721431.07/0.736131.41/0.751331.55/0.749631.76/0.7552
Industry × 8 21.85/0.485322.04/0.487222.44/0.513122.58/0.516422.69/0.5273
River lake × 8 27.42/0.701127.91/0.715428.23/0.726328.35/0.729528.51/0.7353
Filed × 8 30.04/0.677130.97/0.684131.25/0.687431.36/0.689331.44/0.6936
Forest × 8 25.56/0.441825.91/0.443326.11/0.455426.17/0.458526.31/0.4615
Resident × 8 20.78/0.423921.13/0.425121.44/0.448221.54/0.456321.70/0.4624
Parking × 8 21.77/0.453122.05/0.468122.35/0.487322.46/0.494222.59/0.4998
Table 5. The PSNR of DCAN with a scale factor of 4 under different n f on the UC Merced dataset.
Table 5. The PSNR of DCAN with a scale factor of 4 under different n f on the UC Merced dataset.
n f PSNR
3228.41
6428.68
12828.52
Table 6. The PSNR of DCAN which have different DCAB s with a scale factor of 4 on the UC Merced dataset.
Table 6. The PSNR of DCAN which have different DCAB s with a scale factor of 4 on the UC Merced dataset.
DCAB s PSNR
428.68
828.70
1028.71
Table 7. The PSNR of DCAN and DCAN-s with a scale factor of 4 on the UC Merced dataset.
Table 7. The PSNR of DCAN and DCAN-s with a scale factor of 4 on the UC Merced dataset.
ModelPSNR
DCAN28.63
DCAN-S28.70
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ma, Y.; Lv, P.; Liu, H.; Sun, X.; Zhong, Y. Remote Sensing Image Super-Resolution Based on Dense Channel Attention Network. Remote Sens. 2021, 13, 2966. https://doi.org/10.3390/rs13152966

AMA Style

Ma Y, Lv P, Liu H, Sun X, Zhong Y. Remote Sensing Image Super-Resolution Based on Dense Channel Attention Network. Remote Sensing. 2021; 13(15):2966. https://doi.org/10.3390/rs13152966

Chicago/Turabian Style

Ma, Yunchuan, Pengyuan Lv, Hao Liu, Xuehong Sun, and Yanfei Zhong. 2021. "Remote Sensing Image Super-Resolution Based on Dense Channel Attention Network" Remote Sensing 13, no. 15: 2966. https://doi.org/10.3390/rs13152966

APA Style

Ma, Y., Lv, P., Liu, H., Sun, X., & Zhong, Y. (2021). Remote Sensing Image Super-Resolution Based on Dense Channel Attention Network. Remote Sensing, 13(15), 2966. https://doi.org/10.3390/rs13152966

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop