Next Article in Journal
Effects of Assimilating Clear-Sky FY-3D MWHS2 Radiance on the Numerical Simulation of Tropical Storm Ampil
Next Article in Special Issue
Learning Adjustable Reduced Downsampling Network for Small Object Detection in Urban Environments
Previous Article in Journal
Preface: Earth Observations for Environmental Sustainability for the Next Decade
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Making Low-Resolution Satellite Images Reborn: A Deep Learning Approach for Super-Resolution Building Extraction

1
Department of Earth System Science, Tsinghua University, Beijing 100084, China
2
Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
3
CUHK-SenseTime Joint Lab, The Chinese University of Hong Kong, Hong Kong, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(15), 2872; https://doi.org/10.3390/rs13152872
Submission received: 5 July 2021 / Revised: 18 July 2021 / Accepted: 20 July 2021 / Published: 22 July 2021
(This article belongs to the Special Issue Deep Learning in Remote Sensing Application)

Abstract

:
Existing methods for building extraction from remotely sensed images strongly rely on aerial or satellite-based images with very high resolution, which are usually limited by spatiotemporally accessibility and cost. In contrast, relatively low-resolution images have better spatial and temporal availability but cannot directly contribute to fine- and/or high-resolution building extraction. In this paper, based on image super-resolution and segmentation techniques, we propose a two-stage framework (SRBuildingSeg) for achieving super-resolution (SR) building extraction using relatively low-resolution remotely sensed images. SRBuildingSeg can fully utilize inherent information from the given low-resolution images to achieve high-resolution building extraction. In contrast to the existing building extraction methods, we first utilize an internal pairs generation module (IPG) to obtain SR training datasets from the given low-resolution images and an edge-aware super-resolution module (EASR) to improve the perceptional features, following the dual-encoder building segmentation module (DES). Both qualitative and quantitative experimental results demonstrate that our proposed approach is capable of achieving high-resolution (e.g., 0.5 m) building extraction results at 2×, 4× and 8× SR. Our approach outperforms eight other methods with respect to the extraction result of mean Intersection over Union (mIoU) values by a ratio of 9.38%, 8.20%, and 7.89% with SR ratio factors of 2, 4, and 8, respectively. The results indicate that the edges and borders reconstructed in super-resolved images serve a pivotal role in subsequent building extraction and reveal the potential of the proposed approach to achieve super-resolution building extraction.

Graphical Abstract

1. Introduction

With rapid urbanization in recent years, high-resolution building extraction plays an increasingly essential role in urban planning, change monitoring, and population estimation [1,2,3,4]. With a rich set of remotely sensed images, it is possible to infer and distinguish buildings from background objects at pixel level [5]. Such a process is defined as building segmentation or building extraction [6].
In terms of data source, very high resolution (VHR) remotely sensed images were viewed as an essential data source for producing high-resolution building extraction in previous studies, such as 0.1 m airborne images [7,8] and 0.5 m space-borne images [9]. Nevertheless, those VHR images are restricted to a limited spatial extent and temporal availability, thus making the methods which demand VHR images as data sources difficult to apply in large areas. In contrast, relatively low-resolution images such as satellite-based images of WorldView series (1.2–2.4 m) and Planet series (3 m) have better spatiotemporal availability. Notwithstanding, it has been proven by Mariana Belgiu and Lucian Drǎguţ [10] along with Huiping Huang et al. [11] that remotely sensed images with relatively lower resolution could generally lead to lower accuracy and coarser boundaries in segmentation results. The resolution inconsistency between the remotely sensed images and the building extraction target greatly impacts the segmentation results. Ryuhei Hamaguchi and Shuhei Hikosaka [12] pointed out that deep learning models that were trained using low-resolution images could hardly extract buildings with significantly different high-resolution. Haut et al. [13] pointed out that the resolution of remotely sensed images significantly affects the distribution of the spatial features, which is important in distinguishing the pixels of buildings from those of the background. Therefore, it remains challenging to develop an automated framework for achieving super-resolution building extraction results using relatively low-resolution remotely sensed images [14,15,16].
Despite the difficulties, achieving super-resolution building extraction results from relatively low-resolution remotely sensed images can be of great value. First, for long-term building change detection research, relatively low-resolution remotely sensed images are irreplaceable and exclusive, especially for the late 20th century and early 21st century [17,18,19]. In these cases, relatively low-resolution remotely sensed images are the only choice for building extraction. In addition, with the rich diversity of satellites and remote sensor technologies, it is common to observe inconsistent spatial resolutions in source datasets and target results for a certain task. For example, Chen et al. [20] transferred collected low-resolution training material into remotely sensed image pixel classification of another resolution version, making it possible to generate building segmentation results over large areas [10,21,22] or long time periods [23,24,25] via spatiotemporally available low-resolution remotely sensed images.
To conduct such a task, the simplest and most widely used solution is to interpolate all the resolution-inconsistent images into one desired resolution in the preprocess, for example, by bilinear interpolation and bicubic interpolation [26]. However, interpolation-based methods, for which the generated pixels are calculated by adjacent pixels, suffer the loss of spatial information, especially in the edges and high-frequency regions where interpolation will generate insufficient gradients [26,27].
Alternatively, super-resolution (SR) methods aim at reconstructing low-resolution images into high-resolution versions with finer spatial details [28]. SR provides a promising alternative to map remotely sensed images with inconsistent resolution into a version with uniform resolution for high-resolution building extraction. However, existing SR approaches in remote sensing require a number of external high-resolution images to obtain training datasets. Haut et al. [29] retrieved 2100 external high-resolution images for training, while Shao et al. [30] collected more than 100,000 image patches for training. Developing a novel SR approach with no need for external high-resolution images remains challenging but valuable. Moreover, previous studies mainly focus on the perceptual improvements of super-resolved images, with no evaluation regarding how much the improvement of image perceptual quality can be transferred into the improvement of subsequent building extraction.
Hereby, we propose the edge-aware super-resolved building segmentation network (SRBuildingSeg) as a novel framework to achieve super-resolution building extraction. The major contributions in this paper are as follows:
  • We propose a two-stage framework for attaining super-resolution building extraction, named SRBuildingSeg, which can make use of the extracted features of the given low-resolution images to improve the performance of building extraction in high-resolution representation.
  • Considering the self-similarity between each building in remotely sensed images, we develop an internal pairs generation module (IPG) and an edge-aware super-resolution module (EASR). Using the two proposed modules, we can fully utilize the internal information of the given images to improve the perceptional features for subsequent building segmentation without any external high-resolution images.
  • We propose a dual-encoder integration module (DES) for building segmentation tasks which enables our approach to attain super-resolution building extraction by fully utilizing the texture features and enhanced perceptional features.
  • We demonstrate that the reconstructed high frequency information of the super-resolved image can be transferred into the improvement of the super-resolution building extraction task. The assessment results reveal that our proposed approach ranks the best among all eight compared methods.
The rest of the paper is organized as follows: in Section 2, we introduce the related work, including the existing deep learning-based building extraction methods and single image super-resolution techniques. In Section 3, we provide a detailed description of the proposed approach. Experimental results and discussion are given in Section 4 and Section 5. We present our conclusions in Section 6.

2. Related Work

2.1. Building Extraction Using Deep Learning Approaches

Since Sakrapee Paisitkriangkrai et al. [31] first proposed a CNN based framework to extract buildings in multispectral images, deep learning based building extraction approaches were proposed and have proven to be effective using VHR images [32,33,34]. Despite the great success of deep learning approaches in building semantic segmentation, only a few discussions focus on building extraction in which the given images and the target results differ in spatial resolution. Belgiu and Drǎguţ [10], and Hamaguchi and Hikosaka [12] compared the building segmentation results of several different approaches using multi-resolution remotely sensed images. They found that the accuracy of extraction results differs with respect to each building size and each specific resolution. Schuegraf and Bittner [35] proposed a hybrid deep learning network for obtaining high-resolution (0.5 m) building extraction results using low-resolution (2 m) multi-spectral and panchromatic images, but their experimental results only show slight improvement in extracting buildings of small size. Guo et al. [36] proposed a framework to extract buildings from relatively low-resolution remotely sensed images while using relatively high-resolution images as training material. Nevertheless, their proposed framework could only generate low-resolution segmentation results from the given high-resolution training material. In addition, they used 0.5 m remotely sensed images as “low-resolution images”, and their extraction accuracy rapidly declines as the ratio of unaligned resolutions enlarges. Thus, it remains a challenge to obtain fine and high-resolution building extraction from low-resolution remotely sensed images.

2.2. Single Image Super-Resolution

Single image super-resolution (SISR), which aims at reconstructing the image into a high-resolution version while providing finer spatial details than those of the original version [28], has emerged as a promising alternative in mapping low-resolution remotely sensed images into versions of higher resolution [37,38,39]. Although super-resolution (SR) can reconstruct essential details of land features from the original datasets into a specific desired spatial resolution, tremendous external paired high-resolution images for training are also generally required [30,40]. Moreover, the reconstructed images generated by those SR models strongly relied on the external information provided by training material, which made the collection of training samples more difficult [41]. At the same time, the SR based models trained in an unsupervised way, e.g., the unsupervised Generative Adversarial Networks (GAN) model for SR [42], have emerged as practical alternatives. However, the performances of those unsupervised algorithms usually are unsatisfactory in the high-frequency region as compared with the supervised approaches [43,44]. Another unsupervised SR model, zero-shot super-resolution (ZSSR) [41], requires thousands of gradient updates in image reconstruction. In addition, remotely sensed images are usually large in size, but the ZSSR model is designed for natural images of small size. Thus, it is still challenging to generate fine super-resolved images without using external high-resolution images. Furthermore, the contributions of the SR methods for the subsequent building extraction lack qualitative evaluation and discussion.

3. Methodology

In this paper, we aim to utilize the given low-resolution remotely sensed images to achieve building extraction in high-resolution representation. As illustrated in Figure 1, the overall framework SRBuildingSeg is a two-stage architecture. Stage one focuses on reconstructing a high-resolution version from the given low-resolution images. We first propose an internal pairs generation module (IPG) to construct LR-HR training pairs, which can improve the model trained in an unsupervised way. Hereafter, we reconstruct the super-resolved images using an edge-aware super-resolution module (EASR) which is trained on the constructed training dataset. Stage two exploits the dual-encoder building segmentation module (DES) to achieve building extraction in high-resolution representation, which takes both super-resolved images and enhanced perceptional features as input in order to improve the segmentation performance.
We will elaborate on the details of SRBuildingSeg in the following sections. In Section 3.1 and Section 3.2, we respectively introduce the IPG module and EASR module. The description of DES is presented in Section 3.3, the assessment criteria and loss function are presented in Section 3.4.

3.1. Internal Pairs Generation Module

Existing supervised SR methods in the remote sensing domain require a large number of LR-HR pairs as training material. In addition, the performance of supervised approaches strongly relies on the external information extracted from LR-HR pairs, e.g., the representativeness of the training dataset. Considering the limitations of supervised SR approaches, we propose an internal pairs generation module (IPG) to obtain LR-HR training pairs without any external high-resolution images. Different from existing supervised approaches, the IPG can fully exploit the self-similarity of the remotely sensed image, which generally covers a large area and thus contains buildings of nearly all various colors, shapes, surroundings, materials, heights, and forms. The internal information of the remotely sensed images is a generalized and representative information source, which proved its effectiveness in the training of the SR model [23,45].
The proposed IPG consists of four steps to generate the HR and its corresponding LR (LR-HR) training pairs from the given low-resolution image I l o w . First, we obtain the “HR training pairs” by simply splitting and cropping the given low-resolution image I l o w . In other words, the “HR training pairs” I L R is actually presented in relatively low resolution, which is considered as a high-resolution information source in the process of training dataset generation. The corresponding LR training pairs I L L R s are then obtained by downscaling each image in I L R using bicubic interpolation, where the s represents the desired SR scale factor. The “LR training pairs” I L L R s is actually a lower-resolution version of the given I l o w . The generated I L R and I L L R s consist of many LR-HR image pairs, which can be used as input and target in the training process of the subsequent SR module. Furthermore, for the sake of robustness, as well as to enrich the diversity of building sizes, we generate many versions of the LR-HR image pairs using a random upscale factor. Finally, the training material is enriched by randomly transforming each image in LR-HR pairs using four rotations (0°, 90°, 180°, 270°), with a mirror reflecting in the vertical and horizontal directions, flipping, resampling, and adding Gaussian noise.
Taking the training pairs generated with a scale factor of 4 as an example, as illustrated in Figure 2, we first generate HR training pairs I L R (2 m) via cropping and splitting the given images I l o w (2 m). We then downscale each image in I L R and generate the corresponding LR training pairs I L L R s (8 m). After dataset enrichment, the generated LR-HR pairs are used as training material for SR model. Finally, we use the properly trained SR model and the given images I l o w (2 m) as input to generate super-resolved images I H s (0.5 m).

3.2. Edge-Aware Super-Resolution Module for Reconstructing High-Resolution Images

Considering that our LR-HR training pairs are generated using only the given low-resolution images and contain no external information, the high-frequency information of reconstructed images, which plays a vital role in subsequent building extraction [46], remains to be improved. We employ an edge-aware super-resolution module (EASR) to better reconstruct the high frequency of any given low-resolution remotely sensed images. EASR integrates the initial generative adversarial subnetwork and gradient-based enhancement subnetwork. In the training phase, the EASR utilizes the constructed I L R and I L L R s as training material. In the test phase, the EASR takes the given LR image I l o w as input and outputs super-resolved images I H s with given scale factor s as follows:
I H s = E A S R ( I l o w )
The proposed EASR network is illustrated in Figure 3. EASR is a GAN-based architecture consisting of a generator and a discriminator.
The generator, which aims to reconstruct HR image I H s from given LR image I l o w with given scale factor s, consists of an initial reconstruction subnetwork and gradient-based enhancement subnetwork. The reconstruction process contains the following three steps:
The first step reconstructs an initial SR image I i n i t using the initial generative adversarial subnetwork composed of several residual blocks and a reconstruction layer as decoder for generating the intermediate HR result I i n i t , which thus helps to achieve overall performance for the reconstruction of I H s .
The second step focuses on the reconstruction of high-frequency information I e d g e + , which plays an important role in distinguishing the borders and edges of buildings in remotely sensed images. In this step, we first utilize gradient guidance operation to detect and extract gradient information from I i n i t , which is intuitively useful for better inferring the local intensity of sharpness. In addition, a frame branch and mask branch is utilized to extract fine edge maps from the gradient information. These two branches are utilized to learn the noise mask through the attention mechanism so that the network can focus on the real edge information to achieve the purpose of removing noises and artifacts. Specifically, the mask branch consists of three convolutional layers, which aims to adaptively learn specific weight matrices with soft attention to the relevant information. The frame branch contains several residual blocks to infer and extract the sharp edge information. Therefore, the gradient-based enhancement subnetwork reconstructs I e d g e + as follows:
I e d g e + = G E ( I i n i t )
where G E ( · ) denotes the mapping function of the gradient-based enhancement subnetwork, which consists of gradient calculator, frame branch, and masks branch. The enhancement subnetwork can reconstruct the edge while reducing the noises and maintaining sharpness.
The third step concatenates the initial SR image I i n i t and enhanced I e d g e + , and produces the final enhanced SR images I S R as follows:
I S R = I i n i t I e d g e + I e d g e +
while the generator module is dedicated to reconstructing an SR image that is similar to the ground truth HR image, and the discriminator module aims to distinguish the reconstructed SR images from ground truth HR images. For the discriminator, we take the architectural design in [46] as a reference but use the maximum pooling to replace the strided convolution.

3.3. Segmentation Network for Building Extraction

Using the reconstructed HR image I H s and corresponding building footprint label as training material, we train a dual-encoder segmentation module (DES) for building extraction in stage two. The proposed DES is a modified version of DlinkNet, which was firstly proposed by Zhou et al. [47] and proved to be effective in several recent studies [48,49,50]. The proposed DES contains two encoder submodules and one decoder submodule. As discussed above, the high-frequency information of reconstructed images can help define the building boundaries. Hence, we append an extra encoder module which takes the reconstructed high-frequency information I e d g e + as input to assist the segmentation module in distinguishing building area from background. The final building segmentation S e g is produced as follows:
S e g = D E S ( I H s , I e d g e + )
Each encoder of the proposed DES uses a ResNet-34 pre-trained weight on the ImageNet dataset as an initial parameter. In addition, we employ dilated convolution layers with dilation rates of 1, 2, 4, and 8 to improve the global and local representativeness of the buildings. The other submodule is the decoder of the segmentation network, which is in accordance with the decoder in U-net. The decoder uses transposed convolution layers to upscale the feature map to the same size as the size of input images.

3.4. Loss Function

In stage one, we utilize commonly used loss functions for SR methods, including reconstruction loss L r e c , adversarial loss L a d v [46], content loss L c o n t [51], and a total variation (TV) regularization L t v [52] to constrain the smoothness of I S R . The overall loss is defined as:
L t o t a l = L r e c + α L a d v + β L c o n t + γ L t v
where α , β , and γ denote the weights of each loss.
The reconstruction loss L r e c is utilized to preserve the consistency of image content between the super-resolved image I S R and HR image I L R , which is operated pixel-wise and defined as:
L r e c = ( I S R I L R ) 2
The content loss enforces the generator to generate an intermediate I S R image similar to I i n i t , which isoperated in pixel-wise and defined as:
L c o n t = ( I i n i t I L R ) 2
The adversarial loss aims to help the network to improve the perceptual quality of generated images. The adversarial loss of generator L a d v and discriminator L a d v D are defined as follows:
L a d v = l o g ( D ( G ( I L R ) ) )
L a d v D = l o g ( D ( I L R ) ) l o g ( 1 D ( I L R ) )
The total variation (TV) loss aims to constrain the smoothness of I S R , which is defined as:
L t v = | | ( I S R ) | | 2
where ( · ) denotes the gradient operator among the horizontal and vertical directions.
In stage two, we utilize commonly used binary cross entropy loss for the segmentation task.

4. Experiments

4.1. Study Area and Data

The study area contains the main city zone of three megacities in China, including Beijing, Shanghai, and Xi’an. The study area covers a total of approximately 1860 km 2 and contains multiple building types. As shown in Figure 4, the study areas cover variable types, forms, and shapes of buildings, including the most modern buildings in developed areas and factories under development, all of which make the selected areas representative and remarkable for this study.
As for datasets, versions of remotely sensed images of several resolutions are retrieved from Google Earth for the year 2018, including resolutions of 0.5 m, 1.0 m, 2.0 m, and 4.0 m, which are used as given material with scale ratio of 1, 2, 4, and 8 for each experimental case. For the training period in stage one, training and evaluation material were generated via segmenting the input images into patches of size 128 × 128 pixels. For the training period in the segmentation stage, images were cropped into 1024 × 1024 pixel patches. We extract a total of 36,000 group images with the size of 1024 × 1024 as our datasets. The datasets were further divided into training set, test dataset, and validation set according to 7:2:1 proportions.
With regard to the annotated dataset, the building footprint was annotated manually with the referenced remotely sensed images retrieved in 2018. The annotated dataset contains spatial coordinates of all annotated building polygons, and a further rasterization process was conducted in the QGIS platform to generate ground truth labels with the corresponding resolution for each baseline case. Note that a few mismatch cases are inevitable between annotation results and the actual ‘ground truth’ as a result of limitations in human-based interpretation and minor time inconsistency between retrieved images and referenced images for annotating buildings.

4.2. Implementation Details

Two experiments are conducted for verifying the effectiveness of our proposed two-stage SRBuildingSeg. In the first experiment, we compare the building footprint segmentation results of varied unsupervised SR approaches and the same segmentation approach. The other experiment compares the building extraction performance using the proposed SR methods and varied segmentation approach.
In the training phase, our method is implemented in PyTorch. All the networks in this paper are trained by a mini-batch stochastic gradient descent (SGD) with a momentum of 0.9 and the weight decay of 0.0005. The learning rate of the super-resolution stage is initialized as 0.001 and the learning rate of the segmentation stage is initialized as 0.0001. We utilize a reduced rate of 0.9 after every five epochs for both stages. Our network converges in 100 epochs in both stages, and the batch size is set to 5. An NVIDIA 2080Ti GPU is used for training.

4.3. The Effect of Super-Resolution in Building Extraction

In this section, we focus on comparing the effects contributed by the super-resolution stage in achieving super-resolution building extraction. Therefore, we train each DlinkNet [47] for the segmentation stage under the same conditions while using different SR methods in the super-resolution stage. Considering that the IPG module can help train the SR techniques in an unsupervised way, we select two unsupervised SR approaches (i.e., TSE [53], ZZSR [41]) as well as four supervised SR methods (i.e., SRGAN [46], EEGAN [51], waifu [54], DASR [55]) which are trained on the dataset generated by our proposed IPG module. All segmentation networks are trained under the same conditions using 0.5 m super-resolved images. According to the scale ratios in generating those SR images, our building segmentation experiment consists of three cases, including ratio ×2 (the resolution of LR images is 1 m), ratio ×4 (the resolution of LR images is 2 m), and ratio ×8 (the resolution of LR images is 4 m). Note that the case of using the bicubic interpolated remotely sensed images to train the segmentation model (BCI) is viewed as a baseline in this experiment.
Table 1 presents the quantitative evaluation of the super-resolved building extraction with the scale factors of 2, 4, and 8 by using those methods. Note that it presents the average results of all images collected for testing to provide a global view. All original high-resolution images in the test area are used for testing. According to the quantitative assessment reported in Table 1, the proposed EASR achieves better performance over other methods, which reveals that the EASR module helps to improve the representativeness of reconstructed images distinguished from the background. In comparison with unsupervised methods, the building extraction achieved via integration of the supervised SR method and proposed IPG module achieve better performance, indicating that our proposed IPG module offers an advantage in helping supervised SR methods to fully utilize the internal information of low resolution remotely sensed images, which can also be proven through qualitative evaluation of Figure 5. As shown in Figure 5, the proposed approach exhibits a great advantage in extracting borders and primary structures of buildings in remotely sensed images.
As the SR scale ratio enlarges, the IPG module contributes less improvement and even regresses in achieving super-resolution building extraction. The reason for this phenomenon is that the IPG module takes LR images as input to extract useful information for reconstructing edges and borders of super-resolved images. Meanwhile, as the SR scale ratio enlarges, the LR images tend to contain more noise and blurring, which makes it difficult to extract useful information.

4.4. The Effect of the Segmentation Module in Building Extraction

In this section, we compare our proposed DES module with six other state-of-the-art segmentation methods (i.e., Unet [56], DeepLabv3p [57], PSPNet [58], DlinkNet [47], UNet++ [59], and HRNet [60]). For a fair comparison, we train each segmentation network using the same super-resolved images reconstructed via our proposed EASR. For a fair comparison, all segmentation networks are trained under the same conditions using a 0.5 m super-resolved image.
We demonstrate the qualitative evaluation via visualizing the results between predicted results and ground-truth labels. As demonstrated in Figure 6, the segmentation results of the proposed DES can maintain the main structures and borders of buildings, while others fail to extract buildings, especially in a large scale factor of building extraction tasks (the ratio ×8 cases in Figure 6), which reveals that our approach can significantly improve feature representativeness in the process of building extraction, especially in the region of borders of buildings. Furthermore, our proposed approach shows its robustness in extracting buildings of variable density, height, textures, and forms, while the others result in the unclear contour of buildings (the ratio ×4 cases in Figure 6). The quantitative evaluation, as shown in Table 2, indicates that the proposed approach could achieve better performance over other methods with regard to IoU, recall, F1 score, overall accuracy, and kappa coefficient. This signifies that our proposed approach can enhance the comprehensive features and information of super-resolved remotely sensed images with an appropriate SR scale factor. It could be inferred from the segmentation details, as shown in Figure 6, that the extracted results outperform other methods with fewer false positive cases, especially in the vicinity of building boundaries.

5. Discussion

Since the above experimental results show the potential of the proposed approach in achieving high-resolution building extraction using low-resolution images, the mechanism and limits of the proposed approach necessitate further discussion. In this section, we primarily discuss two topics: (1) how enhanced high frequency information influences the super-resolution building extraction and (2) what the limits of the proposed approach are in conducting super-resolution building extraction tasks.

5.1. The Effectiveness of High Frequency Information in Building Extraction

As demonstrated in Table 1 and Table 2, it seems feasible to achieve high-resolution building extraction via integrating super-resolution and building segmentation methods. In comparison to the building extraction results using bicubic interpolated images, all SR-integrated methods achieve better performance with the scaling ratios of 2, 4, and 8. In addition, the details of reconstructed images using our approach are well-maintained in comparison to those of other integrated methods. A simple but vital question is: how does enhanced high frequency information influence the super-resolution building extraction?
As demonstrated in Figure 7, we visualized the penultimate CNN layer of DlinkNet (Figure 7a) and our proposed DES (Figure 7b). Specifically, we visualized the variance between the penultimate CNN layer of DlinkNet and DES (Figure 7c), which highlights the enhanced high frequency information in the segmentation task. Owing to the fusion of enhanced details and input images, it is clear that the features in edges and borders of buildings are well represented, resulting in better building extraction results. Meanwhile, the super-resolved images generated via other SR methods maintain better edges and borders in comparison with that of bicubic interpolated images, thus resulting in better performance. This reveals that edges and borders of buildings serve pivotal roles in building segmentation from remotely sensed images.

5.2. The Limitations of the Proposed Approach

Since we can generate high-resolution building extraction results with acceptable accuracy using training material with resolution around 4 m, it seems theoretically possible that we can use even lower resolution material, such as 6 m, to achieve high-resolution building extraction results following the same approach. As shown in the experimental results of remotely sensed image SR and building extraction, the improvements contributed by SR methods rapidly decline as the ratio factor enlarges, especially in the cases with a ratio factor of 8. This is primarily attributed to the unsatisfactory reconstruction results obtained for the edges and borders. Nevertheless, whether there is a limit scaling ratio in conducting high-resolution building extraction using low-resolution material remains to be determined.
Furthermore, we demonstrate the proposed approach in conducting high-resolution building extraction with ratio factors of 2, 4, 8, and 12. Note that the training material with a ratio factor of 12 was generated from OHGT using bilinear interpolation. As shown in Table 3 and Figure 8, two shortages emerge as the ratio factor of super-resolution in building extraction continually increases, which could lead to the theoretical ratio limits of our proposed approach. On one hand, the higher the ratio is, the harder the training process becomes. The training pairs’ generation module first downscales the given input images, after which much coarser images are generated for the input of the SR module, resulting in a remotely sensed image without sufficient important details for subsequent reconstruction. The strength in providing finer details regarding output images based on the proposed SR module becomes weaker as the scale ratio of the building extraction task enlarges. On the other hand, the improvement contributed by our enhancement module leads to worse performance as the scale ratio of the building extraction task enlarges, which results from the insufficiency of details retrieved from the given low-resolution images. As demonstrated in Figure 8, the high frequency information reconstructed via the proposed EASR becomes coarse, which may even lead to a few artifacts. This indicates that the proposed approach is reaching its limits in conducting the building extraction task at the scale ratio of 12.

6. Conclusions

In this study, we propose a novel two-stage framework (SRBuildingSeg) to achieve super-resolution building extraction using relatively low-resolution remotely sensed images. SRBuildingSeg can fully utilize inherent information of the given low-resolution images to achieve relatively high-resolution building extraction. For generating LR-HR training pairs, we propose an internal pairs generation module (IPG) with no need for external high-resolution images, which can reconstruct super-resolved images with only the given low-resolution images. The edge-aware super-resolution (EASR) module then generates super-resolved images at the desired higher resolution, after which the super-resolution building extraction result is obtained using the dual-encoder building segmentation module (DES). The experimental results demonstrate the capability of the proposed approach in achieving super-resolution building extraction, which outperforms other methods in terms of both the perceptual quality of the super-resolved remotely sensed image and the building extraction accuracy for all small (×2), middle (×4), and large (×8) scale ratios. Furthermore, we demonstrate how the reconstructed high frequency information affects the subsequent building extraction. The assessment results reveal that our proposed approach ranks the best among all SR-integrated methods. In summary, we present the potential of the proposed straightforward approach in demonstrating the use of widely available low-resolution resolution data to obtain high-resolution building extraction results. This approach is practical and especially useful when extra datasets of high-resolution remotely sensed images are unavailable.

Author Contributions

Conceptualization, L.Z.; Data curation, L.Z. and R.D.; Formal analysis, L.Z.; Funding acquisition, H.F.; Investigation, L.Z.; Methodology, L.Z. and R.D.; Software, L.Z.; Supervision, H.F.; Validation, L.Z.; Visualization, L.Z.; Writing—original draft, L.Z.; Writing—review and editing, L.Z., R.D., S.Y., W.L., J.Z. and H.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by R&D Program of China (Grant No. 2017YFA0604500) and National Natural Science Foundation of China (U1839206).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The code of this study is available at https://github.com/xian1234/SRBuildSeg (accessed on 4 June 2021). The data presented in this study are available on request from the corresponding author.

Acknowledgments

We appreciate acdemic editor and the anonymous reviewers for their thoughtful and constructive suggestions and comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Feng, T.; Zhao, J. Review and comparison: Building extraction methods using high-resolution images. In Proceedings of the 2009 Second International Symposium on Information Science and Engineering, Shanghai, China, 26–28 December 2009; pp. 419–422. [Google Scholar] [CrossRef]
  2. Huang, X.; Zhang, L. A multidirectional and multiscale morphological index for automatic building extraction from multispectral GeoEye-1 imagery. Photogramm. Remote Sens. 2011, 77, 721–732. [Google Scholar] [CrossRef]
  3. Huang, Z.; Cheng, G.; Wang, H.; Li, H.; Shi, L.; Pan, C. Building extraction from multi-source remote sensing images via deep deconvolution neural networks. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 1835–1838. [Google Scholar] [CrossRef]
  4. Rottensteiner, F.; Briese, C. A new method for building extraction in urban areas from high-resolution LIDAR data. In International Archives of Photogrammetry Remote Sensing and Spatial Information Sciences; Natural Resources Canada: Ottawa, ON, Canada, 2020; Volume 34, pp. 295–301. [Google Scholar]
  5. Liu, F.; Jiao, L.; Hou, B.; Yang, S. POL-SAR Image Classification Based on Wishart DBN and Local Spatial Information. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3292–3308. [Google Scholar] [CrossRef]
  6. Wang, J.; Song, J.; Chen, M.; Yang, Z. Road network extraction: A neural-dynamic framework based on deep learning and a finite state machine. Int. J. Remote Sens. 2015, 36, 3144–3169. [Google Scholar] [CrossRef]
  7. Ghanea, M.; Moallem, P.; Momeni, M. Building extraction from high-resolution satellite images in urban areas: Recent methods and strategies against significant challenges. Int. J. Remote Sens. 2016, 37, 5234–5248. [Google Scholar] [CrossRef]
  8. Huang, J.; Zhang, X.; Xin, Q.; Sun, Y.; Zhang, P. Automatic building extraction from high-resolution aerial images and LiDAR data using gated residual refinement network. ISPRS J. Photogramm. Remote Sens. 2019, 151, 91–105. [Google Scholar] [CrossRef]
  9. Feng, W.; Sui, H.; Hua, L.; Xu, C.; Ma, G.; Huang, W. Building extraction from VHR remote sensing imagery by combining an improved deep convolutional encoder-decoder architecture and historical land use vector map. Int. J. Remote Sens. 2020, 41, 6595–6617. [Google Scholar] [CrossRef]
  10. Belgiu, M.; Dr Gut, L. Comparing supervised and unsupervised multiresolution segmentation approaches for extracting buildings from very high resolution imagery. ISPRS J. Photogramm. Remote Sens. 2014, 96, 67–75. [Google Scholar] [CrossRef] [Green Version]
  11. Huang, H.; Wu, B.; Fan, J. Analysis to the relationship of classification accuracy, segmentation scale, image resolution. In Proceedings of the IGARSS 2003, IEEE International Geoscience and Remote Sensing Symposium, Proceedings (IEEE Cat. No. 03CH37477), Toulouse, France, 21–25 July 2003; Volume 6, pp. 3671–3673. [Google Scholar] [CrossRef]
  12. Hamaguchi, R.; Hikosaka, S. Building detection from satellite imagery using ensemble of size-specific detectors. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 223–2234. [Google Scholar] [CrossRef]
  13. Haut, J.M.; Paoletti, M.E.; Fernandez-Beltran, R.; Plaza, J.; Plaza, A.; Li, J. Remote Sensing Single-Image Superresolution Based on a Deep Compendium Model. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1432–1436. [Google Scholar] [CrossRef]
  14. Jozdani, S.; Chen, D. On the versatility of popular and recently proposed supervised evaluation metrics for segmentation quality of remotely sensed images: An experimental case study of building extraction. ISPRS J. Photogramm. Remote Sens. 2020, 160, 275–290. [Google Scholar] [CrossRef]
  15. Na, Y.; Kim, J.H.; Lee, K.; Park, J.; Hwang, J.Y.; Choi, J.P. Domain Adaptive Transfer Attack-Based Segmentation Networks for Building Extraction From Aerial Images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5171–5182. [Google Scholar] [CrossRef]
  16. Zhang, L.; Wu, J.; Fan, Y.; Gao, H.; Shao, Y. An Efficient Building Extraction Method from High Spatial Resolution Remote Sensing Images Based on Improved Mask R-CNN. Sensors 2020, 20, 1465. [Google Scholar] [CrossRef] [Green Version]
  17. Bagan, H.; Yamagata, Y. Landsat analysis of urban growth: How Tokyo became the world’s largest megacity during the last 40 years. Remote Sens. Environ. 2012, 127, 210–222. [Google Scholar] [CrossRef]
  18. Dong, L.; Shan, J. A comprehensive review of earthquake-induced building damage detection with remote sensing techniques. ISPRS J. Photogramm. Remote Sens. 2013, 84, 85–99. [Google Scholar] [CrossRef]
  19. Weng, Q. Remote sensing of impervious surfaces in the urban areas: Requirements, methods, and trends. Remote Sens. Environ. 2012, 117, 34–49. [Google Scholar] [CrossRef]
  20. Chen, B.; Xu, B.; Zhu, Z.; Yuan, C.; Suen, H.P.; Guo, J.; Xu, N.; Li, W.; Zhao, Y.; Yang, J. Stable classification with limited sample: Transferring a 30-m resolution sample set collected in 2015 to mapping 10-m resolution global land cover in 2017. Sci. Bull. 2019, 64, 370–373. [Google Scholar]
  21. Shrivastava, N.; Rai, P.K. Automatic building extraction based on multiresolution segmentation using remote sensing data. Geogr. Pol. 2015, 88, 407–421. [Google Scholar] [CrossRef] [Green Version]
  22. Wang, J.; Qin, Q.; Yang, X.; Wang, J.; Ye, X.; Qin, X. Automated road extraction from multi-resolution images using spectral information and texture. In Proceedings of the 2014 IEEE Geoscience and Remote Sensing Symposium, Quebec City, QC, Canada, 13–18 July 2014; pp. 533–536. [Google Scholar] [CrossRef]
  23. Deng, C.; Zhu, Z. Continuous subpixel monitoring of urban impervious surface using Landsat time series. Remote Sens. Environ. 2020, 238, 110929. [Google Scholar] [CrossRef]
  24. Fu, Y.; Liu, K.; Shen, Z.; Deng, J.; Gan, M.; Liu, X.; Lu, D.; Wang, K. Mapping impervious surfaces in town–rural transition belts using China’s GF-2 imagery and object-based deep CNNs. Remote Sens. 2019, 11, 280. [Google Scholar] [CrossRef] [Green Version]
  25. Gong, P.; Li, X.; Zhang, W. 40-Year (1978–2017) human settlement changes in China reflected by impervious surfaces from satellite remote sensing. Sci. Bull. 2019, 64, 756–763. [Google Scholar] [CrossRef] [Green Version]
  26. Zhang, Y.; Zhao, D.; Zhang, J.; Xiong, R.; Gao, W. Interpolation-dependent image downsampling. IEEE Trans. Image Process. 2011, 20, 3291–3296. [Google Scholar] [CrossRef]
  27. Thévenaz, P.; Blu, T.; Unser, M. Image interpolation and resampling. Handb. Med. Imaging Process. Anal. 2000, 1, 393–420. [Google Scholar]
  28. Park, S.C.; Park, M.K.; Kang, M.G. Super-resolution image reconstruction: A technical overview. IEEE Signal Process. Mag. 2003, 20, 21–36. [Google Scholar] [CrossRef] [Green Version]
  29. Haut, J.M.; Fernandez-Beltran, R.; Paoletti, M.E.; Plaza, J.; Plaza, A. Remote Sensing Image Superresolution Using Deep Residual Channel Attention. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9277–9289. [Google Scholar] [CrossRef]
  30. Shao, Z.; Wang, L.; Wang, Z.; Deng, J. Remote Sensing Image Super-Resolution Using Sparse Representation and Coupled Sparse Autoencoder. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 2663–2674. [Google Scholar] [CrossRef]
  31. Paisitkriangkrai, S.; Sherrah, J.; Janney, P.; Hengel, V.D. Effective semantic pixel labelling with convolutional networks and conditional random fields. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 36–43. [Google Scholar] [CrossRef]
  32. Lin, J.; Jing, W.; Song, H.; Chen, G. ESFNet: Efficient Network for Building Extraction From High-Resolution Aerial Images. IEEE Access 2019, 7, 54285–54294. [Google Scholar] [CrossRef]
  33. Mou, L.; Bruzzone, L.; Zhu, X.X. Learning Spectral-Spatial-Temporal Features via a Recurrent Convolutional Neural Network for Change Detection in Multispectral Imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 924–935. [Google Scholar] [CrossRef] [Green Version]
  34. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef] [Green Version]
  35. Schuegraf, P.; Bittner, K. Automatic Building Footprint Extraction from Multi-Resolution Remote Sensing Images Using a Hybrid FCN. ISPRS Int. J. Geo-Inf. 2019, 8, 191. [Google Scholar] [CrossRef] [Green Version]
  36. Guo, Z.; Wu, G.; Song, X.; Yuan, W.; Chen, Q.; Zhang, H.; Shi, X.; Xu, M.; Xu, Y.; Shibasaki, R.; et al. Super-Resolution Integrated Building Semantic Segmentation for Multi-Source Remote Sensing Imagery. IEEE Access 2019, 7, 99381–99397. [Google Scholar] [CrossRef]
  37. Dong, X.; Sun, X.; Jia, X.; Xi, Z.; Gao, L.; Zhang, B. Remote Sensing Image Super-Resolution Using Novel Dense-Sampling Networks. IEEE Trans. Geosci. Remote Sens. 2021, 59, 1618–1633. [Google Scholar] [CrossRef]
  38. Zhang, D.; Shao, J.; Li, X.; Shen, H.T. Remote Sensing Image Super-Resolution via Mixed High-Order Attention Network. IEEE Trans. Geosci. Remote Sens. 2020, 1–14. [Google Scholar] [CrossRef]
  39. Zou, F.; Xiao, W.; Ji, W.; He, K.; Yang, Z.; Song, J.; Zhou, H.; Li, K. Arbitrary-oriented object detection via dense feature fusion and attention model for remote sensing super-resolution image. Neural Comput. Appl. 2020, 32, 14549–14562. [Google Scholar] [CrossRef]
  40. Yang, D.; Li, Z.; Xia, Y.; Chen, Z. Remote sensing image super-resolution: Challenges and approaches. In Proceedings of the 2015 IEEE International Conference on Digital Signal Processing (DSP), Singapore, 21–24 July 2015; pp. 196–200. [Google Scholar] [CrossRef]
  41. Shocher, A.; Cohen, N.; Irani, M. “zero-shot” super-resolution using deep internal learning. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3118–3126. [Google Scholar] [CrossRef] [Green Version]
  42. Haut, J.M.; Fernandez-Beltran, R.; Paoletti, M.E.; Plaza, J.; Plaza, A.; Pla, F. A New Deep Generative Network for Unsupervised Remote Sensing Single-Image Super-Resolution. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6792–6810. [Google Scholar] [CrossRef]
  43. Chen, S.; Han, Z.; Dai, E.; Jia, X.; Liu, Z.; Xing, L.; Zou, X.; Xu, C.; Liu, J.; Tian, Q. Unsupervised image super-resolution with an indirect supervised path. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 468–469. [Google Scholar] [CrossRef]
  44. Lugmayr, A.; Danelljan, M.; Timofte, R. Unsupervised learning for real-world super-resolution. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27–28 October 2019; pp. 3408–3416. [Google Scholar] [CrossRef] [Green Version]
  45. Socher, R.; Ganjoo, M.; Manning, C.D.; Ng, A. Zero-shot learning through cross-modal transfer. arXiv 2018, arXiv:1301.3666. [Google Scholar]
  46. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar] [CrossRef] [Green Version]
  47. Zhou, L.; Zhang, C.; Wu, M. D-LinkNet: LinkNet With Pretrained Encoder and Dilated Convolution for High Resolution Satellite Imagery Road Extraction. In Proceedings of the 2018 CVPR Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 182–186. [Google Scholar] [CrossRef]
  48. Abdollahi, A.; Pradhan, B.; Shukla, N.; Chakraborty, S.; Alamri, A. Deep Learning Approaches Applied to Remote Sensing Datasets for Road Extraction: A State-Of-The-Art Review. Remote Sens. 2020, 12, 1444. [Google Scholar] [CrossRef]
  49. Jiang, Z.; Chen, Z.; Ji, K.; Yang, J. Semantic segmentation network combined with edge detection for building extraction in remote sensing images. In Proceedings of the MIPPR 2019: Pattern Recognition and Computer Vision, International Society for Optics and Photonics, Xi’an, China, 8–11 November 2019; Volume 11430, p. 114300. [Google Scholar] [CrossRef]
  50. Zhang, Z.; Wang, Y. JointNet: A Common Neural Network for Road and Building Extraction. Remote Sens. 2019, 11, 696. [Google Scholar] [CrossRef] [Green Version]
  51. Jiang, K.; Wang, Z.; Yi, P.; Wang, G.; Lu, T.; Jiang, J. Edge-Enhanced GAN for Remote Sensing Image Superresolution. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5799–5812. [Google Scholar] [CrossRef]
  52. Aly, H.A.; Dubois, E. Image up-sampling using total-variation regularization with a new observation model. IEEE Trans. Image Process. 2005, 14, 1647–1659. [Google Scholar] [CrossRef] [PubMed]
  53. Huang, J.B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar] [CrossRef]
  54. Nagadomi. Waifu: Image Super-Resolution. [EB/OL]. Available online: http://waifu2x.udp.jp/ (accessed on 4 June 2021).
  55. Wang, L.; Wang, Y.; Dong, X.; Xu, Q.; Yang, J.; An, W.; Guo, Y. Unsupervised Degradation Representation Learning for Blind Super-Resolution. arXiv 2021, arXiv:2104.00416. [Google Scholar]
  56. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the 2015 International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar] [CrossRef]
  57. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar] [CrossRef] [Green Version]
  58. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar] [CrossRef] [Green Version]
  59. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: Redesigning Skip Connections to Exploit Multiscale Features in Image Segmentation. IEEE Trans. Med. Imaging 2019, 36, 1856–1867. [Google Scholar] [CrossRef] [Green Version]
  60. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep High-Resolution Representation Learning for Human Pose Estimation. arXiv 2019, arXiv:1902.09212. [Google Scholar]
Figure 1. The overall workflow of proposed two-stage building extraction using the LR images approach. Stage one, which consists of internal pairs generation module (IPG) and edge-aware super-resolution module (EASR), focuses on image super resolution. Stage two focuses on building extraction, which contains dual-encoder building segmentation module (DES).
Figure 1. The overall workflow of proposed two-stage building extraction using the LR images approach. Stage one, which consists of internal pairs generation module (IPG) and edge-aware super-resolution module (EASR), focuses on image super resolution. Stage two focuses on building extraction, which contains dual-encoder building segmentation module (DES).
Remotesensing 13 02872 g001
Figure 2. An example workflow of the proposed LR-HR training pairs generation with a scaling factor of 4. In this example, I l o w represents the raw remote sensing images (2 m), which usually consists one or two images with a large size and covering a large area. I L R represents HR patches (2 m), which is generated via cropping and splitting the I l o w . For each image patch in I L R , we generate its downscaled version, which consists the corresponding LR training pairs I L L R s (8 m). Several dataset enrichment schemes were applied to the LR-HR training pairs. After the super-resolution model was properly trained on the LR-HR training pairs, the trained SR model takes the given images I l o w (2 m) as input to generate super-resolved images I H s (0.5 m).
Figure 2. An example workflow of the proposed LR-HR training pairs generation with a scaling factor of 4. In this example, I l o w represents the raw remote sensing images (2 m), which usually consists one or two images with a large size and covering a large area. I L R represents HR patches (2 m), which is generated via cropping and splitting the I l o w . For each image patch in I L R , we generate its downscaled version, which consists the corresponding LR training pairs I L L R s (8 m). Several dataset enrichment schemes were applied to the LR-HR training pairs. After the super-resolution model was properly trained on the LR-HR training pairs, the trained SR model takes the given images I l o w (2 m) as input to generate super-resolved images I H s (0.5 m).
Remotesensing 13 02872 g002
Figure 3. The architectures of the proposed EASR module.
Figure 3. The architectures of the proposed EASR module.
Remotesensing 13 02872 g003
Figure 4. Examples of buildings with various shapes, types, and forms in the study area.
Figure 4. Examples of buildings with various shapes, types, and forms in the study area.
Remotesensing 13 02872 g004
Figure 5. Qualitative examples of segmentation results of each method with ratio factors of 2, 4, and 8.
Figure 5. Qualitative examples of segmentation results of each method with ratio factors of 2, 4, and 8.
Remotesensing 13 02872 g005
Figure 6. Qualitative examples of segmentation results of each method with ratio factors of 2, 4, and 8.
Figure 6. Qualitative examples of segmentation results of each method with ratio factors of 2, 4, and 8.
Remotesensing 13 02872 g006
Figure 7. The visualizations of the penultimate CNN layer (better visualized in color). (a) is the input remote sensing images; (b) is the visualization results of DlinkNet; (c) is the visualization results of DES; (d) is the variance between (b,c), which denotes the information enhanced in the segmentation task.
Figure 7. The visualizations of the penultimate CNN layer (better visualized in color). (a) is the input remote sensing images; (b) is the visualization results of DlinkNet; (c) is the visualization results of DES; (d) is the variance between (b,c), which denotes the information enhanced in the segmentation task.
Remotesensing 13 02872 g007
Figure 8. Qualitative assessment of the improvements contributed by the enhancement module with scale ratios of 2 (column I), 4 (column II), 8 (column III), and 12 (column IV).
Figure 8. Qualitative assessment of the improvements contributed by the enhancement module with scale ratios of 2 (column I), 4 (column II), 8 (column III), and 12 (column IV).
Remotesensing 13 02872 g008
Table 1. Quantitative evaluation results on the super-resolution stage.
Table 1. Quantitative evaluation results on the super-resolution stage.
Stage OneStage TwoScaleIoUPrecisionRecallF1 ScoreKappa
BCIDlinkNet20.62060.74200.79140.76590.6684
TSE0.64030.77200.78960.78070.7157
ZZSR0.64920.77600.79890.78730.7066
SRGAN *0.65380.76380.81950.79070.7537
EEGAN *0.66330.78050.79860.79760.7438
waifu0.64480.78070.76270.78370.7235
DASR *0.65300.78740.78340.79220.7586
EASR(ours)0.67210.79650.82880.80390.7771
BCIDlinkNet40.60690.74990.72210.75540.6069
TSE0.61840.76930.75920.76420.6969
ZZSR0.62240.77860.75630.76730.6972
SRGAN *0.62630.77870.78420.77020.7160
EEGAN *0.63360.78750.76430.77570.6940
waifu0.62200.77690.76120.76300.6863
DASR *0.63910.78140.78500.78890.7234
EASR(ours)0.64130.79190.79160.78140.7361
BCIDlinkNet80.56160.72060.71800.71930.6152
TSE0.58220.69780.77850.73590.6866
ZZSR0.59250.73660.75180.74410.6906
SRGAN *0.58630.72250.75680.73920.6589
EEGAN *0.60600.76470.73390.75470.7041
waifu0.60780.70010.73720.73560.6534
DASR *0.59080.73080.75520.74280.6691
EASR(ours)0.62370.78180.75510.76820.7214
* indicates that this supervised SR method is trained on datasets generated by our proposed IPG module.
Table 2. Quantitative evaluation results on the segmentation stage.
Table 2. Quantitative evaluation results on the segmentation stage.
Stage OneStage TwoScaleIoUPrecisionRecallF1 ScoreKappa
EASR(ours)Unet20.60810.75460.75800.75630.6654
DeepLabv3p0.65360.81620.76640.79050.7160
PSPNet0.65220.79910.78010.78950.7125
DlinkNet0.67210.78050.82880.80390.6684
UNet++0.64310.80660.76140.78460.7270
HRNet0.69860.81030.80350.80190.7614
DES(ours)0.70700.82650.83050.82780.7761
EASR(ours)Unet40.58890.70370.78300.74130.6585
DeepLabv3p0.63360.77160.76430.77570.6940
PSPNet0.63850.76460.79470.77940.6296
DlinkNet0.64130.77870.78420.78140.7206
UNet++0.62310.74730.78080.76860.6772
HRNet0.63010.78140.77110.77000.6963
DES(ours)0.65950.78750.81950.79480.7361
EASR(ours)Unet80.54140.74040.66830.70250.6045
DeepLabv3p0.61550.78490.73130.76200.6786
PSPNet0.63170.78540.76340.77430.6974
DlinkNet0.62370.78180.75510.76820.7214
UNet++0.61210.76320.75840.76130.7036
HRNet0.62750.79760.73210.76350.7172
DES(ours)0.63460.79550.76820.77650.7310
Table 3. Quantitative evaluation of the results of building extraction.
Table 3. Quantitative evaluation of the results of building extraction.
MethodsScaleIoUPrecisionRecallF1 ScoreKappa
BCI20.62060.74200.79140.76590.6684
Ours0.70700.82650.83050.82780.7761
BCI40.60690.74990.72210.75540.6069
Ours0.65950.78750.81950.79480.7361
BCI80.56160.72060.71800.71930.6152
Ours0.63460.79550.76820.77650.7310
BCI120.52790.72700.65840.69100.5837
Ours0.54140.74040.66830.70250.6045
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, L.; Dong, R.; Yuan, S.; Li, W.; Zheng, J.; Fu, H. Making Low-Resolution Satellite Images Reborn: A Deep Learning Approach for Super-Resolution Building Extraction. Remote Sens. 2021, 13, 2872. https://doi.org/10.3390/rs13152872

AMA Style

Zhang L, Dong R, Yuan S, Li W, Zheng J, Fu H. Making Low-Resolution Satellite Images Reborn: A Deep Learning Approach for Super-Resolution Building Extraction. Remote Sensing. 2021; 13(15):2872. https://doi.org/10.3390/rs13152872

Chicago/Turabian Style

Zhang, Lixian, Runmin Dong, Shuai Yuan, Weijia Li, Juepeng Zheng, and Haohuan Fu. 2021. "Making Low-Resolution Satellite Images Reborn: A Deep Learning Approach for Super-Resolution Building Extraction" Remote Sensing 13, no. 15: 2872. https://doi.org/10.3390/rs13152872

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop