Next Article in Journal
Neural Network Based Kalman Filters for the Spatio-Temporal Interpolation of Satellite-Derived Sea Surface Temperature
Next Article in Special Issue
Accuracy Assessment on MODIS (V006), GLASS and MuSyQ Land-Surface Albedo Products: A Case Study in the Heihe River Basin, China
Previous Article in Journal
Quantitative Assessment of Desertification in an Arid Oasis Using Remote Sensing Data and Spectral Index Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Boosting Learning Saliency Method for Built-Up Areas Extraction in Sentinel-2 Images

1
Institute of Remote Sensing and Digital Earth, Chinese Academy of Sciences, Beijing 100101, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Sanya Institute of Remote Sensing, Hainan 572029, China
4
School of Sciences, Qiqihar University, Qiqihar 161006, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2018, 10(12), 1863; https://doi.org/10.3390/rs10121863
Submission received: 14 September 2018 / Revised: 19 November 2018 / Accepted: 20 November 2018 / Published: 22 November 2018
(This article belongs to the Special Issue Advanced Topics in Remote Sensing)

Abstract

:
Built-up areas extraction from satellite images is an important aspect of urban planning and land use; however, this remains a challenging task when using optical satellite images. Existing methods may be limited because of the complex background. In this paper, an improved boosting learning saliency method for built-up area extraction from Sentinel-2 images is proposed. First, the optimal band combination for extracting such areas from Sentinel-2 data is determined; then, a coarse saliency map is generated, based on multiple cues and the geodesic weighted Bayesian (GWB) model, that provides training samples for a strong model; a refined saliency map is subsequently obtained using the strong model. Furthermore, cuboid cellular automata (CCA) is used to integrate multiscale saliency maps for improving the refined saliency map. Then, coarse and refined saliency maps are synthesized to create a final saliency map. Finally, the fractional-order Darwinian particle swarm optimization algorithm (FODPSO) is employed to extract the built-up areas from the final saliency result. Cities in five different types of ecosystems in China (desert, coastal, riverside, valley, and plain) are used to evaluate the proposed method. Analyses of results and comparative analyses with other methods suggest that the proposed method is robust, with good accuracy.

1. Introduction

Population density and resource utilization intensity tend to be very high in built-up areas. Rapid urbanization has resulted in several problems, including the urban heat island effect, air pollution, and unreasonable land use. Therefore, extracting built-up areas is a major topic of interest across numerous fields, including sustainability, remote sensing, and the social sciences. To efficiently distribute information regarding built-up areas to various research disciplines, remote sensing technology is widely used to extract and monitor these areas. The term “built-up areas” is widely used in the literature, and refers to the spatial extent of urbanized areas on a regional scale, but this is a nebulous and inconsistent definition [1]. For the purpose of this study, built-up areas are defined as areas dominated by buildings, streets, and impervious surfaces; golf courses, green urban parks, sparse buildings in suburbs, and rural settlements are not included within this working definition.
Over the last few decades, many methods for extracting built-up areas have been proposed [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15]. These methods can be broadly categorized into four groups: classification-based, index-based, texture-based, and multisensor-based methods. Classification-based methods [1,2,3] primarily consider how suitable classifiers may be used to extract built-up areas. However, these methods are fraught with challenges and limitations when applied at regional and global scales; for example, scene-to-scene data analyses are subjective, while the overall process is time-consuming, and entails complicated computing [4]. Index-based methods [4,5,6,7,8,9] are designed based on spectral bands within which built-up areas exhibit their highest and lowest reflectance values amidst a multispectral dataset. Usually, these methods are unsuccessful in addressing the difficulties in distinguishing between built-up areas and other land cover types. Texture-based methods [10,11] can extract built-up areas based on high texture granularity. However, such methods may fail when ground objects with similar texture features to the built-up areas are encountered. Multisensor-based methods [1,12,13,14] combine the various characteristics of multiple sensors to extract built-up areas. However, owing to difficulties associated with synthesizing different data types, such methods have not been widely used [15].
Saliency detection provides a unique perspective for ground object extraction, because it selects only interesting information related to the current behavior or task to be processed, ignoring irrelevant information [16]. Saliency detection methods have been gradually introduced into the field of remote sensing, in recent years, to detect specific objects [17,18,19] in very high spatial resolution images, or high spatial resolution images, and they have proven to be effective. The Sentinel-2 satellites have a multispectral instrument with 13 spectral bands, and its richer spectrum ensures superior detection and extraction of built-up areas. Due to the unique spectral characteristics of the built-up areas, they can be highlighted in some band combinations, so the built-up areas can easily be identified by saliency detection methods. The saliency detection methods can be mainly categorized as bottom-up and top-down methods. Numerous bottom-up saliency detection methods have been proposed, and these can be broadly divided into four groups: those based on contrast, graph theory, and information theory, prior knowledge, and low-rank matrix recovery theory.
Contrast-based methods consist of local contrast and global contrast methods. The former is used to investigate the rarity of image regions with respect to nearby neighborhoods; the classic saliency method proposed by Itti et al. [20] is a typical local contrast method. Since its proposal, several approaches have also adopted the center-surround contrast strategy to calculate saliency, including the graph-based visual saliency detection method [21], the fuzzy growing method [22], and the discriminant center-surround hypothesis model [23,24], and so on. The global contrast methods calculate saliency using the contrast of a pixel or region with respect to the entire image. Cheng et al. [25] proposed a regional contrast-based saliency extraction algorithm that simultaneously evaluates global contrast differences and spatial coherence. Perazzi et al. [26] developed a contrast-based filtering method, and Shi et al. [27] proposed a generic and fast computational framework called pixelwise image saliency aggregating (PISA), that uses prior spatial information to weight color contrast, and direction contrast to generate a final saliency map.
Graph theory-based methods firstly use a graph model to represent an image, and then apply an established undirected or directed graph to predict the saliency value of each region. Gopalakrishnan et al. [28] performed random walks on graphs to detect salient objects. Wei et al. [29] proposed a model based on the geodesic method. Jiang et al. [30] proposed a saliency detection method using absorbing Markov chains. Yan et al. [30,31] applied a hierarchical model to analyze saliency cues from multiple levels of structure that were then integrated to infer a final saliency map. Qin et al. [32] introduced cellular automata to intuitively detect salient objects based on a dynamic evolution model. Based on information theory-based model, Bruce and Tsotsos [33] used Shannon’s self-information to measure saliency, and proposed a visual attention model based on information maximization. Zhang et al. [34] used the self-information of local features in an image to represent object rarity, and then measured the saliency value.
Prior knowledge-based methods improve the accuracy of saliency detection methods via the general image properties that researchers have summarized through observation or experiment. Common prior knowledge types include center, background, objectness, semantic, color, spatial distribution, and sparse priors, with the center and background priors particularly widely used. Several studies [26,35,36] employed the center prior, and demonstrated that it can enhance model performance. The background prior has also been used in several studies [37,38].
Matrix recovery-based methods, aiming at decomposing a matrix into a low-rank matrix and a sparse one, have shown the potential to address the problem of saliency detection, where the decomposed low-rank matrix naturally corresponds to the background, and the sparse one captures salient objects. Peng et al. [39] introduced sparse structures into the rotated principal component analysis (RPCA) model, and proposed a saliency detection method based on low-rank representation and structural sparse matrix decomposition. Lang et al. [40] proposed a multitask sparse pursuit algorithm based on low-rank representation.
Compared to bottom-up methods, little research has hitherto been conducted regarding the top-down saliency model. Jiang et al. [41] proposed a learning-based method by regarding saliency detection as a regression problem, where the saliency detection model was constructed based on the integration of numerous descriptors extracted from training samples with ground truth labels. Zhang et al. [34] integrated the top-down and bottom-up information to construct a Bayesian-based top-down model, where saliency is computed locally. Yang et al. [42] proposed a method combining conditional random field and sparse coding theory. Cholakkal et al. [43] regarded top-down saliency detection as an image-classification problem, and proposed a saliency detection method based on an image-classification framework.
As each group has different advantages, Tong et al. [44] proposed a bootstrap learning (BL) method to enhance performance; it exploits the strength of both bottom-up contrast-based saliency methods and top-down learning methods. However, more research is needed if the BL method is used to extract the built-up area. First, BL introduces a dark channel prior into the coarse saliency detection model to generate a coarse saliency map, but this prior is not suitable for all images. In images of darker backgrounds or brighter foregrounds, it may produce the opposite effect. Although the authors used adaptive weights to attenuate the inverse effects of the dark channel prior, remote sensing images are highly complex, especially where water bodies occur as dark background; in such cases, the BL algorithm may fail. Second, BL does not take into account the spatial information of ground objects, which may result in the detection of large amounts of background information. In addition, it simply superimposes multiscale saliency maps without fully integrating the information they provide.
In this study, an improved boosting learning saliency method for extracting built-up areas from remote sensing images is proposed. First, we determine the optimal band combination for extracting built-up areas. To overcome the shortcomings associated with a dark channel prior, we introduce a multi-cue fusion and water removal strategy into the coarse saliency model, to improve the accuracy of the coarse saliency map. Then, the GWB model [45] is used to consider spatial information, thereby eliminating the impact of land cover surrounding built-up areas, to further improve accuracy and provide reliable training samples for the strong saliency model. After that, the CCA [46] is employed to effectively integrate multiscale saliency maps to optimize the accuracy of the final saliency map. Finally, FODPSO algorithm [47] is used to segment the final saliency map to accurately capture information on built-up areas.
The contribution of this paper is threefold: (1) We improve the BL saliency method based on the characteristics of the remote sensing image for extracting the built-up areas. (2) GWB and CCA are introduced in the proposed method to suppress background regions and attach more importance to regions which are more likely to be parts of built-up areas. (3) We determine the optimal band combination of Sentinel-2 for built-up areas detection.
The rest of this paper is organized as follows: the proposed method is illustrated in Section 2, Section 3 focuses on the experimental results, and Section 4 and Section 5 provide the discussion and conclusions, respectively.

2. Proposed Method

A flowchart of the proposed method is shown in Figure 1. The method consists of four stages. First, the image is sharpened to 10 m, and the optimal band combination for the built-up area extraction is determined. Then, the false color image generated by optimal band combination is segmented into a group of segmented objects. Subsequently, a coarse saliency map is constructed based on multiple clues fusion, and GWB, to generate training samples for a strong model. Based on the representation of three features, a strong classifier is trained to measure saliency. Next, the coarse and refined saliency maps are weighted, in combination, to generate the final saliency map. Finally, the built-up areas are extracted using the FODPSO method.

2.1. Image Preprocessing

2.1.1. Sentinel-2 Constellation

The Sentinel-2 constellation consists of two polar-orbiting satellites (Sentinel-2A and Sentinel-2B) placed in the same orbit. Sentinel-2A and Sentinel-2B equip with multispectral instruments capable of acquiring 13 bands information at different spatial resolutions (10, 20, and 60 m). Sentinel-2 provides more details in the near-infrared (NIR) band range and short wavelength infrared (SWIR) band range, which is helpful for land cover, land monitoring, and emergency response [48]. A high revisit time (10 days at the equator with one satellite, and 5 days with 2 satellites under cloud-free conditions, which results in 2–3 days at mid-latitudes) provides more cloudless images, and is a good support for the built-up areas extraction.

2.1.2. Atmospheric Correction and Image Sharpening

The bottom-of-atmosphere (surface) reflectance is a basic input to many earth observation applications, ranging from land surface phenology to land cover classification and change detection [49]. To process top-of-atmosphere Level-1C data into atmospherically corrected bottom-of-atmosphere data, the Sen2Cor processor (version 2.4), developed by ESA to perform atmospheric correction, was employed [50].
In higher spatial resolution images, built-up areas tend to be more easily detected because higher spatial resolution images can clearly define the boundaries of the built-up areas, uniformly highlight built-up areas, and eliminate redundant backgrounds in the extracted built-up areas [19]. To sharpen the bands of a Sentinel-2A image with spatial resolutions of 20 m and 60 m to a spatial resolution of 10 m, the modified selected and synthesized band scheme [51] was employed.

2.1.3. Optimal Band Selection

Built-up areas yield a higher reflectance response in the SWIR than in other bands [52], and it may help in alleviating the problem of confusion between built-up areas and other types of land cover, such as artificial open spaces, river gravel, and sand dunes [53]. As Sentinel-2 has two SWIR bands, it is, therefore, inherently advantageous when applied to built-up areas extraction. In this study, both SWIR bands were selected to form the optimal band combination for built-up areas extraction. To select the third band of the optimal band combination, the optimum index factor (OIF) [54] was employed. OIF is a statistic value that can be used to select the optimum combination of three bands in a satellite image with which you want to create a color composite. The optimal band combination of bands, out of all possible 3-band combinations, is the one with the highest amount of “information” (highest sum of standard deviations), with the least amount of duplication (lowest correlation among band pairs). Band 9 and Band 10 are the water vapor band and the cirrus band. Band 8 and band 8A have very high correlations, and the information overlap is large. The standard deviation of Band 1, Band 2, and Band 3 is relatively low and contains less information. Therefore, these bands are not considered. Five candidate band combinations are Bands 12, 11, 8; Bands 12, 11, 7; Bands 12, 11, 6; Bands 12, 11, 5; and Bands 12, 11, 4. The OIF values of candidate band combinations were calculated based on the TIFF images exported by SNAP in ENVI software. The OIF method can reflect the amount of information in the band combination, but it still has some limitations. We needed to further analyze the separability of candidate band combinations for built-up areas and non-built-up areas. Here, the Jeffries–Matusita (J-M) distance [55,56] was used as a separability criterion for optimal band combination selection, whereby the J-M value ranges from 0 to 2. First, we used ENVI to select samples of built-up areas and non-built-up areas from the TIFF images exported by SNAP. Then, we calculated the J-M value and determined the band combination with the largest J-M value as the optimal band combination. The optimal results are shown in Section 3.

2.2. Multiscale Segmentation

The saliency map accuracy level is sensitive to the segmentation scale, so a multiscale strategy was employed. The false color image generated by optimal band combination was first segmented into homogenous and compact regions using a simple linear iterative clustering (SLIC) superpixel segmentation method [57]. In SLIC, the number of superpixels, N’, can affect the effect of segmentation. If N’ is too small, it is often impossible to accurately separate the built-up areas from the background; if N’ is too large, it needs much more computing time [58]. As can be seen from Figure 2a,b, in some superpixels, both the built-up area pixels and the non-built-up area pixels are included, and the contour of the built-up areas cannot be accurately captured. In Figure 2c, the contour of the built-up areas can be accurately captured. Due to the scenes of remote sensing images being highly complex, determining the optimal N’ for each image is very time-consuming. To simplify the problem, we given a large N’ to avoid under segmentation. In this study, we empirically set N’ to 20,000. Then, Hu’s method [59] was adopted to merge similar superpixels into a set of objects Oi, i = 1, ..., N, where N is the number of segmented objects. Figure 2d,e show the results of merging 20,000 superpixels into 2000 and 4000 objects, respectively. It not only reduces the number of superpixels, but also ensures that the outline of the built-up areas is well captured.

2.3. Feature Selection

In this paper, three descriptors, including the color, texture, and spatial features, are used to describe each segmented object. Color feature is an important feature of the saliency detection methods; almost all saliency methods utilize the color feature. In particular, CIELab [60] aspires to perceptual uniformity, and its L component closely matches human perception of lightness, while the a and b channels approximate the human chromatic opponent system, and RGB is often the default choice for scene representation and storage [36]. Both of them are complementary and widely used for saliency detection [41,44,61]. Hence, we calculated the average pixel value of each segmented object Oi in RGB space and CIELab space, and the color feature of the segmented object Oi can be described as
F c o l o r = [ c r , g , b , c L , a , b ] ,
where cr,g,b and cL,a,b represent the average value of each color channel of the pixels in the segmented object Oi in the RGB and CIELab color spaces.
Built-up areas usually have unique texture features. Local binary patterns (LBPs) [62] was utilized to calculate the texture feature of segmented objects. First, the LBP encoding for each pixel in the image was calculated using a 3 × 3 window and, in the uniform pattern [62], each pixel was assigned a value between 0 and 56. It is worth pointing out that although larger window size (such as 5 × 5 or 7 × 7) can utilize more information in the neighborhood, the noise corruption from the pixel away from the center can be more severe, which inevitably deteriorates the discriminative ability of the LBPs’ feature [63]. Then, an LBP histogram for each segmented object Oi was constructed, and the texture feature of the segmented object Oi can be described as
F t e x t u r e = H L B P = { h 0 , , h i , , h 56 } ,
where hi is the value of the i-th bin in an LBP histogram.
For spatial features, the eccentricity and area properties were used to eliminate segmented objects with a large eccentricity and a large area; these segmented objects are often strips of bare rock and river banks (such as yellow river bank). This can be described as
F s p a t i a l = [ O A r e a > t h 1 O E c c e < t h 2 ] ,
where OArea is the area and OEcce is the eccentricity of segmented object Oi, and OEcce is between 0 and 1. To avoid erroneously eliminating the road inside the built-up areas, we only consider the long strip-shaped segmented objects with a large area. We experimentally set the th1 to 500 pixels, and th2 was set to 0.95. The feature vector of segmented object Oi can be obtained by
F = [ F c o l o r , F t e x t u r e , F s p a t i a l ]

2.4. Coarse Saliency Map

In this part, we mainly explain how to obtain coarse saliency maps and training samples for a strong model. In Section 2.4.1, we mainly used the clues, such as the color and texture, to obtain the initial saliency map of the built-up areas. However, the map often has some background and water information. In Section 2.4.2, we considered the spatial information of ground objects, and introduced the GWB model to eliminate the background information similar to the built-up areas, and obtained the coarse saliency map. In Section 2.4.3, we eliminated the water information from the coarse saliency map. In Section 2.4.4, we selected the training samples from the coarse saliency map for the strong model.

2.4.1. Multiple Cues Fusion

Compactness Saliency Using Color Cues

Following the image segmentation, a graph G = (V, E) with N nodes {v1, v2, · · ·, vN} was constructed, and edges E weighted by an affinity matrix W = [wij]N × N. Node vi corresponds to the ith segmented object, and edge eij link nodes vi and vj to each other, and the CIELab color space distance lij between nodes vi and vj is defined as
l i j = | | c i c j | | ,
where ci and cj are the mean of segmented objects corresponding to nodes vi and vj in the CIELab color space. Note that the distance matrix L = [lij]N × N is normalized to the interval [0, 1]. The affinity matrix wij is defined as
w i j = {    e l i j σ 2 ,     i f j = Ω i              0      , o t h e r w i s e                ,
where σ is a constant, and Ωi denotes the set of neighbors of node vi. If vi and vj are adjacent, vj is treated as a neighbor of vi, and the set of neighbors is equal to the number of vj.
Salient objects typically have compact spatial distributions, whereas background regions are widely distributed across the entire image. Therefore, compactness may be determined by calculating the spatial variances of the segmented objects to calculate the compactness saliency map [64]. First, the similarity aij between a pair of segmented objects, vi and vj, is defined as
a i j = e l i j σ 2     
The similarity based on the manifold ranking through the constructed graph is as follows:
H T = ( D α W ) 1 A     ,
where A = [aij]N × N, D = diag {d11, d22, …, dNN}, dii is the degree of nodes vi, and H = [hij]N × N is the similarity matrix after the diffusion process, α balances the smooth and fitting constraints of the manifold ranking algorithm and, empirically, α was set to 0.99 as in [65]. The spatial variance of segmented objects can be calculated as
s v ( i ) = j = 1 N h i j n j | | b j μ i | | j = 1 N h i j n j        ,
where nj represents the number of pixels that belong to segmented object vj, bj = [ b j x , b j y ] represents the centroid of the segmented object vj, and the μi = [ μ i x , μ i y ] represents the spatial mean.
Considering that segmented objects at the center of an image are more noticeable, the spatial distances between segmented objects and the image’s center can be calculated as follows
s d ( i ) = j = 1 N h i j n j | | b j p | | j = 1 N h i j n j      ,
where p = [px, py] is the spatial coordinate of the image center.
The saliency map based on compactness is defined as
S c o m ( i ) = 1 Norm ( s v ( i ) + s d ( i ) ) ,
where Norm(x) is a function that normalizes x to [0, 1].

Foreground Saliency Using Multiple Cues Contrast

Although the compactness saliency method tends to perform well, as attested to by a previous study [64], it only used the spatial variance of the color in the image space. As it primarily depends on color information, when the foreground and background objects are similar in color, the saliency deteriorates. To address this limitation, further aspects, such as texture and position, should be incorporated to refine the results.
First, the foreground seed set was determined by segmenting the compactness saliency map. Then, the contrast of each segmented object with the seeds was calculated using multiple cues, including information on texture and position. The foreground saliency is computed as follows:
S f o r e ( i ) = j Ω s [ D t ( i , j ) exp ( | | b i b j | | / σ 2 ) n j ] ,
where Ωs is the foreground seed set, Dt is the texture similarity between segmented objects based on LBP, and ||bibj|| is the Euclidean distance between position of segmented objects.
Next, the SFG map was propagated using manifold ranking and, then, the propagated map was normalized to [0, 1], and denoted Sfore(i). The Scom(i) and Sfore(i) maps are complementary to one another, and both saliency maps were integrated to define the initial saliency map,
S i c ( i ) = η S c o m ( i ) + ( 1 η ) S f o r e ( i ) ,
where η balances the compactness saliency map Scom(i) and foreground saliency map Sfore(i). In the optimal band combination, the built-up areas can be better identified using color features, while the built-up areas are also sensitive to texture features. Both of them have important contributions, so η was set to 0.5.

2.4.2. Geodesic Weighted Bayesian

Spatial information is a key aspect of geographic information system (GIS) and remote sensing fields, and while spatial relationships have increasingly been incorporated into satellite image processes, less attention has been given to the use of higher-level spatial relationships [66]. To rectify this, a GWB model [45] was introduced to optimize the initial saliency map. The Bayesian inference for estimating the saliency map [67] is calculated as
p ( s a l | v ) = p ( s a l ) p ( v | s a l ) p ( s a l ) p ( v | s a l ) + p ( b k ) p ( v | b k )    ,
p ( b k ) = 1 p ( s a l )    ,
where p(sal) is the prior probability of being salient at pixel v, p(bk) is the prior probability of a pixel belonging to the background, p(v|sal) and p(v|bk) are the likelihood of observations, v is the feature vector of a given pixel. When the spatial relationships were considered, p(v|sal) and p(v|bk) can be rewritten as
p ( v | s a l ) = s i s a l p g e o ( s i | s a l ) p ( v | s i )    ,
p ( v | b k ) = s i b k p g e o ( s i | b k ) p ( v | s i ) ,
where si is segmented object, sal is initial set of salient regions, and bk is the initial set of background regions. pgeo(si) denotes the probability of si, namely, the weight of segmented object si.
Given pixel x, the feature vector was represented by its CIELab color and LBP texture features, and the observation likelihood of the given pixel x in segmented object Oi can be calculated as
p ( v | s i ) = f { L , a , b , L B P } n j ( f ( x ) ) n j ,
where nj denotes the number of pixels within segmented object Oi, nj(f(x)) denotes the number of f(x) values contained in segmented object Oi, and f∈{L, a, b, LBP} denotes the component of feature vector v, substituting observation likelihood (16) and (17) into (14), and utilizing the initial saliency map as a prior distribution to generate a more precise saliency map. Then, the initial saliency map was further refined to obtain the coarse saliency map Scourse based on graph cut method [68].

2.4.3. Removing the Water Bodies

Water bodies in remote sensing images usually belong to dark targets, and are more easily identified as salient objects, which can render several saliency detection methods unsuccessful. To avoid the interference caused by water bodies, they must be removed from the coarse saliency map. In [69], Xu noticed that water bodies have a stronger absorbability, and the built-up class has greater radiation in the SWIR band. Based on this characteristic, we set the segmented objects whose average pixel values are smaller than the given threshold, Tw, to 0, thereby achieving the purpose of removing water bodies. To determine Tw, the histogram of SWIR band was first generated. For cities with more water bodies, water bodies occupy a larger area, so there is a peak on the left side of the histogram. The gray values of other ground objects are usually greater than water bodies, and their peaks on the histogram are to the right of the water peak. We determined the value corresponding to the first trough to the right of the water peak as Tw. Based on statistical results of multiple images, we determined that the threshold Tw is 0.15. The cities with less water are almost unaffected by water bodies, and Tw was set to 0.01. The gray value of the building shadow is also low, but its area is small. To avoid removing building shadow, we only removed segmented objects with a large area and the gray value less than Tw. Considering that buildings in some areas are dense, the area of the shadow is relatively large. We empirically set the area threshold for removing water to 100 pixels.

2.4.4. Training Sample Selection

To select accurate training samples from the coarse saliency map, a set of selection rules was established: first, the average saliency value of each segmented object was computed, and two thresholds, Th and Tl (Th is greater than Tl), were set to generate initial built-up area and non-built-up area training samples. Both thresholds can be adaptively determined by the mean value ϖ of the coarse saliency map, Th was set to ϑ times ϖ, and Tl was set to ϖ, where ϑ is a parameter and ϑ was set to 1.8; more discussions about the values of ϑ can be found in Section 3.1.3. The segmented objects with saliency values above the Th were selected as initial built-up area samples, while those with saliency values below the Tl were selected as initial non-built-up area samples. Next, we constrained the initial training sample set using the spatial feature Fspatial to obtain the training samples {si, li } i = 1 P , where si is the i-th training sample from the coarse saliency map Scourse, li is the binary label of the training sample, P is the number of the samples, built-up areas samples are labeled +1, and non-built-up areas samples are labeled −1.

2.5. Refined Saliency Map

One of the main difficulties using a support vector machine (SVM) is to determine the appropriate kernel for the given image. To select the appropriate kernel function for any input image, a multiple kernel boosting method [70] was employed. In this method, SVMs with different kernels are selected as weak classifiers, then, a strong classifier is learned, based on the boosting method. In this paper, we used Nf (Nfeature by Nkernel) different standard SVM classifiers, where Nfeature is the number of features and Nkernel is the number of kernel functions. The four different kernel functions are linear, polynomial, radial basis function, and sigmoid. For different feature sets, the decision function can be defined as
Y ( r ) = n = 1 N f β n i = 1 P w i l i k n ( s , s i ) + b ¯ ,   n = 1 N f β n = 1 ,       β n +  
where βn is the kernel weight, wi is the Lagrange multiplier, and b ¯ is the bias in the standard SVM algorithm. Equation (19) is a conventional function for the multiple kernel learning method; when the boosting algorithm was used to replace the simple combination of single-kernel SVMs in the multiple kernel learning, Equation (19) can be rewritten as
Y ( r ) = n = 1 N f β n ( w T k n ( s ) + b ¯ ) ,
where kn(s) = [kn(s,s1),kn(s,s2),…,kn(s,sP)]T, w = [w1l1,w2l2,…,wPlP]T, and b =   n   =   1 N f b ¯ . By setting the decision function as Z n ( S ) = w T k n ( S ) + b ¯ n , the AdaBoost method may be employed to train a strong classifier, and Formula (20) can be rewritten as
Y ( s ) = j = 1 J β j z j ( s )
The AdaBoost method was used to calculate βj, and J represents the number of iterations of the boosting process. The process is as follows:
Step 1: Begin with uniform weights, ω1(i) = 1/P, i = 1, 2, …, P, and assign a set of decision functions {Zn(S), n = 1, 2, …, Nf} to each weak classifier.
Step 2: Compute the classification error {εn} for each of the weak classifiers, and ascertain the decision function zj(s) with the minimum error εj; then, the combination coefficient βj is computed by
β j = 1 2 ln 1 ε j ε j 1 2 ( sgn ( ln 1 ε j ε j ) + 1 ) ,
where sgn(x) is the sign function, which equals 1 when x > 0, and is −1 otherwise, and βj must exceed 0.
Step 3: Update the weight according to Equation (23), and repeat step 2 for the next iteration until J iterations are completed.
w j + 1 ( s i ) = w j ( s i ) e β j l i z j ( s i ) 2 ε j ( ε j 1 )  
Following J iterations, all of the βj and zj(s) can be obtained, and then the strong classifier was learned. Subsequently, a pixel-wise saliency map was generated using the strong classifier. Finally, the refined saliency map Srefined was improved based on the graph cut method [68] and guided filter [71].

2.6. Multiscale Saliency

Since the size of the ground objects in the image is different, the saliency objects can appear on a variety of scales. In other words, the accuracy of the saliency map is sensitive to the number of segmented objects [70], a multiscale strategy was employed. In this study, seven layers (M is set to 7) of segmented objects with different granularities were generated, where N was set to 1000, 1500, 2000, 2500, 3000, 3500, and 4000, in each of the layers. More discussions about the values of M can be found in Section 3.1.3. To effectively integrate the results of the multiple scales M, the CCA method [46] was employed, whereby each cell corresponds to a pixel, and the saliency values of all pixels constitute the set of cells’ states. For any cell in a saliency map, there should be 5M − 1 neighbors, including pixels with the same coordinates from different saliency maps, in addition to their 4 connected pixels [46]. The saliency value of pixel i in the m-th saliency map at time t stands for its probability to be the foreground F, represented as S m , i ( t ) , while its possibility to be the background B is denoted as 1 − S m , i ( t ) . Otsu’s method was used to binarize each map using an adaptive threshold. The threshold did not change, and was only related to the initial image. The threshold of the m-th saliency map is represented as γm. Following segmentation, a pixel i can be classified as foreground or background. If a pixel, i, is foreground, the probability that one of its neighboring pixels, j, is measured as foreground is λ, while µ is the probability that j is measured as background when i belongs to the background. We assumed that λ is equal to µ if it was considered equally probable that the pixel belongs to the foreground or to the background. The posterior probability can be denoted as S m , i ( t ) ·λ, which represents the probability of pixel i belonging to the foreground F, on the condition that its neighboring pixel j in the m-th saliency map was binarized as foreground at time t, and posterior probability S m , i ( t + 1 ) can also be used to represent the probability of pixel i belonging to the foreground F at time t + 1. Based on the prior ratio in [46], we have
S m , i ( t + 1 ) 1 S m , i ( t + 1 ) = S m , i ( t ) 1 S m , i ( t ) λ 1 λ      
To deal with the logarithm of Equation (24), we have
l ( s m , i ( t + 1 ) ) = l ( s m , i ( t ) ) + Λ ,
where l( s m , i ( t + 1 ) ) is ln( s m , ( t + 1 ) /1 − s m , ( t + 1 ) ), and Λ is ln(λ/1 − λ). Assuming that each neighbor’s contribution was conditionally independent, the synchronous updating rule can be defined as
l ( s m , i ( t + 1 ) ) = l ( s m , i ( t ) ) + m ( t ) Λ ,
m ( t ) = j = 1 5 k = 1 M δ ( k = m , j > 1 ) s i g n ( s j , k ( t ) γ k 1 )    ,
where s m ( t ) is the m-th saliency map at time t, M is the number of multiscale saliency maps, s j , k ( t ) is the vector containing the saliency values of the j-th neighbor for all pixels in the m-th saliency map at time t, and 1 = [1, 1, . . . , 1]. After TC iterations, the integrated saliency map s(TC) can be integrated as
s ( T C ) = 1 M m = 1 M s m ( T C )

2.7. Integration

Coarse saliency maps boast several advantages for detecting details and capturing local structural information, while refined saliency maps are more adept at describing global shapes. To maximize the complementarity of both salient maps, we integrated them using a weighted combination
S f i n a l = κ S c o u r s e + ( 1 κ ) S r e f i n e ,
where κ is a balance factor for the combination. In the extraction of built-up areas, greater attention was paid to the outer contours of a city, and the fine saliency map is more applicable, so κ was set as 0.2, κ is between 0 and 1.

2.8. Bulit-Up Area Extraction

In the final saliency map, Sfinal, built-up areas usually have the highest values, ground objects similar to built-up areas have the next highest values, and other ground objects have very low values. As such, the final saliency map can also be broadly segmented into three parts based on the gray value. To extract accurate built-up areas, an appropriate threshold segmentation image needs to be set. In [72], the genetic algorithm was used to determine the optimal segmentation threshold and achieved good results. In our paper, a multi-threshold segmentation algorithm, FODPSO [47], was employed. Following segmentation, the highest value part is the binary map of the built-up areas. The pseudo-code for FODPSO is presented in Table 1.

3. Experimental Results

To date, there has been little investigation into saliency detection in remote sensing images; thus, there are no classic testing datasets with existing ground truth (GT) that may be consulted when categories of Sentinel-2 images are introduced to evaluate the effectiveness and novelty of the proposed method. The GT map is obtained by manual segmentation based on the definition of the built-up areas in the first section. Since the built-up areas extraction is greatly affected by the surrounding land cover, based on the ecosystems, the experimental cities are divided into five types: desert, coastal, riverside, valley, and plain cities (Figure 3). Desert cities are distributed in the northwest of China, surrounded by desert and loess, with little vegetation. These conditions have a significant impact on the built-up area extraction. Coastal cities are located in the eastern coastal areas of China, and are the most economically active areas, requiring more timely monitoring of built-up areas. The riverside cities are distributed along the banks of the Yangtze and Yellow Rivers. As dark ground object, water bodies are likely to be detected as saliency targets by the saliency model, which may affect the built-up area extraction by the saliency method. Valley cities are located in western China, and are often affected by bare rock, rendering built-up areas extraction extremely challenging. Plain cities are distributed in eastern China, where the landform of the built-up areas is plain, and surrounded by significant amounts of farmland (including bare land), which may have an impact on the extraction of built-up areas. Details of the study areas are presented in Table 2.
Based on the OIF and J-M methods, the selection of the optimal band combination for different images is shown in Table 3. From Table 3, the optimal band combination for most cities consists of bands 12, 11, and 7, while the optimal band combination for most valley cities is composed of bands 12, 11, and 5.
To evaluate the performance of the proposed method with respect to saliency detection, we compared it to the eight most recent saliency detection methods: dense and sparse reconstruction (DSR) [73], discriminative regional feature integration (DRFI) [41], regional principal color (RPC) [74], diffusion-based compactness and local contrast (DCLC) [64], inner and inter label propagation (LPS) [75], bootstrap learning (BL) [44], diffusion process on a two-layer sparse graph (DPTLSG) [76], and reversion correction and regularized random walks ranking (RCRR) [77]. To ensure fairness, all methods used the optimal band combination as the original image for saliency detection. To further evaluate the extraction accuracy of the proposed method, the saliency map was segmented to obtain the binary map of the built-up areas based on FODPSO; the results were compared to some built-up area extraction methods. Due to index-based methods being sensitive to the built-up areas [72], they are widely used in built-up areas extraction. In this study, two index-based methods, NDBI [7] and NBI [5], were selected. PanTex [11] is a method for extracting built-up areas based on texture feature, and has been evaluated by many experiments [78], so the method was also selected.

3.1. Comparison to the State-of-the-Art Saliency Methods

3.1.1. Qualitative Experiment

The results of saliency maps generated by nine different methods are presented for qualitative comparison in Figure 4. It is clear that our method efficiently detects built-up areas and identifies the contours of built-up areas most accurately, while the results of the other eight methods are inferior. For riverside cities, several of the methods, including DSR, DRFI, RPC, BL, and RCRR, identify water bodies as salient objects rather than built-up areas. Of these methods, DCLC and DPTLSG have superior performance, but still fall short in terms of accuracy. The BL method also produces satisfactory saliency maps, but its performance is generally poor in coastal and riverside cities. DSR, DRFI, and RPC fail to highlight the built-up areas in their entireties. LPS focuses too much on an image’s central information, overlooking the information on built-up areas throughout the image. The RCRR method detects and highlights unnecessary and irrelevant background information.

3.1.2. Quantitative Experiment

To quantitatively evaluate the performance of each saliency method, the receiver operating characteristic (ROC)-area under the curve (AUC) metric, precision, recall, and F-measure, and time comparison were used.

ROC-AUC Metric

The receiver operating characteristic (ROC) curve is derived by thresholding a saliency map at the threshold within the range [0, 255], and further classifying the saliency map into the saliency objects and the background [79]. The ROC graph is generated by plotting the true positive rate (on the y-axis) against the false positive rate (on the x-axis). The true positive and false positive rates are expressed as
T P R = T P T P + F N ,
F P R = F P F P + T N ,
where TPR is the true positive rate, FPR is the false positive rate, TP is the true positives and it is the number of correctly identified built-up areas, FN is the false negatives and it is the number of incorrectly rejected, FP is false positives and it is the number of the incorrectly identified, and TN is the true negatives and it is the number of the correctly rejected. When the FPR value is the same, the higher the TPR, the better a method’s performance; the larger the area under the curve (AUC), the better the performance. The AUC values for the different methods are presented in Table 4, from which it may be seen that the proposed method has the highest AUC value. The ROC curves for the different methods are shown in Figure 5a. We can conclude that the ROC curve generated by our model demonstrates superior performance.

Precision, Recall, and F-Measure

To further evaluate the quality of the saliency maps, precision, recall, and F-measure were employed. They can be computed by
p r e c i s i o n = x y t ( x , y ) s ( x , y ) x y s ( x , y )     ,
r e c a l l = x y t ( x , y ) s ( x , y ) x y t ( x , y ) ,
F n = ( 1 + β 2 ) p r e c i s i o n r e c a l l β 2 p r e c i s i o n + r e c a l l ,
where the term (x, y) denotes the coordinates of the images, t(x, y) is the ground truth, and s(x, y) is the binary image after the thresholding saliency maps. The threshold was set to twice the average gray value; the segmented objects whose average gray value are greater than the threshold was designated foreground, with all others designated background. High recall means that a model returned most of the built-up areas, whereas high precision means that a model returned substantially more built-up areas than background regions. The F-measure is the harmonic mean of precision and recall, and β2 was set as 1 to balance the importance of precision and recall.
Figure 5b shows the precision, recall, and F-measure values of the evaluated methods. According to Figure 5b, our method achieves the highest precision, which means that the redundant background in the residential areas acquired by our method is the smallest among the nine methods. The recall of our method is the highest among the nine methods, which means that our methods can complete built-up areas. The built-up areas extracted by the RPC and BL methods are greatly affected by water bodies and have a low recall and precision. LPS focuses too much on an image’s central information, so it also has a low recall and precision. DCLC, DPTLSG have good accuracy and recall rate, and can also identify the built-up areas well. Although RCRR has a good recall rate, it extracts more background information, so the accuracy is lower. Overall, the proposed method performs well against the state-of-the-art methods.

Time Comparison

We compared the computational time for each method using MATLAB on a PC with 8 G RAM, Intel Core i5-4590 CPU @ 3.30 GHz. The average time comparison of our method and the other competing methods is given in Table 5. As can be seen from Table 5, our method is very time-consuming, mainly due to multiscale segmentation and an ensemble learning strategy.

3.1.3. Important Parameter Settings

There are two important parameters in the proposed method: M, layers of segmented objects; the parameter ϑ used to calculate the thresholds Th and Tl. To determine the optimal parameter M, we compared the accuracy and time of the different M. The results are shown in Table 6. From Table 6, as M increases, the accuracy increases, but the calculation time also rises sharply. If M continues to increase, the computational time cannot be accepted.
To determine the optimal parameter ϑ, we compared the accuracy of different ϑ. The results are shown in Table 7. From Table 7, we can see that when ϑ is 1.8, the accuracy is the highest. Although ϑ = 1.8 was determined to be the optimal ϑ, it was not suitable for images of Xining and Yulin. For the images of Xining and Yulin, ϑ was set to 2.8.

3.2. Comparison to the State-of-the-Art Built-Up Areas Extraction Methods

To evaluate the overall performance of the proposed method in extracting built-up areas, we further compared our method to some advanced built-up areas extraction methods, including two index-based methods [5,7] and one texture-based method [11]. They were calculated according to Equations (35)–(37). To make these equations available, we used ESA’s Sen2Cor atmospheric correction module to process Sentinel-2 Level-1C images into Level-2A bottom of atmosphere (BOA) reflectance images [50].
N D B I = b s w i r b n i r b s w i r b n i r ,
N B I = b r e d b s w i r b b i r     ,
f ( b u i l t _ u p ) = { t x 1 , t x i , , t x n } ,     i [ 1 , 2 , , n ] ,
where bswir is the reflectance of the SWIR band, bnir is the reflectance of the NIR band, bred is the reflectance of the red band, txi = f(w = 9, vi, m = CON), i∈[α1, d1; a2, d2; …; an, dn]. w = window size, α and d are the distance and angle defining the displacement vector v required to select the pairs producing the co-occurrence matrix, and m is the textural measure applied to the given co-occurrence matrix distribution. CON = i   = 1 N g j   = 1 N g ( i j ) 2 · P i j , with Ng is the number of gray levels present in the image, and Pi,j is the (i,j)th entry of the co-occurrence matrix.
Figure 6 presents the results achieved by the different methods. The binary maps shown in Figure 6f,g,h were obtained by automatically determining the segmentation threshold using Otsu’s algorithm to segment maps in Figure 6b,c,d. The binary maps shown in Figure 6j were obtained by segmenting the maps in Figure 6f. based on the optimal thresholds determined by FODPSO algorithm. Since the built-up areas are usually large, we keep the area larger than the Tarea and remove the area smaller than the Tarea, where the Tarea is empirically set to 3000 pixels. From Figure 6, it is clear that the two index-based methods perform poorly on the images of the desert cities and the valley cities, they are almost impossible to use to identify and extract the built-up areas, while the desert and bare rock are clearly extracted. However, they perform very well on the images of coastal cities, riverside cities, and plain cities, because of the high vegetation coverage or large water area in these cities. The PanTex method performs better than both index-based methods in detecting built-up areas, with the locations clearly identifiable. However, the PanTex method only utilizes texture features; some areas with texture features similar to built-up areas may also be extracted. For example, in desert cities, they are surrounded by a large number of loess and desert areas, which have similar texture features to the built-up areas and are, therefore, incorrectly extracted. Although the land cover around cities varies, our proposed method can still efficiently identify the locations and boundaries of built-up areas, and can accurately extract them. To quantitatively evaluate the various methods, three statistical measures were used: overall accuracy, commission error, and omission error. The commission error represents the percentage of pixels that belong to non-built-up areas but have been classified as built-up areas. Omission error represents the percentage of pixels that belong to built-up areas, but have been classified as non-built-up areas. Table 8 shows the average statistical measurement results for five different types of cities. The overall accuracies of our proposed method in five different types of cities are higher than the other three methods, and the commission and omission errors of the proposed method are the lowest among the four methods. NDBI and NBI have low overall accuracy and high commission errors and omission error on images in desert and valley cities. This suggests that these two index-based methods are not suitable for extracting built-up areas surrounded by bare rock and desert. However, they perform well on images in the other three types of cities and achieve high overall accuracy. PanTex performs very well, second only to the proposed method. Its omission error is also relatively low, while the commission error is high in desert cities. This indicates that when PanTex extracted the built-up areas of the desert city, a large amount of non-built-up areas are incorrectly extracted. In summary, the proposed method takes into account the different features of the built-up areas, based on visual salience, and can achieve good results in different types of cities.

4. Discussion

In this paper, a new method for extracting built-up areas from images based on the principles of salient object detection is proposed. Compared to other saliency detection methods, the unique band information of remote sensing is fully exploited. For example, the optimal band combination composed of unique bands of satellite images can highlight the built-up areas very effectively. Water bodies in the SWIR band can be removed to prevent them from being detected as salient objects. Compared to existing built-up area extraction methods using optical images, our method devotes greater attention to the most conspicuous built-up areas in the image, and not only the spectral and textural information of the built-up areas. Therefore, it is more robust and not easily distorted by surrounding ground objects, such as bare rock or desert. Compared to the BL saliency method, we consider a greater number of cues and introduce the GWB model to improve the detection accuracy of the coarse saliency map, thus providing a more reliable training sample for training strong models. CCA is employed to integrate multiscale detection results, as distinct from the simple superposition effect using the BL saliency method, and it can improve the ultimate detection accuracy and reduce the background information. In addition, shape information of ground objects is also utilized. For the multiscale segmentation strategy, we do not set different SLIC parameters for segmenting the image multiple times but, rather, adopt a fixed segmentation parameter, then merge the superpixels based on the various merge parameters to obtain multiscale segmentation images, improving segmentation and algorithm efficiency.
Although our method is precise and robust, it does have some shortcomings and limitations that cannot be ignored. First, it employs multiscale segmentation and an ensemble learning strategy that affects its processing efficiency, resulting in a computational time that is several times as long as some other methods. However, we can select fewer scales if the requirements can be met and the calculation time is shortened. Second, our method also incorrectly detects some non-built-up areas. For example, in the Lhasa image, the river bank is detected. The third limitation concerns the detection accuracy of the coarse saliency map, which affects the reliability of the training sample and the final result. The detection accuracy of the coarse saliency map depends on the detection method and the input image. Overall, the proposed method has the potential to extract built-up areas in different types of cities with adequate accuracy.

5. Conclusions

This paper proposes a new built-up area extraction method based on an improved BL saliency model. First, the band combinations that highlight the built-up areas are explored. Then, we produce coarse saliency maps based on multiple cues, and the GWB model to generate training samples for a strong classification model that is subsequently used to produce a refined saliency map. To further improve detection performance, multiscale saliency maps are integrated by CCA. The final saliency result combines the coarse and refined saliency maps. Finally, the information pertaining to built-up areas is extracted using the FODPSO algorithm. Comparative experimentation using other advanced saliency detection methods indicates that our method outperforms the other eight models in extracting built-up areas from various complex background environments. Comparative analyses with three advanced built-up area extraction methods confirm the superior performance of our proposed method. Therefore, the proposed mothed not only has good precision and robustness, but also has practical value in the extraction of built-up areas.
Future research will focus primarily on three aspects: First, we intend to optimize our method further, shortening the computational duration of the process. Second, we will take more features into consideration to avoid the extraction of other ground objects. Third, we will determine the optimal band combinations of other satellites for extracting built-up areas, ultimately extending our method for use with more satellites.

Author Contributions

Z.S. had the original idea for the study. Q.M. supervised the research and contributed to the article’s organization. W.Z. analyzed the data. Z.S. performed the experiments and drafted the manuscript, which was revised by all of the authors. All of the authors read and approved the submitted manuscript.

Funding

This research was funded by Hainan Provincial Department of Science and Technology grant number ZDKJ2016021, Major Special Project—the China High-Resolution Earth Observation System grant number 30-Y20A07-9003-17/18, Hainan Province Natural Science Foundation grant number 2017CXTD015, Sichuan Province Science and Technology Support Program grant number 2016JZ0027.

Acknowledgments

We would like to thank the Sentinel-2 data were obtained from the European Space Agency.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Taubenböck, H.; Esch, T.; Felbier, A.; Wiesner, M.; Roth, A.; Dech, S. Monitoring urbanization in mega cities from space. Remote Sens. Environ. 2012, 117, 162–176. [Google Scholar] [CrossRef]
  2. Yu, S.; Sun, Z.; Guo, H.; Zhao, X.; Sun, L.; Wu, M. Monitoring and analyzing the spatial dynamics and patterns of megacities along the maritime silk road. J. Remote Sens. 2017, 21, 169–181. [Google Scholar]
  3. Sun, Z.; Guo, H.; Li, X.; Lu, L.; Du, X. Estimating urban impervious surfaces from landsat-5 tm imagery using multilayer perceptron neural network and support vector machine. J. Appl. Remote Sens. 2011, 5, 053501. [Google Scholar] [CrossRef]
  4. Deng, C.; Wu, C. Bci: A biophysical composition index for remote sensing of urban environments. Remote Sens. Environ. 2012, 127, 247–259. [Google Scholar] [CrossRef]
  5. Jieli, C.; Manchun, L.; Yongxue, L.; Chenglei, S.; Wei, H.U. Extract residential areas automatically by new built-up index. In Proceedings of the 18th International Conference on Geoinformatics, Beijing, China, 18–20 June 2010. [Google Scholar]
  6. Xu, H. A new index for delineating built-up land features in satellite imagery. Int. J. Remote Sens. 2008, 29, 4269–4276. [Google Scholar] [CrossRef]
  7. Zha, Y.; Gao, J.; Ni, S. Use of normalized difference built-up index in automatically mapping urban areas from tm imagery. Int. J. Remote Sens. 2003, 24, 583–594. [Google Scholar] [CrossRef]
  8. Sun, G.; Chen, X.; Jia, X.; Yao, Y.; Wang, Z. Combinational build-up index (cbi) for effective impervious surface mapping in urban areas. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 2081–2092. [Google Scholar] [CrossRef]
  9. Zhang, P.; Sun, Q.; Liu, M.; Li, J.; Sun, D. A strategy of rapid extraction of built-up area using multi-seasonal landsat-8 thermal infrared band 10 images. Remote Sens. 2017, 9, 1126. [Google Scholar] [CrossRef]
  10. Shao, Z.; Tian, Y.; Shen, X. Basi: A new index to extract built-up areas from high-resolution remote sensing images by visual attention model. Remote Sens. Lett. 2014, 5, 305–314. [Google Scholar] [CrossRef]
  11. Pesaresi, M.; Gerhardinger, A.; Kayitakire, F. A robust built-up area presence index by anisotropic rotation-invariant textural measure. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2008, 1, 180–192. [Google Scholar] [CrossRef]
  12. Wentz, E.A.; Stefanov, W.L.; Gries, C.; Hope, D. Land use and land cover mapping from diverse data sources for an arid urban environments. Comput. Environ. Urban Syst. 2006, 30, 320–346. [Google Scholar] [CrossRef]
  13. Leinenkugel, P.; Esch, T.; Kuenzer, C. Settlement detection and impervious surface estimation in the mekong delta using optical and sar remote sensing data. Remote Sens. Environ. 2011, 115, 3007–3019. [Google Scholar] [CrossRef]
  14. Zhu, Z.; Woodcock, C.E.; Rogan, J.; Kellndorfer, J. Assessment of spectral, polarimetric, temporal, and spatial dimensions for urban and peri-urban land cover classification using landsat and sar data. Remote Sens. Environ. 2012, 117, 72–82. [Google Scholar] [CrossRef]
  15. Zhang, J.; Li, P.; Wang, J. Urban built-up area extraction from landsat tm/etm+ images using spectral information and multivariate texture. Remote Sens. 2014, 6, 7339–7359. [Google Scholar] [CrossRef]
  16. Borji, A.; Cheng, M.; Jiang, H.; Li, J. Salient object detection: A benchmark. IEEE Trans. Imag. Process. 2015, 24, 5706–5722. [Google Scholar] [CrossRef] [PubMed]
  17. Dong, C.; Liu, J.; Xu, F. Ship detection in optical remote sensing images based on saliency and a rotation-invariant descriptor. Remote Sens. 2018, 10, 400. [Google Scholar] [CrossRef]
  18. Zhang, Y.; Wang, X.; Xie, X.; Li, Y. Salient object detection via recursive sparse representation. Remote Sens. 2018, 10, 652. [Google Scholar] [CrossRef]
  19. Zhang, L.; Li, A.; Zhang, Z.; Yang, K. Global and local saliency analysis for the extraction of residential areas in high-spatial-resolution remote sensing image. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3750–3763. [Google Scholar] [CrossRef]
  20. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef] [Green Version]
  21. Harel, J.; Koch, C.; Perona, P. Graph-based visual saliency. In Advances in Neural Information Processing Systems; Touretzky, D.S., Mozer, M.C., Hasselmo, M.E., Eds.; MIT Press: Cambridge, MA, USA, 2007. [Google Scholar]
  22. Ma, Y.; Zhang, H. Contrast-based image attention analysis by using fuzzy growing. In Proceedings of the Eleventh ACM International Conference on Multimedia, Berkeley, CA, USA, 2–8 November 2003. [Google Scholar]
  23. Gao, D.; Mahadevan, V.; Vasconcelos, N. The discriminant center-surround hypothesis for bottom-up saliency. In Advances in Neural Information Processing Systems; Platt, J.C., Koller, D., Singer, Y., Roweis, S.T., Eds.; Curran Associates Icn: Vancouver, BC, Canada, 2008. [Google Scholar]
  24. Gao, D.; Vasconcelos, N. Bottom-up saliency is a discriminant process. In Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, Rio de Janeiro, Brazil, 14–21 October 2007. [Google Scholar]
  25. Cheng, M.; Mitra, N.J.; Huang, X.; Torr, P.H.; Hu, S.-M.; Intelligence, M. Global contrast based salient region detection. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 569–582. [Google Scholar] [CrossRef] [PubMed]
  26. Perazzi, F.; Krähenbühl, P.; Pritch, Y.; Hornung, A. Saliency filters: Contrast based filtering for salient region detection. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  27. Shi, K.; Wang, K.; Lu, J.; Lin, L. Pisa: Pixelwise image saliency by aggregating complementary appearance contrast measures with spatial priors. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 25–27 June 2013. [Google Scholar]
  28. Gopalakrishnan, V.; Hu, Y.; Rajan, D. Random walks on graphs for salient object detection in images. IEEE Trans. Imag. Process. 2010, 19, 3232–3242. [Google Scholar] [CrossRef] [PubMed]
  29. Wei, Y.; Wen, F.; Zhu, W.; Sun, J. Geodesic saliency using background priors. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012. [Google Scholar]
  30. Jiang, B.; Zhang, L.; Lu, H.; Yang, C.; Yang, M.-H. Saliency detection via absorbing markov chain. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 3–6 December 2013. [Google Scholar]
  31. Yan, Q.; Xu, L.; Shi, J.; Jia, J. Hierarchical saliency detection. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Oregon, Portland, 25–27 June 2013. [Google Scholar]
  32. Qin, Y.; Lu, H.; Xu, Y.; Wang, H. Saliency detection via cellular automata. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 8–10 June 2015. [Google Scholar]
  33. Bruce, N.; Tsotsos, J. Saliency based on information maximization. In Advances in Neural Information Processing Systems; Jordan, M.I., LeCun, Y., Solla, S.A., Eds.; MIT Press: Cambridge, MA, USA, 2006; pp. 155–162. [Google Scholar]
  34. Zhang, L.; Tong, M.H.; Marks, T.K.; Shan, H.; Cottrell, G.W. Sun: A bayesian framework for saliency using natural statistics. J. Vis. 2008, 8, 32. [Google Scholar] [CrossRef] [PubMed]
  35. Shen, X.; Wu, Y. A unified approach to salient object detection via low rank matrix recovery. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 18–20 June 2012. [Google Scholar]
  36. Borji, A.; Itti, L. Exploiting local and global patch rarities for saliency detection. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 18–20 June 2012. [Google Scholar]
  37. Wang, Q.; Zheng, W.; Piramuthu, R. Grab: Visual saliency via novel graph model and background priors. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar]
  38. Zhu, W.; Liang, S.; Wei, Y.; Sun, J. Saliency optimization from robust background detection. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014. [Google Scholar]
  39. Peng, H.; Li, B.; Ji, R.; Hu, W.; Xiong, W.; Lang, C. Salient object detection via low-rank and structured sparse matrix decomposition. IEEE Trans. Patt. Anal. Mach. Intell. 2013, 39, 796–802. [Google Scholar]
  40. Lang, C.; Liu, G.; Yu, J.; Yan, S. Saliency detection by multitask sparsity pursuit. IEEE Trans. Imag. Process. 2012, 21, 1327–1338. [Google Scholar] [CrossRef] [PubMed]
  41. Jiang, H.; Wang, J.; Yuan, Z.; Wu, Y.; Zheng, N.; Li, S. Salient object detection: A discriminative regional feature integration approach. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Oregon, Portland, 25–27 June 2013. [Google Scholar]
  42. Yang, J.; Yang, M. Top-down visual saliency via joint crf and dictionary learning. IEEE Trans. Patt. Anal. Mach. Intell. 2017, 39, 576–588. [Google Scholar] [CrossRef] [PubMed]
  43. Cholakkal, H.; Rajan, D.; Johnson, J. Top-Down Saliency with Locality-Constrained Contextual Sparse Coding. Available online: http://www.bmva.org/bmvc/2015/papers/paper159/paper159.pdf (accessed on 4 August 2018).
  44. Tong, N.; Lu, H.; Ruan, X.; Yang, M.-H. Salient object detection via bootstrap learning. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 8–10 June 2015. [Google Scholar]
  45. Wang, X.; Ma, H.; Chen, X. Geodesic weighted bayesian model for salient object detection. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 397–401. [Google Scholar]
  46. Qin, Y.; Feng, M.; Lu, H.; Cottrell, G.W. Hierarchical cellular automata for visual saliency. Int. J. Comput. Vis. 2018, 1–20. [Google Scholar] [CrossRef]
  47. Couceiro, M.; Ghamisi, P. Fractional Order Darwinian Particle Swarm Optimization: Applications and Evaluation of an Evolutionary Algorithm; Springer: Berlin, Germany, 2015. [Google Scholar]
  48. Drusch, M.; Del Bello, U.; Carlier, S.; Colin, O.; Fernandez, V.; Gascon, F.; Hoersch, B.; Isola, C.; Laberinti, P.; Martimort, P. Sentinel-2: Esa’s optical high-resolution mission for gmes operational services. Remote Sens. Environ. 2012, 120, 25–36. [Google Scholar] [CrossRef]
  49. Vuolo, F.; Żółtak, M.; Pipitone, C.; Zappa, L.; Wenng, H.; Immitzer, M.; Weiss, M.; Baret, F.; Atzberger, C. Data service platform for sentinel-2 surface reflectance and value-added products: System use and examples. Remote Sens. 2016, 8, 938. [Google Scholar] [CrossRef]
  50. Mueller-Wilm, U. Sentinel-2 msi—Level-2a Prototype Processor Installation and User Manual. Available online: http://step.esa.int/thirdparties/sen2cor/2.2.1/S2PAD-VEGA-SUM-0001-2.2.pdf (accessed on 6 July 2018).
  51. Park, H.; Choi, J.; Park, N.; Choi, S. Sharpening the vnir and swir bands of sentinel-2a imagery through modified selected and synthesized band schemes. Remote Sens. 2017, 9, 1080. [Google Scholar] [CrossRef]
  52. Valdiviezo-N, J.C.; Téllez-Quiñones, A.; Salazar-Garibay, A.; López-Caloca, A.A. Built-up index methods and their applications for urban extraction from sentinel 2a satellite data: Discussion. J. Opt. Soc. Am. A 2018, 35, 35–44. [Google Scholar] [CrossRef] [PubMed]
  53. Pesaresi, M.; Corbane, C.; Julea, A.; Florczyk, A.J.; Syrris, V.; Soille, P.; Sensing, R. Assessment of the added-value of sentinel-2 for detecting built-up areas. Remote Sens. 2016, 8, 299. [Google Scholar] [CrossRef] [Green Version]
  54. Chavez, P.; Berlin, G.L.; Sowers, L.B. Statistical method for selecting landsat mss ratios. J. Appl. Photogr. Eng. 1982, 8, 23–30. [Google Scholar]
  55. Richards, J.A.; Richards, J. Remote Sensing Digital Image Analysis; Springer: Berlin, Germany, 1999. [Google Scholar]
  56. Swain, P.H.; Davis, S.M. Remote sensing: The quantitative approach. IEEE Trans. Patt. Anal. Mach. Intell. 1981, 713–714. [Google Scholar] [CrossRef]
  57. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. Slic superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Patt. Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [PubMed]
  58. Moya, M.M.; Koch, M.W.; Perkins, D.N.; West, R.D.D. Superpixel segmentation using multiple sar image products. In Radar Sensor Technology XVIII; International Society for Optics and Photonics: Bellingham, WA, USA, 2014; p. 90770R. [Google Scholar]
  59. Hu, Z.; Wu, Z.; Zhang, Q.; Fan, Q.; Xu, J. A spatially-constrained color–texture model for hierarchical vhr image segmentation. IEEE Geosci. Remote Sens. Lett. 2013, 10, 120–124. [Google Scholar] [CrossRef]
  60. Connolly, C.; Fleiss, T. A study of efficiency and accuracy in the transformation from rgb to cielab color space. IEEE Trans. Imag. Process. 1997, 6, 1046–1048. [Google Scholar] [CrossRef] [PubMed]
  61. Hu, P.; Wang, W.; Zhang, C.; Lu, K. Detecting salient objects via color and texture compactness hypotheses. IEEE Trans. Imag. Process. 2016, 25, 4653–4664. [Google Scholar] [CrossRef] [PubMed]
  62. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Patt. Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef] [Green Version]
  63. Jia, S.; Deng, B.; Zhu, J.; Jia, X.; Li, Q. Local binary pattern-based hyperspectral image classification with superpixel guidance. IEEE Trans. Geosci. Remote Sens. 2018, 56, 749–759. [Google Scholar] [CrossRef]
  64. Zhou, L.; Yang, Z.; Yuan, Q.; Zhou, Z.; Hu, D. Salient region detection via integrating diffusion-based compactness and local contrast. IEEE Trans. Imag. Process. 2015, 24, 3308–3320. [Google Scholar] [CrossRef] [PubMed]
  65. Zhou, D.; Weston, J.; Gretton, A.; Bousquet, O.; Schölkopf, B. Ranking on data manifolds. In Advances in Neural Information Processing Systems; Mit Press: Cambridge, MA, USA, 2004; pp. 169–176. [Google Scholar]
  66. Qiao, C.; Wang, J.; Shang, J.; Daneshfar, B. Spatial relationship-assisted classification from high-resolution remote sensing imagery. Int. J. Dig. Earth 2015, 8, 710–726. [Google Scholar] [CrossRef]
  67. Xie, Y.; Lu, H.; Yang, M. Bayesian saliency via low and mid level cues. IEEE Trans. Imag. Process. 2013, 22, 1689–1698. [Google Scholar]
  68. Boykov, Y.; Kolmogorov, V. An experimental comparison of min-cut/max-flow algorithms for energy minimization in vision. IEEE Trans. Patt. Anal. Mach. Intell. 2004, 26, 1124–1137. [Google Scholar] [CrossRef] [PubMed]
  69. Xu, H. Modification of normalised difference water index (ndwi) to enhance open water features in remotely sensed imagery. Int. J. Remote Sens. 2006, 27, 3025–3033. [Google Scholar] [CrossRef]
  70. Lu, H.; Zhang, X.; Qi, J.; Tong, N.; Ruan, X.; Yang, M.-H. Co-bootstrapping saliency. IEEE Trans. Imag. Process. 2017, 26, 414–425. [Google Scholar] [CrossRef] [PubMed]
  71. He, K.; Sun, J.; Tang, X. Guided image filtering. IEEE Trans. Patt. Anal. Mach. Intell. 2013, 1397–1409. [Google Scholar] [CrossRef] [PubMed]
  72. Li, K.; Chen, Y. A genetic algorithm-based urban cluster automatic threshold method by combining viirs dnb, ndvi, and ndbi to monitor urbanization. Remote Sens. 2018, 10, 277. [Google Scholar] [CrossRef]
  73. Li, X.; Lu, H.; Zhang, L.; Ruan, X.; Yang, M.-H. Saliency detection via dense and sparse reconstruction. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013. [Google Scholar]
  74. Lou, J.; Ren, M.; Wang, H. Regional principal color based saliency detection. PLoS ONE 2014, 9, e112475. [Google Scholar] [CrossRef] [PubMed]
  75. Li, H.; Lu, H.; Lin, Z.; Shen, X.; Price, B. Inner and inter label propagation: Salient object detection in the wild. IEEE Trans. Imag. Process. 2015, 24, 3176–3186. [Google Scholar] [CrossRef] [PubMed]
  76. Zhou, L.; Yang, Z.; Zhou, Z.; Hu, D. Salient region detection using diffusion process on a two-layer sparse graph. IEEE Trans. Imag. Process. 2017, 26, 5882–5894. [Google Scholar] [CrossRef] [PubMed]
  77. Yuan, Y.; Li, C.; Kim, J.; Cai, W.; Feng, D.D. Reversion correction and regularized random walk ranking for saliency detection. IEEE Trans. Imag. Process. 2018, 27, 1311–1322. [Google Scholar] [CrossRef] [PubMed]
  78. Pesaresi, M.; Huadong, G.; Blaes, X.; Ehrlich, D.; Ferri, S.; Gueguen, L.; Halkia, M.; Kauffmann, M.; Kemper, T.; Lu, L. A global human settlement layer from optical hr/vhr rs data: Concept and first results. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 6, 2102–2131. [Google Scholar] [CrossRef]
  79. Zhang, L.; Lv, X.; Liang, X. Saliency analysis via hyperparameter sparse representation and energy distribution optimization for remote sensing images. Remote Sens. 2017, 9, 636. [Google Scholar] [CrossRef]
Figure 1. Flowchart of built-up area extraction.
Figure 1. Flowchart of built-up area extraction.
Remotesensing 10 01863 g001
Figure 2. (a) Segmentation result of SLIC (N = 100) (b) Segmentation result of SLIC (N = 1000) (c) Segmentation result of SLIC (N = 10,000) (d) Result of merging 20,000 superpixels into 2000 objects (e) Result of merging 20,000 superpixels into 4000 objects.
Figure 2. (a) Segmentation result of SLIC (N = 100) (b) Segmentation result of SLIC (N = 1000) (c) Segmentation result of SLIC (N = 10,000) (d) Result of merging 20,000 superpixels into 2000 objects (e) Result of merging 20,000 superpixels into 4000 objects.
Remotesensing 10 01863 g002
Figure 3. Study areas: Alxa (AL), Jinchang (JC), Wuwei (WW), Yulin (YL), Dalian (DL), Haikou (HK), Rizhao (RZ), Shanwei (SW), Chongqing (CQ), Jingzhou (JZ), Sanmenxia (SM), Wuhan (WH), Lanzhou (LZ), Lhasa (LS), Tianshui (TS), Xining (XN), Baoding (BD), Kaifeng (KF), Shangqiu (SQ), and Suzhou (SZ).
Figure 3. Study areas: Alxa (AL), Jinchang (JC), Wuwei (WW), Yulin (YL), Dalian (DL), Haikou (HK), Rizhao (RZ), Shanwei (SW), Chongqing (CQ), Jingzhou (JZ), Sanmenxia (SM), Wuhan (WH), Lanzhou (LZ), Lhasa (LS), Tianshui (TS), Xining (XN), Baoding (BD), Kaifeng (KF), Shangqiu (SQ), and Suzhou (SZ).
Remotesensing 10 01863 g003
Figure 4. Saliency maps produced by our proposed model and eight competing models. (a) Original images; (b) optimal band combinations; (c) ground truth; (d) dense and sparse reconstruction (DSR); (e) discriminative regional feature integration (DRFI); (f) regional principal color (RPC); (g) diffusion-based compactness and local contrast (DCLC); (h) inner and inter label propagation (LPS); (i) bootstrap learning (BL); (j) diffusion process on a two-layer sparse graph (DPTLSG); (k) reversion correction and regularized random walks ranking (RCRR); (l) Ours.
Figure 4. Saliency maps produced by our proposed model and eight competing models. (a) Original images; (b) optimal band combinations; (c) ground truth; (d) dense and sparse reconstruction (DSR); (e) discriminative regional feature integration (DRFI); (f) regional principal color (RPC); (g) diffusion-based compactness and local contrast (DCLC); (h) inner and inter label propagation (LPS); (i) bootstrap learning (BL); (j) diffusion process on a two-layer sparse graph (DPTLSG); (k) reversion correction and regularized random walks ranking (RCRR); (l) Ours.
Remotesensing 10 01863 g004aRemotesensing 10 01863 g004b
Figure 5. Quantitative evaluation results of different methods: (a) ROC curves of different methods on Sentinel-2 images (b) Precision, recall, and F-measure of different methods on Sentinel-2 images.
Figure 5. Quantitative evaluation results of different methods: (a) ROC curves of different methods on Sentinel-2 images (b) Precision, recall, and F-measure of different methods on Sentinel-2 images.
Remotesensing 10 01863 g005
Figure 6. Comparison of the results of the four methods. (a) RGB images, (b) ground truth, (c) NDBI maps, (d) NBI maps, (e) PanTex maps, (f) our saliency maps, (g) built-up areas maps (NDBI), (h) built-up areas maps (NBI), (i) built-up areas maps (PanTex), (j) built-up areas maps (proposed method).
Figure 6. Comparison of the results of the four methods. (a) RGB images, (b) ground truth, (c) NDBI maps, (d) NBI maps, (e) PanTex maps, (f) our saliency maps, (g) built-up areas maps (NDBI), (h) built-up areas maps (NBI), (i) built-up areas maps (PanTex), (j) built-up areas maps (proposed method).
Remotesensing 10 01863 g006aRemotesensing 10 01863 g006b
Table 1. Pseudo-code for the fractional-order Darwinian particle swarm optimization (FODPSO) algorithm.
Table 1. Pseudo-code for the fractional-order Darwinian particle swarm optimization (FODPSO) algorithm.
Start
Set Initial parameters vn [0], xn [0], χ1n [0], χ2n [0]
// vn is position parameter, xn is velocity parameter, χ1n is local best, χ2n is global best
for i = 1:1:Max. Number of the iteration
Generated swarm matrix
 for n = 1:1:Number of swarm matrix row
 Calculate fitness function of each row
 end
 Obtain min. fitness function’s parameter configuration
 If min. fitness function(i)<min. fitness function (i−1)
  Update χ 1 n i [t], χ 2 n i [t]
  Update v n i [t + 1], x n i [t + 1]
  else
  Kill all swarm matrix member
  Go to “generated swarm matrix”
  end
end
end
Table 2. Study areas.
Table 2. Study areas.
City TypeCitySize (pixels)Date
Desert citiesAlxa (Inner Mongolia)1000 × 90011/01/2016
Jinchang (Gansu)858 × 85811/24/2016
Wuwei (Gansu)1000 × 100011/04/2016
Yulin (Shanxi)1500 × 150004/01/2016
Coastal citiesDalian (Liaoning)1300 × 130011/21/2016
Haikou (Hainan)1250 × 125012/20/2016
Rizhao (Shandong)1000 × 100011/03/2016
Shanwei (Guangdong)500 × 50012/13/2016
Riverside citiesChongqing1650 × 165004/14/2017
Jinzhou (Hubei)1250 × 125008/01/2016
Sanmenxia (Henan)1000 × 100009/23/2016
Wuhan (Hubei)1800 × 120008/28/2016
Valley citiesLanzhou (Gansu)1750 × 90012/01/2016
Lhasa (Tibet)1500 × 150010/24/2016
Tianshui (Gansu)1020 × 102006/06/2017
Xining (Qinghai)1750 × 175011/04/2016
Plain citiesBaoding (Hebei)1400 × 140009/01/2016
Kaifeng (Henan)1200 × 120008/23/2016
Shangqiu (Henan)1250 × 125008/28/2016
Suzhou (Anhui)1200 × 120008/28/2016
Table 3. Optimal band combination.
Table 3. Optimal band combination.
Desert CityCitiesAlxaJinchangWuweiYulin
OptimalBands 12, 11, 7Bands 12, 11, 7Bands 12, 11, 5Bands 12, 11, 7
Coastal CityCitiesDalianHaikouShanweiRizhao
OptimalBands 12, 11, 7Bands 12, 11, 7Bands 12, 11, 7Bands 12, 11, 5
Riverside CitiesCitiesChongqingJingzhouSanmenxiaWuhan
OptimalBands 12, 11, 7Bands 12, 11, 7Bands 12, 11, 7Bands 12, 11, 7
Valley CitiesCitiesLanzhouLhasaTianshuiXining
OptimalBands 12, 11, 5Bands 12, 11, 5Bands 12, 11, 5Bands 12, 11, 7
Plain CitiesCitiesBaodingKaifengShangqingSuzhou
OptimalBands 12, 11, 7Bands 12, 11, 7Bands 12, 11, 7Bands 12, 11, 7
Table 4. The area under the curve (AUC) for different methods.
Table 4. The area under the curve (AUC) for different methods.
MethodDSRDRFIRPCDCLCLPSBLDPTLSGRCRROurs
AUC0.88900.87880.75210.91460.84850.83660.91670.84050.9687
Table 5. Running time comparisons for nine methods.
Table 5. Running time comparisons for nine methods.
MethodDSRDRFIRPCDCLCLPSBLDPTLSGRCRROurs
Time(s)178.331547.65514.23325.42137.92135.241583.001363.151226.22
Table 6. Comparison results of different M.
Table 6. Comparison results of different M.
MM = 3
(N = 1000, 1500, 2000)
M = 5
(N = 1000, 1500, 2000, 2500, 3000)
M = 7
(N = 1000, 1500, 2000, 2500, 3000, 3500, 4000)
AUC0.96450.96640.9687
F-Measure0.78530.78680.8038
Time(s)259.60626.421226.22
Table 7. Comparison results of different ϑ.
Table 7. Comparison results of different ϑ.
ϑ1.21.41.61.82.0
AUC0.95340.95810.95920.95970.9487
F-Measure0.71470.72120.72990.73780.7148
Table 8. Accuracy assessment of the resultant images.
Table 8. Accuracy assessment of the resultant images.
NDBINBIPanTexOurs
OACoOmOACoOmOACoOmOACoOm
Desert36.4090.8565.6642.5290.4864.3281.27 65.1833.8795.56 11.2526.26
Costal78.22 48.2730.5986.06 27.6627.1887.88 36.6636.3995.13 8.0411.17
Riverside80.96 38.6424.0784.51 29.3428.9281.57 37.2129.0393.3510.4917.84
Valley48.62 83.0148.8664.62 71.7243.3587.81 29.5425.9196.37 11.6818.79
Plain88.43 30.8716.8788.56 21.6140.4587.80 30.2722.2694.77 8.7217.97
OA = overall accuracy, Co. = commission error, Om = omission error.

Share and Cite

MDPI and ACS Style

Sun, Z.; Meng, Q.; Zhai, W. An Improved Boosting Learning Saliency Method for Built-Up Areas Extraction in Sentinel-2 Images. Remote Sens. 2018, 10, 1863. https://doi.org/10.3390/rs10121863

AMA Style

Sun Z, Meng Q, Zhai W. An Improved Boosting Learning Saliency Method for Built-Up Areas Extraction in Sentinel-2 Images. Remote Sensing. 2018; 10(12):1863. https://doi.org/10.3390/rs10121863

Chicago/Turabian Style

Sun, Zhenhui, Qingyan Meng, and Weifeng Zhai. 2018. "An Improved Boosting Learning Saliency Method for Built-Up Areas Extraction in Sentinel-2 Images" Remote Sensing 10, no. 12: 1863. https://doi.org/10.3390/rs10121863

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop