Next Article in Journal
Marine Resources Application Potential for Biotechnological Purposes
Next Article in Special Issue
On the Deployment of Out-of-the-Box Embedded Devices for Self-Powered River Surface Flow Velocity Monitoring at the Edge
Previous Article in Journal
Correction: Refaat et al. Multi-Stage Dynamic Transmission Network Expansion Planning Using LSHADE-SPACMA. Appl. Sci. 2021, 11, 2155
Previous Article in Special Issue
Combined Geodetic and Seismological Study of the December 2020 Mw = 4.6 Thiva (Central Greece) Shallow Earthquake
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparing Three Machine Learning Techniques for Building Extraction from a Digital Surface Model

by
Nicla Maria Notarangelo
,
Arianna Mazzariello
,
Raffaele Albano
* and
Aurelia Sole
School of Engineering, University of Basilicata, Via dell’Ateneo Lucano 10, 85100 Potenza, Italy
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(13), 6072; https://doi.org/10.3390/app11136072
Submission received: 30 May 2021 / Revised: 24 June 2021 / Accepted: 28 June 2021 / Published: 30 June 2021

Abstract

:
Automatic building extraction from high-resolution remotely sensed data is a major area of interest for an extensive range of fields (e.g., urban planning, environmental risk management) but challenging due to urban morphology complexity. Among the different methods proposed, the approaches based on supervised machine learning (ML) achieve the best results. This paper aims to investigate building footprint extraction using only high-resolution raster digital surface model (DSM) data by comparing the performance of three different popular supervised ML models on a benchmark dataset. The first two methods rely on a histogram of oriented gradients (HOG) feature descriptor and a classical ML (support vector machine (SVM)) or a shallow neural network (extreme learning machine (ELM)) classifier, and the third model is a fully convolutional network (FCN) based on deep learning with transfer learning. Used data were obtained from the International Society for Photogrammetry and Remote Sensing (ISPRS) and cover the urban areas of Vaihingen an der Enz, Potsdam, and Toronto. The results indicated that performances of models based on shallow ML (feature extraction and classifier training) are affected by the urban context investigated (F1 scores from 0.49 to 0.81), whereas the FCN-based model proved to be the most robust and best-performing method for building extraction from a high-resolution raster DSM (F1 scores from 0.80 to 0.86).

1. Introduction

In recent years, the availability of high-spatial-resolution remote sensing data has fostered the development of research methods and applications in several fields [1], such as urban planning, land monitoring, and environmental risk management [2,3,4,5] (e.g., urbanization delineation, vegetation mapping, flood modeling). The accurate information provided by the high-resolution data has been exploited both in large-scale problems, such as land use and land cover types [6,7], and small-scale ones, such as the extraction of urban objects: trees, roads, buildings, etc. [8,9,10]. Laser imaging detection and ranging (LiDAR) airborne mapping systems supply point clouds datasets containing 3-dimensional x, y, and z points and attributes to produce precise digital terrain model (DTM) and digital surface model (DSM) products—within a gridded or raster data format—in both natural and manmade environments [11].
Automatic building extraction from remotely sensed data is a major area of interest for an extensive range of applications but challenging due to difficulties in extracting precise boundaries because of urban morphology complexity [12]. Different methods have been proposed to address this issue [13], such as methods based on template matching, knowledge, object-based image analysis, and machine learning (ML) [14]. Commonly used template-matching-based approaches for automated buildings and urban objects footprints detection use “snake” or active contour model (ACM) [15] improvements [10,16,17,18] and integration with ML and deep learning (DL) [19]. The approaches based on supervised ML can achieve the best results, especially along with DL [14], a subset of ML based on neural networks with representation learning [20]. Traditional computer vision techniques require manually engineered feature descriptors for the desired object detection, whereas DL models automate the process of feature engineering [21,22]. Generally, classical ML-based object detection methods involve two steps: feature extraction and classifier training. Among features descriptors, histogram of oriented gradients (HOG) [23] features have proven to be effective in describing the edge or local shape information of the urban objects [14,24,25,26,27]. Typical ML algorithms for classifier training include, but are not limited to, the support vector machine (SVM) [7,28,29,30], artificial neural network (ANN) [31,32], including the extreme learning machine (ELM) [33] and AdaBoost. Several successful buildings and urban objects detection approaches are based on DL methods [10,34,35,36,37], such as convolutional networks, and in particular, fully convolutional networks (FCNs) have shown good performance on semantic segmentation [38,39,40,41,42], with correct pixel classification and accurate spatial information [43].
Despite the numerous algorithms proposed, building segmentation based only on the geometric information provided by DSM data is still a complicated task [44,45], mainly because objects with similar morphological characteristics and height can create ambiguity, resulting in position inaccuracy and local under-sampling [35,45].
A simplified method to detect and extract the building footprint based only on a DSM as input without any other additional feature could be beneficial in scenarios where only DSM data are available and a more expeditious solution is desirable (e.g., the rapid assessment of building damage). This paper aims to investigate the capability of three different, popular supervised ML models—namely HOG with SVM, HOG with ELM, and FCN—to detect building footprints using only raster DSM data as input and evaluate their performance on a publicly available benchmark dataset. This empirical comparison may highlight the potential usefulness of expeditious methods and help understand which model performs best in different urban environments. The first two methods rely on a HOG feature descriptor and a classical ML classifier (SVM) or a shallow neural network classifier (ELM) [46,47,48], and the FCN model is based on DL [49,50]. The descriptor and classification models were chosen for their accuracy [14,51], especially in segmentation tasks [52], popularity [14], simplicity, reduced computational time, and the ability to extract building footprint masks without reducing the resolution of raster data.

2. Materials and Methods

2.1. Overview

Figure 1 shows the general workflow of our study. The core steps included data retrieval and preparation, model implementation and training, and test performance evaluation. All the algorithms were implemented using the MATLAB environment, except the FCN that was implemented with the Python programming language and the Torch package. The experiments were conducted on a Fujitsu Celsius Workstation with Intel Xeon E5-2643 CPU (3.30 GHz) and 16 GB RAM. The following subsections describe each step.

2.2. Dataset Description

The data used in this study were obtained from the International Society for Photogrammetry and Remote Sensing (ISPRS) “Test Project on Urban Classification, 3-D Building Reconstruction, and Semantic Labeling”. The sets consisting of airborne images and laser scanner data were made publicly available to evaluate and compare different urban object extraction methods, providing benchmark datasets with ground truth [13,53], more updated and complete than the existing ones [54,55,56].
For the sake of generalizability, the available datasets were chosen covering three urban areas that differ in terms of urban morphology.
The first part of the dataset was captured over the city of Vaihingen an der Enz (Germany) and originally obtained from the digital aerial cameras tests carried out by the German Association of Photogrammetry and Remote Sensing (Deutsche Gesellschaft für Photogrammetrie, Fernerkundung und Geoinformation, DGPF) [57]. Each of the 33 different-sized patches (Figure 2) of the Vaihingen dataset contained a labeled true orthophoto (TOP) paired with a DSM [58] with a ground sampling distance of 9 cm. The publishers of the dataset provided a train/test split (30 patches for training, outlined with red, and 3 for testing, outlined with green). The morphology of Vaihingen presents the characteristics of a relatively small Central European village with many detached buildings and small multi-story buildings. It encompasses slightly different urban fabrics: relatively dense patterns formed by historic buildings with complex shapes and sparse trees, loose patterns formed by few high-rising residential buildings surrounded by trees, and regular patterns formed by small detached houses.
The second part of the dataset was captured over the city of Potsdam (Germany) and was originally obtained from the DGPF tests [58]. The Potsdam dataset consisted of 38 patches containing a labeled TOP paired with a DSM with a 6000 × 6000 pixel size and a ground-sampling distance of 5 cm. The provided train/test split for the Potsdam dataset (Figure 3) consisted of 35 patches for training (outlined with red) and 3 for testing (outlined with green). The morphology of Potsdam represents a typical Central European historic city with large building blocks, narrow streets, and a dense settlement structure.
The third part of the dataset covered the city of Toronto (Canada) and consisted of 3 different-sized patches containing a TOP and a DSM interpolated from the airborne laser scanner point cloud with a grid width of 25 cm. The given train/test split ratio for the Toronto dataset (Figure 4) was 2:1. The urban morphology of downtown Toronto exhibits the characteristics of a typical modern North American megacity as the presence of different urban objects and urban fabrics formed by a mixture of low- and high-story buildings with various degrees of shape complexity in rooftop structures and streets, including clusters of high-rise buildings that cast fairly large shadows [59].
The final database retrieved from the different ISPRS challenges was representative of the typical Central European and North American cities with respect to urban complexity, dimensions, and fabric density, especially suitable for method evaluation and comparison.
The supervised learning models require a labeled set of data to learn, during the training process, the underlying patterns that can be used to make predictions on novel data.
To solve the building detection task, the building footprints were considered as a group of pixels that can be distinguished from the pixels representing any other objects that may appear in an urban environment (roads, paved areas, vegetation, etc.) through discriminatory features extracted from the DSM raster images. Thus, the input to the models and test data were lists of DSM tiles and the corresponding binary maps (Figure 5) containing the ground truth for two semantic classes: building and non-building. For the Vaihingen and Potsdam sets, the binary masks were generated by selecting the building category from the provided ground truth (8-bit RGB tif files with one color per land cover class). For the Toronto set, the 8-bit binary masks were obtained by rasterizing the shapefiles describing the building outlines.
The preprocessing phase consisted of rescaling the raster data with an adequate factor and applying nearest interpolation (Figure 1). The initial mismatching resolutions—9 cm/pixel for the Vaihingen set, 5 cm/pixel for the Potsdam set, and 25 cm/pixel for the Toronto set—were conformed to a unique value of 36 cm/pixel to improve computational speed and performances. This value was obtained by progressively down-scaling a sample until a marked computational time reduction was seen without significant loss of information. The adjusted set of DSM tiles and masks was binned into training and test subsets.
Despite representing different and complex urban morphologies, the generated dataset showed a balanced class distribution. The number of pixels belonging to the building class and non-building class was close to the ideal 1:1 ratio (percentage difference << 0.5%) in every tile. Furthermore, the quality of the original datasets ensured the absence of uncertainty in labeling due to noise, low resolution, no-data pixels, inaccurate object edges, or class overlapping. Thus, the possible biases for the building extraction models were minimized, safeguarding the fairness of the test evaluation and the potential transferability of the learned knowledge.

2.3. Models’ Implementation and Training

2.3.1. Shallow ML-Based Building Detection

A shallow ML-based model that can classify pixels as building or non-building requires discriminative and computable engineered features for image segmentation and pixels classification.
Among the existing edge- and gradient-based descriptors, the HOG descriptors appeared especially suitable for detecting the construction footprints, as HOG detectors cue mainly on contours, can be computed quickly, and are fairly invariant to geometric transformations and occlusions [23].
The HOG feature extraction chain computes occurrences of gradient orientation in the detection window, or the region of interest (ROI) across the input, on a regular cell grid and uses overlapping local contrast normalization to enhance the accuracy. For every pixel in the input, the gradient vectors contain information on pixel value changes in x (gx) and y (gy) directions with respect to its four neighbors.
The attributes of the gradient are its magnitude
M x , y = g x 2 + g y 2
and its direction
θ = arctan g y / g x .
Gradient information is then pooled into a 1-D histogram of orientations that can be used as input for ML algorithms.
Thus, the feature extractor encoded the raster inputs into feature vectors to feed a classifier for the object/non-building instance detection, namely ELM and SVM classifiers.
The first method for building detection under examination used HOG as an input to the SVM [60], a supervised ML algorithm that produces accurate classifications of remotely sensed data [28,61,62,63]. The binary SVM classifier models a data point as a multi-dimensional vector belonging to one of two classes and constructs a hyperplane or set of hyperplanes to separate the classes [64] with the maximal margin (a space containing no observations) and the lowest misclassification [65,66,67].
The SVM classifier was fitted on the aforementioned processed training set to find the best-separating hyperplane (i.e., the decision boundary) using as input the predictor vectors Xj along with their class labels Y i { + 1 , 1 } . During the training process, sequential minimal optimization [68] solved the constrained optimization problem by breaking it into a series of sub-problems that could be solved analytically. The final model was then saved to perform prediction on the independent test set in the evaluation phase.
The second method combines the HOG and an ELM [47] classifier instead of the SVM [46]. The ELM trains a shallow feedforward neural network with a single hidden layer, i.e., the feature mapping, which does not need parameter tuning [69,70,71]. The main advantages of the ELM method compared to the previous one include better scalability and similar generalization at a faster learning speed [72].
The ELM classifier was fitted on the same training subset with a single hidden layer of 1000 neurons and a training ratio value of 0.9 to find the combination of nodes, weights, and biases minimizing the error between the actual output of the network (predictions) and the expected one (the ground truth) and obtain a learned model to be used for the test evaluation. The net started with random input weights and calculated the best values using the root-mean-square error to assess the prediction accuracy.

2.3.2. DL-Based Building Detection

The third building detection model was based on a DL architecture. The FCN-based semantic segmentation [50] classifies the building or non-building class for each pixel directly within the image inputs. The FCN gives a pixel-wise output (label map) without needing a hand-engineered feature vector to feed the classifier for building extraction of remotely sensed data [7,42,73,74,75].
The state-of-the-art classification convolutional model VGG-16 [76,77] structure was repurposed for the segmentation task employing the method fully described in [50]: adapting the original architecture into an FCN and transferring its learned representations by fine-tuning to the desired task. The repurposed architecture combines the semantic global information of deep, coarse layers to learn the feature hierarchy and the local knowledge of shallow, fine layers with 32-, 16-, and 8-pixel strides to improve the segmentation accuracy and enable pixel-wise predictions.
We fit the three models on a common portion of the dataset (the training set) by feature extraction and classifier training for the shallow ML-based models and fine-tuning through transfer learning for the DL model and then evaluated their discriminatory ability on the untouched data left (the test set).

2.4. Test Evaluation

The three implemented methods were fitted on a common dataset and then used to generate predictions on the same hold-out test set to obtan an unbiased estimate of each model’s accuracy.
Given the building label as the positive class and non-building label as the negative class assigned to every single pixel of the raster images, the true positives (TPs), false positives (FPs), true negatives (TNs), and false negatives (FNs) were counted to obtain the pixel-wise metrics for the evaluation of the binary classifiers [78,79] and a contingency map from a pixel-to-pixel comparison [32,80,81].
The TPs were the pixels correctly identified as building pixels; the FPs were non-building pixels wrongly labeled as belonging to the building class. Similarly, the TNs were non-building pixels correctly classified, whereas the FNs were building pixels wrongly classified as non-building pixels (undetected building pixels). Totaling the number of TPs, TNs, FPs, and FNs added up to the total pixels of the test set.
Sensitivity, or recall, is the proportion of TPs, pixels that actually are positive (building class). Sensitivity takes values in the range (0, 1); with higher sensitivity, fewer building pixels are undetected (larger footprints).
Sensitivity   or   Recall = TP TP + FN   ,
Specificity is the proportion of TNs, pixels that actually are negative (non-building class). Precision takes values in the range (0, 1); higher specificity leads to fewer pixels mislabeled as building.
Specificity = TN TN + FP   ,
The relationship between sensitivity and specificity was visualized using the receiver operating characteristic (ROC) curve and the area under the curve (AUC) [82] to quantify the performance of each classifier over its range of possible cut-offs (classification thresholds).
Precision or positive predicted value (PPV) is the proportion of relevant pixels (TP) among all the pixels classified as building pixels (positive) in the test. The PPV varies from 0 to 1, corresponding respectively to the value of the worst and the best classifier.
Precision   or   PPV = TP TP + FP   ,
Similarly, the negative prediction value (NPV) represents the proportion of pixels with accurate non-building test results (TN) among all the pixels identified as non-building pixels (negative class). NPV reference values are 0 for the worst classification and 1 for the best classification possible.
NPV = TN TN + FN   ,
The F1 score combines the precision and recall values by taking their harmonic mean. The F1 score takes values in the range (0, 1); higher F1 values correspond to better model performance.
F 1   score = 2 × Precision × Recall Precision + Recall   ,
The mean-square error (MSE) evaluates the mean of the quadratic prediction errors; lower values indicate better performance.
MSE = 1 N pixels imgref     imgout 2   ,
The Matthews correlation coefficient (MCC) effectively and reliably measures the quality of binary classifications [83], even if the classes are unbalanced, by taking into account true and false positives and negatives. MCC values range from the worst value −1 to the best value +1.
MCC = TP   ×   TN   FP   ×   FN TP + FP TP + FN TN + FP TN + FN   ,

3. Results

We reported pixel-based metrics to evaluate the prediction error for each of the presented models for the building detection task in different urban areas.
Figure 6 shows the contingency maps resulting from the pixel-wise results of the binary classification with the SVM, EML, and FCN superimposed on the aerial images of the considered areas: the building pixels correctly detected (TPs) are colored in yellow, FPs are colored in red, and FNs are colored in blue. As seen from examples in Figure 6, ELM tended to underestimate the pixels belonging to the building class, since a fairly high rate of FNs was produced. The resulting segmentation images were not close to reality in the Vaihingen and Toronto patches: in the first case, the smaller footprints were undetected; in the second case, there were “holes” in some footprints, having building pixels surrounding non-building areas, which is a characteristic of urban fabrics formed by specific architectural typologies (e.g., courtyard, siheyuan, and patio buildings). However, the SVM seemed in the Vaihingen case more prone to overestimating the number of building pixels due to the higher occurrence of FPs. The contingency maps produced by the FCN model showed a better correspondence of the predicted footprint position and size to ground-truth masks in all subsets.
For quantitative evaluation of the different classifiers, we reported pixel-wise classification accuracy (Table 1) in terms of sensitivity, specificity, precision, the NPV, F1 scores, the MSE, the MCC, and the AUC using both the aggregate metrics and the metrics split by area.
Considering the Vaihingen test set, MCC values of the HOG-SVM and HOG-ELM models (0.657and 0.513, respectively) were significantly lower than the value of the FCN model (0.833); compared with the ELM classifier, the SVM was more prone to detecting building pixels as it disclosed higher sensitivity (77.3% vs. 41.5%), but the precision was lower (44.6% vs. 60.8%). The metrics relative to non-building pixels were not found to differ markedly. Considering the Potsdam test set, the HOG-SVM and HOG-ELM models showed similarly better outcomes in terms of sensitivity but were slightly outperformed by the FCN model; the difference became greater considering the MCC, a metric that provides a balanced measure of the relationship between reality and prediction. Considering the Toronto test set, however, the MCC values of the HOG-SVM and FCN models were significantly higher than the value of the HOG-ELM model (0.826 and 0.821 vs. 0.664), which suffered from lower sensitivity and NPV due to the presence of small clusters of FN pixels inside the clusters of pixels representing the building footprints, as shown in Figure 6.
Evaluating the aggregated metrics, the best values were achieved by the FCN-based classifier: the FCN-based building detection method exhibited good overall detection reliability. It was found that such an approach produces good-quality results in the different urban contexts considered, as demonstrated by the values of the F1 score (0.831), AUC (0.960), and MCC (0.834), which are fairly close to the ideal value of 1.
Figure 7 and Figure 8 illustrate the detection abilities of the three binary classifiers comparing ROC curves calculated on the output scores. The ROC curves of the SVM-based (blue lines), ELM-based (red lines), and FCN-based (orange line) classifiers plotted the dependency of the true-positive rate (sensitivity) on the false-positive rate (1—specificity) obtained at various thresholds.
Globally, all the final models could successfully detect the building footprints within the study areas using only DSM patches as input, presenting good predictive ability with AUC > 0.8 [84].
The SVM-based classifier produced accurate predictions in Toronto (AUC = 0.96), fairly good predictions in Potsdam (AUC = 0.83), but minor predictions in Vaihingen (AUC = 0.78) patches. The ELM-based classifier performance varied according to the considered urban area: the prediction accuracy on the Potsdam test set slightly surpassed the SVM accuracy (AUC = 0.84 vs. 0.83), whereas the test results on Vaihingen and Toronto datasets were the poorest among the considered methods, albeit being still acceptable (AUC = 0.71 and 0.77). The FCN curves were close to the “perfect” classifier, with AUC values between 0.95 and 0.97 for each urban area.

4. Discussion

The study compared the performance of three common supervised ML techniques—HOG with SVM, HOG with ELM, and FCN—for the task of building footprint extraction from high-resolution DSM data on a benchmark dataset of DSMs of three different urban contexts (Vaihingen an der Enz, Potsdam, and Toronto). Our results showed that all the ML techniques could successfully complete the task, but the FCN appeared more robust to urban fabric diversity.
The performance of both HOG-based models was influenced by the urban context investigated and in particular decreased within the Vaihingen area. This area is more challenging due to the complexity of the urban fabrics that are composed of many small buildings varying in shape, sparse low vegetation and trees, and irregular narrow road networks. The improvements in the Potsdam area could be attributed to the larger footprints of the constructions and the relatively reduced number of trees. The worst generalization ability of the two methods based on classical feature learning and classifier training may be caused by the size of the training dataset [72,85], confirming the worse performance of the ELM classifier compared to the SVM on small datasets [72].
In summary, these results show that a DL approach based on FCNs is the most preferable method as it achieves good classification regardless of the urban context, despite the inaccuracy in contours and boundaries that is a drawback of DL-based segmentation [86], and support evidence from previous studies [35,36,38,39,41,85]. Furthermore, such an approach automates the feature engineering and benefits from transfer learning that limits the impacts of the training data size.
However, with a limited data size, although sufficient for the comparison, caution must be applied, as the findings might not be fully appliable in the case of large datasets or urban scenarios with substantially different morphological structures, e.g., Asian cities [87].
Table 2 summarizes the main advantages and disadvantages of the three methods.

5. Conclusions

In this paper, we performed an empirical comparison of three different supervised ML-based building detection methods—HOG-SVM, HOG-ELM, and FCN models—on a benchmark high-resolution remotely sensed dataset using only raster DSMs. Two of these methods belong to the classical feature learning and classifier training category—shallow ML—whereas the third is a DL network. We used HOG as a feature descriptor and trained the classifiers (SVM and ELM) for the first two methods using publicly available ISPRS datasets. The same data were employed for fine-tuning the FCN architecture through transfer learning. The high quality of the publicly available data resolved the major problem of correctly labeling the training data, on which depend the predictive skills of the final trained models. The methods are easy to implement, as analogous functions are widely accessible in both proprietary and open source software and programming languages (e.g., MATLAB, Python, R).
Our results demonstrate that determining the footprint of buildings from remotely sensed data can produce different results, depending not only on the urban morphology of the context to be surveyed but also on the model choice.
The performances of building detection techniques based on shallow ML were affected by the complexity of the urban context considered, in particular by the presence of vegetation and smaller footprints. The FCN-based model has proven to be the most robust and best-performing method for building extraction from high-resolution DSM data. Furthermore, this DL technique can generate accurate building masks without any manually engineered features with high transferability potential. Due to this, the model has the potential to solve similar pixel classification tasks, such as extraction of a different class (e.g., ground surfaces, vegetation, cars) or multi-class segmentation by being re-trained with adequate ground-truth masks and classes number. Using solely a DSM as input data increases portability between urban areas as such data are widely available, continuously improved, and constantly released [88].
Future work could investigate the potentials in the aforementioned multi-class problem domains, i.e., multi-class semantic segmentation of urban areas, as well as a systematic analysis of the impacts on the accuracy of the raster resolution variation for applications in data-poor environments or at a larger scale. To better explore the robustness to urban fabric heterogeneity and geographic transferability, future experiments may also include the application in different kinds of settlements (e.g., informal, regional, and vernacular settlements) that can present different morphological characteristics. A metrics ensemble based on both raster resolution and urban morphology could also lead to a more complete characterization of the advantages and disadvantages of each algorithm. The knowledge of the urban context, analysis of urban objects, color information from orthophotos [34], and additional data could be used to develop novel methods for classification, pre-processing the inputs, or post-processing the outputs. Thus, studies on combinations of different approaches and data to improve the FCN detection accuracy would be worthwhile.

Author Contributions

Conceptualization, R.A.; methodology, R.A.; software, R.A.; validation, N.M.N. and A.M.; formal analysis, N.M.N.; investigation, A.M. and N.M.N.; data curation, R.A.; writing—original draft preparation, N.M.N., A.M., R.A. and A.S.; writing—review and editing, N.M.N., A.M., R.A. and A.S.; visualization, A.M. and N.M.N.; supervision, R.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The used datasets were created by the ISPRS in the “Test Project on Urban Classification, 3-D Building Reconstruction, and Semantic Labeling” and can be required at https://www2.isprs.org/commissions/comm2/wg4/benchmark/ (accessed on 8 October 2019).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Demir, I.; Koperski, K.; Lindenbaum, D.; Pang, G.; Huang, J.; Basu, S.; Hughes, F.; Tuia, D.; Raskar, R. DeepGlobe 2018: A Challenge to Parse the Earth through Satellite Images. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 172–17209. [Google Scholar]
  2. Li, L.; Liang, J.; Weng, M.; Zhu, H. A Multiple-Feature Reuse Network to Extract Buildings from Remote Sensing Imagery. Remote Sens. 2018, 10, 1350. [Google Scholar] [CrossRef] [Green Version]
  3. Samela, C.; Albano, R.; Sole, A.; Manfreda, S. A GIS Tool for Cost-Effective Delineation of Flood-Prone Areas. Comput. Environ. Urban Syst. 2018, 70, 43–52. [Google Scholar] [CrossRef]
  4. Sole, A.; Giosa, L.; Albano, R.; Cantisani, A. The laser scan data as a key element in the hydraulic flood modelling in urban areas. In Proceedings of the International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences—(ISPRS Archive), London, UK, 29–31 May 2013; Volume XL-4/W1, pp. 65–70. [Google Scholar] [CrossRef] [Green Version]
  5. Tian, J.; Cui, S.; Reinartz, P. Building Change Detection Based on Satellite Stereo Imagery and Digital Surface Models. IEEE Trans. Geosci. Remote Sens. 2014, 52, 406–417. [Google Scholar] [CrossRef] [Green Version]
  6. Li, W.; Dong, R.; Fu, H.; Yu, L. Large-Scale Oil Palm Tree Detection from High-Resolution Satellite Images Using Two-Stage Convolutional Neural Networks. Remote Sens. 2018, 11, 11. [Google Scholar] [CrossRef] [Green Version]
  7. Zhang, B.; Wang, C.; Shen, Y.; Liu, Y. Fully Connected Conditional Random Fields for High-Resolution Remote Sensing Land Use/Land Cover Classification with Convolutional Neural Networks. Remote Sens. 2018, 10, 1889. [Google Scholar] [CrossRef] [Green Version]
  8. Audebert, N.; Le Saux, B.; Lefèvre, S. Segment-before-Detect: Vehicle Detection and Classification through Semantic Segmentation of Aerial Images. Remote Sens. 2017, 9, 368. [Google Scholar] [CrossRef] [Green Version]
  9. Tang, T.; Zhou, S.; Deng, A.; Lei, L.; Zou, H. Arbitrary-Oriented Vehicle Detection in Aerial Imagery with Single Convolutional Neural Networks. Remote Sens. 2017, 9, 1170. [Google Scholar] [CrossRef] [Green Version]
  10. Xu, Y.; Xie, Z.; Feng, Y.; Chen, Z. Road Extraction from High-Resolution Remote Sensing Imagery Using Deep Learning. Remote Sens. 2018, 10, 1461. [Google Scholar] [CrossRef] [Green Version]
  11. Guo, B.; Huang, X.; Zhang, F.; Sohn, G. Classification of Airborne Laser Scanning Data Using JointBoost. ISPRS J. Photogramm. Remote Sens. 2015, 100, 71–83. [Google Scholar] [CrossRef]
  12. Bretar, F. Feature Extraction from LiDAR Data in Urban Areas. In Topographic Laser Ranging and Scanning; Shan, J., Toth, C.K., Eds.; CRC Press: Boca Raton, FL, USA, 2017; pp. 403–420. ISBN 978-1-315-21956-1. [Google Scholar]
  13. Rottensteiner, F.; Sohn, G.; Gerke, M.; Wegner, J.D.; Breitkopf, U.; Jung, J. Results of the ISPRS Benchmark on Urban Object Detection and 3D Building Reconstruction. ISPRS J. Photogramm. Remote Sens. 2014, 93, 256–271. [Google Scholar] [CrossRef]
  14. Cheng, G.; Han, J. A Survey on Object Detection in Optical Remote Sensing Images. ISPRS J. Photogramm. Remote Sens. 2016, 117, 11–28. [Google Scholar] [CrossRef] [Green Version]
  15. Kass, M.; Witkin, A.; Terzopoulos, D. Snakes: Active Contour Models. Int. J. Comput. Vis. 1988, 1, 321–331. [Google Scholar] [CrossRef]
  16. Bypina, S.K.; Rajan, K.S. Semi-Automatic Extraction of Large and Moderate Buildings from Very High-Resolution Satellite Imagery Using Active Contour Model. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1885–1888. [Google Scholar]
  17. Kabolizade, M.; Ebadi, H.; Ahmadi, S. An Improved Snake Model for Automatic Extraction of Buildings from Urban Aerial Images and LiDAR Data. Comput. Environ. Urban Syst. 2010, 34, 435–441. [Google Scholar] [CrossRef]
  18. Liasis, G.; Stavrou, S. Building Extraction in Satellite Images Using Active Contours and Colour Features. Int. J. Remote Sens. 2016, 37, 1127–1153. [Google Scholar] [CrossRef]
  19. Sun, Y.; Zhang, X.; Zhao, X.; Xin, Q. Extracting Building Boundaries from High Resolution Optical Images and LiDAR Data by Integrating the Convolutional Neural Network and the Active Contour Model. Remote Sens. 2018, 10, 1459. [Google Scholar] [CrossRef] [Green Version]
  20. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  21. Notarangelo, N.M.; Hirano, K.; Albano, R.; Sole, A. Transfer Learning with Convolutional Neural Networks for Rainfall Detection in Single Images. Water 2021, 13, 588. [Google Scholar] [CrossRef]
  22. O’Mahony, N.; Campbell, S.; Carvalho, A.; Harapanahalli, S.; Hernandez, G.V.; Krpalkova, L.; Riordan, D.; Walsh, J. Deep Learning vs. Traditional Computer Vision. In Advances in Computer Vision; Arai, K., Kapoor, S., Eds.; Advances in Intelligent Systems and Computing; Springer International Publishing: Cham, Switzerland, 2020; Volume 943, pp. 128–144. ISBN 978-3-030-17794-2. [Google Scholar]
  23. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; IEEE: Piscataway, NJ, USA, 2005; Volume 1, pp. 886–893. [Google Scholar]
  24. Konstantinidis, D.; Stathaki, T.; Argyriou, V.; Grammalidis, N. A Probabilistic Feature Fusion for Building Detection in Satellite Images. In Proceedings of the 10th International Conference on Computer Vision Theory and Applications, Berlin, Germany, 11–14 March 2015; SCITEPRESS—Science and Technology Publications: Setúbal, Portugal, 2015; pp. 205–212. [Google Scholar]
  25. Ilsever, M.; Unsalan, C. Building Detection Using HOG Descriptors. In Proceedings of the 2013 6th International Conference on Recent Advances in Space Technologies (RAST), Istanbul, Turkey, 12–14 June 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 115–119. [Google Scholar]
  26. Cheng, G.; Han, J.; Zhou, P.; Guo, L. Scalable Multi-Class Geospatial Object Detection in High-Spatial-Resolution Remote Sensing Images. In Proceedings of the 2014 IEEE Geoscience and Remote Sensing Symposium, Quebec City, QC, Canada, 13–18 July 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 2479–2482. [Google Scholar]
  27. Tuermer, S.; Kurz, F.; Reinartz, P.; Stilla, U. Airborne Vehicle Detection in Dense Urban Areas Using HoG Features and Disparity Maps. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 6, 2327–2337. [Google Scholar] [CrossRef]
  28. Turker, M.; Koc-San, D. Building Extraction from High-Resolution Optical Spaceborne Images Using the Integration of Support Vector Machine (SVM) Classification, Hough Transformation and Perceptual Grouping. Int. J. Appl. Earth Obs. Geoinf. 2015, 34, 58–69. [Google Scholar] [CrossRef]
  29. Krupinski, M.; Lewiński, S.; Malinowski, R. One Class SVM for Building Detection on Sentinel-2 Images. In Proceedings of the Photonics Applications in Astronomy, Communications, Industry, and High-Energy Physics Experiments, Wilga, Poland, 6 November 2019; Romaniuk, R.S., Linczuk, M., Eds.; SPIE: Bellingham, WA, USA, 2019; p. 6. [Google Scholar]
  30. Mountrakis, G.; Im, J.; Ogole, C. Support Vector Machines in Remote Sensing: A Review. ISPRS J. Photogramm. Remote Sens. 2011, 66, 247–259. [Google Scholar] [CrossRef]
  31. Lari, Z.; Ebadi, H. Automated Building Extraction from High-Resolution Satellite Imagery Using Spectral and Structural Information Based on Artificial Neural Networks. In Proceedings of the International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences, Hannover, Germany, 29 May–1 June 2007; Volume 36. [Google Scholar]
  32. Albano, R.; Samela, C.; Crăciun, I.; Manfreda, S.; Adamowski, J.; Sole, A.; Sivertun, Å.; Ozunu, A. Large Scale Flood Risk Mapping in Data Scarce Environments: An Application for Romania. Water 2020, 12, 1834. [Google Scholar] [CrossRef]
  33. Wang, P.; Zhang, X.; Hao, Y. A Method Combining CNN and ELM for Feature Extraction and Classification of SAR Image. J. Sens. 2019, 2019, 1–8. [Google Scholar] [CrossRef]
  34. Nahhas, F.H.; Shafri, H.Z.M.; Sameen, M.I.; Pradhan, B.; Mansor, S. Deep Learning Approach for Building Detection Using LiDAR–Orthophoto Fusion. J. Sens. 2018, 2018, 1–12. [Google Scholar] [CrossRef] [Green Version]
  35. Maltezos, E.; Doulamis, A.; Doulamis, N.; Ioannidis, C. Building Extraction from LiDAR Data Applying Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2019, 16, 155–159. [Google Scholar] [CrossRef]
  36. Vakalopoulou, M.; Karantzalos, K.; Komodakis, N.; Paragios, N. Building Detection in Very High Resolution Multispectral Data with Deep Learning Features. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1873–1876. [Google Scholar]
  37. Diakogiannis, F.I.; Waldner, F.; Caccetta, P.; Wu, C. ResUNet-a: A Deep Learning Framework for Semantic Segmentation of Remotely Sensed Data. ISPRS J. Photogramm. Remote Sens. 2020, 162, 94–114. [Google Scholar] [CrossRef] [Green Version]
  38. Zhong, Z.; Li, J.; Cui, W.; Jiang, H. Fully Convolutional Networks for Building and Road Extraction: Preliminary Results. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1591–1594. [Google Scholar]
  39. Xu, Y.; Wu, L.; Xie, Z.; Chen, Z. Building Extraction in Very High Resolution Remote Sensing Imagery Using Deep Learning and Guided Filters. Remote Sens. 2018, 10, 144. [Google Scholar] [CrossRef] [Green Version]
  40. Fu, G.; Liu, C.; Zhou, R.; Sun, T.; Zhang, Q. Classification for High Resolution Remote Sensing Imagery Using a Fully Convolutional Network. Remote Sens. 2017, 9, 498. [Google Scholar] [CrossRef] [Green Version]
  41. Maggiori, E.; Tarabalka, Y.; Charpiat, G.; Alliez, P. Convolutional Neural Networks for Large-Scale Remote-Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 645–657. [Google Scholar] [CrossRef] [Green Version]
  42. Shrestha, S.; Vanneschi, L. Improved Fully Convolutional Network with Conditional Random Fields for Building Extraction. Remote Sens. 2018, 10, 1135. [Google Scholar] [CrossRef] [Green Version]
  43. Audebert, N.; Saux, B.L.; Lefèvre, S. Semantic Segmentation of Earth Observation Data Using Multimodal and Multi-Scale Deep Networks. arXiv 2016, arXiv:1609.06846. [Google Scholar]
  44. Arefi, H.; Reinartz, P. Building Reconstruction Using DSM and Orthorectified Images. Remote Sens. 2013, 5, 1681–1703. [Google Scholar] [CrossRef] [Green Version]
  45. Protopapadakis, E.; Doulamis, A.; Doulamis, N.; Maltezos, E. Stacked Autoencoders Driven by Semi-Supervised Learning for Building Extraction from near Infrared Remote Sensing Imagery. Remote Sens. 2021, 13, 371. [Google Scholar] [CrossRef]
  46. Chorowski, J.; Wang, J.; Zurada, J.M. Review and Performance Comparison of SVM- and ELM-Based Classifiers. Neurocomputing 2014, 128, 507–516. [Google Scholar] [CrossRef]
  47. Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme Learning Machine for Regression and Multiclass Classification. IEEE Trans. Syst. Man Cybern. B 2012, 42, 513–529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Yang, L.; Lyu, K.; Li, H.; Liu, Y. Building Climate Zoning in China Using Supervised Classification-Based Machine Learning. Build. Environ. 2020, 171, 106663. [Google Scholar] [CrossRef]
  49. Khan, M.A. HCRNNIDS: Hybrid Convolutional Recurrent Neural Network-Based Network Intrusion Detection System. Processes 2021, 9, 834. [Google Scholar] [CrossRef]
  50. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 3431–3440. [Google Scholar]
  51. Kotsiantis, S. Supervised Machine Learning: A Review of Classification Techniques. Informatica 2007, 31, 249–268. [Google Scholar]
  52. Minaee, S.; Boykov, Y.; Porikli, F.; Plaza, A.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. arXiv 2020, arXiv:2001.05566. [Google Scholar]
  53. Rottensteiner, F.; Sohn, G.; Jung, J.; Gerke, M.; Baillard, C.; Benitez, S.; Breitkopf, U. The ISPRS Benchmark on Urban Object Classification and 3d Building Reconstruction. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. I-3 2012, 1, 293–298. [Google Scholar] [CrossRef] [Green Version]
  54. Champion, N.; Rottensteiner, F.; Matikainen, L.; Liang, X.; Hyyppä, J.; Olsen, B. A Test of Automatic Building Change Detection Approaches. In Proceedings of the CMRT09, Paris, France, 3–4 September 2009. [Google Scholar]
  55. Kaartinen, H.; Hyyppä, J.; Gülch, E.; Vosselman, G.; Hyyppä, H.; Matikainen, L.; Hofmann, A.; Mäder, U.; Persson, Å.; Söderman, U.; et al. Accuracy of 3D City Models: EuroSDR Comparison. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2005, 36, 227–232. [Google Scholar]
  56. Mayer, H.; Hinz, S.; Bacher, U.; Baltsavias, E. A Test of Automatic Road Extraction Approaches. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2006, 36, 209–214. [Google Scholar]
  57. Cramer, M. The DGPF-Test on Digital Airborne Camera Evaluation Overview and Test Design. PFG Photogramm. Fernerkund. Geoinf. 2010, 2010, 73–82. [Google Scholar] [CrossRef]
  58. Haala, N.; Cramer, M.; Jacobsen, K. The German Camera Evaluation Project-Results from the Geometry Group. In Proceedings of the International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences: [2010 Canadian Geomatics Conference and Symposium of Commission I, ISPRS Convergence in Geomatics-Shaping Canada’s Competitive Landscape], Calgary, AB, Canada, 15–18 June 2010; Nr. Part 1; Copernicus GmbH: Göttingen, Germany, 2010; Volume 38. [Google Scholar]
  59. Albano, R. Investigation on Roof Segmentation for 3D Building Reconstruction from Aerial LIDAR Point Clouds. Appl. Sci. 2019, 9, 4674. [Google Scholar] [CrossRef] [Green Version]
  60. Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods, 1st ed.; Cambridge University Press: Cambridge, UK, 2000; ISBN 978-0-521-78019-3. [Google Scholar]
  61. Durst, N.J.; Sullivan, E.; Huang, H.; Park, H. Building Footprint-Derived Landscape Metrics for the Identification of Informal Subdivisions and Manufactured Home Communities: A Pilot Application in Hidalgo County, Texas. Land Use Policy 2021, 101, 105158. [Google Scholar] [CrossRef]
  62. Foody, G.M.; Mathur, A. A Relative Evaluation of Multiclass Image Classification by Support Vector Machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1335–1343. [Google Scholar] [CrossRef] [Green Version]
  63. Huang, C.; Davis, L.S.; Townshend, J.R.G. An Assessment of Support Vector Machines for Land Cover Classification. Int. J. Remote Sens. 2002, 23, 725–749. [Google Scholar] [CrossRef]
  64. Schölkopf, B.; Smola, A.J.; Bach, F. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; MIT Press: Cambridge, MA, USA, 2002. [Google Scholar]
  65. Guenther, N.; Schonlau, M. Support Vector Machines. Stata J. 2016, 16, 917–937. [Google Scholar] [CrossRef] [Green Version]
  66. Vapnik, V. The Nature of Statistical Learning Theory; Springer Science & Business Media: Berlin, Germany, 2013. [Google Scholar]
  67. Wang, L. Support Vector Machines: Theory and Applications; Springer Science & Business Media: Berlin, Germany, 2005; Volume 177. [Google Scholar]
  68. Fan, R.-E.; Chen, P.-H.; Lin, C.-J. Working Set Selection Using Second Order Information for Training Support Vector Machines. J. Mach. Learn. Res. 2005, 6, 1889–1918. [Google Scholar] [CrossRef]
  69. Ertuğrul, Ö.F.; Kaya, Y. A Detailed Analysis on Extreme Learning Machine and Novel Approaches Based on ELM. Am. J. Comput. Sci. Eng. 2014, 1, 43–50. [Google Scholar]
  70. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme Learning Machine: A New Learning Scheme of Feedforward Neural Networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541), Budapest, Hungary, 25–29 July 2004; IEEE: Piscataway, NJ, USA, 2004; Volume 2, pp. 985–990. [Google Scholar]
  71. Huang, G.-B. An Insight into Extreme Learning Machines: Random Neurons, Random Features and Kernels. Cogn. Comput. 2014, 6, 376–390. [Google Scholar] [CrossRef]
  72. Liu, X.; Gao, C.; Li, P. A Comparative Analysis of Support Vector Machines and Extreme Learning Machines. Neural Netw. 2012, 33, 58–66. [Google Scholar] [CrossRef]
  73. Bischke, B.; Helber, P.; Folz, J.; Borth, D.; Dengel, A. Multi-Task Learning for Segmentation of Building Footprints with Deep Neural Networks. arXiv 2017, arXiv:1709.05932. [Google Scholar]
  74. Li, Y.; He, B.; Long, T.; Bai, X. Evaluation the Performance of Fully Convolutional Networks for Building Extraction Compared with Shallow Models. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 850–853. [Google Scholar]
  75. Liu, T.; Abd-Elrahman, A. An Object-Based Image Analysis Method for Enhancing Classification of Land Covers Using Fully Convolutional Networks and Multi-View Images of Small Unmanned Aerial System. Remote Sens. 2018, 10, 457. [Google Scholar] [CrossRef] [Green Version]
  76. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  77. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  78. Murphy, K.P. Machine Learning: A Probabilistic Perspective; Adaptive Computation and Machine LEARNING series; MIT Press: Cambridge, MA, USA, 2012; ISBN 978-0-262-01802-9. [Google Scholar]
  79. Zheng, A. Evaluating Machine Learning Models; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2015; ISBN 978-1-4920-4875-6. [Google Scholar]
  80. Scarpino, S.; Albano, R.; Cantisani, A.; Mancusi, L.; Sole, A.; Milillo, G. Multitemporal SAR Data and 2D Hydrodynamic Model Flood Scenario Dynamics Assessment. IJGI 2018, 7, 105. [Google Scholar] [CrossRef] [Green Version]
  81. Albano, R.; Mancusi, L.; Adamowski, J.; Cantisani, A.; Sole, A. A GIS Tool for Mapping Dam-Break Flood Hazards in Italy. IJGI 2019, 8, 250. [Google Scholar] [CrossRef] [Green Version]
  82. Powers, D.M.W. Evaluation: From Precision, Recall and F-Measure to ROC, Informedness, Markedness and Correlation. arXiv 2020, arXiv:2010.16061. [Google Scholar]
  83. Chicco, D.; Jurman, G. The Advantages of the Matthews Correlation Coefficient (MCC) over F1 Score and Accuracy in Binary Classification Evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef] [Green Version]
  84. Hosmer, D.W.; Lemeshow, S. Applied Logistic Regression; John Wiley & Sons: New York, NY, USA; Toronto, ON, Canada, 2005; ISBN 978-0-471-72214-4. [Google Scholar]
  85. Liu, T.; Abd-Elrahman, A.; Morton, J.; Wilhelm, V.L. Comparing Fully Convolutional Networks, Random Forest, Support Vector Machine, and Patch-Based Deep Convolutional Neural Networks for Object-Based Wetland Mapping Using Images from Small Unmanned Aircraft System. GISci. Remote Sens. 2018, 55, 243–264. [Google Scholar] [CrossRef]
  86. Yang, Y.; Zhou, X.; Liu, Y.; Hu, Z.; Ding, F. Wood Defect Detection Based on Depth Extreme Learning Machine. Appl. Sci. 2020, 10, 7488. [Google Scholar] [CrossRef]
  87. Chen, T.-L.; Chiu, H.-W.; Lin, Y.-F. How Do East and Southeast Asian Cities Differ from Western Cities? A Systematic Review of the Urban Form Characteristics. Sustainability 2020, 12, 2423. [Google Scholar] [CrossRef] [Green Version]
  88. Uuemaa, E.; Ahi, S.; Montibeller, B.; Muru, M.; Kmoch, A. Vertical Accuracy of Freely Available Global Digital Elevation Models (ASTER, AW3D30, MERIT, TanDEM-X, SRTM, and NASADEM). Remote Sens. 2020, 12, 3482. [Google Scholar] [CrossRef]
Figure 1. Overview of the study workflow steps: data retrieval and preparation (labeling, resizing, and train/test split), model implementation and training (building extraction), and test performance evaluation.
Figure 1. Overview of the study workflow steps: data retrieval and preparation (labeling, resizing, and train/test split), model implementation and training (building extraction), and test performance evaluation.
Applsci 11 06072 g001
Figure 2. Digital surface model (DSM) patches of the Vaihingen an der Enz dataset: training (red) and test (green) subsets.
Figure 2. Digital surface model (DSM) patches of the Vaihingen an der Enz dataset: training (red) and test (green) subsets.
Applsci 11 06072 g002
Figure 3. Digital surface model (DSM) patches of the Potsdam dataset: training (red) and test (green) subsets.
Figure 3. Digital surface model (DSM) patches of the Potsdam dataset: training (red) and test (green) subsets.
Applsci 11 06072 g003
Figure 4. Digital surface model (DSM) patches of the Toronto dataset: training (red) and test (green) subsets.
Figure 4. Digital surface model (DSM) patches of the Toronto dataset: training (red) and test (green) subsets.
Applsci 11 06072 g004
Figure 5. Sample of the dataset of digital surface model (DSM) patches and corresponding binary masks (ground truth): (a) Vaihingen test set, (b) Potsdam test set, and (c) Toronto test set.
Figure 5. Sample of the dataset of digital surface model (DSM) patches and corresponding binary masks (ground truth): (a) Vaihingen test set, (b) Potsdam test set, and (c) Toronto test set.
Applsci 11 06072 g005
Figure 6. Contingency maps obtained from a pixel-to-pixel comparison between the masks outputted by the models and the ground truth by urban area: (a) Vaihingen test set, (b) Potsdam test set, and (c) Toronto test set. True positives are shown in yellow, false positives in red, and false negatives (missed building pixels) in blue.
Figure 6. Contingency maps obtained from a pixel-to-pixel comparison between the masks outputted by the models and the ground truth by urban area: (a) Vaihingen test set, (b) Potsdam test set, and (c) Toronto test set. True positives are shown in yellow, false positives in red, and false negatives (missed building pixels) in blue.
Applsci 11 06072 g006
Figure 7. Receiver operating characteristic (ROC) curves for the aggregated data, showing the relationship between true-positive and false-positive rates estimated by the three models compared to the ground truth.
Figure 7. Receiver operating characteristic (ROC) curves for the aggregated data, showing the relationship between true-positive and false-positive rates estimated by the three models compared to the ground truth.
Applsci 11 06072 g007
Figure 8. Receiver operating characteristic (ROC) curves of the three models for the different urban areas: (a) Vaihingen, (b) Potsdam, and (c) Toronto.
Figure 8. Receiver operating characteristic (ROC) curves of the three models for the different urban areas: (a) Vaihingen, (b) Potsdam, and (c) Toronto.
Applsci 11 06072 g008
Table 1. Pixel-wise performance by area for the three models: test evaluation metrics.
Table 1. Pixel-wise performance by area for the three models: test evaluation metrics.
SensitivitySpecificityPrecisionNPVF1 ScoreMSEMCCAUC
HOG-SVMVaihingen0.7730.7670.4460.9330.5660.0440.6570.782
Potsdam0.7370.8760.7180.8860.7280.0790.6200.828
Toronto0.8870.8710.7420.9490.8080.0330.8260.914
Total0.8170.8540.6870.9230.7460.0580.7250.867
HOG-ELMVaihingen0.4150.9350.6080.8680.4930.1140.5130.710
Potsdam0.7200.8880.7350.8810.7270.0840.6200.840
Toronto0.6540.9090.7490.8630.6980.1020.6440.771
Total0.6530.9060.7310.8700.6900.0990.6260.809
FCNVaihingen0.7410.9660.8720.9230.8010.0610.8330.948
Potsdam0.8010.9480.7840.8990.7920.0600.8340.964
Toronto0.8340.9470.8770.9250.8550.0520.8210.973
Total0.7850.9560.8830.9140.8310.0630.8340.960
Table 2. Advantages and disadvantages of the three implemented methods.
Table 2. Advantages and disadvantages of the three implemented methods.
MethodAlgorithmAdvantagesDisadvantages
HOG-SVM[60]
-
Good predictive abilities
-
Training data size sensitivity
-
Manual feature engineering
-
Not suitable for urban context with sparse vegetation, small footprints, and narrow roads
HOG-ELM[70]
-
Average predictive abilities
-
Fastest computational time
-
Possible training data size sensitivity
-
Manual feature engineering
-
Not suitable for urban context with sparse vegetation, small footprints, and narrow roads
-
Not suitable for urban context with large footprints
FCN[50]
-
Best predictive abilities in all urban contexts
-
Transfer learning (training data size robustness)
-
Automatic feature extraction
-
Easy implementation
-
High transferability potential
-
Irregular footprint edges
-
Irrelevant feature susceptibility
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Notarangelo, N.M.; Mazzariello, A.; Albano, R.; Sole, A. Comparing Three Machine Learning Techniques for Building Extraction from a Digital Surface Model. Appl. Sci. 2021, 11, 6072. https://doi.org/10.3390/app11136072

AMA Style

Notarangelo NM, Mazzariello A, Albano R, Sole A. Comparing Three Machine Learning Techniques for Building Extraction from a Digital Surface Model. Applied Sciences. 2021; 11(13):6072. https://doi.org/10.3390/app11136072

Chicago/Turabian Style

Notarangelo, Nicla Maria, Arianna Mazzariello, Raffaele Albano, and Aurelia Sole. 2021. "Comparing Three Machine Learning Techniques for Building Extraction from a Digital Surface Model" Applied Sciences 11, no. 13: 6072. https://doi.org/10.3390/app11136072

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop