Next Article in Journal
Preserving Biodiversity and Ecosystem Services in West African Forest, Watersheds, and Wetlands: A Review of Incentives
Next Article in Special Issue
Extraction of Information on Trees outside Forests Based on Very High Spatial Resolution Remote Sensing Images
Previous Article in Journal
Community-Based Watershed Change: A Case Study in Eastern Congo
Previous Article in Special Issue
Beta-Diversity Modeling and Mapping with LiDAR and Multispectral Sensors in a Semi-Evergreen Tropical Forest
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Discriminating Urban Forest Types from Sentinel-2A Image Data through Linear Spectral Mixture Analysis: A Case Study of Xuzhou, East China

1
School of Environmental Science and Spatial Informatics, China University of Mining and Technology, Daxue Road 1, Xuzhou 221116, China
2
Engineering Research Center of Ministry of Education for Mine Ecological Restoration, China University of Mining and Technology, Daxue Road 1, Xuzhou 221116, China
3
Department of Geography, Earth System Science, Vrije Universiteit Brussel, Pleinlaan 2, 1050 Brussels, Belgium
4
School of Geography, Geomatics and Planning, Jiangsu Normal University, Shanghai Road 101, Xuzhou 221116, China
*
Author to whom correspondence should be addressed.
Forests 2019, 10(6), 478; https://doi.org/10.3390/f10060478
Submission received: 24 April 2019 / Revised: 24 May 2019 / Accepted: 29 May 2019 / Published: 31 May 2019

Abstract

:
Urban forests are an important component of the urban ecosystem. Urban forest types are a key piece of information required for monitoring the condition of an urban ecosystem. In this study, we propose an urban forest type discrimination method based on linear spectral mixture analysis (LSMA) and a support vector machine (SVM) in the case study of Xuzhou, east China. From 10-m Sentinel-2A imagery data, three different vegetation endmembers, namely broadleaved forest, coniferous forest, and low vegetation, and their abundances were extracted through LSMA. Using a combination of image spectra, topography, texture, and vegetation abundances, four SVM classification models were performed and compared to investigate the impact of these features on classification accuracy. With a particular interest in the role that vegetation abundances play in classification, we also compared SVM and other classifiers, i.e., random forest (RF), artificial neural network (ANN), and quick unbiased efficient statistical tree (QUEST). Results indicate that (1) the LSMA method can derive accurate vegetation abundances from Sentinel-2A image data, and the root-mean-square error (RMSE) was 0.019; (2) the classification accuracies of the four SVM models were improved after adding topographic features, textural features, and vegetation abundances one after the other; (3) the SVM produced higher classification accuracies than the other three classifiers when identical classification features were used; and (4) vegetation endmember abundances improved classification accuracy regardless of which classifier was used. It is concluded that Sentinel-2A image data has a strong capability to discriminate urban forest types in spectrally heterogeneous urban areas, and that vegetation abundances derived from LSMA can enhance such discrimination.

1. Introduction

Urban forests are important carriers of urban ecosystems [1,2], which can improve the urban microclimate, maintain the surface water–heat exchange balance [3,4], mitigate rainstorm runoff [5,6], and provide a comfortable habitat for urban residents [7]. Discriminating urban forest types has fundamental implications for planning, management, and protection of urban forests, as well as for forestry studies [8]. It also provides a basis for the estimation of above-ground biomass of urban vegetation [9,10]. Since it was first introduced by Jorgensen (1986) [11], urban forestry received increasing attention from scholars. However, the scope of urban forests was defined from a variety of research perspectives [12,13,14]. An urban forest can refer to all the trees in an urban area, including forest parks, and public and private woodlands [15], while Miller (1996) [16] and other researchers [17,18] consider urban forests as the sum of all the vegetation in the city, not only trees, but also park vegetation and private vegetation. In this study, we adopt the broader definition given by Miller—who also uses urban vegetation and urban forests interchangeably—and refer to an urban forest as a sum of trees in groups or individual trees, shrubs, and grassland within an urban area. Previous studies on urban forests focused on their release of oxygen and carbon [19], cooling and humidification effects [20], and landscape patterns [21,22]. However, discrimination of urban forest types was little studied despite being considered essential for urban forestry.
Field-based inventorying is the traditional and the most accurate method for vegetation survey and monitoring [23], but its use is restricted because it is time-consuming, expensive, and slow in updating [22]. As such, a quick and reliable approach is needed, which now can be addressed by applying remote-sensing (RS) technology. RS provides multi-source earth observation data at varying spatial resolutions from repeated visits, which allows forests to be surveyed and mapped rapidly and dynamically [24]. Based on the source of data used, an RS classification-based discrimination approach can be roughly divided into two categories [8]. One involves urban forest vegetation classification based on optical remote-sensing data, including moderate-resolution Landsat TM (Thematic Mapper)/ETM+ (Enhanced Thematic Mapper Plus) and AVIRIS (Airborne Visible/Infrared Imaging Spectrometer) imagery [25,26,27], and (very) high-resolution IKONOS, Worldview, airborne, and UAV (Unmanned Aerial Vehicle) imagery [28,29,30,31]; the other involves making use of radar data (including spaceborne and airborne radar) [32,33,34,35]. Some researchers also worked on an integrated use of these multi-source remote-sensing data for discriminating urban forest types [36]. Dalponte et al. (2012) [8] combined very-high-resolution multispectral/hyperspectral imagery and LiDAR (Light Detection and Ranging) data to classify forest vegetation in the Southern Alps, and distinguished five forest types. Liu et al. (2017) [37] identified 15 urban vegetation species types based on an aeronautical hyperspectral and airborne LiDAR point cloud. These studies used either free relatively coarse-resolution image data or purchased high-resolution image data, but none took advantage of the latest free multispectral image data with increased spatial resolution, such as Sentinel-2A.
It is noted that, because urban forests are often disturbed by human activity and, in a spectrally heterogeneous urban context, do not spread continuously like natural forests, it is challenging for traditional supervised classification methods to acquire highly separable training samples and get a satisfying classification accuracy of urban forest type discrimination [38]. Linear spectral mixture analysis (LSMA)—which is frequently used for estimating spectral endmember abundances from hyper- or multispectral images—offers an alternative to obtaining training samples for classifying urban vegetation [39,40]. In addition, machine learning algorithms such as support vector machines (SVM) and random forests (RF) can extract effective features from large feature datasets and produce higher classification accuracy than ordinary maximum-likelihood and K-means classifiers [41,42]. These promise a possible improvement in urban forest mapping and discrimination.
In order to contribute to the general objective of estimating the above-ground biomass in urban areas, this study assesses the possibility of mapping urban forest types from a single Sentinel-2A image. We, therefore, propose in this study an innovative method by combing linear spectral mixture analysis (LSMA) and an SVM machine learning algorithm in Xuzhou, east China. Specific objectives were (1) to investigate the capability of Sentinel-2A data and LSMA for extracting urban forest vegetation endmembers; (2) to find out the optimal feature combination for mapping urban forests; and (3) to identify whether vegetation abundances are similarly effective in improving classification accuracy when different machine learning classifiers are performed.

2. Study Area and Data

2.1. Study Area

Bordering the provinces of Shandong, Henan, and Anhui in a counterclockwise direction, Xuzhou (33°43′–34°58′ north (N), 116°22′–118°40′ east (E)) is located in the northwestern part of Jiangsu province with an average altitude below 400 m (Figure 1). It has a warm temperate semi-humid monsoon climate and a frost-free period of 200–220 days, with an average annual temperature of 13–16 °C and an average annual precipitation of 800–900 mm [43]. In 2017, the forest coverage area of Xuzhou was 336,300 ha—a forest coverage rate as high as 30.12%, which ranked Xuzhou first in Jiangsu. It is one of the “National Forest Cities” (awarded in 2012) and the “National Ecological Garden Cities” (awarded in 2016).
The study area is within the Sanhuan Road of Xuzhou, covering an area of approximately 108.51 km2. The northern, eastern, and southern parts of the study area are hilly lands (labeled ①–④ in Figure 1d), dominantly covered by Platycladus orientalis (L.) Franco, whereas the central part is for commercial and residential purposes. Based on our prior knowledge coupled with field observations, urban forests in the study area were concentrated and tended to be fragmented. Different vegetation types exist mostly independent of each other. Therefore, we divided the land cover of the study area into five types: broadleaved forest, coniferous forest, low vegetation (including shrubs, grasslands, and suburban farmlands), water bodies, and non-vegetation area (excluding water bodies).

2.2. Data

2.2.1. Satellite Data

Data used for mapping and discriminating urban forest types involved a single Sentinel-2A image acquired on 24 July 2017 with little cloud contamination (1.74%) and downloaded from the Copernicus Open Access Hub (https://scihub.copernicus.eu/dhus/#/home). The Sentinel-2 satellite was launched by the European Space Agency (ESA) in mid-2015 aimed at earth observation. It travels in a sun-synchronous orbit with an orbit height of 786 km and an inclination angle of 98.5°, providing image data of 290 km in width [44]. The Sentinel-2A satellite carries a multispectral instrument (MSI), providing a total of 13 bands from visible light to shortwave infrared (four bands at 10 m, six bands at 20 m, and three bands at 60 m; for more details about Sentinel-2A bands, please refer to the overview introduction of Sentienl-2 MSI images on the website of the ESA [45]).
Compared with Landsat and SPOT (Systeme Probatoire d’Observation de la Terre) data, Sentinel-2 images have more advantages in discriminating urban forest types due to their increased multispectral bands, increased spatial resolution, and shorter revisit period [46,47]. They are characterized by three unique “vegetation red-edge” bands (bands 5, 6, and 7), which are valuable for remote sensing of vegetation. Although it was widely used in the monitoring of fires [48], vegetation biophysical estimation [49,50], and surface feature extraction analysis [51], the potential of Sentinel-2 data to discriminate urban forest types remains to be fully acknowledged.
The product level of the Sentinel-2A image used in this study was Level-1C, which means that geometric correction, radiation calibration, and top of atmosphere (TOA) correction were already applied [52].
Preprocessing of Sentinel-2A Level-1C products includes atmospheric correction, resampling, and clipping. The atmospheric correction was conducted in the Sen2cor plugin, a Python-based atmospheric correction tool used in SNAP® (Sentinel Application Platform), which is an open-source application developed by ESA for processing Sentinel-1 to -3 data and is freely available from ESA’s website. Through atmospheric correction, the Level-1C data were converted into Level-2A data, such as bottom of atmosphere (BOA), aerosol optical thickness images, and water vapor images [52]. In our study, 10-m bands and 20-m bands were independently corrected before the 20-m bands were resampled to 10-m bands using the nearest neighbor method in SNAP. In total, ten bands were used, except for bands 1, 9, and 10, because they are not relevant to vegetation. Clipping (i.e., extracting the study area from the image) and other data processing (e.g., layer stacking and spectral mixture analysis) were done in ENVI® (remote sensing software by US-based Harris Geospatial Solutions. Inc., Broomfield, Colorado, CO, USA).

2.2.2. Fieldwork

In order to identify forest types on the field and collect validation data for image classification accuracy assessment, we conducted fieldwork from October to December 2017. Despite being three months later than the acquisition data of the Sentinel-2A image, this is considered acceptable for a study area where vegetation does not change much over three months.
A total of 192 sites for fieldwork were randomly selected on the corrected Sentinel-2A image (Universal Transverse Mercator Projection WGS84-50N) in ArcGIS® (geographic information system software by US-based Esri Inc., Redlands, CA, USA). Then, we localized these pre-selected sites on the field using hand-held Hi-Target® Hi-Q5 GPS devices (by China-based Hi-Target Surveying Instrument Co. Ltd, Guangzhou, China), which have a maximal horizontal accuracy of 0.5 m when connected with the continuously operating reference stations (CORS) network of Xuzhou. Due to restricted accessibility of some areas in Xuzhou (such as special education schools, military zones), the number of effective sites was 140 (35 coniferous forest sites, 73 broadleaved forest sites, and 32 low vegetation sites; see Table A1, Appendix A), down from the pre-selected 192.
On the field, we recorded tree species and other parameters, such as tree height, diameter at breast height (DBH), crown width, and vegetation coverage for our further research on urban biomass estimation, within a 10 m × 10 m rectangle centered at the site’s coordinates. The size of the rectangle allows a spatial match with a Sentinel-2A pixel.

3. Methods

A technical flowchart is provided to better illustrate the methods of this study (Figure 2). The left part shows the fieldwork, and the right part details the image processing, the features used for classification, and the classification models.

3.1. Linear Spectral Mixture Analysis

Due to rapid urban expansion and human activity, urban forests are increasingly fragmented and vegetated areas tend to be mixed pixels on satellite images. To address the mixed-pixel issue, linear spectral mixture analysis (LSMA), which treats the pixel spectrum as a linear combination of the endmember spectrums of the objects [53], can be used to extract vegetation endmembers and vegetation abundances (i.e., the proportion of vegetation to the area of a pixel), and to acquire more reliable training samples.
Three steps were required to perform LSMA on Sentinel-2A data. Firstly, a minimum noise fractionation (MNF) transformation, which is superior to principal component analysis (PCA) [54], was conducted to separate the band noise and minimize the intra-band correlations. In our study, the first six MNF components contained 80.94% of the original spectral variability and were, therefore, used for Cartesian coordinate system establishment and endmember selection in the second step. Vegetation endmembers usually lie at the vertices of the feature space constructed by combining any pair of the six MNF components. However, due to the similarity of spectral features between different forest types [55], such selection is not sufficient to get reliable vegetation endmembers. In the last step, these pre-selected endmembers were imported into ENVI’s interactive n-D Visualizer tool to generate pure pixels for spectral unmixing.
Fully constrained least squares (FCLS) is an LSMA method for endmember abundance calculation, which can simultaneously satisfy non-negativity (each abundance ranging between 0 and 1) and sum-to-unity (the sum of abundances for each pixel is 1) [53,56].
R ( λ i ) = k = 1 m f ( k i ) C ( k λ ) + ε ( λ i ) ,
k = 1 m f ( k i ) = 1   and   f ( k i ) > 0 ,
where R(λi) is the reflectance of the i pixel in the λ band, f(ki) is the proportion of the k endmember in the i pixel, C(kλ) is the reflectance of the k endmember in the λ band, m is the number of bands, and ε(λi) is the error value.
The root-mean-square error (RMSE) was used to assess the accuracy of FCLS in this study, which is given by
R M S E = 1 m i = 1 m ε ( λ i ) 2 ,
where m and ε(λi) have the same meanings as in Equations (1) and (2).
Three vegetation endmembers (broadleaved forest, coniferous forest, and low vegetation) were identified by trial and error, and their abundances were estimated through an FCLS spectral unmixing plugin of ENVI. For detailed results and analysis, see Section 4.1.

3.2. Selection of Features for Classification

For image classification, we preliminarily considered 126 features in total, ranging from spectral features and vegetation abundances to topographic features and textural features (Table 1). In addition to each spectral band, spectral features also included 30 spectral indices such as normalized difference vegetation index (NDVI) and normalized difference water index (NDWI). The features and their indicators are listed in Table 1. The abundances of the three vegetation endmembers (B1–B3 in Table 1) were coniferous forests, broadleaved forests, and low vegetation, which were obtained by the LSMA method (Section 4.1). Terrain features included a digital elevation model (DEM), and a DEM-derived slope and aspect. The textural features included eight types of textures such as mean, variance, and homogeneity, and were calculated using the gray-level co-occurrence matrix (GLCM) in ENVI. A total of 80 textural features in 10 bands (shown in Table 1, from D1–D80) were finally obtained.
However, in order to reduce data dimension and improve computational efficiency, the features should be selected prior to classification. As it can rank features in order of importance (a higher value implies a more important feature), random forest (RF) is often used for selecting essential features from a large number of features [58]. The number of decision trees in a random forest (mtree) and the number of features per node (ntry) are two key parameters in RF [59], and how they are combined impacts classification accuracy. Classification accuracy usually increases with the increase of mtree, and an optimal ntry is among p / 2 , p , and 2 p , where p is the number of features. This study set several different combinations of parameters and explored the effects of different parameter combinations on the feature importance. By constraining the range of mtree (100, 300, 500, 1000, 1500) and the range of ntry (5, 11, 22), random forest returned the importance of the 126 features for 15 different parameter combinations (Figure 3). Although the importance of each feature changed with parameter combination, their relative ranking almost always remained the same.
All the features were sorted in order of feature importance, and only those whose importance values were higher than 0.5 were selected to guarantee that each band had textural features to participate in the classification—because textural features proved important for land-cover classification in previous studies [60,61]. The chosen features used for classification are shown in Table 2. They were 54 in total, consisting of 40 spectral features, three vegetation abundance features, one terrain feature (i.e., DEM), and 10 textural features.

3.3. Support Vector Machine Classifier

The support vector machine (SVM) is a machine learning algorithm based on statistical learning theories. By constructing a classifying hyperplane, it can effectively solve the problems of limited, non-linear, and high-dimensional training samples [62]. If the samples are linearly separable, a linear discriminant function is established by constructing the classification surface to ensure the maximum distance between the samples. If the samples are linearly inseparable, the SVM projects the training samples to a high-dimensional space and finds the optimal classifying hyperplane [63].
In our study, we constructed four different SVM classification models differing in the feature used: Model 1 with spectral features (SF) (M1: SF); Model 2 with SF and digital elevation model (DEM) (M2: SF + DEM); Model 3 with SF, DEM, and textural features (TF) (M3: SF + DEM + TF); and Model 4 with SF, DEM, TF, and vegetation abundances (VA) (M4: SF + DEM + TF + VA). These four SVM classification models were tested to identify the best one for mapping and discriminating urban forest types of the study area. Classifications were implemented using an ENVI add-in known as EnMAP-box, which allows SVM and RF classifications [64]. The kernel function of the SVM classifications was Radial Basis Function (RBF), and the optimal penalty parameter (C) and the nuclear parameter (g) were determined by the grid search method [65]. Model accuracy was used to assess which parameter combination was best when constructing the SVM model with the training samples.

4. Results and Discussion

4.1. LSMA Result and Analysis

After MNF and endmember selection, three vegetation endmembers were identified, including broadleaved forest, coniferous forest, and low vegetation (Figure 4). Despite typical spectral signature vegetation with peaks and troughs located at quite similar wavelengths, they showed contrasting reflectance values in the same spectral range (e.g., 800–900 nm). In the visible part of the electromagnetic spectrum, low vegetation had higher reflectance than the other two types. This is because broadleaved forests can effectively use the red light and blue–violet light more efficiently than coniferous forests and low vegetation in photosynthesis [66]. However, highest reflectance was observed for low vegetation, and it was lowest for coniferous forest in the NIR–SWIR (near infrared-shortwave infrared) region, especially in the “vegetation red-edge” band (700–800 nm). This is likely to be explained by the complex canopy structures of low vegetation and broadleaved forests. Light cannot transmit them easily, resulting in increased reflections on the canopy surface. As for coniferous forests, their needle leaves are more prone to transmission and, thus, lower reflectance.
Through the FCLS-based LSMA, three vegetation abundance maps were produced (Figure 5). All vegetation abundance values ranged from 0 to 1—a brighter pixel had a higher vegetation abundance value and vice versa. The RMSE of the FCLS was 0.019, indicating that our LSMA result is reliable [67].
As explained in Section 3.1, an abundance value refers to the ratio of the area of an urban forest type to the total area of a given pixel. Through the LSMA, the abundance of each forest type in every pixel can be derived. Given the spatial resolution of the image used (10 m × 10 m, i.e., the ground area of a pixel was 100 m2), the areas of different urban forest types in the study area were straightforwardly calculated (Table 3 and Figure 5). Results show that coniferous forests covered a maximal area of 15.28 km2 (accounting for 14.09% of the study area) and were mostly distributed on the hills of Yunlong, Zhushan, Zifang, and Jiuli (Figure 1d). They were dominated by Platycladus orientalis, mostly grown from local forestation projects between the 1950s and 1960s. Low vegetation (13.96 km2) was primarily distributed in parks and idle construction land, while broadleaved forests (4.21 km2) were mostly found in parks, rural settlements, and less hilly areas.

4.2. Interpretation of Feature Importance

In this study, to determine a proper number of features used for urban forest discrimination, we first built a dataset of 126 candidate features and selected only 54 from them using the random forest. They covered a variety of spectral features (e.g., vegetation index and soil index), topographic features, and vegetation component abundances (Table 2). Among the features selected, spectral features like N_NIR, VRE2, and VRE3 bands (see their definitions in the note below Table 1) were highly ranked in the parameter combinations, suggesting their important roles in discrimination. This is not unexpected as these bands serve the vegetation monitoring purpose of the Sentinel-2A sensor, one of its major applications [44]. As reflectance at these bands is related to vegetation cellular structure [57] and varies with vegetation type, it is useful to use these bands to discriminate urban forest types [68]. We discuss textural features in Section 4.3.

4.3. SVM Classification Results and Accuracy Assessment

By means of the grid search method, the optimal values of parameters C and g for the four classification models were determined as (125, 0.04), (25, 0.2), (125, 0.4), and (625, 0.04), with model accuracies all being over 97%. Then, we performed the classifications and produced four different land-cover maps (Figure 6). Accuracy assessment based on the validation data acquired from our fieldwork showed that the highest accuracy and Kappa coefficient were achieved by M4 (89.86% and 0.83) and the lowest by M1 (86.96% and 0.79) (Table 4). Overall classification accuracy was improved by 1.45% when adding DEM to M1, and was further improved when textural features and vegetation abundances were added one by one. This suggests that classification accuracy tends to increase with the number of input features, which agrees with the study of Raczko and Zagajewski (2017) [69]. Reasons for the improvements vary. In the case of topography, it helps to improve classification accuracy because the impact of topography on vegetation growth is considered [70,71,72]. Textural features often prove useful in vegetation classification because vegetation texture varies with age, species, and many other factors [60,61]. Although only a few textural features were relatively highly ranked in terms feature importance, they might contribute to the improvement in classification accuracy. Vegetation abundances were shown to have a positive effect on classification, which is in agreement with the finding of Adams (1995) [39].
As M4 produced the best classification result, we here present its confusion matrix for a detailed analysis (Table 5). Both the highest user accuracy and producer accuracy were observed for the coniferous forest type. This is likely attributed to the fact that coniferous forests in the study area consisted mostly of Platycladus orientalis (L.) Franco, and these trees tend to grow in large numbers.
The classification model (M4) resulting in the best accuracy (overall accuracy 89.86% and Kappa 0.83) was not employed in previous studies but it is still interesting to compare it with other urban vegetation classification studies. For example, De Colstoun et al. (2003) [73] mapped vegetation types in the Pennsylvania national forest park using a decision tree (C5.0) and multi-temporal Landsat data with an overall accuracy of 82.05% and a Kappa coefficient of 0.80. Liu and Yang (2013) [74] tested the multiple endmember spectral mixture analysis (MESMA) technique for urban vegetation classification with a maximal classification accuracy of 80.55%. Based on the SVM classifier, Poursanidis et al. (2015) [75] extracted urban land cover in the Greek city of Rafina by combining textural and spectral information, resulting in a highest classification accuracy of 89.23%. Compared with these studies, our method is more capable of discriminating different urban forest types.

4.4. Comparison of Different Classifier Results

In order to evaluate the differences between SVM and other machine learning classifiers for urban vegetation information extraction, and whether vegetation abundances can enhance urban vegetation discrimination similarly, three machine learning algorithms, i.e., RF (random forest), ANN (artificial neural network), and QUEST (quick unbiased efficient statistical tree), were used to discriminate urban forest types before and after adding vegetation abundances. The classification based on the SVM classification without vegetation abundance (i.e., M3, with 51 features) was labeled as SVM, and the SVM classification with vegetation abundances (i.e., M4, with 54 features) was labeled as SVM + VA. Similarly, we also named the RF-, ANN-, and QUEST-based classifications with and without vegetation abundances (Table 6).
As such, there were eight different classification models. These classifications were performed using the same training samples and assessed using validation data as the four SVM classifications in Table 4. For both RF classifications, their optimal mtry and ntry values were 1000 and 22 based on the result of out-of-bag error (OOB) test (see its definition in Li et al. (2017) [76]). The QUEST is a type of decision tree classifier and has a faster calculation and higher accuracy than other types [77]. Classification maps are presented in Figure 7 and the number of features, overall accuracy, and kappa coefficient for each classification model are shown in Table 6.
Among the eight classification models, the SVM + VA model obtained the highest accuracy (89.86%) and Kappa coefficient (0.83). In terms of classifier, SVM produced the best classification results, which agrees well with previous studies [8,36]. It has two evident advantages. Firstly, it can find an optimal hyperplane with the highest classification boundary in the n-dimensional feature space. This prevents the classifier from falling into local minima [63], which is the case for ANN. Secondly, SVM can minimize unseen errors in training samples [78] and, thus, a higher classification accuracy [79].
For each classifier, adding vegetation abundances in classification resulted in increased accuracy. This is particularly remarkable for the RF, ANN, and QUEST classifiers. Their classification accuracies were approximately 82% and the Kappa coefficients were less than 0.72, which rose to above 85% and 0.75, respectively. Classification accuracy of the SVM was also improved by including vegetation abundances, although such improvement was not that prominent. Additionally, our results suggest that ANN and RF could achieve similar classification accuracies, which was also confirmed by previous studies [80]. In our case, classification accuracies of ANN and RF were 82.14% and 81.29% before vegetation abundances were added, and increased by 2.86% and 2.92% for ANN and RF, respectively, after vegetation abundances were included.
In addition, if we examine the SVM- and RF-based classification maps closely (Figure 8), we can find that adding vegetation abundances resulted in more homogeneous classification maps. This is because it could effectively reduce the salt-and-pepper effect that usually occurs in pixel-based classification.

5. Conclusions

This study aimed at mapping and discriminating urban forest using single Sentinel-2A imagery and machine learning algorithms, such as SVM, ANN, RF, and QUEST. Input features were selected based on the feature importance of RF and included vegetation abundances obtained from linear spectral mixture analysis. From the results, we conclude the following:
  • Three urban forest endmembers can be successfully identified from Sentinel-2A image data, and the LSMA method allows accurate mapping of their abundances with a low mean RMSE of 0.019.
  • Classification accuracy of SVM classification tends to increase when spectral, topographic, and textural features and vegetation abundances are added one by one.
  • The SVM classifier outperforms the other three machine learning algorithms based on the same classification samples and field-based validation data.
  • Vegetation abundances help improve classification accuracy regardless of classifier.
Our study provides a basis for urban biomass estimation and has practical implications for forest management. It also demonstrates the capability of 10-m Sentinel-2A image data to discriminate vegetation types in a complex urban context. However, an avenue for future research would be to use different sources of remote-sensing data, such as Sentinel-1 SAR (Synthetic Aperture Radar) imagery. This might enable full use of the textural features of vegetation surfaces.

Author Contributions

X.Z. and L.L. conducted and designed the research plan and discussed it with L.C. who supervised and finalized the study. X.Z., Y.L., and Y.C. contributed to the design and conduction of fieldwork. Y.Z. and T.Z. were responsible for remote-sensing data collection and preprocessing. X.Z., L.L., and L.C. completed data analysis and the interpretation of results. X.Z. and L.L. wrote the manuscript, and all authors contributed to the revision and editing of the manuscript.

Funding

This research was supported by the Fundamental Research Funds for the Central Universities (Grant No.: 2018ZDPY07).

Acknowledgments

We would like to acknowledge the European Space Agency (ESA) for freely providing Sentinel-2A data required for the research, as well as the reviewers’ constructive comments and suggestions for the improvement of our study.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Fieldwork sites for acquiring validation data for classification accuracy assessment.
Table A1. Fieldwork sites for acquiring validation data for classification accuracy assessment.
NumberX-coordinateY-coordinateUF TypesNumberX-coordinateY-coordinateUF Types
15144053787635BF25149453787685BF
35159753787725CF45168753787885CF
55150353787905CF65177253787995BF
75172053788075CF85170153788245CF
95154053788525CF105176353788785CF
115166453788865BF125170053789155BF
135187053789215LW145181153789545LW
155160153790555BF165199853793155BF
175124253793635LW185122053794045CF
195129653794215CF205138653794485CF
215137953794735CF225142053795255BF
235160953796225CF245169953796385BF
255138753797275LW265175553797965LW
275200753786135LW285176353786305CF
295203053786325BF305190353786365LW
315201653786525BF325178053786555CF
335182053786605BF345135853786715BF
355183353786835CF365188653786925BF
375140753787025LW385176153787125BF
395138653787145BF405133453787245LW
415154053787255BF425160453787335CF
435140653787385BF445149553787415BF
455155653787415CF465136553787435BF
475176253787435BF485195453787475CF
495187853787665BF505151253787725CF
515138853787875BF525158453787955CF
535120153787995BF545124053788085CF
555174853788175BF565151253788215BF
575130853788225BF585156853788235CF
595116253788245BF605129653788445BF
615113453788455LW625172253788465CF
635178553788515BF645174753788785LW
655114553788855BF665160753788855BF
675156353788905BF685168853789325BF
695117953789355LW705159953789425BF
715202853789435BF725199053789465BF
735151053789505LW745201353789585BF
755130453789645CF765124653789665BF
775165453789855BF785197553789925BF
795195653790205BF805162053790305BF
815124253791195BF825141553791505BF
835196253792175CF845188853792185CF
855204353792825BF865127153792845BF
875196953793135BF885137853793485LW
895122153793785LW905124853793895BF
915127853793895CF925137553794005BF
935119453794035BF945172753794065BF
955128453794365LW965157353794465BF
975131653794515LW985123053794715BF
995140853794855CF1005122853794885LW
1015121453795265BF1025130053795325BF
1035152553795355BF1045113553795485BF
1055156653795515BF1065157953795765CF
1075149053795785BF1085128053795905BF
1095132453795935BF1105144253795995LW
1115155553795995CF1125165953796145CF
1135112153796195BF1145153953796245LW
1155148253796345BF1165165753796575BF
1175114853796605BF1185170353796625BF
1195123753796685CF1205152953796725LW
1215163553796725LW1225175453796755LW
1235178053796845BF1245185353796855LW
1255172253796985LW1265187053797015CF
1275120253797065BF1285149453797165LW
1295131953797255LW1305131653797415BF
1315177653797495LW1325185053797515LW
1335157553797585CF1345151653797825LW
1355168453797865BF1365187953797975LW
1375191753798005CF1385156553798105LW
1395163853798135BF1405189653798255LW
Note: UF refers to urban forest, BF refers to broadleaved forest, CF refers to coniferous forest, and LW refers to low vegetation.
Table A2. Formulas used for calculating spectral indices [55].
Table A2. Formulas used for calculating spectral indices [55].
Spectral IndicesFormula
Green index (GI) G I = G r e e n / R e d
Green normalized different vegetation index (gNDVI) g N D V I = ( N _ N I R G r e e n ) / ( N _ N I R + G r e e n )
Normalized difference vegetation index (NDVI) N D V I = ( N I R R e d ) / ( N I R + R e d )
Ratio vegetation index (RVI) R V I = N I R / R e d
Difference vegetation index (DVI) D V I = N I R R e d
Enhanced vegetation index 2 (EVI2) E V I 2 = ( N I R R e d ) / ( 1 + N I R + 2.4 × R e d )
Chlorophyll green index (Chlogreen) C h l o g r e e n = N _ N I R / ( G r e e n + V E R 1 )
Normalized difference vegetation index (NDVIre1) N D V I r e 1 = ( N I R V E R 1 ) / ( N I R + V E R 1 )
Normalized difference vegetation index (NDVIre1n) N D V I r e 1 n = ( N _ N I R V E R 1 ) / ( N _ N I R + V E R 1 )
Simple ratio 1 (SR1) S R 1 = N I R / V E R 1
Simple ratio 2 (SR2) S R 2 = N _ N I R / V E R 1
Simple ratio 3 (SR3) S R 3 = N _ N I R / R e d
Simple ratio 4 (SR4) S R 4 = N _ N I R / G r e e n
Simple ratio 5 (SR5) S R 5 = N _ N I R / B l u e
Simple ratio 6 (SR6) S R 6 = B l u e / V E R 1
Simple ratio 7 (SR7) S R 7 = N I R / R e d
Normalized difference water index (NDWI) N D W I = ( G r e e n N I R ) / ( G r e e n + N I R )
Normalized difference water index 1 (NDWI1) N D W I 1 = ( N _ N I R S W I R 1 ) / ( N _ N I R + S W I R 1 )
Normalized difference water index 2 (NDWI2) N D W I 2 = ( G r e e n N _ N I R ) / ( G r e e n + N _ N I R )
Normalized humidity index (NHI) N H I = ( S W I R 1 G r e e n ) / ( S W I R 1 + G r e e n )
Normalized difference infrared index (NDII) N D I I = ( N I R S W I R 1 ) / ( N I R + S W I R 1 )
Modified normalized difference water index (MNDWI) M N D W I = ( G r e e n S W I R 1 ) / ( G r e e n + S W I R 1 )
Normalized difference build-up index (NDBI) N D B I = ( S W I R 1 N I R ) / ( S W I R 1 + N I R )
Build-up area index (BAI) B A I = ( B l u e N _ N I R ) / ( B l u e + N _ N I R )
Enhanced index-based built-up index (EIBI) E I B I = N D B B I ( 4 × E B S I + S A V I + M N D W I ) / 6 N D B B I + ( 4 × E I B I + S A V I + M N D W I ) / 6
Soil-adjusted vegetation index (SAVI) S A V I = N _ N I R R e d N _ N I R + R e d + L × 0.5
Modified soil-adjusted vegetation index 2 (MSAVI2) M S A V I 2 = 0.5 × [ ( 2 × N I R + 1 ) ( 2 × N I R + 1 ) 2 8 × ( N I R R e d ) ]
Optimized soil-adjusted vegetation index (OSAVI) O S A V I = ( N I R R e d ) / ( N I R + R e d + 0.16 )
Bare soil index (BSI) B S I = [ ( S W I R 1 + R e d ) ( N I R + B l u e ) ] / [ ( S W I R 1 + R e d ) + ( N I R + B l u e ) ]
Normalized difference bareness and built-up index (NDBBI) N D B B I = 1.5 × S W I R 2 ( N I R + G r e e n ) / 2 1.5 × S W I R 2 + ( N I R + G r e e n ) / 2
Note: VRE1–VRE3 represent the three vegetation red-edge bands; N_NIR represents the narrow near-infrared bands.

References

  1. Dwivedi, P.; Rathore, C.S.; Dubey, Y. Ecological benefits of urban forestry: The case of Kerwa Forest Area (KFA), Bhopal, India. Appl. Geogr. 2009, 29, 194–200. [Google Scholar] [CrossRef]
  2. Young, R.F. Managing municipal green space for ecosystem services. Urban For. Urban Green. 2010, 9, 313–321. [Google Scholar] [CrossRef]
  3. Oliveira, S.; Andrade, H.; Vaz, T. The cooling effect of green spaces as a contribution to the mitigation of urban heat: A case study in Lisbon. Build. Environ. 2011, 46, 2186–2194. [Google Scholar] [CrossRef]
  4. Park, J.; Kim, J.-H.; Lee, D.K.; Park, C.Y.; Jeong, S.G. The influence of small green space type and structure at the street level on urban heat island mitigation. Urban For. Urban Green. 2017, 21, 203–212. [Google Scholar] [CrossRef]
  5. Inkiläinen, E.N.M.; McHale, M.R.; Blank, G.B.; James, A.L.; Nikinmaa, E. The role of the residential urban forest in regulating throughfall: A case study in Raleigh, North Carolina, USA. Landsc. Urban Plan. 2013, 119, 91–103. [Google Scholar] [CrossRef]
  6. Kirnbauer, M.C.; Baetz, B.W.; Kenney, W.A. Estimating the stormwater attenuation benefits derived from planting four monoculture species of deciduous trees on vacant and underutilized urban land parcels. Urban For. Urban Green. 2013, 12, 401–407. [Google Scholar] [CrossRef]
  7. Jensen, R.R.; Hardin, P.J.; Hardin, A.J. Classification of urban tree species using hyperspectral imagery. Geocarto Int. 2012, 27, 443–458. [Google Scholar] [CrossRef]
  8. Dalponte, M.; Bruzzone, L.; Gianelle, D. Tree species classification in the Southern Alps based on the fusion of very high geometrical resolution multispectral/hyperspectral images and LiDAR data. Remote. Sens. Environ. 2012, 123, 258–270. [Google Scholar] [CrossRef]
  9. He, C.; Convertino, M.; Feng, Z.; Zhang, S. Using LiDAR data to measure the 3D green biomass of Beijing urban forest in China. PLoS ONE 2013, 8, e75920. [Google Scholar] [CrossRef]
  10. Singh, K.K.; Chen, G.; McCarter, J.B.; Meentemeyer, R.K. Effects of LiDAR point density and landscape context on estimates of urban forest biomass. ISPRS J. Photogramm. Remote Sens. 2015, 101, 310–322. [Google Scholar] [CrossRef] [Green Version]
  11. Jorgensen, E. Urban forestry in the rearview mirror. Arboric. J. 1986, 10, 177–190. [Google Scholar] [CrossRef]
  12. McBride, J.; Jacobs, D. Urban forest development: A case study, Menlo Park, California. Urban Ecol. 1976, 2, 1–14. [Google Scholar] [CrossRef]
  13. Konijnendijk, C.C.; Ricard, R.M.; Kenney, A.; Randrup, T.B. Defining urban forestry—A comparative perspective of North America and Europe. Urban For. Urban Green. 2006, 4, 93–103. [Google Scholar] [CrossRef]
  14. Rowntree, R.A. Ecology of the urban forest—Introduction to part I. Urban Ecol. 1984, 8, 1–11. [Google Scholar] [CrossRef]
  15. Steenberg, J.W.N.; Duinker, P.N.; Charles, J.D. The neighbourhood approach to urban forest management: The case of Halifax, Canada. Landsc. Urban Plan. 2013, 117, 135–144. [Google Scholar] [CrossRef]
  16. Miller, R.W.; Hauer, R.J.; Werner, L.P. Urban Forestry: Planning and Managing Urban Greenspaces, 3rd ed.; Waveland Press Inc.: Long Grove, IL, USA, 2015; ISBN 9781478606376. [Google Scholar]
  17. Konijnendijk, C.C. A decade of urban forestry in Europe. For. Policy Econ. 2003, 5, 173–186. [Google Scholar] [CrossRef]
  18. Davies, H.J.; Doick, K.J.; Hudson, M.D.; Schreckenberg, K. Challenges for tree officers to enhance the provision of regulating ecosystem services from urban forests. Environ. Res. 2017, 156, 97–107. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Myeong, S.; Nowak, D.J.; Duggin, M.J. A temporal analysis of urban forest carbon storage using remote sensing. Remote. Sens. Environ. 2006, 101, 277–282. [Google Scholar] [CrossRef]
  20. Gunawardena, K.R.; Wells, M.J.; Kershaw, T. Utilising green and bluespace to mitigate urban heat island intensity. Sci. Total. Environ. 2017, 584–585, 1040–1055. [Google Scholar] [CrossRef]
  21. Nowak, D. Understanding the structure of urban forests. J. For. 1994, 92, 42–46. [Google Scholar]
  22. Ren, Z.; Zheng, H.; He, X.; Zhang, D.; Yu, X.; Shen, G. Spatial estimation of urban forest structures with Landsat TM data and field measurements. Urban For. Urban Green. 2015, 14, 336–344. [Google Scholar] [CrossRef]
  23. Trammell, T.L.E.; Carreiro, M.M. Vegetation composition and structure of woody plant communities along urban interstate corridors in Louisville, KY, U.S.A. Urban Ecosyst. 2011, 14, 501–524. [Google Scholar] [CrossRef]
  24. Fassnacht, F.E.; Latifi, H.; Stereńczak, K.; Modzelewska, A.; Lefsky, M.; Waser, L.T.; Straub, C.; Ghosh, A. Review of studies on tree species classification from remotely sensed data. Remote Sens. Environ. 2016, 186, 64–87. [Google Scholar] [CrossRef]
  25. Haapanen, R.; Ek, A.R.; Bauer, M.E.; Finley, A.O. Delineation of forest/nonforest land use classes using nearest neighbor methods. Remote Sens. Environ. 2004, 89, 265–271. [Google Scholar] [CrossRef]
  26. Martin, M.E.; Newman, S.D.; Aber, J.D.; Congalton, R.G. Determining forest species composition using high spectral resolution remote sensing data. Remote Sens. Environ. 1998, 65, 249–254. [Google Scholar] [CrossRef]
  27. Alonzo, M.; Roth, K.; Roberts, D. Identifying Santa Barbara’s urban tree species from AVIRIS imagery using canonical discriminant analysis. Remote Sens. Lett. 2013, 4, 513–521. [Google Scholar] [CrossRef]
  28. Pu, R.; Landry, S. A comparative analysis of high spatial resolution IKONOS and WorldView-2 imagery for mapping urban tree species. Remote Sens. Environ. 2012, 124, 516–533. [Google Scholar] [CrossRef]
  29. Pu, R.; Landry, S.; Zhang, J. Evaluation of atmospheric correction methods in identifying urban tree species with WorldView-2 imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 1886–1897. [Google Scholar] [CrossRef]
  30. Richter, R.; Reu, B.; Wirth, C.; Doktor, D.; Vohland, M. The use of airborne hyperspectral data for tree species classification in a species-rich Central European forest area. Int. J. Appl. Earth Obs. Geoinf. 2016, 52, 464–474. [Google Scholar] [CrossRef]
  31. Brovkina, O.; Cienciala, E.; Surový, P.; Janata, P. Unmanned aerial vehicles (UAV) for assessment of qualitative classification of Norway spruce in temperate forest stands. Geo-spatial Inf. Sci. 2018, 21, 12–20. [Google Scholar] [CrossRef] [Green Version]
  32. Reitberger, J.; Krzystek, P.; Stilla, U. Analysis of full waveform LiDAR data for the classification of deciduous and coniferous trees. Int. J. Remote Sens. 2008, 29, 1407–1431. [Google Scholar] [CrossRef]
  33. Linders, J. Comparison of three different methods to select feature for discriminating forest cover types using SAR imagery. Int. J. Remote Sens. 2000, 21, 2089–2099. [Google Scholar] [CrossRef]
  34. Alonzo, M.; Bookhagen, B.; McFadden, J.P.; Sun, A.; Roberts, D.A. Mapping urban forest leaf area index with airborne lidar using penetration metrics and allometry. Remote Sens. Environ. 2015, 162, 141–153. [Google Scholar] [CrossRef]
  35. Wang, K.; Wang, T.; Liu, X. A review: individual tree species classification using integrated airborne LiDAR and optical imagery with a focus on the urban environment. Forests 2018, 10, 1. [Google Scholar] [CrossRef]
  36. Ghosh, A.; Fassnacht, F.E.; Joshi, P.K.; Koch, B. A framework for mapping tree species combining hyperspectral and LiDAR data: Role of selected classifiers and sensor across three spatial scales. Int. J. Appl. Earth Obs. Geoinf. 2014, 26, 49–63. [Google Scholar] [CrossRef]
  37. Liu, L.; Coops, N.C.; Aven, N.W.; Pang, Y. Mapping urban tree species using integrated airborne hyperspectral and LiDAR remote sensing data. Remote Sens. Environ. 2017, 200, 170–182. [Google Scholar] [CrossRef]
  38. Tigges, J.; Lakes, T.; Hostert, P. Urban vegetation classification: Benefits of multitemporal Rapid Eye satellite data. Remote Sens. Environ. 2013, 136, 66–75. [Google Scholar] [CrossRef]
  39. Adams, J.B.; Sabol, D.E.; Kapos, V.; Almeida Filho, R.; Roberts, D.A.; Smith, M.O.; Gillespie, A.R. Classification of multispectral images based on fractions of endmembers: Application to land-cover change in the Brazilian Amazon. Remote. Sens. Environ. 1995, 52, 137–154. [Google Scholar] [CrossRef]
  40. Lu, D.; Weng, Q. Spectral mixture analysis of the urban landscape in Indianapolis with Landsat ETM+ imagery. Photogramm. Eng. Remote Sens. 2004, 70, 1053–1062. [Google Scholar] [CrossRef]
  41. Waske, B.; Benediktsson, J.A.; Árnason, K.; Sveinsson, J.R. Mapping of hyperspectral AVIRIS data using machine-learning algorithms. Can. J. Remote Sens. 2009, 35, S106–S116. [Google Scholar] [CrossRef]
  42. Dalponte, M.; Orka, H.O.; Gobakken, T.; Gianelle, D.; Naesset, E. Tree species classification in boreal forests with hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2632–2645. [Google Scholar] [CrossRef]
  43. Zhang, Y.; Chen, L.; Wang, Y.; Chen, L.; Yao, F.; Wu, P.; Wang, B.; Li, Y.; Zhou, T.; Zhang, T. Research on the contribution of urban land surface moisture to the alleviation effect of urban land surface heat based on Landsat 8 data. Remote Sens. 2015, 7, 10737–10762. [Google Scholar] [CrossRef]
  44. Drusch, M.; Del Bello, U.; Carlier, S.; Colin, O.; Fernandez, V.; Gascon, F.; Hoersch, B.; Isola, C.; Laberinti, P.; Martimort, P.; et al. Sentinel-2: ESA’s optical high-resolution mission for GMES operational services. Remote Sens. Environ. 2012, 120, 25–36. [Google Scholar] [CrossRef]
  45. MSI Instrument—Sentinel-2 MSI Technical Guide—Sentinel Online. Available online: https://sentinel.esa.int/web/sentinel/technical-guides/sentinel-2-msi/msi-instrument (accessed on 20 May 2019).
  46. Munyati, C. The potential for integrating Sentinel 2 MSI with SPOT 5 HRG and Landsat 8 OLI imagery for monitoring semi-arid savannah woody cover. Int. J. Remote Sens. 2017, 38, 4888–4913. [Google Scholar] [CrossRef]
  47. Forkuor, G.; Dimobe, K.; Serme, I.; Tondoh, J.E. Landsat-8 vs. Sentinel-2: examining the added value of sentinel-2’s red-edge bands to land-use and land-cover mapping in Burkina Faso. GIScience Remote Sens. 2018, 55, 331–354. [Google Scholar] [CrossRef]
  48. Fernández-Manso, A.; Fernández-Manso, O.; Quintano, C. SENTINEL-2A red-edge spectral indices suitability for discriminating burn severity. Int. J. Appl. Earth Obs. Geoinf. 2016, 50, 170–175. [Google Scholar] [CrossRef]
  49. Mura, M.; Bottalico, F.; Giannetti, F.; Bertani, R.; Giannini, R.; Mancini, M.; Orlandini, S.; Travaglini, D.; Chirici, G. Exploiting the capabilities of the Sentinel-2 multi spectral instrument for predicting growing stock volume in forest ecosystems. Int. J. Appl. Earth Obs. Geoinf. 2018, 66, 126–134. [Google Scholar] [CrossRef]
  50. Puliti, S.; Saarela, S.; Gobakken, T.; Ståhl, G.; Næsset, E. Combining UAV and Sentinel-2 auxiliary data for forest growing stock volume estimation through hierarchical model-based inference. Remote Sens. Environ. 2018, 204, 485–497. [Google Scholar] [CrossRef]
  51. Pesaresi, M.; Corbane, C.; Julea, A.; Florczyk, A.J.; Syrris, V.; Soille, P. Assessment of the added-value of Sentinel-2 for detecting built-up areas. Remote. Sens. 2016, 8, 299. [Google Scholar] [CrossRef]
  52. Main-Knorn, M.; Pflug, B.; Debaecker, V.; Louis, J. Calibration and validation plan for the L2a processor and products of the Sentinel-2 mission. ISPRS - Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2015, XL-7/W3, 1249–1255. [Google Scholar] [CrossRef]
  53. Li, L.; Canters, F.; Solana, C.; Ma, W.; Chen, L.; Kervyn, M. Discriminating lava flows of different age within Nyamuragira’s volcanic field using spectral mixture analysis. Int. J. Appl. Earth Obs. Geoinf. 2015, 40, 1–10. [Google Scholar] [CrossRef]
  54. Zhang, C.; Xie, Z. Combining object-based texture measures with a neural network for vegetation mapping in the Everglades from hyperspectral imagery. Remote Sens. Environ. 2012, 124, 310–320. [Google Scholar] [CrossRef]
  55. Schmidt, K.S.; Skidmore, A.K. Spectral discrimination of vegetation types in a coastal wetland. Remote. Sens. Environ. 2003, 85, 92–108. [Google Scholar] [CrossRef]
  56. Zhang, Y.; Li, L.; Chen, L.; Liao, Z.; Wang, Y.; Wang, B.; Yang, X. A modified multi-source parallel model for estimating urban surface evapotranspiration based on ASTER thermal infrared data. Remote Sens. 2017, 9, 1029. [Google Scholar] [CrossRef]
  57. Njoku, E.G. Encyclopedia of Remote Sensing; Springer Science: New York, NY, USA, 2014; ISBN 9780387366982. [Google Scholar]
  58. Ma, L.; Fan, S. CURE-SMOTE algorithm and hybrid algorithm for feature selection and parameter optimization based on random forests. BMC Bioinform. 2017, 18, 169. [Google Scholar] [CrossRef]
  59. Cui, Y.; Li, L.; Chen, L.; Zhang, Y.; Cheng, L.; Zhou, X.; Yang, X. Land-use carbon emissions estimation for the Yangtze River Delta urban agglomeration using 1994–2016 landsat image data. Remote Sens. 2018, 10, 1334. [Google Scholar] [CrossRef]
  60. Immitzer, M.; Atzberger, C.; Koukal, T. Tree species classification with random forest using very high spatial resolution 8-band WorldView-2 satellite data. Remote. Sens. 2012, 4, 2661–2693. [Google Scholar] [CrossRef]
  61. Dian, Y.; Li, Z.; Pang, Y. Spectral and texture features combined for forest tree species classification with airborne hyperspectral imagery. J. Indian Soc. Remote Sens. 2015, 43, 101–107. [Google Scholar] [CrossRef]
  62. Nello, C.; John, S. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
  63. Cervantes, J.; Garcia-Lamont, F.; Rodriguez, L.; López, A.; Castilla, J.R.; Trueba, A. PSO-based method for SVM classification on skewed data sets. Neurocomputing 2017, 228, 187–197. [Google Scholar] [CrossRef] [Green Version]
  64. Van Der Linden, S.; Rabe, A.; Held, M.; Jakimow, B.; Leitão, P.J.; Okujeni, A.; Schwieder, M.; Suess, S.; Hostert, P. The EnMAP-Box—A toolbox and application programming interface for EnMAP data processing. Remote. Sens. 2015, 7, 11249–11266. [Google Scholar] [CrossRef]
  65. Bao, Y.; Hu, Z.; Xiong, T. A PSO and pattern search based memetic algorithm for SVMs parameters optimization. Neurocomputing 2013, 117, 98–106. [Google Scholar] [CrossRef] [Green Version]
  66. Yu, Q.; Wang, S.; Huang, K.; Zhou, L.; Chen, D. An anlysis of the spectrums between different canopy structures based on Hypersion hyperspectral data in a temperate forest of North China. Spectrosc. Spectr. Anal. 2015, 35, 1980–1985. [Google Scholar]
  67. Wu, C.; Murray, A.T. Estimating impervious surface distribution by spectral mixture analysis. Remote Sens. Environ. 2003, 84, 493–505. [Google Scholar] [CrossRef]
  68. Canetti, A.; Garrastazu, M.C.; De Mattos, P.P.; Braz, E.M.; Netto, S.P. Understanding multi-temporal urban forest cover using high resolution images. Urban For. Urban Green. 2018, 29, 106–112. [Google Scholar] [CrossRef]
  69. Raczko, E.; Zagajewski, B. Comparison of support vector machine, random forest and neural network classifiers for tree species classification on airborne hyperspectral APEX images. Eur. J. Remote Sens. 2017, 50, 144–154. [Google Scholar] [CrossRef] [Green Version]
  70. Fahsi, A.; Tsegaye, T.; Tadesse, W.; Coleman, T. Incorporation of digital elevation models with Landsat-TM data to improve land cover classification accuracy. For. Ecol. Manag. 2000, 128, 57–64. [Google Scholar] [CrossRef]
  71. Gislason, P.O.; Benediktsson, J.A.; Sveinsson, J.R. Random Forests for land cover classification. Pattern Recognit. Lett. 2006, 27, 294–300. [Google Scholar] [CrossRef]
  72. Gartzia, M.; Alados, C.L.; Pérez-Cabello, F.; Bueno, C.G. Improving the accuracy of vegetation classifications in mountainous areas. Mt. Res. Dev. 2013, 33, 63–74. [Google Scholar] [CrossRef]
  73. De Colstoun, E.C.B.; Story, M.H.; Thompson, C.; Commisso, K.; Smith, T.G.; Irons, J.R. National Park vegetation mapping using multitemporal Landsat 7 data and a decision tree classifier. Remote. Sens. Environ. 2003, 85, 316–327. [Google Scholar] [CrossRef]
  74. Liu, T.; Yang, X. Mapping vegetation in an urban area with stratified classification and multiple endmember spectral mixture analysis. Remote Sens. Environ. 2013, 133, 251–264. [Google Scholar] [CrossRef]
  75. Poursanidis, D.; Chrysoulakis, N.; Mitraka, Z. Landsat 8 vs. Landsat 5: A comparison based on urban and peri-urban land cover mapping. Int. J. Appl. Earth Obs. Geoinf. 2015, 35, 259–269. [Google Scholar] [CrossRef]
  76. Li, L.; Solana, C.; Canters, F.; Kervyn, M. Testing random forest classification for identifying lava flows and mapping age groups on a single Landsat 8 image. J. Volcanol. Geotherm. Res. 2017, 345, 109–124. [Google Scholar] [CrossRef] [Green Version]
  77. Hasan, R.C.; Ierodiaconou, D.; Monk, J. Evaluation of four supervised learning methods for benthic habitat mapping using backscatter from multi-beam sonar. Remote. Sens. 2012, 4, 3427–3443. [Google Scholar] [CrossRef]
  78. Mountrakis, G.; Im, J.; Ogole, C. Support vector machines in remote sensing: A review. ISPRS J. Photogramm. Remote Sens. 2011, 66, 247–259. [Google Scholar] [CrossRef]
  79. Petropoulos, G.P.; Kontoes, C.C.; Keramitsoglou, I. Land cover mapping with emphasis to burnt area delineation using co-orbital ALI and Landsat TM imagery. Int. J. Appl. Earth Obs. Geoinf. 2012, 18, 344–355. [Google Scholar] [CrossRef]
  80. Li, W.; Fu, H.; Yu, L.; Gong, P.; Feng, D.; Li, C.; Clinton, N. Stacked Autoencoder-based deep learning for remote-sensing image classification: a case study of African land-cover mapping. Int. J. Remote Sens. 2016, 37, 5632–5646. [Google Scholar] [CrossRef]
Figure 1. The location of the study area: (ac) field photos illustrating three different urban forest types; (d) sites for field investigation (yellow for low vegetation, green for broadleaved forest, and purple for coniferous forest), the border of the study area, and a true-color composition of the Sentinel-2A image used for classification. Labels ①–④ refer to the hills of Zhushan, Yunlong, Zifang, and Jiuli, respectively; (e) study area in Xuzhou.
Figure 1. The location of the study area: (ac) field photos illustrating three different urban forest types; (d) sites for field investigation (yellow for low vegetation, green for broadleaved forest, and purple for coniferous forest), the border of the study area, and a true-color composition of the Sentinel-2A image used for classification. Labels ①–④ refer to the hills of Zhushan, Yunlong, Zifang, and Jiuli, respectively; (e) study area in Xuzhou.
Forests 10 00478 g001
Figure 2. Technical flowchart of discriminating urban forest types in our study. The step in the dashed-line rectangle represents the four support vector machine (SVM) classification models constructed with four different sets of features. SF refers to the spectral features; DEM refers to the digital elevation model, selected from terrain features; TF refers to the textural features; and VA refers to the vegetation abundances.
Figure 2. Technical flowchart of discriminating urban forest types in our study. The step in the dashed-line rectangle represents the four support vector machine (SVM) classification models constructed with four different sets of features. SF refers to the spectral features; DEM refers to the digital elevation model, selected from terrain features; TF refers to the textural features; and VA refers to the vegetation abundances.
Forests 10 00478 g002
Figure 3. Feature importance. A total of 15 parameter combinations were tested and only five (ntry = 11) are shown in the graph for illustration.
Figure 3. Feature importance. A total of 15 parameter combinations were tested and only five (ntry = 11) are shown in the graph for illustration.
Forests 10 00478 g003
Figure 4. Spectra of three vegetation endmembers.
Figure 4. Spectra of three vegetation endmembers.
Forests 10 00478 g004
Figure 5. Vegetation abundance maps: (a) coniferous forest; (b) broadleaved forest; (c) low vegetation.
Figure 5. Vegetation abundance maps: (a) coniferous forest; (b) broadleaved forest; (c) low vegetation.
Forests 10 00478 g005
Figure 6. Land-cover maps produced from the four SVM classification models: (a) M1 (SF); (b) M2 (SF + DEM); (c) M3 (SF + DEM + TF); (d) M4 (SF + DEM + TF + VA).
Figure 6. Land-cover maps produced from the four SVM classification models: (a) M1 (SF); (b) M2 (SF + DEM); (c) M3 (SF + DEM + TF); (d) M4 (SF + DEM + TF + VA).
Forests 10 00478 g006
Figure 7. Land-cover maps produced by eight different classification models: (a) SVM; (b) RF; (c) ANN; (d) QUEST; (e) SVM + VA; (f) RF + VA; (g) ANN + VA; and (h) QUEST + VA.
Figure 7. Land-cover maps produced by eight different classification models: (a) SVM; (b) RF; (c) ANN; (d) QUEST; (e) SVM + VA; (f) RF + VA; (g) ANN + VA; and (h) QUEST + VA.
Forests 10 00478 g007
Figure 8. Close examination of the SVM- and RF-based classification maps.
Figure 8. Close examination of the SVM- and RF-based classification maps.
Forests 10 00478 g008
Table 1. Potential features for classification and their indicators.
Table 1. Potential features for classification and their indicators.
FeaturesIndicatorsFeaturesIndicatorsFeaturesIndicators
A1BlueA19SR3A37NDII
A2GreenA20SR4A38EIBI
A3RedA21SR5A39NDBI
A4VRE1A22SR6A40BAI
A5VRE2A23SR7B1Low vegetation abundance
A6VRE3A24RVIB2Broadleaved forest abundance
A7NIRA25NDVIre1nB3Coniferous forest abundance
A8N_NIRA26NDVIreC1DEM
A9SWIR1A27NDVIC2Slope
A10SWIR2A28gNDVIC3Aspect
A11SAVIA29GID1–D10Mean (A1–A10)
A12BSIA30ChlogreenD11–D20Variance (A1–A10)
A13MSAVI2A31EVI2D20–D30Homogeneity (A1–A10)
A14NDBBIA32NDWID30–D40Contrast (A1–A10)
A15OSAVIA33NDWI1D40–D50Dissimilarity (A1–A10)
A16DVIA34NDWI2D50–D60Entropy (A1–A10)
A17SR1A35NHID60–D70Second Moment (A1–A10)
A18SR2A36MNDWID70–D80Correlation (A1–A10)
Note: A1–A10 are the 10 remote-sensing bands (excluding bands 1, 9 and 10) obtained after atmosphere correction; VRE1–VRE3 represent the three vegetation red-edge bands; N_NIR represents the narrow near-infrared bands; SR1 represents the ratio of near-infrared and vegetation red-edge 1 and the meanings of features A18–A23 are similar to SR1; NDVIre and NDVIre1n represent two different vegetation indices. The reader is referred to Njoku (2014) [57] for a detailed description of the indices listed in Table 1 and to Table A2 (Appendix A) for the corresponding formulas. The DEM used is at the 12.5-m resolution and freely available from the Alaska Satellite Facility (https://vertex.daac.asf.alaska.edu/).
Table 2. Features selected for image classification.
Table 2. Features selected for image classification.
CategoryFeatureNumber of Features
Spectral featuresA1–A4040
Vegetation abundancesB1–B33
Terrain featuresC11
Textural featuresD1–D1010
Total 54
Table 3. Areas of urban forest types in the study area.
Table 3. Areas of urban forest types in the study area.
Urban Forest TypeArea (km2)Percentage of the Study Area (%)
Coniferous forest15.2814.09
Broadleaved forest4.213.88
Low vegetation13.9612.87
Total33.4530.84
Table 4. Accuracy assessment of the four support vector machine (SVM) classification models using the validation data acquired from fieldwork.
Table 4. Accuracy assessment of the four support vector machine (SVM) classification models using the validation data acquired from fieldwork.
SVM Classification ModelsM1:
SF
M2:
SF + DEM
M3:
SF + DEM + TF
M4:
SF + DEM + TF + VA
Number of features40415154
Optimal parameters (C)12525125625
Optimal kernel function (g)0.040.20.040.04
Model accuracy (%)97.4097.6698.3898.35
Overall accuracy (%)86.9688.4189.1389.86
Kappa0.790.810.820.83
Table 5. Confusion matrix of M4 (SF + DEM + TF + VA). Each row designates the classification result, and each column designates the field-based validation data.
Table 5. Confusion matrix of M4 (SF + DEM + TF + VA). Each row designates the classification result, and each column designates the field-based validation data.
Types of ForestsLow VegetationBroadleaved ForestsConiferous ForestsUser Accuracy (%)
Low vegetation265083.87
Broadleaved forest466390.41
Coniferous forest023294.12
Producer accuracy (%)86.6790.4191.43
Overall accuracy (%)89.86
Kappa0.83
Table 6. Accuracy assessment of classifications comparing different machine learning classifiers.
Table 6. Accuracy assessment of classifications comparing different machine learning classifiers.
Classification ModelNumber of FeaturesOverall AccuracyKappaNote
QUEST5179.71%0.66
RF5181.29%0.70
ANN5182.14%0.72
SVM5189.13%0.82M3 in Table 4
QUEST + VA5483.45%0.73
RF + VA5484.21%0.75
ANN + VA5485.00%0.76
SVM + VA5489.86%0.83M4 in Table 4

Share and Cite

MDPI and ACS Style

Zhou, X.; Li, L.; Chen, L.; Liu, Y.; Cui, Y.; Zhang, Y.; Zhang, T. Discriminating Urban Forest Types from Sentinel-2A Image Data through Linear Spectral Mixture Analysis: A Case Study of Xuzhou, East China. Forests 2019, 10, 478. https://doi.org/10.3390/f10060478

AMA Style

Zhou X, Li L, Chen L, Liu Y, Cui Y, Zhang Y, Zhang T. Discriminating Urban Forest Types from Sentinel-2A Image Data through Linear Spectral Mixture Analysis: A Case Study of Xuzhou, East China. Forests. 2019; 10(6):478. https://doi.org/10.3390/f10060478

Chicago/Turabian Style

Zhou, Xisheng, Long Li, Longqian Chen, Yunqiang Liu, Yifan Cui, Yu Zhang, and Ting Zhang. 2019. "Discriminating Urban Forest Types from Sentinel-2A Image Data through Linear Spectral Mixture Analysis: A Case Study of Xuzhou, East China" Forests 10, no. 6: 478. https://doi.org/10.3390/f10060478

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop