Next Article in Journal
Local Freeman Decomposition for Robust Imaging and Identification of Subsurface Anomalies Using Misaligned Full-Polarimetric GPR Data
Previous Article in Journal
Retrieving Freeze/Thaw Cycles Using Sentinel-1 Data in Eastern Nunavik (Québec, Canada)
Previous Article in Special Issue
Crossing the Great Divide: Bridging the Researcher–Practitioner Gap to Maximize the Utility of Remote Sensing for Invasive Species Monitoring and Management
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forest Disturbance Detection with Seasonal and Trend Model Components and Machine Learning Algorithms

1
Posgrado en Geografía, Centro de Investigaciones en Geografía Ambiental, Universidad Nacional Autónoma de México, Morelia 58190, Mexico
2
Centro de Investigaciones en Geografía Ambiental, Universidad Nacional Autónoma de México, Morelia 58190, Mexico
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(3), 803; https://doi.org/10.3390/rs14030803
Submission received: 11 December 2021 / Revised: 3 February 2022 / Accepted: 5 February 2022 / Published: 8 February 2022
(This article belongs to the Special Issue Advances in the Remote Sensing of Forest Cover Change)

Abstract

:
Forest disturbances reduce the extent of natural habitats, biodiversity, and carbon sequestered in forests. With the implementation of the international framework Reduce Emissions from Deforestation and forest Degradation (REDD+), it is important to improve the accuracy in the estimation of the extent of forest disturbances. Time series analyses, such as Breaks for Additive Season and Trend (BFAST), have been frequently used to map tropical forest disturbances with promising results. Previous studies suggest that in addition to magnitude of change, disturbance accuracy could be enhanced by using other components of BFAST that describe additional aspects of the model, such as its goodness-of-fit, NDVI seasonal variation, temporal trend, historical length of observations and data quality, as well as by using separate thresholds for distinct forest types. The objective of this study is to determine if the BFAST algorithm can benefit from using these model components in a supervised scheme to improve the accuracy to detect forest disturbance. A random forests and support vector machines algorithms were trained and verified using 238 points in three different datasets: all-forest, tropical dry forest, and temperate forest. The results show that the highest accuracy was achieved by the support vector machines algorithm using the all-forest dataset. Although the increase in accuracy of the latter model vs. a magnitude threshold model is small, i.e., 0.14% for sample-based accuracy and 0.71% for area-weighted accuracy, the standard error of the estimated total disturbed forest area was 4352.59 ha smaller, while the annual disturbance rate was also smaller by 1262.2 ha year−1. The implemented approach can be useful to obtain more precise estimates in forest disturbance, as well as its associated carbon emissions.

1. Introduction

Forest disturbance has been studied extensively especially after the recognition of its important role as a source of greenhouse gas (GHG) emissions and climate change since approximately 12–20% of the global GHG anthropogenic emissions have been attributed to this process [1,2]. Forest disturbances modify forest canopy structure and biomass content and can result from deforestation and forest degradation. Deforestation usually consists of the complete removal of forest cover at a big scale, while forest degradation involves subtle changes in forest structure and canopy cover [3]. With the future projections of climate change, improving accuracy and reducing uncertainty in mapping forest disturbances is critical, since it assures accurate estimations of forest carbon and climate change modeling [4]. In this study, forest disturbance was defined as “a relatively discrete event causing a change in the physical structure of the vegetation and surface soil” [5]. We focus on identifying disturbances that include both natural events, such as fires and anthropogenic activities—such as slash and burn agriculture and logging—that cause changes in forest structure that are detectable by remote sensing.
Remote sensing has been recognized as an important method to study forest disturbances, mapping their occurrence and quantifying their extent and severity [6,7]. Forest disturbances can be detected using different methods, each one having advantages or disadvantages depending on the dominating cause of forest disturbance. For example, disturbances caused by forest clearing for plantations (i.e., deforestation) can be estimated with relatively simple techniques such as comparing land cover maps of two different dates [8,9]. In turn, for disturbances that cause subtler changes in forests and lead to forest degradation, such as those from shifting cultivation and logging, the accurate estimation usually needs more sophisticated time series analyses and modeling such as LandTrendr or Breaks for Additive Season and Trend (BFAST) [7,10,11,12,13,14]. Regardless of the method used to detect forest disturbances, the final step usually consists of establishing a threshold to classify the disturbed and undisturbed areas [13]. There are generally two approaches to determine this threshold: expert knowledge or data-driven approach [15,16]. The first one usually implies a translation of an expert’s knowledge into a numeric threshold, while in the data-driven approach, an artificial intelligence algorithm determines the decision threshold based on a training dataset. Thus, the latter is commonly referred to as a supervised approach.
With the rapid development of artificial intelligence, the application of machine learning in remote sensing studies and particularly in change detection has become popular [17,18]. The algorithms—such as decision trees, support vector machines, artificial neural networks, random forests, among others—have been used to map land cover [19,20], detect and analyze land cover change patterns [21], and predict forest biomass [22]. For instance, Grinand et al. [20] used a random forests algorithm with a Landsat time series (2000, 2005, and 2010) to classify both land cover and deforestation in a tropical forest and obtained higher accuracy in stable land cover (84.7%) than deforestation (60.7%). Dlamini [21] predicted deforestation patterns and drivers using Bayesian classifiers and variables derived from expert knowledge including fuelwood consumption, population density, and land tenure, among others. For biomass retrieval, in situ measurements are typically associated with the corresponding spectral data from remote sensors using algorithms such as random forests, artificial neural networks, and support vector machines, among others [22].
The first step to implement a machine learning (ML) approach on time series data, usually consists of extracting different metrics that summarize important patterns in the time series, which the ML algorithms subsequently use as predictors for detecting vegetation changes or perform land-use/land-cover classifications [23,24,25]. Nowadays, several disturbance detection algorithms have included data-driven approaches in their workflows, such as Continuous Change Detection and Classification (CCDC) or Satellite Image Time Series Analysis for Earth Observation Data Cubes (SITS) with encouraging results [26,27]. However, other algorithms that have proved useful for disturbance detection, such as BFAST, have seldom been used in a data-driven approach, although the few examples that exist suggest that it could help increase its accuracy [7,28].
BFAST is a time series algorithm that fits a harmonic model to the observed data and then projects this model into the monitoring period [11,29]. Afterward, disturbances are identified as the observations that diverged from the expected model using a moving sum approach [29]. Most studies that have used BFAST for disturbance detection have relied on the use of magnitude (i.e., the difference between the observed and modeled value) to detect disturbance. However, previous studies have shown that other metrics, which we refer to as components, can be extracted as descriptors of the fitted model and when used with magnitude could enable a more accurate detection (e.g., [7,28]). Additionally, fitting separate models for different types of forest have proved useful to increase the accuracy of disturbance detection [25,30,31,32].
In Mexico, the tropical dry forest has been increasingly used for agriculture (mainly shifting cultivation), cattle raising activities and establishing human settlements, which has put pressure on local vegetation, causing not only deforestation but also degradation of the remaining forests [33,34]. As a consequence, the extent of natural habitats, biodiversity, and carbon sequestered in forests have been greatly reduced. With the implementation of the international framework Reduce Emissions from Deforestation and forest Degradation (REDD+), it is important to reduce the uncertainties in the estimation of the carbon emissions from forest disturbances [35]. Since BFAST has been shown effective in the estimation of disturbances by shifting cultivation [7,14,36], we believe the adoption of BFAST with ML algorithms can improve the quantification of forest disturbance [30].
Given the previous context, we wanted to assess the possible beneficial effect of using two different machine learning approaches on several BFAST components to identify disturbed forest areas. We focus on the following three main research questions: (1) if disturbance detection can be enhanced by using other BFAST model components, besides magnitude, under a ML approach or (2) by training separate models for each forest type (i.e., all-forest, only tropical dry forest, or only temperate forest), and (3) how random forests and support vector machines differ in their capabilities for disturbance detection.

2. Materials and Methods

2.1. Study Site

The study area is in the Ayuquila river basin, western Mexico, with elevations ranging from approximately 250 m to 2500 m above mean sea level [37] (Figure 1). The elevation variability in the study area translates into a range of climatic conditions that include tropical semi-dry and sub-humid climates in the lower elevations, and template sub-humid and humid conditions in the higher altitudes. Annual precipitation is found in the 800–1200 mm interval and the average annual temperature is between 18 and 22 °C. The study area shows a clear rainfall seasonal pattern, where most of the precipitation falls from June to October [37].
There are two types of natural forest: temperate forest (TF) and tropical dry forest (TDF): TF covers about 12% of the watershed and is found mostly in higher elevations, while TDF occupies around 24% of the basin and is distributed usually in the lower areas. The main composition of TF includes pines (Pinus spp.), firs (Abies spp.), and oaks (Quercus spp.) and it is exploited mainly for timber, although recently, avocado (Persea americana) plantations have been established in areas previously occupied by TF. The area covered by TDF is mainly used for shifting cultivation, although residents also use TDF as natural resources for fuelwood extraction, cattle grazing, and pole extraction for constructing fences [38].

2.2. General Workflow

In a previous study, forest disturbances were identified by applying a BFAST model to an NDVI time series (1994–2018) derived from Landsat 5, 7, and 8 [30]. The time series were divided into a historical period (1994–2015) and a change monitoring period (2016–2018). We applied a threshold value of | 0.2 | for the magnitude of change, which was derived from field verification, to optimize the detection of disturbances. Finally, we used 624 stratified random points to evaluate the detected disturbances and obtained 96.52% overall accuracy with an area-weighted error matrix [30]. These results represent the baseline model to which the ML models will be compared.
The general workflow of the methodology is presented in Figure 2. In the present study, to balance the number of observations of true and false disturbances in the training and validation datasets, we sampled 238 points based on a gamma distribution of magnitude values centered at | 0.2 | . We manually labeled those points as disturbed and non-disturbed forests using monthly median reflectance composites of Sentinel-2 images (2016–2018) as reference. Then, we trained two machine learning algorithms, random forests (RF) and support vector machines (SVM), to classify forest disturbances using the following components derived from BFAST: magnitude of change, trend, amplitude, goodness-of-fit, model fitting period, and data quality. In the following sections, we provide detailed descriptions of the methods.

2.2.1. Training and Validation Datasets

Although all the breakpoints detected by the BFAST spatial model with negative magnitude values are potential forest disturbances, those with magnitude values close to zero most probably correspond to false disturbances caused by noise or interannual variations, while those with magnitude values farther from zero will tend to correspond to true disturbances [29,39]. In order to balance the number of observations of true and false disturbances to train and validate the models, the raw magnitude output of the BFAST was used to establish a stratified random sampling design. In this procedure, we used a gamma distribution centered at magnitude | 0.2 | to determine the number of observations for the complete range of magnitude values (0– | 0.5 | ) in 0.05-value steps. A total of 238 points were selected following this design and distributed proportionally by the extent of the two types of forest. Thus, most of the visually interpreted points were close to a magnitude = | 0.2 | , while the more distant magnitude classes had fewer points (0– | 0.05 | : 10 pts, | 0.45 | | 0.5 | : 6 pts; Table A1).
These points were labeled manually as disturbed and non-disturbed. For this, we consulted Sentinel-2 images from 2016–2018 in GEE at level 1C, which were orthorectified and radiometrically calibrated. Sentinel-2 images provided by the European Space Agency (ESA) have spatial resolutions ranging from 10 m to 60 m and multi-spectral bands covering the visible, red edge, and infrared spectral range and a temporal resolution of 2–5 days [40]. We first applied a cloud mask using the data quality layer ‘QA60’ and then we constructed monthly time series images using the median value. For each image, we used blue, green, and red channels with a spatial resolution of 10 m to construct natural color composites. Using the monthly natural color composites, we visually interpreted the sample points and labeled them as either disturbed or non-disturbed forest. When the interpretation was not clear with the Sentinel-2 images, we consulted high spatial resolution images in Google Earth. Some examples of Sentinel-2 images for sample points interpretation as disturbance and non-disturbance in TDF and TF are in Figure 1B.
After visually interpreting the 238 points, 143 points corresponded to non-disturbed forest (TF: 53 and TDF: 48), while 95 points, to disturbed forest (TF: 90, TDF: 47). This distribution confirmed that the followed stratified random sample helped in obtaining a more balanced dataset in terms of the number of observations of the disturbance/non-disturbance classes. Afterward, the verified points were split at random into training and validation datasets with a proportion of 0.7/0.3, respectively. It has been reported that forest type is an important factor in classifying forest disturbance using BFAST [30,31,32], therefore, three models were constructed with different forest stratum: all-forest, only TDF, and only TF.

2.2.2. Variables: BFAST Components and Time Series Landsat Data Quality

BFAST has been successfully applied in characterizing spatial–temporal vegetation dynamics and detecting changes in a time series [7,11,36]. BFAST model focuses on near real-time change detection, by which data is divided into a historical and monitoring period. The stable trend, modeled in the historical period, is projected into the monitoring period and a breakpoint of change is detected when the difference between the projected and observed values is larger than a threshold [29]. Usually, magnitude of change is the variable that is adjusted to optimize the detection [29,39,41]. However, a range of other variables can be obtained from the BFAST model such as trend, amplitude of the seasonal model, goodness-of-fit, model fitting period, as well as data quality in the fitted model. These variables have been tested relevant for vegetation change detection. For example, Grogan et al. [31] reported that the difference in slope of the model is important in the disturbance detection of different forest types. De Vries et al. [39] found an enhanced disturbance detection when using the breakpoints along with magnitude, while Dutrieux et al. [7] and Schultz et al. [42] reported data availability as an important factor in disturbance detection.
In this work, the following BFAST model components were extracted: magnitude, trend, amplitude, goodness-of-fit, model fitting period, and data quality (Table 1). The abbreviations in Table 1 are the following: i stands for each observation in a time series, y i and y ^ i for the observed and predicted NDVI value of i, y ¯ stands for the mean NDVI, n for the number of observations in the time series, x for the independent variable (i.e., days), y stands for the dependent variable (NDVI), y b r e a k and y ^ b r e a k for the observed and predicted value in the detected breakpoint, respectively, y ^ m a x and y ^ m i n for the maximum and the minimum predicted value of the detrended model, respectively, n , represents valid observations and N represents the total number of observations in the stable historical period, respectively. We use data quality as a measure of the percentage of valid observations over the total observations, which are 365 per year. The valid observations correspond to cloudless observations, while the total observations include valid observations, no data, cloud pixels, and missing data.

2.2.3. Baseline Model

The baseline model for forest disturbance detection uses only the threshold ( | 0.2 | ) of magnitude of change as a predictive variable, which was determined by field verification. With this model, an area with an absolute magnitude value higher than this threshold is classified as a disturbance. This model is applied in all three datasets with different forest stratum. The purpose of this baseline model is to have a point of comparison with the ML models using other variables of BFAST components.

2.2.4. Machine Learning Algorithm

Two ML algorithms, RF and SVM, were applied in a supervised classification scheme using variables of BFAST components and time series data quality to test if they can improve upon the baseline model. These two algorithms use non-parametric models that have been applied previously for land cover classification and change detection [18].
RF is a classification and regression algorithm based on an ensemble of multiple decision trees, trained using random samples of the data and predictor variables [43]. Predictions are designated by aggregating the estimates made by each tree using a majority vote. RF is usually trained with a proportion of two-thirds of the training data and validated by the remaining one-third, also referred to as out-of-bag dataset [43]. Previous studies have reported that using a high number of decision trees makes the generalization error converge and reduces the overfitting of the model [19,20,43]. Therefore, we use 500 random trees to perform the classification, while the number of predictive variables used in each split of each tree was the square root of the number of predictive variables. Finally, the out-of-bag ratio was one-third.
SVM is also a classification and regression algorithm that tries to find a hyperplane that minimizes the error in the predictions [44]. The hyperplane for the classification is determined by the subset of the data that fall in the decision boundary (these points are referred to as support vectors). SVM assumes that the data are linearly separable in a multidimensional space; but, since this is often not the case, kernel functions are used to transform the feature space to facilitate the classification [17]. SVM can use different types of kernel functions; however, in this study, we applied the radial basis function, which has been frequently used in remote sensing applications [45,46]. On the other hand, a key concept in SVM is the margin, which refers to the distance between the decision hyperplane and the closest observation. Usually, a large margin allows clear discrimination by the hyperplane. Thus, the cost parameter consists of a positive value indicating the penalization for predicting a sample within or on the wrong side of the margin, while the sigma is the precision parameter. In this case, we use cost = 1 and sigma was calculated automatically from the data in a heuristic procedure [47].

2.3. Variable Importance

The importance value of each predictive variable was calculated for both ML algorithms using models that included all the predictive variables. Based on the importance values, multiple models were trained sequentially, including the most important variables first. This sequential model training was applied for the three datasets: all-forest, TDF and TF. The all-forest dataset included forest-type as an additional predictor, besides the BFAST components; while the forest-type specific datasets included only the BFAST components as predictive variables. In total, seven models were trained for the all-forest dataset and six models for each of the forest-type specific datasets.
For RF models, the variable importance was calculated as a linear transformation of the Mean Decrease Gini into a 0–100 range, which measures the total decrease in node impurities resulting from splitting the data according to each predictive variable and then averaging this decrease over all the constructed trees [43]. For SVM models, the variable importance was calculated by permuting the independent variables and fitting a single-predictor model on the complete data [48]. Afterward, the predictor variable that obtained the highest value of the area under the curve (AUC) of the receiver operating characteristic (ROC) curve was ranked as 100 and the variable with the lowest AUC value as zero, while all the other importance values were scaled to this range.

2.4. Best Model Selection

To find the best model configuration in terms of the predictive variables, different models were trained using only the training dataset (n = 168), while leaving the validation data untouched (n = 70). The model training was based on a 3-fold cross-validation with 40 repeats. In this way, each model ran 120 iterations from which the average accuracy and standard error were calculated. Afterward, the best configuration was selected based on a balance between the overall accuracy and the number of variables included, such that models with fewer predictive variables and a non-significant difference with the highest achieved accuracy were prioritized (i.e., within the range of mean accuracy ± 1.96 standard error).
The identical procedure was followed in the three datasets: all-forest, TDF and TF. Although the split ratio between the training and validation datasets was the same for all the three datasets, for TDF and TF datasets, the total number of observations for each split was smaller in comparison with the all-forest dataset. In the case of TDF, training and validation datasets had 97 and 41 observations, respectively; while for TF, the same datasets had 72 and 29 observations, respectively.
The overall accuracy is defined as a metric that summarizes the number of correct predictions over the total predictions [49]. It is calculated as the ratio of the sum of the true positive and true negative cases over the total predictions as summarized in an error matrix (Table 2, Equation (1)).
Overall accuracy = (TP + TN)/(TP + TN + FP + FN)

2.5. Model Validation

The trained model was tested using the validation data, which was kept untouched. Several metrics were calculated including overall accuracy, F1-score, and ROC AUC. Overall accuracy evaluates the correct detection of both disturbance and non-disturbance samples, while F1-score and ROC AUC focus on the class of interest (i.e., disturbance). Therefore, each metric evaluates a different aspect of the results.
F1-score corresponds to the harmonic mean of precision and recall, calculated as Equation (2), where p stands for precision (or user’s accuracy) and r for recall (or producer’s accuracy) [50,51]. Precision is calculated as the ratio of true positives over the total of positive predictions and recall is calculated as the ratio of the true positives over the sum of true positives and false negatives [51].
F1-score = 2 × p × r/(p + r)
The ROC curve is constructed by plotting the false positive rate against the true positive rate of a classification model, using different probability thresholds. True positive rate is a synonym for recall (see Equation (2)), while false positive is calculated as 1 − TN/(TN + FN).
ROC curves can be used to interpret the skill of a model to correctly classify a binary outcome (true or false). In our case, it shows the ability of the model to classify the disturbance class. In turn, the AUC of a ROC plot is calculated as the area under the ROC curve. For a random guess, the AUC is 0.5, while a perfect model will have an AUC of 1. Models with AUC between 0.5 and 1 represent a better prediction than a random guess, while higher AUC values are considered better models.

2.6. Post Classification Processing and Accuracy Assessment

The best model (SVM all-forest) was used to predict forest disturbance in the complete study area. Although the validation dataset was used to verify the models using an independent set of observations, this dataset was created following a gamma distribution of magnitude. Thus, it was not considered as a representative sample of the entire study area. To assess the accuracy of this classification using an independent set of validation data, a stratified random sampling was implemented based on the area occupied by each class of interest in the classification map, following [52]. The number of samples was calculated using Equation (3)
n = (   W i S i ) 2 [ S ( Ô ) ] 2 + ( 1 / N )   W i S i 2 (   W i S i S ( Ô ) ) 2 ,
where S ( Ô ) is the standard error of the estimated overall accuracy. As suggested by Olofsson et al. [52] a value of 0.01 was used. W i is the mapped proportion of the area of class i and S i is the standard deviation of class i , and S i = U i ( 1 U i ) . This formula estimated a sample size of 624 observations. For the rare classes—i.e., disturbances—50 random points were assigned for each forest type. The rest of the validation points were defined according to the predicted proportional area of the remaining strata (i.e., non-disturbed forest in both TF and TDF). Afterward, these points were verified by visual interpretation in Google Earth Engine (GEE) using Sentinel-2 images.

2.7. R Packages

The complete method was carried out in an R 4.0.5 environment. Variable importance was calculated using the caret package [48], the training and validation of the ML algorithms were performed using the tidymodels [53] and yardstick packages [54], the spatial information operations were done using the sf [55] and raster packages [56], while for data wrangling and plots the tidyverse package [57] was used.

3. Results

3.1. Variable Importance

The variable importance values show that the essential variables to correctly identify disturbances vary with the algorithm and dataset used (all-forest, TDF, or TF). Nevertheless, regardless of the dataset and ML algorithm, magnitude was always ranked as the most important predictor, while other important variables included trend, data quality, model fitting period, and amplitude. On the contrary, forest-type and R2 were ranked as the least important (Figure 3).

3.2. Best Model Selection

The exploration of the models with a different number of predictive variables indicated that the accuracy ranged from 61.48% to 87.89% (see Figure 3 and Table A2). From these results, the models that were not significantly different from the highest achieved accuracy and with less predictive variables were selected and verified on the validation set (Table 3). For the all-forest model, both SVM and RF favored three identical predictive variables: magnitude, trend, and model fitting period, in the same order of importance; however, SVM obtained higher overall accuracy than RF on the validation dataset. In turn, in the RF model of TDF, the best model included magnitude, trend, and amplitude; while using SVM, it included magnitude, trend, model fitting period, and data quality. In the TF model, RF and SVM used fewer predictive variables, either magnitude and data quality for RF or magnitude and model fitting period for SVM (Table 3).

3.3. Model Validation

Regardless of the ML algorithm used, the validation of these models showed the same accuracy or an increase in accuracy of up to 21.95% compared to the baseline models (Δ Best vs. Baseline, Table 4). The largest difference was observed in the TDF dataset, where the baseline model achieved an accuracy of 58.54% on the validation dataset, while the RF model had an accuracy of 80.49%. On the contrary, the smallest difference was observed in the TF dataset, where the baseline model and SVM model obtained the same accuracy of 79.31%. When comparing the two ML algorithms, SVM showed a higher accuracy on the all-forest dataset, while RF achieved higher accuracies on the TDF and TF datasets (Table 4). Finally, the model that enabled the highest accuracy was SVM with the all-forest dataset.

3.4. ROC Curve and AUC

The ROC curves of the models with the highest accuracy indicate that in all cases the models showed a higher AUC than a random guess and less than the perfect model. Additionally, both ML algorithms have a higher AUC than the baseline model. The plot shows the possible combinations of true positive rate and false positive rate given the data. For example, when the true positive rate and the false positive rate are both at one (1, 1), all the validation data are classified as disturbed forest; while the opposite combination of (0, 0) happens when none of the validation data is classified as disturbed forest. All the points between these two extremes show a combination of different true and false positive rates in the detection (Figure 4).

3.5. Forest Disturbance Prediction in the Study Area

The all-forest with the SVM model was used to predict forest disturbance for the complete study area (Figure 5). Evaluated by 624 random points, the prediction obtained an overall accuracy of 94.87%, while the user’s and producer’s accuracy for the disturbance and non-disturbance were between 82.69% and 97.31%. When weighted by the area proportion of each class, the overall accuracy increased to 97.23%, while the producer’s accuracy for forest disturbance was lowered from 86% to 13.77% (Table 5). Overall accuracy can be biased by the imbalance in the number of observations between the classes. Therefore, high overall accuracy values can disguise the low classification capabilities for rare classes. Other methods can give further insights into the classification capabilities for both common and rare classes, such as the F1-score. Although the F1-score can give a better idea of the classification capabilities for all the classes, the overall accuracy reports the precision of the entire map. Thus, the two metrics can be used to describe different aspects of the classification performance.
When the accuracy assessment was weighted by area proportion, due to the extremely large area occupied by the non-disturbance class, a small commission error can have a detrimental effect on the area estimation of the disturbance class [48]. In our case, the commission error weighted by the area of the non-disturbance class (99.48%) reduces the producer’s accuracy of the disturbance class from 86% to 13.77%, when weighted by area proportion.

4. Discussion

4.1. Improvement in Disturbance Detection Using Machine Learning Algorithms

This study demonstrates that using ML algorithms and BFAST model components (ML+BFAST) enabled higher accuracy in forest disturbance detection in comparison with the baseline model, which relies only on a magnitude threshold. In all datasets (i.e., all-forest, TDF, TF), the ML algorithms (i.e., SVM or RF) improved the accuracy achieved in the validation dataset, except for SVM with the TF dataset, which obtained the same accuracy as the baseline model (79.31%). Other evaluations reached similar conclusions when using additional BFAST components or ML algorithms to improve the disturbance detection [28,31].
In the all-forest dataset, SVM over-performed RF with a small difference (4.29%). In turn, in both TDF and TF models, RF obtained 4.88–6.90% higher accuracy than SVM. The similar performance of SVM and RF in remote sensing-based classifications has been commonly reported [18,45]. However, it is unclear if there is a particular advantage of one algorithm over the other, as both algorithms can deal with heterogeneous data, data of high dimensionality, and a limited number of observations [18,45].
Compared with the baseline model, ML approaches did not show an advantage when using magnitude as the single predictive variable (Table A2). This suggests that when magnitude is the only variable used, expert knowledge enables a higher accuracy than a ML approach; however, when the number of variables increases, ML algorithms achieved a higher accuracy (Table A2). A prior report on the use of ML and BFAST has shown the advantages of ML over a simple threshold procedure to identify disturbances [28]. In our case, we used a small training sample and therefore, the advantage of expert knowledge vs. ML is still yet to be tested with a larger dataset.
Interestingly, the most accurate model was the one with the all-forest data instead of the forest-type specific models. This seems to contradict the results from previous findings [28,30,31]. However, a smaller sample size in the forest-type specific models may have caused the lower observed accuracies, since the reduction of the size of the training dataset can have repercussions over the accuracy, particularly with an already small dataset [58,59]. Further research is needed to separate the confounding effects of forest type, sample size and data heterogeneity on the accuracy of the tested models. On the other hand, the importance of forest-type as a predictive variable is less than magnitude, trend, and model fitting period; nonetheless, the importance of this variable is relative to the number of predictive variables included in the models. For example, if fewer predictive variables were used, the importance of forest-type would be enhanced. Additionally, because in our study we only included two types of forest over a relatively small area, it is unclear if forest-type might indeed act as an important variable when including a larger variety of forests [31] or with other forest types [60].
In comparison with similar studies, our model had slightly lower accuracy [28,31], which might be related to the data and the data quality. Experiments with better results have used datasets with higher observation density (e.g., MODIS), a harmonized collection from different sensors or data with improved quality through pre-processing with filtering [31,61,62]. Better results have also been reported when using different indices (e.g., NDMI, NDFI, or multiple spectral indices; [7,28,39,42]), or by including SAR together with optical images [60,63].
Reducing the dimensionality of the feature space before training the ML algorithms has been suggested as a necessary preprocess to improve model performance [64]. However, our study used a limited number of predictors, so we used the importance value analysis instead to determine the predictors included in the simplest models, as they offered more critical information for the classification. Additionally, to avoid overfitting, the best model was selected as the one with fewer predictive variables and not significantly different from the model with the highest accuracy.

4.2. Variable Selection

Our results show that magnitude and trend were ranked most important in all the models. Similar results have also been reported [30,31]. Since magnitude represents the difference between the observed and predicted NDVI, disturbances with higher magnitude values represent a larger deviance from a stable trend, and therefore are more likely correspond to a correct disturbance detection [29,39]. On the contrary, lower magnitude values are often related to interannual variations or noise, and they correspond more often to false detection. As for trend, we found that small trend values usually correspond to a long stable period, which facilitates a correct detection; while high trend values (negative or positive) often were fitted either to a short historical data period or very noisy data (Figure 6). Grogan et al. [31] applied a variable of difference in slope (diff.slope) before and after the breakpoint together with amplitude and other variables and found that diff.slope enabled an increase in ROC AUC of 4–0.14% when used with magnitude and depending on forest-type. Thus, trend seems to be the second most important BFAST component to aid in disturbance detection, just after magnitude of change.
Data quality and model fitting period were ranked as either the second or third most important variables. Data quality contributed to accuracy improvement in the RF with TF dataset model in comparison with the baseline model (86.21% vs. 79.31%). The importance of data quality for disturbance detection has also been reported in [7,60]. In turn, the model fitting period is related to the model fitting quality, since models with longer fitting periods are more robust to interannual variation or noise (e.g., in Figure 6A,C). Time series data can have the same percentage of ‘no-data’ values, implying similar data quality, but the ones with a longer model fitting period would normally render better results. We interpret that this is the reason why model fitting period was selected in the SVM all-forest model, instead of data quality.

4.3. Disturbance Area and Rate Estimation

The increase in accuracy of the ML+BFAST compared with the baseline model is small (overall accuracy: 0.14%; area-weighted accuracy: 0.71%). In addition, both the estimated area of forest disturbance and disturbance rate of the ML+BFAST model overlaps with the previous estimate using the baseline model (baseline: 9136.43 ± 5599.49 ha vs. ML+BFAST: 5477.49 ± 1246.90 ha; baseline: 3654.7 ± 2239.8 ha year−1 vs. ML+BFAST: 2191.0 ± 977.6 ha year−1) [30]. However, in both cases, the ML+BFAST model has a smaller standard error, which translates to less uncertainty in the estimated disturbed area and disturbance rate. This is particularly relevant for forest disturbance quantification since it helps reduce uncertainties in estimating carbon and GHG emissions and therefore is significant for REDD+ implementation [1,65].
On the other hand, the results show that regardless of the method (i.e., using magnitude threshold or ML), most of the forest disturbances remain undetected in the final map (see Table 4, mapped disturbance vs. unbiased disturbance area estimate). Therefore, the validation with stratified sampling is essential to estimate the total area of forest disturbance in the study area.

4.4. Study Limitations

Admittedly, the spatial resolution of the Sentinel-2 images might have been too coarse to distinguish fine-scale degradations in the validation process. In these cases, other available imagery such as Google Earth could have been more informative; however, the problem we encountered was that frequently the available images were from very different dates, which introduced uncertainty in distinguishing seasonal changes from actual disturbances. Thus, we prioritized temporal agreement between the comparisons, instead of spatial resolution. A direct consequence of this decision was that fine-scale degradations could remain undetected in the validation process. Therefore, probably most of the detected disturbances correspond to deforestation. Particularly for very seasonal systems, like the one studied here, the availability of freely available images with high spatial and temporal resolution will be critical to help identify and verify fine-scale disturbances.
Another limitation shown by the followed approach is that it lacks a temporal label for each detected disturbance, as the final map only classifies areas into disturbance and non-disturbance in the monitoring period. One way to identify the dates of the disturbances is to consult the breakpoint date of each change detected by BFAST.

5. Conclusions

This study demonstrates that BFAST can benefit from using a machine learning approach to increase its accuracy for disturbance detection, particularly using magnitude of change, trend, and model fitting period. Support vector machines achieved higher accuracy using the all-forest dataset, while random forests over-performed support vector machines in forest-type specific datasets. We found that expert knowledge over-performed machine learning algorithms when using only magnitude as a predictive variable; however, when the number of variables increases, machine learning algorithms performed better than expert knowledge. Since we used a rather small sample size (238 points), the advantage of expert knowledge vs. machine learning algorithm needs to be tested with training data of a larger sample size to be conclusive. Nevertheless, our results suggest that ML algorithms can be used with BFAST to enhance the accuracy achieved to detect disturbances and obtain disturbance area estimates with lower errors. Future studies should address if the advantage of BFAST+ML for disturbance detection is relative to sample size or the scale of disturbances.

Author Contributions

Conceptualization, methodology, and data validation: J.V.S. and Y.G., Original draft preparation, review and editing: J.V.S. and Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Consejo Nacional de Ciencia y Tecnología (CONACYT) ‘Ciencia Básica’ SEP-285349.

Acknowledgments

The first author wishes to thank the Consejo Nacional de Ciencia y Tecnología (CONACyT) for providing a PhD scholarship.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. The extracted BFAST model components for the 238 training and validation data for tropical dry forest (TDF) and temperate forest (TF). The column “Change” includes disturbances, coded as 1, and non-disturbances, as 0.
Table A1. The extracted BFAST model components for the 238 training and validation data for tropical dry forest (TDF) and temperate forest (TF). The column “Change” includes disturbances, coded as 1, and non-disturbances, as 0.
ForestMagnitudeModel Fitting PeriodR2AmplitudeTrendData QualityChange
TDF−0.045313.339730.7425990.4247081.84 × 10−596.30390
TDF−0.0333615.443840.7844020.5280031.69 × 10−596.91380
TDF−0.034915.465750.7978030.5658931.75 × 10−596.528520
TDF−0.047918.794520.7699710.5675952.05 × 10−596.181310
TDF−0.0375623.306850.8200690.5431061.09 × 10−596.344620
TDF−0.0210214.808220.6284960.3196332.39 × 10−597.003330
TDF−0.0086412.857530.686010.3644893.27 × 10−596.825740
TDF−0.0663115.443840.6975060.4738922.15 × 10−596.647751
TDF−0.0591514.895890.6875180.4820082.30 × 10−596.395740
TDF−0.0556913.339730.7385730.4562862.86 × 10−596.221760
TDF−0.0561413.252060.786780.5244163.29 × 10−596.465480
TDF−0.0900813.120550.8029090.5316533.38 × 10−596.597080
TDF−0.070479.4849320.8097260.573344.58 × 10−595.755130
TDF−0.0792211.238360.7477890.4844055.31 × 10−596.295390
TDF−0.0680613.252060.7953910.5272322.91 × 10−596.527490
TDF−0.0773423.306850.8244350.5665051.97 × 10−596.379880
TDF−0.0517122.956170.7630230.4911651.58 × 10−596.730310
TDF−0.0792113.120550.7515320.5067483.74 × 10−596.513570
TDF−0.0960318.926030.7359890.5157021.73 × 10−596.656530
TDF−0.0881711.369860.7419980.4413813.89 × 10−596.217780
TDF−0.1028913.471230.567570.4504913.20 × 10−596.400980
TDF−0.105039.4410960.7430410.492615.59 × 10−595.880480
TDF−0.138124.3589040.37720.1822086.35 × 10−594.535180
TDF−0.109613.383560.6992710.5105794.29 × 10−597.891940
TDF−0.1144813.120550.7353140.502834.16 × 10−596.597080
TDF−0.1004610.843840.6963410.4654794.73 × 10−596.463751
TDF−0.1062312.156160.6352210.4705235.27 × 10−596.48490
TDF−0.114918.2575340.6682420.3887489.26 × 10−594.958540
TDF−0.1165323.306850.8025330.5364941.30 × 10−595.956750
TDF−0.104513.339730.7521790.494322.59 × 10−596.529780
TDF−0.1115811.282190.7686330.5060314.18 × 10−595.921340
TDF−0.1221811.282190.7751960.4624324.91 × 10−596.042730
TDF−0.1122412.857530.7446770.4812454.12 × 10−596.463570
TDF−0.132798.9589040.7196250.4438816.46 × 10−595.383670
TDF−0.128529.3972610.7081840.4130035.80 × 10−596.939670
TDF−0.1039720.895890.7456970.4916352.77 × 10−597.1290
TDF−0.124411.893150.7117110.4614654.04 × 10−596.2690
TDF−0.1109416.164380.6997650.4226343.46 × 10−596.610740
TDF−0.122918.301370.6397680.2704525.85 × 10−595.315080
TDF−0.1251823.306850.7401040.3493122.01 × 10−596.732491
TDF−0.1297423.306850.747040.4638439.89 × 10−696.45040
TDF−0.110589.5287680.6387870.4345275.24 × 10−596.435760
TDF−0.1514723.306850.7463980.5028241.15 × 10−596.638461
TDF−0.187089.4410960.8186640.5840144.80 × 10−595.793441
TDF−0.1721210.230140.6972910.4347035.31 × 10−596.760381
TDF−0.179979.4410960.671160.3600174.48 × 10−595.706411
TDF−0.1823110.843840.7507880.4805624.91 × 10−596.261680
TDF−0.156956.3315070.5878130.2655930.00010495.025960
TDF−0.173382.2136990.4573110.2538589.94 × 10−591.223730
TDF−0.1673818.706850.8175850.5628542.83 × 10−596.119491
TDF−0.1555513.383560.7967060.5748322.33 × 10−596.336471
TDF−0.1648511.849310.8194310.5423924.09 × 10−596.370780
TDF−0.1586313.230140.6137970.4355013.96 × 10−598.074530
TDF−0.154314.435620.5918120.3974663.81 × 10−597.30550
TDF−0.153646.20.6355580.2886528.94 × 10−594.920490
TDF−0.159169.002740.7816140.4213438.11 × 10−595.61910
TDF−0.1629513.252060.6697820.3638894.88 × 10−596.093430
TDF−0.177848.0821910.8067990.4119570.00010795.730260
TDF−0.159062.2136990.6026170.3296518.89 × 10−591.965390
TDF−0.1548713.252060.6813630.4561783.65 × 10−596.58950
TDF−0.1840715.443840.7046410.4555822.65 × 10−596.807381
TDF−0.181268.301370.6696720.3835957.59 × 10−595.249090
TDF−0.152829.4410960.7486140.4647246.52 × 10−596.518710
TDF−0.154598.2136980.6782280.3733258.90 × 10−595.231740
TDF−0.15413.383560.7724060.4910272.40 × 10−596.68440
TDF−0.154388.2136980.73760.3704418.89 × 10−595.49850
TDF−0.169548.2575340.6977360.3448044.68 × 10−595.422881
TDF−0.159482.1260270.3946080.1930959.77 × 10−592.020590
TDF−0.20164.9726030.5375380.2372467.83 × 10−594.438320
TDF−0.2154514.391780.7021730.4222953.38 × 10−596.003040
TDF−0.201379.3095890.7712020.4970916.13 × 10−596.087080
TDF−0.219439.002740.7251470.4613194.80 × 10−595.710370
TDF−0.231519.2219180.3878950.1800289.39 × 10−595.782590
TDF−0.2174515.465750.7568690.4849933.41 × 10−595.908611
TDF−0.210892.1260270.3967370.2513410.00013692.792791
TDF−0.2143112.769860.7361910.473573.91 × 10−596.460751
TDF−0.2127412.156160.7125970.3796222.98 × 10−596.034251
TDF−0.2297913.339730.7372810.5109733.07 × 10−596.837781
TDF−0.2008113.471230.7483930.5000662.53 × 10−596.522980
TDF−0.211949.2219180.5755990.3916826.96 × 10−595.69350
TDF−0.23652.0383560.4172970.1120160.00031491.812080
TDF−0.212859.2219180.8205890.5483745.06 × 10−595.75291
TDF−0.2496712.068490.7466060.4807995.48 × 10−596.096231
TDF−0.224094.3589040.5077920.254950.00012894.472360
TDF−0.225829.002740.7765780.5136657.33 × 10−596.01461
TDF−0.2628813.383560.7403850.4470472.50 × 10−596.725331
TDF−0.252337.6904110.7816090.32828.94 × 10−595.370371
TDF−0.267581.205480.1730270.1271810.00018493.197280
TDF−0.255713.120550.694260.4026741.98 × 10−596.346561
TDF−0.2915115.465750.7132070.3184983.80 × 10−596.794191
TDF−0.257234.9726030.5963360.3243134.65 × 10−594.658591
TDF−0.251249.3534250.701720.3727965.92 × 10−596.076131
TDF−0.252223.306850.7771870.4752742.21 × 10−595.909731
TDF−0.2561713.339730.6133080.3770542.39 × 10−596.447641
TDF−0.264219.320550.777710.495982.09 × 10−596.32781
TDF−0.259945.0164380.5370490.2792990.00011294.650661
TDF−0.2857611.282190.6412840.4077935.45 × 10−595.824231
TDF−0.255077.7780820.698470.3287020.00011295.387320
TDF−0.267361.9287670.4307750.102710.00027492.056740
TDF−0.261814.797260.3452560.2307167.42 × 10−595.262560
TDF−0.2748314.282190.6699390.4947233.56 × 10−597.813580
TDF−0.3123113.252060.646580.306754.80 × 10−596.899541
TDF−0.30479.4849320.6451150.4209376.90 × 10−595.697371
TDF−0.348945.9369860.7454170.1517470.00019594.972330
TDF−0.306964.8191780.5450390.2580978.34 × 10−594.3751
TDF−0.330394.9287670.4788120.2908190.00011795.222221
TDF−0.320615.104110.6060130.1620130.00016995.493560
TDF−0.330841.9726030.4447680.0967480.00032992.233010
TDF−0.310882.0164380.4609190.171070.00025391.723210
TDF−0.311573.2219180.3290940.3057075.59 × 10−593.712831
TDF−0.343174.8191780.4955190.1138640.00016994.659090
TDF−0.3835719.931510.8126950.2636144.40 × 10−596.11051
TDF−0.354451.3150680.0934470.0866690.00025193.347190
TDF−0.3531214.720550.2744820.2483573.63 × 10−599.330110
TDF−0.365749.3095890.6586020.3926947.94 × 10−595.616361
TDF−0.3886423.306850.6797540.3282251.58 × 10−596.285851
TDF−0.3741223.306850.623830.2539922.19 × 10−596.238831
TDF−0.359135.5424660.5301870.1932770.00015995.652180
TDF−0.363522.1917810.6818850.1318290.00052293.133580
TDF−0.357099.4410960.6919550.2960646.17 × 10−595.706411
TDF−0.360664.7534250.6132450.4174674.41 × 10−694.930881
TDF−0.354188.3452050.6706510.3918388.52 × 10−595.438131
TDF−0.359525.8712330.4818780.143120.00014795.895520
TDF−0.353335.104110.5724290.1859310.00016295.278970
TDF−0.430615.0602740.5412950.0885740.00020795.346320
TDF−0.406754.797260.4144650.2926730.00011294.920091
TDF−0.43923.9643840.6250670.1014010.00021294.406070
TDF−0.417484.8191780.4264730.2917890.0001294.772731
TDF−0.434244.7534250.6742890.1342210.00020894.873270
TDF−0.402561.6438360.3962560.2619780.00044292.346090
TDF−0.420825.0164380.5432480.0777930.00019695.196510
TDF−0.4171619.057540.8556860.2885645.29 × 10−596.119021
TDF−0.405725.5424660.7819630.1296290.00024494.960470
TDF−0.464771.2493150.3231130.1273250.00048492.778990
TDF−0.4830113.383560.5910820.1901542.10 × 10−596.479741
TDF−0.4710914.720550.6001440.217341.68 × 10−596.576111
TDF−0.498061.2493150.2838910.1082660.00050193.216630
TF−0.0226212.945210.2900450.1620239.13 × 10−697.672450
TF−0.0473513.339730.5334570.3845512.15 × 10−596.550310
TF−0.0457218.531510.6100690.4354791.48 × 10−596.48190
TF−0.0392214.391780.4590610.3398412.66 × 10−597.01180
TF−0.0147622.956170.1944850.0611851.06 × 10−596.849640
TF−0.0916711.632880.6488290.320885.35 × 10−597.009650
TF−0.085117.084930.7943370.5410892.39 × 10−596.53680
TF−0.0721911.849310.7194950.3050592.03 × 10−596.671290
TF−0.050569.3095890.5273940.2376063.91 × 10−596.410710
TF−0.0595823.306850.8139450.5672061.74 × 10−595.96850
TF−0.0593211.893150.7915640.4915354.10 × 10−596.706590
TF−0.0559714.304110.7004370.4140253.26 × 10−597.357340
TF−0.0603922.956170.6519630.3711851.80 × 10−596.312650
TF−0.0508212.769860.4310050.1519471.39 × 10−597.297290
TF−0.090611.106850.6682540.3114516.02 × 10−596.942050
TF−0.1248821.443840.451620.1029761.95 × 10−599.667860
TF−0.1075621.40.7371560.2373931.74 × 10−599.667180
TF−0.1152211.893150.6847060.4154674.90 × 10−596.798710
TF−0.1017311.238360.7032770.3369833.85 × 10−595.905430
TF−0.1194213.55890.6965240.4457914.17 × 10−596.606060
TF−0.1336511.936990.7325920.3935464.92 × 10−596.718680
TF−0.1069912.989040.6808970.4559623.91 × 10−596.26740
TF−0.1287412.156160.6648490.3644633.99 × 10−596.665160
TF−0.1213212.156160.6942520.4449772.90 × 10−596.777820
TF−0.1194710.80.6286120.4077714.43 × 10−596.474760
TF−0.1028212.769860.7255710.4609123.73 × 10−596.503650
TF−0.1041214.282190.6752290.4206963.12 × 10−597.525890
TF−0.113511.849310.670750.4372913.66 × 10−596.601940
TF−0.111098.8712330.8024050.4739386.59 × 10−596.140790
TF−0.148178.6520550.6185840.3286326.05 × 10−596.58121
TF−0.12039.3534250.6655370.3994025.84 × 10−596.193270
TF−0.1720713.383560.7046620.4523272.81 × 10−596.602541
TF−0.160979.4410960.7852230.5454583.92 × 10−596.025531
TF−0.1503311.632880.6329140.3295454.00 × 10−596.774190
TF−0.1573823.306850.5534280.1653681.99 × 10−596.967560
TF−0.18073.6164380.4465010.2503698.97 × 10−595.760790
TF−0.1626211.849310.7191120.4709875.08 × 10−596.116511
TF−0.1676423.306850.6824120.3940821.45 × 10−596.250591
TF−0.1578910.580820.1075940.091385−8.90 × 10−697.256020
TF−0.1525114.808220.753050.4771274.00 × 10−596.707360
TF−0.1629413.383560.7259560.4593492.95 × 10−596.541141
TF−0.1746311.849310.7719810.423345.44 × 10−596.786870
TF−0.176764.7534250.5586080.2890024.29 × 10−595.391710
TF−0.1976517.961640.7499150.4920952.97 × 10−596.416051
TF−0.1656921.443840.6207880.1974822.65 × 10−599.629530
TF−0.1988612.857530.7612690.5060953.72 × 10−596.442270
TF−0.173354.4904110.6492410.3767169.77 × 10−598.109760
TF−0.16254.797260.6835430.3321433.88 × 10−596.004570
TF−0.170138.0821910.7544890.3775757.03 × 10−596.340230
TF−0.168378.3452050.7838460.3666756.62 × 10−596.75090
TF−0.2163213.120550.769590.5120754.15 × 10−596.054281
TF−0.2255923.306850.677850.2895141.75 × 10−596.191821
TF−0.2115412.857530.7665620.5380635.30 × 10−596.484871
TF−0.232912.989040.7208220.4593142.35 × 10−596.393931
TF−0.2014922.956170.3608880.1536791.54 × 10−596.348451
TF−0.2069815.290410.7104780.4618422.48 × 10−596.166251
TF−0.2248718.663010.7379290.4546783.01 × 10−596.462651
TF−0.229283.0904110.4657880.3965363.93 × 10−595.57130
TF−0.205564.6438360.5184880.2332820.00011596.933960
TF−0.2429923.306850.6845010.2874832.18 × 10−596.203581
TF−0.2263715.334250.6185010.3494967.43 × 10−696.016431
TF−0.2141111.893150.6337480.3828135.96 × 10−596.913860
TF−0.2862323.306850.5390630.2335481.54 × 10−596.215321
TF−0.2542314.435620.8114450.3875436.03 × 10−596.337761
TF−0.264784.7095890.6508020.404377.81 × 10−596.104651
TF−0.2736811.893150.7603290.4816025.42 × 10−596.798711
TF−0.262428.0821910.4974890.222147.84 × 10−596.001360
TF−0.2591813.339730.7264440.2757182.65 × 10−596.509241
TF−0.2965923.306850.6152740.2093822.50 × 10−596.250591
TF−0.2756814.347950.5981750.3080848.78 × 10−696.372661
TF−0.26318.8712330.6464650.2780769.89 × 10−596.171660
TF−0.275517.1643830.4408420.2271699.23 × 10−595.756880
TF−0.259387.1643830.524560.2608049.30 × 10−595.833340
TF−0.2765822.956170.6657630.3733531.70 × 10−596.467781
TF−0.318045.9589040.7208740.4960089.05 × 10−599.356620
TF−0.3024923.306850.5810940.2183891.65 × 10−596.29761
TF−0.3220823.306850.6114730.2388842.01 × 10−596.191821
TF−0.32784.7095890.521520.355638.87 × 10−596.104651
TF−0.3370323.306850.6149240.2520181.54 × 10−596.250591
TF−0.3358223.306850.6296370.2665932.00 × 10−596.262341
TF−0.345224.797260.4982160.2302615.87 × 10−594.977171
TF−0.3427512.375340.7515760.3025514.66 × 10−596.281541
TF−0.358123.306850.4635260.1964571.43 × 10−596.227081
TF−0.360957.3616440.5968330.2095010.00014696.056550
TF−0.3679423.306850.5229180.2240431.50 × 10−596.344621
TF−0.3640812.857530.5031830.2894441.40 × 10−596.250531
TF−0.3593123.306850.5927580.2194931.60 × 10−596.274091
TF−0.3624212.945210.592870.3689992.72 × 10−595.937371
TF−0.3726923.306850.5890950.226551.72 × 10−596.321111
TF−0.3823511.150690.552480.1841921.70 × 10−595.996071
TF−0.3720123.306850.5535740.1124162.19 × 10−597.038081
TF−0.3785412.331510.6652110.2512122.17 × 10−596.246121
TF−0.4155710.273970.5345460.1785082.16 × 10−596.161031
TF−0.4168523.306850.5618890.1990971.80 × 10−596.309361
TF−0.4167210.756160.5604940.1699532.24 × 10−596.154821
TF−0.4032213.339730.2953620.094621.64 × 10−597.412731
TF−0.4397323.306850.5703380.2288541.47 × 10−596.309361
TF−0.4150723.306850.5221610.2072581.56 × 10−596.203581
TF−0.4559823.306850.5606140.2092941.73 × 10−596.321111
TF−0.487164.797260.6914860.3001890.00012494.977170
TF−0.4524723.306850.5604780.2053251.56 × 10−596.415141
Table A2. Mean accuracy in range of confidence interval at 95% level obtained for each of the constructed models using the RF and SVM algorithms. The models were trained using only the random subsets of the training data (not validation data), with 120 iterations. Additionally, the models were applied in three datasets, All-forest, TDF, TF, and the variables for each dataset are indicated. The best model, marked by underlined bold font, is selected by a balance between the accuracy and the number of variables. Given similar accuracy, the models with lower number of predictive variables were preferred.
Table A2. Mean accuracy in range of confidence interval at 95% level obtained for each of the constructed models using the RF and SVM algorithms. The models were trained using only the random subsets of the training data (not validation data), with 120 iterations. Additionally, the models were applied in three datasets, All-forest, TDF, TF, and the variables for each dataset are indicated. The best model, marked by underlined bold font, is selected by a balance between the accuracy and the number of variables. Given similar accuracy, the models with lower number of predictive variables were preferred.
DatasetRF SVM
Predictive VariablesAccuracy (%)Predictive VariablesAccuracy (%)
All-forestBaseline76.78 ± 0.87Baseline76.78 ± 0.87
Magnitude70.64 ± 0.84Magnitude73.7 ± 0.96
Magnitude + trend84.15 ± 0.56Magnitude + trend86.45 ± 0.64
Magnitude + trend + model fitting period85.45 ± 0.66Magnitude + trend + model fitting period87.35 ± 0.59
Magnitude + trend + model fitting period + amplitude85.62 ± 0.68Magnitude + trend + model fitting period + forest-type86.02 ± 0.67
Magnitude + trend + model fitting period + amplitude + data quality85.87 ± 0.76Magnitude + trend + model fitting period + forest-type + data quality87 ± 0.61
Magnitude + trend + model fitting period + amplitude + data quality + R285.52 ± 0.76Magnitude + trend + model fitting period + forest-type + data quality + amplitude87.89 ± 0.71
Magnitude + trend + model fitting period + amplitude + data quality + R2 + forest-type85.04 ± 0.8Magnitude + trend + model fitting period + forest-type + data quality + amplitude + R286.56 ± 0.67
TDFBaseline71.04 ± 1.08Baseline71.04 ± 1.08
Magnitude61.48 ± 1.25Magnitude66.9 ± 1.12
Magnitude + trend 79.17 ± 1.06Magnitude + trend81.93 ± 1.03
Magnitude + trend + amplitude80.96 ± 0.93Magnitude + trend + model fitting period81.9 ± 1.01
Magnitude + trend + amplitude + data quality81.38 ± 1Magnitude + trend + model fittingperiod + data quality84.22 ± 1.02
Magnitude + trend + amplitude + data quality + R281.93 ± 0.93Magnitude + trend + model fitting period+ data quality + R284.58 ± 0.98
Magnitude + trend + amplitude + data quality + R2 + model fitting period80.29 ± 1.04Magnitude + trend + model fitting period+ data quality + R2 + amplitude83.41 ± 1.04
TFBaseline84.75 ± 1.17Baseline84.75 ± 1.17
Magnitude79.46 ± 1.22Magnitude80.04 ± 1.15
Magnitude + data quality85.85 ± 1.2Magnitude + model fitting period85.97 ± 1.15
Magnitude + data quality + trend85.65 ± 1.16Magnitude + model fitting period + trend85.89 ± 1.07
Magnitude + data quality + trend + model fitting period85.28 ± 1.09Magnitude + model fitting period + trend + data quality86.55 ± 1.03
Magnitude + data quality + trend + model fitting period + R284.82 ± 1.16Magnitude + model fitting period + trend + data quality + R286.23 ± 1.1
Magnitude + data quality + trend + model fitting period + R2 + amplitude84.64 ± 1.18Magnitude + model fitting period + trend + data quality + R2 + amplitude85.43 ± 1.11

References

  1. Frolking, S.; Palace, M.W.; Clark, D.B.; Chambers, J.Q.; Shugart, H.H.; Hurtt, G.C. Forest disturbance and recovery: A general review in the context of spaceborne remote sensing of impacts on aboveground biomass and canopy structure. J. Geophys. Res. Biogeosciences 2009, 114, G00E02. [Google Scholar] [CrossRef]
  2. van der Werf, G.R.; Morton, D.C.; DeFries, R.S.; Olivier, J.G.J.; Kasibhatla, P.S.; Jackson, R.B.; Collatz, G.J.; Randerson, J.T. CO2 emissions from forest loss. Nat. Geosci. 2009, 2, 737–738. [Google Scholar] [CrossRef]
  3. FAO. Assessing Forest Degradation. Towards the Development of Globally Applicable Guidelines; FAO: Roma, Italy, 2011. [Google Scholar]
  4. Collins, M.; Knutti, R.; Arblaster, J.; Dufresne, J.-L.; Fichefet, T.; Friedlingstein, P.; Gao, X.; Gutowski, W.J.; Johns, T.; Krinner, G.; et al. Long-term Climate Change: Projections, Commitments and Irreversibility. In Climate Change 2013: The Physical Science Basis. Contribution of Working Group I to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change; Stocker, T.F., Qin, D., Plattner, G.-K., Tignor, M., Allen, S.K., Boschung, J., Nauels, A., Xia, Y., Bex, V., Midgley, P.M., Eds.; Cambridge University Press: Cambridge, UK, 2013. [Google Scholar]
  5. Clark, D. The Role of Disturbance in the Regeneration of Neotropical Moist Forests. In Reproductive Ecology of Tropical Forest Plants; Bawa, K.S., Hadley, M., Eds.; UNESCO, Parthenon Publishing Group: Halifax, UK, 1990. [Google Scholar]
  6. Chuvieco, E.; Aguado, I.; Salas, J.; García, M.; Yebra, M.; Oliva, P. Satellite Remote Sensing Contributions to Wildland Fire Science and Management. Curr. For. Rep. 2020, 6, 81–96. [Google Scholar] [CrossRef]
  7. Dutrieux, L.P.; Jakovac, C.C.; Latifah, S.H.; Kooistra, L. Reconstructing land use history from Landsat time-series: Case study of a swidden agriculture system in Brazil. Int. J. Appl. Earth Obs. Geoinf. 2016, 47, 112–124. [Google Scholar] [CrossRef]
  8. FAO. Global Forest Resources Assessment; FAO: Rome, Italy, 2020. [Google Scholar]
  9. Vieilledent, G.; Grinand, C.; Rakotomalala, F.A.; Ranaivosoa, R.; Rakotoarijaona, J.R.; Allnutt, T.F.; Achard, F. Combining global tree cover loss data with historical national forest cover maps to look at six decades of deforestation and forest fragmentation in Madagascar. Biol. Conserv. 2018, 222, 189–197. [Google Scholar] [CrossRef]
  10. Kennedy, R.E.; Yang, Z.; Cohen, W.B. Detecting trends in forest disturbance and recovery using yearly Landsat time series: 1. LandTrendr—Temporal segmentation algorithms. Remote Sens. Environ. 2010, 114, 2897–2910. [Google Scholar] [CrossRef]
  11. Verbesselt, J.; Hyndman, R.; Newnham, G.; Culvenor, D. Detecting trend and seasonal changes in satellite image time series. Remote Sens. Environ. 2010, 114, 106–115. [Google Scholar] [CrossRef]
  12. Zhu, Z.; Woodcock, C.E.; Olofsson, P. Continuous monitoring of forest disturbance using all available Landsat imagery. Remote Sens. Environ. 2012, 122, 75–91. [Google Scholar] [CrossRef]
  13. Hirschmugl, M.; Gallaun, H.; Dees, M.; Datta, P.; Deutscher, J.; Koutsias, N.; Schardt, M. Methods for Mapping Forest Disturbance and Degradation from Optical Earth Observation Data: A Review. Curr. For. Rep. 2017, 3, 32–45. [Google Scholar] [CrossRef] [Green Version]
  14. Schneibel, A.; Stellmes, M.; Röder, A.; Frantz, D.; Kowalski, B.; Haß, E.; Hill, J. Assessment of spatio-temporal changes of smallholder cultivation patterns in the Angolan Miombo belt using segmentation of Landsat time series. Remote Sens. Environ. 2017, 195, 118–129. [Google Scholar] [CrossRef] [Green Version]
  15. Chi, M.; Plaza, A.; Benediktsson, J.A.; Sun, Z.; Shen, J.; Zhu, Y. Big Data for Remote Sensing: Challenges and Opportunities. Proc. IEEE 2016, 104, 2207–2219. [Google Scholar] [CrossRef]
  16. Xiao, J.; Chevallier, F.; Gomez, C.; Guanter, L.; Hicke, J.A.; Huete, A.R.; Ichii, K.; Ni, W.; Pang, Y.; Rahman, A.F.; et al. Remote sensing of the terrestrial carbon cycle: A review of advances over 50 years. Remote Sens. Environ. 2019, 233, 111383. [Google Scholar] [CrossRef]
  17. Mountrakis, G.; Im, J.; Ogole, C. Support vector machines in remote sensing: A review. ISPRS J. Photogramm. Remote Sens. 2011, 66, 247–259. [Google Scholar] [CrossRef]
  18. Sheykhmousa, M.; Mahdianpari, M.; Ghanbari, H.; Mohammadimanesh, F.; Ghamisi, P.; Homayouni, S. Support Vector Machine Versus Random Forest for Remote Sensing Image Classification: A Meta-Analysis and Systematic Review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 6308–6325. [Google Scholar] [CrossRef]
  19. Camargo, F.F.; Sano, E.E.; Almeida, C.M.; Mura, J.C.; Almeida, T. A comparative assessment of machine-learning techniques for land use and land cover classification of the Brazilian tropical savanna using ALOS-2/PALSAR-2 polarimetric images. Remote Sens. 2019, 11, 1600. [Google Scholar] [CrossRef] [Green Version]
  20. Grinand, C.; Rakotomalala, F.; Gond, V.; Vaudry, R.; Bernoux, M.; Vieilledent, G. Estimating deforestation in tropical humid and dry forests in Madagascar from 2000 to 2010 using multi-date Landsat satellite images and the random forests classifier. Remote Sens. Environ. 2013, 139, 68–80. [Google Scholar] [CrossRef]
  21. Dlamini, W.M. Analysis of deforestation patterns and drivers in Swaziland using efficient Bayesian multivariate classifiers. Modeling Earth Syst. Environ. 2016, 2, 1–14. [Google Scholar] [CrossRef]
  22. Ali, I.; Greifeneder, F.; Stamenkovic, J.; Neumann, M.; Notarnicola, C. Review of machine learning approaches for biomass and soil moisture retrievals from remote sensing data. Remote Sens. 2015, 7, 16398–16421. [Google Scholar] [CrossRef] [Green Version]
  23. Wu, L.; Li, Z.; Liu, X.; Zhu, L.; Tang, Y.; Zhang, B.; Xu, B.; Liu, M.; Meng, Y.; Liu, B. Multi-type forest change detection using BFAST and monthly landsat time series for monitoring spatiotemporal dynamics of forests in subtropical wetland. Remote Sens. 2020, 12, 341. [Google Scholar] [CrossRef] [Green Version]
  24. Xu, Y.; Yu, L.; Peng, D.; Zhao, J.; Cheng, Y.; Liu, X.; Li, W.; Meng, R.; Xu, X.; Gong, P. Annual 30-m land use/land cover maps of China for 1980–2015 from the integration of AVHRR, MODIS and Landsat data using the BFAST algorithm. Sci. China Earth Sci. 2020, 63, 1390–1407. [Google Scholar] [CrossRef]
  25. Fang, X.; Zhu, Q.; Ren, L.; Chen, H.; Wang, K.; Peng, C. Large-scale detection of vegetation dynamics and their potential drivers using MODIS images and BFAST: A case study in Quebec, Canada. Remote Sens. Environ. 2018, 206, 391–402. [Google Scholar] [CrossRef]
  26. Zhu, Z.; Woodcock, C.E. Continuous change detection and classification of land cover using all available Landsat data. Remote Sens. Environ. 2014, 144, 152–171. [Google Scholar] [CrossRef] [Green Version]
  27. Simoes, R.; Camara, G.; Queiroz, G.; Souza, F.; Andrade, P.R.; Santos, L.; Carvalho, A.; Ferreira, K. Satellite Image Time Series Analysis for Big Earth Observation Data. Remote Sens. 2021, 13, 2428. [Google Scholar] [CrossRef]
  28. Schultz, M.; Shapiro, A.; Clevers, J.G.P.W.; Beech, C.; Herold, M. Forest cover and vegetation degradation detection in the Kavango Zambezi Transfrontier Conservation area using BFAST monitor. Remote Sens. 2018, 10, 1850. [Google Scholar] [CrossRef] [Green Version]
  29. Verbesselt, J.; Zeileis, A.; Herold, M. Near real-time disturbance detection using satellite image time series. Remote Sens. Environ. 2012, 123, 98–108. [Google Scholar] [CrossRef]
  30. Gao, Y.; Solórzano, J.V.; Quevedo, A.; Loya-Carrillo, J.O. How BFAST Trend and Seasonal Model Components Affect Disturbance Detection in Tropical Dry Forest and Temperate Forest. Remote Sens. 2021, 2, 2033. [Google Scholar] [CrossRef]
  31. Grogan, K.; Pflugmacher, D.; Hostert, P.; Verbesselt, J.; Fensholt, R. Mapping clearances in tropical dry forests using breakpoints, trend, and seasonal components from modis time series: Does forest type matter? Remote Sens. 2016, 8, 657. [Google Scholar] [CrossRef] [Green Version]
  32. Watts, L.M.; Laffan, S.W. Effectiveness of the BFAST algorithm for detecting vegetation response patterns in a semi-arid region. Remote Sens. Environ. 2014, 154, 234–245. [Google Scholar] [CrossRef]
  33. Trejo, I.; Dirzo, R. Deforestation of seasonally dry tropical forest: A national and local analysis in Mexico. Biol. Conserv. 2000, 94, 133–142. [Google Scholar] [CrossRef]
  34. de la Barreda-Bautista, B.; López-Caloca, A.A.; Couturier, S.; Silván-Cárdenas, J.L. Tropical Dry Forests in the Global Picture: The Challenge of Remote Sensing-Based Change Detection in Tropical Dry Environments; Carayannis, E., Ed.; InTech: Rijeka, Croatia, 2011; pp. 231–256. [Google Scholar]
  35. GOFC-GOLD. A Sourcebook of Methods and Procedures for Monitoring and Reporting Anthropogenic Greenhouse; GOFC-GOLD Report Version COP21-1; Wageningen University: Wageningen, The Netherlands, 2016; pp. 1–266. [Google Scholar]
  36. Jakovac, C.C.; Dutrieux, L.P.; Siti, L.; Peña-Claros, M.; Bongers, F. Spatial and temporal dynamics of shifting cultivation in the middle-Amazonas river: Expansion and intensification. PLoS ONE 2017, 12, e0181092. [Google Scholar] [CrossRef]
  37. Cuevas, R.N.N.; Guzmán, H.L. El Bosque Tropical Caducifolio En La Reserva de La Biosfera Sierra Manantlan, Jalisco-Colima, Mexico. Bol. IBUG 1998, 5, 445–491. [Google Scholar]
  38. Borrego, A.; Skutsch, M. How Socio-Economic Differences Between Farmers Affect Forest Degradation in Western Mexico. Forests 2019, 10, 893. [Google Scholar] [CrossRef] [Green Version]
  39. DeVries, B.; Verbesselt, J.; Kooistra, L.; Herold, M. Robust monitoring of small-scale forest disturbances in a tropical montane forest using Landsat time series. Remote Sens. Environ. 2015, 161, 107–121. [Google Scholar] [CrossRef]
  40. Drusch, M.; Del Bello, U.; Carlier, S.; Colin, O.; Fernandez, V.; Gascon, F.; Hoersch, B.; Isola, C.; Laberinti, P.; Martimort, P.; et al. Sentinel-2: ESA’s Optical High-Resolution Mission for GMES Operational Services. Remote Sens. Environ. 2012, 120, 25–36. [Google Scholar] [CrossRef]
  41. Hamunyela, E.; Verbesselt, J.; Bruin, S.d.; Herold, M. Monitoring deforestation at sub-annual scales as extreme events in landsat data cubes. Remote Sens. 2016, 8, 651. [Google Scholar] [CrossRef]
  42. Schultz, M.; Clevers, J.G.P.W.; Carter, S.; Verbesselt, J.; Avitabile, V.; Quang, H.V.; Herold, M. Performance of vegetation indices from Landsat time series in deforestation monitoring. Int. J. Appl. Earth Obs. Geoinf. 2016, 52, 318–327. [Google Scholar] [CrossRef]
  43. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  44. Boser, B.; Guyon, I.; Vapnik, V. A Training Algorithm for Optimal Margin Classifiers. In Proceedings of the Fifth Annual Workshop on Computational Learning Theory, Pittsburgh, PA, USA, 27–29 July 1992. [Google Scholar]
  45. Dabija, A.; Kluczek, M.; Zagajewski, B.; Raczko, E.; Kycko, M.; Al-Sulttani, A.H.; Tardà, A.; Pineda, L.; Corbera, J. Comparison of support vector machines and random forests for corine land cover mapping. Remote Sens. 2021, 13, 777. [Google Scholar] [CrossRef]
  46. Simoes, R.; Picoli, M.C.A.; Camara, G.; Maciel, A.; Santos, L.; Andrade, P.R.; Sánchez, A.; Ferreira, K.; Carvalho, A. Land use and cover maps for Mato Grosso State in Brazil from 2001 to 2017. Sci. Data 2020, 7, 34. [Google Scholar] [CrossRef] [Green Version]
  47. Karatzoglou, A.; Smola, A.; Hornik, K.; Zeileis, A. kernlab. An S4 Package for Kernel Methods in R. J. Stat. Softw. 2004, 11, 1–20. [Google Scholar] [CrossRef] [Green Version]
  48. Kuhn, M. Caret: Classification and Regression Training. 2021. Available online: https://CRAN.R-project.org/package=caret (accessed on 10 December 2021).
  49. Card, D.H. Using Known Map Category Marginal Frequencies To Improve Estimates of Thematic Map Accuracy. Photogramm. Eng. Remote Sens. 1982, 48, 431–439. [Google Scholar]
  50. Maxwell, A.E.; Warner, T.A.; Guillén, L.A. Accuracy assessment in convolutional neural network-based deep learning remote sensing studies—Part 2: Recommendations and best practices. Remote Sens. 2021, 13, 2591. [Google Scholar] [CrossRef]
  51. Goutte, C.; Gaussier, E. A Probabilistic Interpretation of Precision, Recall and F-Score, with Implication for Evaluation. Lect. Notes Comput. Sci. 2005, 3408, 345–359. [Google Scholar] [CrossRef]
  52. Olofsson, P.; Foody, G.M.; Herold, M.; Stehman, S.V.; Woodcock, C.E.; Wulder, M.A. Good practices for estimating area and assessing accuracy of land change. Remote Sens. Environ. 2014, 148, 42–57. [Google Scholar] [CrossRef]
  53. Kuhn, M.; Wickham, H. Tidymodels: A Collection of Packages for Modeling and Machine Learning Using Tidyverse Principles. 2020. Available online: https://www.tidymodels.org (accessed on 10 December 2021).
  54. Kuhn, M.; Vaughan, D. Yardstick: Tidy Characterizations of Model Performance. 2021. Available online: https://CRAN.R-project.org/package=yardstick (accessed on 10 December 2021).
  55. Pebesma, E. Simple Features for R: Standardized Support for Spatial Vector Data. R J. 2018, 10, 439–446. [Google Scholar] [CrossRef] [Green Version]
  56. Hijmans, R.J. Raster: Geographic Data Analysis and Modeling. 2021. Available online: https://CRAN.R-project.org/package=raster (accessed on 10 December 2021).
  57. Wickham, H.; Averick, M.; Bryan, J.; Chang, W.; D’Agostino McGowan, L.; François, R.; Grolemund, G.; Hayes, A.; Henry, L.; Hester, J.; et al. Tidyverse. J. Open Source Softw. 2019, 4, 1686. [Google Scholar] [CrossRef]
  58. Heydari, S.S.; Mountrakis, G. Effect of classifier selection, reference sample size, reference class distribution and scene heterogeneity in per-pixel classification accuracy using 26 Landsat sites. Remote Sens. Environ. 2017, 204, 648–658. [Google Scholar] [CrossRef]
  59. Vabalas, A.; Gowen, E.; Poliakoff, E.; Casson, A.J. Machine learning algorithm validation with a limited sample size. PLoS ONE 2019, 14, e0224365. [Google Scholar] [CrossRef]
  60. Schultz, M.; Verbesselt, J.; Avitabile, V.; Souza, C.; Herold, M. Error Sources in Deforestation Detection Using BFAST Monitor on Landsat Time Series Across Three Tropical Sites. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 3667–3679. [Google Scholar] [CrossRef]
  61. Geng, L.; Che, T.; Wang, X.; Wang, H. Detecting spatiotemporal changes in vegetation with the BFAST model in the Qilian Mountain region during 2000–2017. Remote Sens. 2019, 11, 103. [Google Scholar] [CrossRef] [Green Version]
  62. Roy, D.P.; Kovalskyy, V.; Zhang, H.K.; Vermote, E.F.; Yan, L.; Kumar, S.S.; Egorov, A. Characterization of Landsat-7 to Landsat-8 reflective wavelength and normalized difference vegetation index continuity. Remote Sens. Environ. 2016, 185, 57–70. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  63. Hirschmugl, M.; Deutscher, J.; Sobe, C.; Bouvet, A.; Mermoz, S.; Schardt, M. Use of SAR and Optical Time Series for Tropical Forest Disturbance Mapping. Remote Sens. 2020, 12, 727. [Google Scholar] [CrossRef] [Green Version]
  64. Ma, W.; Gong, C.; Hu, Y.; Meng, P.; Xu, F. The Hughes Phenomenon in Hyperspectral Classification Based on the Ground Spectrum of Grasslands in the Region around Qinghai Lake. In International Symposium on Photoelectronic Detection and Imaging 2013: Imaging Spectrometer Technologies and Applications; International Society for Optics and Photonics: Bellingham, WA, USA, 2013; Volume 8910, p. 89101G. [Google Scholar] [CrossRef]
  65. Herold, M.; Román-Cuesta, R.M.; Mollicone, D.; Hirata, Y.; Van Laake, P.; Asner, G.P.; Souza, C.; Skutsch, M.; Avitabile, V.; MacDicken, K. Options for monitoring and estimating historical carbon emissions from forest degradation in the context of REDD+. Carbon Balance Manag. 2011, 6, 13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. (A) Study area location and spatial distribution of the temperate and tropical dry forest. (B) Illustrations of sample points manually classified as disturbance and non-disturbance in the tropical dry forest (TDF) and temperate forest (TF) using Sentinel-2 natural color composites available in Google Earth Engine. Here the sample points are depicted by circles with yellow outlines.
Figure 1. (A) Study area location and spatial distribution of the temperate and tropical dry forest. (B) Illustrations of sample points manually classified as disturbance and non-disturbance in the tropical dry forest (TDF) and temperate forest (TF) using Sentinel-2 natural color composites available in Google Earth Engine. Here the sample points are depicted by circles with yellow outlines.
Remotesensing 14 00803 g001
Figure 2. Flowchart of the method.
Figure 2. Flowchart of the method.
Remotesensing 14 00803 g002
Figure 3. Variable importance scaled to a 0–100 range for the random forests (A) and support vector machines (B) models.
Figure 3. Variable importance scaled to a 0–100 range for the random forests (A) and support vector machines (B) models.
Remotesensing 14 00803 g003
Figure 4. ROC curves for the RF and SVM models in the all-forest, TDF, and TF datasets. AUC values are shown in percentages.
Figure 4. ROC curves for the RF and SVM models in the all-forest, TDF, and TF datasets. AUC values are shown in percentages.
Remotesensing 14 00803 g004
Figure 5. Visual comparison of the predicted disturbance by the most accurate models using different datasets. The points indicate the validation data for the disturbance class (orange) and non-disturbance class (blue) (A) Complete study area disturbance prediction using the all-forest baseline model. Examples of the disturbed areas detected using different models and datasets are presented in line; (B) All-forest; (C) Tropical dry forest (TDF); and (D) temperate forest (TF). The three best models are in column 1: Baseline model, 2: random forests (RF), and 3: support vector machines (SVM). The three maps at line B illustrate the disturbance detected by the three models using the all-forest dataset; similarly, the three maps at line C illustrate the disturbance with the TDF dataset and line D with the TF dataset.
Figure 5. Visual comparison of the predicted disturbance by the most accurate models using different datasets. The points indicate the validation data for the disturbance class (orange) and non-disturbance class (blue) (A) Complete study area disturbance prediction using the all-forest baseline model. Examples of the disturbed areas detected using different models and datasets are presented in line; (B) All-forest; (C) Tropical dry forest (TDF); and (D) temperate forest (TF). The three best models are in column 1: Baseline model, 2: random forests (RF), and 3: support vector machines (SVM). The three maps at line B illustrate the disturbance detected by the three models using the all-forest dataset; similarly, the three maps at line C illustrate the disturbance with the TDF dataset and line D with the TF dataset.
Remotesensing 14 00803 g005
Figure 6. Example of areas with correct disturbance detection ((A) TDF and (C) TF) and false detection ((B) TDF and (D) TF) in the baseline model, but correctly classified as non-disturbance in the SVM all-forest model. Each point corresponds to an observed NDVI value in the time series, while its color indicates its correspondence to the historical (green) or monitoring period (red). In turn, the blue solid line shows the model fitted to the data in the historical period and projected to the monitoring window. The dashed lines show the NDVI behavior through time in the historical (green) and monitoring periods (red). Finally, the black dashed line indicates the start of the monitoring period, while the solid yellow line indicates the date of the detected breakpoint.
Figure 6. Example of areas with correct disturbance detection ((A) TDF and (C) TF) and false detection ((B) TDF and (D) TF) in the baseline model, but correctly classified as non-disturbance in the SVM all-forest model. Each point corresponds to an observed NDVI value in the time series, while its color indicates its correspondence to the historical (green) or monitoring period (red). In turn, the blue solid line shows the model fitted to the data in the historical period and projected to the monitoring window. The dashed lines show the NDVI behavior through time in the historical (green) and monitoring periods (red). Finally, the black dashed line indicates the start of the monitoring period, while the solid yellow line indicates the date of the detected breakpoint.
Remotesensing 14 00803 g006
Table 1. Definition and equation of the BFAST components that were used as predictive variables in the machine learning algorithms.
Table 1. Definition and equation of the BFAST components that were used as predictive variables in the machine learning algorithms.
Seasonal and Trend Model ComponentsDefinitionEquation
MagnitudeThe absolute value of the difference between the observed and projected NDVI m a g n i t u d e = | y b r e a k y ^ b r e a k |
TrendThe trend of the linear model that was fitted in the stable historical data period s l o p e = n i ( x i y i ) ( i x i ) ( i y i ) n ( i x i 2 ) ( i x i ) 2
AmplitudeThe difference between the maximum and minimum NDVI of the detrended model a m p l i t u d e = y ^ m a x y ^ m i n
Goodness-of-fitThe goodness-of-fit of the fitted model measured by R2 R 2 = 1 i ( y i y ^ i ) 2 i ( y i y ¯ i ) 2
Model fitting periodThe length in years of the stable historical data period, corresponding to the time window to which the BFAST model is fitted h i s t o r y   l e n g t h = y e a r s t a r t M o n i t o r i n g y e a r l a s t H i s t o r i c a l B r e a k
Data qualityThe number of valid observations over the total observations in the stable historical data period d a t a   q u a l i t y = n N
Table 2. TP (true positive), FP (false positive), FN (false negative), and TN (true negative) in an error matrix for a binary classification of disturbance and non-disturbance. The ground data are located in the columns, and the predicted data are located in the rows.
Table 2. TP (true positive), FP (false positive), FN (false negative), and TN (true negative) in an error matrix for a binary classification of disturbance and non-disturbance. The ground data are located in the columns, and the predicted data are located in the rows.
DisturbanceNon-Disturbance
DisturbanceTPFP
Non-disturbanceFNTN
Table 3. Predictive variables in the best models selected by criteria of the highest accuracy and fewer predictive variables. The order of the variables indicates their importance in the model. Finally, the accuracy achieved on the validation set is shown.
Table 3. Predictive variables in the best models selected by criteria of the highest accuracy and fewer predictive variables. The order of the variables indicates their importance in the model. Finally, the accuracy achieved on the validation set is shown.
Three Forest StratumRandom Forests (RF)Support Vector Machine (SVM)Accuracy (%; RF vs. SVM)
All-forestMagnitude, trend, model fitting period Magnitude, trend, model fitting period 75.71 vs. 80.00
TDFMagnitude, trend, amplitudeMagnitude, trend, model fitting period, data quality80.49 vs. 75.61
TFMagnitude, data qualityMagnitude, model fitting period 86.21 vs. 79.31
Table 4. Evaluation of the baseline model (with magnitude threshold = | 0.2 | ) and the best models using RF and SVM on the validation dataset.
Table 4. Evaluation of the baseline model (with magnitude threshold = | 0.2 | ) and the best models using RF and SVM on the validation dataset.
DatasetEvaluation MetricRF & SVM (%)RF (%)SVM (%)ΔRF
vs.
SVM (%)
BaselineBestΔ Best vs. Baseline BestΔBest vs. Baseline
All-ForestAccuracy68.5775.717.1480.0011.43−4.29
F1-score65.6366.671.0474.078.45−7.40
ROC AUC69.6481.1611.5290.1420.49−8.98
Tropical Dry Forest (TDF)Accuracy58.5480.4921.9575.6117.074.88
F1-score51.4369.2317.8054.553.1214.68
ROC AUC59.9288.3628.4479.6319.718.73
Temperate Forest (TF)Accuracy79.3186.216.9079.310.006.90
F1-score81.2586.675.4278.572.688.10
ROC AUC79.7690.2410.4891.9012.14−1.66
Table 5. Confusion matrix for the accuracy assessment of the classification by the SVM all-forest model for the complete study area; the area-weighted accuracy indices including the producer’s, user’s, and overall accuracy, as well as area estimates for each class are shown. The actual disturbance and non-disturbance are located in the columns, and the predicted disturbance and non-disturbance are located in the rows.
Table 5. Confusion matrix for the accuracy assessment of the classification by the SVM all-forest model for the complete study area; the area-weighted accuracy indices including the producer’s, user’s, and overall accuracy, as well as area estimates for each class are shown. The actual disturbance and non-disturbance are located in the columns, and the predicted disturbance and non-disturbance are located in the rows.
Non-DisturbanceDisturbanceUser’s Accuracy (%)
Non-disturbance5061497.31
Disturbance188682.69
Producer’s accuracy (%)96.5686.00
Overall accuracy (%)94.87
Area and weighted area estimates
Area proportion (%)99.480.52
Map area (ha)175,436.37912.06
Unbiased area estimate (ha)170,870.945477.49
Standard Error (SE)1246.901246.90
Weighted producer’s accuracy (%)99.9113.77
Weighted overall accuracy (%)97.23
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Solórzano, J.V.; Gao, Y. Forest Disturbance Detection with Seasonal and Trend Model Components and Machine Learning Algorithms. Remote Sens. 2022, 14, 803. https://doi.org/10.3390/rs14030803

AMA Style

Solórzano JV, Gao Y. Forest Disturbance Detection with Seasonal and Trend Model Components and Machine Learning Algorithms. Remote Sensing. 2022; 14(3):803. https://doi.org/10.3390/rs14030803

Chicago/Turabian Style

Solórzano, Jonathan V., and Yan Gao. 2022. "Forest Disturbance Detection with Seasonal and Trend Model Components and Machine Learning Algorithms" Remote Sensing 14, no. 3: 803. https://doi.org/10.3390/rs14030803

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop