1. Introduction
Forests play a vital role in ecosystem goods and services to humanity, by providing energy, shelter, and livelihoods [
1]. Human-induced land use and management practices, such as deforestation for agriculture, logging, plantation, or transitional subsistence farming, such as shifting cultivation, have led to forest cover loss [
2]. Reliable information on forest cover and its changes is crucial for policymakers to design effective plans in forest conservation.
Forest disturbances have been detected and quantified using multitemporal spaceborne optical remote sensing, such as Landsat, which has been providing images for over three decades and is being used for characterizing forest extent and change [
3,
4,
5]. The normalized difference vegetation index (NDVI), computed using near infrared and red channel [
6], is related to vegetation photosynthetic activity and measures vegetation greenness and productivity [
7,
8,
9]. It has the advantage of being robust and easily interpretable [
10] and is sensitive to forest change when used in a time series (TS) context [
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21]. A change in NDVI signals in TS can be related to change in green vegetation biomass, cover, and structure [
22,
23].
Both annual or biennial time-steps [
12,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33] and dense intra-annual temporal frequency [
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44] have been used, with the former minimizing vegetation seasonality and seeking deviations from stable conditions, and the latter integrating seasonality and trend analysis to describe complete TS trajectories [
38]. It was found that a seasonal trend model shows better performance than methods that remove the seasonal cycle of the TS [
45].
Trend and seasonal model estimation, such as the Breaks For Additive Seasonal Trend (BFAST), allows the detection of changes in the long-term NDVI trend and in the phenological cycle [
46]. It decomposes the TS into trend, seasonal, and residual components. Major shifts in trends are identified as breakpoints by statistical tests, and the differences between the observed and modeled NDVI correspond to the magnitudes of breakpoints. Since not all the interannual variability detected by breakpoints represents disturbance, different approaches [
16,
34,
38,
44,
47] have been applied to identify a suitable magnitude for disturbance detection, including a simple empirical threshold value [
22,
48], a more sophisticated random forest machine learning algorithm [
49,
50,
51], or a hybrid of these two approaches combined with field-based TS fusion [
50]. It was found that by adding seasonality and trend information, especially magnitude, amplitude, and slope, the BFAST performance improved compared to using breakpoints alone [
38].
Goodness-of-fit is one of the tools to assess the ability of a model to simulate reality [
52]. Often measured by coefficient of determination (
R2), a higher
R2 indicates a higher degree of collinearity between the observed and model-simulated variables [
52]. To ensure a good performance of a model,
R2 is often tested as a threshold to accept or reject a fitted model [
53]. Other statistical measures, such as
F-statistic [
54,
55] and chi-square statistics [
56], have also been adopted for the same purpose. To the knowledge of the authors, no previous studies have been found to associate the goodness-of-fit with the performance of a model for change detection.
Amplitude, as one of the seasonal components in the trend analysis, is related to vegetation phenology change detection [
46]. The trend and seasonal model are capable of phenological change detection with data of different ranges of low, medium, and high amplitude [
45]. However, the signal-to-noise ratio in TS affects phenological change detection; changes can be detected in a large range of land covers, including grassland and forests with different seasonal amplitudes, provided that their seasonal amplitude is larger than the noise level [
46]. A minimum seasonal amplitude is required, for example, to use MODIS NDVI TS; a seasonal amplitude of 0.1 NDVI is necessary to detect phenological change [
46].
In addition, the performance of a trend and seasonal model is affected by the amount of interannual variability; for example, spectral changes caused by variability in illumination, seasonality, or atmospheric scattering in the NDVI TS other than disturbances [
34,
45]. A stable historical period is a time period when the components of the NDVI TS remain stable and without any disturbance [
57]. It has been recommended that the stable historical period length should be at least two years, so that the BFAST model is capable of properly fitting its parameters [
34]. Neither amplitude nor stable historical period has been assessed in previous studies for their effects on forest disturbance detection.
Given the importance of the trend and the seasonal model components in TS model prediction, and the fact that little previous research has been carried out on how these components affect forest disturbance detection in different forest types, we report on a study that focuses on disturbance detection in both tropical dry forests (TDF) and temperate forests (TF), encompassing four research questions:
- (1)
Do TDF and TF differ in accuracy in forest disturbance detection by BFAST trend and seasonal model?
- (2)
Is the difference in accuracy related to the BFAST components, such as the magnitude, goodness-of-fit, amplitude, and length of the stable historical period?
- (3)
Within the same type of forest, do the BFAST trend and seasonal model components affect the correct detection of forest disturbance?
- (4)
How does the percentage of pixels with no data values in a TS, caused by clouds and shadow removals in those pixels, affect forest disturbance detection?
We aim to answer these questions by exploring the potential of Landsat NDVI TS from 1994 to 2018 with the BFAST model for forest disturbance detection in both TF and TDF. These two types of forest in the Ayuquila River Basin have been subject to severe disturbances, with agriculture conversion and shifting cultivation in the TDF and selective logging in the TF. We compare the disturbance detection in these two types of forest by examining components in their trend and seasonal model, such as goodness-of-fit, magnitude, amplitude, length of stable period, and percentage of no data value in the TS, and associate these factors to the difference in disturbance detection.
4. Discussion
4.1. Forest Disturbances in TDF and TF
The disturbance detection in TF obtained a higher accuracy than in TDF. This is in line with the finding from [
38] showing that the trend and seasonal model such as BFAST tends to yield higher accuracy in forests with less seasonality, such as those dominated by conifers, because this facilitates the discrimination between phenological changes and disturbances. Indeed, in our study area, the TDF has a very pronounced seasonality controlled mainly by humidity, with a rapid response to the onset of the rainy season, reflected in the drastic changes in NDVI. The variation in the precipitation from year to year might cause BFAST flags breakpoints that are triggered by interannual variations in the TDF phenology rather than disturbances [
38].
Additionally, in ARB, disturbances in TF are caused mainly by selective logging, forest fires, and recently by permanent conversion to avocado plantations. These activities happen on a larger scale and cause more severe damage to the forest than the subtle changes, such as those caused by shifting cultivation, which is the main type of forest disturbance in TDF. In our experiment, most detected disturbances had led to complete forest clearing instead of vegetation thinning.
4.2. Logistic Regression with All Forests
The logistic regression for all forests shows that amplitude and stable historical period length are significant variables with positive coefficient. Although the odds ratio of amplitude (342.09) is much higher than that of stable historical period length (1.17), since amplitude has a small value range (i.e., 0–1), the effect of its high odds ratio is balanced out. In fact, stable historical period length has higher significance value (p = 0.003) in the regression than amplitude (p = 0.029).
Since amplitude often acts as an indicator of forest type: higher amplitude usually corresponds to more seasonal forest such as TDF (see
Figure 3). At a first glance, the positive relation between amplitude and correct disturbance detection seems contradictory knowing that TDF has lower accuracy than TF. Nevertheless, a closer look at our data shows two key aspects: (1) the false disturbances in TDF demonstrate a similar value and variation as TF false and correct; and (2) correctly identified disturbances had higher amplitude values, especially in TDF (see
Table 5). Therefore, the significant effect of amplitude corresponds to the effect of TDF. This topic will be further discussed in the next section.
4.3. Forest-Type Specific Logistic Regression
The forest-type specific logistic regression shows that for TDF, the single most important predictor of correct or false detection is amplitude. As typical areas covered by TDF correspond to very pronounced seasonal NDVI patterns (i.e., high amplitude values, see
Figure 3); low amplitude values could be used as an indicator of models that were fitted rather inaccurately, and therefore, are more prone to false detections. We consider that a main reason for this inaccurate fit is related to the presence of artifacts in the TS, which is not reflected in the NA percentage.
In the case of TF, stable historical period length was the most important variable, however, it was not statistically significant with p = 0.057. Since we have a low number of observations in the false disturbance category (n = 7), the effect of stable historical period length on correct/false disturbance identification might be obscured. Although it would be desirable to follow a sampling procedure to increase the number of false disturbances, it is difficult to identify these areas prior to the verification procedure.
4.4. General Considerations
The other two BFAST components, NA percentage and magnitude, did not show as significant variables in the disturbance detection. NA percentage was used as an indicator of the data quality in the entire TS; however, it does not necessarily represent the data quality in the modelled period (covered by the stable historical period length). We believe that a more relevant variable would be an indicator for the data quality in the stable historical period length, that will ultimately affect the BFAST model fit and disturbance detection. Additionally, other data quality indicators, such as those related to the presence of artifacts, might be of greater importance to the fit of the BFAST model and therefore affects the disturbance detection.
Magnitude is the variable that is often used in previous studies for eliminating false disturbances. Our study adopted the magnitude threshold of , which resulted from field verification. The lack of explanatory power of magnitude in the logistic regression is related to the fact that a threshold was already applied, and therefore, magnitude values higher than did not gave more information for discriminating between false and correct disturbances.
Previous studies [
38] showed that integrating components such as magnitude, slope, and amplitude for forest-type specific models improved disturbance detection. Based on our results, we found that, in addition, stable historic period length is an important factor as well in refining the results of forest disturbance detection with BFAST model.
5. Conclusions
This study evaluated the contribution of the components of BFAST model and the percentage of NAs in a time series to forest disturbance detection. Time series NDVI from Landsat spanning 1994–2018 was applied, employing BFAST model in both temperate forest and tropical dry forest and its accuracy was evaluated by 624 random sample points. Afterwards, BFAST model components, including goodness-of-fit (R2), magnitude, amplitude, length of the stable historical period, and the percentage of NA, were extracted from the time series data of the verified disturbances. The main findings are the following:
- (1)
The disturbance detection yielded higher accuracy in the temperate forest than in the tropical dry forest.
- (2)
The difference in accuracy is associated with the BFAST model components, especially the amplitude and stable historical period length.
- (3)
Within the same forest type, the components affect disturbance detection. For TDF, the amplitude was the statistically significant factor, and for TF, the stable historical period length contributed most to the detection, however, with no statistical significance.
- (4)
The percentage of NAs in a time series was not an important factor in disturbance detection.
This study confirmed that BFAST model components, though varying with forest type, play an important role in forest disturbance detection. To explore the full potential of these methods to map and quantify forest disturbances, future studies should evaluate different BFAST model components in combination with techniques that help reduce the signal-to-noise ratio. For example, variables that represent data quality in the stable historical period might be more relevant as a data quality indicator in the BFAST model fitting, and therefore, help distinguish between correct and false detected disturbances.