Next Article in Journal
Joint Flying Relay Location and Routing Optimization for 6G UAV–IoT Networks: A Graph Neural Network-Based Approach
Next Article in Special Issue
Enhanced Multi-Stream Remote Sensing Spatiotemporal Fusion Network Based on Transformer and Dilated Convolution
Previous Article in Journal
Maneuvering Extended Object Tracking with Modified Star-Convex Random Hypersurface Model Based on Minimum Cosine Distance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Retrieval of Live Fuel Moisture Content Based on Multi-Source Remote Sensing Data and Ensemble Deep Learning Model

1
School of Technology, Beijing Forestry University, Beijing 100083, China
2
Research Center for Biodiversity Intelligent Monitoring, Beijing Forestry University, Beijing 100083, China
3
The College of Forestry, Beijing Forestry University, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(17), 4378; https://doi.org/10.3390/rs14174378
Submission received: 22 August 2022 / Revised: 31 August 2022 / Accepted: 1 September 2022 / Published: 3 September 2022

Abstract

:
Live fuel moisture content (LFMC) is an important index used to evaluate the wildfire risk and fire spread rate. In order to further improve the retrieval accuracy, two ensemble models combining deep learning models were proposed. One is a stacking ensemble model based on LSTM, TCN and LSTM-TCN models, and the other is an Adaboost ensemble model based on the LSTM-TCN model. Measured LFMC data, MODIS, Landsat-8, Sentinel-1 remote sensing data and auxiliary data such as canopy height and land cover of the forest-fire-prone areas in the Western United States, were selected for our study, and the retrieval results of different models with different groups of remote sensing data were compared. The results show that using multi-source data can integrate the advantages of different types of remote sensing data, resulting in higher accuracy of LFMC retrieval than that of single-source remote sensing data. The ensemble models can better extract the nonlinear relationship between LFMC and remote sensing data, and the stacking ensemble model with all the MODIS, Landsat-8 and Sentinel-1 remote sensing data achieved the best LFMC retrieval results, with R 2   = 0.85, RMSE = 18.88 and ubRMSE = 17.99. The proposed stacking ensemble model is more suitable for LFMC retrieval than the existing method.

Graphical Abstract

1. Introduction

Live fuel moisture content (LFMC) is the ratio of vegetation water content to its dry weight [1]. The research shows that there is a clear correlation between the probability of fire and LFMC [2,3], which is an important index affecting the occurrence probability and propagation rate of forest wildfire. To put it another way, accurate and dynamic retrieval of LFMC is extremely valuable to realize the fire risk assessment and spatial modeling of fire behavior [4]. Remote sensing satellite can provide large-scale, multi-band and near-real-time image data, which makes remote sensing technology one of the main methods to estimate LFMC on a large scale [5]. The method of estimating LFMC based on optical remote sensing data is the most widely studied [6,7,8]. MODIS optical remote sensing data are commonly used in the early stage. Myoung et al. [9] developed an empirical model function of LFMC using an aqua-enhanced vegetation index based on MODIS satellite data for wildfire risk assessment in Southern California. Carmine et al. [10] developed a new spectral index, the perpendicular moisture index (PMI), which is sensitive to LFMC based on MODIS satellite data. The experimental results show that PMI had a linear relationship with LFMC, and the highest R2 was 0.87. Landsat-8 can provide higher spatial resolution than MODIS, which has been introduced to estimate LFMC in recent years. Considering the complexity of upper tree canopy and lower grass canopy, Quan et al. [11] predicted the forest FMC of a two-layer canopy structure in Southwest China by coupling a radiative transfer model and a Landsat-8 product. Mbulisi et al. [5] used Landsat-5 and Landsat-8 data to quantitatively retrieve vegetation canopy FMC in six study areas based on PROSAIL and PROGeoSAIL radiative transfer models. These methods based on optical remote sensing depend on the absorption characteristics of leaf water at near-infrared (NIR) or short-wave infrared (SWIR) wavelengths [12]. Optical and infrared reflectance are highly sensitive to vegetation characteristics such as canopy structure [13] and leaf area index [14,15], and so these models are often only applicable for very specific sites, and the generalization ability of different regions is limited [16,17].
The wavelength of microwave remote sensing is longer than that of optical remote sensing by four orders of magnitude. Microwaves can penetrate the clouds and enter the vegetation canopy, which enables microwave remote sensing to acquire the dynamic changes in vegetation moisture better than optical remote sensing [18,19,20]. In recent years, the prediction ability of active microwave remote sensing technology represented by synthetic aperture radar (SAR) for fire-related variables has been verified [21,22]. Wang et al. [20] coupled the soil backscatter linear model with the vegetation backscatter water cloud model, achieving forest FMC retrieval based on Sentinel-1 SAR data and a better performance than that obtained using Landsat-8 data and empirical methods.
Different remote sensing data have different sensitivities to vegetation water and biomass, and the effect of single-source remote sensing data retrieval of LFMC is limited. Using multi-source remote sensing data to estimate LFMC can avoid the limitations of single-source remote sensing data and provide more comprehensive data for extracting the parameters required for LFMC retrieval [23]. Deep learning can approach the complex nonlinear relationship between various biological, geophysical parameters and remote sensing data through multi-layer learning [24,25], which provides a data-driven alternative for large-scale LFMC retrieval. Rao et al. [19] performed LFMC retrieval based on a long short-term memory (LSTM) network with fused data, i.e., Landsat-8 data, SAR data, terrain, slope and other auxiliary variables. The retrieval of fused data achieved R2 = 0.63, RMSE = 25%, which is better than that of single-source remote sensing data (R2 = 0.44, RMSE = 31.8%). Zhu et al. [26] proposed the LFMC retrieval architecture TempCNN-LFMC based on temporal convolutional networks (TCNs). With MODIS, altitude, slope and other auxiliary data as the input fused data, the retrieval achieved R2 = 0.64, RMSE = 22.74%. The above research shows that the fused data are helpful in improving the performance of LFMC retrieval.
A single model cannot completely extract the features of remote sensing variables in LFMC. To improve the accuracy of LFMC retrieval, it is worthwhile to combine multiple models to extract the features of multi-source remote sensing in time and space dimensions at the same time [25]. Therefore, based on deep learning and ensemble learning methods, this study discusses the LFMC retrieval performance using multi-source remote sensing data. The contents of this study include the following aspects:
(1)
We explore the advantages of LFMC retrieval utilizing multi-source remote sensing data obtained from combing MODIS, Landsat-8, Sentinel-1 and auxiliary data such as canopy height and land cover as data sources, which can provide more comprehensive data and avoid the limitations of single-source remote sensing data.
(2)
We propose a LFMC retrieval model integrating the LSTM and TCN, which exploits the long-time memory capability of LSTM and the superior feature extraction capability of TCN, and finally performs better than LSTM and the TCN alone.
(3)
Based on LSTM, TCN and TCN-LSTM models, two ensemble models (the stacking and Adaboost ensemble models) are designed, and the advantages of stacking ensemble model are confirmed by comparative experiments.

2. Data and Methods

2.1. Study Area

The Western United States was selected as the study area (shown in Figure 1), where wildfires occurred frequently. This area covers more than 3.7 million square kilometers, containing different climates and terrains. The vegetation types are abundant, including broadleaf deciduous forests, needleleaf evergreen forests, shrublands, grasslands and sparse vegetation areas, which made it an ideal area for studying LFMC prediction methods. Considering the integrity and generality of the data, the selected study period was from 1 January 2015 to 31 December 2018.

2.2. Research Data

2.2.1. LFMC Data

The National Fuel Moisture Database (NFMD) [27] is a web-based query system. There have been over 200,000 actual measurements of fuel moisture data since 1977. The database regularly updates monitored fuel moisture data, covering 976 samples mainly located in the Western United States, each covering an area of 5 acres. The measurements were taken in the mid-afternoon and on dry days with no dew or precipitation. In this paper, 133 representative samples were selected, and the specific location is marked by circular points in Figure 1. During the study period, the value of LFMC varied from 16% to 320%, which covers the common water state of live fuels.

2.2.2. MODIS Data

The MODIS data came from MODIS Terra and Aqua joint observation of the MCD43A4 product [28]. The product was the nadir bidirectional reflectance distribution function (BRDF)-adjusted reflectance (NBAR) data, the spatial resolution of which is 500 m. BRDF was fitted using 16-day Terra and Aqua MODIS data and applied to the original reflectance to obtain NBAR. In this study, Band1–Band7 of NBAR reflectivity data were selected as the model input.
In addition, snow cover will lead to abnormal reflectivity. Thus, the snow pixels need to be deleted. The MODIS snow product (MOD10A2-V6) [29] was used to determine whether there was snow. MOD10A2 is a snow cover product synthesized every eight days from the first day of each year. In MOD10A2, if a pixel is classified as snow on any day of the eight days, the pixel is identified as snow.

2.2.3. Landsat Data

Landsat data came from the 16-day surface reflectance data of Landsat-8 [30], which are Level 1T products with a spatial resolution of 30 m. There is a strong correlation between the normalized difference water index (NDWI) and LFMC [31]. Considering that water mainly absorbs the energy of near-infrared (NIR) and short-wave infrared (SWIR) spectral regions, the original band reflectances of red, green, blue, near-infrared and short-wave infrared channels were selected to directly reflect the change in water [32]. The normalized difference vegetation index (NDVI) is a simple, effective and empirical measurement of surface vegetation, and it is also a key factor affecting the prediction of LFMC [33]. The near-infrared vegetation index (NIRV) is an indicator of vegetation biomass level because it is related to carbon assimilation of photosynthesis, so it may help to separate the effects of biomass and LFMC on Sentinel-1 backscattering [34]. To sum up, three vegetation indexes, NDWI, NDVI and NIRV, and the original band reflections of red, green, blue, near-infrared and short-wave infrared channels were selected from Landsat-8 data.

2.2.4. Sentinel-1 Data

Sentinel-1 is a 5.4 GHz C-band synthetic aperture radar (SAR) with a 12-day revisit cycle in the Western United States. The Sentinel-1 data used in this study were derived from the ground-range detector (GRD) data of Sentinel-1, and the data were collected in the wide-strip mode of interferometric measurement with vertical–vertical (VV) and vertical–horizontal (VH) polarization on land [35]. Since the microwave signal has a longer wavelength, is less sensitive to atmospheric conditions, is not susceptible to cloud pollution and can detect deeper vegetation canopy, the microwave remote sensing data can provide more continuous global observation [36]. At the same time, the absorption and scattering of the microwave signal by the surface (including vegetation and soil) is mainly determined by the microwave backscatter σ [37], and microwave backscatter is mainly affected by the moisture content, so the microwave signal is sensitive to vegetation water content [38]. Therefore, in this paper, σ V V , σ V H and σ V V σ V H were used as the microwave input for the model.

2.2.5. Auxiliary Data

Seven kinds of static auxiliary data were chosen to help the model learn the radiative transfer process between time-varying input and LFMC. The specific data can be divided into the following three categories:
The first category is soil data, including silt, sand and clay content, which was used to control the sensitivity of microwave backscattering to soil moisture, so that the retrieval model could separate vegetation-related information from microwave backscattering. The soil data come from the North American soil map of Liu et al. [39].
Vegetation canopy water content has a certain sensitivity in the near-infrared and short-wave infrared bands [5], and the sensitivity of different vegetations to remote sensing data is also different. The canopy height measured by the Global Laser Altimeter System lidar [40] and the land cover information of 300 m spatial resolution obtained by GLOBCOVER [41] were selected as the second auxiliary data.
The third category is terrain data; considering that the local incidence angle will affect the parameterization of backward scattering on vegetation water [37], it was necessary to use the elevation and slope of the National Elevation Dataset [42] to help the model calibrate the local terrain.
Table 1 summarizes all the inputs used in the model.

2.3. Data Process

Given the presence of numerous vegetation species at certain sample points and the lack of information on the abundance of these species, directly averaging the LFMCs of different species will result in significant inaccuracies. We adopted the same strategy as [19], excluding sampling points with multiple species unless the LFMCs of multiple species were similar during the research period (Pearson r between any two species ≥ 0.5). Thus, 2934 samples from 133 sampling points were included in total (shown in Figure 1).
Because the spatial and temporal resolutions of MODIS, Landsat-8 and Sentinel-1 are not the same, spatial and temporal consistency processing was needed. The remote sensing variable data of the sample points were extracted by the Google Earth Engine (GEE) according to the latitude and longitude coordinates. According to the latitude and longitude coordinates, the spatial synchronization of ground data and remote sensing data could be realized. The remote sensing data were unified to the resolution of 250 m using bilinear interpolation.
The sampling period of live fuel samples in each location was roughly one month, so the time series input was linearly interpolated to the end of each month to ensure that the data had the same time phase. The maximum changes in MODIS, Landsat-8 and Sentinel-1 data were only 2.3%, 6.7% and 3.0%, respectively. It can be considered that the interpolation operation had little effect on the input data.

2.4. Dataset

In this work, three-fold cross-validation was used to test the model. To ensure that the performance of the model was tested on samples that were completely different from the training sample points by separating data by sample points, the data were first stratified randomly sampled into training and test sets by a ratio of 2:1 to ensure that the distribution of land cover types in the training and test sets remained the same. This implies that the training set was made up of data from two-thirds of the locations (89 sample points), while the test set was made up of data from the remaining one-third (44 sample points). In addition, the training set was divided into three folds, two for training and one for validation. Finally, the results presented in the paper were calculated based on the estimated value of the test set.

2.5. LFMC Retrieval Models

2.5.1. TCN-LSTM Model

LSTM can effectively deal with the dynamic dependence of complex long-term time series. The TCN has simple structure and strong feature extraction ability. Combined with the ability of TCN feature extraction and LSTM long-time series memory, the TCN-LSTM network is designed to predict the LFMC. The structure of the TCN-LSTM network is shown in Figure 2.
The retrieval process is as follows:
(1)
Firstly, the LFMC data and selected input variables (x1, …, xn) are fed into the TCN. The features of remote sensing variables and LFMC are extracted through the causal convolution layer contained in the TCN.
(2)
Then, multiple LSTM layers combined with the dropout mechanism are used for prediction, which can prevent over fitting.
(3)
Through the flatten layer, the output matrix is compressed into one dimension to facilitate the connection of the later dense layer.
(4)
The nonlinear relationship is mapped to the output space through the dense layer to achieve the LFMC prediction results.

2.5.2. Stacking Ensemble Model

In order to further improve the performance of LFMC retrieval, a two-layer stacking ensemble model integrating LSTM, TCN and TCN-LSTM was further proposed. The model structure is shown in Figure 3.
The first layer extracts the features from the original split dataset through three basic learners. The basic learners of the stacking model should be “accurate but different”, that is, the prediction accuracy of each basic learner is required to be high, and the types of basic learners should also be diverse. So, LSTM, the TCN and TCN-LSTM were introduced as the base learners. In order to avoid over fitting, a simple linear regression (LR) was selected as the meta-learner of the second layer.

2.5.3. Adaboost Ensemble Model

Unlike stacking ensemble, Adaboost ensemble trains several weak learners based on different training subsets randomly selected from the original training dataset. Adaboost ensemble is based on homogeneous integration, which is composed of the same type of basic learners. In this work, the TCN-LSTM model was selected as the weak learner to construct the Adaboost ensemble model. Figure 4 shows the structure of the Adaboost ensemble model.
In each training process, the initial weights are assigned to the samples at first, and the weights are updated after each iteration. The samples with a high error rate obtain higher weights, which makes the algorithm focus on the samples that are more difficult to learn. The sample weight is adjusted to Dn, and passed to the next weak learner Gn for better prediction. Therefore, the features extracted by G1 are transmitted to G2, and then the features estimated with high error can be corrected in the transmission process, which is helpful to improve the prediction accuracy. At last, the weighted average method is utilized to obtain the strong learner HM, the output of which is the final prediction result. Considering the computational efficiency, the number of TCN-LSTM, that is, the number of iterations t, was set to 3.

2.5.4. Model Settings

There are three basic models, LSTM, the TCN and TCN-LSTM. Table 2 lists the architectures of these models. In the ensemble models, the same architectures were used. All the proposed models estimated LFMC for each month using input variables of three previous months. Although predicting one month averaged LFMC value can be error-prone due to the variations in LFMC, we were constrained by the temporal resolution of the remote sensing data.

3. Experiments and Results

3.1. Experimental Setup

The hardware environment of the experiments was: CPU: Intel (R) Core (TM) i7-8565U, Memory: 8 GB. The software environment was: Windows 10 64 operating system, deep learning framework Tensorflow2.3.0 and python 3.7. Adam optimizer was used, and the parameters were the default values. The batch size was 32, the learning rate was 0.01, and the epoch was 300. In order to avoid over fitting, early stopping based on the loss of the validation set was used [43], and patience was 30.

3.2. Evaluating Indicator

Bias, determination coefficient R 2 , root mean square error (RMSE) and unbiased root mean square error (ubRMSE) between estimated and measured LFMCs were chosen to quantitatively evaluate the performance of the models. When R2 was closer to 1 and the RMSE value was lower, the model accuracy was higher and the model was more accurate. The calculations of RMSE and ubRMSE are shown in Formulas (1) and (2):
R M S E = 1 N i N ( L F M C i , m L F M C i , e ) 2
u b R M S E = 1 N i N ( L F M C i , m L F M C i , e ( L F M C m ¯ L F M C e ¯ ) ) 2
where N is the number of measurements; L F M C i , m   and L F M C i , e   are the ith measured and estimated LFMC, respectively; L F M C m ¯ and L F M C e ¯ are the averages of measured and estimated LFMC, respectively.

3.3. Comparison of Different Deep Learning Models

The performances of three different models, LSTM, the TCN and TCN-LSTM, with different remote sensing data were compared. The retrieval results are shown in Figure 5.
By comparing and analyzing the results in Figure 5 and Table 3, it can be concluded that:
(1)
The bias of all the three models was negative, indicating that all the models underestimated LFMC as a whole. The TCN-LSTM model had the lowest bias among all the models on the same dataset. The bias of Sentinel-1 was the largest, and that of Landsat-8 was the lowest. Although microwave remote sensing (Sentinel-1) is more penetrating due to its high sensitivity to surface moisture, it is difficult to distinguish between vegetation and bare soil backscatter only using microwave remote sensing data, which leads to higher bias. The multi-source remote sensing data fuse the microwave remote sensing and optical remote sensing together, which can be essentially seen as the integration of the microwave backscattering characteristics and optical characteristic. Therefore, the retrieval performances of multi-source remote sensing data were higher than those of the single-source remote sensing data.
(2)
The R 2 , RMSE and ubRMSE of the TCN-LSTM model were also better than those of the LSTM and TCN models. The retrieval accuracy of the TCN-LSTM model with all three kinds of remote sensing data was the highest at R 2 = 0.81, RMSE = 21.73 and ubRMSE = 19.93, which means that TCN-LSTM can incorporate the advantages of LSTM and the TCN and effectively extract the features of multi-source remote sensing.
A comparison with the retrieval results of references is shown in Table 4; the TCN-LSTM model with multi-source remote sensing achieved the best results for LFMC retrieval. Compared with the best results of the existing method [20], R 2   and RMSE were improved by 26.56% and 4.44%, respectively.

3.4. Comparison of Different Ensemble Learning Models

Table 5 shows the performance comparison of different ensemble learning models with different remote sensing data. It can be seen that the performances of two ensemble learning models with all three kinds of remote sensing data were better than with other data. The retrieval results for LFMC based on the stacking ensemble model with MODIS, Landsat-8 and Sentinel-1 are the best. This is mainly due to the integration of the advantages of three different models. While the Adaboost ensemble model only uses one kind of basic learner, its performance was poorer than that of the stacking ensemble model.
Together with Table 3, it was found that with different groups of remote sensing data, the trend in the performances of the single model and the ensemble model was almost the same. The more data used, the better the performance. Additionally, based on multi-source data (M+L+S), compared with the TCN-LSTM model, the R 2 , RMSE and ubRMSE of the stacking ensemble model realized an improvement of 4.9%, 15.1% and 10.8%, respectively.

4. Discussion

4.1. Explanation of Estimated LFMC Value

In general, when using Landsat-8, Sentinel-1 and MODIS for LFMC retrieval, the estimated value is higher than the measured value, and the linear fitting is good when the LFMC is low. With the increase in LFMC, the estimated value is lower than the actual value, the points are discrete, and the overall correlation is high [44]. Figure 6 shows the LFMC retrieval results and measured values based on two ensemble learning models combined with MODIS, Landsat-8 and Sentinel-1. We can see that two proposed ensemble learning models underestimated high LFMC values (>120%), and there was a systematic bias for phenological periods with high LFMC values. This can be partly explained by the limited sensitivity of the optical sensing data to wet vegetation and the tendency of the proposed method to globally optimize the solution at the cost of underestimation at high values. Similar underestimations have been observed in other studies using physical or data-driven methods [45]. However, such underestimation is not significant when considering the cause of the fire hazard or behavior [46]. Experience has shown that when LFMC is high (>120%), the probability of fire occurrence is comparatively low, or fire movement through this area is limited, so this has less of an impact on fire managers, who might use this model to assess LFMC.
The proposed models also overestimated low values (<30%), which may have been due to the presence of dead combustibles, such as grass fuel [47] and leaf litter. Nevertheless, the magnitude of the positive bias was very small (as shown in Figure 6). Moreover, when the LFMC value is lower than 60%, the likelihood of fire occurrence increasing dramatically [48]. So, if the LFMC value is less than 30%, fire managers will be more aggressive with the estimated results. The impact introduced by the minor error on fire managers who may use the model is limited to the extent that this is overestimated in the range (<30%).

4.2. Advantages of Multi-Source Remote Sensing Data and Ensemble Learning

Due to the different shortcomings of different remote sensing data, as expected, the LFMC retrieval results with all the Landsat-8, Sentinel-1 and MODIS remote sensing data are much higher than those of other data when using the same model, which can be attributed to the fact that multi-source remote sensing data can reduce the uncertainty of single-source data and provide more valuable features derived from the complementarity of different data.
Furthermore, the ensemble learning method comprises several basic learners together to obtain better performance. The Adaboost ensemble model is a sequential ensemble technique, in which the final prediction is based on the weighted average results of three weak learners (TCN-LSTM) trained on different training subsets, while the stacking ensemble model combines three parallel basic learners (LSTM, TCN and TCN-LSTM) in the first layer to extract abundant features, and then concatenates straightforward logistic regression as the second learner. Three different basic learners combined with sequential concatenation operation produced better features and an improved retrieval performance over the Adaboost ensemble model.
To summarize, the combination of multi-source data fusion and ensemble learning can significantly improve retrieval performance and provide considerable potential for accurate LFMC estimation.

4.3. Limitations of the Proposed Method with Processed Data

As we all know, the estimation of LFMC using remote sensing data (such as optical and microwave data) has the same issue since remote sensing data are dependent not only on LFMC but also on other bio- and geophysical characteristics [26]. Previous studies tried to find the empirical relationships or physical models between LFMC and other factors. Despite the satisfactory results of these methods, carefully handcrafted input variables chosen based on our understanding of radiative transfer processes must be selected; additionally, corresponding field data are needed, making these models challenging to generalize and operationalize on large-scale sites.
In our work, we introduced deep learning to capture the complicated nonlinear relationship among the LFMC and the remote sensing data, hoping to avoid the selection of carefully handcrafted input variables and the collection of corresponding field data, making it easier to realize large-scale LFMC estimation. The results demonstrate that this method performs admirably in large-scale sites (133 sampling points) with diverse vegetations, while during the data processing, considering the time resolution of remote sensing data and the frequency of measured LFMC, we interpolated the data to the end of each month, which means the time resolution was one month, resulting in the misrepresentation of daily or weekly fluctuations in LFMC. However, this limitation would be solved by gathering data with a smaller resolution.
Furthermore, we are all aware that the kind of vegetation has a direct impact on various remote sensing data [19,26]. Here, we simply delegated the task of classifying vegetation types implicitly to the deep learning model. Figure 7 presents the RMSEs of LFMC retrieval results of different vegetations, demonstrating that the RMSEs of four single-vegetations are often lower than that of mixed vegetations. In particular, the worst predictions were made for mixed shrub–grassland. This suggests that our previous strategy of selecting sample points makes sense, and that using the selected samples to train the model is beneficial in improving the accuracy of the predictions. Nevertheless, a fully data-driven model would very likely result in mistakes if detailed vegetation distribution data were not included. In practical application, a feasible option is to collect more data and then create more advanced models.
Finally, while it is widely acknowledged that deep learning is a data-driven nonlinear model with high automated learning and generalization capabilities that have the potential to be applied to other regions, the efficacy of its application in other locations requires more data for validation.

5. Conclusions

In this study, a TCN-LSTM model was firstly designed to improve the effect of feature extraction, and further, two ensemble models were proposed based on the TCN-LSTM model to achieve more accurate retrieval of LFMC. Considering the different shortcomings of separate Landsat-8, Sentinel-1 and MODIS remote sensing data, all the three data were utilized together to obtain higher performance. The results of the experiments on the LFMC data from the Western United States show that the stacking ensemble model with all three remote sensing data achieved the best performance. The proposed stacking ensemble model was trained on historical data, which can automatically extract the nonlinear correlation between remote sensing data and LFMC. This enabled the proposed model’s good generalization ability. Our model is data-driven, which means it has the potential to realize significant accuracy in LFMC estimation for other locations with appropriate training data. Meanwhile, our results reveal that our proposed models had a mixture of predictions with low and high amounts of bias. We believe this is because different vegetations are not explicitly considered. We will study and improve our model on more available data in the future.

Author Contributions

T.Q. analyzed the data and wrote the manuscript. W.H. checked and revised the manuscript. J.X., J.Z., H.H. and B.C. provided guidance for experiments and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was jointly supported by the National Key Research and Development Program of China (No. 2020YFC1511601), the Beijing Municipal Natural Science Foundation (No. 6214040), the Fundamental Research Funds for the Central Universities (No. 2021ZY70) and China Scholarship Council (202106515010).

Data Availability Statement

Not applicable.

Acknowledgments

We would like to express our gratitude to the National Key Research and Development Program of China (no. 2020YFC1511601), the Beijing Municipal Natural Science Foundation (no. 6214040), the Fundamental Research Funds for the Central Universities (no. 2021ZY70) and China Scholarship Council (202106515010).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cunill Camprubí, À.; González-Moreno, P.; Resco de Dios, V. Live fuel moisture content mapping in the Mediterranean Basin using random forests and combining MODIS spectral and thermal data. Remote Sens. 2022, 14, 3162. [Google Scholar] [CrossRef]
  2. Chuvieco, E.; Gonzalez, I.; Verdu, F.; Aguado, I.; Yebra, M. Prediction of fire occurrence from live fuel moisture content measurements in a Mediterranean ecosystem. Int. J. Wildland Fire 2009, 18, 430–441. [Google Scholar] [CrossRef]
  3. Nolan, R.H.; Boer, M.M.; de Dios, V.R.; Caccamo, G.; Bradstock, R.A. Large-scale, dynamic transformations in fuel moisture drive wildfire activity across southeastern Australia. Geophys. Res. Lett. 2016, 43, 4229–4238. [Google Scholar] [CrossRef]
  4. Chuvieco, E.; Aguado, I.; Jurdao, S.; Pettinari, M.L.; Yebra, M.; Salas, J.; Hantson, S.; de la Riva, J.; Ibarra, P.; Rodrigues, M.; et al. Integrating geospatial information into fire risk assessment. Int. J. Wildland Fire 2014, 23, 606–619. [Google Scholar] [CrossRef]
  5. Sibanda, M.; Onisimo, M.; Dube, T.; Mabhaudhi, T. Quantitative assessment of grassland foliar moisture parameters as an inference on rangeland condition in the mesic rangelands of southern Africa. Int. J. Remote Sens. 2021, 42, 1474–1491. [Google Scholar] [CrossRef]
  6. Yebra, M.; Dennison, P.E.; Chuvieco, E.; Riaño, D.; Zylstra, P.; Hunt, E.R., Jr.; Danson, F.M.; Qi, Y.; Jurdao, S. A global review of remote sensing of live fuel moisture content for fire danger assessment: Moving towards operational products. Remote Sens. Environ. 2013, 136, 455–468. [Google Scholar] [CrossRef]
  7. Garcia, M.; Chuvieco, E.; Nieto, H.; Aguado, I. Combining AVHRR and meteorological data for estimating live fuel moisture content. Remote Sens. Environ. 2008, 112, 3618–3627. [Google Scholar] [CrossRef]
  8. García, M.; Riaño, D.; Yebra, M.; Salas, J.; Cardil, A.; Monedero, S.; Ramirez, J.; Martín, M.P.; Vilar, L.; Gajardo, J.; et al. A live fuel moisture content product from Landsat TM satellite time series for implementation in fire behavior models. Remote Sens. 2020, 12, 1714. [Google Scholar] [CrossRef]
  9. Myoung, B.; Kim, S.H.; Nghiem, S.V.; Jia, S.; Whitney, K.; Kafatos, M.C. Estimating live fuel moisture from MODIS satellite data for wildfire danger assessment in Southern California USA. Remote Sens. 2018, 10, 87. [Google Scholar] [CrossRef]
  10. Maffei, C.; Menenti, M. A MODIS-based perpendicular moisture index to retrieve leaf moisture content of forest canopies. Int. J. Remote Sens. 2014, 35, 1829–1845. [Google Scholar] [CrossRef]
  11. Quan, X.; He, B.; Yebra, M.; Yin, C.; Liao, Z.; Li, X. Retrieval of forest fuel moisture content using a coupled radiative transfer model. Environ. Model. Softw. 2017, 95, 290–302. [Google Scholar] [CrossRef]
  12. Yebra, M.; Van Dijk, A.; Leuning, R.; Huete, A.; Guerschman, J.P. Evaluation of optical remote sensing to estimate actual evapotranspiration and canopy conductance. Remote Sens. Environ. 2013, 129, 250–261. [Google Scholar] [CrossRef]
  13. Song, C.H. Optical remote sensing of forest leaf area index and biomass. Prog. Phys. Geogr. -Earth Environ. 2013, 37, 98–113. [Google Scholar] [CrossRef]
  14. Quan, X.; He, B.; Li, X.; Liao, Z. Retrieval of Grassland live fuel moisture content by parameterizing radiative transfer model with interval estimated LAI. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 910–920. [Google Scholar] [CrossRef]
  15. Xiao, Z.; Song, J.; Yang, H.; Sun, R.; Li, J. A 250 m resolution global leaf area index product derived from MODIS surface reflectance data. Int. J. Remote Sens. 2022, 43, 1409–1429. [Google Scholar] [CrossRef]
  16. Al-Moustafa, T.; Armitage, R.P.; Danson, F.M. Mapping fuel moisture content in upland vegetation using airborne hyperspectral imagery. Remote Sens. Environ. 2012, 127, 74–83. [Google Scholar] [CrossRef]
  17. Houborg, R.; Anderson, M.; Daughtry, C. Utility of an image-based canopy reflectance modeling tool for remote estimation of LAI and leaf chlorophyll content at the field scale. Remote Sens. Environ. 2009, 113, 259–274. [Google Scholar] [CrossRef]
  18. Fan, L.; Wigneron, J.P.; Xiao, Q.; Al-Yaari, A.; Wen, J.; Martin-StPaul, N.; Dupuy, J.-L.; Pimont, F.; Al Bitar, A.; Fernandez-Moran, R.; et al. Evaluation of microwave remote sensing for monitoring live fuel moisture content in the Mediterranean region. Remote Sens. Environ. 2018, 205, 210–223. [Google Scholar] [CrossRef]
  19. Rao, K.; Williams, A.P.; Flefil, J.F.; Konings, A.G. SAR-enhanced mapping of live fuel moisture content. Remote Sens. Environ. 2020, 245, 111797. [Google Scholar] [CrossRef]
  20. Wang, L.; Quan, X.W.; He, B.B.; Yebra, M.; Xing, M.; Liu, X. Assessment of the dual polarimetric Sentinel-1A data for forest fuel moisture content estimation. Remote Sens. 2019, 11, 1568. [Google Scholar] [CrossRef] [Green Version]
  21. Bai, X.; He, B.; Li, X.; Zeng, J.; Wang, X.; Wang, Z.; Zeng, Y.; Su, Z. First assessment of Sentinel-1A data for surface soil moisture estimations using a coupled water cloud model and advanced integral equation model over the Tibetan Plateau. Remote Sens. 2017, 9, 714. [Google Scholar] [CrossRef]
  22. Bai, X.; He, B. Potential of Dubois model for soil moisture retrieval in prairie areas using SAR and optical data. Int. J. Remote Sens. 2015, 36, 5737–5753. [Google Scholar] [CrossRef]
  23. Jiao, W.; Wang, L.; McCabe, M.F. Multi-sensor remote sensing for drought characterization: Current status, opportunities and a roadmap for the future. Remote Sens. Environ. 2021, 256, 112313. [Google Scholar] [CrossRef]
  24. Reichstein, M.; Camps-Valls, G.; Stevens, B.; Jung, M.; Denzler, J.; Carvalhais, N.; Prabhat. Deep learning and process understanding for data-driven Earth system science. Nature 2019, 566, 195–204. [Google Scholar] [CrossRef]
  25. Yuan, Q.Q.; Shen, H.F.; Li, T.W.; Li, Z.; Li, S.; Jiang, Y.; Xu, H.; Tan, W.; Yang, Q.; Wang, J.; et al. Deep learning in environmental remote sensing: Achievements and challenges. Remote Sens. Environ. 2020, 241, 111716. [Google Scholar] [CrossRef]
  26. Zhu, L.; Webb, G.I.; Yebra, M.; Scortechini, G.; Miller, L.; Petitjean, F. Live fuel moisture content estimation from MODIS: A deep learning approach. ISPRS J. Photogramm. Remote Sens. 2021, 179, 81–91. [Google Scholar] [CrossRef]
  27. United States Forest Sevices. National Fuel Moisture Database; United States Forest Services: Washington, DC, USA, 2010.
  28. Nietupski, T.C.; Kennedy, R.E.; Temesgen, H.; Kerns, B.K. Spatiotemporal image fusion in Google Earth Engine for annual estimates of land surface phenology in a heterogenous landscape. Int. J. Appl. Earth Obs. Geoinf. 2021, 99, 102323. [Google Scholar] [CrossRef]
  29. Faiz, M.A.; Liu, D.; Tahir, A.A.; Li, H.; Fu, Q.; Adnan, M.; Zhang, L.; Naz, F. Comprehensive evaluation of 0.25° precipitation datasets combined with MOD10A2 snow cover data in the ice-dominated river basins of Pakistan. Atmos. Res. 2020, 231, 104653. [Google Scholar] [CrossRef]
  30. Vermote, E.; Justice, C.; Claverie, M.; Franch, B. Preliminary analysis of the performance of the Landsat 8/OLI land surface reflectance product. Remote Sens. Environ. 2016, 185, 46–56. [Google Scholar] [CrossRef]
  31. Roberts, D.A.; Dennison, P.E.; Peterson, S.; Sweeney, S.; Rechel, J. Evaluation of airborne visible/infrared imaging spectrometer (AVIRIS) and moderate resolution imaging spectrometer (MODIS) measures of live fuel moisture and fuel condition in a shrubland ecosystem in southern California. J. Geophys. Res. -Biogeosci. 2006, 111, 1–16. [Google Scholar] [CrossRef] [Green Version]
  32. Colombo, R.; Merom, M.; Marchesi, A.; Busetto, L.; Rossini, M.; Giardino, C.; Panigada, C. Estimation of leaf and canopy water content in poplar plantations by means of hyperspectral indices and inverse modeling. Remote Sens. Environ. 2008, 112, 1820–1834. [Google Scholar] [CrossRef]
  33. Chuvieco, E.; Aguado, I.; Dimitrakopoulos, A.P. Conversion of fuel moisture content values to ignition potential for integrated fire danger assessment. Can. J. For. Res. 2004, 34, 2284–2293. [Google Scholar] [CrossRef]
  34. Badgley, G.; Field, C.B.; Berry, J.A. Canopy near-infrared reflectance and terrestrial photosynthesis. Sci. Adv. 2017, 3, e1602244. [Google Scholar] [CrossRef]
  35. Torres, R.; Snoeij, P.; Geudtner, D.; Bibby, D.; Davidson, M.; Attema, E.; Potin, P.; Rommen, B.; Floury, N.; Brown, M.; et al. GMES Sentinel-1 mission. Remote Sens. Environ. 2012, 120, 9–24. [Google Scholar] [CrossRef]
  36. Keane, R.E. Fuel concepts. In Wildland Fuel Fundamentals and Applications; Springer: Cham, Switzerland, 2015; pp. 175–184. [Google Scholar]
  37. Brocca, L.; Crow, W.T.; Ciabatta, L.; Massari, C.; de Rosnay, P.; Enenkel, M.; Hahn, S.; Amarnath, G.; Camici, S.; Tarpanelli, A.; et al. A review of the applications of ASCAT soil moisture products. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 2285–2306. [Google Scholar] [CrossRef]
  38. Datta, S.; Das, P.; Dutta, D.; Giri, R.K. Estimation of surface moisture content using Sentinel-1 C-band SAR data through machine learning models. J. Indian Soc. Remote Sens. 2020, 49, 887–896. [Google Scholar] [CrossRef]
  39. Liu, S.; Wei, Y.; Post, W.M.; Cook, R.B.; Schaefer, K.; Thornton, M.M. NACP MsTMIP: Unified North American Soil Map; ORNL Distributed Active Archive Center: Oak Ridge, TN, USA; p. 2014.
  40. Simard, M.; Pinto, N.; Fisher, J.B.; Baccini, A. Mapping forest canopy height globally with spaceborne lidar. J. Geophys. Res. -Biogeosci. 2011, 116, G04021. [Google Scholar] [CrossRef]
  41. Arino, O.; Ramos, J.; Kalogirou, V.; Bontemps, S.; Defourny, P.; Van Bogaert, E. Global Land Cover Map for 2009 (GlobCover 2009); European Space Agency (ESA): Paris, France; Université catholique de Louvain (UCL), PANGAEA: Ottignies-Louvain-la-Neuve, Belgium. [CrossRef]
  42. USGS-NED. National Elevation Dataset; US Geological Survey: Lafayette, LA, USA, 2004.
  43. Pelletier, C.; Webb, G.I.; Petitjean, F. Temporal convolutional neural network for the classification of satellite image time series. Remote Sens. 2019, 11, 523. [Google Scholar] [CrossRef]
  44. Jia, S.; Kim, S.H.; Nghiem, S.V.; Kafatos, M. Estimating live fuel moisture using SMAP L-band radiometer soil moisture for Southern California, USA. Remote Sens. 2019, 11, 1575. [Google Scholar] [CrossRef] [Green Version]
  45. Yebra, M.; Quan, X.; Riano, D.; Larraondo, P.R.; van Dijk, A.I.J.M.; Cary, G.J. A fuel moisture content and flammability monitoring methodology for continental Australia based on optical remote sensing. Remote Sens. Environ. Interdiscip. J. 2018, 212, 260–272. [Google Scholar] [CrossRef]
  46. Pimont, F.; Ruffault, J.; Martin-StPaul, N.K.; Dupuy, J.-L. Why is the effect of live fuel moisture content on fire rate of spread underestimated in field experiments in shrublands? Int. J. Wildland Fire 2019, 28, 127–137. [Google Scholar] [CrossRef]
  47. Yebra, M.; Scortechini, G.; Badi, A.; Beget, M.E.; Boer, M.M.; Bradstock, R.; Chuvieco, E.; Danson, F.M.; Dennison, P.; de Dios, V.R.; et al. Globe-LFMC, a global plant water status database for vegetation ecophysiology and wildfire applications. Sci. Data 2019, 6, e155. [Google Scholar] [CrossRef] [PubMed]
  48. Jurdao, S.; Chuvieco, E.; Arevalillo, J.M. Modelling fire ignition probability from satellite estimates of live fuel moisture content. Fire Ecol. 2012, 8, 77–97. [Google Scholar] [CrossRef]
Figure 1. Geographical location of study area and sample point distribution.
Figure 1. Geographical location of study area and sample point distribution.
Remotesensing 14 04378 g001
Figure 2. Structure of TCN-LSTM model.
Figure 2. Structure of TCN-LSTM model.
Remotesensing 14 04378 g002
Figure 3. Structure of stacking model.
Figure 3. Structure of stacking model.
Remotesensing 14 04378 g003
Figure 4. Structure of Adaboost ensemble model.
Figure 4. Structure of Adaboost ensemble model.
Remotesensing 14 04378 g004
Figure 5. Evaluation indicators of LFMC retrieval results based on LSTM, TCN and TCN-LSTM. (a) Bias; (b) R2; (c) RMSE and ubRMSE. S represents Sentinel-1 data, L represents Landsat8 data, M represents MODIS data.
Figure 5. Evaluation indicators of LFMC retrieval results based on LSTM, TCN and TCN-LSTM. (a) Bias; (b) R2; (c) RMSE and ubRMSE. S represents Sentinel-1 data, L represents Landsat8 data, M represents MODIS data.
Remotesensing 14 04378 g005aRemotesensing 14 04378 g005b
Figure 6. LFMC retrieval results based on MODIS, Landsat-8 and Sentinel-1. The gray dotted line and the black dotted line represent the 1:1 line and the 120% fire risk based on level of moisture, respectively, and the red dotted line is the fitting line of the model retrieval results.
Figure 6. LFMC retrieval results based on MODIS, Landsat-8 and Sentinel-1. The gray dotted line and the black dotted line represent the 1:1 line and the 120% fire risk based on level of moisture, respectively, and the red dotted line is the fitting line of the model retrieval results.
Remotesensing 14 04378 g006
Figure 7. RMSEs of LFMC retrieval results for different vegetations.
Figure 7. RMSEs of LFMC retrieval results for different vegetations.
Remotesensing 14 04378 g007
Table 1. Input variables of LFMC retrieval model.
Table 1. Input variables of LFMC retrieval model.
MODISLandsat-8Sentinel-1Auxiliary Variables
Band1red σ V V Silt content
Band2green σ V H Sand content
Band3blue σ V V σ V H Clay content
Band4NIR Canopy height (m)
Band5SWIR Land cover
Band6NDWI Altitude (m)
Band7NDVI Slope (°)
NIRV
Table 2. The architectures of the used basic prediction models.
Table 2. The architectures of the used basic prediction models.
LSTMTCNTCN-LSTM
LayerOutput ShapeLayerOutput ShapeLayerOutput Shape
LSTM(32,4,10)Conv1D(32,365,64)Conv1D(32,4,32)
LSTM(32,4,10)AvgPool(32,182,64)Conv1D(32,4,32)
LSTM(32,10)Conv1D(32,182,64)MaxPool(32,2,32)
Dense(32,1)AvgPool(32,60,64)Flatten(32,64)
Conv1D(32,60,64)RepeatVector(32,1941,64)
MaxPool(32,15,64)LSTM(32,1941,10)
Flatten(32,960)LSTM(32,1941,10)
Dense(32,256)LSTM(32,10)
Dense(32,1)Dense(32,1)
Table 3. Evaluation results of LFMC retrieval results based on LSTM, TCN and TCN-LSTM.
Table 3. Evaluation results of LFMC retrieval results based on LSTM, TCN and TCN-LSTM.
DataModelBias (%)R2RMSE (%)ubRMSE (%)
SLSTM−9.390.3837.0735.86
TCN−7.420.4235.9735.2
TCN-LSTM−5.930.4434.8134.3
LLSTM−3.930.5132.6732.43
TCN−3.80.5431.5831.35
TCN-LSTM−3.120.6025.8325.64
MLSTM−7.390.5233.0132.17
TCN−5.040.5531.3930.98
TCN-LSTM−4.740.6225.1124.66
L+SLSTM−7.060.6026.3225.35
TCN−4.830.6723.0522.54
TCN-LSTM−4.810.7222.7822.31
M+SLSTM−4.760.6225.3324.88
TCN−4.530.6822.4321.97
TCN-LSTM−4.050.7522.2121.71
M+L+SLSTM−6.530.7324.3923.5
TCN−5.030.7623.0322.48
TCN-LSTM−4.660.8121.7319.93
Table 4. Comparison of different LFMC retrieval methods.
Table 4. Comparison of different LFMC retrieval methods.
Method R 2 RMSE (%)
LSTM (Landsat+SAR) [19]0.6325
TempCNN-LFMC (MODIS+Auxiliary data) [20]0.6422.74
TCN-LSTM model0.8121.73
Table 5. Performances of different ensemble learning models.
Table 5. Performances of different ensemble learning models.
DataStackingAdaboost
Bias
(%)
R 2 RMSE (%)ubRMSE (%)Bias
(%)
R 2 RMSE (%)ubRMSE (%)
S−5.750.5331.8731.35−4.590.5331.6231.29
L−3.350.723.2623.21−1.650.6523.623.54
M−4.550.7423.8223.39−4.160.6822.5322.14
L+S−1.550.8119.9619.95−2.610.762221.31
M+S−1.430.8119.8619.81−2.70.820.520.32
M+L+S−0.5420.8518.8817.99−0.5630.8319.718.8
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xie, J.; Qi, T.; Hu, W.; Huang, H.; Chen, B.; Zhang, J. Retrieval of Live Fuel Moisture Content Based on Multi-Source Remote Sensing Data and Ensemble Deep Learning Model. Remote Sens. 2022, 14, 4378. https://doi.org/10.3390/rs14174378

AMA Style

Xie J, Qi T, Hu W, Huang H, Chen B, Zhang J. Retrieval of Live Fuel Moisture Content Based on Multi-Source Remote Sensing Data and Ensemble Deep Learning Model. Remote Sensing. 2022; 14(17):4378. https://doi.org/10.3390/rs14174378

Chicago/Turabian Style

Xie, Jiangjian, Tao Qi, Wanjun Hu, Huaguo Huang, Beibei Chen, and Junguo Zhang. 2022. "Retrieval of Live Fuel Moisture Content Based on Multi-Source Remote Sensing Data and Ensemble Deep Learning Model" Remote Sensing 14, no. 17: 4378. https://doi.org/10.3390/rs14174378

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop