Next Article in Journal
Integrating Multi-Source Data to Explore Spatiotemporal Dynamics and Future Scenarios of Arid Urban Agglomerations: A Geodetector–PLUS Modelling Framework for Sustainable Land Use Planning
Previous Article in Journal
Total Precipitable Water Retrieval from FY-3D MWHS-II Data
Previous Article in Special Issue
Early Warning for Stepwise Landslides Based on Traffic Light System: A Case Study in China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatio-Temporal Fusion of Landsat and MODIS Data for Monitoring of High-Intensity Fire Traces in Karst Landscapes: A Case Study in China

by
Xiaodong Zhang
1,2,†,
Jingyi Zhao
1,2,†,
Guanzhou Chen
2,*,
Tong Wang
2,
Qing Wang
1,
Kui Wang
3 and
Tingxuan Miao
3
1
School of Geosciences, Yangtze University, Wuhan 430010, China
2
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China
3
Hubei United Transportation Investment & Development Co., Ltd., Wuhan 430040, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2025, 17(11), 1852; https://doi.org/10.3390/rs17111852
Submission received: 6 April 2025 / Revised: 21 May 2025 / Accepted: 23 May 2025 / Published: 26 May 2025
(This article belongs to the Special Issue Remote Sensing Data Application for Early Warning System)

Abstract

:
The surface fragmentation of karst landscapes leads to a high degree of coupling between fire scar site boundaries and topographic relief. However, the applicability of spatio-temporal data fusion methods for fire scar extraction in such geomorphological areas remains systematically unevaluated. This study developed a spatial–temporal adaptive fusion model integrating Landsat 30-m data with MODIS daily observations to generate continuous high-precision dNBR datasets. Using a typical karst fire region in Guizhou and Yunnan, China, as a case study, we validated the method’s effectiveness for fire trace extraction in fragmented landscapes. The proposed fusion technique addresses cloud cover limitations in humid climates by constructing continuous NBR time series, enabling precise fire boundary delineation and severity quantification. We comparatively implemented multiple fusion approaches (FSDAF, STARFM, and STDFA) and evaluated their performance through both spectral (RMSE, AD, and PSNR) and spatial (Edge, LBP, and SSIM) metrics. Key findings include the following: (1) FSDAF outperformed other methods in spectral consistency and spatial adaptation, particularly for heterogeneous mountainous terrain with fragmented vegetation. (2) Comparative experiments demonstrated that pre-calculating vegetation indices before temporal fusion (Strategy I) produced superior results to post-fusion calculation (Strategy II). Moreover, in our karst landscape study area, our proposed Hybrid Strategy selection framework can dynamically optimize the fusion process of multi-source satellite data, which is significantly better than a single fusion strategy. (3) The dNBR-based extraction achieved 90.00% producer accuracy, 69.23% user accuracy, and a Kappa coefficient of 0.718 when validated against field data. This study advances fire monitoring in karst regions by significantly improving both the spatio-temporal resolution and accuracy of burn scar detection compared to conventional approaches. The framework provides a viable solution for fire impact assessment in topographically complex landscapes under cloudy conditions.

1. Introduction

As a special karst ecosystem is widely distributed around the globe, karst landscapes are known for their unique dissolving topography, fragile ecological structure, and complex hydrological processes; but the high degree of fragmentation of their surface morphology, heterogeneity of their vegetation cover, and frequent cloudy and rainy climates make such regions particularly challenging in global ecological and environmental monitoring. Under the double pressure of climate change and human activities, forest fires are frequent in karst regions, and the rate of post-disaster vegetation succession is significantly faster compared to other ecosystems. This feature makes the accurate monitoring of fire trails face a triple challenge as follows: (1) the topographic fragmentation leads to a high proportion of mixed pixels in traditional remote sensing images, which makes it difficult to capture the details of fire boundaries; (2) the cloudy and rainy climate leads to limited access to optical remote sensing data, and insufficient spatial–temporal continuity; and (3) the rapid succession of vegetation cover may conceal fire trails within months, which compresses the window period for effective monitoring. Therefore, ways for achieving the high-precision and time-efficient extraction of fire traces in a dynamically changing and complex environment have become a core scientific issue for fire emergency management and ecological restoration assessment in karst regions. The spatio-temporal fusion technique can generate spatio-temporally continuous high-resolution datasets by constructing a spatio-temporal covariance model of reflectance and interpolating within cloud coverage gaps [1]. This significantly enhances the identification capability of fire trace boundaries [2], which is of considerable practical value for emergency responses to fires in karst areas characterized by fragmented topography and fragile ecology, as well as for ecological restoration assessments [3].
In typical cases around the world, the preferred sensor strategy for fire response mechanisms in different karst landscape development stages and the interpretation of physical mechanisms for multi-scale feature fusion are still scientific bottlenecks that need to be urgently addressed [4]. Conventional methods for extracting fire traces based on a single optical satellite are susceptible to cloud interference [5] and are not sensitive to changes in the carbon fraction of the burned surface [6,7]. While Landsat data feature a high spatial resolution of 30 m, their 16-day revisit cycle makes it difficult to capture the dynamic process of fire spreading. Conversely, while MODIS provides daily coverage [8,9], its coarse pixels (250–1000 m) are prone to interference from the mixed pixel effect caused by heterogeneous surfaces in karst areas [10]. Synthetic aperture radar (SAR) data have the ability to penetrate but are limited by the geometrical distortions in complex terrain [11]. Multi-source satellite data fusion represents a key technology for improving the accuracy of monitoring surface processes in complex environments [12] by synergistically leveraging the complementary advantages of different sensors and effectively overcoming the temporal and spatial resolution limitations of individual data sources [13]. Based on this, the construction of a Landsat-MODIS synergistic fusion framework [14] can effectively integrate the complementary spatial and temporal advantages of the two datasets, which theoretically improves both the spatial identification accuracy of burnt sites and the temporal resolution of the vegetation restoration process [15].
In emergencies such as fires, fusing thermal infrared data (e.g., VIIRS fire point data) [16] with optical imagery can rapidly locate fire scenes and assess the extent of vegetation damage. After years of development, spatio-temporal fusion models [17] have matured considerably. STARFM (Spatial and Temporal Adaptive Reflectance Fusion Model) [18], a linear hybrid model based on weighted filtering, was the first to achieve the fusion of Landsat (30 m) and MODIS (500 m) data. It introduces spatial similarity weights and controls temporal prediction errors within 8–12%. The ESTARFM (Enhanced STARFM) model [19] enhances the search strategy for similar image elements, improves accuracy in heterogeneous regions, and incorporates linear regression correction, reducing RMSE by 23% in farmland areas. The FSDAF (Flexible Spatiotemporal Data Fusion) model [20] integrates the thin-plate spline function (TPS) with endmember change detection and establishes an image element decomposition model to improve the handling of abrupt surface changes, achieving a fire traces detection error of <15% [21]. The CUFSDAF (Convolutional Unit-based FSDAF) model [22] introduces a 3 × 3 convolutional kernel to extract spatial features for constructing a residual learning framework, improving accuracy by 19% (compared with FSDAF) in urbanized regions. The VSDF (Variational Spatiotemporal Data Fusion) model [23], based on a variational optimization framework, establishes an energy function solved using the ADMM algorithm, improving convergence speed by 40% and achieving a PSNR of 32.6 dB in coverage scenarios. The high-precision [24] extraction of fire traces is critically important for forest ecosystem restoration assessments and carbon cycle research. Most current official fire trace extraction methods rely on spectral index-based threshold segmentation to calculate the differential normalized burn ratio (dNBR), quantifying changes in near-infrared and shortwave infrared bands before and after fires for rapid extraction [25]. However, their accuracy is constrained by terrain heterogeneity and the subjectivity of threshold selection [26]. Future research trends emphasize the collaborative inversion of multimodal remote sensing data, which can provide technological support for large-scale disaster emergency responses [27].
Although multi-source data fusion has shown potential in fire-scorched monument monitoring, existing research still lacks systematic validation and heterogeneous model co-optimization methods for karst landscapes. In this paper, by integrating the FSDAF, STARFM, and STDFA models, a multidimensional fusion evaluation system is constructed for the complex karst terrain, which covers band fidelity, NBR index consistency, and terrain correction robustness, solving the applicability of Landsat-MODIS data in the spatial and temporal extension of the karst zone. Further, we propose the “model optimization-feature fusion-dynamic inversion” technology chain to generate a 30 m/1 day high-resolution continuous dataset and couple dNBR with adaptive threshold segmentation to achieve subpixel-level extraction of the fire boundary. This framework not only breaks the technical bottleneck of high-precision dynamic monitoring in complex terrain areas but also provides a relocatable quantitative solution for the post-disaster recovery assessment of karst ecosystems. In view of the lack of the application of fire burn trace monitoring based on multi-source satellite data fusion in typical complex terrain areas such as karst landscapes, this paper mainly makes the following contributions:
  • For the first time, a systematic validation framework for the spatio-temporal fusion technology for karst fire traces has been constructed. By integrating heterogeneous remote sensing fusion algorithms and innovatively establishing a multi-dimensional evaluation standard system, the breakthrough reveals the applicability boundary of spatio-temporal fusion technology in karst landscape areas, laying the methodological foundation for the application of this technology in special landscape scenarios.
  • This study pioneeringly proposed a multi-source remote sensing cooperative optimisation model for karst areas. By deeply analyzing the spectral response characteristics of different algorithms and establishing a mechanism for complementing the advantages of multiple models, we have overcome the technical bottleneck of obtaining remote sensing data with high spatial and temporal resolution in complex terrain areas, and we have provided innovative solutions for monitoring fragile karst ecosystems.
  • We have constructed a dynamic monitoring technology system for karst ecosystem disaster processes. We have achieved the precise identification of the boundaries of fire traces, and we have established a time-series remote sensing inversion model, revealing for the first time the dynamic evolution law of vegetation restoration in karst areas and providing a universal technical paradigm for disaster assessment and ecological restoration of ecologically fragile areas.

2. Materials and Methods

2.1. Experimental Materials

2.1.1. Study Area

The study area of this paper is located in the Yunnan–Guizhou Plateau at the junction of Bijie City, Guizhou Province and Zhaotong City, Yunnan Province, China. This area is primarily characterized by a karst landscape. It spans latitudes from 26°21′ to 27°46′N and longitudes from 103°36′ to 106°43′E. The altitude ranges from 1000 to 2900 m, with terrain predominantly characterized by karstic mountains [28] featuring longitudinal gullies.
The region has a subtropical monsoon climate [29] and karst landform, with an average annual temperature of approximately 12–15 °C and annual precipitation ranging between 900 and 1200 mm. The dry and wet seasons are distinct, and the alternation of spring droughts and intense summer precipitation makes the area prone to forest fires. The main vegetation types include evergreen broad-leaved forests (approximately 35%); coniferous forests dominated by species from the families Crustaceae and Camphoraceae (approximately 28%); Yunnan pine and Huashan pine plantation forests and shrub (20%); vine thorn shrub endemic to the karst region; and farmland and secondary forests (17%). The average annual NDVI value ranges from 0.6 to 0.8, while the biomass density reaches 120–180 t/ha [30]. The annual mean fire-affected area in the study region over the past decade exceeded 2000 ha, with fire risk covering more than 2000 ha [31]. During the fire season (November–April), precipitation remains below 50 mm, and relative humidity falls below 60%. Fire disturbances have resulted in a 40–60% reduction in soil organic matter content. The recovery period for plant communities is 8–12 years, leading to an estimated loss of carbon sinks of approximately 5.8 t/ha [32].
Due to the unique local karst landform geographic environment, a single remote sensing data source often struggles to meet the monitoring requirements for the accurate identification and dynamic restoration assessment of fire traces. There is an urgent need to construct time-series remote sensing data with high spatial and temporal resolution to effectively monitor these areas. Landsat 8 SWIR2 (Band 7) is highly sensitive to charred materials, and when combined with the reflectance characteristics of vegetation in the NIR (Band 5), it enables the quantification of burn severity. Additionally, the extent of fire traces can be extracted by calculating the Differential Normalized Burn Ratio (dNBR). This capability highlights the significance of this study. To conduct the relevant experiments, we selected a specific area, as shown in Figure 1, as the study area.

2.1.2. Data Sources

In this paper, Landsat 8 Level-2 C2 data and the MODIS surface reflectance daily product MOD09GQ1 data covering the study area are selected as the experimental data. Both Landsat 8 Level-2 data and MODIS data were downloaded from Google Earth Engine (GEE) (https://earthengine.google.com/accessed on 14 March 2025). The Band 5 and Band 7 of Landsat 8 images correspond to the Band 2 and Band 7 of MODIS images, which belong to the near-infrared (NIR) and short-wave infrared (SWIR) bands, respectively. The spatial resolution of the two bands of MODIS data is 250 m and 500 m, and the temporal resolution is 1 day. Table 1 shows the specific band correspondence information.
The MODIS imagery on 16 July and 18 September 2023 and the Landsat 8 imagery on 16 July and 18 September 2023 were selected as the two cloud-free data periods for the experiments in this study (Figure 2). The generated data were derived from atmospherically corrected surface albedo and surface temperature. Cropping to obtain the study area data was performed during the data download process on the GEE platform [33]. The MODIS surface reflectance data products, which were atmospherically corrected, needed to be converted to WGS_1984_UTM through projection, resampled to a 30 m resolution to ensure consistency with the Landsat 8 data, and then cropped to the extent of the study area. These steps were also completed while downloading the data from the GEE platform.

2.1.3. Fusion Data Selection

The FSDAF, STARFM, and STDFA models in this study require a high spatial resolution image, a low spatial resolution image, and a high temporal resolution image as input data for modeling. The near-infrared (NIR) band and short-wave infrared (SWIR) band data from Landsat 8 and MODIS and NBR index data from 16 July 2023 are selected to train the FSDAF, STARFM, and STDFA models. All three models are used to predict MODIS data fusion, starting from 18 September 2023, to generate the corresponding high-resolution data. The real Landsat 8 image data from 18 September 2023 are utilized for comparative analyses and accuracy assessment. The experimental data information corresponding to the specific models is shown in Table 2.

2.2. Methods

The experimental process is shown in Figure 3. First, MOD09GA and Landsat 8 images were used to construct an a priori surface reflectance dataset, which was subsequently de-clouded and matched across time series [34]. Three types of image fusion (STARFM, STDFA, and FSDAF) were performed for the near-infrared (NIR) and short-wave infrared bands. To achieve higher fusion accuracy, the reflectance was linearly fitted for spectral consistency. The vegetation index NBR was calculated, and different fusion strategies were adopted for the data fusion of the bands and remote sensing indices. The fusion results were quantitatively evaluated in both spectral and spatial dimensions to select the best-performing model for the study area. For improved fusion accuracy, linear fitting of reflectance (spectral homogenization) was conducted [35]. The vegetation index NBR was computed, and different fusion strategies were applied for the data fusion of the bands and remote sensing indices. The fusion outcomes were quantitatively assessed across both spectral and spatial dimensions, selecting the optimal fusion model and strategy for the study area. Finally, based on the multi-source satellite fusion results, the threshold method was employed to determine the optimal threshold for extracting burnt traces, and the identified threshold was utilized to extract the burned area. Subsequently, heterogeneous official satellite data were introduced for the further extraction of burnt traces. Based on the aforementioned multi-source satellite fusion result data, the optimal threshold value was determined using the threshold method to extract the fire-burned land area, and this threshold was then applied to delineate the fire-burned region [36].

2.2.1. Fusion Method

  • FSDAF
FSDAF is a flexible spatio-temporal data fusion method proposed by Guo et al. [20]. Compared to previous data fusion methods, it incorporates Thin-Plate Spline (TPS) interpolation and linear mixed modeling to improve adaptability to nonlinear variations such as crop harvesting and fire [37]. The method first performs unsupervised classification of high spatial resolution images at one time point, then estimates the class changes corresponding to the low spatial resolution image at another time point, and predicts the high spatial resolution image at the second time point based on these changes. The initial prediction is achieved through the following:
L pred = L base + TPS M pred M base
where L base is the reflectance value of the high-resolution Landsat 8 image for the base date. M pred M base captures global changes in low-resolution imagery from the base date to the target date, and TPS interpolates low-resolution variations into high-resolution space to compensate for the lack of detail in high-resolution images.
In practice, image elements corresponding to two time points are considered to have different effects on classification, and residuals need to be assigned to them. The model introduces a TPS function to improve the accuracy of residual assignment. After calculating the residuals and weights, Equation (2) is used to correct the residuals of the initial prediction using the weighted amount of low-resolution variations of similar image elements, thereby improving local consistency.
L final = L pred + i = 1 n λ i · M pred ( i ) M base ( i )
where λ i is the mixing weight that represents the contribution of the i-th similar image element to the residual correction, and n represents the number of similar image elements obtained through spectral similarity screening.
2.
STARFM
STARFM is a spatio-temporal adaptive reflectance fusion method proposed by Gao et al. [18] in 2006. The core idea is to utilize the spatial details of high spatial resolution images and the time series of high temporal resolution images to generate high spatial and temporal resolution images through similar pixel search and weighted fusion (Equation (3)).
L x i , y j , t k = m = 1 n W m · M x m , y m , t k + Δ L m t 0 , t k
where Δ L m t 0 , t k = L x m , y m , t 0 M x m , y m , t 0 represents the calculation of the difference between high-resolution and low-resolution data at the base date, and W m are the weights determined by temporal and spectral differences.
3.
STDFA
STDFA is a spatio-temporal data fusion analysis method proposed by Zhang et al. [38]. It adaptively adjusts the contribution weights of different spatio-temporal location data to the target image element through the spatio-temporal distance decay function. The closer the spatial distance and the smaller the temporal difference, the higher the weight. Assuming that the target high-resolution image is linearly combined using the spatio-temporal neighboring image elements of the low-resolution image, the weight coefficients are solved by least squares optimization. Statistics such as local variance are introduced to dynamically adjust the spatio-temporal neighborhood range, improving the fusion robustness for complex scenes. The STDFA model can be simplified as follows:
Y ^ x 0 , t 0 = i = 1 N w i · Y x i , t i
where Y ^ x 0 , t 0 is the predicted value of the target spatio-temporal position x 0 , t 0 , Y x i , t i represents observations at the ith image element in the spatio-temporal neighborhood, and w i is the spatio-temporal weight, jointly determined by the spatial kernel K s and the temporal kernel K t as follows:
w i = K s x 0 x i | | · K t t 0 t i j = 1 N K s x 0 x j | | · K t t 0 t j
where the spatial kernel K s is often a Gaussian function K s ( d ) = e d 2 / h s 2 , and the temporal kernel K t can be similarly defined as K t ( Δ t ) = e Δ t 2 / h t 2 . The weighting function explicitly models the spatio-temporal correlation, and the model is based on linear weighting to improve computational efficiency.

2.2.2. Fusion Strategies

When using three models mentioned above for the spatio-temporal fusion of NBR indices, an appropriate fusion strategy must be determined. Both strategies can obtain NBR results with high spatial and temporal resolution; however, their accuracy in different bands and for the NBR index remains unclear. The following two strategies are applied in our study:
  • Strategy I: The NBR is calculated using the reflectance of the NIR and shortwave IR bands, and then the calculated NBR is fused using each of the three models.
  • Strategy II: We utilize the three models to perform spatio-temporal fusion of the NIR (Landsat 8 is band 5 and MOD09GA is band sur-refl-b02) and short-wave IR (Landsat 8 is band 7 and MOD09GA is band sur-refl-b07) bands, which are required for the calculation of the NBR index. The NBR is then computed using the two fused bands.
  • Hybrid Strategy: The NBR is calculated in stages and the fusion weights are dynamically adjusted, with Strategy II adopted in sparse data areas (when cloud cover is severe) to enhance spatial and temporal continuity, and Strategy I adopted in complex terrain areas to preserve spectral details.
For the first two strategies (Strategy I and Strategy II), this study makes the following hypotheses: If the experimental results show that Strategy I (calculating NBR before fusion) is better, it means that fusion at the exponential level is more robust to noise or error. The NBR itself is a normalized exponential, which attenuates the noise in the original band through the normalization process, making it easier for the fusion model to capture the spatio-temporal variation law of the NBR. The fusion model is better at capturing the exponential features and is more sensitive to the NBR index in the low dynamic range (−1 to 1), while it is less adaptive to the heterogeneity of the original bands in the high dynamic range (e.g., NIR, SWIR). If Strategy II (fusing the bands before calculating the NBR) is better, it means that the band-level fusion retains more physical information, and the reflectance of the original bands contains richer radiative properties. The direct fusion of bands can more accurately restore the true reflectance of the surface, thus improving the accuracy of the subsequent NBR calculation.The calculation of NBR involves a nonlinear combination (ratio and difference) between the bands, which may be more completely preserved if the bands are fused first, whereas fusing the NBR first may lead to a loss of information due to the linear assumptions of the model.
For the Hybrid Strategy, a spatio-temporal adaptive fusion framework based on dynamic strategy selection and weighted mixing is developed to address the monitoring challenges of surface fragmentation, frequent cloud and rain disturbances, and rapid vegetation succession in karstic landscapes. By extracting cloud coverage and terrain complexity as regional features, the optimal strategy is dynamically selected using a threshold value. At the same time, the joint modeling of spectral spatial features is achieved by a weighted Hybrid Strategy, in which the weights of the two strategies are dynamically adjusted according to the degree of cloud coverage, with the higher weights of Strategy II for higher cloud coverage or simpler terrain, and the higher weights of Strategy I for more complex terrain or lower cloud coverage. The machine learning random forest method is used to train the threshold model based on historical data to achieve the dynamic weighted fusion strategy selection.

2.2.3. Extraction Methods for Fire Sites

The fire traces studied in this experiment can be further analyzed by comparing the changes in vegetation in the area before and after the disaster. Change Detection (CD) [39,40,41] refers to the process of identifying differences in the state of an object or phenomenon by observing it at different times. The goal of change detection based on Earth observation data is to identify changes on Earth by comparing two or more satellite or aerial images covering the same area at different times. Satellite remote sensing enables the repeated acquisition of images of the same area at short time intervals, making CD one of the primary applications of remote sensing data obtained from Earth-orbiting satellites.
The Normalized Burn Ratio (NBR) effectively identifies differences in surface reflectance before and after a fire, leveraging the sensitivity of near-infrared (NIR) and shortwave infrared (SWIR) bands to changes in vegetation moisture content and biomass. The Difference Normalized Burn Ratio (dNBR), calculated based on the NBR, quantifies fire intensity through difference operations and is now the preferred method for producing many official fire-site data products [42].
The NBR index for the target date was calculated as follows:
NBR past = ρ NIR ρ SWIR 2 ρ NIR + ρ SWIR 2
In order to generate the difference index (dNBR), it is also necessary to calculate the NBR index before vegetation burning. The experiment was conducted using data from 16 July 2023, based on the following:
NBR pre = ρ NIR ρ SWIR 2 ρ NIR + ρ SWIR 2
where ρ NIR indicates reflectance in the near-infrared band (Landsat 8 corresponds to band 5 and MODIS corresponds to band 2), and ρ SWIR 2 indicates short-wave infrared 2-band reflectance (Landsat 8 corresponds to band 7 and MODIS corresponds to band 7).
The Differential Normalized Burn Ratio (dNBR), representing quantitative fire intensity, was obtained by a difference operation (Equation (8)) based on the results of the above equation as follows:
dNBR = NBR pre NBR past
where higher dNBR values indicate more severe fire burning. The USGS standardized grading method was used to set the threshold dNBR > 0.3 for extracting the fire burn sites [43].

2.2.4. Accuracy Evaluation Methods

  • Spatio-temporal fusion accuracy assessment
Fused image quality evaluation methods can be categorized into the following two main types: subjective evaluation methods that use human vision as the primary evaluation index, and objective evaluation methods that employ specific algorithms to provide quantitative metrics. In this study, an objective evaluation method is used to quantitatively assess the similarity between the fused data and the real Landsat 8 image data. The validation image in this study is a high-resolution image from 18 September 2023, which is used to quantitatively evaluate the accuracy of the fusion results.
The accuracy assessment metrics in this study use a new framework proposed by Zhu et al. [44] that combines spectral and spatial details, and the experimental results show that the optimal combination of accuracy assessment metrics includes the root mean square error (RMSE), average difference (AD), peak signal-to-noise ratio (PSNR), Robert’s edge, local binary pattern (LBP), and structural similarity index (SSIM). This combination not only provides a comprehensive and effective assessment of errors in both the spectral and spatial dimensions but also significantly reduces the information redundancy associated with highly correlated metrics.Therefore, the fusion results of this study are evaluated using these six accuracy assessment metrics for quality assessment.
2.
Validation of the results of fire trace extraction
The reference data were constructed using Sentinel-2, VIIRS, and Himawari-8 official fire point data products as ground truth to build the validation dataset, screening for fires that overlapped with Landsat imaging time within ±3 days. Spatial matching was performed by superimposing fire point coordinates onto the NBR results, with true positives awarded if the fire point fell within a fire trace polygon. Confusion matrices [45] were constructed to statistically analyze true positives (TP), false positives (FP), false negatives (FN), and true negatives (TN). The producer accuracy (PA), user accuracy (UA), overall accuracy (OA), and Kappa coefficients were then calculated.

3. Results

3.1. Results of the Spatial-Temporal Fusion of Multi-Source Satellite Data

Figure 4, Figure 5 and Figure 6 show the fusion results of the experimental data for the three models over two bands and one vegetation index. We found that all three models achieved better fusion results over the study area compared to the original images. Visually, the NBR results obtained by fusing Landsat 8 and MODIS data using the FSDAF model (Figure 4d) are closer to the real image in texture details (Figure 4a) and can reproduce the real NBR well. There is not much difference between the fusion results of the three models in the NIR and SWIR bands visually, which necessitates the use of accuracy assessment metrics in later experiments to quantify the accuracy of the fusion results in the spectral and spatial dimensions. We compare the results in spectral and spatial dimensions.
Table 3 and Table 4 show the accuracy assessment results of the experimental group under the two strategies for the three models from spectral (RMSE, PSNR, and AD) and spatial (Edge, LBP, and SSIM) dimensions.
For Table 3, we conclude that for the fusion of NIR and SWIR bands as well as NBR indices, the accuracy of the FSDAF model is mostly higher than that of the other experimental models in the spectral dimension. Specifically, in the spectral dimension of the experiments, the fusion result of the FSDAF model has an RMSE closer to 0, the largest PSNR value, and an AD value of 0.1041, which is only slightly higher in the NIR band compared to the STARFM model’s value of 0.1004.
For Table 4, we conclude that for the fusion of the NIR and SWIR bands as well as the NBR index, the accuracy of the FSDAF model in the spatial dimension is generally higher than that of the other experimental models. Specifically, in the spatial dimension of the experiments, the fusion results of the FSDAF model are closer to or equal to 1 for both Edge and SSIM values, while the LBP is slightly lower at 0.9998 for the SWIR bands with the NBR index compared to 0.9999 for the STARFM model.
Within the study area, we conclude that the FSDAF model exhibits higher temporal and spatial fusion accuracies in both the spectral and spatial dimensions for the NIR, SWIR, and NBR indices.

3.2. Results of Experiments Comparing Fusion Strategies

Figure 7 shows the fusion results of the experimental data under three models and two strategies. We found that both strategies achieved better fusion results for the study area compared to the original images. Visually, the results obtained using fusion Strategy I, which calculates the vegetation index first and then fuses it (Figure 7a–e), as well as the NBR results obtained by fusing Landsat 8 and MODIS data using the FSDAF model (Figure 7a,b), appear closer to the texture details of the real data. Regarding the boundary located on the right side of the original remote sensing image, which is formed by mosaicking during the data preprocessing stage, the fusion results obtained by the STARFM and STDFA models using Strategy I exhibit noticeable errors. This is because the two images used for mosaicking are not from the same day, and the fusion algorithms of these two models are significantly less effective than the FSDAF model in handling the image splicing boundary.
Figure 8 shows the fusion results of the three models using the Hybrid Strategy, and from the visual interpretation results we can see that the fusion results obtained using the Hybrid Strategy are more similar to the original images than those obtained using Strategy I and Strategy II.
Table 5 shows the results of the accuracy assessment of the real experimental groups of the three spatio-temporal fusion models under the three strategies in terms of spectral and spatial dimensions. We conclude that all three multi-source satellite data spatio-temporal fusion models in the study area within the karstic landscape area obtained higher accuracy fusion results, in terms of spectral and spatial dimensions, using a Hybrid Strategy. This confirms that our proposed Hybrid Strategy fusion framework is more suitable for regions with thick cloud cover or complex topography, such as a karstic terrain, than a single fusion strategy.
Comparing and contrasting Strategy I and Strategy II, all three spatio-temporal fusion models perform better with Strategy I, both in the spectral and spatial dimensions. Additionally, we reach the same conclusion as observed in the accuracy evaluation results in Table 3 and Table 4, as follows: FSDAF achieves the highest accuracy using Strategy I in the study area. Compared with Strategy II, the specific performance metrics are as follows: the RMSE value is 0.1749, which is closest to 0; the PSNR value is 94.8161, which is the largest; the values of Edge and SSIM are both 1; and the AD value is −0.1586, which is slightly farther from 0 compared to Strategy II. In addition, from Table 4, we can also conclude that the accuracy of all three spatio-temporal fusion models in terms of spatial aspects using Strategy I to directly compute the NBR index is much higher than that of Strategy II for direct band fusion. Take the FSDAF model as an example, the Edge value (0.8226 at NIR, 0.8269 at SWIR, and 0.000 at NBR), the LBP value (0.9998 at NIR, 0.9999 at SWIR, and 0.9999 at NBR), and the SSIM value (0.8843 at NIR, 0.9648 at SWIR, and 1.0000 at NBR) demonstrate that the model is able to retain spatial and textural details in the image well.
Within the study area, we conclude that the Hybrid Strategy exhibits higher spatio-temporal fusion accuracy in both spectral and spatial dimensions.

3.3. Fire Trace Extraction Results

Figure 9 shows the binarization of the results based on the fusion of multi-source satellite data, calculating the dNBR and extracting the range of fire traces using the threshold method. The black areas indicate unchanged pixels in the change detection, with the unfired area assigned a value of 0, while the white areas represent changed pixels in the change detection, with the overfired area assigned a value of 1.
In this study, based on Landsat 8 and MODIS data, the extent of fire traces was extracted by calculating the normalized burn ratio (NBR), and a validation dataset was constructed by incorporating official fire point products from heterogeneous satellites. Spatial superposition analysis was employed to generate a confusion matrix (Table 6) for quantitatively assessing the producer accuracy (90.00%), user accuracy (69.23%), and Kappa coefficient (0.718). The results indicate that the NBR method is sensitive to medium- to high-intensity large fire areas but is vulnerable to interference from cloud residue and bare ground. The integration of temporal difference NBR fused with machine learning classifiers can be explored further to enhance the robustness of extraction in complex terrains.

4. Discussion

4.1. Methodological Improvements and Application Potential for Fire Monitoring in Karst Landscapes

In this study, the applicability of spatio-temporal data fusion methods in complex terrain areas is systematically verified for the first time, addressing the technical challenges associated with the dynamic monitoring of fire traces in karst terrain regions. Compared with traditional single remote sensing data sources (Landsat or MODIS applied independently), the spatio-temporal adaptive fusion model employed in this study effectively addresses the issue of missing time-series data caused by cloud cover by coupling Landsat’s 30 m spatial resolution with MODIS’s high temporal resolution data. Additionally, by incorporating a terrain heterogeneity correction module, the interference of abrupt surface reflectance changes due to karst fragmentation on dNBR calculations is significantly mitigated. Case-area validation shows that the fused dNBR dataset exhibits high overall performance in fire boundary identification, especially in the extraction of medium-intensity and high-intensity fires at small scales. The cross-tabulation analysis shows that the method maintains a stable recognition capability under complex terrain and heterogeneous surface conditions, and its performance advantages are mainly reflected in the accuracy of the fine-grained fire edges and the sensitivity of the response to low-contrast burning regions. This result verifies the potential of the technical framework in improving the resolution and adaptability of fire monitoring, and it provides a reliable database for subsequent multi-scale disaster assessment.

4.2. Mechanistic Advantages of FSDAF in Topographically Complex Environments

The present study confirms the superiority of the FSDAF model in the application of the NBR-based index for monitoring fire-scorched sites in karstic landscapes, which may stem from its adaptive ability to spatio-temporal non-stationarity, consistent with the findings of [46] in arid zones. In comparison with similar studies, some previous studies concluded that ESTARFM is more suitable for vegetation indices [47]; however, the conclusion obtained in this paper is that the FSDAF model has higher accuracy for NBR index fusion and fire burnt trace extraction in the study area. This can be attributed to the following reasons: the FSDAF model is more dynamically adaptive, introducing a chunking and fusion strategy along with adaptive weight adjustments, which can better capture surface mutation areas such as fire traces. Experimental data show that the NBR fusion RMSE of FSDAF is approximately 15% lower than that of ESTARFM at a 30-day time interval [20].
The study area of this paper, the Yunnan–Guizhou Plateau, is located in a karst landform region, greatly affected by climate change [48]. The strong topographic undulation in the Yunnan–Guizhou Plateau can lead to the distortion of the reflectivity of image elements [49]. FSDAF performs DEM-assisted radiometric terrain correction before fusion [50], while ESTARFM usually assumes the ground is flat, which may lead to NBR fusion value shifts in mountainous areas, as reported by Borini Alves et al. [51], who noted a 32% deviation. Small plots of mixed vegetation are common in highland areas, and FSDAF’s 30 m × 30 m plot fusion strategy is better than ESTARFM’s fixed-window approach in capturing details such as terraces or forest edges.
The experimental results of this paper based on NBR Vegetation Index Fusion for fire-site monitoring show that Strategy I (calculating the vegetation index before fusion) is better, and this result confirms the previous hypothesis, which indicates that the index-level fusion is more robust to noise or errors, and that the spatio-temporal fusion model of multivariate satellite data is more adept at capturing the index features, and the spatio-temporal fusion model can be directly applied to the vegetation index product. This approach breaks the limitation of traditional band fusion prior to index calculation, reduces the accumulation of band fusion errors, and stabilizes the physical significance of the vegetation index. Some scholars have observed a similar pattern in NDVI fusion [52,53]. The high accuracy of Strategy I suggests that the time-smoothing property of the vegetation index product itself may suppress noise transfer during the fusion process. This finding provides a theoretical basis for the direct fusion of vegetation indices. Future studies need to further validate the applicability of the method in continuously cloudy regions and explore the transfer mechanism of atmospheric correction errors during the fusion process.

4.3. Applicability of a Spatio-Temporal Adaptive Fusion Framework Based on Dynamic Strategy Selection and Weighted Mixing in Regions with Thick Cloud Cover or Complex Terrain, Exemplified by Karstic Terrain Areas

The dynamic strategy selection and weighted hybrid framework proposed in this study achieves a remarkable technological breakthrough in response to the special challenges (fragmented surface, frequent cloud and rain disturbances, and rapid vegetation succession) of monitoring fire trails in karstic terrain areas. Through the intelligent decision-making of the cloud coverage ratio and terrain complexity by machine learning random forest, this method systematically couples regional features with remote sensing data processing flow dynamically for the first time, which solves the defect of a single fusion strategy selection in traditional methods. The experiments show that this adaptive strategy can effectively improve the spatial and temporal continuity of the data by selecting the strategy of ‘fusing and then calculating NBR’ to fill the observation gap of Landsat-8 with the high-frequency data from MODIS in areas with severe cloud cover or simple terrain, while the strategy of ‘calculating NBR and then calculating NBR’ is adopted in areas with complex terrain. The ‘NBR calculation before blending’ strategy is adopted in areas with complex terrain, which significantly reduces the interference of the blended pixels on the boundary of fire trails by preserving the spectral details with high spatial resolution. The introduction of the weighted blending strategy further enhances the flexibility of the approach.
The innovation of this method is reflected in the following three aspects: Intelligent strategy selection, quantifying the dynamic effects of clouds and terrain through the random forest model, avoiding the subjectivity of manually setting the thresholds, and providing a reusable decision-making framework for different geomorphological scenarios. Synergistic optimization of multi-source data, combining the spatial adaptability of the FSDAF algorithm with the constraints of the topographic factors, and solving the spectral heterogeneity problem caused by the topography of karst areas, such as caves and peaks, etc. The proportion of mixed image elements is 23% lower than that of the traditional STARFM method. The spectral heterogeneity of karst area due to caves, peak forests, and other topography is solved, and the ratio of mixed pixels is reduced compared with the traditional STARFM method. The temporal dynamic modeling captures the subtle changes of the rapid vegetation succession in the karst area through the tracking of the vegetation restoration process by the dNBR index, which provides a scientific basis for the determination of the window period of post-disaster ecological restoration.

4.4. Innovations and Limitations of Fire Response Paradigm Construction and the Fire Trail Extraction Framework in Karst Regions

This breakthrough provides high spatial and temporal resolution data support for post-fire ecological restoration assessments in ecologically fragile karst areas. Future studies could explore coupling thermal infrared band data to optimize spectral separation algorithms for such specific features. The technical framework developed in this study not only fills the gap in multi-source remote sensing fire monitoring in karst landscapes but also serves as a paradigm reference for disaster monitoring in other complex terrain areas through its spatio-temporal adaptive fusion mechanism.
However, there are still limitations in the study, for example, the spatial and temporal coverage of the thermal infrared band data is limited by the frequency of satellite observations, and the complex three-dimensional structure of karst features may affect the accuracy of the spectral separation algorithms. Moreover, the robustness of the model under extreme climatic conditions needs to be verified in the long term. Future research could explore coupling thermal infrared band data to optimize the spectral separation algorithm for such specific features. The technical framework developed in this study not only fills the gap of multi-source remote sensing fire monitoring in karst landscapes, but it also provides an example reference for disaster monitoring in other complex terrain areas through its spatial-temporal adaptive fusion mechanism. This is significant in the following ways: for the first time, it realizes the synergistic analysis of multi-source data in monitoring fires in karst areas, and it provides migratory technical paths for disaster response in similar fragile ecosystems around the world. The significance of this study is that it realizes the synergistic analysis of multi-source data for fire monitoring in karst areas for the first time, provides a transferable technical path for disaster response in similar fragile ecosystems around the world, and promotes the application of remote sensing technology in ecological restoration assessment.

5. Conclusions

In this study, we systematically validate the spatio-temporal fusion technique for NBR-based fire trace extraction in karst landscapes for the first time, addressing the unique challenges posed by topographically coupled fire patterns in fragmented landscapes. Focusing on the transition zone of the Yunnan–Guizhou Plateau, we conducted a comprehensive evaluation of three fusion models (STARFM, STDFA, and FSDAF) for vegetation monitoring applications, and the response time for post-disaster fire trace extraction was controlled to be within 1 day. Our comparative experiments using a dual-path fusion strategy yielded three main conclusions, outlined as follows: (1) The FSDAF model performs well in band-level fusion, generating 30 m/day resolution data that can effectively support short-wave infrared (IR) fire detection and near-infrared (NIR) vegetation restoration monitoring. (2) In this study, a spatio-temporal adaptive fusion framework based on dynamic strategy selection and weighted mixing is proposed, and the fusion accuracies are all higher than the fusion results using a single strategy (Strategy I or Strategy II) in areas with fragmented surfaces, frequent cloud and rain disturbances, and rapid vegetation succession such as karstic geomorphological zones. Moreover, the spectral and spatial accuracies of the NBR fusion results generated by Strategy I (pre-fusion exponential computation) are significantly higher than those of Strategy II, confirming our hypothesis that exponential fusion enhances robustness to noise and better preserves spectral features. (3) the dNBR thresholding method (Kappa = 0.718) shows particular sensitivity to medium- and high-intensity burned regions (dNBR > 0.3), outperforming conventional single-source methods.
Future research should focus on the following three key directions to extend the current framework: (1) address the particularity of karst landscapes in the study area of this paper, which are fragmented and distributed, resulting in higher producer accuracy than user accuracy, and to combine high-resolution data, such as Sentinel-2, to improve the mixed image decomposition capability, especially in heterogeneous landscapes; (2) optimize the temporal interpolation algorithm to maintain the stability of FSDAF performance at input intervals of more than 40 days, possibly through the dynamic weighting of phenological features; (3) integrate geostationary satellite data (e.g., Himawari-8) to solve the time synchronisation challenge in monitoring fast-spreading fires.

Author Contributions

Conceptualization, X.Z.; methodology, J.Z. and G.C.; investigation, J.Z. and T.W.; data curation, J.Z.; writing—original draft preparation, X.Z. and J.Z.; writing—review and editing, G.C. and T.W.; supervision, X.Z., G.C. and Q.W.; project administration, G.C. and K.W.; funding acquisition, G.C. and T.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China [grant number 42101346]; the Key Technology Research and Demonstration Application of the 3D Scene Platform for Intelligent Highway Inspection [grant number HBJT-LJTGS-243773001]; and the China Postdoctoral Science Foundation [grant number 2020M680109].

Data Availability Statement

The datasets are available upon reasonable request from the corresponding author.

Acknowledgments

During the preparation of this work, the authors used GPT-4o for grammatical modification and polishing. After using this tool/service, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication. We are also thankful for the reviewers’ evaluation of our paper and the constructive comments they made.

Conflicts of Interest

Authors Kui Wang and Tingxuan Mia were employed by the company Hubei United Transportation Investment & Development Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Hu, Y.; Wang, H.; Niu, X.; Shao, W.; Yang, Y. Comparative analysis and comprehensive trade-off of four spatiotemporal fusion models for NDVI generation. Remote Sens. 2022, 14, 5996. [Google Scholar] [CrossRef]
  2. Kurbanov, E.; Vorobev, O.; Lezhnin, S.; Sha, J.; Wang, J.; Li, X.; Cole, J.; Dergunov, D.; Wang, Y. Remote sensing of forest burnt area, burn severity, and post-fire recovery: A review. Remote Sens. 2022, 14, 4714. [Google Scholar] [CrossRef]
  3. Yue, X.; Unger, N. Fire air pollution reduces global terrestrial productivity. Nat. Commun. 2018, 9, 5413. [Google Scholar] [CrossRef] [PubMed]
  4. Lai, C.; Zeng, S.; Guo, W.; Liu, X.; Li, Y.; Liao, B. Forest fire prediction with imbalanced data using a deep neural network method. Forests 2022, 13, 1129. [Google Scholar] [CrossRef]
  5. Zhang, X.; Wang, T.; Chen, G.; Tan, X.; Zhu, K. Convective clouds extraction from Himawari–8 satellite images based on double-stream fully convolutional networks. IEEE Geosci. Remote Sens. Lett. 2019, 17, 553–557. [Google Scholar] [CrossRef]
  6. Chen, G.; Zhang, K.; Zhang, X.; Xie, H.; Yang, H.; Tan, X.; Wang, T.; Ma, Y.; Wang, Q.; Cao, J.; et al. Enhancing Terrestrial Net Primary Productivity Estimation with EXP-CASA: A Novel Light Use Efficiency Model Approach. arXiv 2024, arXiv:2406.19969. [Google Scholar] [CrossRef]
  7. Kong, J.; Ryu, Y.; Huang, Y.; Dechant, B.; Houborg, R.; Guan, K.; Zhu, X. Evaluation of four image fusion NDVI products against in situ spectral-measurements over a heterogeneous rice paddy landscape. Agric. For. Meteorol. 2021, 297, 108255. [Google Scholar] [CrossRef]
  8. Hazaymeh, K.; Hassan, Q.K. Spatiotemporal image-fusion model for enhancing the temporal resolution of Landsat-8 surface reflectance images using MODIS images. J. Appl. Remote Sens. 2015, 9, 096095. [Google Scholar] [CrossRef]
  9. Liu, C.; Zhang, X.; Wang, T.; Chen, G.; Zhu, K.; Wang, Q.; Wang, J. Detection of vegetation coverage changes in the Yellow River Basin from 2003 to 2020. Ecol. Indic. 2022, 138, 108818. [Google Scholar] [CrossRef]
  10. Zhou, X.; Wang, P.; Tansey, K.; Zhang, S.; Li, H.; Tian, H. Reconstruction of time series leaf area index for improving wheat yield estimates at field scales by fusion of Sentinel-2,-3 and MODIS imagery. Comput. Electron. Agric. 2020, 177, 105692. [Google Scholar] [CrossRef]
  11. Wang, Y.; Luo, X.; Wang, Q. A boundary finding-based spatiotemporal fusion model for vegetation index. Int. J. Remote Sens. 2021, 42, 8236–8261. [Google Scholar] [CrossRef]
  12. Liu, M.; Liu, X.; Dong, X.; Zhao, B.; Zou, X.; Wu, L.; Wei, H. An improved spatiotemporal data fusion method using surface heterogeneity information based on estarfm. Remote Sens. 2020, 12, 3673. [Google Scholar] [CrossRef]
  13. Lin, R.; Wei, Z.; Chen, H.; Han, C.; Zhang, B.; Jule, M. Improving the Spatiotemporal Resolution of Land Surface Temperature Using a Data Fusion Method in Haihe Basin, China. Remote Sens. 2024, 16, 2374. [Google Scholar] [CrossRef]
  14. Gao, F.; Anderson, M.C.; Zhang, X.; Yang, Z.; Alfieri, J.G.; Kustas, W.P.; Mueller, R.; Johnson, D.M.; Prueger, J.H. Toward mapping crop progress at field scales through fusion of Landsat and MODIS imagery. Remote Sens. Environ. 2017, 188, 9–25. [Google Scholar] [CrossRef]
  15. Cao, Y.; Du, P.; Zhang, M.; Bai, X.; Lei, R.; Yang, X. Quantitative Evaluation of Grassland SOS Estimation Accuracy Based on Different MODIS-Landsat Spatio-Temporal Fusion Datasets. Remote Sens. 2022, 14, 2542. [Google Scholar] [CrossRef]
  16. Schroeder, W.; Oliva, P.; Giglio, L.; Csiszar, I.A. The New VIIRS 375 m active fire detection data product: Algorithm description and initial assessment. Remote Sens. Environ. 2014, 143, 85–96. [Google Scholar] [CrossRef]
  17. Wang, C.; He, T.; Song, D.X.; Zhang, L.; Zhu, P.; Man, Y. Comparison of change-based and shape-based data fusion methods in fine-resolution land surface phenology monitoring with Landsat and Sentinel-2 data. Sci. Total Environ. 2024, 927, 172014. [Google Scholar] [CrossRef]
  18. Gao, F.; Masek, J.; Schwaller, M.; Hall, F. On the blending of the Landsat and MODIS surface reflectance: Predicting daily Landsat surface reflectance. IEEE Trans. Geosci. Remote Sens. 2006, 44, 2207–2218. [Google Scholar]
  19. Zhu, X.; Chen, J.; Gao, F.; Chen, X.; Masek, J.G. An enhanced spatial and temporal adaptive reflectance fusion model for complex heterogeneous regions. Remote Sens. Environ. 2010, 114, 2610–2623. [Google Scholar] [CrossRef]
  20. Guo, D.; Shi, W.; Hao, M.; Zhu, X. FSDAF 2.0: Improving the performance of retrieving land cover changes and preserving spatial details. Remote Sens. Environ. 2020, 248, 111973. [Google Scholar] [CrossRef]
  21. Xu, C.; Du, X.; Fan, X.; Jian, H.; Yan, Z.; Zhu, J.; Wang, R. FastVSDF: An efficient spatiotemporal data fusion method for seamless data cube. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–22. [Google Scholar] [CrossRef]
  22. Gao, H.; Zhu, X.; Guan, Q.; Yang, X.; Yao, Y.; Zeng, W.; Peng, X. cuFSDAF: An enhanced flexible spatiotemporal data fusion algorithm parallelized using graphics processing units. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
  23. Xu, C.; Du, X.; Yan, Z.; Zhu, J.; Xu, S.; Fan, X. VSDF: A variation-based spatiotemporal data fusion method. Remote Sens. Environ. 2022, 283, 113309. [Google Scholar] [CrossRef]
  24. Yang, M.; Zhou, Y.; Xie, Y.; Shao, W.; Luo, F. Accuracy Evaluation of Four Spatiotemporal Fusion Methods for Different Time Scales. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 8291–8301. [Google Scholar] [CrossRef]
  25. Key, C.H.; Benson, N.C. Landscape assessment (LA). In FIREMON: Fire Effects Monitoring and Inventory System; Gen. Tech. Rep. RMRS-GTR-164-CD; Lutes, D.C., Keane, R.E., Caratti, J.F., Key, C.H., Benson, N.C., Sutherland, S., Gangi, L.J., Eds.; US Department of Agriculture, Forest Service, Rocky Mountain Research Station: Fort Collins, CO, USA, 2006; Volume 164, p. LA-1-55. [Google Scholar]
  26. Miller, J.D.; Thode, A.E. Quantifying burn severity in a heterogeneous landscape with a relative version of the delta Normalized Burn Ratio (dNBR). Remote Sens. Environ. 2007, 109, 66–80. [Google Scholar] [CrossRef]
  27. Irwin, K.; Beaulne, D.; Braun, A.; Fotopoulos, G. Fusion of SAR, optical imagery and airborne LiDAR for surface water detection. Remote Sens. 2017, 9, 890. [Google Scholar] [CrossRef]
  28. Wang, K.; Zhang, C.; Chen, H.; Yue, Y.; Zhang, W.; Zhang, M.; Qi, X.; Fu, Z. Karst landscapes of China: Patterns, ecosystem processes and services. Landsc. Ecol. 2019, 34, 2743–2763. [Google Scholar] [CrossRef]
  29. Su, Z.; Tigabu, M.; Cao, Q.; Wang, G.; Hu, H.; Guo, F. Comparative analysis of spatial variation in forest fire drivers between boreal and subtropical ecosystems in China. For. Ecol. Manag. 2019, 454, 117669. [Google Scholar] [CrossRef]
  30. Zeng, Z.Q.; Wang, S.L.; Zhang, C.M.; Gong, C.; Hu, Q. Carbon storage in evergreen broad-leaf forests in mid-subtropical region of China at four succession stages. J. For. Res. 2013, 24, 677–682. [Google Scholar] [CrossRef]
  31. Pellegrini, A.F.; Harden, J.; Georgiou, K.; Hemes, K.S.; Malhotra, A.; Nolan, C.J.; Jackson, R.B. Fire effects on the persistence of soil organic matter and long-term carbon storage. Nat. Geosci. 2022, 15, 5–13. [Google Scholar] [CrossRef]
  32. Van Der Werf, G.R.; Randerson, J.T.; Collatz, G.J.; Giglio, L. Carbon emissions from fires in tropical and subtropical ecosystems. Glob. Change Biol. 2003, 9, 547–562. [Google Scholar] [CrossRef]
  33. Vijayakumar, S.; Saravanakumar, R.; Arulanandam, M.; Ilakkiya, S. Google Earth Engine: Empowering developing countries with large-scale geospatial data analysis—A comprehensive review. Arab. J. Geosci. 2024, 17, 139. [Google Scholar] [CrossRef]
  34. Xing, H.; Chen, B.; Feng, Y.; Ni, Y.; Hou, D.; Wang, X.; Kong, Y. Mapping irrigated, rainfed and paddy croplands from time-series Sentinel-2 images by integrating pixel-based classification and image segmentation on Google Earth Engine. Geocarto Int. 2022, 37, 13291–13310. [Google Scholar] [CrossRef]
  35. Guo, Y.; Chen, S.; Li, X.; Cunha, M.; Jayavelu, S.; Cammarano, D.; Fu, Y. Machine learning-based approaches for predicting SPAD values of maize using multi-spectral images. Remote Sens. 2022, 14, 1337. [Google Scholar] [CrossRef]
  36. Zhao, R.; Li, Y.; Ma, M. Mapping paddy rice with satellite remote sensing: A review. Sustainability 2021, 13, 503. [Google Scholar] [CrossRef]
  37. Zheng, T.; Chen, Z.; Bai, J.; Xie, H.; Jiang, Y.G. Tps++: Attention-enhanced thin-plate spline for scene text recognition. arXiv 2023, arXiv:2305.05322. [Google Scholar]
  38. Zhang, H.; Song, Y.; Han, C.; Zhang, L. Remote sensing image spatiotemporal fusion using a generative adversarial network. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4273–4286. [Google Scholar] [CrossRef]
  39. Roy, D.P.; Jin, Y.; Lewis, P.; Justice, C. Prototyping a global algorithm for systematic fire-affected area mapping using MODIS time series data. Remote Sens. Environ. 2005, 97, 137–162. [Google Scholar] [CrossRef]
  40. Wang, Q.; Zhang, X.; Chen, G.; Dai, F.; Gong, Y.; Zhu, K. Change detection based on Faster R-CNN for high-resolution remote sensing images. Remote Sens. Lett. 2018, 9, 923–932. [Google Scholar] [CrossRef]
  41. Tan, X.; Chen, G.; Wang, T.; Wang, J.; Zhang, X. Segment change model (scm) for unsupervised change detection in vhr remote sensing images: A case study of buildings. In Proceedings of the IGARSS 2024—2024 IEEE International Geoscience and Remote Sensing Symposium, Athens, Greece, 7–12 July 2024; pp. 8577–8580. [Google Scholar]
  42. Miller, J.D.; Knapp, E.E.; Key, C.H.; Skinner, C.N.; Isbell, C.J.; Creasy, R.M.; Sherlock, J.W. Calibration and validation of the relative differenced Normalized Burn Ratio (RdNBR) to three measures of fire severity in the Sierra Nevada and Klamath Mountains, California, USA. Remote Sens. Environ. 2009, 113, 645–656. [Google Scholar] [CrossRef]
  43. Key, C.H. Ecological and sampling constraints on defining landscape fire severity. Fire Ecol. 2006, 2, 34–59. [Google Scholar] [CrossRef]
  44. Zhu, X.; Duan, S.B.; Li, Z.L.; Zhao, W.; Wu, H.; Leng, P.; Gao, M.; Zhou, X. Retrieval of land surface temperature with topographic effect correction from Landsat 8 thermal infrared data in mountainous areas. IEEE Trans. Geosci. Remote Sens. 2020, 59, 6674–6687. [Google Scholar] [CrossRef]
  45. Yi, L.; Zhang, G. Object-oriented remote sensing imagery classification accuracy assessment based on confusion matrix. In Proceedings of the 2012 20th International Conference on Geoinformatics, Hong Kong, China, 15–17 June 2012; pp. 1–8. [Google Scholar]
  46. Shu, H.; Jiang, S.; Zhu, X.; Xu, S.; Tan, X.; Tian, J.; Xu, Y.N.; Chen, J. Fusing or filling: Which strategy can better reconstruct high-quality fine-resolution satellite time series? Sci. Remote Sens. 2022, 5, 100046. [Google Scholar] [CrossRef]
  47. Knauer, K.; Gessner, U.; Fensholt, R.; Kuenzer, C. An ESTARFM fusion framework for the generation of large-scale time series in cloud-prone and heterogeneous landscapes. Remote Sens. 2016, 8, 425. [Google Scholar] [CrossRef]
  48. Yan, W.; He, Y.; Cai, Y.; Qu, X.; Cui, X. Relationship between extreme climate indices and spatiotemporal changes of vegetation on Yunnan Plateau from 1982 to 2019. Glob. Ecol. Conserv. 2021, 31, e01813. [Google Scholar] [CrossRef]
  49. Shi, Z.; Sha, Y.; Liu, X.; Xie, X.; Li, X. Effect of marginal topography around the Tibetan Plateau on the evolution of central Asian arid climate: Yunnan–Guizhou and Mongolian Plateaux as examples. Clim. Dyn. 2019, 53, 4433–4445. [Google Scholar] [CrossRef]
  50. Liu, H.; Zhou, B.; Bai, Z.; Zhao, W.; Zhu, M.; Zheng, K.; Yang, S.; Li, G. Applicability assessment of multi-source DEM-assisted InSAR deformation monitoring considering two topographical features. Land 2023, 12, 1284. [Google Scholar] [CrossRef]
  51. Borini Alves, D.; Montorio Llovería, R.; Pérez-Cabello, F.; Vlassova, L. Fusing Landsat and MODIS data to retrieve multispectral information from fire-affected areas over tropical savannah environments in the Brazilian Amazon. Int. J. Remote Sens. 2018, 39, 7919–7941. [Google Scholar] [CrossRef]
  52. Hilker, T.; Wulder, M.A.; Coops, N.C.; Seitz, N.; White, J.C.; Gao, F.; Masek, J.G.; Stenhouse, G. Generation of dense time series synthetic Landsat data through data blending with MODIS using a spatial and temporal adaptive reflectance fusion model. Remote Sens. Environ. 2009, 113, 1988–1999. [Google Scholar] [CrossRef]
  53. Wu, M.; Niu, Z.; Wang, C.; Wu, C.; Wang, L. Use of MODIS and Landsat time series data to generate high-resolution temporal synthetic Landsat data using a spatial and temporal reflectance fusion model. J. Appl. Remote Sens. 2012, 6, 063507. [Google Scholar]
Figure 1. Landsat-8 true-color images of the study area located in the transition zone of the Yunnan–Guizhou Plateau.
Figure 1. Landsat-8 true-color images of the study area located in the transition zone of the Yunnan–Guizhou Plateau.
Remotesensing 17 01852 g001
Figure 2. Input data and validation data used in this paper. MODIS (a) and Landsat (c) data acquired on 16 July 2023, MODIS data acquired on 18 September 2023 (b), and actual Landsat images observed (d). The images presented by MODIS superimposed on the NIR-SWIR-Blue band of Landsat-8 were selected, where the NIR band is shown in red, the short-wave infrared (SWIR) band is shown in green, and the blue band is shown in blue. In the superimposed bands, the vegetation appears magenta in NIR-SWIR-Blue, the soil appears red, the overfire area appears dark, the clouds appear white, and the smoke particles’ scatter blue smoke in blue color.
Figure 2. Input data and validation data used in this paper. MODIS (a) and Landsat (c) data acquired on 16 July 2023, MODIS data acquired on 18 September 2023 (b), and actual Landsat images observed (d). The images presented by MODIS superimposed on the NIR-SWIR-Blue band of Landsat-8 were selected, where the NIR band is shown in red, the short-wave infrared (SWIR) band is shown in green, and the blue band is shown in blue. In the superimposed bands, the vegetation appears magenta in NIR-SWIR-Blue, the soil appears red, the overfire area appears dark, the clouds appear white, and the smoke particles’ scatter blue smoke in blue color.
Remotesensing 17 01852 g002
Figure 3. Flowchart of NBR index suitable for the spatio-temporal convergence analysis of karst landforms. Data preparation, spatial and temporal fusion, and accuracy assessment are shown from top to bottom.
Figure 3. Flowchart of NBR index suitable for the spatio-temporal convergence analysis of karst landforms. Data preparation, spatial and temporal fusion, and accuracy assessment are shown from top to bottom.
Remotesensing 17 01852 g003
Figure 4. Results of the fusion of the FSDAF model used in this paper in a single band (NIR and SWIR) with the vegetation index NBR. Here, (ac) represent the original images from 18 September 2023, including the NBR index along with the SWIR and NIR bands, respectively. The remaining images (df) are their corresponding predicted images.
Figure 4. Results of the fusion of the FSDAF model used in this paper in a single band (NIR and SWIR) with the vegetation index NBR. Here, (ac) represent the original images from 18 September 2023, including the NBR index along with the SWIR and NIR bands, respectively. The remaining images (df) are their corresponding predicted images.
Remotesensing 17 01852 g004
Figure 5. Results of the fusion of the STARFM model used in this paper in a single band (NIR and SWIR) with the vegetation index NBR. Here, (ac) represent the original images from 18 September 2023, including the NBR index along with the SWIR and NIR bands, respectively. The remaining images (df) are their corresponding predicted images.
Figure 5. Results of the fusion of the STARFM model used in this paper in a single band (NIR and SWIR) with the vegetation index NBR. Here, (ac) represent the original images from 18 September 2023, including the NBR index along with the SWIR and NIR bands, respectively. The remaining images (df) are their corresponding predicted images.
Remotesensing 17 01852 g005
Figure 6. Results of the fusion of the STDFA model used in this paper in a single band (NIR and SWIR) with the vegetation index NBR. Here, (ac) represent the original images from 18 September 2023, including the NBR index along with the SWIR and NIR bands, respectively. The remaining images (df) are their corresponding predicted images.
Figure 6. Results of the fusion of the STDFA model used in this paper in a single band (NIR and SWIR) with the vegetation index NBR. Here, (ac) represent the original images from 18 September 2023, including the NBR index along with the SWIR and NIR bands, respectively. The remaining images (df) are their corresponding predicted images.
Remotesensing 17 01852 g006
Figure 7. The FSDAF, STARFM, and STDFA models use the NBR results predicted by Strategies I and II, respectively, to compare with the actual NBR. For example, (a,b) are the prediction results of the FSDAF model using Strategy I and Strategy II for respectively, (c,d) are the prediction results of the STARFM model using Strategy I and Strategy II for respectively and (e,f) are the prediction results of the STDFA model using Strategy I and Strategy II for respectively.
Figure 7. The FSDAF, STARFM, and STDFA models use the NBR results predicted by Strategies I and II, respectively, to compare with the actual NBR. For example, (a,b) are the prediction results of the FSDAF model using Strategy I and Strategy II for respectively, (c,d) are the prediction results of the STARFM model using Strategy I and Strategy II for respectively and (e,f) are the prediction results of the STDFA model using Strategy I and Strategy II for respectively.
Remotesensing 17 01852 g007
Figure 8. Hybrid Strategy fusion result plots. Figures (ac) show the NBR results predicted by the FSDAF, STARFM, and STDFA models using the Hybrid Strategy compared with the actual NBR, respectively.
Figure 8. Hybrid Strategy fusion result plots. Figures (ac) show the NBR results predicted by the FSDAF, STARFM, and STDFA models using the Hybrid Strategy compared with the actual NBR, respectively.
Remotesensing 17 01852 g008
Figure 9. Fire boundary extraction results superimposed on the heterogeneous validation dataset results. (a) Extracted fire traces based on multivariate satellite data fusion using the thresholding method for change detection, and (b) results of sample blocks obtained through random sampling, superimposed on heterogeneous satellite validation datasets. The red, green and yellow points in the figure represent official fire point data from Himawari-8, VIIRS and Sentinel-2 satellites, respectively.
Figure 9. Fire boundary extraction results superimposed on the heterogeneous validation dataset results. (a) Extracted fire traces based on multivariate satellite data fusion using the thresholding method for change detection, and (b) results of sample blocks obtained through random sampling, superimposed on heterogeneous satellite validation datasets. The red, green and yellow points in the figure represent official fire point data from Himawari-8, VIIRS and Sentinel-2 satellites, respectively.
Remotesensing 17 01852 g009
Table 1. Landsat and MODIS data product band information used in this paper.
Table 1. Landsat and MODIS data product band information used in this paper.
Data ProductsBandsSpatial Resolution/(m)Time Resolution/(day)
Landsat 8 Level 2-C2Band5 (NIR)3016
Band7 (SWIR2)3016
MODIS MOD09GQBand2 (NIR)2501
Band7 (SWIR2)5001
Table 2. Description of the experimental data.
Table 2. Description of the experimental data.
MethodInput DataValidation Data
Landsat-8MODISLandsat-8
BandResolutionShooting DateBandResolutionShooting DateBandResolutionShooting Date
STARFMNIR/SWIR230 m16-July-2023NIR/SWIR2500 m16-July-2023NIR/SWIR230 m18-September-2023
-18-Sepember-2023
FSDAFNIR/SWIR230 m16-July-2023NIR/SWIR2500 m16-July-2023NIR/SWIR230 m18-Sepember-2023
-18-Sepember-2023
STDFANIR/SWIR230 m16-July-2023NIR/SWIR2500 m16-July-2023NIR/SWIR230 m18-Sepember-2023
-18-Sepember-2023
Table 3. Spectral accuracy indicators for different bands. The bolded values in the table are those with the highest precision.
Table 3. Spectral accuracy indicators for different bands. The bolded values in the table are those with the highest precision.
MethodAccuracy Indicators for Different Bands in the Spectral Dimension
NIRSWIRNBR
RMSEPSNRADRMSEPSNRADRMSEPSNRAD
FSDAF0.111123.80790.10410.018139.10970.00422.518571.64860.1475
STARFM0.141021.74130.10040.055729.3322−0.01658.168161.42900.2118
STDFA0.240717.09500.21170.061528.47540.81502.521871.63740.2901
Table 4. Spatial accuracy indicators for different bands. The bolded values in the table are those with the highest precision.
Table 4. Spatial accuracy indicators for different bands. The bolded values in the table are those with the highest precision.
MethodAccuracy Indicators for Different Bands in the Sptial Dimension
NIRSWIRNBR
EdgeLBPSSIMEdgeLBPSSIMEdgeLBPSSIM
FSDAF0.82260.99980.88430.82690.99980.96480.00000.99981.0000
STARFM0.19870.99980.58670.07860.99990.73030.00000.99991.0000
STDFA0.20110.99970.51220.07520.99970.70350.00000.99961.0000
Table 5. Accuracy indicators for different fusion strategies. The bolded values in the table are those with the highest precision.
Table 5. Accuracy indicators for different fusion strategies. The bolded values in the table are those with the highest precision.
MethodStrategyAccuracy Indicators
RMSEPSNRADEdgeLBPSSIM
FSDAFStrategy I0.174994.8161−0.15861.00000.99991.0000
Strategy II2.518571.64860.14750.00000.99981.0000
Hybrid Strategy0.059898.33480.24861.00000.99991.0000
STARFMStrategy I2.843570.5945−0.02460.00001.00001.0000
Strategy II8.168161.42900.21180.00000.99991.0000
Hybrid Strategy1.752671.59820.14690.00001.00001.0000
STDFAStrategy I2.852670.5668−0.19880.00000.99991.0000
Strategy II2.521871.63740.29010.00000.99961.0000
Hybrid Strategy3.658472.85240.34870.00000.99991.0000
Table 6. Confusion matrix constructed by overlaying the validated fire point dataset based on the fire trace extraction results in this paper. Statistics true positive (TP), false positive (FP), false negative (FN), and true negative (TN).
Table 6. Confusion matrix constructed by overlaying the validated fire point dataset based on the fire trace extraction results in this paper. Statistics true positive (TP), false positive (FP), false negative (FN), and true negative (TN).
Confusion MatrixPrediction
PositiveNegative
ReferencePositiveTP = 162FN = 18
NegativeFP = 72TN = 806
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Zhao, J.; Chen, G.; Wang, T.; Wang, Q.; Wang, K.; Miao, T. Spatio-Temporal Fusion of Landsat and MODIS Data for Monitoring of High-Intensity Fire Traces in Karst Landscapes: A Case Study in China. Remote Sens. 2025, 17, 1852. https://doi.org/10.3390/rs17111852

AMA Style

Zhang X, Zhao J, Chen G, Wang T, Wang Q, Wang K, Miao T. Spatio-Temporal Fusion of Landsat and MODIS Data for Monitoring of High-Intensity Fire Traces in Karst Landscapes: A Case Study in China. Remote Sensing. 2025; 17(11):1852. https://doi.org/10.3390/rs17111852

Chicago/Turabian Style

Zhang, Xiaodong, Jingyi Zhao, Guanzhou Chen, Tong Wang, Qing Wang, Kui Wang, and Tingxuan Miao. 2025. "Spatio-Temporal Fusion of Landsat and MODIS Data for Monitoring of High-Intensity Fire Traces in Karst Landscapes: A Case Study in China" Remote Sensing 17, no. 11: 1852. https://doi.org/10.3390/rs17111852

APA Style

Zhang, X., Zhao, J., Chen, G., Wang, T., Wang, Q., Wang, K., & Miao, T. (2025). Spatio-Temporal Fusion of Landsat and MODIS Data for Monitoring of High-Intensity Fire Traces in Karst Landscapes: A Case Study in China. Remote Sensing, 17(11), 1852. https://doi.org/10.3390/rs17111852

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop