1. Introduction
Potatoes are recognized for their reduced greenhouse gas emissions and water requirements, conferring substantial environmental advantages over other key agricultural crops [
1]. In light of these benefits, China has implemented initiatives to establish potatoes as a dietary staple, with the goal of reducing the disparity in production yields in comparison to leading producers. It is anticipated that these measures could lead to a reduction in the carbon-land-water impact of staple crops by 17–25% by 2030 [
1,
2]. Late blight, induced by the pathogen Phytophthora infestans, stands as a principal disease affecting potato cultivation, resulting in considerable losses in yield [
3,
4], and thus, impacting the role of potatoes in sustainable development endeavors. Consequently, extensive mapping of the prevalence of potato late blight (PLB) is essential for crisis management, facilitating effective surveillance, and achieving optimal disease management strategies. However, traditional disease monitoring is primarily dependent on field surveys, which are laborious and time-consuming, and are not aligned with the real-time monitoring requirements for extensive regions [
5].
Given the limitations of traditional monitoring methods, advanced technologies are being increasingly explored. Previous studies have shown that, at the leaf organ level and under laboratory conditions, hyperspectral imaging technology can be used to achieve early detection of PLB, even in the invisible stage of leaf disease, and hyperspectral monitoring can be achieved non-destructively [
6,
7,
8]. After remote sensing spectroscopy proved to be useful for the study of PLB, subsequent studies mostly combined this with UAV for monitoring. For example, some studies have used drones for large-scale surveillance of PLB. One approach to assessing PLB severity involves using RGB images captured by drones [
9]. Another method evaluates the extent of late blight using high-resolution multispectral UAV images combined with machine learning (ML) algorithms [
10]. A deep learning model known as CropdocNet has been proposed for accurately and automatically diagnosing late blight from UAV hyperspectral images [
11]. Additionally, research by Sun et al. (2023) explores collaborative PLB monitoring using multiple drone sensors [
12]. These studies demonstrate the feasibility of using remote sensing spectral information combined with ML or deep learning algorithms to detect PLB [
12]. These methods include the use of multi-layer perceptrons, convolutional neural networks (CNNs), support vector regression (SVR), and Random Forest (RF) applied to multi-spectral UAV imagery. The results indicate that ML algorithms can achieve high accuracy, suggesting their potential to replace traditional visual estimation methods with an acceptable mean absolute error of 11.72% [
12]. Despite the success of UAV technology and ML algorithms, their ability for large-scale disease monitoring remains limited in continuity and coverage. In contrast, satellite remote sensing offers an economical and efficient method for large-scale characterization of ground objects and long-term change detection [
13], making it a valuable tool for disease monitoring. The existing studies of PLB monitoring by satellite remote sensing have confirmed that the relationship between spectrum and PLB has scale variability, but is feasible and has great potential for the use of Sentinel-2 multispectral monitoring of PLB [
14].
Furthermore, crop distribution data serve as the foundational information for precise disease monitoring. Currently, there are few methods and models for mapping potato distribution; nevertheless, inspiration can be drawn from summarizing identification algorithms for other crops [
15,
16,
17,
18,
19]. The Google Earth Engine (GEE) platform, in conjunction with Sentinel-1/2 data, effectively utilized phenology data and ML techniques to create continuous, inter-annual distribution maps of oilseed rape with a 10-m resolution [
20]. Object-based methods and deep learning algorithms were employed to provide a novel perspective on the mapping of crop types in the northwest Ardabil region of Iran [
21]. A multi-temporal Gaussian mixture model developed on the GEE platform was applied to efficiently map maize distributions [
22]. Higher resolution of corn and soybean mapping in China [
23] was achieved by PlanetScope data and a RF-based method. These successful cases mainly utilize data from optical satellites, which offer high spatial and temporal resolution, covering the visible, near-infrared (NIR), and shortwave-infrared (SWIR) bands. These bands can capture different spectral characteristics of crops, which are helpful for crop classification and monitoring [
24,
25].
While optical remote sensing images have garnered considerable attention, the full potential of Synthetic Aperture Radar (SAR) technology remains largely untapped. SAR sensors can obtain images under various weather conditions, making them ideal tools for long-term and multi-seasonal monitoring. In addition, SAR holds extensive potential for various applications in agricultural monitoring [
26], for it can provide important information about crop growth and soil moisture conditions [
25]. Variations in radar backscatter during different growth seasons can also aid in detecting crop types, as different crop growth stages produce distinct scattering mechanisms, including surface scattering, corner scattering, and volume scattering. Notably, the RADARSAT satellite derived C-band SAR data have exhibited superior capability in differentiating rice and potato crops within the Ganges Plain. This methodology has attained remarkable classification accuracy, reaching 94% for rice and 93% for potatoes during the initial phases of their developmental cycle [
27]. By fusing multi-temporal Sentinel-1 SAR data and Sentinel-2 optical data, and applying the 3D U-Net deep learning network, scientists have significantly improved the accuracy of crop type mapping, providing an efficient crop monitoring and classification method for data-driven agriculture [
28]. SAR and multispectral data have been successfully used to extract crop-specific land phenology from major crops in Europe. Validation against ground truth data confirmed the accuracy, demonstrating the complementary nature of these satellite data types in delivering crop phenology insights, which bolsters the precision and dependability of crop surveillance [
29]. These successful cases of crop mapping using time series multi-source data and phenological information provide solid foundations for PLB distribution mapping [
30,
31].
Moreover, conventional techniques for processing remote sensing data for large-scale analyses are intricate, involving complexities across spatial and temporal dimensions [
32]. The task of large-scale crop mapping entails the processing and handling of extensive volumes of diverse satellite imagery sourced from multiple sensors, which poses considerable challenges associated with ’big data’. As remote sensing technology advances, monitoring crop disease using large volumes of satellite imagery has become feasible, especially with the emergence of the GEE platform. GEE provides powerful capabilities for big data analysis and access to high-resolution imagery on a global scale. It has shown excellent performance in accessing remote sensing products and pre-processing through the cloud platform [
33,
34], enabling extensive surveillance and cartographic representation of PLB [
35]. The unrestricted access to satellite-based global observations offers a supplementary method to consistently acquire large-scale PLB information at a low cost, both spatially and temporally, providing a new solution for potato production management [
36,
37]. Despite previous success in crop monitoring and distribution mapping based on GEE satellite remote sensing technology, there is limited research on the identification and mapping of large-scale crop disease. Most existing studies rely on UAV-scale hyperspectral and multispectral modeling for PLB monitoring, thereby restricting the utilization of reliable and freely available satellite data for large-scale and efficient disease monitoring.
At present, there is a growing trend in the utilization of remote sensing technology for agricultural monitoring, but challenges remain in large-scale detection and spatial visualization of PLB [
14,
18]. Moreover, there is no remote sensing-based potato distribution mapping in China, and only limited research on large-scale PLB monitoring methods. This study aims to develop a large-scale PLB monitoring method based on GEE to improve the accuracy and monitoring efficiency of disease identification, so as to provide real-time information support for agricultural disease management. The specific objectives are (1) precise identification of potato planting areas by using multi-source time series remote sensing data and K-Means clustering algorithm; (2) evaluation of the application effect of Sentinel-1 and Sentinel-2 data fusion in disease surveillance to reveal the complementarity of different data sources; and (3) provision of PLB monitoring framework based on machine learning algorithm, in order to provide a scientific basis for sustainable agricultural development and to reduce the yield loss caused by PLB.
3. Results
3.1. Spatial Clustering Analysis Based on K-Means Algorithm and Sentinel 1/2 Time-Series Data
The monthly time series data from Sentinel-1 and Sentinel-2 satellites were processed within the study area throughout 2021. The K-Means clustering algorithm was utilized to perform spatial clustering analysis on the time series data, categorizing them into 20 unique clusters within the study area, as depicted in
Figure 3. The final clustering result curves illustrate how the cluster centers evolve during the iterative process and how data points progressively stabilize within specific clusters (
Figure 4). The clustering results were interpreted by examining the various vegetation conditions and growth states, as detailed in
Table A1 and
Table A2.
Table A3 integrates the results of the NDVI change analysis (
Table A1) with those from the VH band analysis (
Table A2), which are combined as different land cover types.
Comprehensively considering the land cover classification map and the time series clustering graph of VH and NDVI, we can carry out a more in-depth analysis and understanding.
1. Correlation between land cover type and VH/NDVI value:
Urban areas may show stable low values in the VH chart (around −25.5 dB) because artificial surfaces reflect less radar waves. At the same time, in the NDVI map, these areas may show low values because there is less vegetation cover. Agricultural regions may show seasonal fluctuations in the VH map, reflecting the effects of irrigation and crop growth. In the NDVI map, these areas may show high values during the growing season. Forest areas are likely to show high values in both maps, with higher VH values, because the forest has a strong reflection of radar waves and higher NDVI values because the forest has rich vegetation cover.
2. Seasonal changes:
The time series maps of both VH and NDVI show significant seasonal variations, which echo the color distribution in the land cover classification maps. For example, an increase in VH values in the summer may correspond to the expansion of green areas in the land cover map, which may represent areas that are heavily vegetated in the summer. The peak of NDVI occurs in spring and summer, which is consistent with the vegetation growth cycle shown in the land cover map.
3. Impact of human activities:
Agricultural irrigation and crop growth are clearly reflected in both VH and NDVI maps. Seasonal fluctuations in VH values may match the distribution of agricultural land in the land cover map. The influence of urbanization on VH and NDVI is also obvious. The urban area shows a stable low value in the VH map and a low value in the NDVI map, which is consistent with the distribution of the urban area in the land cover map. These charts and data offer insights into land cover and vegetation changes, aiding potato field identification.
3.2. Potato Identification Model Based on Spatial Clustering and Multi-Indices
Spatial clustering results identified using multi-source time-series remote sensing features and the K-Means clustering algorithm were used in combination with measured potato distribution samples (based on the
Section 2.4.2 and
Section 3.1) to finally determine the distribution of potatoes in 2021 (
Figure 5). The F1 score for the model in monitoring the potato distribution area was 0.95, as detailed in
Table 3. This score, being close to 1, signifies that the model was adept at identifying the majority of true-positive samples while keeping the false alarm rate low. The analysis of the experimental outcomes demonstrated that the integration of the K-Means algorithm with the time-series data from Sentinel-1/2 was highly effective for spatial cluster analysis.
Four scenes were selected to illustrate the spatial details of the classification. The first row in
Figure 6 shows real Google Earth images, while the second row displays the results of potato classification, with the green area representing the model’s identified distribution area of potatoes and black dots indicating the ground-measured distribution points.
Figure 6a,b provides a detailed comparison of the classification results at different spatial resolutions: (a) shows the overall distribution of potato fields at a low spatial resolution, capturing broad patterns; while (b) highlights the model’s ability to identify smaller, more dispersed fields at a high spatial resolution, demonstrating its versatility in both broad and detailed classifications.
Figure 6c shows the algorithm’s classification performance in mountainous regions, where errors are primarily due to terrain-induced shadows.
Figure 6d displays the algorithm’s classification of areas with towns and water bodies, noting some remaining fragmented areas. Nevertheless, the algorithm effectively distinguishes between water bodies and vegetation near water, despite the similar morphological color characteristics of some farmland and water bodies in RGB images. By comparing with Google Earth images and field survey data, we find that the classification accuracy of the K-Means method is slightly decreased in some edge regions and mixed pixel regions, but the overall effect is small.
3.3. PLB Separability Results Based on Euclidean Distance
In this study, the Euclidean distance between the Sentinel-2 bands and PLB severity was used to quantify the distinguishability of spectral features. The color gradient of the heatmap ranges from white (low) to blue (high), representing the magnitude of the distances (
Figure 7). The Euclidean distance matrix shows the distance between different Sentinel-2 bands. A smaller distance indicates that the two bands are more similar in terms of spectral features; a larger distance indicates that they are more different in terms of spectral features. As shown in
Figure 7, in general, the spectra were more effective in distinguishing between different PLB severities when there was a large difference in disease severity. Moreover, it was observed that the spectral difference between PLB 85 and other PLB < 85 was particularly large.
3.4. The Characteristics and Importance of the PLB Surveillance Model
Feature importance analysis in
Figure 8 shows that the top-ranked features in the model were mainly concentrated in the fourth month (August), which suggests that this period may be a critical stage for disease development, where physiological and morphological features of the vegetation are critical for PLB prediction. In particular, near-infrared bands (e.g., B8_3 and B8A_3) shows consistently high importance throughout the time series, highlighting the key role of vegetation biomass and health in disease prediction. The SAVI_3 and NDWI_3 indices are crucial in the fourth month, highlighting the role of vegetation water status in disease progression. The B2_1 band’s importance in the first month indicates early disease spectral characteristics for early detection. Other significant indicators include B7_3, B6_3, and DWSI3_3, which may relate to vegetation moisture and physiological activity. Vegetation texture features, like B8_savg_3, show the effects of disease on structure. However, features like MSK_CLDPRB, MSK_SNWPRB, and NDBI had zero importance at various times, suggesting they were less relevant for prediction.
3.5. Performance of PLB Surveillance Model
Table 4 compares seven models: CART, GTB, RF, TS–RF, STS–RF, MSTS–GTB, and MSTS–RF. The comparison results of the three machine learning models show that RF performs best in PLB monitoring and can be used as the main model for PLB monitoring research, followed by GTB. It can be seen from the experimental results that the STS–RF method using single source radar data has poor performance on the verification set, and the verification for R
2 is only 0.10, indicating that its prediction ability for new data is weak. In contrast, the TS–RF method using single-source optical data performed well on both the training set and the validation set, with a validation R
2 of 0.54, indicating that it has some predictive power for new data. MSTS–RF, which combines multi-source data with time series, has the best performance on multiple evaluation indexes, the lowest RMSE and the highest R
2, and has good prediction accuracy and generalization ability. These results suggest that MSTS–RF be prioritized as the final model, as it not only performs well in the training phase but also has the best adaptability to new data in actual deployments. The MSTS–RF model not only has high fitting accuracy to the training data but also has good generalization ability for the unseen validation data. In contrast, the MSTS–GTB model performed well on the training data (RMSE 16.53, R
2 0.79), but poorly on the validation data (RMSE 30.50, R
2 0.45).
The training set of the MSTS–RF model displayed in
Figure 9 had good monitoring performance. The RMSE value was 14.69, which indicated the deviation between the percentage of PLB monitored by the model and the actual observations. The R² value was 0.89, which suggests that the model explains 89% of the variation in the actual observations.
Figure 10 shows the performance of the PLB monitoring model on an independent validation set. The validation set provided a more rigorous test of the model’s generalization ability than the training set. The RMSE on the validation set was 20.50, which was higher than the RMSE of 14.69 on the training set. The R²on the validation set was 0.71, which was lower than the R
2 of 0.89 on the training set.
From the learning curve (
Figure 11), we can observe the following: With the increase in the proportion of the training set, the RMSE of the training set decreases gradually, which indicates that the model fits the training data better and better. The RMSE of the validation set is the lowest when the ratio of training set is 0.70, which indicates that the generalization ability of the model is the strongest at this time. When the proportion of the training set decreases further, the RMSE of the validation set increases significantly, indicating that the generalization ability of the model decreases.
The R² of the training set is high at all scales, indicating that the model fits well on the training data. The R² of the validation set reaches the highest when the ratio of the training set is 0.70, indicating that the generalization ability of the model is the strongest at this time. When the proportion of the training set is reduced to 0.10, the R² of the validation set decreases significantly, indicating that the generalization ability of the model decreases significantly.
From these learning curves, we can draw the following conclusions: Optimal data partition ratio: When the training set ratio is 0.70, the model has the best performance on the validation set, the lowest RMSE and the highest R².
Overfitting and underfitting: When the proportion of the training sets is too high or too low, the model is prone to overfitting or underfitting. When the training set ratio is 0.90, the model performs well on the training set, but the performance on the validation set decreases, indicating overfitting. When the training set ratio is 0.10, the performance of the model on the validation set decreases significantly, indicating underfitting.
3.6. The Distribution of PLB Monitoring Results
This figure (
Figure 12) shows the distribution of pixel values of PLB severity detected by the surveillance model. The histogram shows that the distribution of PLB values ranges from 4.75 to 68.25, indicating that the monitored PLB severity ranges from very mild to very severe. Most of the pixel values were clustered around 50, indicating that most of the monitored PLB severity was moderate. At the lower (near 4.75) and higher (near 68.25) ends of the PLB values, the number of pixels is lower, indicating that extremely mild or extremely severe PLB conditions are less common. Since most of the predicted values are concentrated in the early and middle stages, this suggests that the monitoring model has the potential to monitor PLB in its early stages, thus helping to take timely measures to control the development of the disease. Overall, this graph shows that the surveillance model is able to better identify moderate severity of PLB and monitor it at early stages, which is important for the management and control of potato late blight.
Figure 13 illustrates the severity of the disease monitored at various locations within the study area, with higher values denoting a greater disease risk as evaluated by the model. The distribution plot highlights substantial spatial heterogeneity in disease risk across the study area. High-risk areas, characterized by monitoring values nearing 80, tend to be clustered in particular geographical locations. This clustering may be associated with local environmental conditions, crop cultivation practices, or historical disease incidence patterns. The disease probability distribution map constitutes a potent tool for agricultural managers, facilitating the identification and prioritization of disease control measures. By targeting these high-risk areas, resources can be more efficiently allocated, thereby enhancing the efficacy and effectiveness of disease management strategies.
Most of the validation points in
Figure 14 have small errors, with the majority of errors concentrated in lower intervals (below 25). Points with larger errors tend to be more spatially homogeneous. With an RMSE of 26.52, these results suggest that the model performs well.