Next Article in Journal
Mechanical Properties of Frozen Loess Subject to Directional Shear Effects from Multiple Principal Stress Directions
Previous Article in Journal
Practical Design of Lattice Cell Towers on Compact Foundations in Mountainous Terrain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparative Assessment of Regular and Spatial Cross-Validation in Subfield Machine Learning Prediction of Maize Yield from Sentinel-2 Phenology

by
Dorijan Radočaj
*,
Ivan Plaščak
and
Mladen Jurišić
Faculty of Agrobiotechnical Sciences Osijek, Josip Juraj Strossmayer University of Osijek, Vladimira Preloga 1, 31000 Osijek, Croatia
*
Author to whom correspondence should be addressed.
Eng 2025, 6(10), 270; https://doi.org/10.3390/eng6100270
Submission received: 19 September 2025 / Revised: 5 October 2025 / Accepted: 8 October 2025 / Published: 9 October 2025

Abstract

The aim of this study is to determine the reliability of regular and spatial cross-validation methods in predicting subfield-scale maize yields using phenological measures derived by Sentinel-2. Three maize fields from eastern Croatia were monitored during the 2023 growing season, with high-resolution ground truth yield data collected using combine harvester sensors. Sentinel-2 time series were used to compute two vegetation indices, Enhanced Vegetation Index (EVI) and Wide Dynamic Range Vegetation Index (WDRVI). These features served as inputs for three machine learning models, including Random Forest (RF) and Bayesian Generalized Linear Model (BGLM), which were trained and evaluated using both regular and spatial 10-fold cross-validation. Results showed that spatial cross-validation produced a more realistic and conservative estimate of the performance of the model, while regular cross-validation overestimated predictive accuracy systematically because of spatial dependence among the samples. EVI-based models were more reliable than WDRVI, generating more accurate phenomenological fits and yield predictions across parcels. These results emphasize the importance of spatially explicit validation for subfield yield modeling and suggest that overlooking spatial structure can lead to misleading conclusions about model accuracy and generalizability.

1. Introduction

Understanding and predicting crop yield variability at the subfield scale is a cornerstone of precision agriculture [1]. Subfield yield variability arises from spatial heterogeneity in complex biotic and abiotic factors, including soil properties, topography, drainage, fertility, and microclimate conditions [2,3]. Traditional field-scale yield assessments overlook this fine-scale variability, limiting the effectiveness of input management in precision agriculture [4]. Subfield-level yield prediction improves several aspects of site-specific management, including variable rate fertilization, hybrid selection, irrigation planning, and pest and disease control [5]. Yield maps derived from remote sensing can identify consistently high- or low-yielding zones within fields, allowing for more effective resource allocation and risk mitigation [6]. Over time, these predictive maps support adaptive management by informing decisions in subsequent seasons. As climate variability increases and input costs rise, yield prediction using free satellite imagery becomes an essential tool for optimizing productivity and sustainability across diverse agricultural systems [7].
While combine harvester yield monitors provide accurate, high-resolution yield data, their use is limited by high equipment costs, data inconsistencies, and restricted availability, particularly in smallholder or resource-limited farming systems [8]. These systems often lack access to the advanced machinery and technical infrastructure required to implement such monitoring, making widespread adoption challenging [9]. Furthermore, yield data collected from combine harvesters can be affected by calibration errors, mechanical delays, and incomplete spatial coverage, particularly at field edges or in irregularly shaped plots [10]. In contrast, satellite-based remote sensing missions, such as Sentinel-2, provide a globally consistent, freely accessible, and reliable source of high-resolution imagery [11,12]. Sentinel-2 delivers multispectral data at up to 10 m spatial resolution, with revisit frequencies of 5 days at the equator and even more frequent at higher latitudes when combined with both Sentinel-2A and 2B satellites [13]. This allows for regular and timely monitoring of crop development and spatial variability within and between fields over the entire growing season [6]. Previous studies observed that vegetation indices derived from Sentinel-2 imagery correlate with crop vigor, canopy structure, and biomass [14,15]. Nevertheless, vegetation indices from individual satellite observations tend to reflect only immediate canopy conditions. Crop phenology modeling has become a particularly effective approach to determine temporal dynamics by fitting smooth curves on vegetation index time series to derive important seasonal transition values [16,17]. It has been demonstrated that phenology-based indicators using MODIS, Landsat, and Sentinel-2 enhance the understanding of space and time variations in vegetation development and yield [18]. When combined with machine learning methods, these metrics can support yield prediction at the subfield scale [19]. Although these models may be less precise than direct yield measurements from harvester monitors, they offer a highly scalable and cost-effective alternative for monitoring large agricultural areas without requiring expensive machinery or dense ground-based observations. This approach also supports temporal analysis, allowing for yield monitoring over multiple seasons, which is critical for long-term agricultural planning and resilience due to climate variability [20].
Machine learning methods in previous studies enabled accurate modeling of the complex relationships between crop yield and various remote sensing, environmental, and management variables [21]. However, the validity of model performance assessments heavily depends on the choice of cross-validation approach used during model development [22]. A commonly applied approach is regular (random) cross-validation, where training and testing data are randomly split without considering the spatial structure of the dataset [23]. While this method is dominantly used even for geospatial predictions, it can produce overly optimistic accuracy estimates in geospatial modeling contexts, particularly in agricultural landscapes characterized by strong spatial autocorrelation and underlying heterogeneity in soil, topography, and plant growth conditions [24]. Especially in precision agriculture, which is based on geospatial variability in crop-growing conditions, using inappropriate validation methods can lead to misleading conclusions about the real-world performance of crop yield prediction models [25]. Regular cross-validation often results in spatial leakage, where training and test samples are drawn from nearby or overlapping locations, thus artificially inflating predictive accuracy [26]. This undermines the reliability of the models when applied to unseen areas, such as different fields or growing seasons [27]. Despite this, many existing yield prediction studies do not explicitly address the implications of spatial structure in their validation protocols. Even with technological advances in satellite-based phenological modeling, there remains a critical methodological issue, as the spatial autocorrelation of training data can skew the model validation results [28]. The standard (regular) cross-validation randomly partitions data without respect to spatial structure and can commonly lead to spatial leakage, whereby neighboring samples are observed in the training data and test data [29]. Spatial cross-validation techniques might be used to solve this problem by dividing data into geographically different folds, providing an improved generalization performance [23]. There is a research gap in understanding how the performance of crop yield prediction models changes when evaluated under spatial cross-validation, which partitions data to ensure spatial independence between training and test sets. Spatial cross-validation provides a more realistic assessment of generalization performance and is theoretically more suitable for geospatial applications [30,31].
The main aim of this study was to address this research gap by comparing the performance of machine learning yield prediction models under both regular and spatial cross-validation, highlighting the consequences for accuracy assessment and practical deployment in precision agriculture.

2. Materials and Methods

The workflow of the study included the following: (1) creation and preprocessing of ground truth maize yield samples from combine harvester yield mapping sensors; (2) creation of Sentinel-2 time series data for each yield sample and calculation of vegetation indices; (3) phenological modeling of vegetation index time series; (4) machine learning prediction of subfield maize yield based on phenological metrics as covariates; and (5) a comparison of accuracy assessment of evaluated yield prediction models according to regular and spatial cross-validation approaches.

2.1. Study Area and Ground Truth Maize Yield Data

The study area included three maize parcels located in the proximity of near Koška in Eastern Croatia (Figure 1). The Koška region in eastern Croatia is characterized by a continental climate with hot summers and moderately cold winters (Dfb according to Köppen climate classification), making it suitable for maize cultivation [32]. Annual precipitation averages 650–800 mm, with the majority occurring during the growing season (April–October) [33], while soils in the study area are predominantly clay loam [32]. Elevation ranges from 90 to 120 m above sea level, usually with minimal slope.
Maize yield samples were collected during harvest from 13 to 14 October 2023 using a Claas Lexion 6900 combine harvester (Harsewinkel, Germany) with a Quantimeter yield sensor and Claas Connect software v1.0 (Harsewinkel, Germany) [34], which was calibrated according to the standard procedure stated by the manufacturer. The maize yield was measured in t/ha. The combined area of the three analyzed parcels was 18.0 ha, with a total of 1566 collected georeferenced yield samples before preprocessing (Table 1). The preprocessing was performed using the paar package in R v4.5.0 [35] according to procedure outlined by Paccioretti et al. [36]. It included spatial-statistical filtering in three steps: (1) edge removal, which excluded removing samples distanced 10 m or less from parcel boundaries to reduce border effects from mechanical operations; (2) outlier removal, which identified and removed extreme values outside three standard deviations from the mean yield value per parcel; and (3) inlier removal, which detected and excluded spatially inconsistent values that deviate from local trends quantified by the local Moran index. Additionally, the Shapiro–Wilk normality test was performed for all input samples, suggesting that only yield samples from Parcel 3 had normal distribution, while the coefficient of variation was notably lower for all three parcels after preprocessing.

2.2. Sentinel-2 Time Series Data and Vegetation Indices

Sentinel-2 bottom-of-atmosphere (BOA) Level-2A surface reflectance data [13] were acquired using Google Earth Engine to generate time series for maize yield samples between 1 January 2023 and 31 December 2023. Only scenes with less than 50% cloud cover were retained, while cloud, snow, shadow, and cirrus effects were mitigated using pixel-level quality information, including the MSK_CLDPRB, MSK_SNWPRB, and SCL bands. Pixels with cloud or snow probability higher than 5% or being classified as cloud shadow (SCL = 3) or cirrus (SCL = 10) were removed from the time series. The cleaned image collection was mapped to sample reflectance values for bands B1–B12 at a 10 m spatial resolution, with each observation being tagged with the acquisition date. After preprocessing, surface reflectance values from bands B2, B4, and B8 were used to compute the Enhanced Vegetation Index (EVI) and the Wide Dynamic Range Vegetation Index (WDRVI). Previous studies noted that EVI and WDRVI are well-suited for maize yield prediction due to their enhanced sensitivity to high-biomass conditions and improved performance in dense canopy environments typical of maize during peak growth [37]. EVI minimizes atmospheric and soil background effects [38], while WDRVI effectively distinguishes subtle changes in vegetation vigor during key phenological stages due to its increased resistance to saturation effect due to high biomass [39]. EVI and WDRVI were calculated according to Equations (1) and (2) as follows:
EVI = 2.5 × NIR R NIR + 6 R 7.5 B + 1 ,
WDRVI = 0.1 NIR R 0.1 NIR + R ,
where B—BOA surface reflectance in blue band (Band 2); R—BOA surface reflectance in red band (Band 4); and NIR—BOA surface reflectance in near-infrared band (Band 8).

2.3. Phenological Modeling

For each parcel and vegetation index, time series were separated by unique point identifiers and converted into daily observations. Phenological modeling was performed using the phenofit v0.3.10 R package [40], which provides a comprehensive framework for smoothing, fitting, and extracting seasonal dynamics from vegetation index time series. It supports multiple curve-fitting methods and is designed to handle irregular or noisy data, making it well-suited for remote sensing-based phenology analysis [41]. The workflow included temporal smoothing using the Whittaker smoother and phenological curve fitting using six methods, including Asymmetric Gaussian (AG), Beck, Elmore, Zhang, Klos, and Gu. The lambda parameter in the Whittaker smoother was calculated individually from cross-validation and differed for each sample, while the weight updating function used was the method used in TIMESAT, according to [40].
For each fitted curve, eleven phenological stages were extracted, including start of season, peak of season, end of season, greenup, maturity, senescence, dormancy, and transitional points, including upturn date, stabilization date, downturn date, and recession date. For each stage, the modeled vegetation index value and corresponding calendar date were calculated. The total number of Sentinel-2 observations per fit and goodness-of-fit metrics, including coefficient of determination, root mean square error, and Nash-Sutcliffe efficiency coefficient (NSE), were calculated for each model to assess fitting quality [41]. In cases of detection of multiple phenological cycles during 2023, the model with the longest vegetative period, quantified with the difference between the end of the season and the start of the season, was selected to ensure robust extraction of maize seasonal dynamics. In the study area, the maize growing season typically spans from April to October and thus covers most of the calendar year [42].

2.4. Machine Learning Prediction and Accuracy Assessment Using Regular and Spatial Cross-Validation

The evaluation of machine learning methods for subfield maize yield prediction based on phenological metrics was performed using the caret package in R [43]. The input dataset consisted of phenological metrics, which included transition dates converted to day of year (DOY) and corresponding vegetation index values, with a total of 22 covariates per model. For each combination of field parcel, vegetation index, and phenological fitting method, two machine learning methods were evaluated, including Random Forest (RF) from the ranger R package [44] and Bayesian Generalized Linear Model (BGLM) from the arm R package [45]. Before modeling, input data were cleaned to remove outliers in yield after phenological modeling and standardized. Hyperparameter tuning for each model was performed using a random search in 10 repetitions, and optimal hyperparameters were selected with the criterion of lowest root mean square error among repetitions. The hyperparameter ranges for random search during tuning were 2–22 for mtry, “variance” and “extratrees” for splitrule, and [1, 3, 5, 7, 10] for min.node.size for RF, while all hyperparameters for BGLM were selected from the prior scale and thus were not tuned.
Model training and accuracy assessment were performed using 10-fold cross-validation in five repetitions to reduce variance and improve model stability. To provide a comparative assessment for general and spatial aspects of data structure, both regular and spatial cross-validation were evaluated. Regular folds were determined randomly, while spatial folds were determined using the CAST R package [46], based on k-means clustering on E and N coordinates of yield samples in the Croatian Terrestrial Reference System (HTRS96/TM, EPSG: 3765), enabling evaluation across spatially distinct subfield sections (Figure 2). Input samples were clustered in 10 clusters to match the k-fold number of cross-validation, which was the sole parameter used for the creation of spatial folds. Wilcoxon tests were used to evaluate if there were significant differences in accuracy assessment metrics between regular and spatial cross-validation, using metrics from all calculated folds in the study.
Three accuracy assessment metrics were used for the evaluation of maize yield prediction using machine learning, including the coefficient of determination (R2), root-mean-square error (RMSE), and mean absolute error (MAE), which were calculated according to Equations (3)–(5) as follows:
R 2 = 1 i = 1 n y i y i 2 i = 1 n y i y ¯ 2 ,
R M S E = i = 1 n y i y i 2 n ,
M A E = i = 1 n y i y i n ,
where y i —actual crop yield; y i —predicted crop yield; y ¯ —mean of actual crop yield data; and n—sample count. The most accurate machine learning model for maize yield prediction had the highest R2 and lowest RMSE and MAE per parcel, vegetation index and phenological fitting method.

3. Results

All three parcels in the study area had their median value in the range of 8.5–9.3 t/ha (Figure 3). Additionally, subfield variability in maize yield is notable, as collected maize yields across the study area ranged from 5.5 to 12.2 t/ha, with the highest frequency in yield samples of around 8.6 t/ha for Parcel 1, around 9.5 t/ha for Parcel 2, and 8.3 t/ha for Parcel 3.
The accuracy assessment of the evaluated fitting methods for phenological modeling varied significantly between parcels and vegetation indices (Table 2). EVI produced notably higher phenological fitting accuracy than WDRVI due to higher number of successfully fitted samples, particularly in Parcel 1 and Parcel 3. At these parcels, very high R2 and NSE indicated high model agreement with observed data, with more than 0.92 and 0.91, respectively.
All six evaluated phenological fitting methods produced very similar fitting accuracy assessment metrics for both vegetation indices, with Gu, Klos, and Zhang being among the most accurate ones for Parcels 1 and 3 based on all three metrics. However, the main issue of these methods, in comparison to AG, Beck, and Elmore, was the number of successfully fitted samples during phenological modeling. For EVI, these three methods achieved percentages of successful fitting of 81.4%, 77.7%, and 75.5% in comparison to 34.6%, 29.3%, and 14.9% of Gu, Klos, and Zhang methods for Parcels 1, 2, and 3, respectively.
Table 3 presents the accuracy assessment metrics of the evaluated machine learning models for maize yield prediction across three parcels, vegetation indices, phenology fitting methods, and cross-validation approaches. The optimal hyperparameters for RF are presented in Table A1, while BGLM did not require hyperparameter tuning. Overall, the most accurate models according to regular cross-validation results indicated notably higher prediction accuracy than the most accurate models based on spatial cross-validation. This observation was the most pronounced for Parcel 1, but these results also might not be representative of maize yield value distribution for Parcels 1 and 2, in which machine learning models based on WDRVI produced the highest prediction accuracy but were trained on a very small number of available yield samples. Moreover, relative (R2) and absolute (RMSE, MAE) accuracy assessment metrics were dominantly in disagreement on the most accurate machine learning model per parcel. Wilcoxon tests comparing the accuracy assessment statistical metrics from regular and spatial cross-validation for evaluated machine learning models proved that there were significant differences in accuracy assessment metrics between regular and spatial cross-validation for all calculated folds (Table 4).
As for machine learning methods, both RF and BGLM produced the most accurate prediction models in three instances, with RF producing higher prediction accuracy on Parcel 2 for both regular and spatial cross-validation, with the same being applicable for BGLM on Parcel 3. Among evaluated phenological fitting methods, AG most frequently produced the most accurate prediction model per parcel, also producing the highest number of samples with successfully fitted phenological models, slightly outperforming Beck and Elmore methods, whose sample counts were in some cases reduced after outlier removal prior to machine learning prediction.
Figure 4 presents value ranges of accuracy assessment metrics per fold in 10-fold cross-validation across all five repetitions, quantifying the effect of randomness in training and test data split during cross-validation on machine learning prediction accuracy. The robustness of predictions increased with the total number of available yield samples per parcel and vegetation index, with EVI in Parcel 1 producing the most stable prediction accuracy values. On the contrary, while WDRVI produced the most accurate prediction models for Parcels 1 and 2, very low sample counts produced very high ranges in resulting accuracy assessment metrics, which was the most pronounced for R2.

4. Discussion

The smaller phenological fitting accuracy of Parcel 2 likely occurred not due to model inadequacy but to quality differences between data and local growing conditions. Local management variability in the form of hybrid type or variable nitrogen application could have changed canopy dynamics, weakening the temporal coherence of the vegetation index time series [47]. Moreover, sensor noise of yield monitors in this parcel might have added to the increased variance in the training data, as sampling density on Parcel 2 (42.7 samples per ha) was lower in comparison to Parcel 1 (65.7 samples per ha). This observation suggests that the additional property of input yield samples is a major determinant of phonological fitting accuracy and should be explored in future studies. The median yield of three parcels evaluated in the study is above the mean maize yield in the Pannonian biogeoregion of Croatia of 7.4 t/ha [48] but considerably below the maximum possible maize yield in the study area, which can reach up to 14 t/ha for some maize hybrids [33].
The inequality of evaluated phenological modeling methods in terms of fitting accuracy and the percentage of successfully fitted samples is a major observation on the future selection of fitting methods in crop yield prediction based on phenological metrics. AG, Beck, and Elmore methods ensured notably higher reliability in securing adequate yield sample count, thus providing more reliable options for yield prediction. Moreover, the potential trade-off of those methods against Gu, Klos, and Zhang’s methods in terms of fitting accuracy has only a minor effect, as it varied far less in comparison to the number of successfully fitted samples. While recent similar studies did not report phenological fitting accuracy per method, maximum achieved fitting accuracy was comparable [49] or slightly lower [18] than fitting results based on EVI time series from this study. The discrepancy between the fitting accuracy and amount of successfully fitted samples from the results of this study is disproportionate, suggesting that increased methods to variations in vegetation indices in time series data outweigh their slightly lower fitting accuracy. However, further studies based on increased ground truth data quantity and diverse study areas are required to confirm this observation.
The difference in predictive accuracy under regular cross-validation is attributable to spatial leakage [27], caused by the proximity of test samples relative to training samples, thus negating the assumption of independent and identically distributed data. In spatially structured environments, particularly in agricultural fields where soil and topographic characteristics typically vary gradually, neighboring samples tend to share similar phenological and yield characteristics [50]. Due to this occurrence, regular cross-validation was likely strongly affected by localized effects rather than generalized phenological maize properties, resulting in overestimated yield prediction accuracy. Due to the differences in prediction accuracy between regular and spatial cross-validation, it is very likely that these models exploited localized spatial effects caused by different micro-scale differences in soil management and microclimate that are correlated with maize yield. However, it does not demonstrate that evaluated machine learning models were incapable of learning the relationship between generalizable phenology and maize yield but that they are dependent on spatial dependence existing in the training data.
The previous studies dominantly focused on exploring the temporal component of cross-validation instead of the spatial, as a study by Croci et al. [51] emphasized temporal generalization using year-based splits, and, although their models showed good transferability across years, spatial heterogeneity was not accounted for. Moreover, Perich et al. [52] predicted winter wheat yields using Sentinel-2 time series and validated models using leave-one-year-out cross-validation, in which R2 values resulted up to 0.88, but a lower prediction accuracy was noted in years with atypical weather. However, Fernando et al. [53], who applied leave-one-field-out cross-validation for canola yield prediction, reported a lower R2 of 0.46, noting the possible magnitude in which regular cross-validation inflated metrics in spatially structured problems. Additionally, Crusiol et al. [54] compared different scales of model validation for soybean yield across 15 fields, reporting an R2 of 0.82 under global validation but only 0.32 when using field-specific models. This observation reinforces the notion that models trained on aggregated data may exploit spatial dependencies rather than capturing physiologically meaningful responses to environmental variation. As was observed from the results presented in Table 3, spatial cross-validation indicated that prediction accuracy was actually lower than regular cross-validation results suggest, but the display of achieved value ranges of all three accuracy assessment metrics also indicates that prediction results are more affected by the randomness in training and test data split, which is especially apparent for smaller datasets [55].
Previous studies [56,57] demonstrated the predictive ability of phenological metrics derived from Sentinel-2, such as normalized difference vegetation index (NDVI), normalized difference red-edge index (NDRE), or EVI for yield estimation. The results from this study confirmed their effectiveness in crop yield prediction, while their prediction accuracy metrics heavily depended on the cross-validation approach. This was partially noted by Darra et al. [56], which utilized repeated 5-fold random cross-validation to predict tomato yield and reported high explained variance, but without a spatially blocked validation, the generalization of these results remains untested. Furthermore, while the inclusion of temporal metrics enhanced model performance under interpolation scenarios, their behavior under spatial extrapolation remains mixed [58].
The main limitations of this study were caused by the relatively short temporal span and restricted study area. Longer-term studies covering broader spatial gradients, including multiple agroclimatic zones, are needed for better understanding of the scalability of predictive models based on Sentinel-2 time series for crop yield prediction. Furthermore, more comprehensive evaluation of machine learning methods, especially deep learning or transfer learning, might further improve the understanding of the generalization gap observed under spatial cross-validation. Another important direction for future studies is the inclusion of ancillary data representing environmental covariates, such as climate, soil, and topography data, which may improve performance under spatial cross-validation by capturing site-specific effects that are otherwise contained in spatial autocorrelation [59]. However, climate and soil properties also pose a challenge to geospatial modeling at the subfield scale, as soil data are frequently insufficiently spatially resolved to resolve fine-scale variability in soil texture, organic matter, or nutrient levels in small parcels, and their extrapolation to sparse samples may lead to significant errors. Likewise, topography- or micro-management-motivated microclimatic gradients cannot be reliably captured by climate data based on meteorological stations or coarse-resolution gridded products, which could be improved in future studies. Given that the combination of EVI time series with more robust phenological fitting methods like AG, Beck, and Elmore enabled the most successful maize yield prediction in this study, the main challenge for practical use of the performed methodology is standardizing and validating this approach for crops with varying phenological dynamics, especially winter crops, which shall be evaluated in future studies.

5. Conclusions

This study aims to evaluate the predictive accuracy and reliability of regular and spatial cross-validation approaches in subfield-scale maize yield prediction using phenology-derived metrics from Sentinel-2 imagery. The results indicated that regular cross-validation overestimated model accuracy in comparison to the results of spatial cross-validation. Based on produced results, the main conclusions of this study are:
Spatial cross-validation likely captured the effects of spatial autocorrelation, unlike regular cross-validation, thus providing a more realistic and conservative estimate of model generalizability in geospatial applications. Cross-validation results were sensitive to training/test split randomness, especially under low sample counts, reinforcing the importance of repeated validation with sufficient sample size. Therefore, ignoring spatial structure during model evaluation can lead to misleading conclusions, which may negatively affect precision agriculture decision-making.
EVI consistently outperformed WDRVI in terms of phenological model fitting accuracy and yield prediction reliability. Most notably, EVI produced notably more samples whose time series data enabled successful phenological modeling, which produced more robust yield prediction in comparison to WDRVI. Yield prediction models based on WDRVI were occasionally more accurate but were highly unstable due to low sample sizes after phenology modeling.
Among phenological fitting methods, AG, Beck, and Elmore demonstrated higher robustness due to their greater number of successfully fitted samples, making them more suitable for subfield crop yield prediction. Conversely, despite small differences in fitting accuracy, Gu, Klos, and Zhang’s methods had poor reliability due to a high failure rate in curve fitting, limiting their applicability in real-world scenarios.
Machine learning model performance varied across parcels, with RF and BGLM producing similar predictive accuracy depending on the parcel and index used.
Future studies should include longer-term and broader spatial datasets and consider integrating environmental covariates, including climate, soil, and topography data, to improve spatial model generalization.

Author Contributions

Conceptualization, D.R.; methodology, D.R.; software, D.R.; validation, D.R.; formal analysis, D.R.; investigation, D.R.; resources, I.P.; data curation, D.R.; writing—original draft preparation, D.R.; writing—review and editing, D.R., I.P. and M.J.; visualization, D.R.; supervision, M.J.; project administration, I.P.; funding acquisition, D.R. and I.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors are thankful to Jerković d.o.o. (Koška, Croatia) for providing ground truth maize yield data for the research. This research was supported by the scientific project “Prediction of maize yield potential using machine learning models based on vegetation indices and phenological metrics from Sentinel-2 multispectral satellite images (AgroVeFe)”.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Optimal hyperparameters for all RF models used in the study.
Table A1. Optimal hyperparameters for all RF models used in the study.
Parcel IDVegetation IndexFitting MethodCross-Validation MethodRF Optimal Hyperparameters
Parcel 1EVIAGregularmtry = 15, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 19, splitrule = “extratrees”, min.node.size = 5
Beckregularmtry = 13, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 4, splitrule = “extratrees”, min.node.size = 5
Elmoreregularmtry = 19, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 22, splitrule = “extratrees”, min.node.size = 5
Guregularmtry = 22, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 13, splitrule = “extratrees”, min.node.size = 5
Klosregularmtry = 22, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 19, splitrule = “extratrees”, min.node.size = 5
Zhangregularmtry = 4, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 6, splitrule = “extratrees”, min.node.size = 5
WDRVIAGregularmtry = 6, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 10, splitrule = “extratrees”, min.node.size = 5
Parcel 2EVIAGregularmtry = 19, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 13, splitrule = “extratrees”, min.node.size = 5
Beckregularmtry = 6, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 19, splitrule = “extratrees”, min.node.size = 5
Elmoreregularmtry = 22, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 15, splitrule = “extratrees”, min.node.size = 5
Guregularmtry = 10, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 15, splitrule = “extratrees”, min.node.size = 5
Klosregularmtry = 13, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 13, splitrule = “extratrees”, min.node.size = 5
Zhangregularmtry = 13, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 10, splitrule = “extratrees”, min.node.size = 5
WDRVIAGregularmtry = 4, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 6, splitrule = “extratrees”, min.node.size = 5
Parcel 3EVIAGregularmtry = 22, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 8, splitrule = “extratrees”, min.node.size = 5
Beckregularmtry = 15, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 13, splitrule = “extratrees”, min.node.size = 5
Elmoreregularmtry = 2, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 6, splitrule = “extratrees”, min.node.size = 5
WDRVIAGregularmtry = 2, splitrule = “extratrees”, min.node.size = 5
spatialmtry = 2, splitrule = “extratrees”, min.node.size = 5

References

  1. Burdett, H.; Wellen, C. Statistical and Machine Learning Methods for Crop Yield Prediction in the Context of Precision Agriculture. Precis. Agric. 2022, 23, 1553–1574. [Google Scholar] [CrossRef]
  2. Kitchen, N.R.; Clay, S.A. Understanding and Identifying Variability. In Precision Agriculture Basics; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 2018; pp. 13–24. ISBN 978-0-89118-367-9. [Google Scholar]
  3. Wang, N.; Wu, Q.; Gui, Y.; Hu, Q.; Li, W. Cross-Modal Segmentation Network for Winter Wheat Mapping in Complex Terrain Using Remote-Sensing Multi-Temporal Images and DEM Data. Remote Sens. 2024, 16, 1775. [Google Scholar] [CrossRef]
  4. Wang, Y.; Yuan, Y.; Yuan, F.; Liu, X.; Tian, Y.; Zhu, Y.; Cao, W.; Cao, Q. Optimizing Management Zone Delineation through Advanced Dimensionality Reduction Models and Clustering Algorithms. Precis. Agric. 2025, 26, 68. [Google Scholar] [CrossRef]
  5. McFadden, J.R.; Rosburg, A.; Njuki, E. Information Inputs and Technical Efficiency in Midwest Corn Production: Evidence from Farmers’ Use of Yield and Soil Maps. Am. J. Agric. Econ. 2022, 104, 589–612. [Google Scholar] [CrossRef]
  6. Radočaj, D.; Plaščak, I.; Jurišić, M. Phenology-Based Maize and Soybean Yield Potential Prediction Using Machine Learning and Sentinel-2 Imagery Time-Series. Appl. Sci. 2025, 15, 7216. [Google Scholar] [CrossRef]
  7. Getahun, S.; Kefale, H.; Gelaye, Y. Application of Precision Agriculture Technologies for Sustainable Crop Production and Environmental Sustainability: A Systematic Review. Sci. World J. 2024, 2024, 2126734. [Google Scholar] [CrossRef]
  8. Aarif KO, M.; Alam, A.; Hotak, Y. Smart Sensor Technologies Shaping the Future of Precision Agriculture: Recent Advances and Future Outlooks. J. Sens. 2025, 2025, 2460098. [Google Scholar] [CrossRef]
  9. Dhillon, R.; Moncur, Q. Small-Scale Farming: A Review of Challenges and Potential Opportunities Offered by Technological Advancements. Sustainability 2023, 15, 15478. [Google Scholar] [CrossRef]
  10. Longchamps, L.; Tisseyre, B.; Taylor, J.; Sagoo, L.; Momin, A.; Fountas, S.; Manfrini, L.; Ampatzidis, Y.; Schueller, J.K.; Khosla, R. Yield Sensing Technologies for Perennial and Annual Horticultural Crops: A Review. Precis. Agric. 2022, 23, 2407–2448. [Google Scholar] [CrossRef]
  11. Phiri, D.; Simwanda, M.; Salekin, S.; Nyirenda, V.R.; Murayama, Y.; Ranagalage, M. Sentinel-2 Data for Land Cover/Use Mapping: A Review. Remote Sens. 2020, 12, 2291. [Google Scholar] [CrossRef]
  12. Sun, H.; Ma, X.; Liu, Y.; Zhou, G.; Ding, J.; Lu, L.; Wang, T.; Yang, Q.; Shu, Q.; Zhang, F. A New Multiangle Method for Estimating Fractional Biocrust Coverage From Sentinel-2 Data in Arid Areas. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4404015. [Google Scholar] [CrossRef]
  13. Sentinel-2 L2A—Documentation. Available online: https://documentation.dataspace.copernicus.eu/APIs/SentinelHub/Data/S2L2A.html (accessed on 3 June 2025).
  14. Zou, X.; Zhu, S.; Mõttus, M. Estimation of Canopy Structure of Field Crops Using Sentinel-2 Bands with Vegetation Indices and Machine Learning Algorithms. Remote Sens. 2022, 14, 2849. [Google Scholar] [CrossRef]
  15. Swoish, M.; Da Cunha Leme Filho, J.F.; Reiter, M.S.; Campbell, J.B.; Thomason, W.E. Comparing Satellites and Vegetation Indices for Cover Crop Biomass Estimation. Comput. Electron. Agric. 2022, 196, 106900. [Google Scholar] [CrossRef]
  16. Arshad, A.; Raza, M.A.; Zhang, Y.; Zhang, L.; Wang, X.; Ahmed, M.; Habib-ur-Rehman, M. Impact of Climate Warming on Cotton Growth and Yields in China and Pakistan: A Regional Perspective. Agriculture 2021, 11, 97. [Google Scholar] [CrossRef]
  17. Arshad, A.; Zhang, Y.; Zhang, P.; Wang, X.; Chen, Y.; Ahmed, M.; Zhang, L. APSIM-Cotton Model Calibration for Phenology-Driven Sowing and Yield Optimization in Drip Irrigated Arid Climate. Smart Agric. Technol. 2025, 12, 101325. [Google Scholar] [CrossRef]
  18. Diao, C.; Li, G. Near-Surface and High-Resolution Satellite Time Series for Detecting Crop Phenology. Remote Sens. 2022, 14, 1957. [Google Scholar] [CrossRef]
  19. Ma, Y.; Liang, S.-Z.; Myers, D.B.; Swatantran, A.; Lobell, D.B. Subfield-Level Crop Yield Mapping without Ground Truth Data: A Scale Transfer Framework. Remote Sens. Environ. 2024, 315, 114427. [Google Scholar] [CrossRef]
  20. Dossa, K.F.; Bissonnette, J.-F.; Barrette, N.; Bah, I.; Miassi, Y.E. Projecting Climate Change Impacts on Benin’s Cereal Production by 2050: A SARIMA and PLS-SEM Analysis of FAO Data. Climate 2025, 13, 19. [Google Scholar] [CrossRef]
  21. Radočaj, D.; Plaščak, I.; Jurišić, M. A Machine-Learning Approach for the Assessment of Quantitative Changes in the Tractor Diesel-Engine Oil During Exploitation. Poljoprivreda 2024, 30, 108–114. [Google Scholar] [CrossRef]
  22. Yates, L.A.; Aandahl, Z.; Richards, S.A.; Brook, B.W. Cross Validation for Model Selection: A Review with Examples from Ecology. Ecol. Monogr. 2023, 93, e1557. [Google Scholar] [CrossRef]
  23. Tziachris, P.; Nikou, M.; Aschonitis, V.; Kallioras, A.; Sachsamanoglou, K.; Fidelibus, M.D.; Tziritis, E. Spatial or Random Cross-Validation? The Effect of Resampling Methods in Predicting Groundwater Salinity with Machine Learning in Mediterranean Region. Water 2023, 15, 2278. [Google Scholar] [CrossRef]
  24. Koldasbayeva, D.; Tregubova, P.; Gasanov, M.; Zaytsev, A.; Petrovskaia, A.; Burnaev, E. Challenges in Data-Driven Geospatial Modeling for Environmental Research and Practice. Nat. Commun. 2024, 15, 10700. [Google Scholar] [CrossRef]
  25. Wang, J.; Wang, Y.; Li, G.; Qi, Z. Integration of Remote Sensing and Machine Learning for Precision Agriculture: A Comprehensive Perspective on Applications. Agronomy 2024, 14, 1975. [Google Scholar] [CrossRef]
  26. Beigaitė, R.; Mechenich, M.; Žliobaitė, I. Spatial Cross-Validation for Globally Distributed Data. In Proceedings of the Discovery Science; Pascal, P., Ienco, D., Eds.; Springer Nature Switzerland: Cham, Switzerland, 2022; pp. 127–140. [Google Scholar]
  27. John, K.; Saurette, D.D.; Heung, B. The Problematic Case of Data Leakage: A Case for Leave-Profile-out Cross-Validation in 3-Dimensional Digital Soil Mapping. Geoderma 2025, 455, 117223. [Google Scholar] [CrossRef]
  28. Kattenborn, T.; Schiefer, F.; Frey, J.; Feilhauer, H.; Mahecha, M.D.; Dormann, C.F. Spatially Autocorrelated Training and Validation Samples Inflate Performance Assessment of Convolutional Neural Networks. ISPRS Open J. Photogramm. Remote Sens. 2022, 5, 100018. [Google Scholar] [CrossRef]
  29. Roberts, D.R.; Bahn, V.; Ciuti, S.; Boyce, M.S.; Elith, J.; Guillera-Arroita, G.; Hauenstein, S.; Lahoz-Monfort, J.J.; Schröder, B.; Thuiller, W.; et al. Cross-Validation Strategies for Data with Temporal, Spatial, Hierarchical, or Phylogenetic Structure. Ecography 2017, 40, 913–929. [Google Scholar] [CrossRef]
  30. Wang, Y.; Khodadadzadeh, M.; Zurita-Milla, R. Spatial+: A New Cross-Validation Method to Evaluate Geospatial Machine Learning Models. Int. J. Appl. Earth Obs. Geoinf. 2023, 121, 103364. [Google Scholar] [CrossRef]
  31. Meyer, H.; Pebesma, E. Predicting into Unknown Space? Estimating the Area of Applicability of Spatial Prediction Models. Methods Ecol. Evol. 2021, 12, 1620–1633. [Google Scholar] [CrossRef]
  32. Radočaj, D.; Jurišić, M. A Phenology-Based Evaluation of the Optimal Proxy for Cropland Suitability Based on Crop Yield Correlations from Sentinel-2 Image Time-Series. Agriculture 2025, 15, 859. [Google Scholar] [CrossRef]
  33. Banaj, A.; Banaj, Đ.; Stipešević, B.; Horvat, D. The Impact of Planting Technology on the Maize Yield. Poljoprivreda 2024, 30, 100–107. [Google Scholar] [CrossRef]
  34. CLAAS Connect|CLAAS. Available online: https://www.claas.com/en-us/smart-farming/claas-connect (accessed on 24 July 2025).
  35. Paccioretti, P.; Córdoba, M.; Giannini-Kurina, F.; Balzarini, M. Paar: Precision Agriculture Data Analysis. Available online: https://cran.r-project.org/web/packages/paar/index.html (accessed on 24 July 2025).
  36. Paccioretti, P.; Córdoba, M.; Balzarini, M. FastMapping: Software to Create Field Maps and Identify Management Zones in Precision Agriculture. Comput. Electron. Agric. 2020, 175, 105556. [Google Scholar] [CrossRef]
  37. Radočaj, D.; Plaščak, I.; Jurišić, M. Fusion of Sentinel-2 Phenology Metrics and Saturation-Resistant Vegetation Indices for Improved Correlation with Maize Yield Maps. Agronomy 2025, 15, 1329. [Google Scholar] [CrossRef]
  38. Huete, A.; Didan, K.; Miura, T.; Rodriguez, E.P.; Gao, X.; Ferreira, L.G. Overview of the Radiometric and Biophysical Performance of the MODIS Vegetation Indices. Remote Sens. Environ. 2002, 83, 195–213. [Google Scholar] [CrossRef]
  39. Gitelson, A.A. Wide Dynamic Range Vegetation Index for Remote Quantification of Biophysical Characteristics of Vegetation. J. Plant Physiol. 2004, 161, 165–173. [Google Scholar] [CrossRef] [PubMed]
  40. Kong, D.; Xiao, M.; Zhang, Y.; Gu, X.; Cui, J. Phenofit: Extract Remote Sensing Vegetation Phenology. Available online: https://cran.r-project.org/web/packages/phenofit/index.html (accessed on 4 June 2025).
  41. Kong, D.; Zhang, Y.; Wang, D.; Chen, J.; Gu, X. Photoperiod Explains the Asynchronization Between Vegetation Carbon Phenology and Vegetation Greenness Phenology. J. Geophys. Res. Biogeosciences 2020, 125, e2020JG005636. [Google Scholar] [CrossRef]
  42. Rapčan, I.; Radočaj, D.; Jurišić, M. A Length-of-Season Analysis for Maize Cultivation from the Land- Surface Phenology Metrics Using the Sentinel-2 Images. Poljoprivreda 2025, 31, 92–98. [Google Scholar] [CrossRef]
  43. Kuhn, M.; Wing, J.; Weston, S.; Williams, A.; Keefer, C.; Engelhardt, A.; Cooper, T.; Mayer, Z.; Kenkel, B.; R Core Team; et al. Caret: Classification and Regression Training. Available online: https://cran.r-project.org/web/packages/caret/index.html (accessed on 17 July 2025).
  44. Wright, M.N.; Wager, S.; Probst, P. Ranger: A Fast Implementation of Random Forests. Available online: https://cran.r-project.org/web/packages/ranger/index.html (accessed on 4 June 2025).
  45. Gelman, A.; Su, Y.-S.; Yajima, M.; Hill, J.; Pittau, M.G.; Kerman, J.; Zheng, T.; Dorie, V. Arm: Data Analysis Using Regression and Multilevel/Hierarchical Models. Available online: https://cran.r-project.org/web/packages/arm/index.html (accessed on 28 July 2025).
  46. Meyer, H.; Ludwig, M.; Milà, C.; Linnenbrink, J.; Schumacher, F. The CAST Package for Training and Assessment of Spatial Prediction Models in R. arXiv 2024, arXiv:2404.06978. [Google Scholar] [CrossRef]
  47. Sáenz, C.; Cicuéndez, V.; García, G.; Madruga, D.; Recuero, L.; Bermejo-Saiz, A.; Litago, J.; de la Calle, I.; Palacios-Orueta, A. New Insights on the Information Content of the Normalized Difference Vegetation Index Sentinel-2 Time Series for Assessing Vegetation Dynamics. Remote Sens. 2024, 16, 2980. [Google Scholar] [CrossRef]
  48. POLJ-2023-2-6 Area and Production of Cereals and Other Crops in 2023—Provisional Data|State Bureau of Statistics. Available online: https://podaci.dzs.hr/2023/hr/58457 (accessed on 25 July 2025).
  49. Mo, Y.; Zhang, X.; Liu, Z.; Zhang, J.; Hao, F.; Fu, Y. Effects of Climate Extremes on Spring Phenology of Temperate Vegetation in China. Remote Sens. 2023, 15, 686. [Google Scholar] [CrossRef]
  50. Zhang, D.; Hou, L.; Lv, L.; Qi, H.; Sun, H.; Zhang, X.; Li, S.; Min, J.; Liu, Y.; Tang, Y.; et al. Precision Agriculture: Temporal and Spatial Modeling of Wheat Canopy Spectral Characteristics. Agriculture 2025, 15, 326. [Google Scholar] [CrossRef]
  51. Croci, M.; Ragazzi, M.; Grassi, A.; Impollonia, G.; Amaducci, S. Assessing the Temporal Transferability of Machine Learning Models for Predicting Processing Pea Yield and Quality Using Sentinel-2 and ERA5-Land Data. Smart Agric. Technol. 2025, 12, 101207. [Google Scholar] [CrossRef]
  52. Perich, G.; Turkoglu, M.O.; Graf, L.V.; Wegner, J.D.; Aasen, H.; Walter, A.; Liebisch, F. Pixel-Based Yield Mapping and Prediction from Sentinel-2 Using Spectral Indices and Neural Networks. Field Crops Res. 2023, 292, 108824. [Google Scholar] [CrossRef]
  53. Fernando, H.; Ha, T.; Nketia, K.A.; Attanayake, A.; Shirtliffe, S. Machine Learning Approach for Satellite-Based Subfield Canola Yield Prediction Using Floral Phenology Metrics and Soil Parameters. Precis. Agric. 2024, 25, 1386–1403. [Google Scholar] [CrossRef]
  54. Crusiol, L.G.T.; Sun, L.; Sibaldelli, R.N.R.; Junior, V.F.; Furlaneti, W.X.; Chen, R.; Sun, Z.; Wuyun, D.; Chen, Z.; Nanni, M.R.; et al. Strategies for Monitoring Within-Field Soybean Yield Using Sentinel-2 Vis-NIR-SWIR Spectral Bands and Machine Learning Regression Methods. Precis. Agric. 2022, 23, 1093–1123. [Google Scholar] [CrossRef]
  55. An, C.; Park, Y.W.; Ahn, S.S.; Han, K.; Kim, H.; Lee, S.-K. Radiomics Machine Learning Study with a Small Sample Size: Single Random Training-Test Set Split May Lead to Unreliable Results. PLoS ONE 2021, 16, e0256152. [Google Scholar] [CrossRef] [PubMed]
  56. Darra, N.; Anastasiou, E.; Kriezi, O.; Lazarou, E.; Kalivas, D.; Fountas, S. Can Yield Prediction Be Fully Digitilized? A Systematic Review. Agronomy 2023, 13, 2441. [Google Scholar] [CrossRef]
  57. Mancini, A.; Solfanelli, F.; Coviello, L.; Martini, F.M.; Mandolesi, S.; Zanoli, R. Time Series from Sentinel-2 for Organic Durum Wheat Yield Prediction Using Functional Data Analysis and Deep Learning. Agronomy 2024, 14, 109. [Google Scholar] [CrossRef]
  58. Shi, Q.; Dai, W.; Santerre, R.; Liu, N. A Modified Spatiotemporal Mixed-Effects Model for Interpolating Missing Values in Spatiotemporal Observation Data Series. Math. Probl. Eng. 2020, 2020, 1070831. [Google Scholar] [CrossRef]
  59. Hounkpatin, K.O.L.; Stendahl, J.; Lundblad, M.; Karltun, E. Predicting the Spatial Distribution of Soil Organic Carbon Stock in Swedish Forests Using a Group of Covariates and Site-Specific Data. Soil 2021, 7, 377–398. [Google Scholar] [CrossRef]
Figure 1. A display of the study area, including three maize parcels from the year 2023 located in eastern Croatia.
Figure 1. A display of the study area, including three maize parcels from the year 2023 located in eastern Croatia.
Eng 06 00270 g001
Figure 2. Schematic representation of training and datasets from regular and spatial 10-fold cross-validation per fold.
Figure 2. Schematic representation of training and datasets from regular and spatial 10-fold cross-validation per fold.
Eng 06 00270 g002
Figure 3. Violin plots representing value distribution of ground truth maize yield samples per parcel during the 2023 harvest after preprocessing and outlier removal.
Figure 3. Violin plots representing value distribution of ground truth maize yield samples per parcel during the 2023 harvest after preprocessing and outlier removal.
Eng 06 00270 g003
Figure 4. Boxplots of all resulting accuracy assessment metrics after machine learning prediction and 10-fold cross-validation across all repetitions, representing variability of model accuracy across cross-validation folds and repetitions.
Figure 4. Boxplots of all resulting accuracy assessment metrics after machine learning prediction and 10-fold cross-validation across all repetitions, representing variability of model accuracy across cross-validation folds and repetitions.
Eng 06 00270 g004
Table 1. Descriptive statistics of ground truth maize yield samples before and after preprocessing.
Table 1. Descriptive statistics of ground truth maize yield samples before and after preprocessing.
Parcel IDAreaPreprocessing StageSample CountMeanMedianCVShapiro–Wilk Test
p-Value
Parcel 111.4 haBefore preprocessing10947.698.160.242<0.0001
After preprocessing7498.348.540.135<0.0001
Parcel 24.4 haBefore preprocessing3028.258.440.229<0.0001
After preprocessing1888.918.980.1340.0217
Parcel 32.2 haBefore preprocessing1708.208.460.275<0.0001
After preprocessing949.279.240.1290.4941
CV: coefficient of variation.
Table 2. Accuracy assessment of evaluated fitting methods during phenological modeling per parcel and vegetation index.
Table 2. Accuracy assessment of evaluated fitting methods during phenological modeling per parcel and vegetation index.
Parcel IDFitting MethodEVIWDRVI
R2RMSENSEnfitR2RMSENSEnfit
Parcel 1AG0.9240.0660.9176100.2510.500−0.94135
Beck0.9220.0670.9166100.2660.500−0.93935
Elmore0.9210.0670.9156100.2650.500−0.94135
Gu0.9280.0640.9222590.1600.488−1.0713
Klos0.9280.0640.9232590.1600.490−1.0873
Zhang0.9280.0640.9222590.2120.489−1.0783
Parcel 2AG0.6030.2840.5821460.2580.491−0.75565
Beck0.5990.2840.5821460.2840.491−0.74965
Elmore0.6000.2840.5831460.2850.491−0.75565
Gu0.5770.3000.558550.2520.476−0.70010
Klos0.5830.2990.564550.2450.476−0.70210
Zhang0.5770.3000.559550.2650.477−0.70610
Parcel 3AG0.9410.0640.927710.2210.474−0.92728
Beck0.9400.0620.935710.2300.474−0.92828
Elmore0.9390.0620.935710.2370.473−0.92128
Gu0.9380.0640.934140.2030.473−0.82411
Klos0.9410.0620.937140.2750.472−0.81711
Zhang0.9330.0680.925140.2370.474−0.83611
nfit—the number of successfully fitted samples during phenological modeling.
Table 3. Accuracy assessment of evaluated machine learning models for maize yield prediction per parcel and vegetation index according to regular and spatial cross-validation.
Table 3. Accuracy assessment of evaluated machine learning models for maize yield prediction per parcel and vegetation index according to regular and spatial cross-validation.
Parcel IDVegetation IndexFitting MethodCross-Validation MethodRFBGLMn
R2RMSEMAER2RMSEMAE
Parcel 1EVIAGregular0.5690.7540.5780.4490.8630.654610
spatial0.1340.9020.7200.1490.9160.706610
Beckregular0.5710.7490.5640.4570.8410.651608
spatial0.1620.8750.6980.1480.8720.696608
Elmoreregular0.5050.8010.6110.3290.9320.740606
spatial0.1410.9500.7640.1170.9970.817606
Guregular0.5680.7760.5880.4710.8620.671259
spatial0.1720.9200.7620.2500.9430.776259
Klosregular0.4980.8390.6420.3600.9640.724258
spatial0.1311.0420.8510.1971.0620.844258
Zhangregular0.5760.7720.5820.4500.8840.683259
spatial0.1950.9230.7470.1980.9540.772259
WDRVIAGregular0.4610.4330.3870.5510.5960.52834
spatial0.5880.4490.3930.6240.7360.64334
Parcel 2EVIAGregular0.3091.0230.8230.3621.0080.800146
spatial0.1751.0690.8850.1791.0250.825146
Beckregular0.3540.9870.7810.3601.0090.814144
spatial0.1601.0350.8440.1731.0130.833144
Elmoreregular0.3570.9920.7960.2381.1110.894146
spatial0.2511.0250.8450.2001.1970.988146
Guregular0.6770.8970.7670.7000.8370.72455
spatial0.2160.9770.8540.3320.8940.78355
Klosregular0.7510.7540.6430.6400.9150.79155
spatial0.3560.8330.7370.3291.0410.87855
Zhangregular0.6560.9040.7500.6790.8690.74155
spatial0.1801.0870.9370.3191.0780.91255
WDRVIAGregular0.2770.7270.6040.2351.2340.88264
spatial0.2400.7400.6470.2031.2670.90064
Parcel 3EVIAGregular0.3360.9530.7600.4360.8990.70871
spatial0.1501.1350.9370.2821.0000.81071
Beckregular0.3970.9250.7490.3951.0040.80671
spatial0.1951.0620.8880.2851.0810.89071
Elmoreregular0.2981.0430.8420.2371.1350.89671
spatial0.2031.1941.0170.1941.2621.04771
WDRVIAGregular0.5941.0440.8650.5981.6501.35728
spatial0.7251.1391.0090.7201.9221.71428
Accuracy assessment metrics indicating the most accurate prediction per parcel for both regular and spatial cross-validation are bolded.
Table 4. The results of Wilcoxon tests comparing the accuracy assessment statistical metrics from regular and spatial cross-validation for evaluated machine learning models.
Table 4. The results of Wilcoxon tests comparing the accuracy assessment statistical metrics from regular and spatial cross-validation for evaluated machine learning models.
Test StatisticsRFBGLM
R2RMSEMAER2RMSEMAE
p-value<0.0001<0.0001<0.0001<0.0001<0.0001<0.0001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Radočaj, D.; Plaščak, I.; Jurišić, M. A Comparative Assessment of Regular and Spatial Cross-Validation in Subfield Machine Learning Prediction of Maize Yield from Sentinel-2 Phenology. Eng 2025, 6, 270. https://doi.org/10.3390/eng6100270

AMA Style

Radočaj D, Plaščak I, Jurišić M. A Comparative Assessment of Regular and Spatial Cross-Validation in Subfield Machine Learning Prediction of Maize Yield from Sentinel-2 Phenology. Eng. 2025; 6(10):270. https://doi.org/10.3390/eng6100270

Chicago/Turabian Style

Radočaj, Dorijan, Ivan Plaščak, and Mladen Jurišić. 2025. "A Comparative Assessment of Regular and Spatial Cross-Validation in Subfield Machine Learning Prediction of Maize Yield from Sentinel-2 Phenology" Eng 6, no. 10: 270. https://doi.org/10.3390/eng6100270

APA Style

Radočaj, D., Plaščak, I., & Jurišić, M. (2025). A Comparative Assessment of Regular and Spatial Cross-Validation in Subfield Machine Learning Prediction of Maize Yield from Sentinel-2 Phenology. Eng, 6(10), 270. https://doi.org/10.3390/eng6100270

Article Metrics

Back to TopTop