Next Article in Journal
ESG Performance and Digital Transformation: Evidence from Chinese A-Listed Companies
Previous Article in Journal
Can Low-Carbon City Pilot Policies Promote Green Total Factor Energy Efficiency in China?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Models for Yield Estimation of Hybrid and Conventional Japonica Rice Cultivars Using UAV Imagery

1
College of Agriculture, Northeast Agricultural University, Harbin 150030, China
2
National Key Laboratory of Smart Farm Technologies and Systems, Harbin 150030, China
3
Institute of Crops Research, Xinjiang Academy of Agricultural and Reclamation Sciences, Shihezi 832000, China
4
College of Electrical Engineering and Information, Northeast Agricultural University, Harbin 150030, China
*
Authors to whom correspondence should be addressed.
Sustainability 2025, 17(18), 8515; https://doi.org/10.3390/su17188515
Submission received: 3 July 2025 / Revised: 11 September 2025 / Accepted: 18 September 2025 / Published: 22 September 2025

Abstract

Advancements in unmanned aerial vehicle (UAV) multispectral systems offer robust technical support for the precise and efficient estimation of japonica rice yield in cold regions within the framework of precision agriculture. These innovations also present a viable alternative to conventional yield estimation methods. However, recent research suggests that reliance solely on vegetation indices (VIs) may result in inaccurate yield estimations due to variations in crop cultivars, growth stages, and environmental conditions. This study investigated six fertilization gradient experiments involving two conventional japonica rice varieties (KY131, SJ22) and two hybrid japonica rice varieties (CY31, TLY619) at Yanjiagang Farm in Heilongjiang Province during 2023. By integrating UAV multispectral data with machine learning techniques, this research aimed to derive critical phenotypic parameters of rice and estimate yield. This study was conducted in two phases: In the first phase, models for assessing phenotypic traits such as leaf area index (LAI), canopy cover (CC), plant height (PH), and above-ground biomass (AGB) were developed using remote sensing spectral indices and machine learning algorithms, including Random Forest (RF), XGBoost, Support Vector Regression (SVR), and Backpropagation Neural Network (BPNN). In the second phase, plot yields for hybrid rice and conventional rice were predicted using key phenotypic parameters at critical growth stages through linear (Multiple Linear Regression, MLR) and nonlinear regression models (RF). The findings revealed that (1) Phenotypic traits at critical growth stages exhibited a strong correlation with rice yield, with correlation coefficients for LAI and CC exceeding 0.85 and (2) the accuracy of phenotypic trait evaluation using multispectral data was high, demonstrating practical applicability in production settings. Remarkably, the R2 for CC based on the RF algorithm exceeded 0.9, while R2 values for PH and AGB using the RF algorithm and for LAI using the XGBoost algorithm all surpassed 0.8. (3) Yield estimation performance was optimal at the heading (HD) stage, with the RF model achieving superior accuracy (R2 = 0.86, RMSE = 0.59 t/ha) compared to other growth stages. These results underscore the immense potential of combining UAV multispectral data with machine learning techniques to enhance the accuracy of yield estimation for cold-region japonica rice. This innovative approach significantly supports optimized decision-making for farmers in precision agriculture and holds substantial practical value for rice yield estimation and the sustainable advancement of rice production.

1. Introduction

Amid the continuously increasing demands for food production and consumption, coupled with the significant challenges posed by climate change to agricultural systems, ensuring sustainable agricultural production has become a critical research priority. This necessitates the efficient allocation and optimization of various factors in food production, guided by accurate crop yield estimation [1,2]. Traditional growth monitoring methods primarily rely on field measurements, which are labor-intensive, time-consuming, and prone to crop damage and can also be heavily influenced by weather conditions [3]. Consequently, these conventional approaches have been empirically shown to be inefficient and inadequate for effective crop growth monitoring in agricultural fields. In contrast, remote sensing technologies enable farmers and technicians to detect and monitor a wide range of crop traits [4,5]. Unmanned aerial vehicle (UAV)-based remote sensing, in particular, offers prominent advantages, including broad spatial coverage, high cost-effectiveness, and a well-established technological framework, leading to its extensive application in precision agriculture [6,7]. As a staple food crop of global importance, rice production stability and improvement are directly linked to food security worldwide. Therefore, accurate pre-harvest prediction of rice yield holds significant practical value for ensuring stable and high production while promoting the sustainable use of agricultural resources such as water, fertilizers, and pesticides [8,9].
Spectral indices in crop remote sensing can reflect crop growth status and stress levels, making them valuable for yield assessment during critical growth stages. Wang F. et al. [10] identified the optimal timing for yield evaluation using UAV-based multispectral data, while Jian Wang et al. [11] assessed the impact of lodging on the accuracy of rice yield estimation. Xi Su et al. [12] combined spectral mixture analysis with multispectral indices to enhance estimation accuracy by deriving parameters related to leaves, soil, and water. However, the limited resolution of multispectral cameras restricts the acquisition of organ-level phenotypic information in rice plants. To address this, Yang et al. [13] improved prediction accuracy by integrating multispectral sensor data with color and texture features from RGB images. Wang F. et al. [14] further enhanced yield forecasting by incorporating fluorescence spectroscopy, and Hou Longfei et al. [15] integrated phenological stages and agronomic management practices into rice yield prediction models. In terms of yield estimation modeling, ensemble learning has emerged as a prominent research focus. Sarkar et al. [16] demonstrated the advantages of ensemble learning approaches after evaluating multiple algorithms. Similarly, with advances in artificial intelligence, Yang Q. et al. [17] reviewed CNN- and ANN-based algorithms for yield estimation using imagery.
Although yield estimation based on spectral indices has shown promising results, Sun et al. [18] noted that the correlation between these indices and rice yield is often weak, making direct yield predictions highly sensitive to variations in experimental data. Furthermore, because vegetation indices are influenced by multiple factors related to crop growth, relying solely on these parameters makes it challenging to provide actionable agricultural guidance. In contrast, Zha H. et al. [19] emphasized that plant height (PH), leaf area index (LAI), and biomass exhibit strong correlations with rice yield throughout the growth stages.
This study conducted a field experiment involving two conventional and two hybrid rice varieties, tested under six fertilizer gradients with three replicates each. Both UAV-based remote sensing data and manual measurements were collected to support yield modeling and analysis. The objectives of this research are to (i) assess the accuracy of spectral index-based models in estimating rice biomass, yield, and PH; (ii) evaluate the feasibility of predicting yield through the preliminary assessment of key growth stage phenotypes; and (iii) identify the optimal growth stage and most effective modeling approach for rice yield estimation.

2. Materials and Methods

Figure 1 provides a comprehensive overview of the methodology and multi-stage procedures employed in this study, including field experiments, computational modeling, machine learning regression modeling, and statistical regression analyses. The illustrated workflow begins with field data collection and UAV data acquisition, followed by Spearman correlation analysis to determine the relationship between phenotypic parameters and yield. Then, machine learning regression models are constructed to invert crop phenotypic parameters, and linear and nonlinear regression analyses are conducted for yield estimation. Details regarding data acquisition equipment, model construction, software utilized, and analytical techniques applied at each stage are further described in the following.

2.1. Study Area and Experimental Setup

This field experiment was conducted in 2023 at Yanjia Gang Farm, Harbin, Heilongjiang Province, China (126°18′ E, 45°36′ N), which features a mid-temperate continental monsoon climate. The site has an average annual precipitation of 380 mm and an average annual temperature of 3.4 °C, and the annual average accumulated temperature is 2851 °C, belonging to the first accumulated temperature zone. The average annual frost-free period is approximately 160 days. An overview of the experimental site and plot layout is presented in Figure 2.
In this experiment, two conventional japonica rice varieties (Kongyu131 [KY131] and Songjing22 [SJ22]) and two hybrid japonica rice varieties (Chuangyou31 [CY31] and Tianlongyou619 [TLY619]) served as plant materials. Detailed information on these rice varieties is provided in Table 1. For each variety, 6 plots with different nitrogen concentrations were setup (among the 6 nitrogen fertilizer treatments, the nitrogen application rates of different treatments are 0 (N1), 45 (N2), 90 (N3), 135 (N4), 180 (N5), and 225 (N6) kg/ha, respectively), with 3 replicates for each concentration, resulting in a total of 72 experimental plots. Each plot measured 25 m2 in area, and adjacent plots were separated by ridges approximately 0.5 m wide. Seeds underwent a standard soaking and germination process prior to sowing. Sowing was performed on 15 April 2023, and seedlings were transplanted on 15 May 2023, at 30 days of age. Plant spacing was set at 13.3 cm × 30.0 cm. Three seedlings per hill were transplanted for conventional japonica rice, and two seedlings per hill for hybrid japonica rice. To minimize yield reduction, refined management strategies—including appropriate pest, disease, and weed control—were implemented throughout the growing season in line with local agricultural practices. The data in this study were collected from June to October 2023. During the data collection period of this study (June–October 2023), the climate in Harbin showed significant seasonal variations, specifically as follows: In June, the average temperature was 15–25 °C, with moderate rainfall, and the temperature and humidity conditions were coordinated, which was suitable for the growth and development of the tested crops. From July to August, the average temperature rose to 20–28 °C, with concentrated precipitation and sufficient light and heat resources, and the risk of waterlogging needed to be prevented during this stage. In September and October, the average temperature dropped to 10–20 °C, the rainfall decreased, and the tested crops were harvested one after another, so that the yield data could be accurately obtained. The aforementioned climatic characteristics provided typical regional environmental conditions for field experiments and data collection.

2.2. Data Collection and Preprocessing

2.2.1. Data Collection

During field data collection, six rice plants were randomly selected from the inner rows of each plot at four growth stages: tillering (TL), panicle initiation (PI), heading (HD), and maturity (MT). The selected plants showed uniform growth and were representative of the plot condition. LAI, CC, PH, and AGB were measured using a crop growth monitor. All measurements were performed between 08:00 and 10:00 local time to minimize the influence of variations in light intensity on data collection.
LAI was measured non-destructively through PAR interception using AccuPAR LP-80 ceptometers (METER Group, Inc., Pullman, WA, USA). The AccuPAR LP-80 calculates LAI by integrating PAR measurements taken above and below the canopy, alongside additional parameters such as canopy structure and solar position. The instrument automatically computes the zenith angle based on input parameters, including longitude, latitude, and local time at the measurement location. To obtain LAI, photosynthetically active radiation (PAR) intensity was first measured at the base of the rice canopy (approximately 5 cm above the water or soil surface), and immediately afterward, the incident PAR at the top of the canopy was measured. The instrument automatically computes the LAI value. Measurements were performed under diffuse light conditions to reduce artifacts caused by solar angle variations. All sensors underwent cosine correction and a 5-point inter-calibration process, achieving an inter-sensor variability of less than 3%.
Chlorophyll content (CC) was assessed using a SPAD-502 chlorophyll meter (Konica Minolta Sensing, Inc., Tokyo, Japan). In each plot, six fully expanded functional leaves exhibiting consistent growth were selected from the top of the canopy, specifically the first to third leaves from the apex, which contribute more than 70% to yield. Each leaf was measured three times at different positions to avoid repeated measurements at the same position, and the average of these three readings was recorded as the CC value for that leaf. Prior to measurement, leaf surfaces were cleaned of dust, and care was taken to avoid veins and damaged areas. The mean and variance of CC readings are provided in the table.
Aboveground biomass (AGB) was determined through destructive sampling following standardized protocols (GB/T 32725-2016; General Administration of Quality Supervision, Inspection and Quarantine of the People’s Republic of China (AQSIQ), Standardization Administration of China (SAC): Beijing, China, 2016.). For each plot, three representative rice plants were selected and cut at ground level. PH was immediately measured from the base to the tip of the highest panicle using a graduated pole with an accuracy of ±0.1 cm. All aboveground tissues, including stems, leaves, and panicles, were sealed in pre-weighed paper bags and dried in a forced-air oven at 75 °C for 48 h until a constant mass was achieved (mass variation < 0.5% between consecutive weightings at 2 h intervals). Dry weight was determined using an analytical balance with a resolution of 0.01 g.
Multispectral UAV data were collected using a drone-mounted multispectral sensor, synchronized with the dates of field data acquisition. A UAV (DJI Matrice 300 RTK) served as the remote sensing platform, equipped with a multispectral camera (MS600 Pro; camera specifications are detailed in Table A5). Image acquisition was conducted under clear weather conditions, with flight times scheduled between 10:30 and 14:00 to minimize the impact of variable light conditions on the imagery. The flight altitude was set at 25 m, with a forward overlap of 80% and a side overlap of 70%, and images were captured at regular time intervals.

2.2.2. Data Preprocessing

Following UAC multispectral imagery acquisition, Yusense Map (V2.2.0) software was employed for image stitching and radiometric calibration. Prior to each flight, a calibrated gray panel was photographed for radiometric calibration. Multispectral images underwent radiometric correction through processes including band registration, image stitching, and calibration using YUSENSE Map software.
The stitched multispectral images were then segmented according to the plot boundaries to calculate the vegetation indices per plot. The stitched images were further processed on ENVI 5.6 software. Initially, the soil background was removed from the stitched images. Subsequently, the Region of Interest (ROI) tool was used to delineate each measurement area, with careful attention paid to exclude sampling zones and edge effect areas to enhance result accuracy. When calculating vegetation indices, the average reflectance of all pixels within the marked area was used as the representative canopy reflectance for the plot.
Several vegetation indices were employed in this study, each with specific sensitivities and applications: the Normalized Difference Vegetation Index (NDVI) exhibits strong sensitivity to chlorophyll concentration and leaf density; the Green Normalized Difference Vegetation Index (GNDVI) optimizes chlorophyll detection through enhanced green spectrum bands; the Soil-Adjusted Vegetation Index (SAVI) reduces soil background interference via integrated correction factors; the Optimized Soil-Adjusted Vegetation Index (OSAVI) adjusts soil adjustment coefficients for rice canopy architecture; the Modified Simple Ratio (MSR) mitigates saturation effects in dense vegetation through nonlinear transformation; the Difference Vegetation Index (DVI) quantifies biomass via linear red–NIR spectral contrast; the Nitrogen Reflectance Index (NRI) offers sensitivity to foliar chlorosis progression (particularly effective for CC inversion); the Normalized Difference Red Edge Index (NDRE) detects subtle chlorophyll variations through red edge spectral dynamics (720 nm); and the Structure-Insensitive Pigment Index (SIPI) indicates photosynthetic stress via carotenoid/chlorophyll ratios. The calculation formulas and references for these vegetation indices are presented in Table 2. Plot segmentation and vegetation index calculations were performed using ENVI 5.3.

2.3. Construction of Rice Phenotypic Parameter Inversion Models

2.3.1. Correlation Analysis

Spearman’s rank correlation coefficient was used as a non-parametric statistical measure to quantify the relationship between variables. This method provides particular advantages when datasets violate parametric assumptions, such as cases involving limited sample sizes or non-normally distributed variables, where Pearson correlation analysis can yield biased estimates [29].
The dataset as a whole displays non-normal distributions. Spearman’s correlation analysis is particularly well-suited for non-normally distributed data; hence, this study employed the Spearman method to analyze the correlation between predicted phenotypic parameters and measured yield. This approach is consistent with established best practices in agricultural data analysis, ensuring reliable statistical inferences even when data deviate from normal distributions.

2.3.2. Phenotypic Inversion Model Development

The extracted vegetation indices served as input data, while rice phenotypic parameters including LAI, CC, PH, and AGB served as target variables, forming the training dataset for machine learning regression modeling. The dataset was split into 70% training and 30% testing sets to ensure independent model training and evaluation.
A rigorous data cleaning protocol was implemented: (1) missing value analysis confirmed the completeness of data across all variables; (2) outliers were identified using the Median Absolute Deviation (MAD) method with a threshold of |modified Z-score| > 3.5; and (3) four outlier plots, deemed measurement artifacts, were permanently excluded. This conservative strategy retained 96% of the original observations while effectively removing technical noise.
During the training stage, several machine learning algorithms, including RF, XG Boost, SVR, and BPNN, were utilized to establish predictive relationships between input features and rice phenotypic data. The modeling procedure involved two main stages. First, feature selection and optimization were performed through correlation analysis or model-based feature importance evaluation methods. This step identified the most influential spectral features and vegetation indices, helping to reduce redundant information and improve model efficiency. Next, model training was performed by fitting parameters with the training set data. RF constructs multiple decision trees and employs ensemble learning techniques to enhance stability and mitigate overfitting [30]. XG Boost is a boosting-tree model that uses gradient-based optimization strategies to improve prediction accuracy while applying regularization to prevent overfitting [31]. SVR is suitable for modeling small-sample, high-dimensional datasets, effectively capturing complex relationships between spectral features and rice phenotypic data [32]. BPNN is a multilayer feedforward neural network that iteratively optimizes weights via the backpropagation algorithm, making it particularly effective at capturing nonlinear functions and providing strong fitting capabilities [33]. Each of these four models has distinct advantages; thus, all were selected for modeling in this study.
To mitigate overfitting and enhance the robustness of the evaluation outcomes, 3-fold cross-validation was adopted in this study. Field plots were used as grouping units to preserve spatial integrity; each treatment was set with 3 repetitions to ensure the accuracy of the data. To optimize model performance, parameter tuning and debugging were conducted for each model. Details of the adjustment and debugging of specific parameters can be found in Appendix A, Table A1, Table A2, Table A3 and Table A4.
Finally, after accuracy validation, the best-performing model was selected to predict rice phenotypic parameters for all plots, making it possible to generate the predicted phenotypic data.

2.4. Construction of Rice Yield Estimation Models

2.4.1. Model Construction

The predicted rice phenotypic data from inversion models were matched with corresponding measured yield data for each growth stage, generating comprehensive stage-specific datasets.
In this study, the dataset was split into 70% training and 30% testing to ensure objective model evaluation and generalization. Subsequently, MLR and RF were applied to model linear and nonlinear relationships between rice phenotypic data and yield at each stage (TL, PI, HD, and MT).
For comparative accuracy assessment, a consistent methodology was applied to develop a yield estimation model using VIs and actual yield data.
During modeling, each stage–model combination was trained and evaluated. The optimal model from the best-performing stage was selected as the final yield estimation model, identifying both the optimal growth stage and modeling approach. The model uses remotely sensed rice phenotypic data to estimate yield at the field scale across multiple growth stages (e.g., based on imagery acquired at key phenological stages). The established model was applied to both conventional and hybrid japonica rice, with yield estimation differences between the two compared to provide a scientific basis for precision agricultural management and rice yield forecasting.

2.4.2. Model’s Performance

At each rice growth stage, inversion models for rice phenotypic data were developed using RF, XG Boost, SVR, and BPNN, while yield estimation models were established employing MLR and RF methods. The performances of the four phenotypic inversion models and the two yield estimation models were evaluated and compared based on their coefficients of determination (R2) and root mean square errors (RMSEs).
The formula used to calculate RMSE is as follows [34]:
R M S E = i = 1 n ( y i y i ^ ) 2 / n
The formula used to calculate R2 is as follows [34]:
R 2 = 1 i = 1 n ( y i y i ^ ) 2 / i = 1 n ( y i y ¯ ) 2
where n denotes the total number of samples in the testing set and y ¯ represents the average measured yield within the testing set.
In these metrics, higher R2 values indicate greater model accuracy, whereas lower RMSE values indicate higher prediction precision. In this study, both evaluation criteria were considered comprehensively to select the model exhibiting the best overall performance.

3. Results

3.1. Comparison of Phenotypic Parameters

PH, as a key agronomic trait, can effectively reflect crop growth vigor and provide an important basis for decision-making in crop layout. The determination results of PH at MT (Figure 3a) showed that among conventional japonica rice varieties, the average PH of KY131 was 67.8 cm and that of SJ22 was 95.8 cm, with an overall average PH of 81.8 cm for this type of variety. Among hybrid japonica rice varieties, the average PH of CY31 was 82.1 cm, and that of TLY619 was 99.4 cm, with an overall average PH of 90.8 cm, which was 9 cm higher than that of conventional japonica rice in total.
The trend of CC change was basically consistent across all varieties, peaking at the HD. The determination data at the HD stage (Figure 3b) showed that among conventional japonica rice varieties, the average CC of KY131 was 51.7 SPAD, the average CC of SJ22 was 42.8 SPAD, and the overall average CC of this type of variety was 47.3 SPAD; among hybrid japonica rice varieties, the average CC of CY31 was 45.9 SPAD and the average CC of TLY619 was 45.4 SPAD, with an overall average CC of 45.6 SPAD, which was 1.7 SPAD lower than that of conventional japonica rice.
During the TL, PI, HD, and MT stages, the effect of variety on LAI was significant. Except for in the PI stage, the LAI of hybrid japonica rice was higher than that of conventional japonica rice, with the greatest difference (0.25 m2/m2) observed at the MT stage (see Figure 3c for details).
Similarly, there were significant differences in AGB among different variety types at several key growth stages, and the dry matter accumulation of each variety reached its maximum and remained stable at the MT stage. The determination data at the maturity stage (Figure 3d) showed that among conventional japonica rice varieties, the average AGB of KY131 was 7.93 t/ha, the average AGB of SJ22 was 13.42 t/ha, and the overall average AGB of this type of variety was 10.67 t/ha; among hybrid japonica rice varieties, the average AGB of CY31 was 11.76 t/ha, the average AGB of TLY619 was 12.62 t/ha, and their overall average AGB was 12.19 t/ha. On average, AGB of hybrid japonica rice was 1.52 t/ha higher than that of conventional japonica rice.

3.2. Correlation Analysis Between Rice Phenotypes and Yield

Rice phenotypic parameters such as LAI, AGB, CC, and PH can directly or indirectly influence rice yield. Therefore, correlation analysis was conducted to examine the relationships between these parameters and yield.
Spearman correlation analysis revealed that AGB, LAI and PH exhibited extremely significant and positive correlations with yield, with correlation coefficients greater than 0.60—specifically, AGB (ρ = 0.784, p < 0.01) and LAI (ρ = 0.622, p < 0.01). PH values also showed an extremely significant positive correlation with yield (ρ = 0.603, p < 0.01). CC had the lowest correlation with yield (ρ = −0.041, p > 0.05), where ρ denotes the correlation coefficient and p denotes significance. Correlation analysis results are shown in Figure 4a.
Additionally, the correlation between nine vegetation indices and yield was examined, with findings illustrated in Figure 4b. The analysis revealed that among the nine vegetation indices, the Structure-Insensitive Pigment Index (SIPI) exhibited the strongest correlation with yield, achieving a correlation coefficient of 0.58. The correlation coefficients for the remaining indices were all below 0.48, with the Nitrogen Reflectance Index (NRI) demonstrating the weakest correlation at 0.375.

3.3. Inversion Models of Vegetation Indices for Rice Phenotypic Data

Using data collected from four rice growth stages, including a total of 96 sets of measurements for LAI, CC, PH, and AGB, along with their corresponding UAV multispectral data for each stage and plot, inversion models were constructed to predict rice phenotypic parameters. Four modeling techniques—RF, XG Boost, SVR, and BPNN—were employed. For each phenotypic parameter, the model demonstrating the highest accuracy, defined by the highest R2 and the lowest RMSE, was selected as the final prediction model. Table 3 summarizes the performance metrics of these models in predicting each rice variety’s phenotypic data. Among these, XG Boost achieved the highest accuracy for predicting LAI (RMSE = 0.31 m2/m2, R2 = 0.83). Meanwhile, RF exhibited the best overall accuracy for estimating CC, PH, and AGB, achieving RMSE values of 1.70 SPAD, 7.38 cm, and 1.74 t/ha and respective R2 values of 0.91, 0.84, and 0.86. In contrast, SVR had the lowest prediction accuracy for all four phenotypic parameters. While BPNN accuracy was generally lower than the best-performing models, it showed stable prediction results with R2 values ranging from 0.74 to 0.79 and consistently low RMSE values across all parameters. More detailed accuracy comparison of each model and parameter are depicted in Figure 5, Figure 6, Figure 7 and Figure 8.

3.4. Construction of Yield Estimation Models

3.4.1. Comparison of Yield Estimation Model Accuracy

This study developed yield estimation models using two distinct approaches: (1) integrating actual yield data with predicted phenotypic parameter values across four growth stages as input variables and (2) combining actual yield data with vegetation indices (VIs) across the same four growth stages as inputs. For both approaches, Multiple Linear Regression (MLR) and Random Forest (RF) algorithms were applied for modeling, and the model demonstrating the highest accuracy was selected as the final rice yield estimation model. Table 4 presents the performance metrics of each model for yield estimation across the four rice growth stages. Among them, the RF model achieved the highest prediction accuracy during the HD stage, with RMSE = 0.59 t/ha and R2 = 0.86, followed by the PI stage, with RMSE = 0.69 t/ha and R2 = 0.79. The yield estimation model for the TL stage exhibited the lowest accuracy, with RMSE = 0.81 t/ha and R2 = 0.75. Overall, the MLR model consistently demonstrated lower prediction accuracy compared to the RF model, with its highest accuracy also occurring at the HD stage (RMSE = 0.83 t/ha, R2 = 0.77), followed by the MT stage (RMSE = 0.84 t/ha, R2 = 0.72). The lowest accuracy for MLR was observed at the MT stage (RMSE = 1.17 t/ha, R2 = 0.62). Detailed accuracy comparisons for each growth stage are shown in Figure 9. And the importance ranking of each feature in the optimal prediction model during the HD period is shown in Figure 10.

3.4.2. Comparison of Yield Between Conventional and Hybrid Japonica Rice

The average yield of hybrid japonica rice varieties was 5.50 t ha−1, while the average yield of conventional japonica rice was 4.38 t ha−1, representing a 25.6% yield advantage for hybrid japonica rice over conventional varieties. Among the tested varieties, the hybrid japonica rice variety CY31 achieved the highest average yield at 6.00 t ha−1, whereas the conventional japonica rice variety KY131 exhibited the lowest average yield at only 3.59 t ha−1. Yield results are presented in Figure 11.
Based on the estimates from the highest accuracy model, the predicted yield for KY131 was 3.86 t ha−1 and for SJ22 it was 4.48 t ha−1, resulting in an average predicted yield of 4.17 t ha−1 for conventional japonica rice varieties. The predicted yield for CY31 was 5.31 t ha−1 and for TLY619 it was 5.18 t ha−1, yielding an average predicted yield of 5.25 t ha−1 for hybrid japonica rice varieties.

4. Discussion

4.1. Comparison Between Direct and Phased Estimation Approaches

As shown in Table 4, the indirect estimation approach, first assessing key phenotypic parameters (PH, LAI, CC, and AGB) at critical rice growth stages using spectral indices, followed by yield estimation based on these phenotypes, demonstrated higher accuracy than direct yield estimation using spectral indices alone. Moreover, evaluating phenotypic traits at key growth stages during rice cultivation provides interpretable insights [16], offering greater practical value for agronomists in field management and decision-making.
These insights may prove useful for the integration of remotely sensed LAI and leaf nitrogen accumulation with the RiceGrow model, using a particle swarm optimization algorithm, for rice grain yield assessment or the evaluation of the feasibility of yield estimation based on leaf area dynamics measurements in farmer-managed rice paddy fields.

4.2. Optimal Growth Stage for Yield Estimation

The spectral reflectance characteristics of crop canopies at different phenological stages comprehensively reflect their internal biophysical and biochemical properties throughout growth [35]. During the vegetative stage (e.g., tillering), the rapid increase in LAI and CC enhances absorption in the red band while significantly increasing near-infrared (NIR) reflectance. As the crop enters the reproductive stage (e.g., heading), the canopy closes and LAI reaches its maximum, potentially causing saturation in broadband vegetation indices. By maturity, chlorophyll degradation and water loss lead to increased red-band reflectance and pronounced changes in short-wave infrared (SWIR) reflectance [35,36].
As shown in Table 4, yield estimation using the RF method at the HD stage achieved the highest accuracy, consistent with findings reported in [19]. This can be attributed to two main factors: (1) the heading stage represents the transition from vegetative to reproductive growth in rice, directly affecting yield-related traits such as seed-setting rate and 1000-grain weight, and (2) at this stage, the canopy is fully developed, enabling UAV-acquired imagery to effectively capture the grain-filling status of panicles. In contrast, by the maturity stage, leaf senescence and environmental interference in UAV imagery reduce prediction accuracy [10].

4.3. Comparative Analysis Machine Learning Models for Yield Estimation

Among the evaluated algorithms, linear regression, ensemble learning, and neural networks, the RF algorithm outperformed the others in phenotypic assessment across rice growth stages and in yield estimation, whereas XGBoost achieved the highest accuracy in LAI estimation. Although Multiple Linear Regression provides interpretability for individual parameters, its predictive accuracy is lower than that of ensemble learning methods due to its simplicity. Neural network algorithms, despite their computational complexity, require large datasets for effective training, leading to suboptimal performance in small-sample agronomic studies [37].
Residuals for each rice variety were calculated as the differences between predicted and actual yields, with the results presented in Figure 12. Considering the RMSE of the yield estimation model (0.59 t/ha), the residuals for conventional japonica rice varieties mostly fell within the RMSE range, with only two treatments of KY131 exceeding this threshold. In contrast, for hybrid japonica rice varieties, residuals for approximately half of the treatments exceeded the RMSE range, indicating relatively larger prediction errors.
The UAV-based remote sensing RF yield estimation model developed in this study demonstrated excellent accuracy in the Northeast region. However, its capacity for spatial and temporal generalization is limited due to region-specific environmental factors. The model was trained exclusively on data from cold-region rice-growing areas in Northeast China, without accounting for other regional rice varieties. As soil temperature critically influences material accumulation during the grain-filling stage [38], this thermal dependency may restrict the model’s applicability across diverse environmental conditions. Furthermore, the model does not consider the impacts of extreme climatic events, such as heatwaves or hail, nor does it incorporate key physiological parameters, such as the photosynthetic thermal inhibition coefficient [39]. Despite the robust performance achieved through cross-validation, our machine learning model inherits common limitations of data-driven approaches. These include dependence on large training datasets, potential performance degradation when extrapolated to regions or growing seasons with contrasting environmental conditions (e.g., different climate zones or extreme weather years), and the ongoing risk of overfitting due to site-specific noise, underscoring the need for continued validation with independent datasets.
Therefore, although the current model exhibits limitations in generalization, future research could incorporate Geographically Weighted Regression (GWR) to better capture spatially varying relationships, thereby enhancing the model’s adaptability and predictive stability across diverse regions [40]. Additionally, further studies should investigate the use of individual UAV spectral bands as direct inputs for yield modeling, alongside traditional vegetation indices, to improve feature interpretability and leverage underutilized spectral information [41].
By integrating spatially explicit modeling approaches such as GWR with richer spectral feature sets, future models are expected to achieve substantially improved spatiotemporal transferability and broader applicability under varying regional conditions.

5. Conclusions

By integrating UAV multispectral data with machine learning, this study systematically explored the feasibility of early-stage rice yield estimation. Key findings include (1) a strong correlation between rice phenotypes at critical growth stages and yield, with correlation coefficients for LAI and CC exceeding 0.85; (2) high evaluation accuracy of rice phenotypic traits using multispectral data, with the RF algorithm achieving an R2 > 0.9 for CC and R2 > 0.8 for PH and AGB and the XGBoost algorithm yielding R2 > 0.8 for LAI; and (3) optimal yield estimation performance at the HD stage, where the RF model achieved R2 = 0.86 and RMSE = 0.59 t/ha, outperforming other growth stages. In summary, this study underscores the critical influence of rice developmental stage, particularly the HD stage, and model selection on the accuracy of UAV-based yield estimation, with the RF model demonstrating superior performance. Integrating the HD stage with the RF model provides a reliable, high-throughput, and non-destructive phenotyping approach for precise early yield estimation in rice breeding and field production. These findings offer essential methodological support for improving rice breeding efficiency and advancing precision agriculture practices.

Author Contributions

Conceptualization, L.Z. and Z.Z.; methodology, X.L. (Xueyu Liang) and Z.Z.; software, L.Z. and Z.Z.; validation, X.L. (Xiao Li) and K.Z.; formal analysis, X.L. (Xueyu Liang) and X.L. (Xiao Li); investigation, X.L. (Xueyu Liang); resources, L.Z. and Z.Z.; data curation, L.Z. and Z.Z.; writing—original draft preparation, L.Z.; writing—review and editing, Z.Z. and L.Z.; visualization, L.Z. and K.Z.; supervision, Z.Z.; project administration, Z.Z.; funding acquisition, Q.C. and Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Department of Science and Technology of Heilongjiang Province, China (Grant numbers: [No. JD2023GJ01-13], [No. 2024ZX01A07], [No.2022ZX01A23]), with Northeast Agricultural University’s Zhenqing Zhao as the Principal Investigator.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest. And no author is affiliated with the UAV or sensor vendor.

Abbreviations

The following abbreviations are used in this manuscript:
UAVUnmanned Aerial Vehicle
TLtillering
PIpanicle initiation
HDheading
MTmaturity
LAIleaf area index
CCchlorophyll content
PHplant height
AGBabove-ground biomass
VIsvegetation indices
RFRandom Forest
XGBoostExtreme Gradient Boosting
SVRSupport Vector Regression
BPNNBack Propagation Neural Network
MLRMultiple Linear Regression
DSMdigital surface model
DEMdigital elevation model
NDVINormalized Difference Vegetation Index
GNDVIGreen Normalized Difference Vegetation Index
SAVISoil Adjusted Vegetation Index
OSAVIOptimized Soil Adjusted Vegetation Index
MSRModified Simple Ratio
DVIDifference Vegetation Index
NRINitrogen Reflectance Index
NDRENormalized Difference Red Edge Index
SIPIStructure Insensitive Pigment Index

Appendix A

Table A1. Optimized hyperparameters for the RF model.
Table A1. Optimized hyperparameters for the RF model.
Hyperparameter NameValue
n_estimators70
max_depth10
min_samples_split2
min_samples_leaf1
bootstrapTRUE
Table A2. Optimized hyperparameters for the XGBoost model.
Table A2. Optimized hyperparameters for the XGBoost model.
Hyperparameter NameValue
n_estimators70
learning_rate0.05
max_depth10
gamma0
subsample1
colsample_bytree1
colsample_bylevel1
lambda1
alpha0
objectivereg
tree_methodauto
Table A3. Optimized hyperparameters for the SVR model.
Table A3. Optimized hyperparameters for the SVR model.
Hyperparameter NameValue
kernelrbf
C60
gammaauto
degree3
coef00
tol0.015
Table A4. Optimized hyperparameters for the BPNN model.
Table A4. Optimized hyperparameters for the BPNN model.
Hyperparameter NameValue
learning_rate0.07
hidden_units100
epochs1000
l2_lambda0.5
activationlogistic
optimizerlbfgs
Table A5. Key specifications of the MS600pro multispectral camera.
Table A5. Key specifications of the MS600pro multispectral camera.
ParameterParameter ValueParameterParameter Value
Configuration6 Spectral BandsDimensions129 mm × 157 mm × 148 mm
Spectral Bands450 nm, 555 nm, 660 nm, 720 nm, 750 nm, 840 nmMounting InterfaceX-Port
Effective Pixels1.2 MpxPower Consumption7 W/10 W
Shutter TypeGlobal ShutterImage Format16-bit Raw TIFF
Radiometric Resolution (Bit Depth)12 bitStorage MediumStandard 64 GB, supports up to 128 GB micro SD car
Field of View (FOV)49.6° × 38°Processing SoftwareYusense Map
Ground Sampling Distance8.65 cm @ h120 mParameter ConfigurationDJI Pilot2
Swath Width110 m × 83 m @ h120 mShooting TriggerInterval Timer Trigger
ConfigurationSapphire Optical Glass WindowCapture Frequency1 Hz

References

  1. Sanyaolu, M.; Sadowski, A. The Role of Precision Agriculture Technologies in Enhancing Sustainable Agriculture. Sustainability 2024, 16, 6668. [Google Scholar] [CrossRef]
  2. Kabato, W.; Getnet, G.T.; Sinore, T.; Nemeth, A.; Molnár, Z. Towards Climate-Smart Agriculture: Strategies for Sustainable Agricultural Production, Food Security, and Greenhouse Gas Reduction. Agronomy 2025, 15, 565. [Google Scholar] [CrossRef]
  3. Weiss, M.; Jacob, F.; Duveiller, G. Remote sensing for agricultural applications: A meta-review. Remote Sens. Environ. 2020, 236, 111402. [Google Scholar] [CrossRef]
  4. Garofalo, S.P.; Modugno, A.F.; De Carolis, G.; Sanitate, N.; Negash Tesemma, M.; Scarascia-Mugnozza, G.; Tekle Tegegne, Y.; Campi, P. Explainable Artificial Intelligence to Predict the Water Status of Cotton (Gossypium hirsutum L., 1763) from Sentinel-2 Images in the Mediterranean Area. Plants 2024, 13, 3325. [Google Scholar] [CrossRef]
  5. Choudhary, K.; Shi, W.; Dong, Y.; Paringer, R. Random Forest for rice yield mapping and prediction using Sentinel-2 data with Google Earth Engine. Adv. Space Res. 2022, 70, 2443–2457. [Google Scholar] [CrossRef]
  6. Shaikh, T.A.; Rasool, T.; Lone, F.R. Towards leveraging the role of machine learning and artificial intelligence in precision agriculture and smart farming. Comput. Electron. Agric. 2022, 198, 107119. [Google Scholar] [CrossRef]
  7. Maes, W.H.; Steppe, K. Perspectives for Remote Sensing with Unmanned Aerial Vehicles in Precision Agriculture. Trends Plant Sci. 2019, 24, 152–164. [Google Scholar] [CrossRef]
  8. De Clercq, D.; Mahdi, A. Feasibility of machine learning-based rice yield prediction in India at the district level using climate reanalysis and remote sensing data. Agric. Syst. 2024, 220, 104099. [Google Scholar] [CrossRef]
  9. Basso, B.; Liu, L. Chapter Four—Seasonal crop yield forecast: Methods, applications, and accuracies. In Advances in Agronomy; Sparks, D.L., Ed.; Academic Press: Cambridge, MA, USA, 2019; Volume 154, pp. 201–255. ISBN 9780128174067. [Google Scholar] [CrossRef]
  10. Wang, F.; Wang, F.; Zhang, Y.; Hu, J.; Huang, J.; Xie, J. Rice Yield Estimation Using Parcel-Level Relative Spectral Variables From UAV-Based Hyperspectral Imagery. Front. Plant Sci. 2019, 10, 453. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  11. Wang, J.; Wu, B.; Kohnen, M.V.; Lin, D.; Yang, C.; Wang, X.; Qiang, A.; Liu, W.; Kang, J.; Li, H.; et al. Classification of Rice Yield Using UAV-Based Hyperspectral Imagery and Lodging Feature. Plant Phenomics 2021, 2021, 9765952. [Google Scholar] [CrossRef] [PubMed]
  12. Su, X.; Wang, J.; Ding, L.; Lu, J.; Zhang, J.; Yao, X.; Cheng, T.; Zhu, Y.; Cao, W.; Tian, Y. Grain yield prediction using multi-temporal UAV-based multispectral vegetation indices and endmember abundance in rice. Field Crops Res. 2023, 299, 108992. [Google Scholar] [CrossRef]
  13. Yang, G.; Li, Y.; Yuan, S.; Zhou, C.; Xiang, H.; Zhao, Z.; Wei, Q.; Chen, Q.; Peng, S.; Xu, L. Enhancing direct-seeded rice yield prediction using UAV-derived features acquired during the reproductive phase. Precis. Agric. 2024, 25, 1014–1037. [Google Scholar] [CrossRef]
  14. Wang, F.; Yao, X.; Xie, L.; Zheng, J.; Xu, T. Rice Yield Estimation Based on Vegetation Index and Florescence Spectral Information from UAV Hyperspectral Remote Sensing. Remote Sens. 2021, 13, 3390. [Google Scholar] [CrossRef]
  15. Zhou, L.; Meng, R.; Yu, X.; Liao, Y.; Huang, Z.; Lü, Z.; Xu, B.; Yang, G.; Peng, S.; Xu, L. Improved Yield Prediction of Ratoon Rice Using Unmanned Aerial Vehicle-Based Multi-Temporal Feature Method. Rice Sci. 2023, 30, 247–256. [Google Scholar] [CrossRef]
  16. Sarkar, T.K.; Roy, D.K.; Kang, Y.S.; Jun, S.R.; Park, J.W.; Ryu, C.S. Ensemble of Machine Learning Algorithms for Rice Grain Yield Prediction Using UAV-Based Remote Sensing. J. Biosyst. Eng. 2024, 49, 1–19. [Google Scholar] [CrossRef]
  17. Yang, Q.; Shi, L.; Han, J.; Zha, Y.; Zhu, P. Deep convolutional neural networks for rice grain yield estimation at the ripening stage using UAV-based remotely sensed images. Field Crops Res. 2019, 235, 142–153. [Google Scholar] [CrossRef]
  18. Sun, X.; Zhang, P.; Wang, Z.; Wang, Y. Potential of multi-seasonal vegetation indices to predict rice yield from UAV multispectral observations. Precis. Agric. 2024, 25, 1235–1261. [Google Scholar] [CrossRef]
  19. Zha, H.; Miao, Y.; Wang, T.; Li, Y.; Zhang, J.; Sun, W.; Feng, Z.; Kusnierek, K. Improving Unmanned Aerial Vehicle Remote Sensing-Based Rice Nitrogen Nutrition Index Prediction with Machine Learning. Remote Sens. 2020, 12, 215. [Google Scholar] [CrossRef]
  20. Rouse, J.W.; Haas, R.H.; Schell, J.A.; Deering, D.W. Monitoring Vegetation Systems in the Great Plains with Erts. In Proceedings of the Third NASA Earth Resources Technology Satellite Symposium, Washington, DC, USA, 10–14 December 1973; p. 351. [Google Scholar]
  21. Gitelson, A.A.; Kaufman, Y.J.; Merzlyak, M.N. Use of a green channel in remote sensing of global vegetation from EOS-MODIS. Remote Sens. Environ. 1996, 58, 289–298. [Google Scholar] [CrossRef]
  22. Huete, A.R. A soil-adjusted vegetation index (SAVI). Remote Sens. Environ. 1988, 25, 295–309. [Google Scholar] [CrossRef]
  23. Rondeaux, G.; Steven, M.; Baret, F. Optimization of soil-adjusted vegetation indices. Remote Sens. Environ. 1996, 55, 95–107. [Google Scholar] [CrossRef]
  24. Tucker, C.J. Red and photographic infrared linear combinations for monitoring vegetation. Remote Sens. Environ. 1979, 8, 127–150. [Google Scholar] [CrossRef]
  25. Filella, I.; Serrano, L.; Serra, J.; Peñuelas, J. Evaluating Wheat Nitrogen Status with Canopy Reflectance Indices and Discriminant Analysis. Crop Sci. 1995, 35, 1400–1405. [Google Scholar] [CrossRef]
  26. Jordan, C.F. Derivation of Leaf-Area Index from Quality of Light on the Forest Floor. Ecology 1969, 50, 663–666. [Google Scholar] [CrossRef]
  27. Barnes, E.M.; Clarke, T.R.; Richards, S.E.; Colaizzi, P.D.; Haberland, J.; Kostrzewski, M.; Waller, P.; Choi, C.; Riley, E.; Thompson, T.; et al. Coincident Detection of Crop Water Stress, Nitrogen Status and Canopy Density Using Ground-Based Multispectral Data. In Proceedings of the Fifth International Conference on Precision Agriculture and Other Resource Management, Bloomington, MN, USA, 16–19 July 2000; ASA-CSSA-SSSA: Madison, WI, USA, 2000. [Google Scholar]
  28. Peñuelas, J.; Filella, I.; Gamon, J.A. Assessment of photosynthetic radiation-use efficiency with spectral reflectance. New Phytol. 1995, 131, 291–296. [Google Scholar] [CrossRef]
  29. Field, A. Discovering Statistics Using IBM SPSS Statistics, 5th ed.; Sage: Thousand Oaks, CA, USA, 2018. [Google Scholar]
  30. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  31. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ’16), San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  32. Smola, A.J.; Schölkopf, B. A tutorial on support vector regression. Stat. Comput. 2004, 14, 199–222. [Google Scholar] [CrossRef]
  33. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  34. Wooldridge, J.M. Introductory Econometrics: A Modern Approach, 7th ed.; Cengage Learning: Boston, MA, USA, 2019; pp. 39–40, 84–85. [Google Scholar]
  35. Zheng, H.; Cheng, T.; Yao, X.; Deng, X.; Tian, Y.; Cao, W.; Zhu, Y. Detection of rice phenology through time series analysis of ground-based spectral index data. Field Crops Res. 2016, 198, 131–139. [Google Scholar] [CrossRef]
  36. Sakamoto, T.; Wardlow, B.D.; Gitelson, A.A.; Verma, S.B.; Suyker, A.E.; Arkebauer, T.J. A Two-Step Filtering approach for detecting maize and soybean phenology with time-series MODIS data. Remote Sens. Environ. 2010, 114, 2146–2159. [Google Scholar] [CrossRef]
  37. Kang, Y.; Nam, J.; Kim, Y.; Lee, S.; Seong, D.; Jang, S.; Ryu, C. Assessment of Regression Models for Predicting Rice Yield and Protein Content Using Unmanned Aerial Vehicle-Based Multispectral Imagery. Remote Sens. 2021, 13, 1508. [Google Scholar] [CrossRef]
  38. Li, H.; Wang, X.; Duan, J.; Yin, C. Heading stage as the critical window for UAV-based yield prediction in cold-region rice. Precis. Agric. 2023, 24, 1021–1037. [Google Scholar]
  39. Yin, C.; Zhang, K.; Zhao, H.; Li, T. Robustness of machine learning models with limited samples: A case study of rice yield prediction in Heilongjiang Province. Eur. J. Agron. 2023, 142, 126702. [Google Scholar]
  40. Haghighattalab, A.; Crain, J.; Mondal, S.; Rutkoski, J.; Singh, R.P.; Poland, J. Application of geographically weighted regression to improve grain yield prediction from unmanned aerial system imagery. Crop Sci. 2017, 57, 2478–2489. [Google Scholar] [CrossRef]
  41. Delegido, J.; Verrelst, J.; Meza, C.; Rivera, J.; Alonso, L.; Moreno, J. A red-edge spectral index for remote sensing estimation of green LAI over agroecosystems. Eur. J. Agron. 2013, 46, 42–52. [Google Scholar] [CrossRef]
Figure 1. Workflow diagram of the study. The diagram illustrates the key steps of the research process: (1) collection of field data, including chlorophyll content (CC), plant height, biomass, and leaf area index (LAI); (2) acquisition and preprocessing of UAV multispectral data, involving image stitching, cropping, and extraction of vegetation indices; (3) development of inversion models (RF, XGBoost, SVR, BPNN) by correlating vegetation indices with field data, followed by selection of the most accurate model for field inversion; (4) construction of a rice yield estimation model using correlation analysis between yield and inversion data, comparing linear and nonlinear models across different phenological stages via time-series analysis.
Figure 1. Workflow diagram of the study. The diagram illustrates the key steps of the research process: (1) collection of field data, including chlorophyll content (CC), plant height, biomass, and leaf area index (LAI); (2) acquisition and preprocessing of UAV multispectral data, involving image stitching, cropping, and extraction of vegetation indices; (3) development of inversion models (RF, XGBoost, SVR, BPNN) by correlating vegetation indices with field data, followed by selection of the most accurate model for field inversion; (4) construction of a rice yield estimation model using correlation analysis between yield and inversion data, comparing linear and nonlinear models across different phenological stages via time-series analysis.
Sustainability 17 08515 g001
Figure 2. Overview of the experimental study area. The UAV image depicts the layout of the experimental plots. (a) Map of China; (b) map of Heilongjiang Province; and (c) map of Yanjiagang Farm, where V1, V2, V3, and V4 denote the four rice varieties, Kongyu131, Songjing22, Chuangyou31, and Tianlongyou619, respectively, and N1–N6 indicate different nitrogen treatments with increasing concentrations. Projection system: WGS 84.
Figure 2. Overview of the experimental study area. The UAV image depicts the layout of the experimental plots. (a) Map of China; (b) map of Heilongjiang Province; and (c) map of Yanjiagang Farm, where V1, V2, V3, and V4 denote the four rice varieties, Kongyu131, Songjing22, Chuangyou31, and Tianlongyou619, respectively, and N1–N6 indicate different nitrogen treatments with increasing concentrations. Projection system: WGS 84.
Sustainability 17 08515 g002
Figure 3. Comparison of measured phenotypic parameters among four varieties (KY131, SJ22, CY31, TLY619) at four key growth stages (TL, PI, HD, MT). (a) PH; (b) CC; (c) LAI; (d) AGB. Note: The red line in the figure represents the mean value of each variety.
Figure 3. Comparison of measured phenotypic parameters among four varieties (KY131, SJ22, CY31, TLY619) at four key growth stages (TL, PI, HD, MT). (a) PH; (b) CC; (c) LAI; (d) AGB. Note: The red line in the figure represents the mean value of each variety.
Sustainability 17 08515 g003aSustainability 17 08515 g003b
Figure 4. Spearman correlation analysis: (a) between rice phenotypic traits and yield; (b) between vegetation indices and yield. Note: *** and ** indicate significant correlations at p < 0.01 and p < 0.05, respectively.
Figure 4. Spearman correlation analysis: (a) between rice phenotypic traits and yield; (b) between vegetation indices and yield. Note: *** and ** indicate significant correlations at p < 0.01 and p < 0.05, respectively.
Sustainability 17 08515 g004
Figure 5. Validation of UAV-derived LAI prediction models using vegetation indices. Scatter plots of ground-truth vs. predicted values: (A) RF; (B) XGBoost; (C) SVR; (D) BPNN.
Figure 5. Validation of UAV-derived LAI prediction models using vegetation indices. Scatter plots of ground-truth vs. predicted values: (A) RF; (B) XGBoost; (C) SVR; (D) BPNN.
Sustainability 17 08515 g005
Figure 6. Validation of UAV-derived CC prediction models using vegetation indices. Scatter plots of ground-truth vs. predicted values: (A) RF; (B) XGBoost; (C) SVR; (D) BPNN.
Figure 6. Validation of UAV-derived CC prediction models using vegetation indices. Scatter plots of ground-truth vs. predicted values: (A) RF; (B) XGBoost; (C) SVR; (D) BPNN.
Sustainability 17 08515 g006
Figure 7. Validation of UAV-derived PH prediction models using vegetation indices. Scatter plots of ground-truth vs. predicted values: (A) RF; (B) XGBoost; (C) SVR; (D) BPNN.
Figure 7. Validation of UAV-derived PH prediction models using vegetation indices. Scatter plots of ground-truth vs. predicted values: (A) RF; (B) XGBoost; (C) SVR; (D) BPNN.
Sustainability 17 08515 g007aSustainability 17 08515 g007b
Figure 8. Validation of UAV-derived AGB prediction models using vegetation indices: Scatter plots of ground-truth vs. predicted values: (A) RF; (B) XGBoost; (C) SVR; (D) BPNN.
Figure 8. Validation of UAV-derived AGB prediction models using vegetation indices: Scatter plots of ground-truth vs. predicted values: (A) RF; (B) XGBoost; (C) SVR; (D) BPNN.
Sustainability 17 08515 g008
Figure 9. Performance comparison of rice yield models across growth stages, evaluating RF and MLR algorithms with phenotypic parameters. (A) TL stage, (B) PI stage, (C) HD stage, (D) MT stage.
Figure 9. Performance comparison of rice yield models across growth stages, evaluating RF and MLR algorithms with phenotypic parameters. (A) TL stage, (B) PI stage, (C) HD stage, (D) MT stage.
Sustainability 17 08515 g009aSustainability 17 08515 g009b
Figure 10. Feature importance of the RF yield estimation model at the HD stage.
Figure 10. Feature importance of the RF yield estimation model at the HD stage.
Sustainability 17 08515 g010
Figure 11. Comparison of actual yield differences among varieties. Note: Yield differences were observed between conventional japonica rice varieties (KY131, SJ22) and hybrid japonica rice varieties (CY31, TLY619). The bars represent mean ± standard deviation (SD), with data based on 6 treatments and 3 field replicates.
Figure 11. Comparison of actual yield differences among varieties. Note: Yield differences were observed between conventional japonica rice varieties (KY131, SJ22) and hybrid japonica rice varieties (CY31, TLY619). The bars represent mean ± standard deviation (SD), with data based on 6 treatments and 3 field replicates.
Sustainability 17 08515 g011
Figure 12. Residual plot of the yield estimation model across rice varieties. Each point represents the average yield across six treatments for the corresponding variety. The area between the dashed lines indicates the interval within which the prediction error is less than the model’s RMSE.
Figure 12. Residual plot of the yield estimation model across rice varieties. Each point represents the average yield across six treatments for the corresponding variety. The area between the dashed lines indicates the interval within which the prediction error is less than the model’s RMSE.
Sustainability 17 08515 g012
Table 1. Information on tested rice varieties used in the experiment.
Table 1. Information on tested rice varieties used in the experiment.
GroupVarietyAbbreviationVariety CharacteristicsDays to Maturity
Conventional japonica riceKongyu131KY131Early-maturing127 D
Songjing22SJ22Late-maturing144 D
Hybrid japonica riceChuangyou31CY31Early-maturing129 D
Tianlongyou619TLY619Late-maturing139 D
Information was retrieved from the China Rice Data Center (https://www.ricedata.cn/variety/) Accessed on 1 May 2025.
Table 2. Vegetation indices and their formulas.
Table 2. Vegetation indices and their formulas.
Vegetation IndexFormulaReference
NDVI ( N I R R E D ) / ( N I R + R E D ) Rouse, J. W et al., 1974 [20]
GNDVI ( N I R G R E E N ) / ( N I R + G R E E N ) Anatoly A et al., 1996 [21]
SAVI ( ( N I R R E D ) × ( 1 + L ) ) / ( N I R + R E D + L ) ,   L = 0.5 A.R Huete et al., 1988 [22]
OSAVI ( N I R R E D ) / ( N I R + R E D + 0.16 ) Geneviève Rondeaux et al., 1988 [23]
MSR ( ( N I R / R E D ) 1 ) / ( ( N I R / R E D ) + 1 ) Compton J et al., 1979 [24]
DVI N I R R E D Filella, 1995 [25]
NRI ( G R E E N R E D ) / ( G R E E N + R E D ) Jordan et al., 1969 [26]
NDRE ( N I R R E ) / ( N I R + R E ) Barnes et al., 2000 [27]
SIPI ( N I R B L U E ) / ( N I R R E D ) Peñuelas et al., 1995 [28]
Table 3. Comparison of the RMSE and R2 obtained by RF, XG Boost, SVR, and BPNN regression models for predicting four rice phenotypic parameters: LAI, CC, PH, and AGB.
Table 3. Comparison of the RMSE and R2 obtained by RF, XG Boost, SVR, and BPNN regression models for predicting four rice phenotypic parameters: LAI, CC, PH, and AGB.
Model Type LAICCPHAGB
RFRMSE0.38 (m2/m2)1.70 (SPAD)7.38 (cm)1.74 (t/ha)
R20.780.910.840.86
XG BoostRMSE0.31 (m2/m2)2.76 (SPAD)7.52 (cm)1.94 (t/ha)
R20.830.840.810.82
SVRRMSE0.40 (m2/m2)4.30 (SPAD)10.32 (cm)3.46 (t/ha)
R20.750.540.670.61
BPNNRMSE0.48 (m2/m2)2.74 (SPAD)8.67 (cm)1.85 (t/ha)
R20.780.740.760.73
Both R2 and RMSE in this table refer to the accuracy of the test set.
Table 4. Accuracy comparison of yield estimation models: predicted phenotypic parameters vs. direct vegetation indices (VIs) using RF and MLR algorithms across four growth stages: TL, PI, HD, and MT.
Table 4. Accuracy comparison of yield estimation models: predicted phenotypic parameters vs. direct vegetation indices (VIs) using RF and MLR algorithms across four growth stages: TL, PI, HD, and MT.
Model Type TLPIHDMT
Phenotypic ParametersVIsPhenotypic ParametersVIsPhenotypic ParametersVIsPhenotypic ParametersVIs
MLRRMSE1.17 (t/ha)1.35 (t/ha)0.88 (t/ha)1.10 (t/ha)0.83 (t/ha)1.44 (t/ha)0.84 (t/ha)0.80 (t/ha)
R20.620.310.700.390.770.410.720.37
RFRMSE0.81 (t/ha)1.24 (t/ha)0.69 (t/ha)0.82 (t/ha)0.59 (t/ha)1.06 (t/ha)0.73 (t/ha)1.46 (t/ha)
R20.750.440.790.380.860.500.780.38
Both R2 and RMSE in this table refer to the accuracy of the test set.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Liang, X.; Li, X.; Zeng, K.; Chen, Q.; Zhao, Z. Machine Learning Models for Yield Estimation of Hybrid and Conventional Japonica Rice Cultivars Using UAV Imagery. Sustainability 2025, 17, 8515. https://doi.org/10.3390/su17188515

AMA Style

Zhang L, Liang X, Li X, Zeng K, Chen Q, Zhao Z. Machine Learning Models for Yield Estimation of Hybrid and Conventional Japonica Rice Cultivars Using UAV Imagery. Sustainability. 2025; 17(18):8515. https://doi.org/10.3390/su17188515

Chicago/Turabian Style

Zhang, Luyao, Xueyu Liang, Xiao Li, Kai Zeng, Qingshan Chen, and Zhenqing Zhao. 2025. "Machine Learning Models for Yield Estimation of Hybrid and Conventional Japonica Rice Cultivars Using UAV Imagery" Sustainability 17, no. 18: 8515. https://doi.org/10.3390/su17188515

APA Style

Zhang, L., Liang, X., Li, X., Zeng, K., Chen, Q., & Zhao, Z. (2025). Machine Learning Models for Yield Estimation of Hybrid and Conventional Japonica Rice Cultivars Using UAV Imagery. Sustainability, 17(18), 8515. https://doi.org/10.3390/su17188515

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop