Next Article in Journal
Genetic Diversity and Population Structure of Wheat Germplasm for Grain Nutritional Quality Using Haplotypes and KASP Markers
Previous Article in Journal
Harvest Date Monitoring in Cereal Fields at Large Scale Using Dense Stacks of Sentinel-2 Imagery Validated by Real Time Kinematic Positioning Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Framework for Predicting Daily Reference Evapotranspiration Using Interpretable Machine Learning Techniques

by
Elsayed Ahmed Elsadek
1,2,†,
Mosaad Ali Hussein Ali
3,†,
Clinton Williams
4,
Kelly R. Thorp
5 and
Diaa Eldin M. Elshikha
1,*
1
Biosystems Engineering Department, University of Arizona, Tucson, AZ 85721, USA
2
Agricultural and Biosystems Engineering Department, College of Agriculture, Damietta University, Damietta 34517, Egypt
3
Mining and Metallurgical Engineering Department, Faculty of Engineering, Assiut University, Assiut 71511, Egypt
4
United States Department of Agriculture (USDA)—Agricultural Research Service (ARS), Arid Land Agricultural Research Center, Maricopa, AZ 85138, USA
5
Grassland Soil & Water Research Laboratory, United States Department of Agriculture (USDA)—Agricultural Research Service (ARS), Temple, TX 76502, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Agriculture 2025, 15(18), 1985; https://doi.org/10.3390/agriculture15181985
Submission received: 28 August 2025 / Revised: 16 September 2025 / Accepted: 16 September 2025 / Published: 20 September 2025
(This article belongs to the Section Agricultural Water Management)

Abstract

Accurate estimation of daily reference evapotranspiration (ETo) is crucial for sustainable water resource management and irrigation scheduling, especially in water-scarce regions like Arizona. The standardized Penman–Monteith (PM) method is costly and requires specialized instruments and expertise, making it generally impractical for commercial growers. This study developed 35 ETo models to predict daily ETo across Coolidge, Maricopa, and Queen Creek in Pinal County, Arizona. Seven input combinations of daily meteorological variables were used for training and testing five machine learning (ML) models: Artificial Neural Network (ANN), Random Forest (RF), Extreme Gradient Boosting (XGBoost), Categorical Boosting (CatBoost), and Support Vector Machine (SVM). Four statistical indicators, coefficient of determination (R2), the normalized root-mean-squared error (RMSEn), mean absolute error (MAE), and simulation error (Se), were used to evaluate the ML models’ performance in comparison with the FAO-56 PM standardized method. The SHapley Additive exPlanations (SHAP) method was used to interpret each meteorological variable’s contribution to the model predictions. Overall, the 35 ETo-developed models showed an excellent to fair performance in predicting daily ETo over the three weather stations. Employing ANN10, RF10, XGBoost10, CatBoost10, and SVM10, incorporating all ten meteorological variables, yielded the highest accuracies during training and testing periods (0.994 ≤ R2 ≤ 1.0, 0.729 ≤ RMSEn ≤ 3.662, 0.030 ≤ MAE ≤ 0.181 mm·day−1, and 0.833 ≤ Se ≤ 2.295). Excluding meteorological variables caused a gradual decline in ET-developed models’ performance across the stations. However, 3-variable models using only maximum, minimum, and average temperatures (Tmax, Tmin, and Tave) predicted ETo well across the three stations during testing (17.655 ≤ RMSEn ≤ 13.469 and Se ≤ 15.45%). Results highlighted that Tmax, solar radiation (Rs), and wind speed at 2 m height (U2) are the most influential factors affecting ETo at the central Arizona sites, followed by extraterrestrial solar radiation (Ra) and Tave. In contrast, humidity-related variables (RHmin, RHmax, and RHave), along with Tmin and precipitation (Pr), had minimal impact on the model’s predictions. The results are informative for assisting growers and policymakers in developing effective water management strategies, especially for arid regions like central Arizona.

1. Introduction

In the desert U.S. Southwest, agroecosystems are vulnerable to water stress due to the ongoing drought in the Colorado River basin [1,2], depletion of groundwater and reservoir water supplies, and increased distribution of water for industrial and municipal needs [2,3,4,5]. A great challenge for agriculture in this region is to find solutions for sustaining crop production under the reduced water allocations mandated for many of the water districts [6,7] and climate change, which significantly impacts the agricultural system and accelerates the inevitable risk of water scarcity [8,9,10]. Given the current and future water scarcity and climate change challenges facing the agriculture sector [11], improving existing irrigation management practices for obtaining economic yields with less water is critical. Water scarcity requires more precise use of irrigation water [12], where irrigation timing and amount are synchronized to meet crop water requirements and sustain plant-available water status [13,14]. The crop water requirement is closely related to crop evapotranspiration (ET), which is the sum of (1) the losses of water to air from the soil surface evaporation and (2) plant transpiration of soil water from the crop root zone [15]. In arid and semi-arid climates in the U.S., ET constitutes a large majority of the water consumption from irrigated fields [16]. Thus, improvements in irrigation management and water use efficiency require accurate estimates of ET to help guide appropriate irrigation scheduling decisions for growers [17].
During the past few decades, various measurement techniques have been shown to accurately quantify the actual crop ET in fields, including soil water balance (SWB) methods, lysimeters, Bowen ratio systems, and eddy covariance (EC) systems [18]. However, use of these methods is costly and time-consuming, and generally impractical for commercial growers. Moreover, they require expertise and knowledge of the instruments used [19]. Nevertheless, some cost-effective ET tools have been developed, such as widely used crop coefficient (Kc) methods, where the ET for a crop is estimated by multiplying crop-specific, time-based Kc values by a weather-based reference ET [20]. The Penman–Monteith model (PM) is the standard method for estimating reference evapotranspiration (ETo) and a benchmark model for calibrating other ETo models. Measuring all the input variables needed for the PM is costly and labor-intensive, requiring specialized instrumentation and expertise. Moreover, most of these variables may not be available or of low quality, especially in developing countries, impeding its universal application [21]. Therefore, empirical models such as the Hargreaves and Samani model [22], a simplified temperature-based equation that utilizes maximum, minimum, and average temperatures, have been developed for calculating ETo [9,10]. However, ETo estimates based on such equations have shown inconsistencies compared to the PM in different geographical regions worldwide [23,24,25].
To date, artificial intelligence techniques have exhibited rapid expansion, offering powerful alternative tools for ETo modeling [26,27,28,29], especially through machine learning (ML) models [30]. However, no cited studies have used advanced ML techniques to accurately predict daily ETo under arid climate conditions in Arizona, USA. To fill the scientific gap and provide a reliable and interpretable data-driven framework to support irrigation planning and water resource management in water-scarce regions, such as Arizona, the main objectives of this study were to (1) develop and evaluate five advanced ML models, namely Artificial Neural Network (ANN), Random Forest (RF), Extreme Gradient Boosting (XGBoost), Categorical Boosting (CatBoost), and Support Vector Machine (SVM) for the accurate prediction of daily ETo based on ten weather variables (maximum, minimum, and average temperature [Tmax, Tmin, and Tave], precipitation [Pr], maximum, minimum, and average relative humidity [RHmax, RHmin, and RHave], wind speed at 2 m height [U2], extraterrestrial solar radiation [Ra], and solar radiation [Rs]), (2) employ the SHapley Additive exPlanations (SHAP) method to interpret the contribution of each meteorological variable to model predictions, and (3) use each ML algorithm to develop ETo models with different weather input variables, resulting in varying levels of model complexity. To ensure their accuracy, the 35 ETo-developed models were trained and tested at three weather stations in Pinal County, Arizona, from January 1995 to June 2025.

2. Materials and Methods

2.1. Study Area and Datasets

Pinal County, located in central Arizona, USA (latitude 32.90431° N, longitude 111.34471° W), covers an area of approximately 13,900 km2 (Figure 1). The region has an arid climate, where average annual precipitation is about 150–200 mm. Hot and dry conditions prevail from late spring through fall, with maximum daily air temperatures typically reaching 43.0 °C to 48.0 °C during summer. Pinal County was selected because agriculture accounts for nearly 92.3% of the county’s water consumption [31]. Most of this water was sourced from the Colorado River [32]. Irrigation water is predominantly consumed for irrigating highly water-intensive crops, such as alfalfa and cotton [33]. Recently, Arizona’s Colorado River supply was reduced by 18–21% due to the mandated water rationing in the state, where private growers in the central and southern counties of the state are primarily bearing the brunt of the reductions [34].
Daily meteorological data, including Tmax, Tmin, Tave, Pr, RHmax, RHmin, RHave, U2, Ra, Rs, and ETo, during the period from January 1995 to June 2025, were collected from Arizona Meteorological Network (AZMET) stations in Coolidge, Maricopa, and Queen Creek (https://cals.arizona.edu/AZMET/26.htm, accessed on 1 July 2025). The meteorological variables were input to the five ML models to project daily ETo as a target output variable across the three weather stations in Pinal County, Arizona. The meteorological variables were selected based on their Kendall correlation (Figure 2) and significance in predicting ETo [23,35,36]. The statistics of the datasets used in the current study are summarized in Table 1 and Figure 2.

2.2. Machine Learning Models

2.2.1. Artificial Neural Network (ANN)

The ANN is a mathematical ML model that simulates the human brain learning from trials and errors. Like the human brain, an ANN consists of neurons that arrange processing elements (PEs) in three layers: input, hidden, and output. In the input layer, data are introduced into the ANN, and then the data are processed in the hidden layer. Finally, the results are produced in the output layer. The three layers are fully interconnected by interconnection weights, similar to the synapses in biological neurons. Each PE (neuron) consists of a part to sum the weighted inputs, add a bias term, and then pass the results (activation values) by another part called the activation or transfer function. The activation function (f) works as a nonlinear filter to squash the output values of an artificial neuron to values between two asymptotes, typically between zero and 1 or −1 and 1.
PE output , j = f ( I j )
I j = i = 1 n ( w i j   x i + b i )
where PEoutput,j is the output of a processing element or a neuron j, Ij is the sum of weighted inputs for a neuron j in the i-th layer, wij is the weight between the i-th input and the neuron j of the hidden layer. xi and bi are the i-th input value and bias term.
ANN has been used in hydrological studies to approximate unknown functions or predict values based on a muddled time series of data [37]. For further information about ANN, see Kumar et al. [38].

2.2.2. Random Forest (RF)

The RF, developed by Breiman [39], is a powerful ML algorithm widely used for regression and classification tasks. The concept of RF is to ensemble a set of regression or classification trees. Each tree in the forest grows upon a diverse learning dataset, which is called bagging or bootstrap sampling. In each bootstrap, a random sample is drawn from the original dataset with replacement to train each decision tree and generate a regression tree during the prediction by the RF. Then, each tree is split randomly, using a subset of features, to construct multiple trees.
The splitting algorithm follows a hierarchical structure and is implemented in a binary manner. Thus, at each step, it identifies the optimal split points and the best splitting variable among all input variables and all possible split points, aiming to minimize the sum of squared residuals.
Three training factors control how the forest is built: (1) the number of trees (ntree), (2) the number of input variables randomly tried at each split (mtry), and (3) the maximum number of terminal nodes per tree (maxnode). A full description of RF can be found in Breiman [39].

2.2.3. XGBoost

Extreme Gradient Boosting (XGBoost) is a scalable end-to-end tree boosting algorithm, widely recognized as an effective machine learning method. The original framework of XGBoost involves aggregating numerous “weak” learners into a “strong” learner through an iterative process known as gradient boosting decision trees. In each iteration, the model updates by correcting previous predictions based on residuals. The algorithm allows for the independent selection of the loss function appropriate for model evaluation. Furthermore, a regularization term is incorporated into the model to mitigate the risk of overfitting. The mean score of each tree serves as the predictive value in regression tasks. For the m-th decision tree, the XGBoost formula is expressed as follows:
y i ^ = m = 1 M f m x i ,   f m Ƒ
where y i ^ is the predicted value for the i-th sample, M is the total number of trees (boosting rounds), f m x i is the prediction from the m-th regression tree for input xi, and Ƒ is the space of all possible regression trees. Each f m is a tree structure that maps xi to a leaf score. A full description of XGBoost can be found in Chen and Guestrin [40].

2.2.4. CatBoost

CatBoost is a powerful new gradient boosting (GB) toolkit, especially designed to handle categorical features with minimal loss of information. CatBoost has both a Graphics Processing Unit (GPU) and a Central Processing Unit (CPU) implementation of learning and scoring algorithms, respectively, making it outperform other GB algorithms on ensembles of similar sizes [41].
CatBoost is distinguished by two crucial algorithmic advances: (1) implementing ordered boosting, a permutation-driven alternative to overcome the overfitting caused by the classical GB algorithms, and (2) an innovative algorithm for handling categorical features. Both were developed to address the prediction shift caused by a specific form of target leakage inherent in all existing implementations of GB algorithms [42].
Binary decision trees are used as a base predictor in CatBoost, whereas a decision tree (h) is written as follows [42].
f x i = j = 1 J b j   1 { x R j }
where f(xi) is the prediction/forecast made by a decision tree for the i-th variable x, J is the total number of terminal nodes/leaves in the decision tree, Rj is the region/subset of feature space corresponding to the j-th leaf, bj is the predicted value in region Rj, 1{xRj} is an indicator function, equal to one if x falls in the region Rj or zero otherwise.

2.2.5. Support Vector Machine (SVM)

The SVM, originally introduced by Cortes and Vapnik [43], is a machine learning algorithm designed for two-class issues. Theoretically, SVM is a kernel trick-based algorithm [44], whereas it works by mapping input vectors nonlinearly into a very high-dimensional feature space. Within this space, a linear decision boundary is established. The distinctive characteristics of this decision boundary ensure the learning machine has a high generalization ability.
f x = i = 1 n i K ( x ,   x i ) + b
where f(x) is the predicted/forecasted value for an input x, ∝i is the Lagrange multiplier for under/overestimation, K(x, xi) is the Kernel function, and b is a bias term (intercept).
As a powerful and good function for nonlinear data, the Kernel Radial Basis Function (RBF) was used in the current study to predict ETo.
K x ,   x i = e x p ( γ   x x i 2 )

2.3. Models’ Hyperparameter Settings

Table 2 summarizes the default and tuned hyperparameters for each machine learning model used in ETo prediction. The hyperparameters of each machine learning model were selected based on the best practices and empirical performance. For the ANN, a two-hidden-layer architecture was adopted using MLPRegressor (hidden_layer_sizes = (100, 50), max_iter = 1000, random_state = 42), where the increased layer depth and iteration count enhance the model’s capacity to capture complex, nonlinear relationships in meteorological data. The SVM model employed a radial basis function kernel [SVR (kernel = ‘rbf’)], which is well-suited for handling nonlinear regression problems, with other parameters left at default to maintain simplicity. The RF was configured with n_estimators = 100 and random_state = 42, striking a balance between accuracy and computational efficiency while leveraging the ensemble’s robustness to hyperparameter settings. For the XGBoost, the model was initialized as XGBRegressor (n_estimators = 100, learning_rate = 0.1, random_state = 42), reflecting standard configurations that offer strong performance across a wide range of tabular datasets. Lastly, the CatBoost was used with verbose = 0 to suppress console output and random_state = 42 for reproducibility, as it typically achieves high accuracy with minimal tuning. These configurations were chosen to establish a strong baseline performance while ensuring stability and comparability across the ML models. In this study, all ML models were implemented using the open-source scikit-learn library (Version 1.7.1) within the Python 3.13.3 environment.

2.4. Input Combinations

The current study used seven input combinations of daily meteorological variables as a running scenario for training and testing the five ML models on predicting ETo across Pinal County, Arizona (Table 3). The seven combinations consist of a complete combination with all meteorological variables and the incomplete combinations with part of the variables, reflecting different levels of complexity: (1) Tmax, Tmin, Tave, Pr, RHmax, RHmin, RHave, U2, Ra, and Rs; (2) Tmax, Tmin, Tave, Pr, RHave, U2, Ra, and Rs; (3) Tmax, Tmin, Tave, Pr, RHave, U2, and Rs; (4) Tmax, Tmin, Tave, Pr, RHave, and U2; (5) Tmax, Tmin, Tave, Pr, and RHave; (6) Tmax, Tmin, Tave, and Pr; while the last is a simplified temperature-based combination [(7) Tmax, Tmin, and Tave]. Recently, different approaches have been adopted for splitting datasets into training and testing subsets, with a 70:30 division being the most frequently applied [23,45]. Following this approach, the daily meteorological data from January 1995 to June 2025 were randomly partitioned into 70% for training and 30% for testing, ensuring a reliable basis for developing and validating the ETo prediction models.

2.5. SHAP: SHapley Additive exPlanations

SHAP is a framework used for understanding a ML model’s prediction by attributing the contribution of each feature to the predicted outcome [46]. SHAP values, which are rooted in cooperative game theory, endeavor to provide a comprehensive approach to explicate the output of intricate models [47]. These values capture the marginal impact of each feature on the prediction while considering the interplay between features and assessing all feasible combinations. Consequently, SHAP represents a valuable tool for comprehending and quantifying the significance of each feature in a predictive model and thus discerning which features propel specific predictions and how much they contribute.

2.6. Evaluation Indices

Four statistical indicators, namely, coefficient of determination (R2), the normalized root-mean-squared error (RMSEn), mean absolute error (MAE), and simulation error (Se), were used to evaluate the performance of the five ML algorithms.
R 2 = i = 1 n E T o b s E T o b s ( E T p r e E T p r e ) i = 1 n E T o b s E T o b s 2 i = 1 n ( E T p r e E T p r e ) 2 2
R M S E n = 100 E T o b s i = 1 n E T p r e E T o b s 2 n
M A E = i = 1 n | E T p r e E T o b s | n
S e = E T p r e E T o b s E T o b s × 100 %
where ETobs and ETpre refer to observed and predicted ET. E T o b s and E T p r e are the means of observed and predicted ET, respectively. The R2 values reflect the strength of the predicted datasets. The R2 close to 1.0 indicates a good match between the predicted and observed datasets. The RMSEn value ranges between 0 and ∞. The simulation is excellent if the estimation result of RMSEn <   10% and good if the result ranges between 10 and 20%. A result from 20 to 30% indicates fair performance, while RMSEn > 30% indicates poor performance [48]. The MAE reflects the difference between the mean absolute value of the projection/modeling and the observed value. The lower the MAE value, the greater the projection/modeling accuracy. The Se between ±15% is acceptable [49].

3. Results and Discussion

3.1. Coolidge Station

Table 4 summarizes the performance evaluation results of the 35 developed ETo models across the Coolidge weather station, Pinal County, Arizona. Considering the 10-variable models, which require Tmax, Tmin, Tave, Pr, RHmax, RHmin, RHave, U2, Ra, and Rs, the average evaluation indices were 1.305 ≤ RMSEn ≤ 3.485, and 0.890 ≤ Se ≤ 2.295 during the testing period (Table 4). Meanwhile, all 10-variable models showed a slight tendency to overestimate ETo by 0.045–0.115 mm·day−1.
For all ML models, the scatter plots of the test set across Coolidge station showed a linear-positive distribution between ETobs and ETpre with 0.994 ≤ R2 ≤ 0.999 (Table 4 and Figure 3).
Based on violin diagrams and performance indices (Figure 4 and Table 4), there were no obvious differences in the structure and distribution of all ML algorithms compared with the actual datasets, indicating excellent accuracy in predicting ETo.
The CatBoost algorithm, CatBoost10 model, outperformed other models with R2 = 0.999, RMSEn = 1.305%, MAE = 0.045 mm·day−1, and Se = 0.890%, making it the best choice to predict ETo over Coolidge using the ten weather parameters.
Figure 5 presents the SHAP beeswarm plot for the CatBoost10 model, illustrating each input feature’s contribution to the ETo prediction. The features are ranked by importance from top to bottom, based on their average absolute SHAP values. The SHAP values on the x-axis represent the impact of each feature on the model output in terms of the ETo prediction in mm·day−1, where positive values indicate an increase and negative values a decrease in predicted ETo. Each dot corresponds to a sample, with its color reflecting the feature’s actual value (red for high, blue for low). These values quantify how much each feature shifts the model’s output from the base value, i.e., the average model prediction, either positively or negatively. For instance, a SHAP value of +2 for Tmax indicates that this feature has increased the predicted ETo by 2 mm·day−1 for a specific instance. Among all variables, Tmax, Rs, and U2 had the most significant positive influence on ETo predictions, as high values of these parameters tend to enhance evapotranspiration. Ra and Tave also showed moderate positive effects. Conversely, the variables related to relative humidity (RHmin, RHmax, and RHave) showed an adverse effect on ETo, resulting in reduced SHAP values. Tmin and Pr had the least influence, suggesting limited direct relevance to the model’s prediction of ETo under the given conditions.
Figure 6 displays a SHAP bar plot illustrating the mean absolute SHAP values for each input feature in predicting ETo. The x-axis represents the average magnitude of a feature’s contribution to the model’s output, expressed in mm·day−1. Importantly, these values reflect the absolute influence of each variable regardless of whether the effect increases or decreases the average ETo prediction across all data points. Thus, a higher value on the x-axis signifies a more substantial and consistent role in shaping the model’s predictions, rather than a directional change. For instance, a mean absolute SHAP value of 0.6 mm·day−1 for Rs means that, on average, solar radiation contributes ± 0.6 mm·day−1 to the model’s prediction of daily ETo, regardless of whether it increases or decreases the prediction. Among the considered variables, maximum temperature emerged as the most influential factor, followed by solar radiation and wind speed at 2 m height (Tmax, Rs, and U2). Ra and Tave also contributed significantly, to a lesser extent. Meanwhile, humidity-related variables (RHmin, RHmax, and RHave) along with Tmin and Pr, were found to have minimal impact on the model’s predictions.
Related to the 8-variable models, which require the same input variables except RHmax and RHmin, all ML models showed an excellent performance in predicting ETo with 0.994 ≤ R2 ≤ 0.999, 1.685 ≤ RMSEn ≤ 3.546,1.191 ≤ Se ≤ 2.389, and 0.060 ≤ MAE ≤ 0.120 mm·day−1 during the testing period (Table 4).
The ANN algorithm, model ANN8, outperformed other models with R2 = 0.999, RMSEn = 1.685%, MAE = 0.045 mm·day−1, and the lowest simulation error (Se = 1.191%), making it the best choice to predict ETo over Coolidge station using the eight weather parameters.
Considering the 7-variable models (Tmax, Tmin, Tave, RHave, Rs, Pr, U2), all ML models showed a slight tendency to overestimate ETo by 0.105–0.143 mm·day−1. The lowest simulation error during the testing period was recorded for the CatBoost algorithm, model CB 7 (Se = 2.099%), making it the best choice to predict ETo using the seven weather parameters over the Coolidge station.
The absence of Rs from the 6-variable models (Tmax, Tmin, Tave, RHave, Pr, U2) led to increasing Se up to 8.107% (RF6 model); however, all ML models performed well during the testing period with 0.948 ≤ R2 ≤ 0.954, 10.125 ≤ RMSEn ≤ 10.811, and a slight overestimation of ETo by 0.381 to 0.406 mm·day−1. The ANN 6 model outperformed all 6-variable models with a Se of 7.615%.
Excluding U2 from the 5-variable models (Tmax, Tmin, Tave, RHave, Pr) resulted in a slight overestimation of ETo, ranging from 0.563 (ANN6 model) to 0.601 mm·day−1 (RF6 model). The ANN 5-variable model outperformed other ML models with R2 = 0.903, RMSEn = 14.718%, and a Se of 11.249%.
Even after excluding Pr and RHave from the 4-variable and 3-variable models, respectively, all tested ML algorithms performed well in predicting ETo over Coolidge station (0.832 ≤ R2 ≤ 0.902 and 14.748 ≤ RMSEn ≤ 19.327). All ML models overestimated ETo by 0.567 (ANN 4 model) to 0.727 mm·day−1 (RF3 model) and Se between 11.325% (ANN 4 model) and 14.525% (RF3 model), making ANN 4 and ANN3 models the best-performing models when using the 4-variable and 3-variable models, respectively, to predict ETo over Coolidge station.

3.2. Maricopa Station

The performance evaluation results of the 35 developed ETo models across the Maricopa weather station are summarized in Table 5. Similarly to the Coolidge weather station, considering 10-variable parameters, all ML algorithms showed a linear-positive distribution between ETobs and ETpre with 0.995 ≤ R2 ≤ 0.999 during the testing period (Figure 7 and Table 5). All 10-variable models showed a slight tendency to overestimate ETo by 0.043–0.114 mm·day−1. RMSEn values ranged between 1.224 and 3.521%, while the Se values were 0.833 ≤ Se ≤ 2.219.
Our analysis revealed no notable differences in the structure and distribution of violin diagrams of the observed and predicted datasets, indicating a highly accurate prediction of ETo during the testing period (Figure 8 and Table 5).
Likewise, the CatBoost algorithm, CatBoost10 model, outperformed other models with R2 = 0.999, RMSEn = 1.224%, MAE = 0.043 mm·day−1, and Se = 0.833% when considering the ten weather parameters, making it the best choice to predict ETo over Maricopa.
Like Coolidge station, Tmax, U2, and Rs had the most significant positive impact on ETo predictions. Additionally, Ra and Tave showed a moderate positive effect on predicting ETo. However, RHmin, RHmax, and RHave diminishing ETo, leading to lower SHAP values. RHave and Pr had the lowest impact on ETo, suggesting limited direct relevance to the model’s predictions (Figure 9).
Among the ten weather parameters, Tmax was identified as the most influential variable, followed by U2 and Rs. Additionally, both Ra and Tave contributed notably to the model’s predictions; however, RHmin, RHmax, Tmin, RHave, and Pr had minimal impacts (Figure 10).
Considering the 8, 7, 6, 5, 4, and 3-variable models, all ML algorithms showed excellent (RMSEn = 1.587%) to fair (RMSEn = 20.737%) accuracy in predicting daily ETo over the Maricopa station. The ANN algorithm outperformed all ML models with Se values ranging between 1.063 and 14.747%. The RMSEn values were between 1.587% and 19.457%, reflecting an excellent-to-good performance of ANN in predicting ETo over the Maricopa station.

3.3. Queen Creek Station

Evaluation indices of the 35 developed ETo models across the Queen Creek weather station are summarized in Table 6. Like the Coolidge and Maricopa weather stations, all ML algorithms showed a linear-positive distribution between ETobs and ETpre (0.994 ≤ R2 ≤ 0.999) with a slight overestimation of daily ETo ranging between 0.043 and 0.110 mm·day−1 when considering 10-variable parameters (Figure 11 and Table 6).
No notable differences in the structure and distribution of violin diagrams of the observed and predicted datasets (Figure 12). RMSEn values ranged between 1.224 and 3.521%, while the Se values were 0.877 ≤ Se ≤ 2.235, indicating excellent accuracy in the prediction of ETo during the testing period (Table 6).
The CatBoost algorithm, CB10 model, outperformed other models with a slight overestimation of predicted ETo (MAE = 0.043 mm·day−1) when considering the ten weather parameters. R2, Se, and RMSEn were 0.999, 0.877%, and 1.450%, indicating an excellent performance of the CB10 model in predicting ETo Queen Creek.
Like Coolidge and Maricopa, Tmax, Rs, and U2 significantly impacted ETo predictions while Ra and Tave had moderate effects. However, RHmin, RHmax, and RHave have lessened ETo, with RHave and Pr showing minimal impact, indicating limited relevance to the model (Figure 13).
Tmax stood out as the most influential variable, with U2 and Rs also playing significant roles among the ten weather parameters. Meanwhile, Ra and Tave made notable contributions to the model’s predictions, while RHmin, RHmax, Tmin, RHave, and Pr had only minor impacts (Figure 14).
All ML algorithms showed excellent (RMSEn = 1.748%) to good (RMSEn = 18.603%) accuracy in predicting daily ETo over the Queen Creek when considering the 8, 7, 6, 5, 4, and 3-variable models. The ANN algorithm outperformed all ML models with R2 values between 0.862 and 0.999 and Se ≤ 13.489%. The RMSEn ranged between 1.748% and 17.655%, reflecting an excellent-to-good performance of the ANN-based models in predicting ETo over the Queen Creek station.

3.4. Overall Discussion

In this study, 35 ETo models were developed using five ML algorithms and 7 input combinations of meteorological variables, as listed in Table 3. Statistical indices assessing the performance of 35 ETo-developed models in comparison with the FAO-56 PM standard model across Coolidge, Maricopa, and Queen Creek are summarized in Table 4, Table 5 and Table 6, respectively. Generally speaking, the 35 ETo-developed models showed an excellent to fair performance in predicting daily ETo across the three tested weather stations in Pinal County, Arizona. Employing the ANN10, RF10, XGBoost10, CatBoost10, and SVM10 ETo models using all the meteorological variables resulted in the highest accuracies in predicting daily ETo during training and testing periods at the three tested stations. However, excluding meteorological variables led to a gradual reduction in the performance of the ET-developed models across the three tested weather stations, whereas the highest Se were recorded under 3-variable models. The ET models executing complete meteorological variables have the best prediction accuracy, as compared with the incomplete input parameters [50,51]. These findings are consistent with previously cited studies, which reported that using more meteorological variables as inputs can generally boost the accuracy of ML models in estimating daily ETo [35,36,52,53].
The CatBoost algorithm, CatBoost10 model, slightly outperformed other 10-variable models with R2 = 0.999, 1.227 ≤ RMSEn ≤ 1.479, 0.043 ≤ MAE ≤ 0.073 mm·day−1, and 0.833 ≤ Se ≤ 0.890% during the testing period (Table 4, Table 5 and Table 6), making it the best choice to predict ETo over the three weather stations. Our findings agreed with Huang et al. [54], who reported a higher performance of CatBoost than SVM and RF-based models when all input combinations are available. Also, Zhang et al. [55] concluded that the CatBoost model has a great ability in modeling daily ETo in the arid and semi-arid regions of Northern China and recommended CatBoost for ETo estimations with similar climates.
The ANN-based models were slightly superior to the SVM-based models across the three weather stations (Table 4, Table 5 and Table 6). These were inconsistent with Kişi and Cimen [56] and Wen et al. [57], who reported a better performance of SVM than ANN in modeling daily ETo at North Coast Valleys, California, in the USA, and under extreme arid regions in Ejina City, China. This inconsistency might be a result of factors involved with the performance of the SVM and ANN models, like the training algorithm used and the models’ hyperparameter settings.
Meanwhile, the SVM models were more robust and capable than the RF models in predicting ETo across Pinal weather stations. This was in agreement with Rai et al. [58], who revealed that SVM outperforms RF for estimating monthly ETo in Uttar Pradesh and Uttarakhand States, India. Similar observations were reported by Abdallah et al. [35], Chen et al. [59], Fan et al. [60], and Huang et al. [54], who evaluated different deep and machine learning models, including SVM and RF, in predicting daily ETo under different climate conditions in China and Sudan.
Similarly, XGBoost models outperformed RF models in predicting ETo across the three weather stations. This agreed with Abdallah et al. [35], who reported that XGBoost models have a higher accuracy in estimating daily ETo than RF-based models under the hyper-arid regions in Sudan. Additionally, Fan et al. [60] concluded that the XGBoost models are more efficient and capable than the RF models in predicting daily ETo over China. Similar findings were reported in Brazil [26,61] and Bengaluru, Karnataka State, in India [62].
Based on our analyses, the RF8, RF7, RF6, RF5, RF4, and RF3 models showed the highest accuracy in predicting ETo during the training period (see R2, RMSEn, MAE, Se values, Table 4, Table 5 and Table 6). In contrast, compared with other ML models, the RF models had the highest simulation errors across the three weather stations during the testing period (Table 4, Table 5 and Table 6). Our observations were in agreement with Abdallah et al. [35], Fan et al. [60], Feng et al. [63] and Huang et al. [54], who reported that the RF algorithm could show a higher accuracy in predicting daily ETo than other ML models during the training period. Meanwhile, they had the highest MAE and root mean square errors during the testing period [35].
Taking the CatBoost10 models as an example, the SHAP-based interpretability across the three weather stations aligns well with established physical understanding of ETo drivers (Figure 5, Figure 9 and Figure 13). Among all variables, Tmax, Rs, and U2 had the most significant positive influence on ETo predictions, as high values of these parameters tend to enhance ETo. Ra and Tave also showed moderate positive effects. This was consistent with Allen et al. [20], who reported that during hot, dry weather conditions, ET demand increases due to the aridity of the atmosphere and the abundance of energy from direct Rs and latent heat. Under such conditions, the air can retain a considerable amount of water vapor, while wind may facilitate water transportation, thereby enabling greater water vapor uptake. Conversely, under humid conditions, the wind can only replace saturated air with slightly less saturated air and remove heat energy, leading to a far lesser extent than under arid conditions.
In contrast, the variables related to relative humidity (RHmin, RHmax, and RHave) showed an adverse effect on ETo, resulting in reduced SHAP values (Figure 5, Figure 9 and Figure 13). RH is inversely proportional to temperature. If the humidity remains constant, increasing temperature should increase ET. Increased humidity can partially offset the impact of increasing temperature on ET. Tmin and Pr had the least influence, suggesting limited direct relevance to the model’s prediction of ETo under the arid climate conditions.
Based on mean absolute SHAP values, Tmax emerged as the most influential factor, followed by Rs and U2 across the three tested weather stations (Figure 6, Figure 10 and Figure 14). Ra and Tave also contributed significantly, to a lesser extent. These variables are well-known drivers of evapotranspiration due to their direct physical relationship with atmospheric energy and vapor transport [20]. Meanwhile, humidity-related variables (RHmin, RHmax, and RHave) along with Tmin and Pr were found to have minimal impact on the model’s predictions (Figure 6, Figure 10 and Figure 14). This ranking provides a clear interpretation of the relative importance of meteorological parameters, reinforcing the physical understanding that Tmax, Rs, U2, followed by Ra, and Tave are the dominant factors influencing ETo under arid and semi-arid conditions.
The clear hierarchy of feature influence revealed by the SHAP values (Figure 6, Figure 10 and Figure 14) aligns well with established agro-meteorological principles. It provides a transparent and quantitative basis for prioritizing critical input variables in model development or field sensor deployment, especially in settings where data collection may be constrained. Overall, this bar plot enhances model interpretability and supports the conclusion that radiative and aerodynamic factors are the key drivers of ETo within the studied region.

4. Conclusions, Recommendations, and Outlooks

The current study developed 35 ETo models to predict daily ETo across Pinal County, Arizona. Seven input combinations of daily meteorological variables were used as a running scenario for training and testing the five ML models: Artificial Neural Network (ANN), Random Forest (RF), Extreme Gradient Boosting (XGBoost), Categorical Boosting (CatBoost), and Support Vector Machine (SVM). Then, SHapley Additive exPlanations (SHAP) was used to interpret each meteorological variable’s contribution to the model predictions. Results highlighted that the ML models could accurately predict daily ETo under the arid climate in Pinal County, Arizona. Based on the statistical indices, the models ranked with all 10 meteorological variables as follows: CatBoost, ANN, SVM, XGBoost, and RF. However, for incomplete input sets, the ranking was: ANN, SVM, CatBoost, XGBoost, and RF. Employing the ANN10, RF10, XGBoost10, CatBoost10, and SVM10 ETo models using all the meteorological variables resulted in the highest accuracies in predicting daily ETo during training and testing periods at the three tested stations. While excluding meteorological variables decreased model performance, models using only Tmax, Tmin, and Tave still predicted ETo well across the three tested weather stations. Therefore, the three-variable temperature-based models are highly recommended as a simplified technique for predicting daily ETo, especially in areas with limited climatic data, such as developing countries. These models can assist in water resource management and irrigation scheduling when meteorological data is limited. Additionally, our findings highlighted that Tmax, Rs, and U2 are the most influential factors affecting ETo in arid conditions, followed by Ra and Tave. In contrast, humidity-related variables (RHmin, RHmax, and RHave), along with Tmin and Pr, had minimal impact on the model’s predictions.
Selecting only three weather stations with daily meteorological data is insufficient to capture the spatial-temporal variability across Arizona and the USA. Thus, to confirm their accuracy, further studies are recommended to evaluate the performance of the 35 ETo-developed models under different spatial and temporal (hourly, monthly, and yearly) scales in the USA and even worldwide. Moreover, the ET-developed ML models rely on historical datasets, which may not reflect future climate change. Therefore, future studies integrating ET-developed models with global climate models would help improve water budget analysis and support growers and policymakers in the face of water scarcity.

Author Contributions

Conceptualization, E.A.E.; methodology, E.A.E., M.A.H.A. and D.E.M.E.; software, E.A.E. and M.A.H.A.; validation, E.A.E., M.A.H.A. and D.E.M.E.; investigation, E.A.E., M.A.H.A. and D.E.M.E.; data curation, E.A.E., M.A.H.A. and D.E.M.E.; writing—original draft preparation, E.A.E. and M.A.H.A.; writing—review and editing, E.A.E., M.A.H.A., C.W., K.R.T. and D.E.M.E.; visualization, E.A.E.; supervision, D.E.M.E.; project administration, D.E.M.E.; funding acquisition, K.R.T., C.W. and D.E.M.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the University of Arizona Cooperative Extension Service, The University of Arizona, Tucson, AZ 85721, USA. Additionally, it was supported by the Agricultural Research Service of the U.S. Department of Agriculture and carried out in collaboration with the Arid Land Agricultural Research Center.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author(s).

Acknowledgments

The authors greatly appreciate the University of Arizona Cooperative Extension Service for supporting irrigation research in arid regions in the USA, such as Arizona.

Conflicts of Interest

The authors declare that they have no known competing financial interest or personal relationships that could have appeared to influence the work reported in this study.

References

  1. Elshikha, D.E.; Attalah, S.; Waller, P.; Hunsaker, D.J.; Thorp, K.R.; Bautista, E.; Williams, C.; Wall, G.W.; Orr, E.; Elsadek, E.A. Can OpenET Transform Irrigation Management in the Southwestern, U.S.? College of Agriculture, Life, and Environmental Sciences, University of Arizona: Tucson, AZ, USA, 2025. [Google Scholar]
  2. Bennett, K.E.; Talsma, C.; Boero, R. Concurrent Changes in Extreme Hydroclimate Events in the Colorado River Basin. Water 2021, 13, 978. [Google Scholar] [CrossRef]
  3. Holdren, G.C.; Turner, K. Characteristics of Lake Mead, Arizona–Nevada. Lake Reserv. Manag. 2010, 26, 230–239. [Google Scholar] [CrossRef]
  4. Castle, S.L.; Reager, J.T.; Thomas, B.F.; Purdy, A.J.; Lo, M.-H.; Famiglietti, J.S.; Tang, Q. Remote Detection of Water Management Impacts on Evapotranspiration in the Colorado River Basin. Geophys. Res. Lett. 2016, 43, 5089–5097. [Google Scholar] [CrossRef]
  5. Elsadek, E.A.; Attalah, S.; Waller, P.; Norton, R.; Hunsaker, D.J.; Williams, C.; Thorp, K.R.; Orr, E.; Elshikha, D.E.M. Simulating Water Use and Yield for Full and Deficit Flood-Irrigated Cotton in Arizona, USA. Agronomy 2025, 15, 2023. [Google Scholar] [CrossRef]
  6. Thorp, K.R.; Calleja, S.; Pauli, D.; Thompson, A.L.; Elshikha, D.E. Agronomic Outcomes of Precision Irrigation Management Technologies with Varying Complexity. J. ASABE 2022, 65, 135–150. [Google Scholar] [CrossRef]
  7. Elshikha, D.E.; Attalah, S.; Elsadek, E.A.; Waller, P.; Thorp, K.; Sanyal, D.; Bautista, E.; Norton, R.; Hunsaker, D.; Williams, C.; et al. The Impact of Gravity Drip and Flood Irrigation on Development, Water Productivity, and Fiber Yield of Cotton in Semi-Arid Conditions of Arizona. In Proceedings of the 2024 ASABE Annual International Meeting, Anaheim, CA, USA, 28–31 July 2024; American Society of Agricultural and Biological Engineers: St. Joseph, MI, USA, 2024; pp. 1–16. [Google Scholar]
  8. Elshikha, D.E.M.; Wang, G.; Waller, P.M.; Hunsaker, D.J.; Dierig, D.; Thorp, K.R.; Thompson, A.; Katterman, M.E.; Herritt, M.T.; Bautista, E.; et al. Guayule Growth and Yield Responses to Deficit Irrigation Strategies in the U.S. Desert. Agric. Water Manag. 2023, 277, 108093. [Google Scholar] [CrossRef]
  9. Elsadek, E.A.; Zhang, K.; Hamoud, Y.A.; Mousa, A.; Awad, A.; Abdallah, M.; Shaghaleh, H.; Hamad, A.A.A.; Jamil, M.T.; Elbeltagi, A. Impacts of Climate Change on Rice Yields in the Nile River Delta of Egypt: A Large-Scale Projection Analysis Based on CMIP6. Agric. Water Manag. 2024, 292, 108673. [Google Scholar] [CrossRef]
  10. Elsadek, E.A. Study on the In-Field Water Balance and the Projected Impacts of Climate Change on Rice Yields in the Nile River Delta. Ph.D. Thesis, Hohai University, Nanjing, China, 2023. [Google Scholar]
  11. Elsadek, E.A.; Elshikha, D.E.M.; Awad, A.; Hamoud, Y.A.; Elsheikha, A.M.; Williams, C.; Orr, E.; Shaghaleh, H.; Hamad, A.A.A.; Thorp, K.R.; et al. Projecting Rice Water Footprint for Different Shared Socioeconomic Pathways under Arid Climate Conditions. Irrig. Sci. 2025, 43, 955–969. [Google Scholar] [CrossRef]
  12. Elsadek, E.; Zhang, K.; Mousa, A.; Ezaz, G.T.; Tola, T.L.; Shaghaleh, H.; Hamad, A.A.A.; Alhaj Hamoud, Y. Study on the In-Field Water Balance of Direct-Seeded Rice with Various Irrigation Regimes under Arid Climatic Conditions in Egypt Using the AquaCrop Model. Agronomy 2023, 13, 609. [Google Scholar] [CrossRef]
  13. Adeyemi, O.; Grove, I.; Peets, S.; Norton, T. Advanced Monitoring and Management Systems for Improving Sustainability in Precision Irrigation. Sustainability 2017, 9, 353. [Google Scholar] [CrossRef]
  14. Abioye, E.A.; Abidin, M.S.Z.; Mahmud, M.S.A.; Buyamin, S.; Ishak, M.H.I.; Rahman, M.K.I.A.; Otuoze, A.O.; Onotu, P.; Ramli, M.S.A. A Review on Monitoring and Advanced Control Strategies for Precision Irrigation. Comput. Electron. Agric. 2020, 173, 105441. [Google Scholar] [CrossRef]
  15. Volk, J.M.; Huntington, J.L.; Melton, F.S.; Allen, R.; Anderson, M.; Fisher, J.B.; Kilic, A.; Ruhoff, A.; Senay, G.B.; Minor, B.; et al. Assessing the Accuracy of OpenET Satellite-Based Evapotranspiration Data to Support Water Resource and Land Management Applications. Nat. Water 2024, 2, 193–205. [Google Scholar] [CrossRef]
  16. Melton, F.S.; Huntington, J.; Grimm, R.; Herring, J.; Hall, M.; Rollison, D.; Erickson, T.; Allen, R.; Anderson, M.; Fisher, J.B.; et al. OpenET: Filling a Critical Data Gap in Water Management for the Western United States. JAWRA J. Am. Water Resour. Assoc. 2022, 58, 971–994. [Google Scholar] [CrossRef]
  17. Wanniarachchi, S.; Sarukkalige, R. A Review on Evapotranspiration Estimation in Agricultural Water Management: Past, Present, and Future. Hydrology 2022, 9, 123. [Google Scholar] [CrossRef]
  18. French, A.N.; Sanchez, C.A.; Hunsaker, D.J.; Anderson, R.G.; Saber, M.N.; Wisniewski, E.H. Lettuce Evapotranspiration and Crop Coefficients Using Eddy Covariance and Remote Sensing Observations. Irrig. Sci. 2024, 42, 1245–1272. [Google Scholar] [CrossRef]
  19. Bawazir, A.S.; Luthy, R.; King, J.P.; Tanzy, B.F.; Solis, J. Assessment of the Crop Coefficient for Saltgrass under Native Riparian Field Conditions in the Desert Southwest. Hydrol. Process. 2014, 28, 6163–6171. [Google Scholar] [CrossRef]
  20. Allen, R.G.; Pereira, L.S.; Raes, D.; Smith, M. Crop Evapotranspiration-Guidelines for Computing Crop Water Requirements-FAO Irrigation and Drainage Paper 56; FAO: Rome, Italy, 1998. [Google Scholar]
  21. Shiri, J.; Marti, P.; Karimi, S.; Landeras, G. Data Splitting Strategies for Improving Data Driven Models for Reference Evapotranspiration Estimation among Similar Stations. Comput. Electron. Agric. 2019, 162, 70–81. [Google Scholar] [CrossRef]
  22. Hargreaves, G.H.; Samani, Z.A. Reference Crop Evapotranspiration from Temperature. Appl. Eng. Agric. 1985, 1, 96–99. [Google Scholar] [CrossRef]
  23. Elbeltagi, A.; Katipoğlu, O.M.; Kartal, V.; Danandeh Mehr, A.; Berhail, S.; Elsadek, E.A. Advanced Reference Crop Evapotranspiration Prediction: A Novel Framework Combining Neural Nets, Bee Optimization Algorithm, and Mode Decomposition. Appl. Water Sci. 2024, 14, 256. [Google Scholar] [CrossRef]
  24. Raja, P.; Sona, F.; Surendran, U.; Srinivas, C.V.; Kannan, K.; Madhu, M.; Mahesh, P.; Annepu, S.K.; Ahmed, M.; Chandrasekar, K.; et al. Performance Evaluation of Different Empirical Models for Reference Evapotranspiration Estimation over Udhagamandalm, The Nilgiris, India. Sci. Rep. 2024, 14, 12429. [Google Scholar] [CrossRef] [PubMed]
  25. Celestin, S.; Qi, F.; Li, R.; Yu, T.; Cheng, W. Evaluation of 32 Simple Equations against the Penman–Monteith Method to Estimate the Reference Evapotranspiration in the Hexi Corridor, Northwest China. Water 2020, 12, 2772. [Google Scholar] [CrossRef]
  26. Ferreira, L.B.; da Cunha, F.F. New Approach to Estimate Daily Reference Evapotranspiration Based on Hourly Temperature and Relative Humidity Using Machine Learning and Deep Learning. Agric. Water Manag. 2020, 234, 106113. [Google Scholar] [CrossRef]
  27. Sowmya, M.R.; Kumar, M.B.S.; Ambat, S.K. Comparison of Deep Neural Networks for Reference Evapotranspiration Prediction Using Minimal Meteorological Data. In Proceedings of the 2020 Advanced Computing and Communication Technologies for High Performance Applications (ACCTHPA), Cochin, India, 2–4 July 2020; pp. 27–33. [Google Scholar]
  28. Bellido-Jiménez, J.A.; Estévez, J.; García-Marín, A.P. New Machine Learning Approaches to Improve Reference Evapotranspiration Estimates Using Intra-Daily Temperature-Based Variables in a Semi-Arid Region of Spain. Agric. Water Manag. 2021, 245, 106558. [Google Scholar] [CrossRef]
  29. Kaya, Y.Z.; Zelenakova, M.; Üneş, F.; Demirci, M.; Hlavata, H.; Mesaros, P. Estimation of Daily Evapotranspiration in Košice City (Slovakia) Using Several Soft Computing Techniques. Theor. Appl. Climatol. 2021, 144, 287–298. [Google Scholar] [CrossRef]
  30. Chia, M.Y.; Huang, Y.F.; Koo, C.H.; Fung, K.F. Recent Advances in Evapotranspiration Estimation Using Artificial Intelligence Approaches with a Focus on Hybridization Techniques—A Review. Agronomy 2020, 10, 101. [Google Scholar] [CrossRef]
  31. Duval, D.; Bickel, A.K.; Frisvold, G. County Agricultural Economy Profiles for Southern Arizona. Available online: https://mapazdashboard.arizona.edu/article/county-agricultural-economy-profiles-southern-arizona (accessed on 27 June 2025).
  32. Migoya, C. With Colorado River Water Cuts, Some Pinal Farmers Drill Wells. For Others, Fields Sit Dry. Available online: https://www.azcentral.com/story/news/local/arizona-environment/2023/02/08/cap-water-cuts-increase-disparities-among-pinal-county-farmers/69827876007/ (accessed on 26 August 2025).
  33. Duval, D.; Montanía, C.V.; Quintero, J.H. Arizona County Agricultural Economy Profiles; College of Agriculture, Life, and Environmental Sciences, University of Arizona: Tucson, AZ, USA, 2022. [Google Scholar]
  34. Nabhan, G.P.; Richter, B.D.; Riordan, E.C.; Tornbom, C. Toward Water-Resilient Agriculture in Arizona: Future Scenarios Addressing Water Scarcity; Lincoln Institute of Land Policy: Cambridge, MA, USA, 2023. [Google Scholar]
  35. Abdallah, M.; Mohammadi, B.; Zaroug, M.A.H.; Omer, A.; Cheraghalizadeh, M.; Eldow, M.E.E.; Duan, Z. Reference Evapotranspiration Estimation in Hyper-Arid Regions via D-Vine Copula Based-Quantile Regression and Comparison with Empirical Approaches and Machine Learning Models. J. Hydrol. Reg. Stud. 2022, 44, 101259. [Google Scholar] [CrossRef]
  36. Mehdizadeh, S.; Mohammadi, B.; Pham, Q.B.; Duan, Z. Development of Boosted Machine Learning Models for Estimating Daily Reference Evapotranspiration and Comparison with Empirical Approaches. Water 2021, 13, 3489. [Google Scholar] [CrossRef]
  37. Bisoyi, N.; Gupta, H.; Padhy, N.P.; Chakrapani, G.J. Prediction of Daily Sediment Discharge Using a Back Propagation Neural Network Training Algorithm: A Case Study of the Narmada River, India. Int. J. Sediment Res. 2019, 34, 125–135. [Google Scholar] [CrossRef]
  38. Kumar, M.; Raghuwanshi, N.S.; Singh, R. Artificial Neural Networks Approach in Evapotranspiration Modeling: A Review. Irrig. Sci. 2011, 29, 11–25. [Google Scholar] [CrossRef]
  39. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  40. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 785–794. [Google Scholar]
  41. Dorogush, A.V.; Ershov, V.; Gulin, A. CatBoost: Gradient Boosting with Categorical Features Support. arXiv 2018, arXiv:1810.11363. [Google Scholar] [CrossRef]
  42. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased Boosting with Categorical Features. In Proceedings of the Advances in Neural Information Processing Systems, 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
  43. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  44. Wang, L. Support Vector Machines: Theory and Applications; Springer Science & Business Media: Boston, NY, USA, 2005; Volume 177, ISBN 3540243887. [Google Scholar]
  45. Muraina, I. Ideal Dataset Splitting Ratios in Machine Learning Algorithms: General Concerns for Data Scientists and Data Analysts. In Proceedings of the 7th International Mardin Artuklu Scientific Research Conference, Mardin, Turkey, 10–12 December 2021; pp. 496–504. [Google Scholar]
  46. Štrumbelj, E.; Kononenko, I. Explaining Prediction Models and Individual Predictions with Feature Contributions. Knowl. Inf. Syst. 2014, 41, 647–665. [Google Scholar] [CrossRef]
  47. Antwarg, L.; Miller, R.M.; Shapira, B.; Rokach, L. Explaining Anomalies Detected by Autoencoders Using Shapley Additive Explanations. Expert Syst. Appl. 2021, 186, 115736. [Google Scholar] [CrossRef]
  48. Jamieson, P.D.; Porter, J.R.; Wilson, D.R. A Test of the Computer Simulation Model ARCWHEAT1 on Wheat Crops Grown in New Zealand. Field Crops Res. 1991, 27, 337–350. [Google Scholar] [CrossRef]
  49. Brisson, N.; Ruget, F.; Gate, P.; Lorgeou, J.; Nicoullaud, B.; Tayot, X.; Plenet, D.; Jeuffroy, M.-H.; Bouthier, A.; Ripoche, D.; et al. STICS: A Generic Model for Simulating Crops and Their Water and Nitrogen Balances. II. Model Validation for Wheat and Maize. Agronomie 2002, 22, 69–92. [Google Scholar] [CrossRef]
  50. Mattar, M.A. Using Gene Expression Programming in Monthly Reference Evapotranspiration Modeling: A Case Study in Egypt. Agric. Water Manag. 2018, 198, 28–38. [Google Scholar] [CrossRef]
  51. Mehdizadeh, S.; Behmanesh, J.; Khalili, K. Using MARS, SVM, GEP and Empirical Equations for Estimation of Monthly Mean Reference Evapotranspiration. Comput. Electron. Agric. 2017, 139, 103–114. [Google Scholar] [CrossRef]
  52. Droogers, P.; Allen, R.G. Estimating Reference Evapotranspiration Under Inaccurate Data Conditions. Irrig. Drain. Syst. 2002, 16, 33–45. [Google Scholar] [CrossRef]
  53. Wang, S.; Lian, J.; Peng, Y.; Hu, B.; Chen, H. Generalized Reference Evapotranspiration Models with Limited Climatic Data Based on Random Forest and Gene Expression Programming in Guangxi, China. Agric. Water Manag. 2019, 221, 220–230. [Google Scholar] [CrossRef]
  54. Huang, G.; Wu, L.; Ma, X.; Zhang, W.; Fan, J.; Yu, X.; Zeng, W.; Zhou, H. Evaluation of CatBoost Method for Prediction of Reference Evapotranspiration in Humid Regions. J. Hydrol. 2019, 574, 1029–1041. [Google Scholar] [CrossRef]
  55. Zhang, Y.; Zhao, Z.; Zheng, J. CatBoost: A New Approach for Estimating Daily Reference Crop Evapotranspiration in Arid and Semi-Arid Regions of Northern China. J. Hydrol. 2020, 588, 125087. [Google Scholar] [CrossRef]
  56. Kişi, O.; Cimen, M. Evapotranspiration Modelling Using Support Vector Machines/Modélisation de l’évapotranspiration à l’aide de ‘Support Vector Machines’. Hydrol. Sci. J. 2009, 54, 918–928. [Google Scholar] [CrossRef]
  57. Wen, X.; Si, J.; He, Z.; Wu, J.; Shao, H.; Yu, H. Support-Vector-Machine-Based Models for Modeling Daily Reference Evapotranspiration With Limited Climatic Data in Extreme Arid Regions. Water Resour. Manag. 2015, 29, 3195–3209. [Google Scholar] [CrossRef]
  58. Rai, P.; Kumar, P.; Al-Ansari, N.; Malik, A. Evaluation of Machine Learning versus Empirical Models for Monthly Reference Evapotranspiration Estimation in Uttar Pradesh and Uttarakhand States, India. Sustainability 2022, 14, 5771. [Google Scholar] [CrossRef]
  59. Chen, Z.; Zhu, Z.; Jiang, H.; Sun, S. Estimating Daily Reference Evapotranspiration Based on Limited Meteorological Data Using Deep Learning and Classical Machine Learning Methods. J. Hydrol. 2020, 591, 125286. [Google Scholar] [CrossRef]
  60. Fan, J.; Yue, W.; Wu, L.; Zhang, F.; Cai, H.; Wang, X.; Lu, X.; Xiang, Y. Evaluation of SVM, ELM and Four Tree-Based Ensemble Models for Predicting Daily Reference Evapotranspiration Using Limited Meteorological Data in Different Climates of China. Agric. For. Meteorol. 2018, 263, 225–241. [Google Scholar] [CrossRef]
  61. Ferreira, L.B.; da Cunha, F.F.; Fernandes Filho, E.I. Exploring Machine Learning and Multi-Task Learning to Estimate Meteorological Data and Reference Evapotranspiration across Brazil. Agric. Water Manag. 2022, 259, 107281. [Google Scholar] [CrossRef]
  62. Agrawal, Y.; Kumar, M.; Ananthakrishnan, S.; Kumarapuram, G. Evapotranspiration Modeling Using Different Tree Based Ensembled Machine Learning Algorithm. Water Resour. Manag. 2022, 36, 1025–1042. [Google Scholar] [CrossRef]
  63. Feng, Y.; Cui, N.; Gong, D.; Zhang, Q.; Zhao, L. Evaluation of Random Forests and Generalized Regression Neural Networks for Daily Reference Evapotranspiration Modelling. Agric. Water Manag. 2017, 193, 163–173. [Google Scholar] [CrossRef]
Figure 1. Arizona linear hydrography features and geographic distribution of the Pinal weather stations, Arizona, USA.
Figure 1. Arizona linear hydrography features and geographic distribution of the Pinal weather stations, Arizona, USA.
Agriculture 15 01985 g001
Figure 2. Kendall correlation analysis between the variables used in the Pinal weather stations.
Figure 2. Kendall correlation analysis between the variables used in the Pinal weather stations.
Agriculture 15 01985 g002
Figure 3. Scatter plots of predicted reference evapotranspiration (ETo, mm·day−1) at the Coolidge weather station while considering all meteorological variables (10-variable models).
Figure 3. Scatter plots of predicted reference evapotranspiration (ETo, mm·day−1) at the Coolidge weather station while considering all meteorological variables (10-variable models).
Agriculture 15 01985 g003
Figure 4. Violin diagrams for actual reference evapotranspiration (ETo, mm·day−1) versus predicted ETo at the Coolidge weather station while considering all meteorological variables (10-variable models).
Figure 4. Violin diagrams for actual reference evapotranspiration (ETo, mm·day−1) versus predicted ETo at the Coolidge weather station while considering all meteorological variables (10-variable models).
Agriculture 15 01985 g004
Figure 5. SHapley Additive exPlanations for the Coolidge weather station.
Figure 5. SHapley Additive exPlanations for the Coolidge weather station.
Agriculture 15 01985 g005
Figure 6. Mean absolute SHapley Additive exPlanations (SHAP) values for each input feature used to predict daily reference evapotranspiration (ETo, mm·day−1) across the Coolidge weather station.
Figure 6. Mean absolute SHapley Additive exPlanations (SHAP) values for each input feature used to predict daily reference evapotranspiration (ETo, mm·day−1) across the Coolidge weather station.
Agriculture 15 01985 g006
Figure 7. Scatter plots of predicted reference evapotranspiration (ETo, mm·day−1) at the Maricopa weather station while considering all meteorological variables (10-variable models).
Figure 7. Scatter plots of predicted reference evapotranspiration (ETo, mm·day−1) at the Maricopa weather station while considering all meteorological variables (10-variable models).
Agriculture 15 01985 g007
Figure 8. Violin diagrams for actual reference evapotranspiration (ETo, mm·day−1) versus predicted ETo at the Maricopa weather station while considering all meteorological variables (10-variable models).
Figure 8. Violin diagrams for actual reference evapotranspiration (ETo, mm·day−1) versus predicted ETo at the Maricopa weather station while considering all meteorological variables (10-variable models).
Agriculture 15 01985 g008
Figure 9. SHapley Additive exPlanations for the Maricopa weather station.
Figure 9. SHapley Additive exPlanations for the Maricopa weather station.
Agriculture 15 01985 g009
Figure 10. Mean absolute SHapley Additive exPlanations (SHAP) values for each input feature used to predict daily reference evapotranspiration (ETo, mm·day−1) across the Maricopa weather station.
Figure 10. Mean absolute SHapley Additive exPlanations (SHAP) values for each input feature used to predict daily reference evapotranspiration (ETo, mm·day−1) across the Maricopa weather station.
Agriculture 15 01985 g010
Figure 11. Scatter plots of predicted reference evapotranspiration (ETo, mm·day−1) at the Queen Creek weather station while considering all meteorological variables (10-variable models).
Figure 11. Scatter plots of predicted reference evapotranspiration (ETo, mm·day−1) at the Queen Creek weather station while considering all meteorological variables (10-variable models).
Agriculture 15 01985 g011
Figure 12. Violin diagrams for actual reference evapotranspiration (ETo, mm·day−1) versus predicted ETo at the Queen Creek weather station while considering all meteorological variables (10-variable models).
Figure 12. Violin diagrams for actual reference evapotranspiration (ETo, mm·day−1) versus predicted ETo at the Queen Creek weather station while considering all meteorological variables (10-variable models).
Agriculture 15 01985 g012
Figure 13. SHapley Additive exPlanations for the Queen Creek weather station.
Figure 13. SHapley Additive exPlanations for the Queen Creek weather station.
Agriculture 15 01985 g013
Figure 14. Mean absolute SHapley Additive exPlanations (SHAP) values for each input feature used to predict daily reference evapotranspiration (ETo, mm·day−1) across the Queen Creek weather station.
Figure 14. Mean absolute SHapley Additive exPlanations (SHAP) values for each input feature used to predict daily reference evapotranspiration (ETo, mm·day−1) across the Queen Creek weather station.
Agriculture 15 01985 g014
Table 1. Statistics of the daily meteorological variables (1995–2025) used to predict reference evapotranspiration (ETo, mm·day−1) across Pinal County, Arizona.
Table 1. Statistics of the daily meteorological variables (1995–2025) used to predict reference evapotranspiration (ETo, mm·day−1) across Pinal County, Arizona.
ParameterStationStatistics
MeanStandard
Deviation
MinimumMaximumStandard
Error
Range
Tmax, °CCoolidge30.28.55.848.70.142.9
Maricopa30.48.85.848.00.142.2
Queen Creek29.88.65.747.90.142.2
Tmin, °CCoolidge10.88.4−1031.10.141.1
Maricopa12.78.9−8.732.50.141.2
Queen Creek12.18.3−17.332.20.149.5
Tave, °CCoolidge20.48.3−1.037.60.138.6
Maricopa21.58.90.139.20.139.1
Queen Creek21.18.4−0.638.40.139.0
Pr, mmCoolidge0.42.30.051.00.0151.0
Maricopa0.42.50.056.90.0156.9
Queen Creek0.52.90.073.40.0173.4
RHmax, %Coolidge77.316.117.0100.30.283.3
Maricopa69.119.519.21000.280.8
Queen Creek71.917.416.9100.20.283.3
RHmin, %Coolidge18.612.31.497.00.195.6
Maricopa18.212.10.191.00.190.9
Queen Creek18.812.11.893.00.191.2
RHave, %Coolidge45.117.67.4100.00.292.6
Maricopa40.518.48.598.00.289.5
Queen Creek42.117.76.1100.00.293.9
U2, m·s−1Coolidge1.80.80.08.00.018.0
Maricopa1.80.80.27.60.017.5
Queen Creek1.70.70.06.60.016.6
Ra,
MJ·m−2·day−1
Coolidge30.68.50.041.50.141.5
Maricopa30.68.417.841.50.123.7
Queen Creek30.68.50.041.50.141.5
Rs,
MJ·m−2·day−1
Coolidge20.67.11.234.30.133.1
Maricopa20.77.11.233.80.132.6
Queen Creek20.27.00.332.60.132.3
ETo, mm·day−1Coolidge5.02.40.312.50.0112.2
Maricopa5.22.60.412.80.0112.3
Queen Creek5.02.30.312.60.0112.2
Notes: Tmax, Tmin, and Tave refer to maximum, minimum, and average temperature, respectively. Pr refers to precipitation. RHmax, RHmin, and RHave are the maximum, minimum, and average relative humidity, respectively, while U2 is the wind speed at 2 m height. Ra is extraterrestrial solar radiation, and RS is solar radiation. ETo is the reference evapotranspiration.
Table 2. List of the default and tuned hyperparameters for each machine learning model used in ETo prediction.
Table 2. List of the default and tuned hyperparameters for each machine learning model used in ETo prediction.
ModelHyperparametersDefaultTuned
ANNhidden_layer_sizes100,100, 50
activation‘relu’‘relu’
solver‘adam’‘adam’
alpha (L2 penalty)0.00010.0001
learning_rate‘constant’‘constant’
learning_rate_init0.0010.001
max_iter2001000
batch_size‘auto’‘auto’
random_stateNone42
RFn_estimators100100
max_depthNoneNone
min_samples_split22
min_samples_leaf11
max_features‘sqrt’‘sqrt’
bootstrapTrueTrue
random_stateNone42
XGBoostbooster‘gbtree’‘gbtree’
learning_rate0.30.1
max_depth66
n_estimators100100
subsample11
colsample_bytree11
gamma00
reg_alpha00
reg_lambda11
random_stateNone42
CatBoostiterations10001000
learning_rateAuto tunedAuto tuned
depth66
l2_leaf_reg33
loss_functionRMSERMSE
bootstrap_type‘Bayesian’‘Bayesian’
random_strength11
bagging_temperature11
random_stateNone42
SVMkernel‘rbf’‘rbf’
C1.01.0
epsilon0.10.1
gamma‘scale’‘scale’
degree33
coef00.00.0
random_stateNone42
Notes: ANN, RF, XGBoost, CatBoost, and SVM refer to Artificial Neural Network (ANN), Random Forest, Extreme Gradient Boosting, Categorical Boosting, and Support Vector Machine, respectively.
Table 3. Input combination of the five machine learning models used for predicting daily reference evapotranspiration (ETo, mm·day−1) across Pinal County, Arizona.
Table 3. Input combination of the five machine learning models used for predicting daily reference evapotranspiration (ETo, mm·day−1) across Pinal County, Arizona.
Model (No. of Inputs)Input Combination
ANN10RF10XGBoost10CatBoost10SVM10Tmax, Tmin, Tave, Pr, RHmax, RHmin, RHave, U2, Ra, Rs
ANN8RF8XGBoost8CatBoost8SVM8Tmax, Tmin, Tave, Pr, RHave, U2, Ra, Rs
ANN7RF7XGBoost7CatBoost7SVM7Tmax, Tmin, Tave, Pr, RHave, U2, Rs
ANN6RF6XGBoost6CatBoost6SVM6Tmax, Tmin, Tave, Pr, RHave, U2
ANN5RF5XGBoost5CatBoost5SVM5Tmax, Tmin, Tave, Pr, RHave
ANN4RF4XGBoost4CatBoost4SVM4Tmax, Tmin, Tave, RHave
ANN3RF3XGBoost3CatBoost3SVM3Tmax, Tmin, Tave
Notes: ANN, RF, XGBoost, CatBoost, and SVM refer to Artificial Neural Network (ANN), Random Forest, Extreme Gradient Boosting, Categorical Boosting, and Support Vector Machine, respectively. Tmax, Tmin, and Tave refer to maximum, minimum, and average temperature, respectively. Pr refers to precipitation. RHmax, RHmin, and RHave are the maximum, minimum, and average relative humidity, respectively, while U2 is the wind speed at 2 m height. Ra is extraterrestrial solar radiation, and RS is solar radiation.
Table 4. Statistical indices assessing the performance of five machine learning models in predicting daily reference evapotranspiration (ETo, mm·day−1), in comparison with the FAO-56 Penman Monteith (PM) standardized method at Coolidge, Pinal County, Arizona.
Table 4. Statistical indices assessing the performance of five machine learning models in predicting daily reference evapotranspiration (ETo, mm·day−1), in comparison with the FAO-56 Penman Monteith (PM) standardized method at Coolidge, Pinal County, Arizona.
ModelTrainingTesting
R2MAE, mmRMSEn,
%
Se,
%
R2MAE,
mm
RMSEn,
%
Se,
%
Tmax, Tmin, Tave, Pr, RHmax, RHmin, RHave, U2, Ra, Rs
ANN100.9990.0481.3400.9410.9990.0491.4030.983
SVM100.9990.0521.7501.0320.9980.0562.2381.123
RF100.9990.0451.3590.8950.9940.1153.4852.295
XGBoost100.9990.0631.6341.2550.9970.1002.8421.999
CatBoost101.0000.0310.7880.6090.9990.0451.3050.890
Tmax, Tmin, Tave, Pr, RHave, U2, Ra, Rs
ANN80.9990.0551.5271.0890.9990.0601.6851.191
SVM80.9980.0561.9561.0980.9970.0622.4741.229
RF80.9990.0461.3780.9120.9940.1203.5462.389
XGBoost80.9990.0691.7901.3570.9960.1103.0972.189
CatBoost80.9990.0441.1520.8690.9990.0621.8361.243
Tmax, Tmin, Tave, Pr, RHave, U2, Rs
ANN70.9960.1032.9052.0470.9960.1073.1102.142
SVM70.9960.1023.1152.0250.9940.1093.5752.173
RF70.9990.0541.5561.0660.9920.1434.1462.867
XGBoost70.9970.0932.4941.8480.9930.1393.9102.775
CatBoost70.9980.0822.2521.6250.9960.1053.0592.099
Tmax, Tmin, Tave, Pr, RHave, U2
ANN60.9570.3739.7267.3810.9540.38110.1257.615
SVM60.9540.37310.0637.3790.9510.38210.4917.646
RF60.9920.1544.0833.0420.9480.40610.8118.107
XGBoost60.9700.3068.0366.0480.9490.40010.6767.991
CatBoost60.9690.3148.2426.2220.9530.38310.2587.647
Tmax, Tmin, Tave, Pr, RHave
ANN50.9010.56414.70111.1530.9030.56314.71811.249
SVM50.8960.57315.07111.3430.8990.56914.97311.371
RF50.9840.2275.9574.4950.8890.60115.75312.025
XGBoost50.9290.48112.4519.5150.8920.59115.48411.812
CatBoost50.9270.48912.5839.6780.8970.57915.14511.573
Tmax, Tmin, Tave, RHave
ANN40.9010.56614.68511.1970.9020.56714.74811.325
SVM40.8960.57515.07711.3670.8990.56914.99111.367
RF40.9840.2295.9804.5290.8870.60815.88912.149
XGBoost40.9290.48212.4279.5310.8920.59315.51811.848
CatBoost40.9280.48912.5549.6790.8960.58015.18011.586
Tmax, Tmin, Tave
ANN30.8500.68618.10513.5740.8500.68318.28413.649
SVM30.8460.69418.30213.7250.8490.68718.35513.731
RF30.9770.2687.1165.2920.8320.72719.32714.525
XGBoost30.8920.58915.35911.6580.8390.71418.90114.280
CatBoost30.8860.60515.73711.9710.8460.70118.52414.011
Notes: ANN, RF, XGBoost, CatBoost, and SVM refer to Artificial Neural Network (ANN), Random Forest, Extreme Gradient Boosting, Categorical Boosting, and Support Vector Machine, respectively. The R2 close to 1.0 indicates a good match between the predicted and observed datasets. The simulation is excellent if the estimation result of RMSEn  < 10% and good if the result ranges between 10 and 20%. A result from 20 to 30% indicates fair performance, while RMSEn > 30% indicates poor performance. The lower the MAE value, the greater the projection/modeling accuracy. The Se between ±15% is acceptable.
Table 5. Statistical indices assessing the performance of five machine learning models in predicting daily reference evapotranspiration (ETo), in comparison with the FAO-56 PM standard model at Maricopa, Pinal County, Arizona.
Table 5. Statistical indices assessing the performance of five machine learning models in predicting daily reference evapotranspiration (ETo), in comparison with the FAO-56 PM standard model at Maricopa, Pinal County, Arizona.
ModelTrainingTesting
R2MAE,
mm
RMSEn,
%
Se,
%
R2MAE,
mm
RMSEn,
%
Se,
%
Tmax, Tmin, Tave, Pr, RHmax, RHmin, RHave, U2, Ra, Rs
ANN101.0000.0401.0630.7620.9990.0441.2240.848
SVM100.9990.0511.8160.9770.9980.0572.2501.101
RF100.9990.0451.3460.8590.9950.1143.5212.219
XGBoost100.9990.0591.5061.1370.9970.1012.8791.955
CatBoost101.0000.0300.7290.5680.9990.0431.2270.833
Tmax, Tmin, Tave, Pr, RHave, U2, Ra, Rs
ANN80.9990.0511.3930.9870.9990.0551.5871.063
SVM80.9990.0551.9561.0540.9980.0622.5301.196
RF80.9990.0461.3560.8760.9950.1183.6072.294
XGBoost80.9990.0631.6251.2200.9970.1042.9882.015
CatBoost81.0000.0411.0550.7920.9990.0601.7301.155
Tmax, Tmin, Tave, Pr, RHave, U2, Rs
ANN70.9970.1022.7231.9640.9970.1032.8321.995
SVM70.9960.1033.0511.9770.9950.1063.5102.061
RF70.9990.0551.5591.0600.9940.1394.0142.704
XGBoost70.9980.0922.4061.7630.9950.1313.6012.534
CatBoost70.9980.0812.1391.5520.9970.1042.8902.014
Tmax, Tmin, Tave, Pr, RHave, U2
ANN60.9660.3649.2477.0060.9620.3799.7677.361
SVM60.9610.3769.8627.2390.9590.38210.1317.416
RF60.9940.1544.0062.9600.9570.40410.4577.848
XGBoost60.9740.3087.9775.9260.9590.39410.1947.660
CatBoost60.9740.3148.0596.0460.9620.3819.8537.401
Tmax, Tmin, Tave, Pr, RHave
ANN50.9040.60915.43811.7190.9040.61015.59711.848
SVM50.9000.61915.77811.8990.8990.61915.95712.025
RF50.9850.2436.2044.6770.8960.63316.23212.294
XGBoost50.9310.51713.0639.9480.8990.62315.98212.095
CatBoost50.9290.52613.22410.1190.9010.61915.84112.025
Tmax, Tmin, Tave, RHave
ANN40.9030.61315.50611.7970.9040.61115.56811.858
SVM40.8990.61915.79811.9130.9000.61715.89811.976
RF40.9840.2446.2194.6890.8940.63716.35412.368
XGBoost40.9320.51813.0149.9540.8970.62916.12112.207
CatBoost40.9290.53113.30010.2130.9000.62215.92112.089
Tmax, Tmin, Tave
ANN30.8510.75719.21714.5570.8500.75919.45714.747
SVM30.8450.77219.59214.8420.8450.76919.78114.939
RF30.9750.3027.8265.8010.8300.79520.73715.450
XGBoost30.8880.65916.69512.6800.8410.77720.01815.089
CatBoost30.8830.67317.03312.9520.8460.76719.71814.904
Notes: ANN, RF, XGBoost, CatBoost, and SVM refer to Artificial Neural Network (ANN), Random Forest, Extreme Gradient Boosting, Categorical Boosting, and Support Vector Machine, respectively. The R2 close to 1.0 indicates a good match between the predicted and observed datasets. The simulation is excellent if the estimation result of RMSEn  < 10% and good if the result ranges between 10 and 20%. A result from 20 to 30% indicates fair performance, while RMSEn > 30% indicates poor performance. The lower the MAE value, the greater the projection/modeling accuracy. The Se between ±15% is acceptable.
Table 6. Statistical indices assessing the performance of five machine learning models in predicting daily reference evapotranspiration (ETo), in comparison with the FAO-56 PM standard model at Queen Creek, Pinal County, Arizona.
Table 6. Statistical indices assessing the performance of five machine learning models in predicting daily reference evapotranspiration (ETo), in comparison with the FAO-56 PM standard model at Queen Creek, Pinal County, Arizona.
ModelTrainingTesting
R2MAE,
mm
RMSEn,
%
Se,
%
R2MAE,
mm
RMSEn,
%
Se,
%
Tmax, Tmin, Tave, Pr, RHmax, RHmin, RHave, U2, Ra, Rs
ANN100.9990.0621.2400.8330.9990.0721.4500.932
SVM100.9990.0861.7331.0040.9970.1342.7051.138
RF100.9990.0631.2770.8270.9940.1813.6622.235
XGBoost100.9990.0711.4321.0870.9960.1472.9851.906
CatBoost101.0000.0380.7590.5880.9990.0731.4790.877
Tmax, Tmin, Tave, Pr, RHave, U2, Ra, Rs
ANN80.9990.0741.4881.0070.9990.0861.7481.092
SVM80.9980.0921.8561.0490.9960.1422.8791.186
RF80.9990.0651.3160.8540.9940.1863.7562.339
XGBoost80.9990.0801.6041.2100.9960.1523.0692.016
CatBoost80.9990.0551.1000.8260.9990.0921.8631.165
Tmax, Tmin, Tave, Pr, RHave, U2, Rs
ANN70.9960.1412.8342.0130.9960.1573.1822.218
SVM70.9960.1563.1482.0430.9930.2034.1072.300
RF70.9990.0761.5201.0410.9910.2254.5643.042
XGBoost70.9970.1222.4501.7790.9930.2024.0922.775
CatBoost70.9980.1142.2961.6550.9950.1633.2992.244
Tmax, Tmin, Tave, Pr, RHave, U2
ANN60.9490.52610.5918.2000.9460.54411.0108.469
SVM60.9420.56011.2848.3050.9400.57311.5968.455
RF60.9910.2204.4213.3430.9370.59011.9549.041
XGBoost60.9630.4478.9966.8830.9390.58111.7608.919
CatBoost60.9620.4569.1757.0580.9430.56211.3718.651
Tmax, Tmin, Tave, Pr, RHave
ANN50.8970.74615.01911.5800.8930.76515.49911.854
SVM50.8940.75815.26011.6010.8920.76915.58011.781
RF50.9840.2985.9984.5350.8830.80116.21912.284
XGBoost50.9270.62712.6299.6720.8900.77915.77411.950
CatBoost50.9260.63212.7299.8160.8920.77115.60611.803
Tmax, Tmin, Tave, RHave
ANN40.8990.56414.87311.3660.8940.58015.46111.750
SVM40.8930.57815.30511.6430.8920.58415.63211.835
RF40.9830.2276.0394.5700.8810.61216.33912.398
XGBoost40.9270.48412.6599.7430.8870.59815.93712.103
CatBoost40.9260.48912.7429.8420.8900.59015.75211.950
Tmax, Tmin, Tave
ANN30.8650.85317.18713.1420.8620.87217.65513.489
SVM30.8620.86517.41813.1870.8600.87717.75913.469
RF30.9780.3487.0185.2690.8460.91918.60314.184
XGBoost30.9020.72914.68911.2170.8540.89618.14313.814
CatBoost30.8960.74915.08711.5340.8570.88717.95713.681
Notes: ANN, RF, XGBoost, CatBoost, and SVM refer to Artificial Neural Network (ANN), Random Forest, Extreme Gradient Boosting, Categorical Boosting, and Support Vector Machine, respectively. The R2 close to 1.0 indicates a good match between the predicted and observed datasets. The simulation is excellent if the estimation result of RMSEn  < 10% and good if the result ranges between 10 and 20%. A result from 20 to 30% indicates fair performance, while RMSEn > 30% indicates poor performance. The lower the MAE value, the greater the projection/modeling accuracy. The Se between ±15% is acceptable.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Elsadek, E.A.; Ali, M.A.H.; Williams, C.; Thorp, K.R.; Elshikha, D.E.M. A Novel Framework for Predicting Daily Reference Evapotranspiration Using Interpretable Machine Learning Techniques. Agriculture 2025, 15, 1985. https://doi.org/10.3390/agriculture15181985

AMA Style

Elsadek EA, Ali MAH, Williams C, Thorp KR, Elshikha DEM. A Novel Framework for Predicting Daily Reference Evapotranspiration Using Interpretable Machine Learning Techniques. Agriculture. 2025; 15(18):1985. https://doi.org/10.3390/agriculture15181985

Chicago/Turabian Style

Elsadek, Elsayed Ahmed, Mosaad Ali Hussein Ali, Clinton Williams, Kelly R. Thorp, and Diaa Eldin M. Elshikha. 2025. "A Novel Framework for Predicting Daily Reference Evapotranspiration Using Interpretable Machine Learning Techniques" Agriculture 15, no. 18: 1985. https://doi.org/10.3390/agriculture15181985

APA Style

Elsadek, E. A., Ali, M. A. H., Williams, C., Thorp, K. R., & Elshikha, D. E. M. (2025). A Novel Framework for Predicting Daily Reference Evapotranspiration Using Interpretable Machine Learning Techniques. Agriculture, 15(18), 1985. https://doi.org/10.3390/agriculture15181985

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop