Next Article in Journal
Evaluation of Water Social Service and Comprehensive Water Management Linked with Integrated River Evaluation
Next Article in Special Issue
Dynamics of Volumetric Moisture in Sand Caused by Injection Irrigation—Physical Model
Previous Article in Journal
Ontogenetic Structure and Temporal Patterns of Summer Ichthyoplankton in Upper Course of the Xijiang River, SW China
Previous Article in Special Issue
Infiltration Characteristics and Spatiotemporal Distribution of Soil Moisture in Layered Soil under Vertical Tube Irrigation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Saturated Hydraulic Conductivity Estimation Using Artificial Neural Networks

by
Josué Trejo-Alonso
1,
Carlos Fuentes
2,*,
Carlos Chávez
1,*,
Antonio Quevedo
2,
Alfonso Gutierrez-Lopez
3 and
Brandon González-Correa
4
1
Water Research Center, Department of Irrigation and Drainage Engineering, Autonomous University of Queretaro, Cerro de las Campanas SN, Col. Las Campanas, Queretaro 76010, Mexico
2
Mexican Institute of Water Technology, Paseo Cuauhnáhuac Núm. 8532, Jiutepec 62550, Mexico
3
Water Research Center, Centro de Investigaciones del Agua-Queretaro (CIAQ), International Flood Initiative, Latin-American and the Caribbean Region (IFI-LAC), International Hydrological Programme (IHP-UNESCO), Universidad Autonoma de Queretaro, Queretaro 76010, Mexico
4
Engineering Faculty, Autonomous University of Queretaro, Cerro de las Campanas SN, Col. Las Campanas, Queretaro 70610, Mexico
*
Authors to whom correspondence should be addressed.
Water 2021, 13(5), 705; https://doi.org/10.3390/w13050705
Submission received: 10 February 2021 / Revised: 26 February 2021 / Accepted: 27 February 2021 / Published: 5 March 2021
(This article belongs to the Special Issue Study of the Soil Water Movement in Irrigated Agriculture)

Abstract

:
In the present work, we construct several artificial neural networks (varying the input data) to calculate the saturated hydraulic conductivity (K S ) using a database with 900 measured samples obtained from the Irrigation District 023, in San Juan del Rio, Queretaro, Mexico. All of them were constructed using two hidden layers, a back-propagation algorithm for the learning process, and a logistic function as a nonlinear transfer function. In order to explore different arrays for neurons into hidden layers, we performed the bootstrap technique for each neural network and selected the one with the least Root Mean Square Error (RMSE) value. We also compared these results with pedotransfer functions and another neural networks from the literature. The results show that our artificial neural networks obtained from 0.0459 to 0.0413 in the RMSE measurement, and 0.9725 to 0.9780 for R 2 , which are in good agreement with other works. We also found that reducing the amount of the input data offered us better results.

1. Introduction

Soil water movement is important in several fields, like irrigation, drainage, hydrology, and agriculture [1]. Among all the measurable quantities in soils, one of the most important is the saturated hydraulic conductivity (K S ), defined as the ability to transmit water throughout the saturated zone [2,3], which is highly correlated with the optimization of the flow rate applied to the border or furrow in the gravity irrigation [2,3,4,5,6]. Although this property is measured easily in a laboratory, or in the field, it needs to be applied at a small scale, and most of the time, it is required to be used on a large scale [5,6]. This is inconvenient due to the fact that all these tests and measurements are time-consuming, impractical, and not cost-effective [7,8].
In order to solve the inconveniences mentioned above, a great number of studies about pedotransfer functions (PTFs) were published [7,8,9,10]. These mathematical models allow us to estimate the K S from some soil characteristics, such as texture, field capacity, the permanent wilting point, bulk density, porosity, and organic matter, among others [7,8,9,10]. The robustness of the model is linked to the number of physical parameters used to calculate the saturated hydraulic conductivity; the more parameters, the more accurate the prediction. However, as it was mentioned before, depending on the measurements, the PTFs are difficult to get, due to economic resources and the time it takes to measure all the variables, which presents as a limitation for this kind of function and the predictive capacity. Besides, some works have been questioned because the soil in which they want to apply are different from the soil used for their development, such as [11].
In recent years, another alternative has been explored—Artificial Neural Networks (ANNs), which have become a common tool used as a special class of PTFs, for example, [12,13]. ANNs are an artificial intelligence that simulate the behavior of the human brain, and their structures consist of a number of interconnected elements called neurons which are logically arranged in layers, known as input, output, and hidden (see, e.g., [12] and references therein). Each neuron connects to all the neurons in the next layer via weighted connections. In Figure 1, we show a schematic structure for an ANN.
Until now, there has been no analytical way to obtain the ideal network structure (number of hidden layers and neurons inside them) as a function of the complexity of the problem. The structure must be selected by performing a trial-and-error process. ANNs with one or two hidden layers and an adequate number of hidden neurons are found to be sufficient for most problems (e.g., [12,14]). There are also several works studying the ideal number of neurons in the hidden layers [15,16], but these methods present general guidelines for the selection in the number of neurons only.
The ANNs’ name comes not only from of their structure, but because they “learn”. The most common algorithm used in ANNs for the learning process is back-propagation (e.g., [17,18]). Each neuron belonging to a layer receives weighted inputs from the neurons in the previous layer and processes them to transmit its output to the neurons in the next layer, and this is done through links. Each link is assigned a weight that is no more than a numerical estimate of the strength of the connection. This weighted sum of inputs in a neuron are converted into the numerical estimate that we see, according to the nonlinear transfer function (the most commonly used is the sigmoid function). The ANNs then modify the weights of neurons in response to errors between the actual output values and the target output values, using what is known as gradient descent [19,20]. This is then applied on the sum of the squares of the errors for all the training patterns, until the mean error of the sum squared of all the training patterns is minimal or within the tolerance specified for the problem.
In this work, we use ANNs to obtain the K S in Irrigation District 023 placed in Queretaro Mexico using a sample of 900 plots. The sample, ANN configurations, and validation tests are described in Section 2. Finally, in Section 3, we show the results, and a comparison between the several configurations obtained in this work and comparisons of our results with PTFs and other ANNs in the literature.

2. Materials and Methods

2.1. Study Area

The Irrigation District 023 is located in Queretaro, Mexico, at 20 18 to 20 34 N, 99 56 to 100 12 W with an altitude of 1892 M.A.S.L, and it has an area of 11 048 ha. It includes the municipalities of San Juan del Rio and Pedro Escobedo (Figure 2). Its predominant climate is semiarid with summer rains, with an annual precipitation avarage of 599 mm and annual average temperature of 20 C [21]. The water is conducted through open channels. The main channels are lined with concrete, but all the lateral channels that carry water to the plots are unlined. The separation of the plots in some cases are by trees, unlined channels, drains, or roads [22].

2.2. Soil Database

An extensive and detailed database description can be found at [23]. In summary, the database used in this study was developed from samplings in 900 plots at Irrigation District 023, San Juan del Rio Queretaro. These samples were sent to the laboratory to obtain the following parameters: soil texture by the Bouyucos hydrometer, bulk density ( ρ a ) by the cylinder method of known volume, moisture content at saturation ( θ S ), field capacity (FC) and permanent wilting point (PWP) by the method of the pressure membrane pot, and the saturated hydraulic conductivity by the variable head permeameter method. From these measurements, we took seven and considered them through the paper as the input data: percentage of clay, sand and silt, bulk density, permanent wilting point, moisture content at saturation, and field capacity (Table 1).

2.3. The ANNs’ Setup

As mentioned in Section 1, there is no analytical way to define the ANN structure, but based on similar several works (e.g., [12,14,16]) we noticed that the ANNs have one or two hidden layers only. With this in mind, we tested several structures for our ANNs with two hidden layers and an additional one with three hidden layers as a test. Details about the ANNs’ settings are described below.
We began with all the input data (seven) and tested several configurations as follows: we start with two neurons in the first layer and two neurons in the second. Then we continue to vary the number of neurons in the first layer from two to ten, and leave the second layer constant. Next, the number of neurons in the second layer is increased to three and we vary the number of neurons in the first, again, from two to ten, and so on until the number of the second layer varies from two to ten, just like the first one. This input layer has the seven input data mentioned before. Finally, the output layer contains the K S predicted value.
Then, we changed the number of data in the input layer (decreasing it by one) and repeated the process as explained before. The choice of which parameter has to be removed is based on the importance plot (details in Section 2.4), except for the last configuration, where we remove θ S instead of the percentage of silt, because θ S is closely related with the clay percentage. We kept removing input data until we had three measurements only. For each configuration, we varied the number of neurons in the hidden layers from two to ten.
The ANN was programmed using the neuralnet package [24] and the caret package [25], both of them provided by the R software [26].
The neuralnet package trains neural networks using backpropagation, resilient backpropagation (RPROP) with [27] or without weight backtracking [28], or the modified globally convergent version (GRPROP) by [29]. The function allows flexible settings through custom-choice of error and activation function. The caret package (short for Classification And REgression Training) is a set of functions that attempts to streamline the process for creating predictive models. There are many different modeling functions in R. Some have different syntax for model training and/or prediction. The package started off as a way to provide a uniform interface for the functions themselves, as well as a way to standardize common tasks (such as parameter tuning and variable importance). Specifically, we used the train function of this package to perform the cross-validation process.
Furthermore, the calculation of generalized weights [30] is implemented. In this work, we use RPROP as an algorithm type to calculate the neural network, the sum of squared errors as a differentiable function for the calculation of the error and a logistic differentiable function for smoothing the result of the cross-product of the covariate or neurons and the weights.
Using the RMSE measurement, we selected the optimal ANN structure in each case, and finally, we kept this last ANN structure as ideal. The Mean Absolute Error (MAE) measurement was calculated as:
MAE = 1 n i = 1 n | y i x i | ,
where y i are the measurement values, x i are the predicted values, and n is the total measurement. We also calculate the RMSE as:
RMSE = 1 n i = 1 n ( y i x i ) 2 .

2.4. Cross-Validation

In order to validate our ANN results, we made a cross-validation analysis using the train function from the caret package. This function sets up a grid of tuning parameters for a number of classification and regression routines, fits each model, and calculates a resampling-based performance measure. In this case, we use the parameters optimized to neuralnet, and we generated 25 bootstrap replications for each ANN configuration. Finally, for the prediction of new samples, we used the predict function.
The train function returned several results: the ideal ANN configuration, the best RMSE, MAE and R 2 values for each tested configuration, a RMSE matrix, and an importance plot. This last plot indicates which parameter contributed the most to the K S final approximation. Based on this last plot, we decided which input data would be removed on the next run, with the exception already mentioned above.

3. Results and Discussion

As mentioned in Section 2, we generated several ANNs which contained all the combinations from two to ten neurons in each hidden layer. In Figure 3 we show the results for these tests. In the x axis we have the number of neurons in the first hidden layer, the y axis represents the RMSE value obtained from the 25 bootstrap replications, and each color represents the number of neurons in the second hidden layer. The election of the best configuration is based on the smallest RMSE value in these plots. Additionally, from the top to the bottom and from the left to the right, we show the variation in the input data. Another result is shown in Figure 4. This plot is a 10 × 10 matrix where the color represents the RMSE value for each configuration, the rows are the number of neurons in the second hidden layer, and the columns are the number of neurons in first hidden layer. Remember that each plot differs from the other in the number of input data in the same way that Figure 3. In general, we can see that a small number of neurons in the first layer presents higher values of RMSE. Another important result is presented in the Figure 5 where we show a density plot representation (varying the input data from top to bottom) for RMSE, MAE, and R 2 measurements derived from the bootstrap analysis. From these plots we have small variations for each measurement which indicate that the results are not highly dependent on the ANN configurations. Finally, in Figure 6, we show the importance plots, which are described in the cross-validation section presented previously. These plots help us to decide which parameter we must keep or eliminate in each run when we have different numbers of input data.
In order to explore another possibility and to get more confident results, we made a test increasing the number of hidden layers to three. We applied the same process explained previously for this new configuration, and we got an improvement of ∼9% in the RMSE values, but with 15 times more computation time. This was the reason for dismissal of these configurations.
In Table 2 we present the three best configurations of each run, where we have the number of input data, the ANN structure, and the RMSE, MAE, and R 2 measurements.
Following the results, we present Figure 7, where we compare the K S measurements with the K S predicted by the ANNs for all tested configurations. The dotted line is a 1:1 relation and the R 2 is obtained from the train function. The plotted values were obtained applying the best ANN configuration for each run (Figure 7). The histograms in the top-left corner show the residuals ( Δ K S ), which is the difference between the predicted and measured values.
Recall that to obtain the ideal neuron configuration, we apply the train package, which allows us to obtain the RMSE, MAE, and R 2 data for each arrangement of neurons in the hidden layers. The only variation was the input data (ranging from 7 to 3). Therefore, it was possible to obtain five different boxplots for the RMSE, whose only difference would be the aforementioned input data. This is shown in Figure 8, where a trend is observed for the RMSE to fall, while the number of input data decreases.
In Table 3 we show the results obtained from the literature with PTFs or ANNs and compare them with this work.
The results for RMSE obtained in this work are better in, at least, 35% compared with the ones presented by [31], and we reported the fifth-best value for R 2 .
Another advantage of ANN is that the initial shape for the function to get the relation between the variables or a principal component analysis is not needed. Additionally, as we can see in Figure 3 and Table 2, the results are independent of the ANN structure. This is supported by the fact that the RMSE values for different configurations are very similar (the largest difference is ∼10%), and for R 2 values the difference is ∼5%, and we got ∼11% for MAE. Besides, the Figure 5 presents an almost gaussian distribution for these three statistical measurements, which is in agreement with a non-biased result. In this Figure, we also see that the density distributions are narrower, while the number of input data is smaller. This tells us that in contrast with the PTF models, we found a tendency which indicates that the less input data we have, the more accurate our prediction of K S is, as well as the Figure 8 shown this tendency too. This result can be explained by the following reasons. Based on the Principal Component Analysis of [23], we noted that the principal variables contributing the most to the sample were K S , the percentage of clay, θ S , PWP, and FC, which is supported by the importance plots (Figure 6). Additionally, for the 900 analyzed samples contained in 10 of the 12 existing types of soil (according to the USDA Textural Soil Classification), they showed that the infiltration rate depended directly on the percentage of clay and the ρ a .

4. Conclusions

In this work, we developed five Artificial Neural Networks in order to calculate the saturated hydraulic conductivity based on the sample used by [23]. All networks consist of one input layer, two hidden layers, and one output layer. We tested a network with three hidden layers, but with little better results. We took 75% of the sample for training and 25% for validation. We also tested all the possible combinations for the number of neurons in each hidden layer, taking into account that the number of neurons for each hidden layer will vary from 2 to 10 neurons. Finally, we selected the best number of neurons in each layer based on RMSE measurements obtained from a cross-validation analysis.
The results show that, compared with other works, we get better or similar results for RMSE and R 2 measurements and similar configurations for our ANN. Finally, we can say that if the necessary resources are available to obtain a large number of data in the field, it is necessary to develop a study of PTF as well as ANN to compare the results of each process and be able to choose the best option between both of them. The latter will not only be based on the RMSE or R 2 measurements, but also on the desired application (a statistical ground property study or prediction for irrigation proposes). A more detailed study to define an exact range of the amount of data needed from a reliable artificial neural network study should be carried out, but the latter is beyond the objective of this work. Besides, we have to be more careful in the characteristics of the sample where the models come out. In our case, we analyzed 10 of the 12 types of soils where the bulk density and the percentage of clay became more important parameters compared to others. This made our models more reliable for almost any type of soil.
The coupling of the Saint Venant and Richards equations is the complete mathematical model for modeling gravity irrigation [41]. However, its use requires detailed information on the physical properties of the soil, as well as a series of field and laboratory experiments that can be expensive [42,43]. In this way, the results used in this article, combined with some rapid field and laboratory tests, can be an excellent alternative to reduce costs and the time used to obtain that information.
Finally, the application of artificial neural networks have been demonstrated to successfully solve classification and prediction problems, and this is probably for the nonlinear relation between the variables. The calculation of the saturated hydraulic conductivity in this work proves that we need only three variables to predict new values, but the soil properties are crucial for the correct application of these models in contrast with the ANN configuration, which has been proved to play a minor role in the final results.

Author Contributions

Conceptualization, C.C. and C.F.; methodology, C.C., J.T.-A. and A.Q.; software, J.T.-A. and B.G.-C.; validation, B.G.-C., J.T.-A. and A.G.-L.; formal analysis, C.C. and J.T.-A.; investigation, A.G.-L. and J.T.-A.; resources, C.C., C.F. and A.Q.; data curation, C.C. and A.Q.; writing—original draft preparation, J.T.-A.; writing—review and editing, C.C., C.F., A.Q. and A.G.-L.; visualization, J.T.-A. and B.G.-C.; supervision, C.C., C.F. and A.Q.; project administration, C.C.; funding acquisition, C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported as part of a collaboration between the National Water Commission (CONAGUA, according to its Spanish acronym); The Water Users Association of Second Unit of Module Two, Irrigation District No. 023; the Irrigation District 023, San Juan del Río, Querétaro; and the Autonomous University of Querétaro, under the program RIGRAT 2015–2019.

Conflicts of Interest

The authors declare no conflict of interest regarding the publication of this paper.

References

  1. Fuentes, S.; Trejo-Alonso, J.; Quevedo, A.; Fuentes, C.; Chávez, C. Modeling Soil Water Redistribution under Gravity Irrigation with the Richards Equation. Mathematics 2020, 8, 1581. [Google Scholar] [CrossRef]
  2. Chávez, C.; Fuentes, C. Design and evaluation of surface irrigation systems applying an analytical formula in the irrigation district 085, La Begoña, Mexico. Agric. Water Manag. 2019, 221, 279–285. [Google Scholar] [CrossRef]
  3. Di, W.; Xue, J.; Bo, X.; Meng, W.; Wu, Y.; Du, T. Simulation of irrigation uniformity and optimization of irrigation technical parameters based on the SIRMOD model under alternate furrow irrigation. Irrig. Drain 2017, 66, 478–491. [Google Scholar]
  4. Gillies, M.H.; Smith, R.J. SISCO: Surface irrigation simulation, calibration and optimization. Irrig. Sci. 2015, 33, 339–355. [Google Scholar] [CrossRef]
  5. Saucedo, H.; Zavala, M.; Fuentes, C. Complete hydrodynamic model for border irrigation. Water Technol. Sci. 2011, 2, 23–38. [Google Scholar]
  6. Weibo, N.; Ma, X.; Fei, L. Evaluation of infiltration models and variability of soil infiltration properties at multiple scales. Irrig. Drain 2017, 66, 589–599. [Google Scholar]
  7. Zhang, Y.; Schaap, M.G. Estimation of saturated hydraulic conductivity with pedotransfer functions: A review. J. Hydrol. 2019, 575, 1011–1030. [Google Scholar] [CrossRef]
  8. Abdelbaki, A.M. Evaluation of pedotransfer functions for predicting soil bulk density for U.S. soils. Ain Shams Eng. J. 2018, 9, 1611–1619. [Google Scholar] [CrossRef]
  9. Brakensiek, D.; Rawls, W.J.; Stephenson, G.R. Modifying SCS Hydrologic Soil Groups and Curve Numbers for Rangeland Soils; Paper No. PNR-84203; ASAE: St. Joseph, MN, USA, 1984. [Google Scholar]
  10. Rasoulzadeh, A. Estimating Hydraulic Conductivity Using Pedotransfer Functions. In Hydraulic Conductivity-Issues, Determination and Applications; Elango, L., Ed.; InTech: Rijeka, Croatia, 2011; pp. 145–164. [Google Scholar]
  11. Moreira, L.; Righetto, A.M.; Medeiros, V.M. Soil hydraulics properties estimation by using pedotransfer functions in a northeastern semiarid zone catchment, Brazil. International Environmental Modelling and Software Society, 2004, Osnabrueck. Complexity and Integrated Resources Management. In Transactions of the 2nd Biennial Meeting of the International Environmental Modelling and Software Society, iEMSs 2004; International Environmental Modelling and Software Society: Manno, Switzerland, 2004; Volume 2, pp. 990–995. [Google Scholar]
  12. Erzin, Y.; Gumaste, S.D.; Gupta, A.K.; Singh, D.N. Artificial neural network (ANN) models for determining hydraulic conductivity of compacted fine-grained soils. Can. Geotech. J. 2009, 46, 955–968. [Google Scholar] [CrossRef]
  13. Agyare, W.A.; Park, S.J.; Vlek, P.L.G. Artificial Neural Network Estimation of Saturated Hydraulic Conductivity. VZJAAB 2007, 6, 423–431. [Google Scholar] [CrossRef]
  14. Sonmez, H.; Gokceoglu, C.; Nefeslioglu, H.A.; Kayabasi, A. Estimation of rock modulus: For intact rocks with an artificial neural network and for rock masses with a new empirical equation. Int. J. Rock Mech. Min. Sci. Geomech. Abstr. 2006, 43, 224–235. [Google Scholar] [CrossRef]
  15. Grima, M.A.; Babuska, R. Fuzzy model for the prediction of unconfined compressive strength of rock samples. Int. J. Rock Mech. Min. Sci. Geomech. Abstr. 1999, 36, 339–349. [Google Scholar] [CrossRef]
  16. Haque, M.E.; Sudhakar, K.V. ANN back-propagation prediction model for fracture toughness in microalloy steel. Int. J. Fatigue 2002, 24, 1003–1010. [Google Scholar] [CrossRef]
  17. Singh, T.N.; Gupta, A.R.; Sain, R. A comparative analysis of cognitive systems for the prediction of drillability of rocks and wear factor. Geotech. Geol. Eng. 2006, 24, 299–312. [Google Scholar] [CrossRef]
  18. Erzin, Y.; Rao, B.H.; Singh, D.N. Artificial neural network models for predicting soil thermal resistivity. Int. J. Therm. Sci. 2008, 47, 1347–1358. [Google Scholar] [CrossRef]
  19. Rumelhart, D.H.; Hinton, G.E.; Williams, R.J. Learning internal representation by error propagation. In Parallel Distributed Processing; Rumelhart, D.E., McClelland, J.L., Eds.; MIT Press: Cambridge, MA, USA, 1986; Volume 1, Chapter 8. [Google Scholar]
  20. Goh, A.T.C. Back-propagation neural networks for modelling complex systems. Artif. Intel. Eng. 1995, 9, 143–151. [Google Scholar] [CrossRef]
  21. Comisión Nacional del Agua (CONAGUA). Estadísticas del Agua en México; CONAGUA: Coyoacán, México, 2018; p. 306. [Google Scholar]
  22. Chávez, C.; Limón-Jiménez, I.; Espinoza-Alcántara, B.; López-Hernández, J.A.; Bárcenas-Ferruzca, E.; Trejo-Alonso, J. Water-Use Efficiency and Productivity Improvements in Surface Irrigation Systems. Agronomy 2020, 10, 1759. [Google Scholar] [CrossRef]
  23. Trejo-Alonso, J.; Quevedo, A.; Fuentes, C.; Chávez, C. Evaluation and Development of Pedotransfer Functions for Predicting Saturated Hydraulic Conductivity for Mexican Soils. Agronomy 2020, 10, 1516. [Google Scholar] [CrossRef]
  24. Fritsch, S.; Guenther, F; Wright, M.N. Neuralnet: Training of Neural Networks. R Package Version 1.44.2. Available online: https://CRAN.R-project.org/package=neuralnet (accessed on 19 October 2020).
  25. Kuhn, M. Caret: Classification and Regression Training. R Package Version 6.0-86. 2020. Available online: https://CRAN.R-project.org/package=caret (accessed on 22 January 2021).
  26. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2020; Available online: https://www.R-project.org/ (accessed on 12 January 2021).
  27. Riedmiller, M. Rprop—Description and Implementation Details; Technical Report; University of Karlsruhe: Karlsruhe, Germany, 1994. [Google Scholar]
  28. Riedmiller, M.; Braun, H. A direct adaptive method for faster backpropagation learning: The RPROP algorithm. In Proceedings of the IEEE International Conference on Neural Networks (ICNN), San Francisco, CA, USA, 28 March–1 April 1993; pp. 586–591. [Google Scholar]
  29. Anastasiadis, A.; Magoulas, G.D.; Vrahatis, M.N. New globally convergent training scheme based on the resilient propagation algorithm. Neurocomputing 2005, 64, 253–270. [Google Scholar] [CrossRef]
  30. Intrator, O.; Intrator, N. Using Neural Nets for Interpretation of Nonlinear Models. In Proceedings of the Statistical Computing Section; American Statistical Society: San Francisco, CA, USA, 1993; pp. 244–249. [Google Scholar]
  31. Tamari, S.; Wösten, J.H.M.; Ruiz-Suárez, J.C. Testing an Artificial Neural Network for Predicting Soil Hydraulic Conductivity. Soil Sci. Soc. Am. J. 1996, 60, 1732–1741. [Google Scholar] [CrossRef]
  32. Saxton, K.; Rawls, W.J.; Romberger, J.S.; Papendick, R.I. Estimating generalized soil water characteristics from texture. Soil Sci. Soc. Am. J. 1986, 5, 1031–1036. [Google Scholar] [CrossRef]
  33. Parasuraman, K.; Elshorbagy, A.; Si, B. Estimating Saturated Hydraulic Conductivity In Spatially Variable Fields Using Neural Network Ensembles. Soil Sci. Soc. Am. J. 2006, 70, 1851–1859. [Google Scholar] [CrossRef] [Green Version]
  34. Cosby, B.; Hornberger, G.; Clapp, R.; Ginn, T. A statistical exploration of the relationship of soil moisture characteristics to the physical properties of soils. Water Resour. Res. 1984, 20, 682–690. [Google Scholar] [CrossRef] [Green Version]
  35. Ahuja, L.R.; Naney, J.W.; Green, R.E.; Nielsen, D.R. Macroporosity to characterize spatial variability of hydraulic conductivity and effects of land management. Soil Sci. Soc. Am. J. 1984, 48, 699–702. [Google Scholar] [CrossRef]
  36. Schaap, M.G.; Leij, F.J. Using neural networks to predict soil water retention and soil hydraulic conductivity. Soil Tillage Res. 1998, 47, 37–42. [Google Scholar] [CrossRef]
  37. Vereecken, H.; Schnepf, A.; Hopmans, J.W.; Javaux, M.; Or, D.; Roose, T.; Vanderborght, J.; Young, M.H.; Amelung, W.; Aitkenhead, M.; et al. Modeling Soil Processes: Review, Key Challenges, and New Perspectives. Vadose Zone J. 2016, 15. [Google Scholar] [CrossRef] [Green Version]
  38. Minasny, B.; Hopmans, J.; Harter, T.; Eching, S.; Tuli, A.; Denton, M. Neural Networks Prediction of Soil Hydraulic Functions for Alluvial Soils Using Multistep Outflow Data. Soil Sci. Soc. Am. J. 2004, 68, 417–429. [Google Scholar] [CrossRef]
  39. Ferrer-Julià, M.; Estrela-Monreal, T.; Sánchez-del Corral-Jiménez, A.; García-Meléndez, E. Constructing a saturated hydraulic conductivity map of spain using pedotransfer functions and spatial prediction. Geoderma 2004, 123, 275–277. [Google Scholar] [CrossRef]
  40. Merdun, H.; Cinar, O.; Meral, R.; Apan, M. Comparison of artificial neural network and regression pedotransfer functions for prediction of soil water retention and saturated hydraulic conductivity. Soil Tillage Res. 2006, 90, 108–116. [Google Scholar] [CrossRef]
  41. Richards, L.A. Capillary conduction of liquids through porous mediums. Physics 1931, 1, 318–333. [Google Scholar] [CrossRef]
  42. Saucedo, H.; Fuentes, C.; Zavala, M. The Saint-Venant and Richards equation system in surface irrigation: (2) Numerical coupling for the advance phase in border irrigation. Ing. Hidraul. Mexico 2005, 20, 109–119. [Google Scholar]
  43. Fuentes, C.; Chávez, C. Analytic Representation of the Optimal Flow for Gravity Irrigation. Water 2020, 12, 2710. [Google Scholar] [CrossRef]
Figure 1. Schematic representation for an ANN structure. Each circle represents a neuron, where the I j means the neurons in the input layer, and the O j are the neurons in the output layer. All H j circles are the neurons in the hidden layers. The arrows represent the weighted connections.
Figure 1. Schematic representation for an ANN structure. Each circle represents a neuron, where the I j means the neurons in the input layer, and the O j are the neurons in the output layer. All H j circles are the neurons in the hidden layers. The arrows represent the weighted connections.
Water 13 00705 g001
Figure 2. Location map of the sampling points in Irrigation District 023 San Juan del Río, Querétaro.
Figure 2. Location map of the sampling points in Irrigation District 023 San Juan del Río, Querétaro.
Water 13 00705 g002
Figure 3. ANN test for choosing the ideal number of hidden neurons varying the input data. From top to bottom and left to right the number of input data goes from 7 to 3. The x axis represents the number of neurons in the first hidden layer and the y axis is the RMSE value. Each color is a different number of neurons in the second hidden layer.
Figure 3. ANN test for choosing the ideal number of hidden neurons varying the input data. From top to bottom and left to right the number of input data goes from 7 to 3. The x axis represents the number of neurons in the first hidden layer and the y axis is the RMSE value. Each color is a different number of neurons in the second hidden layer.
Water 13 00705 g003
Figure 4. RMSE matrix for different number of neurons in the hidden layers. From top to bottom and left to right, the number of input data goes from 7 to 3. The x axis represents the number of neurons in the first hidden layer and the y axis is the number of neurons in the second hidden layer.
Figure 4. RMSE matrix for different number of neurons in the hidden layers. From top to bottom and left to right, the number of input data goes from 7 to 3. The x axis represents the number of neurons in the first hidden layer and the y axis is the number of neurons in the second hidden layer.
Water 13 00705 g004
Figure 5. Density plots for (from left to right) RMSE, R 2 , and MAE resulting from the 25 bootstrap replications. From top to bottom, the number of input data goes from 7 to 3.
Figure 5. Density plots for (from left to right) RMSE, R 2 , and MAE resulting from the 25 bootstrap replications. From top to bottom, the number of input data goes from 7 to 3.
Water 13 00705 g005
Figure 6. Importance plots referring to the weight of each variable in the calculations. From top to bottom and from right to left, the number of input data goes from 7 to 3.
Figure 6. Importance plots referring to the weight of each variable in the calculations. From top to bottom and from right to left, the number of input data goes from 7 to 3.
Water 13 00705 g006
Figure 7. Plots for the comparison between K S measurement in the field with the K S value obtained applying the different ANN configurations. The dotted line is a 1:1 desirable relation. The histograms represent the residual distribution ( Δ K S ).
Figure 7. Plots for the comparison between K S measurement in the field with the K S value obtained applying the different ANN configurations. The dotted line is a 1:1 desirable relation. The histograms represent the residual distribution ( Δ K S ).
Water 13 00705 g007
Figure 8. Boxplot for each run of the train package. The names means two hidden layers and the number of input data used in each run.
Figure 8. Boxplot for each run of the train package. The names means two hidden layers and the number of input data used in each run.
Water 13 00705 g008
Table 1. Statistical properties of data measured in the laboratory.
Table 1. Statistical properties of data measured in the laboratory.
VariableMinMaxMedianMeanSDQ1Q3
Sand (%)0.0777.8328.3531.1420.2213.7552.00
Clay (%)2.1259.4621.7421.9512.0613.4430.00
Silt (%)0.8092.0045.2746.9123.4827.3059.79
ρ a (g/cm 3 )1.181.701.401.410.111.321.47
PWP (cm 3 /cm 3 )0.070.350.130.150.050.100.17
θ S (cm 3 /cm 3 )0.350.560.470.470.040.450.50
FC (cm 3 /cm 3 )0.170.470.290.300.060.250.32
K S   (cm/h)0.055.150.781.421.420.401.80
Table 2. The top three configurations for ANN structure and their statistical measurements. (1) # Input data (2) contains the ANN neurons’ structure (Input-Hidden1-Hidden2-Ouput), where each number represents the quantity of neurons used in each layer, (3) the RMSE measurements, (4) the MAE measurements, and (5) the R 2 measurements.
Table 2. The top three configurations for ANN structure and their statistical measurements. (1) # Input data (2) contains the ANN neurons’ structure (Input-Hidden1-Hidden2-Ouput), where each number represents the quantity of neurons used in each layer, (3) the RMSE measurements, (4) the MAE measurements, and (5) the R 2 measurements.
# Input DataANN StructureRMSEMAER 2
(cm/h)(cm/h)
7-9-3-10.04590.01590.9725
77-7-6-10.04600.01640.9720
7-10-4-10.04650.01620.9715
6-5-7-10.04450.01710.9740
66-6-3-10.04550.01710.9742
6-8-4-10.04470.01630.9739
5-8-3-10.04130.01520.9780
55-4-9-10.04170.01560.9774
5-8-8-10.04180.01520.9777
4-9-10-10.04490.01520.9736
44-8-8-10.04500.01560.9735
4-9-9-10.04520.01550.9734
3-9-6-10.04330.01550.9757
33-8-9-10.04340.01540.9757
3-8-6-10.04360.01600.9755
Table 3. Comparison between several works for obtaining K S .
Table 3. Comparison between several works for obtaining K S .
ModelRMSE R 2 Type
This work0.04130.9780ANN
Tamari et al. [31]0.0707NAANN
Brakensiek et al. [9]0.13700.9953PTF
Erzin et al. [12]0.17000.9970ANN
Saxton et al. [32]0.18950.9915PTF
Parasuraman et al. [33]0.1900NAANN
Trejo-Alonso et al. [23]0.19830.9901PTF
Cosby et al. [34]0.43250.9546PTF
Ahuja et al. [35]0.64980.8910PTF
Schaap & Leij [36]0.7130NAANN
Vereecken et al. [37]0.71430.9307PTF
Minasny et al. [38]0.7330NAANN
Ferrer-Julià et al. [39]1.30180.4083PTF
Merdun et al. [40]3.51100.5240ANN
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Trejo-Alonso, J.; Fuentes, C.; Chávez, C.; Quevedo, A.; Gutierrez-Lopez, A.; González-Correa, B. Saturated Hydraulic Conductivity Estimation Using Artificial Neural Networks. Water 2021, 13, 705. https://doi.org/10.3390/w13050705

AMA Style

Trejo-Alonso J, Fuentes C, Chávez C, Quevedo A, Gutierrez-Lopez A, González-Correa B. Saturated Hydraulic Conductivity Estimation Using Artificial Neural Networks. Water. 2021; 13(5):705. https://doi.org/10.3390/w13050705

Chicago/Turabian Style

Trejo-Alonso, Josué, Carlos Fuentes, Carlos Chávez, Antonio Quevedo, Alfonso Gutierrez-Lopez, and Brandon González-Correa. 2021. "Saturated Hydraulic Conductivity Estimation Using Artificial Neural Networks" Water 13, no. 5: 705. https://doi.org/10.3390/w13050705

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop