Integration of Visible and Thermal Imagery with an Artiﬁcial Neural Network Approach for Robust Forecasting of Canopy Water Content in Rice

: A total of 120 rice plant samples were scanned by visible and thermal proximal sensing systems under different water stress levels to evaluate the canopy water content (CWC). The oven-drying method was employed for assessing the canopy’s water state. This CWC is of great importance for irrigation management decisions. The proposed framework is to integrate visible and thermal imaging data using an artiﬁcial neural network as a valuable promising implement for accurately estimating the water content of the plant. The RGB-based features included 20 color vegetation indices (VI) and 6 gray level co-occurrence matrix-based texture features (GLCMF). The thermal imaging features were two thermal indicators (T), namely normalized relative canopy temperature (NRCT) and the crop water stress index (CWSI), that were deliberated by plant temperatures. These features were applied with a back-propagation neural network (BPNN) for training the samples with minimal loss on a cross-validation set. Model behavior was affected by ﬁltering high-level features and optimizing hyperparameters of the model. The results indicated that feature-based modeling from both visible and thermal images achieved better performance than features from the individual visible or thermal image. The supreme prediction variables were 21 features: 14VI, 5GLCMF, and 2T. The fusion of color–texture–thermal features greatly improved the precision of water content evaluation (99.40%). Its determination coefﬁcient (R 2 = 0.983) was the most satisﬁed with an RMSE of 0.599. Overall, the methodology of this work can support decision makers and water managers to take effective and timely actions and achieve agricultural water sustainability.


Introduction
About 70-90% of global water resources are consumed through agriculture [1]. Extreme water use in the agricultural sector also causes water scarcity in semi-arid and arid regions [2]. Rice is one of the most economical crops that requires large amounts of water. Decreasing water availability for agriculture threatens the productivity of the irrigated rice ecosystem and ways to save water and increase the water productivity of rice must be found [3]. By optimizing the timing, duration, and amount of irrigation and drainage, water-saving irrigation practice has been proven effective in saving water and increasing rice yield [4]. To improve the water use efficiency of plants, it is important to accurately and timely predict crop water status. Canopy water content (CWC) is a significant factor in regard to the water use efficiency of plants [5], drought assessment [6], a key input variable in irrigation management decisions, and crop ripening monitoring [7]. Moreover, it is a vital parameter that reflects plant physiological status and health [8]. It is one of the indicators widely used to evaluate a plant's water condition [9]. Traditionally, there are 3 of 23 in improving the prediction model because of the direct control of the behaviors of training algorithms [28]. In this research, the ANN building possibility was used as a tool to support accurate irrigation management decision making by means of visible and thermal sensors. Dawson et al. [29] developed an ANN model for predicting water content in leaves depending on single leaf reflectance and transmittance, which are important input variables to vegetation canopy reflectance models. The R 2 was defined as 0.86 with an RMSE of 1.3%. Marti et al. [30] described the application of ANN to estimate stem water potential from soil moisture at different depths and standard meteorological variables, such as temperature, relative humidity, and solar radiation. The ANN presented high-performance accuracy with an optimum R 2 of 0.926.
To maximize irrigation in rice fields, the plant's water content should be explored accurately via proximal sensing systems. Although visible and thermal imagery are widely used to estimate water content in plants, the best performing algorithm has a high priority in research interests. Based on our knowledge, no prior research has implemented the hybrid methodology of BPNN with features extracted from thermal and RGB imaging to develop a predictive model for accurate estimation of the water content status of rice under different irrigation water treatments and climate change. Hence, the main objectives of this study were (i) to create a well-structured CWC modeling of rice relying on high-level features derived from RGB and thermal images, (ii) to define the superior variants of thermal and RGB features for robust CWC prediction, and (iii) to evaluate the performance of BPNN models using the best-combined features and to adopt the finest model that could be recommended for precision irrigation in the future.

Experimental Design
The experiment was conducted on 120 potted plants of rice under natural conditions at Zhejiang University (120 • 09 E, 30 • 14 N), Hangzhou City, Zhejiang Province, P.R. China. Often, a plastic cover was set up over the plants during precipitation to control the water added for them, and to obtain the heterogeneous water condition from the tested leaves. To prepare the germination, the rice seeds (Xiushui 134) were submerged in the water for four days at a temperature of 28-30 • C, and the water was replaced twice a day. Then, the 10 rice plants were sown into a single PVC (polyvinyl chloride) pot, which had dimensions of 140 × 95 × 125 mm. Each pot was filled with 300 g of black peat moss (HAWITA Gruppe GmbH, Germany) and was irrigated to achieve saturation during the first growing month. The compound fertilizer treatment, N-P 2 O 5 -K 2 O (15-15-15), with a rate of 100 kg ha −1 was applied at 10-day intervals, produced by Qingdao SONEF Chemical Company. Potted rice plants were transplanted on July 10, 2018 and divided into 3 groups, each with 40 pots. It was harvested on August 12, August 30, and September 21, 2018 for the 1st, 2nd, and 3rd groups, respectively. Table 1 shows the summary statistics of air temperature, relative humidity, and vapor pressure deficit (VPD). The values of VPD were determined from the formula stated by Abtew and Melesse [31]. In this work, there were three different growth stages of the rice plants: tillering, stem elongation, and panicle initiation. The water deficit was implemented within a period of 27-34 days, 34-52 days, and 52-74 days of plant life for the 1st, 2nd, and 3rd groups, respectively. As declared in Figure 1, an experiment was carried out with a completely randomized block design in a factorial experiment. The experimental design was divided into three groups, each comprising four irrigation treatments: fully-irrigated (T 1 : 100%), mild (T 2 : 80-70%), moderate (T 3 : 60-50%), and severe water stress (T 4 : 40-30% field capacity), respectively. The number of replicates (R 1 , R 2 , etc.) of treatment was 10 samples. Generally, the gravimetric approach was used to measure the volume of water by manually weighing pots twice a day, supplemented by replacing the transpiration of water to preserve the respective moisture stress conditions. Gross water applied volumes were estimated for each group, 1st stage values were 44.35, 41.65, 39.4, and 37.15 L (liter) for well-controlled, mild, moderate, and severe stress conditions, respectively. Their corresponding values at

Digital Image Acquisition
A near-surface portable digital camera (PowerShot SX720 HS, Canon Inc., Tokyo, Japan) was used to capture digital images of the rice plants. RGB images were acquired in three different visions: elevation, side, and plan view, as displayed in Figure 2. For the plan view, the camera was installed on a tripod and fixed perpendicular with a height of 1m above the plants. In other directions, the camera was mounted at a height of 0.95m from the ground level and kept a distance of 0.8m from the crop canopy with an angle of view of 90°. The pots were moved indoors with a fluorescent background light source for RGB image acquisition, and the background color was a black cover. About 400 RGB images were captured and saved in JPG image quality with a spatial resolution of 3888 × 5184 pixels. We found 40 images with uneven illumination and noise that were excluded, while the remaining 360 images were used to extract the features. The average value of the extracted feature was calculated for each pot. Therefore, the total readings were 120 points that were applied to the data analysis.

Digital Image Acquisition
A near-surface portable digital camera (PowerShot SX720 HS, Canon Inc., Tokyo, Japan) was used to capture digital images of the rice plants. RGB images were acquired in three different visions: elevation, side, and plan view, as displayed in Figure 2. For the plan view, the camera was installed on a tripod and fixed perpendicular with a height of 1 m above the plants. In other directions, the camera was mounted at a height of 0.95 m from the ground level and kept a distance of 0.8 m from the crop canopy with an angle of view of 90 • . The pots were moved indoors with a fluorescent background light source for RGB image acquisition, and the background color was a black cover. About 400 RGB images were captured and saved in JPG image quality with a spatial resolution of 3888 × 5184 pixels. We found 40 images with uneven illumination and noise that were excluded, while the remaining 360 images were used to extract the features. The average value of the extracted feature was calculated for each pot. Therefore, the total readings were 120 points that were applied to the data analysis.

Thermal Image Acquisition
The thermal infrared camera is Tau2-640 (FLIR systems, Inc., USA), which had a resolution of 640 × 512 pixels, and a focal length of 19mm. The sensitivity of the Tau2-640

Thermal Image Acquisition
The thermal infrared camera is Tau2-640 (FLIR systems, Inc., USA), which had a resolution of 640 × 512 pixels, and a focal length of 19 mm. The sensitivity of the Tau2-640 camera is 0.05 • C (50 mK), and the accuracy of the measurement is ±5% after calibration. The thermal sensor has a spectral response in the range of 7.5-13.5 µm. The thermal camera was mounted on a tripod at a distance of 1 m above the crop, and the angular field of view was perpendicular to the plants. The thermal images were taken from the top view of the plant canopy. The water status measurements were conducted between 12:00 and 13:00 local standard time on all plants, when the crop plants had a maximum daytime temperature, were in perfect sun, and with minimal wind movement. The crops exhibit high emissivity as well as low reflectivity [32]. In the thermal images, a significant difference in soil temperature was noticed, but when the soil was masked out, it was not important when remaining focused on the crop. The thermal value obtained by thermal infrared images was originally stored in the raw image in which each pixel is 14 bits. The raw format can be converted to JPG file format to explain the temperature distribution of plants in each part of the image. The following is the equation for calculating the transformed temperature from the thermal raw image [33]: where Th raw and Tem ( • C) represent the thermal value and the corresponding pixel temperature, respectively. The first constant (0.04) is to calibrate the thermal camera and obtain the image pixel temperature ( • K), and the second value (−273.15) is to convert Kelvin to Celsius.

Image Segmentation
Various objects such as rice plants, soil, and pots were included in the thermal and RGB images collected at the different growth stages of the rice plants. It was necessary to separate the plants and backgrounds for the efficient computation of the canopy water content. Firstly, for the RGB images, a thresholding technique was employed, which is a process of image segmentation to convert the image to grayscale [34] and produce a binary image [34], which has two possible pixel values, namely the intensity value of the image that is more than or equal to the threshold value of 1 (white or foreground) and when it is less than the threshold value or value 0 (black or background) that can be removed. Each pixel in the binary image has a size of 1 bit. After separating the rice pixels from the background objects, a set of candidate features, including color and texture, were extracted for further analysis. Secondly, for the thermal images, python software was employed to generate a code for segmenting the rice plants. The ground temperature was higher than any parts in the image, despite the soil and pot being lower than others. Accordingly, we can separate the temperature of the objects as either higher or lower than that of rice plants. Average plant temperature was calculated for all pixels in each thermal image. different irrigation treatments in the three growth stages of the rice. At the 1st stage, the canopy temperatures were observed in the ranges of 32°C-33.5°C (well-watered), 33.8°C-35.4°C (mild stress), 35.75°C-36.5°C (moderate stress), and 36.9°C-37.55°C (severe stress). The canopy temperature ranges for such treatments at the 2nd stage were 37.5°C-39.1°C, 39.22°C-41.72°C, 42.39°C-45.8°C, and 46.24°C-48°C, respectively. Their corresponding values at the 3rd stage were 22.85°C-24.5°C, 25°C-27.7°C, 28°C-30.1°C, and 30.5°C-33°C, respectively.

Feature Extraction from RGB Images
RGB images contain large amounts of data, thereby analyzing the image requires a large volume of memory and leads to more time. Consequently, to decrease the volume of data, an image is depicted using a series of features. The purpose is to capture the image's certain visual properties. Feature extraction has a crucial role in different image processing systems. The most common ways to extract features from RGB images are RGB color space percentage and GLCM-based texture features. The RGB imagery involves three bands of red, green, and blue, where each band has a value of each pixel [35]. The rice plants modified the content of chlorophyll to balance photosynthesis and transpiration under conditions of water stress. The RGB channel values were affected, owing to different levels of canopy water content. The percentage values of the RGB color space were determined as typical features using the following formulas: where R i , G i , and B i are the pixel value for the red, green, and blue band in the digital image, respectively; i and S num are the first pixel and the maximum number of pixels, respectively; R, G, and B are mean values of RGB color space. About 20 color vegetation indices for plant water status estimation were evaluated. Table 2 reveals RGB-based vegetation indices as claimed by numerous previous studies. In our study, these indices were examined to explore their feasibility for CWC modeling and compared as input to the model for adoption of the best indices.

Gray Level Co-Occurrence Matrix-Based Texture Features (GLCMF)
Several studies related to assessing the water status of the plant have used GLCMF: Wenting et al. [21], to predict maize leaf moisture content, and Ushada et al. [46], to estimate the moisture content and leaf water potential of moss. This is based on the fact that changes in the textural appearance of a plant canopy due to the water status reflects tonal variations over the community of plants [47]. The co-occurrence matrix is a statistical method for examining the texture of a grayscale image [48]. The GLCMF analyzes the grayscale relationship between two pixels in the image space that are separated by a certain distance and then randomly extracts the canopy texture feature information premised on probability characteristics. There were six variants of GLCMF in this study such as contrast, dissimilarity, homogeneity, angular second moment (ASM), energy, and correlation. The contrast clearly represents the depth of the image and texture. Dissimilarity measures the distance between pairs of pixels in the region of interest. Homogeneity estimates the closeness of the distribution of elements in the GLCMF. ASM discovers the roughness distribution and texture image. Energy measures textural uniformity. The correlation finds out the degree of correlation in the local grayscale image. The RGB-based GLCMF variables are explicated, as shown in Table 3. Table 3. Description of the GLCMF derived from digital images.

Variable Type
Formula Reference Where P i,j are probabilities calculated for values in GLCMF when i and j are row and column references, respectively. µ i and µ j are the mean values. σ i and σ j are the standard deviation of values. N is the number of rows or columns.

Feature Extraction from Thermal Images
Thermal indicators (T) are features extracted from thermal images depending on the rice plant temperatures, which consist of two indices in this study: the crop water stress index (CWSI) and normalized relative canopy temperature (NRCT).

Canopy Temperature-Based Crop Water Stress Index (CWSI)
The CWSI deduced from canopy temperature for many crops has been widely nominated as an instrument to elucidate plant water conditions and schedule irrigation [50]. The CWSI concept is linked to the standard in which the leaf surface is cooled during the transpiration process, and as soil moisture in the root zone is also depleted, leaf temperature increases and stomatal transpiration and conductance decrease. An empirical approach was proposed to determine crop water stress as reported by Idso et al. [51]: where T a , T c , T nws , and T dry are air temperature ( • C), canopy temperature ( • C), non-waterstressed canopy temperature ( • C), and water-stressed canopy temperature ( • C), respectively. To calculate the NRCT index, wet-bulb and dry-bulb temperatures were suggested by Jackson et al. [23] as the lower and upper baseline, respectively. As stated in the formula by Elsayed and Darwish [9], the NRCT was computed as follows: where T min and T max are the lowest and highest temperatures measured in the experiment ( • C), respectively.

Canopy Water Content Computation
After scanning by the visible and thermal imaging system, the plants in each pot were picked. The roots were removed and the rest was studied to provide a reference for data analysis. Thus, the above-ground part of the plants in one pot was defined as one sample in this study. To obtain the fresh weight (FW), the samples were weighed before drying. Then, the dry weight (DW) was obtained after drying the samples for 24 h in an oven at a temperature of 70 • C. The percentage of canopy water content was estimated by the following formula:

Dataset and Data Analysis Software
A total of 120 readings were split into training, validation, and testing, where 80% (96 samples) were used for the training and validation process of the regression model, while the other 20% (24 samples) were used to verify the model's performance by comparing the expected CWC values with the calculated CWC values. A leave-one-out crossvalidation (LOOCV) approach was utilized to train and validate the model. LOOCV excludes one sample for validation and uses the rest of the samples for training in every trial. This method can decrease over-fitting and permits a more accurate assessment of model prediction strength [52]. The feature selection, model establishment, and data analysis were implemented using Python 3.7.3. The ANN module from the Scikit-learn library version 0.20.2 was used for the regression task. The software was run on a PC with Intel Core i7-3630QM, 2.4 GHz CPU, and 8 GB RAM.

Implementation of a Backpropagation Neural Network (BPNN)
The BPNN employs the multi-layer perceptrons (MLPs) as a supervised learning algorithm that contains a flexible function to train a specific data set [53]. MLP is one of the neural network models that has the same back-propagation structure for supervised training. Typically, back-propagation is used for the training of feed-forward. The neural network is structured from 3 types of layers: (1) the input layer is the primary data for the neural network, (2) the hidden layer is an intermediate layer between the independent inputs layer and dependent output layer, and (3) the output layer produces the results of the specified inputs. Figure 4 displays the architecture of an artificial neural network, which is a class of machine learning algorithms. It uses multiple layers to gradually extract high-level features from the raw input. Five squares as input layers are noted as vector I. The network contains two hidden layers, the number of nodes determined according to regression accuracy. Four squares as hidden layers represent the "activation" nodes and usually are noted as weight; W, W h , and Wo are with layers 1-2, 2-3, and 3-4, respectively. The final square represents the output layer that shows the predicted value of canopy water content. The bias neuron (B) is a special neuron added to each layer in the neural network, which is usually taken to be 1 [54]. The artificial neural networks are generalized mathematical models that use a series of neurons or nodes interlinked by weighted connections to simulate human cognition as it applies to pattern identification and prediction [55].
The generalized weight (W) can be realized using the equation below, where i is the index for each covariate of x, o(x) is the predicted outcome probability by covariate vector, and log-odds is a link function for the neural network. At training, the maximum number of neural network iterations was 1000 iterations. A neural network can stop at exact iterations with the highest network performance. To choose the number of neurons in the hidden layer, the cross-validation technique with the LOOV method was performed on the training dataset. The parameter of limited memory Broyden-Fletcher-Goldfarb-Shanno (lbfgs) was used as a weight optimizer to implement the algorithm efficiently [56]. Otherwise, there are several methodologies to select premium features depending on feature selection algorithms that have become an obvious requirement for prediction and modeling [57]. Glorfeld [58] concluded that an index, such as the following BPNN algorithm in this analysis, can be used to select the most important variables and improves the predictive capacity of the regression model. This index is explicated in the following formula: Input variants involve RGB-based features (color: 20VI and texture: 6GLCMF) and thermal properties (2T). To facilitate the model training process, a transformation was performed to rescale the features. Normalization is transformed across individual features (f ) to adjust for differences in magnitude between different features. Feature normalization (f norm ) is computed by subtracting the minimum image data (f min ) and by dividing the difference between the maximum (f max ) and the minimum feature value as shown by the following formula: The neuron is the primary component of ANN, which involves its elementary mathematical functions. The neuron output (y) can be expressed by the following equation: where W n indicates the nth weights of the combination that produces the output, θ is the unit step function, W is the weight related with the nth input, and µ is the average.
The generalized weight (W) can be realized using the equation below, where i is the index for each covariate of x, o(x) is the predicted outcome probability by covariate vector, and log-odds is a link function for the neural network.
At training, the maximum number of neural network iterations was 1000 iterations. A neural network can stop at exact iterations with the highest network performance. To choose the number of neurons in the hidden layer, the cross-validation technique with the LOOV method was performed on the training dataset. The parameter of limited memory Broyden-Fletcher-Goldfarb-Shanno (lbfgs) was used as a weight optimizer to implement the algorithm efficiently [56]. Otherwise, there are several methodologies to select premium features depending on feature selection algorithms that have become an obvious requirement for prediction and modeling [57]. Glorfeld [58] concluded that an index, such as the following BPNN algorithm in this analysis, can be used to select the most important variables and improves the predictive capacity of the regression model. This index is explicated in the following formula: where M is the important measure for the input variable, n p is the number of input variables, n H is the number of hidden layer nodes, | I | P j is the absolute value of the hidden layer weight corresponding to the pth input variable and the jth hidden layer, and | O | j is the absolute value of the output layer weight corresponding to the jth hidden layer.

Hyperparameter Optimization and High-Level Variants Selection
One of the purposes of this research was to compare different features during training the BPNN model. The three types of the dataset (samples×features) for analysis were 120 × 20, 120 × 6, and 120 × 2. These data were employed as input for the neural network. We optimized the model by selecting the prominent hyperparameters. The parameters were the number of neurons in two hidden layers (nr 1 and nr 2 ) and activation function (fun). The main steps of neural network training, appropriate hyperparameters estimation, and feature arrangement are described in Figure 5. This figure illustrates the pseudo-code for training a neural net with various variants. The number of features in each loop was with color indices: 20, 19, 18, etc.; texture features: 6, 5, 4, etc.; and thermal indicators: 2, 1. Generally, the variants were fed to the model randomly in the 1st loop, the lowlevel features were dropped during each loop, and the superb features were organized in ascending order concerning the highest contribution. Then, the ANN outputs were compared to decide high-ranking variants at minimum RMSECV. where M is the important measure for the input variable, is the number of input variables, is the number of hidden layer nodes, | | is the absolute value of the hidden layer weight corresponding to the pth input variable and the jth hidden layer, and | | is the absolute value of the output layer weight corresponding to the jth hidden layer.

Hyperparameter Optimization and High-Level Variants Selection
One of the purposes of this research was to compare different features during training the BPNN model. The three types of the dataset (samples×features) for analysis were 120 × 20, 120 × 6, and 120 × 2. These data were employed as input for the neural network. We optimized the model by selecting the prominent hyperparameters. The parameters were the number of neurons in two hidden layers (nr1 and nr2) and activation function (fun). The main steps of neural network training, appropriate hyperparameters estimation, and feature arrangement are described in Figure 5. This figure illustrates the pseudocode for training a neural net with various variants. The number of features in each loop was with color indices: 20, 19, 18, etc.; texture features: 6, 5, 4, etc.; and thermal indicators: 2, 1. Generally, the variants were fed to the model randomly in the 1st loop, the low-level features were dropped during each loop, and the superb features were organized in ascending order concerning the highest contribution. Then, the ANN outputs were compared to decide high-ranking variants at minimum RMSECV.

Statistical Analysis Methods
The performance of the artificial neural network was measured through the following selected statistical indicators: mean square error (MSE), root mean square error (RMSE), mean absolute percentage error (MAPE), prediction accuracy (Acc), and coefficient of determination (R 2 ) [59]. All parameters are explicated as follows: CWCact is the

Statistical Analysis Methods
The performance of the artificial neural network was measured through the following selected statistical indicators: mean square error (MSE), root mean square error (RMSE), mean absolute percentage error (MAPE), prediction accuracy (Acc), and coefficient of determination (R 2 ) [59]. All parameters are explicated as follows: CWC act is the actual value that was estimated from laboratory calculations, CWC p is the predicted or simulated value, CWC ave is the average value, and NP is the total number of data points.

Optimization Combinations of Independent Variables Selection
Various combinations of variables were examined to predict CWC, and the top neural network architectures for hidden configurations were presented. The numbers of hidden neurons were different due to the variations in the feature values extracted from the visible and thermal images. The proposed features of color indices, texture characteristics, and thermal indicators were evaluated to adopt the best-combined variants.

Color Vegetation Indices (VI)
The optimal VI identified in each loop during neural network training are clarified in Figure 6. The neural network performance was estimated with different VI. As explained in Table 4, model performance metrics (MSE, RMSE, and MAPE) were evaluated through training, cross-validation, and a test set. Moreover, the prediction accuracy (Acc) was calculated. The best hyperparameters such as hidden neurons, activation function, and the maximum number of iterations of the neural network were chosen. Several architectures of the BPNN model relying on VI were collected. The best neural network results were attained by fusing data of 14VI at the 7th loop, as displayed in Figure 6. The finest hyperparameters of activation function, hidden neurons, and maximum iterations were logistic, (19,15) tectures of the BPNN model relying on VI were collected. The best neural network results were attained by fusing data of 14VI at the 7th loop, as displayed in Figure 6. The finest hyperparameters of activation function, hidden neurons, and maximum iterations were logistic, (19,15), and 119 iterations, respectively. The RMSECV decreased to 2.291 vs. 2. 444 compared to 14VI vs. 20VI. The prediction ability of the BPNN-VI-14 model increased equivalent to the model with full VI, with prediction R 2 (0.632 vs. 0.556), MAPE (2.832% vs. 3.273%), RMSE (2.788 vs. 3.065), MSE (7.776 vs. 9.394), and Acc (0.972 vs. 0.967).   Where nf is number of features, Iter is the maximum iteration of the neural network, CV is a cross-validation dataset, and * denotes best-combined VI with the lowest RMSECV.

GLCM-Based Texture Features (GLCMF)
The highest variants of GLCMF selected in each loop are exposed in Figure 7. The different configurations of the GLCMF-based BPNN model and its performance output (MSE, RMSE, and MAPE) with cross-validation, training, and test set are mentioned in Table 5. The optimal neural network was polished by conjoining 5GLCMF at the 2nd loop. The best GLCMF were ASM, homogeneity, contrast, dissimilarity, and correlation, respectively. The ideal parameters of activation function, hidden neurons, and maximum iterations at the fewest value of RMSECV (2.148) were hyperbolic tan, (6,10), and 175 repetitions, respectively. The BPNN-GLCMF-5 model performed better than the BPNN-GLCMF-6 model. The high-level model enhanced the R 2 from 0.641 to 0.679, MAPE from 3.151% to 2.938%, RMSE from 2.757 to 2.607, MSE from 7.599 to 6.794, and Acc from 0.968 to 0.971. different configurations of the GLCMF-based BPNN model and its performance output (MSE, RMSE, and MAPE) with cross-validation, training, and test set are mentioned in Table 5. The optimal neural network was polished by conjoining 5GLCMF at the 2nd loop. The best GLCMF were ASM, homogeneity, contrast, dissimilarity, and correlation, respectively. The ideal parameters of activation function, hidden neurons, and maximum iterations at the fewest value of RMSECV (2.148) were hyperbolic tan, (6,10), and 175 repetitions, respectively. The BPNN-GLCMF-5 model performed better than the BPNN-GLCMF-6 model. The high-level model enhanced the R 2 from 0.641 to 0.679, MAPE from 3.151% to 2.938%, RMSE from 2.757 to 2.607, MSE from 7.599 to 6.794, and Acc from 0.968 to 0.971.   Where * refers to best-combined GLCMF at the smallest value of RMSCV.

Thermal Indicators (T)
The rice plants, at three growth stages, were grown with four levels of irrigation under outdoor conditions. The temperature of the rice canopy was affected by the water stress. The temperatures of non-water stressed and water-stressed canopy were 31 • C and 38 • C at the 1st stage, 37 • C and 49.1 • C at the 2nd stage, and 22 • C and 34.07 • C at the 3rd stage, respectively. We exploited the previous temperatures for computing the two thermal indicators: CWSI and NRCT. The relationship of both CWSI and NRCT with the CWC of rice at each growth stage is designated in Figure 8. It was observed to be inversely related, and the R 2 at each stage independently was higher than the R 2 at three stages. By applying ANN with thermal indicators, the prediction formula to compute the CWC at any stage of rice growth was explored with high accuracy. The neural network outputs with thermal indicators at the 1st and 2nd loop are illuminated in Table 6. It depicts the optimal parameters of the T-based BPNN model and its outputs (R 2 , MAPE, RMSE, MSE, and Acc) through cross-validation, training, and test set. The results approved that the BPNN-CWSI-NRCT was a high-quality model for CWC computation (R 2 = 0.803, MAPE = 1.989%, RMSE = 2.043, MSE = 4.172, and Acc = 0.980) as opposed to BPNN-CWSI (R 2 = 0.505, MAPE = 3.782%, RMSE = 3.239, MSE = 10.473, and Acc = 0.962). The RMSECV decreased from 2.549 to 1.116 against 1 and 2T, respectively. The model of BPNN-T-2 was excellent for estimating the CWC by (19,18) neurons in two hidden layers, logistic as activation function, and 217 duplications. and Acc) through cross-validation, training, and test set. The results approved that the BPNN-CWSI-NRCT was a high-quality model for CWC computation (R 2 =0.803, MAPE=1.989%, RMSE=2.043, MSE=4.172, and Acc=0.980) as opposed to BPNN-CWSI (R 2 =0.505, MAPE=3.782%, RMSE=3.239, MSE=10.473, and Acc=0.962). The RMSECV decreased from 2.549 to 1.116 against 1 and 2T, respectively. The model of BPNN-T-2 was excellent for estimating the CWC by (19,18) neurons in two hidden layers, logistic as activation function, and 217 duplications.   Where a is the best-combined variables of CWSI-NRCT, and b is the indicator of CWSI.

Best-Combined Features Extracted from Visible and Thermal Imagery
The findings inferred that in comparison with the BPNN-GLCMF-5 and BPNN-VI-14 models, the BPNN-T-2 model had dependable foretelling. Further, the thermal indicators (T) were joined with VI and GLCMF to improve the forecasting performance and discover the unmatched model. The three approaches were espoused as a new expectation procedure of the rice water content, incorporating VI-T, GLCMF-T, and VI-GLCMF-T features with the BPNN model. Their outputs (MSE, RMSE, and MAPE) and notable parameters are disclosed in Table 7 The distinguishing parameters of the most robust model were with hidden layer neurons (8,9), activation function (logistic), and maximum iterations (422). The BPNN-VI-GLCM-T-21 model had quick predictive outcomes within 31.911 s for a single sample compared to the oven-drying method [17,60]; a test time of 24 h is required to dry the sample. The findings are in agreement with Meeradevi et al. [61] who used ANN as a fast tool with a prediction time of 9.74 s. The results elucidated that the improvement in prediction accuracy of CWC can be fulfilled if adequate neural network parameters and higher variables are assigned.  Where a is 5 GLCMF and 2 T, b is 14 VI and 2 T, c is 14 VI and 5 GLCMF and 2 T, and C.T. is the consumption time required to analyze the data across a single sample to predict the CWC (sec).

Neural Network Learning Curves with Super Features
The learning curve of the regression model was promoted in this study through the following considerations: (1) number and characteristics of nominated features, (2) picking the supreme parameters, and (3) neural network definite iterations to avoid model over-fitting. The first two points are discussed in the previous sections. The last factor refers to the neural network stopping at specific iterations that can progress the network performance, as indicated in Figure 9. These curves exhibit an evaluation of the model's behavior with the higher variants at training, cross-validation, and testing. As the number of iterations increases, the accuracies of training, cross-validation, and testing increase gradually to reach a point where the learning curves show a good model. As publicized in Figure 9a-d, the optimal repetitions of the BPNN training process through the highest variables (14VI, 5GLCMF, 2T, and VI-GLCMF-T-21) were 119, 175, 217, and 422 steps, respectively. The learning curves of the innovative models reached a cross-validation accuracy of 0.625, 0.656, 0.851, and 0.984, respectively. The learning outcomes verified that the accuracy measure steadily increases over iterations of training. When the permissible iterations of the neural network are exceeded, the model generally takes the form of making an overly complex model. Moreover, training accuracy was usually higher than crossvalidation accuracy. The three-stage model of BPNN-VI-GLCMF-T was accomplished with 21 variables and 422 steps. Its prediction accuracy was 99.4% (R 2 = 0.983 with an RMSE of 0.599), and the learning curve performance was impressive (Figure 9d). The expected performance was improved as said by Thawornwong and Enke [62]; to avoid over-fitting, the network was trained by the back-propagation algorithm with early stopping.

Neural Network Topology with Higher Variants
The neural network architectures after gathering senior features are introduced in Figure 10. This figure showed the best neural network structure with the variants chosen. Each network topology offers basic information such as the synaptic weights trained, a range of hidden neuron layers, steps for converging, and the overall errors. The network topology is constructed with a specific combination of input variables with a number of hidden neuron layers. For instance, the model of BPNN-VI-14 had hidden neuron layers (19,15), BPNN-GLCMF-5 needed (6,10), BPNN-T-2 required (19,18), and BPNN-VI-GLCMF-T-21 desired (8,9). The layout presented in Figure 10a depicts the BPNN-VI-14 model; the training process needed 119 steps to achieve a lower error function. The process has an overall error of roughly 2.358. The learning process of the BPNN-GLCMF-5 model (Figure 10b) demands 175 steps, containing an overall error of about 3.130. At the BPNN-T-2 model (Figure 10c), the training process obligated 217 steps with a total error of about 0.960. At the BPNN-VI-GLCMF-T-21 model (Figure 10d), the learning process requested 422 steps, and the overall error was 0.0317.

Canopy Water Content Prediction and Validation
From the results, the thermal features had a high ranking for measuring the CWC of rice in contrast to VI and GLCMF. The super thermal variables were linked to acceptablelevel features as follows: VI-GLCMF-T, VI-T, and GLCM-T. The VI-GLCMF-T features were the premium integration to filter the uppermost variables. The neural network was trained with the previous features (independent variables) for predicting the CWC (dependent variable). The expected CWC values were then compared with the reserved values not implemented for the neural network. Figure 11 illustrates the scatter plots of observed and predicted CWC in rice exploiting the proposed features. This study evaluated multivariate methods and compared the results clearly, so the use of multivariate methods greatly enhances predictability. Independent validation can also be considered the most robust method for evaluating the accuracy of the regression model since validation data are not involved in the model development process. The BPNN-VI-GLCMF-T-21 was the first best predictive model as evidenced by the performance and showed a stronger relationship between the superlative features and CWC. These features involved in this model are of great significance for predicting water content. Its outputs with R 2 were 0.984 (cross-validation) and 0.983 (test set). The BPNN-GLCMF-T-7 model ranked second in performance. The R 2 value was 0.982 and 0.972 in the cross-validation and test set, respectively. The BPNN-VI-T-16 was the third highest accurate model (R 2 = 0.956 and 0.978 for cross-validation and test set, respectively). Models constructed with individual features of either VI or GLCMF had lower performance expectations, while these features combined with 2T achieved robust forecasting. Otherwise, the residual value plays a vital role to validate the obtained regression model. The difference between the actual CWC and the predicted CWC value is the residual. We calculated the residual value with the first three higher models. At BPNN-VI-GLCM-T-21, −1.223% and 1.916% were the lowest and highest residuals for the CWC prediction, respectively. At BPNN-VI-T-16, the lowest and highest residuals were −1.128% and 1.380%, while the lowest residuals of −0.583% and highest of 2.859% were calculated with BPNN-GLCMF-T-7.  (Figure 9d). The expected performance was improved as said by Thawornwong and Enke [62]; to avoid over-fitting, the network was trained by the back-propagation algorithm with early stopping.

Neural Network Topology with Higher Variants
The neural network architectures after gathering senior features are introduced in Figure 10. This figure showed the best neural network structure with the variants chosen. Each network topology offers basic information such as the synaptic weights trained, a range of hidden neuron layers, steps for converging, and the overall errors. The network  (Figure 9d). The expected performance was improved as said by Thawornwong and Enke [62]; to avoid over-fitting, the network was trained by the back-propagation algorithm with early stopping.

Neural Network Topology with Higher Variants
The neural network architectures after gathering senior features are introduced in Figure 10. This figure showed the best neural network structure with the variants chosen. Each network topology offers basic information such as the synaptic weights trained, a range of hidden neuron layers, steps for converging, and the overall errors. The network

Canopy Water Content Prediction and Validation
From the results, the thermal features had a high ranking for measuring the CWC of rice in contrast to VI and GLCMF. The super thermal variables were linked to acceptablelevel features as follows: VI-GLCMF-T, VI-T, and GLCM-T. The VI-GLCMF-T features were the premium integration to filter the uppermost variables. The neural network was trained with the previous features (independent variables) for predicting the CWC (dependent variable). The expected CWC values were then compared with the reserved values not implemented for the neural network. Figure 11 illustrates the scatter plots of observed and predicted CWC in rice exploiting the proposed features. This study evaluated multivariate methods and compared the results clearly, so the use of multivariate methods greatly enhances predictability. Independent validation can also be considered the most robust method for evaluating the accuracy of the regression model since validation data are not involved in the model development process. The BPNN-VI-GLCMF-T-21 was the first best predictive model as evidenced by the performance and showed a stronger relationship between the superlative features and CWC. These features involved in this model are of great significance for predicting water content. Its outputs with R 2 were 0.984 (cross-validation) and 0.983 (test set). The BPNN-GLCMF-T-7 model ranked second in performance. The R 2 value was 0.982 and 0.972 in the cross-validation and test set, respectively. The BPNN-VI-T-16 was the third highest accurate model (R 2 = 0.956 and 0.978 for bined with 2T achieved robust forecasting. Otherwise, the residual value plays a vital role to validate the obtained regression model. The difference between the actual CWC and the predicted CWC value is the residual. We calculated the residual value with the first three higher models. At BPNN-VI-GLCM-T-21, −1.223% and 1.916% were the lowest and highest residuals for the CWC prediction, respectively. At BPNN-VI-T-16, the lowest and highest residuals were −1.128% and 1.380%, while the lowest residuals of −0.583% and highest of 2.859% were calculated with BPNN-GLCMF-T-7.

Discussion
Color and texture are two important aspects in digital imagery. Plant temperature is also a remarkable tool in thermal imaging, which can be used as an indicator to identify a plantʹs water condition. The plant color feature can be used for plant stress assessment [63]. Texture analysis is significant in many areas such as remote sensing and its common applications include image classification and pattern recognition [64]. These features of VI and GLCMF were applied with a neural network for CWC quantification in rice. The results confirmed that the neural network behavior with GLCMF was superior to VI for predicting the water status of plants. This is similar to applying classification in the following studies: Jana et al. [65], to classify eight varieties of fruits, and Dubey and Jalal [66], to recognize diseases in fruits. Texture features performed better than color features for classification via support vector machine algorithm with accuracy, 32.29% and 69.79% and 83.5% and 88.56%, respectively. Moreover, the thermal analysis results agreed with the findings of Jones [67], who identified that water content or transpiration has an inverse relationship with leaf temperature. The CWC values were highly influenced owing to the

Discussion
Color and texture are two important aspects in digital imagery. Plant temperature is also a remarkable tool in thermal imaging, which can be used as an indicator to identify a plant's water condition. The plant color feature can be used for plant stress assessment [63]. Texture analysis is significant in many areas such as remote sensing and its common applications include image classification and pattern recognition [64]. These features of VI and GLCMF were applied with a neural network for CWC quantification in rice. The results confirmed that the neural network behavior with GLCMF was superior to VI for predicting the water status of plants. This is similar to applying classification in the following studies: Jana et al. [65], to classify eight varieties of fruits, and Dubey and Jalal [66], to recognize diseases in fruits. Texture features performed better than color features for classification via support vector machine algorithm with accuracy, 32.29% and 69.79% and 83.5% and 88.56%, respectively. Moreover, the thermal analysis results agreed with the findings of Jones [67], who identified that water content or transpiration has an inverse relationship with leaf temperature. The CWC values were highly influenced owing to the loss of water through transpiration; CWSI and NRCT reflected normalized temperature values for the crop. These results are in agreement with Blum et al. [68], who indicated that the CWC decreases with increasing canopy temperature as a result of increased water stress.
The current work has allowed a more accurate determination of the water conditions in rice, using the best-combined features extracted from RGB and thermal images. The proposed approaches use all of the sensitive features to changes in the CWC that greatly increase the model performance. Our outcomes achieved greater accuracy than those of Alchanatis et al. [69], who demonstrated that fusion of thermal and visible imaging can provide precise data on the crop water status of grapevines, and the CWSI was highly correlated with measured stomatal conductance with an R 2 value of 0.97. The developed model outperformed Leinonen and Jones [70], who stated that the combination of thermal and visible imagery was a more accurate tool for estimating canopy temperature and could be used to identify water deficit stress in a plant. They investigated the relationship between the measured stomatal conductance and the calculated CWSI with an R 2 of 0.867. Likewise, the first-order model of the BPNN-VI-GLCM-T-21 was very precise in contrast to that of Sun et al. [71], who referred to a reasonable model for computing CWC in wheat based on the ratio vegetation index (RVI: 1605 and 1712 nm) and the normalized difference vegetation index (NDVI: 1712 and 1605 nm), having the highest R 2 and lowest RMSE in model calibration and validation (R 2 c = 0.74 and 0.73; RMSEC = 0.026 and 0.027; R 2 v = 0.72 and 0.71; RMSEV = 0.028 and 0.029). Furthermore, the proposed model is better than Ge et al. [72], who concluded that leaf water content in maize at a pot-scale is successfully predicted with the hyperspectral images using the PLSR model for two genotypes in model cross-validation (R 2 = 0.81 and 0.92; RMSE = 3.7 and 2.3; MAPE = 3.6 and 2.2). The created model achieved high performance compared to Pandey et al. [73], who reported that PLSR analysis can be performed to predict the leaf water content of pot-grown maize and soybean plants with the highest accuracy (R 2 = 0.93, RMSE = 1.62, and MAPE = 1.6%) for validation. In addition, this research is in agreement with the following study in that the combination of color-GLCM-thermal features are high-quality variants in regression or classification applications. Bhole et al. [74] used different features such as the color (CM: color moments and CCV: color coherence vector), texture (GLCMF), and thermal (T) with a random forest for classification of eleven categories of fruits. The results showed that the model accuracies for integrating GLCM-T, CM-T, CCV-T, and CM-CCV-GLCM-T features were 84.26%, 91.17%, 92.95%, and 93.4%, respectively.
Finally, remote and proximal sensing images acquired with high-resolution cameras, mounted at ground level or on unmanned aerial vehicles (UAV), have spatial resolutions of a few centimeters. These images can provide sufficient accurate information for both assessing plant water status in the field and implementing appropriate irrigation management strategies [75]. Thus, this work can assist in improving inexpensive, effective high-throughput phenotyping platforms for large numbers of breeding plants at different levels of irrigation. A practical sharing of research may improve site-specific rice irrigation management through designing an intelligent irrigation system based on the best-proposed model. The methodology of this study that relied on ground-based cameras can be scaled up to UAV-based applications to increase productivity [76] and monitor the water condition of crops in large-scale areas.

Conclusions
Estimation of canopy water content (CWC) is highly important in precision plant reproduction and agricultural development. Low-cost outdoor cameras, such as visible and thermal imaging systems, would be an applicable implement for predicting water content in the plant. Therefore, the present study explored the ability to incorporate toplevel features retrieved from visible and thermal imaging with a back-propagation neural network (BPNN) to adopt a three-stage model of CWC for rice. Hand-crafted features, including 20 vegetation indices (VI), 6 GLCM texture features (GLCMF), and 2 thermal indicators (T), were identified for analysis. The experimental results showed that the proposed model of BPNN-VI-GLCMF-T provided effective recognition of CWC in the rice crop. Expectation accuracy increased to 99.4% by conjoining 21 superlative features. At hidden neuron layers (8,9), the R 2 raised to 0.983 with an RMSE of 0.599. The models with separate features performed lower than the best built-in features. Their corresponding values with prediction R 2 were 0.632 (RMSE = 2.788), 0.679 (RMSE = 2.607), and 0.803 (RMSE = 2.043) using 14VI, 5GLCM, and 2T, respectively. Ultimately, the superlative model has a high level of confidence and reliable outcomes. In the future, this tool may open an avenue for rapid, high-throughput assessments of the water condition of plants, as well as being equally important for procedures related to agricultural water management.