Next Article in Journal
Fault Diagnosis of Brake Train Based on Multi-Sensor Data Fusion
Previous Article in Journal
Emitter Identification of Digital Modulation Transmitter Based on Nonlinearity and Modulation Distortion of Power Amplifier
Previous Article in Special Issue
Calibration of Electrochemical Sensors for Nitrogen Dioxide Gas Detection Using Unmanned Aerial Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Novel Feature-Extraction Methods for the Estimation of Above-Ground Biomass in Rice Crops

by
David Alejandro Jimenez-Sierra
1,†,
Edgar Steven Correa
2,†,
Hernán Darío Benítez-Restrepo
1,
Francisco Carlos Calderon
2,
Ivan Fernando Mondragon
2 and
Julian D. Colorado
2,*
1
Department of Electronics and Computer Science, Pontificia Universidad Javeriana Cali, Cali 760031, Colombia
2
School of Engineering, Pontificia Universidad Javeriana Bogota, Cra. 7 No. 40-62, Bogota 110311, Colombia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2021, 21(13), 4369; https://doi.org/10.3390/s21134369
Submission received: 14 May 2021 / Revised: 17 June 2021 / Accepted: 23 June 2021 / Published: 25 June 2021
(This article belongs to the Special Issue Emerging Robots and Sensing Technologies in Geosciences)

Abstract

:
Traditional methods to measure spatio-temporal variations in above-ground biomass dynamics (AGBD) predominantly rely on the extraction of several vegetation-index features highly associated with AGBD variations through the phenological crop cycle. This work presents a comprehensive comparison between two different approaches for feature extraction for non-destructive biomass estimation using aerial multispectral imagery. The first method is called GFKuts, an approach that optimally labels the plot canopy based on a Gaussian mixture model, a Montecarlo-based K-means, and a guided image filtering for the extraction of canopy vegetation indices associated with biomass yield. The second method is based on a Graph-Based Data Fusion (GBF) approach that does not depend on calculating vegetation-index image reflectances. Both methods are experimentally tested and compared through rice growth stages: vegetative, reproductive, and ripening. Biomass estimation correlations are calculated and compared against an assembled ground-truth biomass measurements taken by destructive sampling. The proposed GBF-Sm-Bs approach outperformed competing methods by obtaining biomass estimation correlation of 0.995 with R 2 = 0.991 and R M S E = 45.358 g. This result increases the precision in the biomass estimation by around 62.43 % compared to previous works.

1. Introduction

Rice is an essential grain to ensure global food security [1], contributing to 20 % of food energy requirements worldwide. Major efforts have emerged to develop novel high-throughput methods for enabling precision biomass characterization aimed at improving crop yield, grain quality and crop management [2,3,4].
Monitoring rice biomass at larger crop scales requires remote sensing approaches for precision sampling, mostly based on unmanned aerial vehicle (UAV-driven) multispectral imagery [3,4,5,6]. For this purpose, Above-Ground Biomass Estimation (AGBE) methods have recently gained significant traction [7,8,9,10], since machine learning models can be trained by features extracted according to different plant reflectance wavelengths, namely vegetation indices (VIs). These VIs, are computed from multispectral imagery captured by UAVs or satellites. Although the combination of several VIs tends to avoid saturation issues with higher values of accumulated biomass [11], in some cases the estimation of the biomass (based on VI extracted features) requires independent models that must be characterized and calibrated dependent on the crop growth stages. Furthermore, VI features are highly affected by the genotype (rice variety) and external abiotic conditions [5,6,12,13,14].
Most of the AGBE methods reported in the literature require image segmentation algorithms to extract the VI formulas from the corresponding canopy [3,4,5,6,15]. Therefore, the accuracy of the extracted VI-features strictly depends on the quality of the segmentation, where several factors such as image resolution, saturation and radiation noise, play an important role in achieving an accurate correlation between the segmented pixels (VI-features) with the biomass. To overcome this problem, recent approaches based on graphs [4] propose the fusion of different wavelengths images into a single graph, where the eigenvectors can be used as features. As a result, graph-based methods do not require image segmentation or other photogrammetry corrections that can be computational expensive.
In this paper, we compare two approaches for the extraction of relevant features from multispectral aerial imagery, allowing the estimation of above-ground biomass based on training classical machine learning models. In previous work reported in [3,4], we introduced the aforementioned methods. The former method is called GFKuts, which was primarily used as an image segmentation method to optimally segment the crop canopy. The latter was based on graphs data fusion, which instead of using VIs or a mask image to extract features, it use the structural information captured by the graphs.

2. Related Work

Several approaches are used to address the above-ground biomass estimation (AGBE) problem [3,4,5,6,15]. For instance, authors in [15] conducted a comprehensive survey to identify which VIs were suitable for estimating rice biomass as a function of the growth stages of the crop. Furthermore, in [15,16], crop features were extracted by using classical k-means binary clustering for segmentation, where linear multi-variable regression models were applied to each crop stage independently for the biomass characterization. In [3], a more sophisticated method named GFKuts was presented, aimed at combining a Monte Carlo K-means with the well-known GrabCut segmentation method [17,18]. Overall, the GFKuts approach is based on three image processing methods: Magic Wand, Intelligent Scissors [19,20], and GrabCut that is grounded on the Graph-Cut method [17,18]. In GFKuts, the energy function of the Graph-Cut method was adjusted, to achieve a global optimization for a N-dimensional image through a Gaussian Mixture Model. Also, the proposed GFKuts algorithm is fully automatic [21,22].
In [23], a deep learning approach was used for the estimation of biomass in forage grass crops. Convolutional neural networks were trained with extensive datasets that are computational expensive to process, while over-fitting issues were also observed. Other approaches based on structural information combine multiple feature sources through a fused graph [4,24,25]. These methods relies on the approximation of the graphs given by the Nyström extension. In this paper, we extend the original graph method presented in [4], by adding a new sampling technique known as blue-noise sampling. This is combined with a graph signal processing approach to enhance the estimation given by the Nyström extension of the graph in [4]. As mentioned, this method uses the eigenvector from the graphs as the features to extract from the multispectral images. As a result, we expect to improve on the estimation of the above-ground biomass, by correlating the estimations against biomass measurements directly obtained from the crop by destructive sampling.

3. Materials and Methods

The general architecture for UAV-driven remote sensing of above-ground biomass in rice crops is presented in Figure 1. An UAV is equipped with a multispectral camera onboard that captures the reflectance of the light spectrum in four different bands (Green, Red, Red-Edge, and NIR). A set of 1868 (i.e., one image per band) images was acquired for the crop’s three growth stages: vegetative, reproductive, and ripening. Further details on the dataset acquisition and crop characteristics can be found in [3].
As mentioned, this paper presents two approaches for feature extraction. The first approach is GFKuts, which is based on the image processing area. This approach is an entirely automatic proposal that integrates a binary classification technique with an optimization approach based on a Gaussian mixture model, followed by a filtering stage, to extract a mask related to pixels that belong to the crop canopy. The second approach is a graph-based framework that aims to fuse data by taking advantage of the structural relationship between the data captured by independent graphs (i.e., the structural relationship of the pixels of an image with one graph per image channel). The method generates a resultant fused graph with the most relevant information. Unlike GFkuts, this method uses eigenvectors of the fused-graph as features rather than vegetation indices. Lastly, the extracted features obtained from both methods are the input of a regression model to estimate the above-ground biomass. The next subsections explain the features extraction approaches and the regressors.

3.1. GFKuts

Figure 2 details the proposed GFKuts approach. The first contribution of GFKuts is to apply a random sampling approach based on Monte Carlo method and integrate these samples into a binary classification method [26]. This random arrangement and the binary classification algorithm solve a time-demanding computational iteration problem as they provide an efficient stochastic numerical method based on a photogrammetric image technique used in precision agriculture.
An array I = ( i 1 , , i n ) of size N, denotes the input image in the standard RGB color space (sRGB). The binary non-supervised classification method implemented in this work is the k-means clustering algorithm that cluster the set of N samples into K groups. The semantic labeling of the groups is random according to the nature of the algorithm. Thus, there is no control of the true class of the label assigned. Given that the results are highly repeatable, the identification of each group is carried out through the characterization of the centroid. This characterization runs only once and operates as long as the conditions of the images do not change (e.g., the type of crop, growing, or environmental conditions).
The Montecarlo sample K-means classification strategy generates a selection of pixels with a uniform random distribution of the spatial coordinates n ( x , y ) of the image. The result is a subset of sRGB values for each selected pixel N. These values come from the refractive bands of the light spectrum captured by the sensor. The implementation of this first stage is shown in Algorithm 1. The result is a pair of masks that allow us to initialize regions of the image to implement the optimization process.
Algorithm 1: Montecarlo Sampled K-means.
Input: the image I , the number of samples N, and two heuristic values associated with the mean expected
  radiance of the canopy: C 1 , and the ground: C 2 .
  for Each pixel in range (1 n) do
   Random (x, y) pixel selection from I to P
   Store sRGB value from P to Feature n
   Store pixel coordinates I x , y
  end for
  Run K-means over F e a t u r e to get labeling I n , k = 1 and I n , k = 2
  calculate the Euclidean distance between C1 and the centroid of the group K = 1 to m 1 ;
  calculate the Euclidean distance between C1 and the centroid of the group K = 2 to m 2 ;
  if m 1 < m 2 then
   group K = 1 to c a n o p y
   group K = 2 to g r o u n d
  else
   group K = 1 to g r o u n d
   group K = 2 to c a n o p y
  end if
  Create a mask T F and set the coordinates in I x , y of each pixel in I n , c a n o p y as the foreground.
  Create a mask T B and set the coordinates in I x , y of each pixel in I n , g r o u n d as the background.
The image segmentation consists of inferring the unknown opacity variables, denoted as α , from the given image data I and the model θ [18]. The opacity term α = ( α 1 , , α N ) is the image segmentation weighed by each pixel with 0 α N 1 for soft segmentation and α N ( 0 , 1 ) for hard segmentation, with 1 denoted as foreground and 0 for background. The parameters θ describe image foreground and background distributions modeled by GMMs values, that introduces the covariance parameter k = ( k 1 , , k N ) with k n ( 1 , , K ) assigned for each pixel. The “Gibbs” energy function of Equation (1) denoted as E , models a trend of solidity of the objects through the opacity parameter, whose minimum corresponds to an optimal segmentation.
E ( α ̲ , k , θ ̲ , I ) = U ( α ̲ , k , θ ̲ , I ) + V ( α ̲ , I ) .
The function U ( ) in Equation (1) is explained in Equation (2); this function estimate the fitness of the opacity distribution α of the image, given the model θ , in which p ( ) is a Gaussian probability distribution and π ( · ) are mixed weights coefficients.
U ( α ̲ , k , θ ̲ , I ) = n l o g p ( z n | α n , k n , θ ) l o g π ( α n , k n )
The remaining function V ( ) shown in Equation (3) is called the smoothness term. The constant γ is defined empirically as 50 by [18], given images in a 0–255 range. β is a constant that regulates the smoothness term.
V ( α ̲ , z ) = γ ( m , n ) d i s ( m , n ) 1 [ α m α n ] β e x p ( Z m Z n ) 2
The segmentation can be estimated as a global minimum of α parameter over the energy model, as seen in Equation (4)
α = a r g m i n E α ( α , θ ̲ )
After the optimization process, an image filtering stage is introduced. Linear and time-invariant filters are usually used. A powerful approach, is to optimize a quadratic function that directly imposes constraints on the unknown output. The solution is obtained by solving a dispersion matrix encoded with the information from the guide image [27,28]. Processing is done around each pixel using a weighted average of the nearby pixels. The processing information is based on the color and intensity of the guide image [27]. This proposal is developed with the bilateral filter, starting with the guided filter. The filter can preserve the edges and smooth out small fluctuations [29]. This approach is based on explicitly constructing the cores using an orientation image. The output is a linear transform of the guide image. This filter has the edge-preserving and anti-aliasing property such as a bilateral filter, but does not suffer from gradient inversion problems [28,29].
The filter stage has two inputs, a guide image ‘I’ and an input image ‘p’. The output is expressed as a weighted average ‘q’. The parameters i and j are the pixel indices, as seen in Equation (5)
q i = j W i j ( I ) p j
The guide image ‘I’ is the reference of the filter kernel W i j independent of ‘p’. The kernel weights can be explicitly expressed by Equation (6)
W i j G F ( I ) = 1 | ω | 2 k ; ( i , j ) ω k 1 + ( I i μ k ) ( I j μ k ) σ k 2 + ϵ
The parameters μ k and σ k 2 are the mean and variance of w k in image ‘I’ respectively, ϵ is a regularization parameter and w is the number of pixels in w k [28].

3.2. Graph-Based Data Fusion

Multi-channel input images denoted as X n c R m × n , with c = 1 , , C h and n = 1 , , N , where C h is the number of channels of each image and N is the total number of images. Here, the goal is to extract relevant features given by each channel of the images, in order to train a model that captures the biomass behavior across the crop stages. As explained in [4], each channel of images are represented by a corresponding graph, with the purpose of obtaining a single fused-graph from all the channels. The fused-graph is then used as an embedded space of images to extract relevant features.
The graphs used in the approach in [4] are undirected, and they are denoted as a triplet G = ( V , E , w ) , consisting of vertexes V , and edges E V × V , and a non-negative weight function w : V × V [ 0 , ) . Therefore, a multichannel image X n c can be represented as graph G by defining a vertex set V , an edge set E V × V , where each edge represents the relationship between two pixels, and a weight function w that measures the strength of that relationship. Typically w is defined by:
w ( v i , v j ) = exp dist v i , v j 2 σ 2 ,
where dist ( v i , v j ) is the Euclidean distance between the nodes v i and v j associated to the pixels X n c , and σ > 0 is a scaling parameter.
Given the graphs for all channels in the image ( X n c ) G c = ( V , E c , w c ) R m n × m n , a fused graph G f = ( V , E f , w f ) R m n × m n can be defined by combining the weight functions of G c as follows:
w f ( v i , v j ) = min ( w 1 ( v i , v j ) , w 2 ( v i , v j ) , , w C h ( v i , v j ) ) ,
Therefore, the eigenvectors ( u ) and eigenvalues ( λ ) are computed from W f R m n × m n by solving the eigen problem of the normalized Laplacian, defined as:
Lu = λ Du ,
where the Laplacian L = I D 1 / 2 W f D 1 / 2 and D is the degree matrix that is diagonal has elements d j = i w f ( v i , v j ) ).

3.2.1. Nyström Extension

Since the matrix W f R P × P with P = m n cannot be computed the authors in [4] leveraged the Nyström extension to compute the eigenvectors of W f without computing the complete matrix, while using some samples ( n s ) of X n c , as follows:
W f = A B B C ,
where A R n s × n s , B R n s × ( P n s ) and C R ( P n s ) × ( P n s ) . This method approximates C by using n s samples from the total number of pixels P X n c with n s P . Thus, the eigenvectors of the matrix W f can be spanned by eigenvalues and eigenvectors of A , by solving the diagonalization of A ( A = U λ U ). Hence, the eigenvectors of W f can be spanned by U ^ = U ; B U λ 1 . Since the approximated eigenvectors U ^ are not orthogonal, as explained in [30], they can be obtained with S = A + A 1 2 B B A 1 2 . Then, by the diagonalization of S ( S = U s λ s U s ) the final approximated eigenvectors of W are given by:
U ^ = A B A 1 2 U s λ s 1 2 .
Up to this point we have discussed the approach proposed by authors in [4] in which, they used the common Gaussian kernel-based graph that might not provide the best representation of the intrinsic behavior of the data. In addition, the uniformly spaced distributed sampling method used to extract samples for the Nyström extension can overlook relevant samples that contain structural information about the change detection. Moreover, as defined by [31], the approximation of W f given by the Nyström extension is highly affected by the way samples are selected. Thus, we propose to use a recent graph-based sampling method named as Blue noise sampling [25] alongside a graph using prior smoothness learning, which is based on graph signal representation [24]. This is detailed in the following.

3.2.2. Graph Signal Processing: Smoothness Prior

The problem of learning a graph with a prior of smoothness, is to learn the Laplacian matrix ( L ), so that the signal variation on the resulting graph ( Q ( L ) ), is small. Here a small value of Q ( L ) , means that the signal on the graph takes similar values to its neighbors, resulting in edge disconnections from the graph [32]. The measure of smoothness of a signal x on a graph is given by:
Q ( L ) = 1 2 i , j w i j ( x ( i ) x ( j ) ) 2 ,
where w i j is the i j th entry of matrix W . To do so, we use the recent approach in [24] for large scale graphs with prior smoothness, where the authors leverage the desired graph sparsity to reduce computational cost. This model needs as inputs the number of neighbors K that are related to the edges per node, and the parameters α and β . Therefore, the minimization problem of [24] can be computed as:
min W R P × P tr ( W T Z ) α i log ( j w i j ) + β 2 W F 2 + c 2 W W 0 F 2 s . t . w i j = w j i 0 , i j , w i j = 0 , i = j
where Z is a pairwise distances matrix, α is a log prior constant ( > α > weights in W ), β is a W F 2 prior constant ( > β → less sparsity in W ), and the parameter c encourages adjacency matrices close to a pre-specified adjacency matrix W 0 . To reduce the complexity of Equation (11) the authors in [24] fix the parameters α = β = 1 and multiply the pairwise distances matrix ( Z ) by a parameter θ to make the sparsity analysis simpler. The parameter θ is computed by each column b of Z , and it has upper and lower bounds, as described by Equation (12).
θ l o w e r = j = 1 n 1 n K z K + 1 , j 2 b k , j z K + 1 , j θ u p p e r = j = 1 n 1 n K z K , j 2 b k , j z k , j
The process is summarized in Algorithm 2, as follows:
Algorithm 2: Graph learning with prior smoothness [24].
Input: Matrix of distances Z, K edges per node (sparsity level)
Output: Graph learned with prior smoothness W s
Step to compute θ
  Compute bounds of θ with Equation (12).
  Compute θ as a geometric mean between the bounds θ l o w e r and θ u p p e r .
Step to compute adjacency matrix W s .
  Compute W s from Equation (11) with Z = θ Z

3.2.3. Blue-Noise Sampling

Many methods to perform a sampling over graphs can be found in the literature [33] in order to find a suitable sampling set S V G . However, as explained in [34], is desirable to get samples from different structures or objects that are present in the image, that from the graph perspective this structures/objects can be capture by sub-graphs. Therefore, we will focus on the sampling method known as blue-noise sampling since it focus on extracting samples related to sub-graphs. This kind of sampling has been widely used in digital half-toning [35], but has recently been extended to graphs [25].
The blue-noise sampling gives as a result a subset S of vertices V that are as far as possible from each other in terms of the geodesic distances on G . The work in [25] demonstrated that the final subset given by blue-noise sampling leads to an accurate reconstruction of signals that is energy is mostly concentrated at the lowest eigenvectors of the graph Laplacian of G .
In order to assemble all the aforementioned methods into a single approach, we need a graph that contains the most relevant samples (i.e., pixels). To do this, we use a down-sampled version of the image taken from the GFKuts method [3] which contains probabilities of the pixels that belong or do not belong to the crop (see Figure 3). The down-sampling is conducted to avoid computer memory saturation, by applying a square grid M R m 25 × n 25 . Secondly, to learn a graph by using the Algorithm 2 of the prior of smoothness, we use, as nodes, the vectorized version ( M i R 625 ) of each square in the image, computing the distances between the nodes ( Z R | V | × | V | ) and apply the prior of smoothness [24]. Lastly, we apply the blue-noise sampling over the smoothed graph W s and inject these samples to the graph based fusion algorithm proposed in [4], in order to obtain the eigenvector with the highest eigenvalue of the fused-graph W f . Since the eigenvectors are embedded in a high-dimensional space equal to the resolution of the images (i.e., 960 × 1280 ), we use the same dimensionality reduction technique used in [4], based on the t distributed stochastic neighbor embedding (t-SNE) [36]. Figure 3 summarizes proposed method.

3.3. Non-Linear Regression Models

3.3.1. Support Vector Machine Regression (SVM-R)

The purpose of a support vector machine regression is to solve a linear regression problem by mapping the input data into a high dimensional space (feature space). The following model is considered:
f ( w e , x ) = ( w e ϕ ( x ) ) + b ,
where w e are the regression coefficients (weights), ϕ ( x ) : R i n R o u t with o u t i n , and b is the bias term. To obtain the optimal values of W e , a loss function can be defined such as the Laplacian, Huber’s, Gaussian or ϵ -intensive. We will use the most common loss function that is the ϵ -intensive with a regularization parameter C:
min W e , ξ , b 1 2 W e + C i = 1 n ξ i + ξ i
s . t . y i f ( w e , x ) b ϵ + ξ i f ( w e , x ) + b y i ϵ + ξ i ξ i , ξ i 0 ,
where ξ i and ξ i are slack variables with measure deviations larger than ϵ . By transforming the optimization problem in Equation (14) into a dual problem, the final solution is obtained by using Lagrange, the Karush–Kuhn–Tucker conditions, and the kernel trick [37], giving:
f ( x ) = α α K ( x ) + b ,
where α , α are the Lagrange multipliers and K ( x ) is a Kerrnel function. In this case we use the Radial Basis Function (RBF) Kernel (similar to Equation (7)).

3.3.2. A Nonlinear Autoregressive Exogenous (NARX)

An exogenous autoregressive model (NARX) aimed at the identification of non-linear systems involves both current and past values of the impulsive series that models the dynamics of the system. NARMAX (Nonlinear Autoregressive Moving Average with Exogenous Inputs) methods provide models that are transparent, and easily solve many problems such as contour errors in CNC machines, forecasting, network traffic, and prediction of daily solar radiation [38,39,40,41]. The NARMAX model is defined in Equation (16). Where y(k) is the output system, u(k) is the input system and e(k) is the noise sequence. n y , n u y n e are the maximum lags for the system output, input and noise respectively; F [ . ] refers to the function and d is a time delay usually set to d = 1 .
Y ( k ) = F [ y ( k 1 ) , , y ( k n y ) , u ( k d ) , , u ( k d n u ) , e ( k 1 ) , , e ( k n e ) ] + e ( k ) ]
In this work, the function F [ . ] has been implemented using artificial neural networks (ANN) with two hidden layers and one output layer.

4. Results and Discussion

4.1. Experimental Setup

In order to compare the effectiveness of the proposed feature extraction based on graphs, the results were compared with two approaches: (i) the former graph method introduced in [4], namely GBF, and (ii) the GFKuts approach [3]. We used the datasets and Ground-Truth reported in [3,4], containing 314 images for the vegetation stage, 82 for the reproductive, and 71 for the ripening. The captured images have a resolution of 960 × 1280 pixels, geo-referenced with the corresponding biomass measurements in grams (g) from the Ground-Truth (further information of the experimental protocol can be found at https://www.protocols.io/view/protocol-bjxskpne, accessed date: 24 June 2021), which was assembled as follows: 1 linear meter of the plants were cut from each plot of the crop and weighted to obtain the fresh biomass. Subsequently, the samples are placed inside an oven at 65 degrees Celsius for 4 days or until a constant weight is reached. This is known as the dry biomass.
Having the imagery dataset and the Ground-Truth, we trained two estimation models: classical SVM-R regression and a robust nonlinear autoregressive network with exogenous inputs (NARX), both accounting for 70 % of the dataset for the training, and the remaining 30 % for testing and validation. Lastly, the performance of the models was measured in terms of the root mean squared error ( R M S E ), the linear correlation (r), and the coefficient of determination ( R 2 ). Additionally, we conducted several experiments with the SVM regressor considering different kernel functions (i.e., linear, polynomial, and RBF), while the best performance was achieved by the RBF Kernel (similar to the work in [4]).
The algorithms were tested in a server with two Intel(R) Xeon(R) CPUs E5-2650 v4 @ 2.20 GHz, with 24 physical cores, 48 threads of processes, and 252 GB of RAM. Parameters are detailed in Table 1.

4.2. Biomass Estimation

Figure 4, Figure 5 and Figure 6 show the experimental results obtained from the 3 methods in the following order: (i) GFKuts, (ii) former graph method (GBF), and (iii) proposed graph method with blue-noise and prior smoothness (GBF-Sm-Bs). Numerical data containing the above-ground biomass estimation results for each model are presented in Table 2.
According to Table 2, the GKFkuts approach achieved the lowest R M S E = 129.490 g, followed by the proposed graph-based model with R M S E = 155.498 g. However, by applying the robust Narx regressor, the best performance was achieved by our proposal, with a R M S E = 45.358 g. Results were also correlated to the linear fit, presented from Figure 4, Figure 5 and Figure 6. There we can observe that the Narx regressor improves the fitting to a linear model, in which the proposed method achieved a linear correlation value of r = 0.995 .
Contrasting with the previous work from [3], the Narx model enabled a slight improvement of about 3 % over the former GFKuts method, due mainly to the configuration of the auto regressive inputs, and the use of vegetative indices with the soft mask generated by the GFKuts approach. Regarding the former graph-based method from [4] (GBF), results were improved by around 62.43 % ( 155.498 / 249.058 ) and 45.42 % ( 45.358 / 99.855 ) for the SVM and Narx regressors in terms of the R M S E metric, respectively. For the SVM regressor, the GBF and the GBF-Sm-Bs methods achieved a R M S E = 259.058 g and R M S E = 155.498 g respectively, as observed in Table 2, while for the Narx regressor, the GBF and the GBF-Sm-Bs methods achieved a R M S E = 99.855 g and R M S E = 45.358 g respectively, as presented in Table 2. This demonstrates the effectiveness of using a structured sampling method combined with graph learning, based on a prior of smoothness, for the extraction of relevant samples to be apply into Nyström extension.
Overall, the proposed GBF-Sm-Bs approach using Narx achieved an improvement in the R M S E metric of about 50 % , while also improving the biomass correlations reported from previous works. Part of the improvement achieved with the GBF-Sm-Bs method comes from extracting more relevant features than GFKuts, precisely 89 features. While GFKuts works with only 7 VI-features (per image) highly sensitive to biomass variations, the graph-based method is characterized by the eigenvectors with the highest eigenvalue associated with the fused-graph W f , yielding 89 features per image.

5. Conclusions

The proposed methods for above-ground biomass estimation, not only enabled the precise characterization of the biomass behavior of the rice crops through the entire phenological cycle, but also worked as both image segmentation and feature extraction techniques, by associating relevant features from the canopy. It is worth mentioning that most of the existing body of work in remote sensing methods for high-throughput biomass estimations based on multispectral imagery, requires dedicated photogrammetry methods for image correction, segmentation and feature extraction. In this study, both GFKuts and the GBF-Sm-Bs methods solved those stages by combining them into one single approach.
According to the findings reported in Table 2, the proposed GBF-Sm-Bs approach obtained an average biomass correlation of 0.995 with R 2 = 0.991 and R M S E = 45.358 g, increasing the precision in the estimation by around 45.42 % ( 45.358 / 99.855 ), compared to the GBF method, and about 52.27 % ( 45.358 / 86.769 ) compared to the GFKuts approach for the Narx regressor. This is a promising result towards GWAS gene characterization (Genome-wide Association Study), that requires larger amounts of precise and accurate phenotyping data for the association of gene functions with a specific trait. In this regard, future work is aimed at the inclusion of clustering approaches within the proposed algorithms, to enable the extraction of features according to several plant varieties (genotypes).
As future work, it would be interesting to explore the graph convolutional network, which combines the structural information captured by the graphs with the high level of abstraction given by neural networks. Additionally, a strong approach for segmentation (i.e., different for square windows over the image) could lead to improved results, since the segmented regions are used to generated a graph with smoothness prior.

Author Contributions

Conceptualization, D.A.J.-S., E.S.C., H.D.B.-R. and F.C.C.; methodology, H.D.B.-R., F.C.C., J.D.C., I.F.M., D.A.J.-S. and E.S.C.; software, D.A.J.-S. and E.S.C.; validation, D.A.J.-S., E.S.C., H.D.B.-R., F.C.C. and J.D.C.; formal analysis and investigation, H.D.B.-R., F.C.C., J.D.C., I.F.M., D.A.J.-S. and E.S.C.; software, D.A.J.-S. and E.S.C.; data curation, David Alejandro Jimenez-Sierra), E.S.C.; writing—original draft preparation, D.A.J.-S. and E.S.C.; writing—review and editing, J.D.C., H.D.B.-R., F.C.C. and I.F.M.; supervision, J.D.C., H.D.B.-R. and F.C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the OMICAS program: “Optimización Multiescala In-silico de Cultivos Agrícolas Sostenibles (Infraestructura y validación en Arroz y Caña de Azúcar)”, anchored at the Pontificia Universidad Javeriana in Cali and funded within the Colombian Scientific Ecosystem by The World Bank, the Colombian Ministry of Science, Technology and Innovation, the Colombian Ministry of Education, the Colombian Ministry of Industry and Tourism, and ICETEX, under grant ID: FP44842-217-2018 and OMICAS Award ID: 792-61187.

Data Availability Statement

Datasets supporting the experimental results presented in Figure 4, Figure 5 and Figure 6 are available at the Open Science Framework: https://osf.io/cde6h/?view_only=1c4e5e03b9a34d3b96736ad8ab1b2774 folder Raw Data—MDPI Sensors. The experimental protocol for crop monitoring is also available at https://www.protocols.io/view/protocol-bjxskpne (accessed on 24 June 2021).

Acknowledgments

The authors would like to thank all CIAT staff that supported the experiments conducted above the crops located at CIAT headquarters in Palmira, Valle del Cauca, Colombia; in particular to Eliel Petro, Camila Rebolledo, Yolima Ospina and Cecile Grenier for their support in upland and lowland trials.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ahmed, M.; Ahmad, S.; Ahmad, S. Climate Variability Impact on Rice Production: Adaptation and Mitigation Strategies. In Quantification of Climate Variability, Adaptation and Mitigation for Agricultural Sustainability; Springer: Cham, Switzerland, 2017; pp. 91–111. [Google Scholar]
  2. Alebele, Y.; Zhang, X.; Wang, W.; Yang, G.; Yao, X.; Zheng, H.; Zhu, Y.; Cao, W.; Cheng, T. Estimation of Canopy Biomass Components in Paddy Rice from Combined Optical and SAR Data Using Multi-Target Gaussian Regressor Stacking. Remote Sens. 2020, 12, 2564. [Google Scholar] [CrossRef]
  3. Colorado, J.D.; Calderon, F.; Mendez, D.; Petro, E.; Rojas, J.P.; Correa, E.S.; Mondragon, I.F.; Rebolledo, M.C.; Jaramillo-Botero, A. A novel NIR-image segmentation method for the precise estimation of above-ground biomass in rice crops. PLoS ONE 2020, 15, e0239591. [Google Scholar] [CrossRef]
  4. Jimenez-Sierra, D.A.; Benítez-Restrepo, H.D.; Vargas-Cardona, H.D.; Chanussot, J. Graph-Based Data Fusion Applied to: Change Detection and Biomass Estimation in Rice Crops. Remote Sens. 2020, 12, 2683. [Google Scholar] [CrossRef]
  5. Yue, J.; Feng, H.; Jin, X.; Yuan, H.; Li, Z.; Zhou, C.; Yang, G.; Tian, Q. A comparison of crop parameters estimation using images from UAV-mounted snapshot hyperspectral sensor and high-definition digital camera. Remote Sens. 2018, 10, 1138. [Google Scholar] [CrossRef] [Green Version]
  6. Yue, J.; Feng, H.; Yang, G.; Li, Z. A comparison of regression techniques for estimation of above-ground winter wheat biomass using near-surface spectroscopy. Remote Sens. 2018, 10, 66. [Google Scholar] [CrossRef] [Green Version]
  7. Xiao, Z.; Bao, Y.-X.; Lin, W.; Du, Z.-Z.; Qian, T.; Can, C. Hyperspectral Features of Rice Canopy and SPAD Values Estimation under the Stress of Rice Leaf Folder. Chin. J. Agrometeorol. 2020, 41, 173. [Google Scholar]
  8. Cheng, T.; Song, R.; Li, D.; Zhou, K.; Zheng, H.; Yao, X.; Tian, Y.; Cao, W.; Zhu, Y. Spectroscopic estimation of biomass in canopy components of paddy rice using dry matter and chlorophyll indices. Remote Sens. 2017, 9, 319. [Google Scholar] [CrossRef] [Green Version]
  9. Yang, X.; Jia, Z.; Yang, J.; Kasabov, N. Change Detection of Optical Remote Sensing Image Disturbed by Thin Cloud Using Wavelet Coefficient Substitution Algorithm. Sensors 2019, 19, 1972. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Li, J.; Wu, Z.; Hu, Z.; Li, Z.; Wang, Y.; Molinier, M. Deep Learning Based Thin Cloud Removal Fusing Vegetation Red Edge and Short Wave Infrared Spectral Information for Sentinel-2A Imagery. Remote Sens. 2021, 13, 157. [Google Scholar] [CrossRef]
  11. Gitelson, A.A.; Kaufman, Y.J.; Stark, R.; Rundquist, D. Novel algorithms for remote estimation of vegetation fraction. Remote Sens. Environ. 2002, 80, 76–87. [Google Scholar] [CrossRef] [Green Version]
  12. Lin, F.; Guo, S.; Tan, C.; Zhou, X.; Zhang, D. Identification of Rice Sheath Blight through Spectral Responses Using Hyperspectral Images. Sensors 2020, 20, 6243. [Google Scholar] [CrossRef]
  13. Harrell, D.; Tubana, B.; Walker, T.; Phillips, S. Estimating rice grain yield potential using normalized difference vegetation index. Agron. J. 2011, 103, 1717–1723. [Google Scholar] [CrossRef]
  14. Campos, J.; García-Ruíz, F.; Gil, E. Assessment of Vineyard Canopy Characteristics from Vigour Maps Obtained Using UAV and Satellite Imagery. Sensors 2021, 21, 2363. [Google Scholar] [CrossRef] [PubMed]
  15. Devia, C.A.; Rojas, J.P.; Petro, E.; Martinez, C.; Mondragon, I.F.; Patino, D.; Rebolledo, M.C.; Colorado, J. High-throughput biomass estimation in rice crops using UAV multispectral imagery. J. Intell. Robot. Syst. 2019, 96, 573–589. [Google Scholar] [CrossRef]
  16. Colorado, J.D.; Cera-Bornacelli, N.; Caldas, J.S.; Petro, E.; Rebolledo, M.C.; Cuellar, D.; Calderon, F.; Mondragon, I.F.; Jaramillo-Botero, A. Estimation of Nitrogen in Rice Crops from UAV-Captured Images. Remote Sens. 2020, 12, 3396. [Google Scholar] [CrossRef]
  17. Hernández-Vela, A.; Reyes, M.; Ponce, V.; Escalera, S. Grabcut-based human segmentation in video sequences. Sensors 2012, 12, 15376–15393. [Google Scholar] [CrossRef] [PubMed]
  18. Rother, C.; Kolmogorov, V.; Blake, A. “GrabCut” interactive foreground extraction using iterated graph cuts. ACM Trans. Graph. (TOG) 2004, 23, 309–314. [Google Scholar] [CrossRef]
  19. Mortensen, E.N.; Barrett, W.A. Intelligent scissors for image composition. In Proceedings of the 22nd Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 6–11 August 1995; pp. 191–198. [Google Scholar]
  20. Xiong, J.; Po, L.M.; Cheung, K.W.; Xian, P.; Zhao, Y.; Rehman, Y.A.U.; Zhang, Y. Edge-Sensitive Left Ventricle Segmentation Using Deep Reinforcement Learning. Sensors 2021, 21, 2375. [Google Scholar] [CrossRef]
  21. Liu, B.; Liu, Z.; Li, Y.; Zhang, T.; Zhang, Z. Iterative Min Cut Clustering Based on Graph Cuts. Sensors 2021, 21, 474. [Google Scholar] [CrossRef]
  22. Boykov, Y.Y.; Jolly, M.P. Interactive graph cuts for optimal boundary & region segmentation of objects in ND images. In Proceedings of the Eighth IEEE International Conference on Computer Vision (ICCV 2001), Vancouver, BC, Canada, 7–14 July 2001; Volume 1, pp. 105–112. [Google Scholar]
  23. Castro, W.; Marcato Junior, J.; Polidoro, C.; Osco, L.P.; Gonçalves, W.; Rodrigues, L.; Santos, M.; Jank, L.; Barrios, S.; Valle, C.; et al. Deep learning applied to phenotyping of biomass in forages with UAV-based RGB imagery. Sensors 2020, 20, 4802. [Google Scholar] [CrossRef]
  24. Kalofolias, V.; Perraudin, N. Large Scale Graph Learning From Smooth Signals. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  25. Parada-Mayorga, A.; Lau, D.L.; Giraldo, J.H.; Arce, G.R. Blue-noise sampling on graphs. IEEE Trans. Signal Inf. Process. Netw. 2019, 5, 554–569. [Google Scholar] [CrossRef] [Green Version]
  26. Shapiro, A. Monte Carlo sampling methods. In Handbooks in Operations Research and Management Science; Elsevier: Amsterdam, The Netherlands, 2003; Volume 10, pp. 353–425. [Google Scholar]
  27. Petschnigg, G.; Szeliski, R.; Agrawala, M.; Cohen, M.; Hoppe, H.; Toyama, K. Digital photography with flash and no-flash image pairs. ACM Trans. Graph. (TOG) 2004, 23, 664–672. [Google Scholar] [CrossRef]
  28. Correa, E.S.; Francisco Calderon, J.D.C. GFkuts: A novel multispectral image segmentation method applied to precision agriculture. In Proceedings of the Virtual Symposium in Plant Omics Sciences (OMICAS), Cali, Colombia, 23–27 November 2020. [Google Scholar]
  29. He, K.; Sun, J.; Tang, X. Guided image filtering. In Proceedings of the European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 1–14. [Google Scholar]
  30. Fowlkes, C.; Belongie, S.; Chung, F.; Malik, J. Spectral grouping using the Nystrom method. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 214–225. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Kumar, S.; Mohri, M.; Talwalkar, A. Sampling methods for the Nyström method. J. Mach. Learn. Res. 2012, 13, 981–1006. [Google Scholar]
  32. Dong, X.; Thanou, D.; Rabbat, M.; Frossard, P. Learning graphs from data: A signal representation perspective. IEEE Signal Process. Mag. 2019, 36, 44–63. [Google Scholar] [CrossRef] [Green Version]
  33. Tanaka, Y.; Eldar, Y.C.; Ortega, A.; Cheung, G. Sampling Signals on Graphs: From Theory to Applications. IEEE Signal Process. Mag. 2020, 37, 14–30. [Google Scholar] [CrossRef]
  34. Iyer, G.; Chanussot, J.; Bertozzi, A.L. A Graph-Based Approach for Data Fusion and Segmentation of Multimodal Images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4419–4429. [Google Scholar] [CrossRef]
  35. Lau, D.L.; Arce, G.R. Modern Digital Halftoning; CRC Press: Boca Raton, FL, USA, 2018; Volume 1. [Google Scholar]
  36. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  37. Laref, R.; Losson, E.; Sava, A.; Siadat, M. On the optimization of the support vector machine regression hyperparameters setting for gas sensors array applications. Chemom. Intell. Lab. Syst. 2019, 184, 22–27. [Google Scholar] [CrossRef]
  38. Huo, F.; Poo, A.N. Nonlinear autoregressive network with exogenous inputs based contour error reduction in CNC machines. Int. J. Mach. Tools Manuf. 2013, 67, 45–52. [Google Scholar] [CrossRef]
  39. Men, Z.; Yee, E.; Lien, F.S.; Yang, Z.; Liu, Y. Ensemble nonlinear autoregressive exogenous artificial neural networks for short-term wind speed and power forecasting. Int. Sch. Res. Not. 2014, 2014, 972580. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Haviluddin; Alfred, R. Performance of modeling time series using nonlinear autoregressive with eXogenous input (NARX) in the network traffic forecasting. In Proceedings of the 2015 International Conference on Science in Information Technology (ICSITech), Yogyakarta, Indonesia, 27–28 October 2015; pp. 164–168. [Google Scholar]
  41. Boussaada, Z.; Curea, O.; Remaci, A.; Camblong, H.; Mrabet Bellaaj, N. A nonlinear autoregressive exogenous (NARX) neural network model for the prediction of the daily direct solar radiation. Energies 2018, 11, 620. [Google Scholar] [CrossRef] [Green Version]
Figure 1. (a) UAV-driven remote sensing of above-ground biomass in rice crops. (b) Multispectral imagery. (c) Dataset amount & crop stages (d) First methodology—GFKuts. (e) Second methodology—Graph based. (f) Destructive biomass sampling. (g) Validation and correlation stage.
Figure 1. (a) UAV-driven remote sensing of above-ground biomass in rice crops. (b) Multispectral imagery. (c) Dataset amount & crop stages (d) First methodology—GFKuts. (e) Second methodology—Graph based. (f) Destructive biomass sampling. (g) Validation and correlation stage.
Sensors 21 04369 g001
Figure 2. GFKuts approach. (a) Preprocessing stage, (b) binary classification approach, (c) GMM modeling & optimization, and (d) filter stage.
Figure 2. GFKuts approach. (a) Preprocessing stage, (b) binary classification approach, (c) GMM modeling & optimization, and (d) filter stage.
Sensors 21 04369 g002
Figure 3. Proposed method based on the three stages: (a) Graph learning with prior smoothness, (b) blue-noise sampling to inject the samples to Nyström extension and (c) the fusion of the the multispectral images to extract features of the crop.
Figure 3. Proposed method based on the three stages: (a) Graph learning with prior smoothness, (b) blue-noise sampling to inject the samples to Nyström extension and (c) the fusion of the the multispectral images to extract features of the crop.
Sensors 21 04369 g003
Figure 4. Biomass estimation using the GFKuts approach.
Figure 4. Biomass estimation using the GFKuts approach.
Sensors 21 04369 g004
Figure 5. Biomass estimations using the graph-based approach with uniform sampling features (GBF).
Figure 5. Biomass estimations using the graph-based approach with uniform sampling features (GBF).
Sensors 21 04369 g005
Figure 6. Biomass estimations using the proposed graph-based approach with blue-noise and prior smoothness features (GBF-Sm-Bs).
Figure 6. Biomass estimations using the proposed graph-based approach with blue-noise and prior smoothness features (GBF-Sm-Bs).
Sensors 21 04369 g006
Table 1. Parameters used for the models GFKuts [3], GBF [4], and our proposal GBF-Sm-Bs model.
Table 1. Parameters used for the models GFKuts [3], GBF [4], and our proposal GBF-Sm-Bs model.
ModelParameters
GFKuts [3]k-neighbors = 2 window radius = 60 ϵ = 10 6 -
GBF [4]--t-SNE d i m = 16 n s = 100
GBF-Sm-Bsedges/node = | V | 2 window size = 25 t-SNE d i m = 89 n s = 100
Table 2. Performance in terms of R M S E , r and R 2 for the models: GFKuts [3], GBF [4], and our proposed GBF-Sm-Bs model with both Narx and SVM regressors.
Table 2. Performance in terms of R M S E , r and R 2 for the models: GFKuts [3], GBF [4], and our proposed GBF-Sm-Bs model with both Narx and SVM regressors.
Model/Regressor RMSE (in grams [g])r R 2
GFKuts [3]/SVM 129 . 490 0 . 963 0 . 929
GBF [4]/SVM 249.058 0.857 0.735
GBF-Sm-Bs/SVM 155.498 0.954 0.910
GFKuts [3]/Narx 86.769 0.983 0.968
GBF [4]/Narx 99.855 0.9856 0.971
GBF-Sm-Bs/Narx 45 . 358 0 . 995 0 . 991
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jimenez-Sierra, D.A.; Correa, E.S.; Benítez-Restrepo, H.D.; Calderon, F.C.; Mondragon, I.F.; Colorado, J.D. Novel Feature-Extraction Methods for the Estimation of Above-Ground Biomass in Rice Crops. Sensors 2021, 21, 4369. https://doi.org/10.3390/s21134369

AMA Style

Jimenez-Sierra DA, Correa ES, Benítez-Restrepo HD, Calderon FC, Mondragon IF, Colorado JD. Novel Feature-Extraction Methods for the Estimation of Above-Ground Biomass in Rice Crops. Sensors. 2021; 21(13):4369. https://doi.org/10.3390/s21134369

Chicago/Turabian Style

Jimenez-Sierra, David Alejandro, Edgar Steven Correa, Hernán Darío Benítez-Restrepo, Francisco Carlos Calderon, Ivan Fernando Mondragon, and Julian D. Colorado. 2021. "Novel Feature-Extraction Methods for the Estimation of Above-Ground Biomass in Rice Crops" Sensors 21, no. 13: 4369. https://doi.org/10.3390/s21134369

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop