Next Article in Journal
Face Pose Alignment with Event Cameras
Next Article in Special Issue
Development and Application of a Vehicle-Mounted Soil Texture Detector
Previous Article in Journal
Autodyne Sensor Signals with Amplitude-Frequency Modulation of Radiation
Previous Article in Special Issue
Visual Guidance and Egg Collection Scheme for a Smart Poultry Robot for Free-Range Farms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparison of Soil Total Nitrogen Content Prediction Models Based on Vis-NIR Spectroscopy

1
Key Laboratory of Modern Precision Agriculture System Integration Research, Ministry of Education, China Agricultural University, Beijing 100083, China
2
Key Laboratory of Agricultural Informatization Standardization, Ministry of Agriculture and Rural Affairs, China Agricultural University, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(24), 7078; https://doi.org/10.3390/s20247078
Submission received: 29 October 2020 / Revised: 24 November 2020 / Accepted: 7 December 2020 / Published: 10 December 2020
(This article belongs to the Special Issue Sensing Technologies for Agricultural Automation and Robotics)

Abstract

:
Visible-near-infrared spectrum (Vis-NIR) spectroscopy technology is one of the most important methods for non-destructive and rapid detection of soil total nitrogen (STN) content. In order to find a practical way to build STN content prediction model, three conventional machine learning methods and one deep learning approach are investigated and their predictive performances are compared and analyzed by using a public dataset called LUCAS Soil (19,019 samples). The three conventional machine learning methods include ordinary least square estimation (OLSE), random forest (RF), and extreme learning machine (ELM), while for the deep learning method, three different structures of convolutional neural network (CNN) incorporated Inception module are constructed and investigated. In order to clarify effectiveness of different pre-treatments on predicting STN content, the three conventional machine learning methods are combined with four pre-processing approaches (including baseline correction, smoothing, dimensional reduction, and feature selection) are investigated, compared, and analyzed. The results indicate that the baseline-corrected and smoothed ELM model reaches practical precision (coefficient of determination (R2) = 0.89, root mean square error of prediction (RMSEP) = 1.60 g/kg, and residual prediction deviation (RPD) = 2.34). While among three different structured CNN models, the one with more 1 × 1 convolutions preforms better (R2 = 0.93; RMSEP = 0.95 g/kg; and RPD = 3.85 in optimal case). In addition, in order to evaluate the influence of data set characteristics on the model, the LUCAS data set was divided into different data subsets according to dataset size, organic carbon (OC) content and countries, and the results show that the deep learning method is more effective and practical than conventional machine learning methods and, on the premise of enough data samples, it can be used to build a robust STN content prediction model with high accuracy for the same type of soil with similar agricultural treatment.

1. Introduction

Soil total nitrogen (STN) has a significant impact on plant growth [1,2], and predicting STN content is vital for crop production as well as income generation for farmers. Spectroscopic techniques have the advantages of fast detection speed, low cost, and non-destructive [3], and it has been widely used in evaluation of crop yield [4,5], detection of fruit maturity [6,7,8], monitoring plant diseases and pests [9,10], and other areas of agricultural production. The spectral absorptions are associated with the stretching and bending of bonds forming the O-H, N-H, and C-H groups in the visible-near-infrared spectrum (Vis-NIRS) bands [11], so Vis-NIRS has been one of the most primary means of predicting STN content. Building predictive models is an essential link for identifying and predicting STN content by Vis-NIRS, the selection and application of models will directly affect the effectiveness of prediction [12,13].
Traditional modeling approaches to evaluate STN content often involve complicated data preprocessing since there exist a nonlinear relationship between STN and soil spectra. Among them, partial least square (PLS) and principal component regression (PCR) are the most common modelling approaches for quantitative spectroscopy analyses in soils [14,15]. Veum et al. [16] predicted total nitrogen (TN) using a large, regional dataset of in situ profile DRS spectra and compare the performance of traditional PLS analysis, PLS on external parameter orthogonalization (EPO) transformed spectra (PLS-EPO), PLS-EPO with the Bayesian Lasso (PLS-EPO-BL), and covariate-assisted PLS-EPO-BL models. The results shown the PLS-EPO transformation dramatically improved model performance relative to PLS alone, reducing RMSEP by 53% for TN. Support vector regression (SVR) is another common conventional machine learning modeling method in this case. Debanen et al. [17] selected SVR to predict soil properties from texturally homogeneous samples (loamy sand), the results shown that Vis-NIRS is suitable for the prediction of properties of texturally homogeneous samples. The best results were obtained for N content prediction using the full dataset with cross-validation (R2 = 0.81, root mean square error of cross-validation (RMSECV) = 0.01, RPD = 2.20) and with an independent validation dataset (R2 = 0.73, RMSEP = 0.03, RPD = 1.22).
In addition, with the development of nonlinear modeling methods and machine learning technology, random forest (RF) and neural network have achieved good results in STN content prediction. RF method shows great potential to deal with the nonlinear relationship between soil total nitrogen and spectrum. By establishing a RF model, Zhang et al. [18] determined the spatial distribution of STN and investigated the importance of different predictors. This work employed 21 predictors (including the original bands (O), normal spectral indices (S), red-edge indices (R), and environment variables (E)) to estimate the spatial distribution of the STN content. The results suggested that the performance of the prediction model could be improved by providing enough types of suitable predictors. Neural network has been becoming a popular method used in STN content prediction due to its good performance, including BP neural network [13] and extreme learning machine (ELM) [19]. Li et al. [20] used uninformative variable elimination (UVE) to extract characteristic wavelengths, and then ELM was applied to establish the calibration models with full spectra and characteristic wavelengths. They concluded that the UVE-ELM model performed better (R2 is 0.9408, RMSEP is 0.0075 and residual prediction deviation (RPD) is 2.97).
However, the above conventional methods are more suitable for smaller sample size, and usually a set of complicated data preprocessing and/or spectral transformation are required before modeling. Misuse of data preprocessing or spectral transformation may introduce artifacts of interferences or remove useful information patterns and result in worse model performance [21].
With the rise of deep learning which is based on the improvement and innovation of neural network technology, researchers are introducing deep learning method to predict soil properties. Convolutional neural networks (CNNs) is one of the most popular deep learning approach, and it can discover intricate structures in high-dimensional data but reducing the need from prior knowledge and human effort in feature engineering [22]. Zhang et al. [21] proposed a method called DeepSpectra which incorporated Inception module based on a CNN, and tested the performance on four open accessed visible and near infrared spectroscopic datasets (corn, tablet, wheat, and soil). The paper showed that DeepSpectra approach provided better results than conventional linear and nonlinear calibration approaches in most scenarios. Riess et al. [23] developed and implemented three one-dimensional CNN: LucasCNN, LucasResNet, and LucasCoordConv, to carry out soil texture classification. They evaluated the performance of the CNN approaches and compared them by testing the freely available LUCAS topsoil dataset. The results showed that the LucasCoordConv achieved the best performance regarding the average accuracy. We can infer from the above successful studies that deep learning modeling method can be used to establish STN content prediction model when a massive soil sample data is available.
From a practical point of view, different modeling approaches may be suitable for different application conditions and lead to different performances. In order to investigate which modeling approach can get higher accuracy without complicated spectral transformation and clarify model effectiveness based on different data characteristics, several mainstream conventional STN content modeling methods, along with deep learning modeling techniques, are implemented and compared in this paper, and their effectiveness are explored according to data feature. The objectives of this paper are: (1) establish reasonable STN prediction models based on Vis-NIRS using both conventional machine learning and deep learning methods with LUCAS topsoil dataset; (2) compare the performance of the built models to find which method suitable for predicting STN content with different conditions; and (3) evaluate the effectiveness and applicability of the models due to data characteristics. Based on the results, we believe that this paper can help the researchers and engineers to quickly build a robust and effective STN content prediction model for their own soil samples.

2. Materials and Methods

Figure 1 is the workflow for STN prediction models based on Vis-NIRS. Firstly, the raw data is read-in, and data screening is carried out to cleanse and segment the dataset. Secondly, the datasets are subjected to preprocessing in conventional machine learning approaches. Finally, through each modeling process, the best model can be obtained from each specific method according to performance evaluation. The following sections expand upon the workflow in greater detail.
All the models are implemented on the Python platform using Keras and Scikit-learn library. The process is run on a Linux server (version is Ubuntu 16.04) with 64 GB of RAM, and an Nvidia Geforce GTX 1080Ti graphics card with 11 GB of RAM.

2.1. Dataset Description

In this paper, we use the open accessed Land Use/Cover Area Frame Statistical Survey (LUCAS) Soil dataset. The project collected 19,036 samples of topsoil and determined their soil parameters and measured their Vis-NIR spectra after oven-dried. The published dataset consists of soil properties (clay, silt and sand content, coarse fragments, pH, organic carbon content, nitrogen, etc.) and Vis-NIRS absorbance. The spectra covered a range from 400 to 2499.5 nm with a resolution of 0.5 nm [23]. Table 1 shows the description of the LUCAS Soil dataset and it is available at: https://esdac.jrc.ec.europa.eu/content/lucas-2009-topsoil-data#tabs-0-description=0.

2.2. Data Preparation and Pre-Processing

2.2.1. Data Cleansing and Data Segmentation

Given the experimental environment changes and variability of sample origins, there exist various noise signals and artifacts derived from sample pretreatment, instrument, and experiment. To reduce noise, data cleansing was carried out and those samples with STN = 0 g/kg were removed. Accordingly, 19,019 samples were obtained as raw experimental data to be used to establish the prediction models.
In most cases, just two data sets are commonly needed to calibrate and validate model in conventional machine learning methods. In this paper, in order to maximize the utilization of reasonably distributed data to obtain the most reliable model, instead of random data division, the samples were sorted in descending order according to their STN content at first. Then a set of data was then extracted at regular intervals (every 5 samples) as the test subset (3804 samples), and the other samples as train subset (15,215 samples). These two data subsets were used in conventional methods-based modeling. However, when CNN modeling was investigated, the above train subset was randomly partitioned into two subsets: the validation subset (25% of the data, 3804 samples), and the train subset (75% of the data, 11,411 samples).

2.2.2. Baseline Correction and Smoothing

Preprocessing is usually required to enhance calibration model accuracy by reducing un-modeled variations such as instrumental and experimental artifacts [24,25]. Baseline correction and smoothing are the typical preprocessing approaches for Vis-NIRS analysis. In this paper, we selected asymmetric least squares baseline correction method (AsLS) and Savitzy–Golay smoothing method (SGS) to remove the noise from the raw spectral data.
Eliser et al. [26] proposed the AsLS approach, and the target function is presented in Equation (1).
z = a r g min z { i w i ( y i z i ) 2 + λ i ( Δ 2 z i ) 2 }
where z represents the fitting spectrum; y i is the raw data at wavelength of i nm; z i is the fitting information at wavelength of i nm, and w i is the weight at wavelength of i nm. λ is an adjustment coefficient, and Δ 2 z i is differential operation and presented in Equation (2).
Δ 2 z i = ( z i z i 1 ) ( z i 1 z i 2 ) = z i 2 z i 1 + z i 2
When y i is much higher than z i , let w i = p ; when y i equal to or less than z i , w i = 1 p . Here p is introduced as another adjustment coefficient. So, there are two free parameters to tune: λ and p. In general, p takes values in the range (0.001–0.1), and λ takes values in the range (102–109). p is set to 0.01 and λ is set to 109 in this paper.
SGS convolution smoothing algorithm is one of the most commonly used spectral smoothing methods. The algorithm uses mathematical methods to establish polynomial regression of local interval data and evaluate its value in the center of approximate window. It not only inherits the advantages of moving average algorithm for fast and simple denoising, but also can effectively retain the relative maximum, minimum, and changeable information of the signal. Compared to traditional filters, SGS is faster and can handle missing data quite well [27]. In this work, the window size is set to 9 and uses a third order polynomial.

2.2.3. Dimensionality Reduction

Conventional machine learning approaches attempt to assign weight to each feature or band (in the case of Vis-NIRS) based on how much information is contained in each feature or band. However, they do not perform well in high dimensional spectral data, so that dimensionality reduction is required. Principal components analysis (PCA) is one of the most popular and effective dimension reduction methods used to explore variation in complex datasets [28].
PCA can transform a high dimensional dataset into a lower dimensional orthogonal feature set while retaining maximum information from the original high dimension dataset. To this goal, the samples (xi) are projected with WTxi in new hyperplane (W), and the location of the projection point determine the information of the new samples. The optimization function (Target) is expressed as Equation (3).
T a r g e t =   max w     t r ( W T X X T W )
In the formula,
W T   W =   I
where I is the unit matrix, XXT is the covariance matrix. Eigen-decomposition is performed on the covariance matrix to have the eigenvectors and corresponding eigenvalues, and the eigenvalues are then sorted in descending order. The required numbers of eigenvalues are taken, and then form a new eigenvector. The new eigenvector is the solution of PCA. In this paper, the number of principal components set to 2–50.

2.2.4. Feature Selection

Feature selection is another way to reduce a high-dimensional feature space, and it ranks the features by a metric and eliminates all features that do not achieve an adequate score. Usually, feature selection methods are classified into three general groups: filter, wrapper, and embedded methods [29]. Because wrapper and embedded methods are related to prediction models, we instead chose the mutual information (MI) method to select feature, which falls under the category of the filter.
With the MI method, the feature subset can be obtained by calculating the mutual information between the features themselves and between the features and the class variables [30]. To measure the value of mutual information, information entropy is incorporated. Given a discrete random variable X, the entropy of X is denoted as H ( X ) (Equation (5))
H ( X ) = x X p ( x ) log 2 p ( x )
where p ( x ) is the prior probability of X. MI has different expressions, and this paper chooses the expression (Equation (6)) proposed by Ross [31].
M I ( X , Y ) = Ψ ( k ) Ψ ( n x + 1 ) + Ψ ( n y + 1 ) + Ψ ( N )
where Ψ(.) is the digamma function, X = N 1 i = 1 N x i , the parameter k represents the number of neighbors in the KNN (K-nearest neighbors) algorithm, n x and n y , respectively represent the sample data falling within the range of the KNN, and N is the number of samples.

2.3. STN Prediction Models Based on Conventional Machine Learning Approaches

2.3.1. Ordinary Least Square Estimation (OLSE) Regression

The linear regression is one of the most commonly used models, which is simple and easy to be built. Given a dataset of Vis-NIRS (marked as x = (x1; x2; …; xd)) with d wavelengths, the linear regression model of STN content can be presented as Equation (7).
f ( x ) = w T x + b
where w = (w1; w2; …; wd). The model is defined when w and b are obtained, and the most common approach to obtain w and b is to minimize the mean-squared error. The linear regression model was selected with the ordinary least square estimation (OLSE) in this paper.

2.3.2. RF Regression

Random forest is an ensemble algorithm based on learning a collection of decision trees [32], and it is characterized by its ease of modeling, fast to calculate, and low computational overhead. RF builds several randomized decision trees at first, each of which is trained to classify the STN content. Then by using the decision tree as base learner, RF randomly chooses a specific number of wavelengths at each node and finds the best split among these wavelengths. Finally, RF generates predictions by voting over a collection of randomized decision trees.

2.3.3. ELM Regression

Extreme learning machine is a kind of feed-forward neural network. Once the parameters (such as the weight between input layer and hidden layer, the threshold of the hidden layer) are determined, they do not change to adapt to new data. As a result, ELM performs fast-learning. In this paper, three-layer ELM was built, the number of hidden layer nodes ranged from 2 to 1000, and sine is selected as activation function in ELM.

2.4. STN Prediction Models Based on Convolutional Neural Networks

Different from conventional machine learning, a convolution layer is used for feature extraction in CNN, together with a pooling to compress the features. Figure 2 represents an example of the feature processing with convolution and max pooling, the convolutional layers labeled as Conv and max pooling layer labeled as Pooling.
The prediction effect of CNN can be affected by the structures of the networks. To evaluate the performance of different models, we built 3 STN predicted models incorporated Inception module [33], whose structures are shown in Figure 3. The major difference of these models lies in the numbers of 1 × 1 convolutions, which will impact the nonlinear fitting ability.
The input of each model is 4200-dimensional raw spectroscopic data and the output is the estimated STN content(s). The convolutional layers are labeled as Conv, flatten layer is labeled as Flatten, and fully connected layer is labeled as F1. The purple modules and yellow modules represent different sizes of general convolution to extract the feature. The green module represents 1 × 1 convolution to reduce the network parameters and enhance model fitting. The gray module represents max pooling (pool_size = 3) to compress the features. The first CNN model (Model 1) is the simplest one with introducing the minimum number of 1 × 1 convolution. The second CNN model (Model 2) and the third CNN model (Model 3) have the same number of 1 × 1 convolution, but model 3 is deeper than model 2. In this paper, zero-padding is used to retain the edge information, and the rectified linear unit (ReLU) is selected as the activation function for the convolutional layers and fully connected layer in each model. To search the local minimum of the objective function, we selected adam optimizer to train the model. Mean squared error (MSE) is adopted as the loss function, which is presented in Equation (8).
  L o s s =   i = 1 n ( y i y i ^ ) 2 n
where yi and y i ^ are measured values and predicted values, respectively. n is the number of samples in the training set, and i is the i-th sample. The detail hyperparameters set is shown in Table 2.

2.5. Model Evaluation

The model performance is evaluated by root mean squared error of prediction (RMSEP), the coefficient of determination (R2), and the residual prediction deviation (RPD), their calculations are shown in Equations (9)–(11).
R 2 ( y , y ^ ) = 1 i = 1 n ( y i y i ^ ) 2 i = 1 n ( y i y ¯ ) 2
R M S E P = i = 1 n ( y i y i ^ ) 2 n    
R P D =   s y R M S E P
s y = i = 1 N ( y i y ¯ ) 2 N  
where   y i and y i ^ are measured values and predicted values, respectively; n is the number of samples in the training set; and s y is the standard deviation of the observed values, which is calculated using Equation (12). In which,   y ¯ is the arithmetic mean of y i ; N is the number of samples in the test set. Note that in the Equation (9), R2 can vary from −∞ to 1 (i.e., it can yield negative values).

3. Results and Discussion

3.1. Impact of Preprocessing on Conventional Machine Learning Approaches

To study the effect of different pre-treatments on conventional machine learning approaches, different pre-treatments (including baseline correction (marked as Baseline), smoothing, dimensionality reduction (marked as Reduction), and feature selection (marked as Selection)) are adopted to pretreat the Vis-NIRS data. After preprocessing, the predicted models (including ordinary least square estimation (OLSE), random forest (RF), and extreme learning machine (ELM)) are built based on the new independent variable data, and the results are showed in Table 3.
By comparing the impact of different pretreatments to three modeling methods, we can conclude that baseline correction is optimal but MI feature selection performs worst. MI feature selection, as a preprocessing stage, is the method used to select a subset of relevant features for model construction. However, no subset can contain the full information of the raw spectral data. Thus, models built based on MI feature selection has the worst results among all models. PCA transforms the raw data into a lower dimensional orthogonal feature set while retaining maximum amount of information from the original high dimensional dataset. It preforms better than MI feature selection, but worse than the raw data without any pre-treatments. Baseline correction and smoothing filter the noise without reducing the information of raw data. Minor differences were observed between them with OLSE model, but baseline correction works better than smoothing with RF and ELM models. Both models perform better than that based on the raw data.
Different prediction models with the same pre-treatment result in different performance. To observe and evaluate the stability of different models, the coefficient of determination (R2) for three models are plotted in Figure 4. As we can see in Figure 4, ELM showed the best adaptation and OLSE is the worst, this is due to non-linearity of ELM.
To observe and evaluate the usability of different models, the RDP heatmap was plotted in Figure 5. If RPD is below 1.5, the model performance cannot be used for STN prediction due to poor performance. If it is between 1.5 and 1.8, the model needs to be improved to get better results. For values of RPD in the range of 1.8–2, the STN prediction model is considered to be practical, and if it is higher than 2.0, the model performance is considered to be very good [28].
For OLSE model, the RPD value is 1.62 with the raw data, but it decreases to 0.56 due to the MI feature selection. In this case, OLSE cannot be used for STN prediction. Baseline correction and PCA pretreatments can improve the availability of RF model, but the RPDs will be significantly reduced if MI feature selection involved. The best usability is obtained from the ELM. The RPD value is 2.34 with baseline correction, and can be used to prediction.

3.2. The Comparison among Three CNN Models and Conventional Models

To evaluate the model performance of different CNNs, the training process was repeated 25 times for each model. The results are shown in Table 4. Overall, CNNs based on raw data can obtain better results than conventional machine learning models. Model 3 provides the best prediction for STN (R2 is 0.93; RMSEP is 0.95 g/kg; and RPD is 3.85 in optimal case). The feature processing with convolution and pooling can retain the information of raw data, and it is not excised the features (or wavelengths). Meanwhile, nonlinear model (ELM) preforms better than linear model (OLSE), and the accuracy is improved by introducing nonlinear activation functions. So, among all the built models, the deep learning method shows encouraging support and the CNN model with deeper convolution layers illustrates the best performance among three different CNN structures built in this paper.

3.3. Impact of Data Properties on Prediction Models

Data properties, such as sample size, may have an impact on the predicted models. To study and reveal the impact, this paper reformed datasets with different data properties (Table 5) to analyze. What needs illustration is that these datasets are randomly sampled from raw dataset, in order to avoid the introduction of instrumental and experimental artifacts due to different collections.
According to the above modeling results, four models with different pre-treatments are selected to compare the performance on different data properties: baseline-corrected OLSE, baseline-corrected RF, baseline-corrected and smoothed ELM; and Model 3 of CNN. The performance of each model is shown in Figure 6. R2 is marked by the lines and RMSEP is marked by the histograms. Of which, Inception model is labeled as red, ELM model is labeled as blue, RF model is labeled yellow, and OLSE model is labeled purple.
The prediction accuracy of the models is poor with small sample data, and the accuracy improves as the amount of data increases. The effects of ELM and RF changed little on different datasets, but OLSE was fluctuated significantly. To evaluate the usability of different models based on different datasets, the RDP heatmap was plotted in Figure 7. Figure 7 shows that, as the amount of data increases, the value of RPD increases. The CNN incorporated Inception module performs better than the other models with different datasets.
In order to clarify the effectiveness and applicability of STN prediction models according to data characteristics, we separated the samples from different countries and made the new dataset (Table 6). Of these, France has the largest number of samples (2807 samples), and Luxembourg has the least number of samples (three samples). Based on the datasets of different countries, we compared the performance of different models (baseline-corrected OLSE, baseline-corrected RF, baseline-corrected and smoothed ELM, and Model 3 of CNN), the results are shown in Figure 8. The exceptional is that the results of the 15th dataset are set 0 due to the small number of samples.
The performance of the OLSE model is the worst among four models. RF model is predicted by dichotomy and presents some nonlinearity, but its performance needs to be improved. For the CNN incorporated Inception module, the accuracy depends on sample size and data distribution, and its credibility decreases as the sample size decreases. ELM model preforms the best in the four predicted models with the dataset divided based on different countries.
From a practical point of view, the CNN model is always trained by the large size of samples and it can be used in each country soil samples. So, we selected the CNN model with total samples in optimal case (R2 is 0.93; RMSEP is 0.95 g/kg; and RPD is 3.85) to predict the STN content with different datasets from different countries, and compared with ELM model (Figure 9). Overall, the CNN model preforms better than ELM, but it is apparently worse in smaller size.
According to the research [34], a close relationship exists between organic carbon (OC) and N levels in soil. The higher OC concentration, the greater the N concentration. Moreover, the C-to-N ratio is relatively stable across different soil types. In order to investigate reliability of different STN prediction model, we divided the LUCAS dataset into 24 soil groups according to different OC concentrations by K-means clustering algorithm (Table 7), and the modeling results are shown in Table 8.
According to Table 8, the accuracy of each prediction model becomes worse based on OC content. We do not think it is because no obvious relationship exists between OC and STN; instead, it is because different countries have different agricultural production and farming histories, so that their soil physical structures and properties differ from each other. So, it is unreasonable and unscientific to divide data by OC concentration alone, it has to be combined country with OC concentration in order to get a high STN prediction accuracy model. Another conclusion can be drawn is that, due to the poor performance of OLSE, it should be noted that the original soil spectra does not have a linear relationship with the STN content.

3.4. Discussion

Agricultural production [35,36,37], soil structure [38], soil properties [39], soil sample data characteristics [40], along with the methods of data preprocessing [16,41], all affect the STN concentration prediction modeling and model performance. Due to the results we achieved in this paper, the STN prediction models with good performance and generalization come from the data set with greater size and more evenly distributed within a country. Data size [22] and data distribution [42] have a major impact on the predictive performance. The prediction accuracy of the model is poor with small sample data size and/or unevenly data distributed, and the accuracy improves as the amount of data increase and/or evenly distributed. Since different countries have different farming histories, and the soil physical structure differ with each other too, so that more suitable models can be built for each country, even for each different farmland in the country. According to the clustering results in this paper, we can see that C-to-N ratio is relatively stable across different countries since the distribution of STN content is following with OC concentration. However, because the data segmentation breaks the national boundaries, the predicted models perform significantly worse. It indicates that different types of soil from different country, even if the carbon content is in the same range, a robust STN prediction is hard to be built. Therefore, the ideal dataset is that include enough same type of soil samples with similar physical structure, evenly distributed and accurate STN concentrations and accurate soil spectra data. When this kind of ideal dataset is not available, then deep learning modeling approach would be very helpful.
As for the modeling method of soil total nitrogen concentration, both conventional and deep learning approaches can achieve practical accuracy due to this study. Data preprocessing method is very important to conventional modeling approaches [43]. In terms of data preprocessing, baseline correction is optimal and MI feature selection performs worst according to the results of different pre-treatments on conventional machine learning approaches. The reason for this is because MI feature selection has removed the useful information, and results in worse model performance. For deep learning approach, network construction is critical to the model performance. In terms of CNN structure, the predictive accuracy of CNN increases with the increase of the number of convolution layers due to the nonlinear activation functions. By modeling approaches comparing results, the deep learning method, CNN, provides good performance and robust generalization for STN content modeling, and it can be used as a benchmark model for all soil type and all countries as long as there are enough training samples from different soil type and different countries.

4. Conclusions

STN content prediction based on Vis-NIRS is becoming more and more feasible, but the performance of prediction is affected by various factors, such as soil properties, modeling method, characteristics of datasets, etc. In order to find a practical way to build STN content prediction model, three conventional machine learning methods and one deep learning approach are investigated and their predictive performances are compared and analyzed. Based on the results, the following conclusions are achieved:
(1)
Misuse of preprocessing (such as feature selection) may introduce artifacts or remove useful patterns and result in worse model performance, but some preprocessing (such as baseline correction) may improve the results.
(2)
Different data characteristics show different impact on modeling approach, so the ideal dataset is that include enough same type of soil samples with similar physical structure, evenly distributed and accurate STN concentrations and accurate soil spectra.
(3)
The deep learning method, CNN, provides good performance and robust generalization for STN content modeling, and it can be used as a benchmark model for all soil type and all countries when there are enough training samples from different soil type and different countries. Otherwise, ELM is the best choice.

Author Contributions

Conceptualization, L.Z. and Y.W.; methodology, M.L. and R.J.; software, Y.W.; validation, Y.W. and M.W.; formal analysis, R.J.; investigation, Y.W. and L.Z.; resources, M.L. and R.J.; data curation, M.W.; writing—original draft preparation, Y.W.; writing—review and editing, L.Z.; visualization, Y.W.; supervision, L.Z.; project administration, M.L.; funding acquisition, M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key Research and Development Program, grant number 2019YFE0125500 and National Natural Science Funds, grant number 31471409.

Acknowledgments

We are thankful to Han Li’s team of China Agricultural University for help with providing the GPU servers.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, J.; Lü, S.; Zhang, Z.; Zhao, X.; Li, X.; Ning, P.; Liu, M. Environmentally friendly fertilizers: A review of materials used and their effects on the environment. Sci. Total. Environ. 2018, 829–839. [Google Scholar] [CrossRef] [PubMed]
  2. Otto, R.; Castro, S.A.D.Q.; Mariano, E.; Franco, H.C.J.; Trivelin, P.C.O. Nitrogen Use Efficiency for Sugarcane-Biofuel Production: What Is Next? BioEnergy Res. 2016, 9, 1272–1289. [Google Scholar] [CrossRef]
  3. Liu, Y.; Pu, H.; Sun, D.-W. Hyperspectral imaging technique for evaluating food quality and safety during various processes: A review of recent applications. Trends Food Sci. Technol. 2017, 69, 25–35. [Google Scholar] [CrossRef]
  4. Kwan, C. Remote Sensing Performance Enhancement in Hyperspectral Images. Sensors 2018, 18, 3598. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Iqbal, M.A.; Shen, Y.; Stricevic, R.; Pei, H.; Sun, H.; Amiri, E.; Penas, A.; Del Rio, S. Evaluation of the FAO AquaCrop model for winter wheat on the North China Plain under deficit irrigation from field experiment to regional yield simulation. Agric. Water Manag. 2014, 135, 61–72. [Google Scholar] [CrossRef]
  6. Jie, D.; Xie, L.; Fu, X.; Rao, X.; Ying, Y. Variable selection for partial least squares analysis of soluble solids content in watermelon using near-infrared diffuse transmission technique. J. Food Eng. 2013, 118, 387–392. [Google Scholar] [CrossRef]
  7. Tian, X.; Fan, S.; Huang, W.; Wang, Z.; Li, J. Detection of early decay on citrus using hyperspectral transmittance imaging technology coupled with principal component analysis and improved watershed segmentation algorithms. Postharvest Biol. Technol. 2020, 161, 111071. [Google Scholar] [CrossRef]
  8. Sun, X.; Xu, S.; Huazhong, L. Non-Destructive Identification and Estimation of Granulation in Honey Pomelo Using Visible and Near-Infrared Transmittance Spectroscopy Combined with Machine Vision Technology. Appl. Sci. 2020, 10, 5399. [Google Scholar] [CrossRef]
  9. Zhang, J.; Pu, R.; Yuan, L.; Huang, W.; Nie, C.; Yang, G. Integrating Remotely Sensed and Meteorological Observations to Forecast Wheat Powdery Mildew at a Regional Scale. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2014, 7, 4328–4339. [Google Scholar] [CrossRef]
  10. Zhang, J.; Huang, Y.; Pu, R.; Gonzalez-Moreno, P.; Yuan, L.; Wu, K.; Huang, W. Monitoring plant diseases and pests through remote sensing technology: A review. Comput. Electron. Agric. 2019, 165, 104943. [Google Scholar] [CrossRef]
  11. Clark, R.N.; King, T.V.V.; Klejwa, M.; Swayze, G.A.; Vergo, N. High spectral resolution reflectance spectroscopy of minerals. J. Geophys. Res. Space Phys. 1990, 95, 12653–12680. [Google Scholar] [CrossRef] [Green Version]
  12. Zheng, L.; Li, M.-Z.; Pan, L.; Sun, J.-Y.; Tang, N. Application of wavelet packet analysis in estimating soil parameters based on NIR spectra. Spectrosc. Spectr. Anal. 2009, 29, 1549–1552. [Google Scholar]
  13. Zheng, L.; Li, M.-Z.; Pan, L.; Sun, J.-Y.; Tang, N. Estimation of soil organic matter and soil total nitrogen based on NIR spectroscopy and BP neural network. Spectrosc. Spectr. Anal. 2008, 28, 1160–1164. [Google Scholar]
  14. Kooistra, L.; Wehrens, R.; Leuven, R.; Buydens, L. Possibilities of visible–near-infrared spectroscopy for the assessment of soil contamination in river floodplains. Anal. Chim. Acta 2001, 446, 97–105. [Google Scholar] [CrossRef]
  15. Vasques, G.; Grunwald, S.; Sickman, J. Comparison of multivariate methods for inferential modeling of soil carbon using visible/near-infrared spectra. Geoderma 2008, 146, 14–25. [Google Scholar] [CrossRef]
  16. Veum, K.; Parker, P.A.; Sudduth, K.A.; Holan, S.H. Predicting Profile Soil Properties with Reflectance Spectra via Bayesian Covariate-Assisted External Parameter Orthogonalization. Sensors 2018, 18, 3869. [Google Scholar] [CrossRef] [Green Version]
  17. Debaene, G.; Pikuła, D.; Niedźwiecki, J. Use of VIS-NIRS for land management classification with a support vector machine and prediction of soil organic carbon and other soil properties. Cienc. E Investig. Agrar. 2014, 41, 21–32. [Google Scholar] [CrossRef] [Green Version]
  18. Zhang, Y.; Sui, B.; Shen, H.; Ouyang, L. Mapping stocks of soil total nitrogen using remote sensing data: A comparison of random forest models with different predictors. Comput. Electron. Agric. 2019, 160, 23–30. [Google Scholar] [CrossRef]
  19. Zhou, P.; Yang, W.; Li, M.; Zheng, L.; Chen, Y. Soil Total nitrogen content pre- diction based on gray correlation-extreme learning machine. Trans. Chin. Soc. Agric. Mach. 2017, 48, 271–276. [Google Scholar] [CrossRef]
  20. Li, H.; Jia, S.; Le, Z. Quantitative Analysis of Soil Total Nitrogen Using Hyperspectral Imaging Technology with Extreme Learning Machine. Sensors 2019, 19, 4355. [Google Scholar] [CrossRef] [Green Version]
  21. Zhang, X.; Lin, T.; Xu, J.; Luo, X.; Ying, Y. DeepSpectra: An end-to-end deep learning approach for quantitative spectral analysis. Anal. Chim. Acta 2019, 1058, 48–57. [Google Scholar] [CrossRef] [PubMed]
  22. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  23. Riese, F.M.; Keller, S. Soil texture classification with 1D convolutional neural networks based on hyperspectral data. ISPRS Ann. Photogramm. Remote. Sens. Spat. Inf. Sci. 2019, 4, 615–621. [Google Scholar] [CrossRef] [Green Version]
  24. Gerretzen, J.; Szymanska, E.; Bart, J.; Davies, A.N.; Van Manen, H.-J.; Heuvel, E.R.V.D.; Jansen, J.J.; Van Kollenburg, G. Boosting model performance and interpretation by entangling preprocessing selection and variable selection. Anal. Chim. Acta 2016, 938, 44–52. [Google Scholar] [CrossRef] [PubMed]
  25. Bi, Y.; Yuan, K.; Xiao, W.; Wu, J.; Shi, C.; Xia, J.; Chu, G.; Zhang, G.; Zhou, G. A local pre-processing method for near-infrared spectra, combined with spectral segmentation and standard normal variate transformation. Anal. Chim. Acta 2016, 909, 30–40. [Google Scholar] [CrossRef] [PubMed]
  26. Eilers, P.H.C. A Perfect Smoother. Anal. Chem. 2003, 75, 3631–3636. [Google Scholar] [CrossRef]
  27. Pan, G. Local Regression and Likelihood. Technometrics 2000, 42, 428–429. [Google Scholar] [CrossRef]
  28. Morellos, A.; Pantazi, X.; Moshou, D.; Alexandridis, T.; Whetton, R.; Tziotzios, G.; Wiebensohn, J.; Bill, R.; Mouazen, A.M. Machine learning based prediction of soil total nitrogen, organic carbon and moisture content by using VIS-NIR spectroscopy. Biosyst. Eng. 2016, 152, 104–116. [Google Scholar] [CrossRef] [Green Version]
  29. Hoque, N.; Bhattacharyya, D.; Kalita, J. MIFS-ND: A mutual information-based feature selection method. Expert Syst. Appl. 2014, 41, 6371–6385. [Google Scholar] [CrossRef]
  30. Li, Z.; Zhou, X.; Dai, Z.; Zou, X. Classification of G-protein coupled receptors based on support vector machine with maximum relevance minimum redundancy and genetic algorithm. BMC Bioinform. 2010, 11, 325. [Google Scholar] [CrossRef] [Green Version]
  31. Ross, B.C. Mutual Information between Discrete and Continuous Data Sets. PLoS ONE 2014, 9, e87357. [Google Scholar] [CrossRef] [PubMed]
  32. Petralia, F.; Wang, P.; Yang, J.; Tu, Z. Integrative random forest for gene regulatory network inference. Bioinformatics 2015, 31, i197–i205. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  34. Fernández-Ugalde, O.; Ballabio, C.; Lugato, E.; Scarpa, S.; Jones, A. Assessment of Changes in Topsoil Properties in LUCAS Samples between 2009/2012 and 2015 Surveys; EUR 30147 EN; Publications Office of the European Union: Luxembourg, 2020; pp. 11–12. ISBN 978-92-76-17430-1. JRC120138. [Google Scholar] [CrossRef]
  35. Liu, Z.; Sun, K.; Liu, W.; Gao, T.; Li, G.; Han, H.; Li, Z.; Ning, T.-Y. Responses of soil carbon, nitrogen, and wheat and maize productivity to 10 years of decreased nitrogen fertilizer under contrasting tillage systems. Soil Tillage Res. 2020, 196, 104444. [Google Scholar] [CrossRef]
  36. Fiorini, A.; Boselli, R.; Maris, S.C.; Santelli, S.; Ardenti, F.; Capra, F.; Tabaglio, V. May conservation tillage enhance soil C and N accumulation without decreasing yield in intensive irrigated croplands? Results from an eight-year maize monoculture. Agric. Ecosyst. Environ. 2020, 296, 106926. [Google Scholar] [CrossRef]
  37. Boselli, R.; Fiorini, A.; Santelli, S.; Ardenti, F.; Capra, F.; Maris, S.C.; Tabaglio, V. Cover crops during transition to no-till maintain yield and enhance soil fertility in intensive agro-ecosystems. Field Crop. Res. 2020, 255, 107871. [Google Scholar] [CrossRef]
  38. Zhu, H.; Sun, R.; Bi, R.; Li, T.; Jing, Y.; Hu, W. Unraveling the local and structured variation of soil nutrients using two-dimensional empirical model decomposition in Fen River Watershed, China. Arch. Agron. Soil Sci. 2020, 66, 1556–1569. [Google Scholar] [CrossRef]
  39. Zhang, Z.; Sun, J. Root Features Determine the Increasing Proportion of Forbs in Response to Degradation in Alpine Steppe, Tibetan Plateau. Front. Environ. Sci. 2020, 8, 1–12. [Google Scholar] [CrossRef]
  40. Liu, Z.-P.; Shao, M.-A.; Wang, Y.-Q. Spatial patterns of soil total nitrogen and soil total phosphorus across the entire Loess Plateau region of China. Geoderma 2013, 197–198, 67–78. [Google Scholar] [CrossRef]
  41. Zhang, Y.; Li, M.; Zheng, L.; Zhao, Y.; Pei, X. Soil nitrogen content forecasting based on real-time NIR spectroscopy. Comput. Electron. Agric. 2016, 124, 29–36. [Google Scholar] [CrossRef]
  42. Fathololoumi, S.; Vaezi, A.R.; Alavipanah, S.K.; Ghorbani, A.; Biswas, A. Comparison of spectral and spatial-based approaches for mapping the local variation of soil moisture in a semi-arid mountainous area. Sci. Total. Environ. 2020, 724, 138319. [Google Scholar] [CrossRef]
  43. Engel, J.; Gerretzen, J.; Szymanska, E.; Jansen, J.J.; Downey, G.; Blanchet, L.; Buydens, L.M. Breaking with trends in pre-processing? TrAC Trends Anal. Chem. 2013, 50, 96–106. [Google Scholar] [CrossRef]
Figure 1. Workflow for soil total nitrogen (STN) prediction model based on visible-near-infrared spectrum (Vis-NIRS).
Figure 1. Workflow for soil total nitrogen (STN) prediction model based on visible-near-infrared spectrum (Vis-NIRS).
Sensors 20 07078 g001
Figure 2. Feature processing of the receptive spectra region by one convolutional layer and one max pooling layer. This is an example of two layers with the convolutional layer having a filter size of three and stride of three and the max pooling layer having a pooling size of three. A green rectangle in the input layer represents one of the input spectral variables, a rectangle in the Conv or Pooling layer represents a neuron. One neuron in the pooling layer covers a receptive field of 9 original spectral variables, respectively.
Figure 2. Feature processing of the receptive spectra region by one convolutional layer and one max pooling layer. This is an example of two layers with the convolutional layer having a filter size of three and stride of three and the max pooling layer having a pooling size of three. A green rectangle in the input layer represents one of the input spectral variables, a rectangle in the Conv or Pooling layer represents a neuron. One neuron in the pooling layer covers a receptive field of 9 original spectral variables, respectively.
Sensors 20 07078 g002
Figure 3. Architectures of three convolutional neural network (CNN) models incorporated Inception module. (a) Model 1. (b) Model 2. (c) Model 3.
Figure 3. Architectures of three convolutional neural network (CNN) models incorporated Inception module. (a) Model 1. (b) Model 2. (c) Model 3.
Sensors 20 07078 g003aSensors 20 07078 g003b
Figure 4. Stability of different predicted models with different preprocessing approaches.
Figure 4. Stability of different predicted models with different preprocessing approaches.
Sensors 20 07078 g004
Figure 5. Usability of different predicted models with different preprocessing approaches.
Figure 5. Usability of different predicted models with different preprocessing approaches.
Sensors 20 07078 g005
Figure 6. Performance of different predicted models with different size datasets. The models included baseline-corrected OLSE, baseline-corrected RF, baseline-corrected and smoothed ELM, and Model 3 of CNN. (For interpretation of the references to color in this figure legend, the reader is referred to the Web version of this article.).
Figure 6. Performance of different predicted models with different size datasets. The models included baseline-corrected OLSE, baseline-corrected RF, baseline-corrected and smoothed ELM, and Model 3 of CNN. (For interpretation of the references to color in this figure legend, the reader is referred to the Web version of this article.).
Sensors 20 07078 g006
Figure 7. Effects of different predicted models with different size datasets.
Figure 7. Effects of different predicted models with different size datasets.
Sensors 20 07078 g007
Figure 8. Effects of different predicted models with different datasets from different countries. (a) Accuracy of different predicted models. (b) Credibility of different predicted models.
Figure 8. Effects of different predicted models with different datasets from different countries. (a) Accuracy of different predicted models. (b) Credibility of different predicted models.
Sensors 20 07078 g008aSensors 20 07078 g008b
Figure 9. Effects of predicted with CNN and ELM.
Figure 9. Effects of predicted with CNN and ELM.
Sensors 20 07078 g009
Table 1. Description of the Land Use/Cover Area Frame Statistical Survey (LUCAS) Soil dataset.
Table 1. Description of the Land Use/Cover Area Frame Statistical Survey (LUCAS) Soil dataset.
Total Number of SamplesThe Number of CountriesMethod/StandardSTN (g/kg)
MinMaxMean
19,03623ISO 11261. 1995038.62.92
Table 2. Hyperparameters set in three CNNs.
Table 2. Hyperparameters set in three CNNs.
HyperparameterModel 1Model 2Model 3
Kernel size 1999
Kernel size 2555
Kernel size 3775
Kernel size 4------7
Stride 1333
Stride 2222
Hidden number323232
Batch size256256256
Dropout rate0.40.40.4
Regularization coefficient0.0010.0010.001
Learning rate0.010.010.01
Learning rate decay0.0010.0010.001
Table 3. The performance of three conventional models based on different pretreatment methods.
Table 3. The performance of three conventional models based on different pretreatment methods.
ModelExperiment No.BaselineSmoothingReductionSelectionR2RMSEP (g/kg)
OLSE1NoNoNoNo0.842.31
2YesNoNoNo0.842.30
3YesYesNoNo0.832.32
4YesNoYesNo0.753.56
5YesNoNoYes0.684.45
6NoYesNoNo0.842.31
7NoYesYesNo0.704.23
8NoYesNoYes0.526.69
9NoNoYesNo0.704.23
10NoNoNoYes0.526.71
11YesYesYesNo0.753.56
12YesYesNoYes0.684.46
RF1NoNoNoNo0.802.77
2YesNoNoNo0.832.37
3YesYesNoNo0.832.37
4YesNoYesNo0.822.51
5YesNoNoYes0.704.20
6NoYesNoNo0.802.78
7NoYesYesNo0.832.40
8NoYesNoYes0.674.64
9NoNoYesNo0.832.38
10NoNoNoYes0.674.63
11YesYesYesNo0.822.52
12YesYesNoYes0.704.20
ELM1NoNoNoNo0.822.49
2YesNoNoNo0.891.60
3YesYesNoNo0.891.60
4YesNoYesNo0.881.71
5YesNoNoYes0.842.25
6NoYesNoNo0.822.49
7NoYesYesNo0.793.00
8NoYesNoYes0.753.53
9NoNoYesNo0.793.00
10NoNoNoYes0.753.51
11YesYesYesNo0.881.71
12YesYesNoYes0.842.23
Note: Total samples of the dataset is 19,019, which the train dataset is 15,215 and the test dataset is 3804. ‘Yes’ indicates that this pre-treatment is used, and ‘No’ indicates this pre-treatment is not used.
Table 4. Predictive results obtained by three CNNs based on raw data.
Table 4. Predictive results obtained by three CNNs based on raw data.
Evaluation MetricModel 1Model 2Model 3
R2RMSEPR2RMSEPR2RMSEP
Mean0.901.400.891.530.930.99
Std0.0130.1820.0140.1930.0030.037
Max0.921.740.911.930.931.07
Min0.881.150.861.330.920.95
Table 5. Description of the datasets with different size.
Table 5. Description of the datasets with different size.
Dataset No.Total SamplesTraining Samples (Validation Samples)Test SamplesMaxMinStd
119031522 (381)38130.20.23.73
238013040 (760)76134.40.23.71
376096087 (1522)152234.20.23.74
495087606 (1902)190238.60.23.70
513,31410,651 (2663)266338.60.23.74
619,01915,215 (3804)380438.60.23.76
Table 6. Description of the datasets with different countries.
Table 6. Description of the datasets with different countries.
Dataset No.CountrySamplesMaxMinAverageStd
1AUSTRIA42117.20.53.112.21
2BELGIUM7112.40.92.892.54
3CZECH REPUBLIC40724.20.52.161.9
4GERMANY181936.20.22.482.6
5DENMARK22225.70.62.262.7
6ESTONIA21934.40.65.677.21
7SPAIN260112.30.21.571.29
8FINLAND166232.30.25.085.9
9FRANCE280723.10.22.351.55
10GREECE4879.10.21.591.14
11HUNGARY43412.60.21.931.31
12IRELAND21824.10.77.035.33
13ITALY117915.70.21.921.36
14LITHUANIA35532.80.42.964.76
15LUXEMBOURG33.12.12.670.42
16LATVIA34728.30.23.815.33
17NETHERLANDS20018.60.32.32.22
18POLAND161038.60.21.93.17
19PORTUGAL4757.60.21.671.1
20SWEDEN223234.20.25.045.91
21SLOVENIA11219.70.842.77
22SLOVAKIA26711.10.42.311.39
23UNITED KINGDOM87127.30.44.033.3
24TOTAL19,01938.60.22.933.76
Table 7. Description of the datasets with different organic carbon (OC) content.
Table 7. Description of the datasets with different organic carbon (OC) content.
Dataset No.SamplesOC (g/kg)STN (g/kg)
MaxMinAverageStdMaxMinAverageStd
176022.820.921.810.553.50.61.870.45
278278.558.367.15.728.11.24.261.52
379311.410.310.890.3120.41.120.23
484516.41515.690.43.40.41.470.33
583635.631.333.321.284.10.72.540.7
67546.204.491.472.40.20.620.24
7728586.8333.7449.0257.338.6617.535.5
888013.712.513.120.334.10.31.310.27
982527.925.226.490.795.40.72.150.56
1075719.317.818.520.442.80.61.660.37
1179448.24144.42.085.30.93.110.99
12773127.878.598.6714.3212.61.85.491.97
137409.388.660.391.80.30.950.21
1484231.327.929.540.953.70.82.330.62
158841513.714.360.382.10.41.370.28
1674520.819.320.070.442.70.41.750.4
17739333.7128.2212.2962.0125.82.39.613.9
1884712.511.411.960.3320.31.20.23
1979717.816.417.070.423.10.51.560.34
2075586.27.180.492.50.20.820.21
2178658.248.252.742.837.60.33.571.25
2281440.935.638.081.54.70.22.760.82
2379325.222.823.980.693.30.61.970.5
2475010.39.39.830.321.80.31.040.2
Table 8. Performance of different models with different datasets based on OC content.
Table 8. Performance of different models with different datasets based on OC content.
Dataset No.InceptionELMRFOLSE
R2RMSEPRPDR2RMSEPRPDR2RMSEPRPDR2RMSEPRPD
10.69 0.06 7.17 0.63 0.07 6.07 0.48 0.10 4.31 −0.07 0.22 2.08
20.63 0.87 1.76 0.71 0.69 2.23 0.44 1.31 1.17 0.35 1.52 1.01
30.28 0.04 6.09 0.53 0.02 9.42 0.44 0.03 7.86 −0.38 0.07 3.20
40.63 0.04 8.55 0.58 0.04 7.39 0.49 0.05 6.11 −0.47 0.15 2.12
50.32 0.34 2.08 0.65 0.18 4.01 0.60 0.20 3.54 0.33 0.34 2.11
60.09 0.05 4.80 0.28 0.04 6.05 0.15 0.04 5.15 −3.25 0.22 1.02
70.79 6.26 0.88 0.79 6.27 0.88 0.52 14.51 0.38 0.59 12.30 0.45
80.36 0.04 6.24 0.37 0.04 6.31 0.37 0.04 6.33 −1.40 0.15 1.65
90.57 0.13 4.18 0.55 0.14 3.97 0.59 0.13 4.33 −14.99 4.99 0.11
100.58 0.06 6.32 0.54 0.06 5.81 0.38 0.09 4.31 −1.12 0.30 1.26
110.78 0.22 4.51 0.75 0.24 4.09 0.60 0.39 2.55 0.28 0.70 1.41
120.73 1.08 1.84 0.65 1.38 1.43 0.39 2.38 0.83 0.06 3.69 0.54
130.45 0.02 8.62 0.44 0.02 8.58 0.44 0.02 8.59 −1.28 0.10 2.10
140.82 0.07 8.98 0.70 0.12 5.30 0.52 0.19 3.34 0.14 0.34 1.86
150.44 0.05 6.24 0.49 0.04 6.85 0.47 0.04 6.58 −0.49 0.12 2.35
160.09 0.15 2.75 0.63 0.06 6.68 0.52 0.08 5.18 −0.06 0.17 2.35
170.67 4.96 0.78 0.49 7.65 0.50 0.31 10.23 0.38 −0.36 20.33 0.19
180.29 0.04 5.95 0.56 0.02 9.52 0.50 0.03 8.40 −0.56 0.09 2.71
190.28 0.08 4.08 0.59 0.05 7.15 0.56 0.05 6.64 −0.19 0.14 2.46
200.29 0.03 7.23 0.37 0.02 8.19 0.05 0.04 5.42 −8.22 0.35 0.56
210.82 0.30 4.29 0.81 0.32 4.02 0.58 0.70 1.84 0.54 0.76 1.68
220.86 0.10 8.74 0.72 0.19 4.35 0.57 0.30 2.78 0.43 0.40 2.10
230.64 0.09 5.40 0.58 0.11 4.68 0.49 0.13 3.83 −7.92 2.30 0.22
240.19 0.03 6.06 0.54 0.02 10.59 0.51 0.02 9.93 −0.99 0.08 2.46
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, Y.; Li, M.; Ji, R.; Wang, M.; Zheng, L. Comparison of Soil Total Nitrogen Content Prediction Models Based on Vis-NIR Spectroscopy. Sensors 2020, 20, 7078. https://doi.org/10.3390/s20247078

AMA Style

Wang Y, Li M, Ji R, Wang M, Zheng L. Comparison of Soil Total Nitrogen Content Prediction Models Based on Vis-NIR Spectroscopy. Sensors. 2020; 20(24):7078. https://doi.org/10.3390/s20247078

Chicago/Turabian Style

Wang, Yueting, Minzan Li, Ronghua Ji, Minjuan Wang, and Lihua Zheng. 2020. "Comparison of Soil Total Nitrogen Content Prediction Models Based on Vis-NIR Spectroscopy" Sensors 20, no. 24: 7078. https://doi.org/10.3390/s20247078

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop