Next Article in Journal
Compensation for Vanadium Oxide Temperature with Stereo Vision on Long-Wave Infrared Light Measurement
Previous Article in Journal
Innovative IntraValvular Impedance Sensing Applied to Biological Heart Valve Prostheses: Design and In Vitro Evaluation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection Method for Walnut Shell-Kernel Separation Accuracy Based on Near-Infrared Spectroscopy

1
School of Engineering, Anhui Agricultural University, Hefei 230036, China
2
School of Tea and Food Science, Anhui Agricultural University, Hefei 230036, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(21), 8301; https://doi.org/10.3390/s22218301
Submission received: 14 September 2022 / Revised: 19 October 2022 / Accepted: 25 October 2022 / Published: 29 October 2022
(This article belongs to the Section Optical Sensors)

Abstract

:
In this study, Near-infrared (NIR) spectroscopy was adopted for the collection of 1200 spectra of three types of walnut materials after breaking the shells. A detection model of the walnut shell-kernel separation accuracy was established. The preprocessing method of de-trending (DT) was adopted. A classification model based on a support vector machine (SVM) and an extreme learning machine (ELM) was established with the principal component factor as the input variable. The effect of the penalty value (C) and kernel width (g) on the SVM model was discussed. The selection criteria of the number of hidden layer nodes (L) in the ELM model were studied, and a genetic algorithm (GA) was used to optimize the input layer weight (W) and the hidden layer threshold value (B) of the ELM. The results revealed that the classification accuracy of SVM and ELM models for the shell, kernel, and chimera was 97.78% and 97.11%. The proposed method can serve as a reference for the detection of walnut shell-kernel separation accuracy.

1. Introduction

Walnuts are extremely nutritious nuts [1]. The preliminary processing of walnuts is mainly divided into the following steps: Green husk removal, walnut drying, walnut size classification, walnut shell-breaking, and walnut shell-kernel separation [2]. At present, some factories use manual sorting for the walnut shell-kernel separation, leading to increased labor costs and contamination of walnut kernels [3]. Effective, clean shell-kernel separation technology is critical for improving the quality and economic efficiency of the walnut industry. The traditional shell-kernel separation technology of nuts can be mainly divided into two categories-screened selection method [4] and the wind selection method [5]. The screened method uses the difference in particle size between the shell and the kernel to separate the material. The wind method is based on the different suspension speeds of the shell, kernel, and chimeras. Walnuts contain a Diaphragma juglandis Fructus (DJF) when compared with other nuts, which leads to the presence of shell-kernel chimera in the material after breaking the shell. Chimeras were when the kernel was sandwiched between the shell and the DJF (Figure 1). It increases the difficulty of walnut shell–kernel separation.
The NIR analysis method is simple, non-destructive, rapid, and accurate [6]. In recent years, NIR spectroscopy has been implemented in a wide range of applications for the analysis of agricultural products [7,8,9,10]. Few studies were conducted on the application of NIR spectrum instruments to walnuts, which were focused on the origin classification and composition detection of walnuts. In reference [11], partial least squares regression and support vector machine regression algorithms were used to predict the fat content of walnuts. In reference [12], fast prediction of antioxidant properties of walnut by NIR spectroscopy combined with multivariate analysis. There was no report on the problem of walnut shell kernel separation using NIR spectroscopy. Limited by its own experimental principle, NIR is not suitable for large-scale separation of walnut shell-kernel, but it can sample and detect the sorted materials to judge the separation accuracy of walnut shell–kernel. This provided a feasible scheme for on-line sampling and detection of walnut kernel separation accuracy.
In this study, NIR information from shells, kernels, and chimeras was analyzed. Spectral feature information was extracted by conducting principal component analysis (PCA). The effects of different preprocessing methods on the accuracy of the model were compared, and the parameter selection and optimization problems of the walnut shell-kernel separation effect detection model based on the support vector machine (SVM) and extreme learning machine (ELM) were investigated.

2. Materials and Methods

2.1. Materials and Instruments

The walnuts selected in the experiment were Anhui Ningguo Carya cathayensis Sarg with diameters of approximately 16–20 mm, clean surfaces, and no foreign matter. Carya cathayensis Sarg has a smaller volume and thicker DJF. More chimeras after the shell were broken compared with ordinary walnut. After the walnuts were broken using a professional shelling machine, 400 kernels, 400 shells, and 400 chimeras were selected as samples. All the samples were stored under ventilated, dry, and insect-free conditions. The spectrum acquisition instrument used was a FieldSpec4 ground object spectrometer (Analytica Spectra Devices., Inc, Boulder, CA 80301, USA). The spectrum data-processing software ViewSpecPro (Version 5.6, Analytica Spectra Devices., Inc, Boulder, CA 80301, USA), which was matched with the ASD spectrometer, was used to process the data.

2.2. Empirical Method

The NIR spectrum acquisition device built according to experimental requirements is shown in Figure 2. The device included a darkroom, a FieldSpec4 (with a spectrum acquisition range of 350–2500 nm and standard resolution of 1 nm), pistol handle, fiber optic cable, halogen lamp (50 W), and computer. The height and angle of the pistol handle were adjusted to avoid shadows and ensure that the sample was the only object to be measured in the field angle. The spectrum reference plate was collected at 30 min intervals to obtain the optimum signal-to-noise ratio of the spectrum data, and a black velvet cloth was used as the background in the darkroom.
The spectrum acquisition parameters were set as follows: the number of pre-average times of the white board was 10, the number of pre-average times of the sample was 10, and the number of average times of the dark current was 25. The spectrometer was warmed for 30 min before collection. The sample was placed at the center of the field, the optical fiber probe was set perpendicular to the sample, and the RS3 software was activated for spectrum collection. To reduce the influence of the storage time and environment temperature on the sample moisture, the spectrum collection of all the samples was completed within 24 h.

2.3. Spectra Data Preprocessing

The original spectra contain useful chemical information, significant background noise, and irrelevant information. It was necessary to remove the noise and uninformative variables from the spectra by using preprocessing methods. Multiplicative scatter correction (MSC), standard normal variate (SNV), de-trending (DT) were adopted to preprocess the original spectra of the three materials to eliminate interference. MSC was used to eliminate the scattering phenomenon caused by uneven particle distribution and particle size. SNV was used to eliminate the influence of solid particle size, surface scattering, and optical path transformation on diffuse reflection [13]. DT was used to eliminate the baseline drift of the spectra and was used after SNV. Four methods, MSC, SNV, DT, and SNV + DT, were chosen to preprocess the original spectra. The effects of the different methods on the accuracy of the models were compared.
The Kennard stone (KS) method was used to divide the data set. The ratio of the training set to the validation set was set at 5:3. For the three samples of shell, kernel, and chimera, 250 samples were extracted as the training set, and 150 samples were extracted as the validation set. There were a total of 750 samples in the training set and 450 samples in the validation set. The wavelength collection interval of the FieldSpec4 spectrometer was 350–2500 nm, the original training set was a matrix with dimensions of 750 × 2151, and the validation set was a matrix with dimensions of 450 × 2151. Principal component analysis (PCA) was adopted to reduce the dimensionality of the data [14]. PCA synthesizes high-dimensional variables into linear independent low-dimensional variables, which are called PCs. PCs explain the variance of the data matrix. These selected PCs can replace the original data when the cumulative explained variance reaches 95%.

2.4. Support Vector Machine Model

The underlying principle of the Support vector machine (SVM) is to separate various feature types in the feature space by searching for a specific hyperplane [15,16]. The objective of SVM is to minimize the cost function:
λ 2 ω 2 + [ 1 η i = 1 n max ( 0 , 1 y ( ω X + b ) ) ]
The first part of the cost function is referred to as the regularization term, and the second part is referred to as the loss term. Moreover, λ is removed from the regularization term and replaced with the penalty value C, which is multiplied by the loss term. The cost function is transformed into:
1 2 ω 2 + C [ 1 η i = 1 n max ( 0 , 1 y ( ω X + b ) ) ]
Adding kernel functions to traditional SVM models to solve non-linear problems. The RBF kernel function is widely used in SVM classification models. The expression of the RBF kernel function is:
K = ( x i , x ) = exp ( x i x 2 / 2 σ )
By replacing 1 / 2 σ with g, this formula can be transformed into:
K = ( x i , x ) = exp ( g x i x 2 / 2 σ )
Two parameters need to be selected in this model, including the penalty value C and the kernel width g [17]. Exploring the effect of C, g parameters on SVM models. Find options for the choice of C, g parameters.

2.5. Extreme Learning Machine Classification Model

The Extreme learning machine (ELM)does not use an algorithm based on the gradient descent method [18]. The output ELM of a single hidden layer feedforward neural network is:
f L ( x ) = i = 1 L β i h i ( x ) = H ( x ) β h i ( x ) = g ( ω i , b i , x )
g ( ω i , b i , x )   is the activation function. ω i is the weight of the input layer (W). b i is the threshold of the hidden layer (B). β is the output weight between the hidden layer and the output layer. Further, W and B are randomly generated and remain unchanged during the training process, and output layer node is determined. Therefore, only the activation function and the number of hidden layer nodes (L) need to be set in the ELM training process. There are many kinds of activation functions. In this experiment, sigmod function was used as the activation function of ELM model. Sigmod is easy to derive and suitable for feedforward neural network. The influence of different hidden layer nodes number on the accuracy of the model is discussed.
The random selection of W and B enables ELM to have the advantages of a rapid learning rate and high generalization performance. However, it also causes randomness in the model results. Exploring the role of L in improving this randomness. In addition to this, GA algorithm was used to select the best W and B to improve the stability of the model.

2.6. GA-ELM

The optimization of the values of W and B in the ELM model using the GA algorithm is divided into the following three steps: (1) Artificially set the initial parameters of the GA [19]. (2) Setting the fitness function. The predicted output error of the ELM on the sample is used as the fitness function to calculate the fitness value of individuals in the initial population. (3) The better individuals are optimized using selection, crossover, and mutation to obtain new populations. The operation ends when the maximum number of iterations is reached. Select the optimal population individuals and identify them as the optimal W and B. Complete the optimization of the ELM model.

2.7. Evaluation of Model Performance

Accuracy is used to evaluate model performance. The formula is:
A c c u r a c y = n m × 100 %
n is the number of predicted and actual values that are equal. m is the number of actual values.

3. Results and Discussion

3.1. Spectra Data Preprocessing

The NIR absorption band is mainly formed by the vibration of different groups of molecules in organic matter. The main components of walnut kernels are fat, protein, water, and sugar. The main components of walnut shells are lignin, cellulose, and hemicellulose. The main components of DJF are amino acids, proteins, and flavonoids. Figure 3 presents the original spectra of the samples. The blue, red, and black lines indicate the spectra of the kernel, shell, and chimera, respectively. The absorption peak intensities of the three sample types were significantly different. In the wavelength range of 750–2500 nm, the order of the overall relative reflectance values was as follows: kernel > shell > chimera. There are overlaps in the spectra of the three materials. The overlap is not conducive to model classification. Therefore, further preprocessing should be conducted.
MSC, SNV, DT, and SNV + DT were adopted to preprocess the original spectra of the three materials to eliminate interference. The classification model based on ELM was established for the preprocessed full-band spectra. Figure 4 presents the model classification results of the different preprocessing methods. The DT preprocessing results were superior. The accuracy of the training sets reached 75.07%. The accuracy of the validation sets reached 75.09%. The results of MSC preprocessing were slightly poorer, and the results of the SNV and SNV + DT preprocessing were the poorest. DT was used as a preprocessing method. Figure 5 shows the spectra after different preprocessing.

3.2. Principal Component Analysis Dimensionality Reduction

Dimension reduction of spectra data was carried out by using PCA. The results are shown in Table 1.
The variance contribution rates of the first six PCs were 47.07%, 36.92%, 7.04%, 4.09%, 1.75%, and 1.04%, respectively. Figure 6 displays the projection of the validation set in the PCA space built with the training set. Different types of samples tended to exhibit trends of separation, and samples of the same type exhibited significant aggregation. The cumulative contribution rate of the first PC (PC1) and second PC (PC2) was 83.99%, which was not enough to fully explain all the information of the original data. The cumulative contribution rates of the first six PC was 97.91%. The first six PCs were extracted for the classification model training [20].

3.3. Model Development: SVM

There were two parameters that needed to be selected in this model, including the penalty value C and the kernel width g [21]. Moreover, C results in a trade-off between the error classification samples and the simplicity of the classification surface. High values of C impose high penalties on errors and make the model more rigorous, thus increasing the bias and reducing variance. Low values of C result in a high tolerance to errors, which expands the boundary. However, the variance is increased. g could be considered as the reciprocal of the radius of influence of the samples selected by the SVM model, which affects the model complexity. This experiment explored the ideas for finding the optimal C and g parameters. The optimal C and g were also determined. The results are shown in Table 2.
In experiments 1–5, g was set as 0.01, and C was set as 1, 2, 3, 4, and 5, respectively. The accuracies of the training set were 98.27%, 99.47%, 99.73%, 100%, and 100%. As the value of C increased, the model exhibited a greater penalty for errors, the size of the boundary surface decreased, and the tolerance for errors decreased. For the training set to be accurately classified, a large C value was required. However, an excessively large C value posed a high risk of overfitting. The accuracies of the validation set decreased as the value of C increased. In experiments 5–10, with a C value of 5, the influence of the g value on the model was investigated. In particular, g was decreased from 0.01 to 0.00005, The accuracies of the validation set changed from 90.22% to 91.33%, 95.11%, 97.11%, 97.78%, and 97.56%. The accuracies of the training set decreased from 100% to 99.47%, 97.2%, 96.8%, 95.73%, and 95.07%. In this model, the optimal values of C and g were C = 5 and g = 0.0001. The accuracies of the training set and validation set were 95.73% and 97.78%.

3.4. Model Development: ELM

An ELM classification model with sigmod function as activation function was established. Experiments were carried out to explore the selection criteria of the number of hidden layer nodes (L). Given that the weight of the input layer and the threshold of the hidden layer of the ELM were randomly selected, the experimental results exhibited features of randomness. The average value extraction method was used to reduce the contingency of the model in multiple experiments. If the L is excessively large or small, the learning ability and complexity of the network structure are affected. At present, there are several empirical methods for selecting L, which lack theoretical support. The L obtained is, therefore, for reference only. Table 3 presents the training and validation results of the ELM model with different values of L.
In Experiment 1, according to Fangfagorman’s theorem, the relationship between L and the number of modes N is L = log2 N. The value of N was 750. Therefore, L = 10. The average accuracy of the training set was 85.69%. The average accuracy of the validation set was 82.37%. In Experiment 2, Kolmogorov highlighted that the L can be expressed as L = 2n + 1 (where n is the number of input layer nodes). Six principal components were extracted from the data using PCA dimension reduction. Therefore, L = 13 for n = 6. The average accuracy of the training set was 87.78%, and the average accuracy of the validation set was 87.11%. L was excessively small, and the network learning ability was insufficient. In Experiments 3–8, the accuracy of the training set increased as L increased. For L = 70, the average accuracy of the training set was 94.31%. However, the accuracy of the validation set increased first and then decreased as L increased. The highest accuracy of the validation set was achieved when the value of L was 40. The highest accuracy rate was 93.04%. The L of the ELM model was set to 40 to avoid overfitting. The average accuracy of the training set and validation set were 92.89% and 93.04%, respectively.
GA was adopted to optimize W and B of the ELM to create the optimal network [22]. The model comparison results of the GA-ELM and ELM are shown in Table 4. It can be seen from the results that the GA determined the optimal W and B under the current L and improved the identification accuracy of the ELM. The accuracy of the training set and validation set in the ELM model optimized by GA was 95.87% and 97.11% when L was 40.

4. Conclusions

In this study, the separation accuracy of walnut shell and kernel was detected by NIR spectroscopy. A shell kernel classification model was established based on an SVM and ELM. A comparison of the influences of different preprocessing methods on the accuracy of model recognition was presented. DT can effectively remove spectra changes caused by the baseline drift, and the obtained walnut shell kernel classification model was the best. Dimensionality reduction was achieved by conducting PCA. The first six PCs were extracted for the classification model training.
The method of choosing the C and g parameters in the SVM was explored. The accuracy of the training set can be improved by increasing C, but too large a value of C will lead to overfitting. The accuracy of the validating set can be improved by decreasing g. However, the decrease of g is limited. Too small g will reduce the accuracy of the training set and verification set. The selection criteria of L were explored in the ELM. Increasing L can improve model accuracy by increasing the learning capability of the network. Too large L will lead to overfitting and reduce the accuracy of the validation set. To improve the uncertainty caused by the random selection of W and B. The GA algorithm was used to select the best W and B.
Overall, the shell-kernel classification model based on SVM and ELM developed in this study has 97.78% and 97.11% accuracy for the validation set. It can effectively evaluate and detect the accuracy of the existing shell-kernel separation method.

Author Contributions

Conceptualization, M.A., C.C. and Z.W.; methodology, M.A., C.C. and Z.W.; software, M.A. and K.L.; validation, M.A. and Z.W.; formal analysis, M.A.; investigation, M.A.; data curation, M.A.; writing—original draft preparation, M.A.; writing—review and editing, M.A. and C.C.; supervision, M.A. and K.L.; funding acquisition, C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 52075003.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank Chengmao Cao and Zhengmin Wu for providing equipment and technical support during the experiments. We also thank other colleagues in the laboratory, such as Kun Luo and Shishun Wang, for their help during the experiments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cao, C.M.; Jiang, L.; Wu, C.Y. Design and Test on Hammerhead of Pecan Shell-breaking Machine. Trans. Chin. Soc. Agric. Mach. 2017, 48, 307–315. [Google Scholar]
  2. Liu, M.Z.; Li, C.H.; Cao, C.M. Walnut Fruit Processing Equipment: Academic Insights and Perspectives. Food Eng. Rev. 2021, 13, 822–857. [Google Scholar] [CrossRef]
  3. Zhu, Z.J.; Li, Z.X.; Yang, L.L. Research Progress on the Separation Technology of Walnut Shell-kernel. Food Ind. 2014, 35, 216–219. [Google Scholar]
  4. Turuhong, T.; Yang, H.; AyimunishaBaikere. Experimental Research on Shell and Kernel of Almonds. J. Agric. Mech. Res. 2018, 40, 167–171. [Google Scholar]
  5. Cao, C.M.; Luo, K.; Peng, M.L. Experiment on Winnowing Mechanism and Winnowing Performance of Hickory Material. Trans. Chin. Soc. Agric. Mach. 2019, 50, 105–112. [Google Scholar]
  6. Oliveira, M.M.; Cruz-Tirado, J.P.; Roque, J.V. Portable near-infrared spectroscopy for rapid authentication of adulterated paprika powder. J. Food Compos. Anal. 2020, 87, 103403. [Google Scholar] [CrossRef]
  7. Wang, Y.J.; Li, L.Q.; Liu, Y. Enhanced quality monitoring during black tea processing by the fusion of NIRS and computer vision. J. Food Eng. 2021, 304, 110599. [Google Scholar] [CrossRef]
  8. Guo, Z.M.; Barimah, A.O.; Shujat, A. Simultaneous quantification of active constituents and antioxidant capability of green tea using NIR spectroscopy coupled with swarm intelligence algorithm. LWT 2020, 129, 109510. [Google Scholar] [CrossRef]
  9. Huang, L.X.; Meng, L.W.; Zhu, N. A primary study on forecasting the days before decay of peach fruit using near-infrared spectroscopy and electronic nose techniques. Postharvest Biol. Technol. 2017, 133, 104–112. [Google Scholar] [CrossRef]
  10. Zhang, K.; Xu, Y.J.; Johnson, L. Development of near-infrared spectroscopy models for quantitative determination of cellulose and hemicellulose contents of big bluestem. Renew. Energy 2017, 109, 101–109. [Google Scholar] [CrossRef]
  11. Ma, W.Q.; Zhang, M.; Li, Y. Non-destructive detection of fat content in walnut kernel by near-infrared spectroscopy. Trans. Chin. Soc. Agric. Mach. 2019, 50, 381–386. [Google Scholar]
  12. Zareef, M.; Arslan, M.; Hassan, M.M. Application of benchtop NIR spectroscopy coupled with multivariate analysis for rapid prediction of antioxidant properties of walnut. Food Chem 2021, 359, 129928. [Google Scholar] [CrossRef]
  13. Li, L.Q.; Wang, Y.J.; Jin, S.S. Evaluation of black tea by using smartphone imaging coupled with micro-near-infrared spectrometer. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2021, 246, 118991. [Google Scholar] [CrossRef] [PubMed]
  14. Li, L.Q.; Jin, S.S.; Wang, Y.J. Potential of smartphone-coupled micro NIR spectroscopy for quality control of green tea. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2021, 247, 119096. [Google Scholar] [CrossRef] [PubMed]
  15. Li, L.Q.; Xie, S.M.; Ning, J.M. Evaluating green tea quality based on multisensor data fusion combining hyperspectral imaging and olfactory visualization systems. Sci. Food Agric. 2019, 99, 1787–1794. [Google Scholar] [CrossRef]
  16. Hao, Y.; Wang, Q.M.; Zhang, S.M. Study on Online Detection Method of “Yali” Pear Black Heart Disease Based on Vis-Near Infrared Spectroscopy and AdaBoost Integrated Model. Spectrosc. Spectr. Anal. 2021, 41, 2764–2769. [Google Scholar]
  17. Huang, F.R.; Song, H.; Guo, L. Detection of adulteration in Chinese honey using NIR and ATR-FTIR spectral data fusion. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2020, 235, 118297. [Google Scholar] [CrossRef]
  18. Luo, Y.J.; Zhu, H.; Li, X.H. Quantitative Analysis of Total Phenol Conten in Cabernet Sauvignon Grape Based on Near-Infrared Spectroscopy. Spectrosc. Spectr. Anal. 2021, 41, 2036–2042. [Google Scholar]
  19. Guo, Y.; Guo, J.X.; Shi, Y. Estimation of Leaf Moisture Content in Cantaloupe Canopy Based on SiPLS-CARS and GA-ELM. Spectrosc. Spectr. Anal. 2022, 42, 2565–2571. [Google Scholar]
  20. Quelal-Vásconez, M.A.; Lerma-García, M.J.; Pérez-Esteve, É. Fast detection of cocoa shell in cocoa powders by near infrared spectroscopy and multivariate analysis. Food Control 2019, 99, 68–72. [Google Scholar] [CrossRef]
  21. Zareef, M.; Chen, Q.S.; Hassan, M.M. An Overview on the Applications of Typical Non-linear Algorithms Coupled with NIR Spectroscopy in Food Analysis. Food Eng. Rev. 2020, 12, 173–190. [Google Scholar] [CrossRef]
  22. Wang, W.X.; Ma, B.X.; Luo, X.Z. Study on the Moisture Content of Dried Hami Big Jujubes by Near-infrared Spectroscopy Combined with Variable Preferred and GA-ELM Model. Spectrosc. Spectr. Anal. 2020, 40, 543–549. [Google Scholar]
Figure 1. The phenomenon of misclassification after walnut shell–kernel separation.
Figure 1. The phenomenon of misclassification after walnut shell–kernel separation.
Sensors 22 08301 g001
Figure 2. Spectrum acquisition test bench.
Figure 2. Spectrum acquisition test bench.
Sensors 22 08301 g002
Figure 3. Original spectra.
Figure 3. Original spectra.
Sensors 22 08301 g003
Figure 4. Model classification accuracy of different preprocessing methods.
Figure 4. Model classification accuracy of different preprocessing methods.
Sensors 22 08301 g004
Figure 5. Spectra after different preprocessing methods.
Figure 5. Spectra after different preprocessing methods.
Sensors 22 08301 g005
Figure 6. Projection of the validation set in the PCA.
Figure 6. Projection of the validation set in the PCA.
Sensors 22 08301 g006
Table 1. Results of Principal Component Analysis (PCA).
Table 1. Results of Principal Component Analysis (PCA).
PCEigenvalueVariance Contribution (%)Cumulative Contribution (%)
11012.4747.0747.07
2749.114236.9283.99
3151.427.0491.03
487.954.0995.12
537.551.7596.87
622.441.0497.91
Table 2. Results of SVM model with different C and g parameters.
Table 2. Results of SVM model with different C and g parameters.
ExperimentC and gTraining SetValidation Set
Accuracy (%)MisjudgmentAccuracy (%)Misjudgment
1C = 1 g = 0.0198.271394.2226
2C = 2 g = 0.0199.47493.3330
3C = 3 g = 0.0199.73292.2235
4C = 4 g = 0.01100091.3339
5C = 5 g = 0.01100090.2244
6C = 5 g = 0.00599.47491.3339
7C = 5 g = 0.00197.22195.1122
8C = 5 g = 0.000596.82497.1113
9C = 5 g = 0.000195.733297.7810
10C = 5 g = 0.0000595.073797.5611
Table 3. Results of ELM model with different values of L.
Table 3. Results of ELM model with different values of L.
ExperimentLDatasetFirst (%)Second (%)Third (%)Average (%)
110Training87.0782.8087.2085.69
Validation82.4479.7884.8982.37
213Training87.8787.3388.1387.78
Validation86.8986.8987.5687.11
320Training86.6087.0790.4089.02
Validation88.6785.1188.4487.41
430Training92.2791.8792.2792.13
Validation90.8989.3388.6789.63
540Training92.8093.6092.2792.89
Validation92.4494.8991.7893.04
650Training95.0793.0793.4793.87
Validation94.0092.0089.7891.93
760Training93.2094.4094.2793.96
Validation91.3393.7889.5691.56
870Training94.8094.1394.0094.31
Validation88.8991.1193.3391.11
Table 4. Results of comparison between the GA-ELM and ELM models.
Table 4. Results of comparison between the GA-ELM and ELM models.
LAccuracy of Training Set (%)Accuracy of Validation Set (%)
ELMGA-ELMELMGA-ELM
3092.1394.9389.6396.89
4092.8995.8793.0497.11
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

An, M.; Cao, C.; Wu, Z.; Luo, K. Detection Method for Walnut Shell-Kernel Separation Accuracy Based on Near-Infrared Spectroscopy. Sensors 2022, 22, 8301. https://doi.org/10.3390/s22218301

AMA Style

An M, Cao C, Wu Z, Luo K. Detection Method for Walnut Shell-Kernel Separation Accuracy Based on Near-Infrared Spectroscopy. Sensors. 2022; 22(21):8301. https://doi.org/10.3390/s22218301

Chicago/Turabian Style

An, Minhui, Chengmao Cao, Zhengmin Wu, and Kun Luo. 2022. "Detection Method for Walnut Shell-Kernel Separation Accuracy Based on Near-Infrared Spectroscopy" Sensors 22, no. 21: 8301. https://doi.org/10.3390/s22218301

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop