Next Article in Journal
Steady State Response Analysis of a Tubular Piezoelectric Print Head
Next Article in Special Issue
Building SDN-Based Agricultural Vehicular Sensor Networks Based on Extended Open vSwitch
Previous Article in Journal
Design Optimization for the Measurement Accuracy Improvement of a Large Range Nanopositioning Stage
Previous Article in Special Issue
Technological Advancement in Tower-Based Canopy Reflectance Monitoring: The AMSPEC-III System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Ensemble Successive Project Algorithm for Liquor Detection Using Near Infrared Sensor

1
College of Computer and Information Technology, Three Gorges University, Yichang 443002, China
2
Beijing Research Center for Agricultural Standards and Testing, Beijing 100097, China
*
Author to whom correspondence should be addressed.
Sensors 2016, 16(1), 89; https://doi.org/10.3390/s16010089
Submission received: 24 November 2015 / Revised: 27 December 2015 / Accepted: 7 January 2016 / Published: 11 January 2016
(This article belongs to the Special Issue Sensors for Agriculture)

Abstract

:
Spectral analysis technique based on near infrared (NIR) sensor is a powerful tool for complex information processing and high precision recognition, and it has been widely applied to quality analysis and online inspection of agricultural products. This paper proposes a new method to address the instability of small sample sizes in the successive projections algorithm (SPA) as well as the lack of association between selected variables and the analyte. The proposed method is an evaluated bootstrap ensemble SPA method (EBSPA) based on a variable evaluation index (EI) for variable selection, and is applied to the quantitative prediction of alcohol concentrations in liquor using NIR sensor. In the experiment, the proposed EBSPA with three kinds of modeling methods are established to test their performance. In addition, the proposed EBSPA combined with partial least square is compared with other state-of-the-art variable selection methods. The results show that the proposed method can solve the defects of SPA and it has the best generalization performance and stability. Furthermore, the physical meaning of the selected variables from the near infrared sensor data is clear, which can effectively reduce the variables and improve their prediction accuracy.

1. Introduction

With the development of the agriculture and agricultural products processing industry (such as food, beverage, feed, tobacco, etc.), attention is not only on the product yield, but also the quality and safety of agricultural products. Wherein, liquor is a kind of agricultural product that is usually made from grain, wheat and sorghum by cooking, saccharification, fermentation, and distillation, and it has huge economic benefits. However, in commerce, the contents of alcohol in liquor is not standard, which seriously damages the interests of the consumers. A series of methods have been proposed to detect the contents of alcohol in liquor, such as using carbon nanotubes acoustic and optical sensors [1], co-immobilized peroxidase and alcohol oxidase in carbon paste [2], and rhythm and formant features [3]. However, most of these are classical chemical methods, the analysis process is complex, and the analysis period is long. Therefore, a fast and accurate method to detect the contents of alcohol in liquor, based on a near infrared (NIR) sensor [4,5], is urgently needed.
With the development of sensor technology, sensor information processing technology has received more and more attention [6,7]. The spectra retrieved from the near infrared (NIR) sensors have the potential to extract chemical information about the composition of a sample [8,9]. It has been widely used for the qualitative and quantitative analysis of complex products in agriculture. Alcohol concentration is one of the main quality and technical indicators in the production and sales of liquor, which can be detected quickly and accurately using an NIR sensor. However, NIR spectroscopy from NIR sensor mainly concerns the molecular absorption of multiplication and combination frequencies [10]. It not only reflects the chemical composition and content of the tested material, but also contains a spectrum response that is caused by many factors such as the temperature, surface texture, density, and uneven distribution of internal components of the measured object [11,12]. As a result, spectral information overlaps and has a high degree of collinearity. Accordingly, redundant information needs to be excluded from the complex spectral information before the useful information is extracted to improve prediction accuracy and efficiency while simplifying the model [13].
The successive projections algorithm (SPA) is a forward selection method that uses vector projection analysis in a vector space to minimize variable collinearity [14,15]. It can effectively eliminate the effects of redundant variables, singularity, and instability while reducing the number of variables and complexity of the model [16]. Hence, it can increase model speed and efficiency [17,18]. In addition, SPA selects effective wavelength, which has more physical meaning than the full spectrum of partial least squares (PLS) because it selects the variables with minimum collinearity directly from the original variables [19]. In contrast, PLS uses latent variables to extract useful information from the spectral data [20]. A latent variable is a linear combination of the original variables that can reflect the information better, but its physical meaning is not clear [21]. The advantages of SPA mean that it has been widely used in spectral variable selection. However, SPA has two disadvantages: (1) If the sample size of the calibration set is small, the samples for modeling are unrepresentative. Although the variable collinearity of the calibration set is minimized, on the validation set, an inappropriate selection of variables could mean that the prediction results are not satisfactory [22,23,24]. (2) SPA is an unsupervised variable selection method, therefore, the selected variables do not necessarily reflect the information of the measured component well [25,26].
To overcome these problems, an ensemble SPA variable selection method (EBSPA) based on a new variable evaluation index (EI) is proposed in this paper. First, using the bagging ensemble strategy, the bootstrap method is used to resample with replacement [27]. The union set of variables selected in parallel by SPA on different sample sets, bootstrap ensemble SPA (BSPA), is used for ensemble modeling to solve the problem of model instability caused by variable selection on small sample sets. Second, a new EI that is associated with the analyte is proposed to evaluate the importance of variables. The EI is used to sort the importance of the variables in the BSPA. Finally, the cross-validation PLS method is used to select the best subset of variables from the sorted variables, thus ensuring that the ultimately selected variables not only have low autocorrelation, but also have a certain crosscorrelation with the analyte. In this paper, we use three methods to establish the models of variables that are selected by EBSPA: the EBSPA multiple linear regression model (EBSPA-MLR), EBSPA PLS regression model (EBSPA-PLS), and EBSPA least-squares support vector machine model (EBSPA-LS-SVM). An experimental comparison of the proposed methods with traditional SPA and five other state-of-the-art methods, the Forward interval PLS method (FiPLS) [28], Backward interval PLS method (BiPLS) [28], elimination of uninformative variables method (UVE) [29,30], Monte-Carlo UVE (MC-UVE) [31,32], and competitive adaptive reweighted sampling (CARS) [33,34] is presented. The results show that, combined with MLR, PLS, and LS-SVM, the proposed EBSPA method can effectively reduce the number of variables while increasing model accuracy.

2. Materials and Data

2.1. Sample Preparation

Liquor and deionized pure water were used to exactly formulate 162 samples of 2 mL each. The concentrations varied from 4.5% to 85.0% in intervals of 0.5%. The 162 samples were divided into two groups using the sample set partitioning based on joint x-y distances (SPXY) method [35] with a ratio of 2:1. Thus, there were 108 and 54 samples in the calibration and validation sets, respectively. The calibration set was used for training the samples, and the validation set was used for testing the samples. Table 1 shows the statistical results of the alcohol content in the samples. Note that the concentration range of the validation set was included in the concentration range of the calibration set. Thus, it is compliant with modeling standards.
Table 1. Descriptive statistics for sample measurements.
Table 1. Descriptive statistics for sample measurements.
DatasetNumber of SamplesConcentration Range (%)Mean Value (%)Standard Deviation
Calibration 1080.045–0.8500.4190.2425
Validation 540.075–0.8350.4680.2266

2.2. Spectral Acquisition from NIR Sensor

An infrared spectrometer produced by PerkinElmer, Inc. (Waltham, MA, USA) was used for the experiments, which installs multiple sensors (e.g., DTGS and MCT) and supports fast and reliable sensor switch. The wavenumber ranged from 12,000 to 4000 cm−1. A total of 32 scans with a resolution of 4 cm−1 and interval of 2 cm−1 were performed. Thus, each spectrum had 4001 variables. The experimental instruments also included a PC and a manual pipette (Eppendorf, Germany). The spectrometer software used to collect the spectral data was Spectrum Version 10.4.1. The indoor temperature was kept at about 25 °C, and the humidity remained basically unchanged (less than 60%). Each sample was collected three times in parallel, and the final spectrum of the sample is the average of these three samples. To ensure the consistency of the measurement environment and manual operations, the background was scanned every 10 samples to eliminate drift.

2.3. Spectral Preprocessing

Different spectral processing methods have a different impact on model performance. The following methods were considered to determine which was best for all 162 samples: raw spectra without processing (RAW), multiplicative scatter correction (MSC), standard normal variable transformation (SNV), SNV plus the trend method (SNV+DT), Savitzky-Golay smoothing convolution (SG), sliding window smoothing (SW), first-order derivative (1-Der), and second-order derivative (2-Der) spectra methods. Table 2 presents the results calculated by the PLS model. As can be seen, SNV produced the best performance, achieving an R value of 0.9521 and RMSECV of 0.0715.
Table 2. Modeling results of different processing methods.
Table 2. Modeling results of different processing methods.
MethodRAWMSCSNVSNV + DTSGSW1-Der2-Der
R0.89940.93250.95210.94440.89930.89910.95070.9512
RMSECV0.10200.08450.07150.07690.10200.10200.07530.0771
Figure 1A depicts the NIR spectra of different concentrations of liquor. It shows the maximum absorption peaks are at 5162 cm−1, which mainly reflects the O–H stretching vibration, bending vibration, and a combination of C–H bending vibration of the absorption band [36]. These characteristic peaks have been widely used for the quantitative analysis of the alcohol content in liquor [37]. Figure 1B shows the spectra from NIR sensor that have been processed by SNV. The spectral absorption peaks have increased and are more obvious, making them more conducive to spectral analysis. Therefore, SNV was selected as the final processing method for the comparative experiments in this study.
Figure 1. Spectra of samples: (A) RAW spectra and (B) SNV spectra.
Figure 1. Spectra of samples: (A) RAW spectra and (B) SNV spectra.
Sensors 16 00089 g001

3. The Proposed Method

3.1. SPA

For spectral matrix X N × M of the calibration set, where N is the number of samples, M is the number of variables, and H is the maximum number of the selected variables, the SPA algorithm is as follows:
  • Step 1: In the initial iteration t = 1 , a column vector x j is arbitrarily selected and denoted as x k ( 0 ) , where k ( 0 ) is the starting position of the first selected variable. The location of the remaining columns are defined as s , where s = { j ,     1 j M ,     j { k ( 0 ) , , k ( H 1 ) } } .
  • Step 2: Calculate the projection of the remaining column vectors x j ( j s ) with respect to the orthogonal vector space that consists of the selected vectors x k ( t 1 ) :
    { P = I x k ( t 1 ) ( x k ( t 1 ) ) T ( x k ( t 1 ) ) T x k ( t 1 ) x j = P x j
    where I is the identity matrix and P is the projection operator.
  • Step 3: Extract the variable that has the maximum projection value arg [ max ( P x j ) ] , ( j s ) , and add it to the set of selected variables.
  • Step 4: Let t = t + 1 . If t < H , return to Step 2 until t = H .
A crucial aspect of SPA is the selection of k ( 0 ) and H . Because there is collinearity between variables, the value of H generally cannot be too large. Otherwise, all of the projection values of the spectra will become zero [38]. For each selection of k ( 0 ) , a method such as MLR or PLS is used to conduct the cross-validation analysis. To obtain the minimum value of the standard error of cross validation (RMSECV), the corresponding k ( 0 ) and the actual number of selected variables h ( h H ) are the final optimal choice.

3.2. EI

To ensure that the selected variables have both lower autocorrelation and some cross-correlation with the analyte, a new EI w i is introduced in this paper to select the best subset of variables, and is defined as
w i = α i p i b i
where α i is the weight coefficient of the i -th variable. The order of the selected variables represents variable importance, which is sorted in descending order. The ordinal number of the variable corresponds to its weight, thus the more important variables have a larger weight. In the variable set that is obtained by multiple resampling, the weights of the same variable are summed when forming the union set. This further reflects the importance of recurring variables.
In addition, p i is the spectral purity value. It expresses the contribution of the i -th variable to the full spectrum. Larger values indicate greater contribution. Here, p i is defined as p i = σ i / μ i , where σ i is the standard deviation of the i -th variable and μ i is its mean value.
Further, b i is the absolute value of the regression coefficient of the i -th variable. For the multivariate calibration model y = X b + e , where y is the measured property vector, X is the spectral matrix, b is the regression coefficient vector, and e is the residual vector. The regression coefficient reflects the change of the spectral signal that is caused by the change of the unit concentration of the analyte. If b i is large, it indicates there is a good linear relationship between y and X .
Finally, w i combines the properties of α i , p i , and b i in Equation (2). This equation is a more comprehensive evaluation of the variables. Therefore, selecting the variables that have large w i will help improve the prediction accuracy of the model [39].

3.3. EBSPA

In the EBSPA method that is proposed in this paper, a bootstrap method is used to obtain T sample sets from the original training set. SPA is then used to select variables from these sample sets. The invalid variables are removed from each sample set to obtain T sets of the selected variables. The union of these T sets without duplicated variables is then obtained. A new EI is used to evaluate the variables of the union set, and these variables are sorted in order of their importance. Finally, the PLS cross-validation method is used to select the final variables for modeling. The details of EBSPA are as follows:
  • Step 1: Set the number of iterations to T . Use the bootstrap method to randomly select samples with replacement from the calibration set. In each iteration, the number of picked samples is the same as the size of the calibration set. The set of selected sample is S i , ( i = 1 , , T ) .
  • Step 2: Use SPA to select the variable subset F i from S i .
  • Step 3: Let i = i + 1 . If i < T , then return to Step 1 to continue until the end of the iterations.
  • Step 4: Take union set F i ( i = 1 , T ) of the T sets of the selected variables, remove the repeated variables, and obtain the ensemble set of variables F B .
  • Step 5: Calculate the importance of each variable in F B according to Equation (2), and arrange the EI values w in descending order.
  • Step 6: Use the PLS cross-validation method to successively accumulate the sorted variables starting with maximum w . When the minimum value of RMSECV is acquired, use the accumulated variables as the final selected set F E B for EBSPA modeling.
Figure 2 outlines the framework of EBSPA. In this paper, new sample sets are obtained by multiple resampling. The characteristics of the bootstrap approach show that, in the original calibration set, some of the samples may be repeated several times, while others may never be selected at all. The ensemble method can increase the difference of the models by bootstrap resampling [40]. This ensemble strategy can be used to enhance the accuracy of small sample sizes when coupled with the calculation power of modern computing hardware. Furthermore, for the ensemble set of the variables, a new EI is proposed in this paper. The final valid variables associated with the measured substance are selected according to this index.
Figure 2. Flowchart of the proposed method.
Figure 2. Flowchart of the proposed method.
Sensors 16 00089 g002

4. Experiments and Discussion

4.1. Parameter Selection

4.1.1. Maximum Number of Selected Variables

The main parameter of both SPA and EBSPA is the maximum number of the selected variables H . When H is large, the projected effect changes and the amount of calculation is increased. When H is too small, the information of the selected variables is insufficient and the model will have poor accuracy. With the iterations of EBSPA set to 10, the prediction performances of SPA and EBSPA for H values of 10, 15, 20, 25, and 30 are listed in Table 3.
Table 3 shows the prediction performance of EBSPA-MLR, SPA-MLR, EBSPA-PLS, and SPA-PLS. The results of EBSPA-MLR and EBSPA-PLS are better than those of SPA-MLR and SPA-PLS, which indicates that the proposed EBSPA method can effectively improve the prediction accuracy of the model. When H is 10, the results of all four methods are relatively poor. This may be caused by a lack of information of the selected variables. As a result, the models cannot achieve optimum results. When H is over 15, the results of SPA-PLS remain unchanged, and the final actual number of the selected variables h is 13. When H is over 20, the results of SPA-MLR are stable, and the final actual h is 17. Hence, the prediction performance of SPA does not continue to improve when H is greater than 20. On the contrary, it increases the computational cost of the model.
Additionally, the experiments also compared EBSPA-LS-SVM and SPA-LS-SVM for the five values of H . A radial basis function was selected as the kernel function of LS-SVM. A grid search combined with leave-one-out cross-validation was used to determine the regularization parameter γ and kernel parameter σ 2 [41]. With a training set of 108 samples for modeling, the optimal SPA-LS-SVM parameters ( γ , σ 2 ) were determined to be (5.278, 0.0039). Different values of H have no impact on EBSPA-LS-SVM and SPA-LS-SVM. The h , R, and standard error of prediction (RMSEP) of EBSPA-LS-SVM are 10, 0.9024, and 0.0882, respectively, regardless of the value of H . In addition, the h , R, and RMSEP of SPA-LS-SVM are 2, 0.8373, and 0.1123, respectively, for all values of H . Therefore, given the results of this comprehensive analysis, the parameter H was set to 20 in this study.
Table 3. Prediction performance for various H .
Table 3. Prediction performance for various H .
H EBSPA-MLRSPA-MLREBSPA-PLSSPA-PLS
R2RMSEPR2RMSEPR2RMSEPR2RMSEP
100.95870.05820.91830.08180.95480.06080.91540.0824
150.95990.05730.91290.08710.96110.05650.95420.0612
200.96140.05630.92690.07880.96540.05340.95420.0612
250.96250.05550.92690.07880.96710.05210.95420.0612
300.94450.06720.92690.07880.96770.05160.95420.0612
The maximum number of the selected variables is H , and R2 and RMSEP are the correlation coefficient and standard deviation of the validation set, respectively.

4.1.2. Number of Iterations

For BSPA and EBSPA, there is another parameter, the number of iterations T . If T is too large, it will increase redundant information and the calculation of the models. If T is too small, it can have an effect on the ensemble strategy. Figure 3 depicts the results obtained for an initial number of iterations from 1 to 15, where the maximum number of selected variables is 20.
Figure 3. Model results for different values of T : (A) selected numbers of variables; and (B) RMSEP values.
Figure 3. Model results for different values of T : (A) selected numbers of variables; and (B) RMSEP values.
Sensors 16 00089 g003
As can be seen in Figure 3A, as the number of iterations increases, the selected variables of BSPA increase. EBSPA uses EI to select variables once again that are based on BSPA. These variables increase then gradually stabilize. Figure 3B shows that the RMSEP of EBSPA decreases and then becomes stable. When T is small, the ensemble effect is not obvious. The poor results are caused by a lack of sufficient information for the EBSPA modeling. When T reaches a certain value, the variable information of EBSPA reaches saturation. The selected variables and prediction performance then remain stable. A comprehensive analysis of Figure 3, shows that the performance of EBSPA-MLR, EBSPA-PLS, and EBSPA-LS-SVM begin to stay stable as T reaches 9, 6, and 10, respectively. On the one hand, these results reflect the importance of EI. On the other hand, they show the stability of the proposed EBSPA method. In this paper, T is set to 10 in the following experiments, as it is a value that makes the model stable without too much additional computation.

4.2. Performance Analysis of Small Sample Sizes

To verify the effect of the ensemble strategy on the performance of SPA for small sample sizes, the SPXY method was used to select (27, 14), (54, 27), and (81, 41) samples from the original training and testing sets (108, 54) to form new sample sets. These three small sample sets are in line with modeling standards. Setting T = 10 and H = 20 with the ensemble method, the union set of the variables that were selected by SPA were used for BSPA modeling. The prediction performances of SPA and BSPA were compared for each of these four sample sizes.
Figure 4 shows the results of the model predictions for the four groups of samples. It can be seen that the model performance of SPA is not stable at small sample sizes. Among the methods, the SPA-LS-SVM method is the most sensitive to the number of samples. The small number of selected variables leads to a lack of useful information for modeling. Hence, when the sample size is small, the prediction accuracy is poor. However, as the number of samples increases, the accuracy significantly improves. The prediction performance of SPA-MLR is poor for the sample set (27, 14), but its accuracy is quite good for the other three sample sets. The performance of SPA-PLS is good and stable at all four different sample sizes, the model is more accurate than that of SPA-MLR and SPA-LS-SVM, especially at sample set (27, 14), which implies insufficient samples.
As can be seen in Figure 4, the accuracy of BSPA combined with PLS is lower than that of SPA-PLS for the sample set (108, 54). Because the number of variables selected by SPA-PLS is large, redundant information may exist after variable integration. In the rest of the cases, the accuracy of BSPA is higher than that of the corresponding SPA. BSPA-MLR is more stable among these different sample sizes and has the best performance. Especially, the prediction performance of BSPA-LS-SVM is better for the sample set (27, 14) than the other three sample sets. In general, BSPA is not sensitive to the size of samples. Even for small sample sizes, BSPA can still achieve higher prediction accuracy. Therefore, these results show that the method based on BSPA can increase the specificity of samples to a certain extent, which can help solve the problem of small sample sizes and improve the accuracy of the model.
Figure 4. Experimental comparison of small sample sizes: (A) correlation coefficient and (B) RMSEP.
Figure 4. Experimental comparison of small sample sizes: (A) correlation coefficient and (B) RMSEP.
Sensors 16 00089 g004

4.3. Performance Analysis of EI

EBSPA uses EI to evaluate the importance of the BSPA variables. The variables are sorted according to importance in descending order, and PLS cross validation is used to investigate the changes of RMSEP value with respect to the number of reserved variables. Finally, the minimum RMSEP is used to select the final variables of EBSPA. Table 4 lists the secondary selection results of EBSPA based on BSPA.
As Table 4 shows, for the four sample sets and three modeling methods, EBSPA can significantly reduce the number of variables in the final model by using the EI for the quadratic selection. On the one hand, it can effectively avoid the redundant information in BSPA that is caused by an excessive number of iterations. On the other hand, it can further improve the prediction accuracy of the model. The results show the feasibility and effectiveness of the proposed EI and further demonstrate the effectiveness of the ensemble variable selection for resolving the problem of small sample sizes. The use of EI can select the optimized variables that are important and relevant to the analyte, and it can compensate for the deficiency of SPA that arises from unsupervised variable selection.
Table 4. Comparison of EBSPA and BSPA.
Table 4. Comparison of EBSPA and BSPA.
(27, 14)(54, 27)(81, 41)(108, 54)
BSPA-MLRR20.96450.96820.95910.9592
RMSEP0.05990.05300.05940.0671
F B 117116107105
EBSPA-MLRR20.96870.97670.96060.9614
RMSEP0.05630.04550.05830.0563
F E B 50574087
BSPA-PLSR20.95880.97160.95180.9192
RMSEP0.06440.05010.06430.0806
F B 119839492
EBSPA-PLSR20.97040.97540.96650.9654
RMSEP0.05470.04670.05380.0534
F E B 20283638
BSPA-LS-SVMR20.91660.88830.88430.8972
RMSEP0.08180.09730.09790.0904
F B 15172221
EBSPA-LS-SVMR20.92040.95100.89850.9024
RMSEP0.08000.06330.09200.0882
F E B 13101110
Metrics R2 and RMSEP are the correlation coefficient and standard deviation of the validation set, respectively; F B and F E B are the number of variables in BSPA and EBSPA, respectively; and (m, n) denotes the sample set, where m and n are the number of samples of the calibration and validation sets, respectively.

4.4. Comparison of Different Spectral Variable Selection Methods

To verify the validity of the EBSPA method proposed in this paper, we compared it with FiPLS, BiPLS, UVE, MC-UVE, and CARS, all efficient spectral variable selection methods. FiPLS and BiPLS are interval variable selection methods based on a spectral segmentation of PLS. UVE and MC-UVE are variable selection methods based on a leave-one-out cross validation and Monte Carlo sampling of the PLS regression coefficients, respectively. CARS imitates the “survival of the fittest” principle in Darwin’s evolution theory, and introduces an exponential decay function to control the variable retention rate for variable selection. These methods are all based on the PLS method, therefore, EBSPA-PLS is used for comparison. In this experiment, the spectra of FiPLS and BiPLS are divided into 30 intervals. The cutoff threshold of UVE was set to 0.9. The number of samples for CARS was 500. The number of iterations of EBSPA-PLS and MC-UVE were 10. These methods are modeled on the original calibration set (108, 54). The variable selection results are shown in Figure 5.
Figure 5A,B show the selected variable intervals for FiPLS and BiPLS, respectively. In these methods, the empirical values of spectral segmentation are usually 20–40 sections. When the full spectrum that has a total of 4001 variables is divided into 30 sections, for the first 19 intervals, each has 133 variables, and for the last 11 intervals, each has 134 variables. FiPLS selects two intervals of 266 variables and BiPLS selects nine intervals of 1199 variables, respectively. Figure 5C expresses the stability coefficient of the variable for each wavelength of UVE. The dotted lines are cutoff lines that are determined by the added random numbers. The variables between the two cutoff lines are considered to be uninformative variables that need to be eliminated, and, ultimately, 214 variables were reserved. Figure 5D presents the change of RMSEP with respect to the reserved variables. The RMSEP values are calculated at every 10 variables, from 1 to 4001. When 1571 variables are retained, the minimum value of RMSEP is 0.0616. Figure 5E describes the variable selection process of CARS. In the first 418 sampling models, RMSECV presents a decreasing trend, which indicates that the eliminated variables are useless. RMSECV then starts to increase, and it may eliminate useful variables. The minimum RMSECV is obtained at the 418th sampling, where there are 29 final variables selected. Figure 5F shows the 38 variables that EBSPA-PLS finally selected.
Figure 5. Variable selection: (A) FiPLS; (B) BiPLS; (C) UVE; (D) MC-UVE; (E) CARS; and (F) EBSPA-PLS.
Figure 5. Variable selection: (A) FiPLS; (B) BiPLS; (C) UVE; (D) MC-UVE; (E) CARS; and (F) EBSPA-PLS.
Sensors 16 00089 g005
Table 5 shows the performance of PLS without variable selection and the six variable selection models. It can be seen that the method proposed in this paper has the highest accuracy and its number of selected variables is small. This can effectively reduce the redundant information of variables, simplify the model, and improve its prediction accuracy. Figure 6 plots the corresponding regression rates. We can see that the sample points of EBSPA-PLS are more concentrated and closer to the regression line, which indicates that the prediction performance is better.
Table 5. Comparison of model performances.
Table 5. Comparison of model performances.
MethodCalibration SetValidation SetVariable Numbers
R1RMSECR2RMSEP
PLS0.95620.07070.95530.06054001
FiPLS0.96960.05940.94400.0685266
BiPLS0.97110.05780.96070.06331199
UVE0.95660.07040.93630.0718214
MC-UVE0.95360.06140.95350.06161571
CARS0.95680.07020.94440.067329
EBSPA-PLS0.97340.05230.96540.053438
Metrics R1 and RMSEC are the correlation coefficient and standard deviation of the calibration set, respectively; and R2 and RMSEP are the correlation coefficient and standard deviation of the validation set, respectively.
Figure 6. Comparison of regression rates: (A) FiPLS; (B) BiPLS; (C) UVE; (D) MC-UVE; (E) CARS; and (F) EBSPA-PLS.
Figure 6. Comparison of regression rates: (A) FiPLS; (B) BiPLS; (C) UVE; (D) MC-UVE; (E) CARS; and (F) EBSPA-PLS.
Sensors 16 00089 g006

5. Conclusions

In this paper, the near infrared sensor has been implemented for obtaining the spectral data of liquor, and an ensemble successive project algorithm is proposed for variable selection and alcohol content detection in these liquor data. The proposed EBSPA can address two defects of SPA and improve the prediction accuracy of alcohol concentrations. The experimental results show that whether combined with linear MLR and PLS or nonlinear LS-SVM for modeling, the proposed EBSPA method can effectively address the disadvantages of SPA. It can simplify the model and improve prediction accuracy. In addition, when compared with FiPLS, BiPLS, UVE, MC-UVE, and CARS, the proposed EBSPA-PLS behaves better. Furthermore, it is shown that the use of NIR sensor and the proposed EBSPA can improve the performance of models with high prediction accuracy and stability, which can be applied for online and real-time detection of alcohol. It can also be effectively used to select variables and applied to NIR sensor data analysis in agriculture.

Acknowledgments

This work was supported in part by the National Science and Technology Projects in Rural Areas (2014BAD04B05), Natural Science Foundations of China (Grant No. 41371349), and National High Technology Research and Development Program of China (Grant No. 2013AA102302).

Author Contributions

The work presented here was carried out in collaboration between all authors. Fangfang Qu and Dong Ren and Jihua Wang conceived the idea. Fangfang Qu, Zhong Zhang, Na Lu and Lei Meng worked together on associated data collection and carried out the experimental work. Fangfang Qu drafted the manuscript. Dong Ren and Jihua Wang provided their experience and co-wrote the paper jointly with Fangfang Qu. All authors contributed, reviewed and improved the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Penza, M.; Cassano, G.; Aversa, P.; Antolini, F.; Cusano, A.; Cutolo, A.; Nicolais, L. Alcohol detection using carbon nanotubes acoustic and optical sensors. Appl. Phys. Lett. 2004, 85, 2379–2381. [Google Scholar] [CrossRef]
  2. Johansson, K.; Jönsson-Pettersson, G.; Gorton, L.; Marko-Varga, G.; Csöregi, E. A reagentless amperometric biosensor for alcohol detection in column liquid chromatography based on co-immobilized peroxidase and alcohol oxidase in carbon paste. J. Biotechnol. 1993, 31, 301–316. [Google Scholar] [CrossRef]
  3. Schiel, F.; Heinrich, C.; Neumeyer, V. Rhythm and formant features for automatic alcohol detection. In Proceedings of the INTERSPEECH 2010—11th Annual Conference of the International Speech Communication Association, Makuhari, Japan, 26–30 September 2010; pp. 458–461.
  4. Ridder, T.D.; Hendee, S.P.; Brown, C.D. Noninvasive alcohol testing using diffuse reflectance near-infrared spectroscopy. Appl. Spectrosc. 2005, 59, 181–189. [Google Scholar] [CrossRef] [PubMed]
  5. Castritius, S.; Kron, A.; Schäfer, T.; Rädle, M.; Harms, D. Determination of alcohol and extract concentration in beer samples using a combined method of near-infrared (NIR) spectroscopy and refractometry. J. Agric. Food. Chem. 2010, 58, 12634–12641. [Google Scholar] [CrossRef] [PubMed]
  6. Kim, S. Sea-Based Infrared Scene Interpretation by Background Type Classification and Coastal Region Detection for Small Target Detection. Sensors 2015, 15, 24487–24513. [Google Scholar] [CrossRef] [PubMed]
  7. Lim, J.; Kim, G.; Mo, C.; Kim, M.S. Design and Fabrication of a Real-Time Measurement System for the Capsaicinoid Content of Korean Red Pepper (Capsicum annuum L.) Powder by Visible and Near-Infrared Spectroscopy. Sensors 2015, 15, 27420–27435. [Google Scholar] [CrossRef] [PubMed]
  8. Sinelli, N.; Casiraghi, E.; Barzaghi, S.; Brambilla, A.; Giovanelli, G. Near infrared (NIR) spectroscopy as a tool for monitoring blueberry osmo-air dehydration process. Food. Res. Int. 2011, 44, 1427–1433. [Google Scholar] [CrossRef]
  9. Faassen, S.M.; Hitzmann, B. Fluorescence Spectroscopy and Chemometric Modeling for Bioprocess Monitoring. Sensors 2015, 15, 10271–10291. [Google Scholar] [CrossRef] [PubMed]
  10. Balabin, R.M.; Smirnov, S.V. Variable selection in near-infrared spectroscopy: Benchmarking of feature selection methods on biodiesel data. Anal. Chim. Acta. 2011, 692, 63–72. [Google Scholar] [CrossRef] [PubMed]
  11. Yong, H.; Xudong, S.; Hao, W. Spectral quantitative model optimization by modified successive projection algorithm. J. Jiangsu Univ. 2013, 34, 49–53. [Google Scholar]
  12. Guo, Z.M.; Huang, W.Q.; Peng, Y.K.; Wang, X.; Tang, X.Y. Adaptive Ant Colony Optimization Approach to Characteristic Wavelength Selection of NIR Spectroscopy. Chin. J. Anal. Chem. 2014, 42, 513–518. [Google Scholar]
  13. Mehmood, T.; Liland, K.H.; Snipen, L.; Sæbø, S. A review of variable selection methods in partial least squares regression. Chemometr. Intell. Lab. 2012, 118, 62–69. [Google Scholar] [CrossRef]
  14. Araújo, M.C.U.; Saldanha, T.C.B.; Galvão, R.K.H.; Yoneyama, T.; Chame, H.C.; Visani, V. The successive projections algorithm for variable selection in spectroscopic multicomponent analysis. Chemometr. Intell. Lab. 2001, 57, 65–73. [Google Scholar] [CrossRef]
  15. Du, G.; Cai, W.; Shao, X. A variable differential consensus method for improving the quantitative near-infrared spectroscopic analysis. Sci. China Chem. 2012, 55, 1946–1952. [Google Scholar] [CrossRef]
  16. Wu, D.; Ning, J.F.; Liu, X. Determination of anthocyanin content in grape skins using hyperspectral imaging technique and successive projections algorithm. Food Sci. 2014, 35, 57–61. [Google Scholar]
  17. Diniz, P.H.G.D.; Gomes, A.A.; Pistonesi, M.F.; Band, B.S.F.; de Araújo, M.C.U. Simultaneous Classification of Teas According to Their Varieties and Geographical Origins by Using NIR Spectroscopy and SPA-LDA. Food Anal. Methods 2014, 7, 1712–1718. [Google Scholar] [CrossRef]
  18. Zou, X.B.; Zhao, J.W.; Povey, M.J.; Holmes, M.; Mao, H. Variables selection methods in near-infrared spectroscopy. Anal. Chim. Acta 2010, 667, 14–32. [Google Scholar]
  19. Hong, Y.; Hong, T.; Dai, F.; Zhang, K.; Chen, H.; Li, Y. Successive projections algorithm for variable selection in nondestructive measurement of citrus total acidity. Trans. CSAE 2010, 26, 380–384. [Google Scholar]
  20. Liu, F.; He, Y. Application of successive projections algorithm for variable selection to determine organic acids of plum vinegar. Food Chem. 2009, 115, 1430–1436. [Google Scholar] [CrossRef]
  21. Wu, D.; Chen, X.; Zhu, X.; Guan, X.; Wu, G. Uninformative variable elimination for improvement of successive projections algorithm on spectral multivariable selection with different calibration algorithms for the rapid and non-destructive determination of protein content in dried laver. Anal. Methods 2011, 3, 1790–1796. [Google Scholar] [CrossRef]
  22. Soares, S.F.C.; Galvão, R.K.H.; Araújo, M.C.U.; Da Silva, E.C.; Pereira, C.F.; de Andrade, S.I.E.; Leite, F.C. A modification of the successive projections algorithm for spectral variable selection in the presence of unknown interferents. Anal. Chim. Acta 2011, 689, 22–28. [Google Scholar] [CrossRef] [PubMed]
  23. Soares, S.F.; Galvão, R.K.; Pontes, M.J.; Araújo, M.C. A new validation criterion for guiding the selection of variables by the successive projections algorithm in classification problems. J. Brazil. Chem. Soc. 2014, 25, 176–181. [Google Scholar] [CrossRef]
  24. Goodarzi, M.; Saeys, W.; de Araujo, M.C.U.; Galvão, R.K.H.; vander Heyden, Y. Binary classification of chalcone derivatives with LDA or KNN based on their antileishmanial activity and molecular descriptors selected using the successive projections algorithm feature-selection technique. Eur. J. Pharm. Sci. 2014, 51, 189–195. [Google Scholar] [CrossRef] [PubMed]
  25. Marreto, P.D.; Zimer, A.M.; Faria, R.C.; Mascaro, L.H.; Pereira, E.C.; Fragoso, W.D.; Lemos, S.G. Multivariate linear regression with variable selection by a successive projections algorithm applied to the analysis of anodic stripping voltammetry data. Electrochim. Acta 2014, 127, 68–78. [Google Scholar] [CrossRef]
  26. Xu, Z.B.; Si, X.B.; Li, C.; Chen, F. Study on the High-Speed Analysis of Coal Qualities by FT-NIR Method Based on Improved Successive Projections Algorithm. Adv. Mater. Res. 2015, 1094, 174–180. [Google Scholar] [CrossRef]
  27. Schapire, R.E. The strength of weak learnability. Mach. Learn. 1990, 5, 197–227. [Google Scholar] [CrossRef]
  28. Zou, X.; Zhao, J.; Li, Y. Selection of the efficient wavelength regions in FT-NIR spectroscopy for determination of SSC of “Fuji”apple based on BiPLS and FiPLS models. Vib. Spectrosc. 2007, 44, 220–227. [Google Scholar] [CrossRef]
  29. Centner, V.; Massart, D.L.; de Noord, O.E.; de Jong, S.; Vandeginste, B.M.; Sterna, C. Elimination of uninformative variables for multivariate calibration. Anal. Chem. 1996, 68, 3851–3858. [Google Scholar] [CrossRef] [PubMed]
  30. Gottardo, P.; de Marchi, M.; Cassandro, M.; Penasa, M. Technical note: Improving the accuracy of mid-infrared prediction models by selecting the most informative wavelengths. J. Dairy Sci. 2015, 98, 4168–4173. [Google Scholar] [CrossRef] [PubMed]
  31. Han, Q.J.; Wu, H.L.; Cai, C.B.; Xu, L.; Yu, R.Q. An ensemble of Monte Carlo uninformative variable elimination for wavelength selection. Anal. Chim. Acta 2008, 612, 121–125. [Google Scholar] [CrossRef] [PubMed]
  32. Lin, Z.; Pan, X.; Xu, B.; Zhang, J.; Shi, X.; Qiao, Y. Evaluating the reliability of spectral variables selected by subsampling methods. J. Chemometr. 2015, 29, 87–95. [Google Scholar] [CrossRef]
  33. Li, H.; Liang, Y.; Xu, Q.; Cao, D. Key wavelengths screening using competitive adaptive reweighted sampling method for multivariate calibration. Anal. Chim. Acta 2009, 648, 77–84. [Google Scholar] [CrossRef] [PubMed]
  34. Yun, Y.H.; Wang, W.T.; Deng, B.C.; Lai, G.B.; Liu, X.B.; Ren, D.B.; Xu, Q.S. Using variable combination population analysis for variable selection in multivariate calibration. Anal. Chim. Acta 2015, 862, 14–23. [Google Scholar] [CrossRef] [PubMed]
  35. Zhang, Z.Y. Determination of hesperidin in tangerine leaf by near-infrared spectroscopy with SPXY algorithm for sample subset partitioning and Monte Carlo cross validation. Spect. Anal. 2009, 29, 964–968. [Google Scholar]
  36. Dorado, M.P.; Pinzi, S.; de Haro, A.; Font, R.; Garcia-Olmo, J. Visible and NIR Spectroscopy to assess biodiesel quality: Determination of alcohol and glycerol traces. Fuel 2011, 90, 2321–2325. [Google Scholar] [CrossRef]
  37. Nordon, A.; Mills, A.; Burn, R.T.; Cusick, F.M.; Littlejohn, D. Comparison of non-invasive NIR and Raman spectrometries for determination of alcohol content of spirits. Anal. Chim. Acta 2005, 548, 148–158. [Google Scholar] [CrossRef]
  38. Moreira, E.D.T.; Pontes, M.J.C.; Galvão, R.K.H.; Araújo, M.C.U. Near infrared reflectance spectrometry classification of cigarettes using the successive projections algorithm for variable selection. Talanta 2009, 79, 1260–1264. [Google Scholar] [CrossRef] [PubMed]
  39. Zhang, S.Z.; Zhang, M.J. Wavelength selection from near infrared spectra by ensemble variable selection method. Comput. Appl. Chem. 2014, 31, 499–502. [Google Scholar]
  40. Fuchs, K.; Gertheiss, J.; Tutz, G. Nearest Neighbor Ensembles for Functional Data with Interpretable Feature Selection. Chemometr. Intell. Lab. 2015. [Google Scholar] [CrossRef]
  41. Zhang, H.L.; He, Y. Measurement of Soil Organic Matter and Available K Based on SPA-LS-SVM. Spect. Anal. 2014, 34, 1348–1351. [Google Scholar]

Share and Cite

MDPI and ACS Style

Qu, F.; Ren, D.; Wang, J.; Zhang, Z.; Lu, N.; Meng, L. An Ensemble Successive Project Algorithm for Liquor Detection Using Near Infrared Sensor. Sensors 2016, 16, 89. https://doi.org/10.3390/s16010089

AMA Style

Qu F, Ren D, Wang J, Zhang Z, Lu N, Meng L. An Ensemble Successive Project Algorithm for Liquor Detection Using Near Infrared Sensor. Sensors. 2016; 16(1):89. https://doi.org/10.3390/s16010089

Chicago/Turabian Style

Qu, Fangfang, Dong Ren, Jihua Wang, Zhong Zhang, Na Lu, and Lei Meng. 2016. "An Ensemble Successive Project Algorithm for Liquor Detection Using Near Infrared Sensor" Sensors 16, no. 1: 89. https://doi.org/10.3390/s16010089

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop