Next Article in Journal
Distributed Task Offloading in Heterogeneous Vehicular Crowd Sensing
Next Article in Special Issue
A Novel Optimization Technique to Improve Gas Recognition by Electronic Noses Based on the Enhanced Krill Herd Algorithm
Previous Article in Journal
Hazardous Traffic Event Detection Using Markov Blanket and Sequential Minimal Optimization (MB-SMO)
Previous Article in Special Issue
A Novel Pre-Processing Technique for Original Feature Matrix of Electronic Nose Based on Supervised Locality Preserving Projections
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Valid Probabilistic Predictions for Ginseng with Venn Machines Using Electronic Nose

1
State Key Laboratory of Industrial Control Technology, Institute of Cyber Systems and Control, Zhejiang University, Hangzhou 310027, Zhejiang, China
2
Computer Learning Research Centre, Royal Holloway, University of London, Egham Hill, Egham, Surrey TW20 0EX, UK
*
Author to whom correspondence should be addressed.
Sensors 2016, 16(7), 1088; https://doi.org/10.3390/s16071088
Submission received: 18 April 2016 / Revised: 22 June 2016 / Accepted: 22 June 2016 / Published: 13 July 2016
(This article belongs to the Special Issue E-noses: Sensors and Applications)

Abstract

:
In the application of electronic noses (E-noses), probabilistic prediction is a good way to estimate how confident we are about our prediction. In this work, a homemade E-nose system embedded with 16 metal-oxide semi-conductive gas sensors was used to discriminate nine kinds of ginsengs of different species or production places. A flexible machine learning framework, Venn machine (VM) was introduced to make probabilistic predictions for each prediction. Three Venn predictors were developed based on three classical probabilistic prediction methods (Platt’s method, Softmax regression and Naive Bayes). Three Venn predictors and three classical probabilistic prediction methods were compared in aspect of classification rate and especially the validity of estimated probability. A best classification rate of 88.57% was achieved with Platt’s method in offline mode, and the classification rate of VM-SVM (Venn machine based on Support Vector Machine) was 86.35%, just 2.22% lower. The validity of Venn predictors performed better than that of corresponding classical probabilistic prediction methods. The validity of VM-SVM was superior to the other methods. The results demonstrated that Venn machine is a flexible tool to make precise and valid probabilistic prediction in the application of E-nose, and VM-SVM achieved the best performance for the probabilistic prediction of ginseng samples.

1. Introduction

Electronic noses (E-noses) are devices designed to precisely detect and distinguish volatiles in complex samples by mimicking the mammalian sensory system. After thirty years of development, E-noses has been applied in the field of environmental quality monitoring [1], food and beverage quality control [2,3,4], medical diagnosis [5,6,7] and others.
Ginseng (Panax ginseng C.A. Meyer, mainly cultivated in Korea and Northeast China) and American ginseng (Panax quinquefolius, mainly cultivated in North America and China) are two main categories of ginseng, which can be subdivided into different subcategories according to species and production places Ginsengs are widely used in Chinese traditional medicine with great medicinal value. The increasing demand and consumption of ginsengs have led to some species substitution or adulteration with other species because of the high profits involved. However, it is not easy to identify the origin of ginseng species just based just on their morphology. Besides, many products are provided commercially in the form of a powder or shredded slices, which increases the difficulties of identification. Traditional identification of ginsengs is implemented with sensory analysis by a panel of experts, which is a costly process. Meanwhile, the validation of identification only depends on the various levels of experts’ experience. Normal analytical techniques, such as gas chromatography-mass spectrometry (GC-MS) [8] and high performance liquid chromatography (HPLC) [9] require long analysis times, complicated sample pretreatment and the use of sophisticated equipment. In this sense, electronic noses are a promising analysis system for the identification of traditional Chinese medicines due to their precise and stable performance, simple sample pretreatment, short measurement time and low cost.
Classification is one of major types of data treatment used for E-noses, which is also one of standard subjects in machine learning. In the applications of medical diagnosis, military usage and others, only predicting the label of testing sample is not enough, the probability estimation, which measure our trust in the prediction, should also be provided since it provides important information for risk control. Probabilistic prediction methods can help solve the problem.
Several classical probabilistic prediction algorithms have been proposed and widely used for classification cases, including Platt’s method [10] (based on support vector machine, SVM), logistic regression (for binary case) or Softmax regression (SR, for multi-classification case), Naive Bayes classification (NB) [11] and etc. However, most of these methods are based on strong statistical assumptions about the example distribution. Usually, examples distribution in Naïve Bayes was assumed to be a Gaussian distribution. Softmax Regression assumes the estimated probability to be the specific mathematic form shown in Section 2.3. Platt’s method assumes pairwise class probabilities take the specific mathematic form that will be shown in Section 2.4. Hence, once the assumed statistical model is unavailable or even incorrect, the estimated probability will be biased and the classification rate will be poor. This case can easily happen to E-nose data, because usually it’s difficult to find an accurate statistical model for real-world data.
The validity of probabilistic prediction is very important for probabilistic prediction methods. Validity means the estimated probability for predicted label is unbiased, which means the estimated probability is equal or close to the observed frequency that predictions are correct.
In this work, a flexible machine learning framework, Venn machine [12,13,14] was introduced to make valid probabilistic prediction for different species of ginseng samples, which means the estimated probability is unbiased. The only assumption required for Venn machine is that the example distribution is an independent identical distribution (I.I.D assumption), which can be easily satisfied by E-nose data, and do not need specific distribution of E-nose data Once the I.I.D assumption is satisfied, the validity of predicted probabilities is guaranteed.
A homemade electronic nose embedded with a sensor array of 16 metal-oxide semi-conductive gas sensors was used for the discrimination of nine ginsengs of different species or production places. Three Venn predictors separately based on three classical probabilistic prediction methods (Platt’s method, Softmax regression and Naive Bayes classification) and these three classical probabilistic prediction methods were developed to make probabilistic predictions for the ginseng samples. The performance of six methods were compared in the aspect of classification rate and validity of probabilistic predictions in offline mode. The validity and precision of probabilistic prediction by Venn predictors in online mode were also investigated. Electronic nose with VM-SVM demonstrated its strong ability to make precise and valid probability estimate with good classification rate.

2. Probabilistic Prediction Methods

2.1. Venn Machine

Consider a training set z i = ( x i , y i ) , i = 1 , , n 1 , consists of objects x i X and its label y i Y , The task is to predict y n for a new object x n and estimate the probability that the prediction is correct.
Firstly, hypothesizing y n = y , y Y . y is finite, and for every attempt ( x n , y ) , a Venn taxonomy, which is a measurable function A n , n N , divide each example in { z 1 , , z n } into the finite categories t i , t i T (T is the set of all sample categories) as follows:
t i = A n ( { z 1 , , z i 1 , z i + 1 , , z n } , z i ) , i = 1 , , n
z i and z j are assigned to the same category if and only if:
A n ( { z 1 , , z i 1 , z i + 1 , , z n } , z i ) = A n ( { z 1 , , z j 1 , z j + 1 , , z n } , z j )
Let p y be the empirical probability distribution of the labels in category τ , where τ is the category that z n is divided to:
p y { y } = | { ( x * , y * ) τ : y * = y } | | τ |
| τ | is the number of examples in category τ . p y , which is a row vector, is a probability distribution on Y consisting of K probabilities, where K = | Y | . After trying all y, we get a K × K matrix P. We define the minimum entry of a column as the quality of the column. The best column with the highest quality j best is our prediction for x n and the probability interval that the prediction is correct is:
[ Pl , Pu ] = [ min i = 1 , , K P i , j best , max i = 1 , , K P i , j best ]
where Pl and Pu is the lower and upper bound of the probability interval.
Any classical classification algorithm can be used as Venn taxonomy A n to assign a category t to z i with a training set { z 1 , , z i 1 , z i + 1 , , z n } . The validity of Venn machine is guaranteed in theory if the assumption that set z i = ( x i , y i ) , i = 1 , , n is generated from independent identical distribution is satisfied.
To further clarify how a Venn machine works, a simple example is given. Hypothesizing a dataset consisting of three categories labelled with A, B and C with 30 samples, 10 samples in each category. The aim is to make probabilistic prediction for new sample x with unknown label y.
Firstly, setting y = A, new sample (x, A) was added to the dataset, and total number of samples was 31. Then leave-one-out cross-validation was applied to predict the label for each sample. Finally, 31 prediction were obtained. Hypothesizing the label of x is predicted to be B, also 1 sample with true label A, 7 samples with true label B and 1 sample with true label C were also predicted to be B. then we got a row vector p A = [ 1 / 9   7 / 9   2 / 9 ] = [ 0.11   0.78   0.11 ] .
Secondly, setting y = B, then new sample (x, B) were added to the dataset. Leave-one-out cross-validation was applied to predict the label of each sample. Hypothesizing we got p B = [ 0.00   0.91   0.09 ] . Thirdly, setting y = C, then new sample (x, C) were added to the dataset. Leave-one-out cross-validation was applied to predict the label of each sample. Hypothesizing we got p c = [ 0.10   0.70   0.20 ] .
Finally, we got matrix P:
P = [ 0.11 0.78 0.11 0.00 0.91 0.09 0.10 0.70 0.20 ]
j best is B with best quality of the second column being 0.70. The predicted label for new sample x is B with probability interval [0.70 0.91].

2.2. Naive Bayes Classification

The Naïve Bayes classifier is designed to be used when features are independent of one another within each class. It appears to still work well in some cases when the independence assumption is not satisfied. It is based on Bayes’ theorem and defined the probability that example x belong to class k, k = 1,..., K is:
P ( y = k | x ) = P ( y = k ) P ( x | y = k ) P ( x ) = P ( y = k ) j = 1 m P ( x j | y = k ) P ( x )
Appropriate specific distribution should be assumed for the probability density P ( x j | y = k ) of examples and the Gaussian distribution is most widely used.

2.3. Softmax Regression

Softmax regression is a generalization of logistic regression for multi-classification. Given a test input x, to estimate the probability that P ( y = k | x ) for each class k = 1, …, K, we assume the K estimated probability take the form:
h θ ( x ) = [ P ( y = 1 | x ; θ ) P ( y = 2 | x ; θ ) P ( y = k | x ; θ ) ] = 1 j = 1 k exp ( θ ( j ) T x ) [ exp ( θ ( 1 ) T x ) exp ( θ ( 2 ) T x ) exp ( θ ( k ) T x ) ]
where θ ( 1 ) , θ ( 2 ) , …, θ ( k ) are parameter in the model needed to be optimized. The optimization objective is to minimize J ( θ ) :
min J ( θ ) = [ i = 1 l t = 1 K 1 { y ( i ) = t } log exp ( θ ( t ) T x ( i ) ) j = 1 k exp ( θ ( j ) T x ( i ) ) ]

2.4. Platt’s Method

Platt’s method was proposed to make SVM output probabilistic prediction [15]. An improved implementation was finished by Lin et al. [10]. Given a test input x, to estimate the probability that P ( y = k | x ) for each class k, k = 1, …, K, we firstly estimate pairwise class probabilities:
r ij P ( y = i | y = i   or   j , x )
If f is the decision value at x, then assuming pairwise class probabilities take the form:
r ij 1 1 + e Af + B
where A and B are estimated by:
min A , B y l = i   or   j 1 { y l = i } · r ij + 1 { y l = j } · r ji
where 1 { a   true   statement } = 1 , 1 { a   false   statement } = 0 . After collecting all r ij values, the following optimization problem is solved as:
min p 1 2 i = 1 k i : j i ( r ji p i r ij p j ) 2
Subject   to   p i 0 , i = 1 , , k ,   i = 1 k p i = 1

2.5. The Validity of Probabilistic Predictions

Probabilistic prediction can provide reliability estimate on the prediction. However, the estimated probability should be valid. In this work, we used two methods to examine the validity of probabilistic predictions. One is to use the standard loss function, the other one is to compare the cumulative probability value for predictions with cumulative number of correct predictions.

2.5.1. Loss Function

Two loss functions, log loss and square loss, were applied in this paper. Supposing p is the probability value for predicted label of testing example x. q is equal to 1 if the prediction is correct and equal to 0 otherwise. The log loss function is:
λ ln ( p , q ) = { ln ( 1 p )   if   q = 0 ln p   if   q = 1
The square loss function is:
λ sq ( p , q ) = ( p q ) 2
However, Venn machine outputs a probability interval (Pl, Pu) instead of single probability value p. So the corresponding minimax probabilistic prediction p is defined as follows:
For log loss function:
p = p u 1 p l + p u
For square loss function:
p = p u + p l 2 2 p u 2 2
p ( p l , p u ) . For more details about the derivation of Equations (12) and (13), please refer to the work [16]. Given n testing examples, for different methods, we will calculate and compare the mean log error ( d ln ):
d ln = 1 n i = 1 n λ ln ( p i , q i )
and the root mean square error ( d sq ):
d sq = 1 n i = 1 n λ sq ( p i , q i )
For each method, the smaller the d ln and d sq are, the better the validity of the method is.

2.5.2. Cumulative Probability Values versus Cumulative Correct Predictions

If the probability estimated by certain method is valid, with large number of n testing examples, the cumulative probability value (CP) should be close or equal to cumulative number of correct predictions (CN):
i = 1 n p ( y i * ) i = 1 n 1 i { y i * = y i }
y i * is the predicted label for x i , and y i is the true label. So the second assessment criteria for the validity of probabilistic prediction is defined to be the absolute difference between the average probability value and the ratio of correct prediction:
d 1 = | i = 1 N 1 i { y i * = y i } N i = 1 N p ( y i * ) N |
However, for Venn machines, which outputs a probability interval (Pl, Pu) instead of a single probability value, the cumulative number of correct predictions should be between the cumulative lower probability and the cumulative upper probability. This is an overall estimate on the validation of probabilistic prediction methods:
i = 1 n Pl ( y i * ) i = 1 n 1 i { y i * = y i } i = 1 n Pu ( y i * )
Firstly, we calculated the absolute difference between ratios of correct predictions, with cumulative upper bounds of probability intervals:
d 11 = | i = 1 N 1 i { y i * = y i } N i = 1 N pl ( y i * ) N |
and with cumulative lower bounds of probability intervals:
d 12 = | i = 1 N 1 i { y i * = y i } N i = 1 N pu ( y i * ) N |
Then, the second assessment criteria for the validity of probabilistic predictions by Venn machine is defined as follows:
d 1 = max ( d 11 , d 12 )
For each method, the smaller d 1 is, the better the validity of the method is.

3. Materials and Methods

3.1. Sample Preparation

Nine categories of ginsengs, thirty-five pieces of ginseng roots for each category, were randomly purchased from Changchun Medicinal Material Market (Changchun, China) and the details of materials were shown in Table 1. Prior to E-nose measurement, every ginseng root was ground into powder separately, and 10 g powder of each sample was placed and sealed in 100 mL headspace vials. Then the vials were placed in a thermostat at 50 °C for 30 min. 10 ml headspace gas from each vial was extracted with s syringe for measurement.

3.2. E-Nose Equipment and Measurement

A homemade E-nose embedded with sensor array of 16 metal-oxide gas sensors, was used in our experiment. The schematic diagram of the E-nose system was introduced in early work [17]. The gas sensors were purchased from Figaro Engineering Inc. (Osaka, Japan), including TGS800, TGS813*2, TGS816, TGS821, TGS822*2, TGS826, TGS830, TGS832, TGS880, TGS2600, TGS2602, TGS2610, TGS2611, TGS2620. All sensors were fixed on a printed circuit board, which was placed in a 200 mL stainless chamber. A three-way valve (Chengdu Qihai Electromechanical Equipment Manufacturing Co. Ltd., Chengdu, China) was designed to switch between target gas and clean dry air. Two mini vacuum pumps (Chengdu Qihai Electromechanical Equipment Manufacturing Co. Ltd.) were used for gas washing at a constant flow of 1 L/min. A data acquisition (DAQ) unit USB6211, purchased from National Instrument Inc. (Austin, TX, USA), was equipped to acquire the signal of sensors and control pumps. Heater voltage of 5V DC, recommended by Figaro Inc., was applied for each sensor to attain the best performance of sensor.
The measurement process was as follows: the chamber loaded with the sensor array was washed with a clean-dry-air flow of 1 L/min for 360 s to allow the sensors to return to baseline. Then the air flow was stopped and the target gas was injected into the chamber with a syringe. The responses of the sensors lasted for 180 s and then the air flow was turned on to wash away the target gas. The signals of 16 sensors were recorded for 340 s at 2 Hz for one measurement, including 20 s before and 140 s after the responses of sensors. The experiment was conducted at room temperature of 20–25 °C and humidity of 50%–70%. Finally, 9 × 35 = 315 samples were obtained (nine categories, 35 samples for each category).

3.3. Data Processing

A typical response curves of 16 sensors to ginseng samples were shown in Figure 1. Firstly, the voltage signal of each sensor was calibrated separately by:
V = V s V o
where V s is resistance signal, V o is the baseline.
Then, eight common used features were extracted from each sensor as follows:
1.
The maximal absolute response, V max = max ( | V | ) , which is most efficient and widely-used steady feature.
2.
The area under the full response curve, V int = 0 T V ( t ) dt , where T (=340 s) is the total measurement time, which is also widely- used steady feature.
3–8.
Exponential moving average of derivative [18,19] of V, E a ( V ) = [ min ( y ( k ) ) , max ( y ( k ) ) , where the discretely sampled exponential moving average y ( k ) = ( 1 a ) y ( k 1 ) + a ( V ( k ) V ( k 1 ) ) with smoothing factors a = 1 / ( 100 × SR ) ,   1 / ( 10 × SR ) ,   1 / SR . SR is the sampling rate, SR = 10 Hz. y ( 1 ) = aV ( 1 ) . For each smoothing factor, two features were extracted. A total of six transient feature were extracted. Besides steady features, transient features were considered to contain much effective information that should be made the best of [20].
Finally, 16 × 8 = 128 features are extracted from each sample and all the features are scaled to [0 1]:
x i , j = x i , j min j ( x i , j ) max ( j x i , j ) min ( j x i , j )
where x i , j is the jth feature from ith sensor. x i , j is the feature after normalization.
Three Venn predictors, VM-SVM, VM-SR and VM-NB, separately based on SVM, Softmax regression and Naïve Bayes, were developed for the probabilistic prediction of ginseng samples in offline and online modes. Three classical probabilistic prediction methods: Platt’s method, Softmax regression and Naïve Bayes were also used for the probabilistic prediction of ginseng samples in offline and online modes. The classification rate and validity of probabilistic predictions by Venn predictors and classical probabilistic prediction methods were compared in offline mode. Platt’s method and SVM were performed with libsvm toolbox [21] using C-SVC (SVM classification with cost parameter C) with nonlinear kernel of Radial Basis Function (RBF). In the training stage, C was searched within [22, 24 … 214] and γ (parameter of kernel RBF) was searched within [2−9, 2−7 … 2−1] with 5-fold cross-validation to obtain optimum parameters for the model. Then, the trained model was performed on the testing examples. For Naïve Bayes, the distribution was set to be Gaussian distribution. All algorithm were implemented in MATLAB 2014a.

4. Results and Discussion

In this work, the performance of probabilistic predictors was investigated in both offline and online mode. In offline mode, the dataset was divided in to two parts: training set and testing set. The model was trained with the training set to form a fixed decision rule. Then the model was applied for the testing set. The prediction performance of testing set was treated as criterion of the model. This is the usual way in most works. In online mode, every time one sample was tested, the sample with its true label was added in the training set, then the model was rebuilt and used to predict for another testing sample. As the number of samples in training set increased, the performance of the model improved gradually and tended to be stable. Online prediction is also very meaningful in practical applications. Firstly, we compared the classification rate and validity of six probabilistic predictors in offline mode. Then, we investigated the validity of Venn predictors in online mode.

4.1. Performance of Probabilistic Predictors in Offline Mode

In offline mode, three Venn predictors, VM-SVM, VM-SR, VM-NB, and corresponding underlying probabilistic prediction methods, Platt’s method, Softmax Regression, Naïve Bayes were applied for predictions of ginseng samples with ‘leave-one-out’ cross-validation: in every cycle, one sample from sample set was treated as testing set, the other samples were treated as training set. Model of prediction method trained with training set was applied to predict for the testing set. The cycle was repeated until all samples in sample set were treated as testing set once. Then the classification rates and assessment criteria of validity of probabilistic prediction results by each method were investigated and shown in Table 2.
From Table 2, we can see that Platt’s method based on SVM achieved the highest classification rate, and classification rate of VM-SVM was just 2.22% lower than that of Platt’s method. The classification rates of VM-SR were a little higher than that of corresponding underlying algorithm SR. However, Venn machine greatly improved the classification rate of Naïve Bayes from 40% to 60%, which demonstrated the ability of Venn machine to improve the classification performance of the underlying method, though it didn’t work for every method. The assessment criteria of validity of Venn predictors were all smaller than that of corresponding underlying methods, which indicated that the probabilistic prediction conducted by Venn predictors were more valid than corresponding underlying probabilistic prediction methods.
VM-SVM, which benefits from the strong classification performance of SVM, achieved the best classification rate and assessment criteria of validity among three Venn predictors, though d 1 of VM-SVM was a litter bigger than that of VM-NB. We can conclude that the VM-SVM achieved the best performance among three Venn predictors, VM-SR came next, and VM-NB performed worst because of poor classification ability of Naïve Bayes for ginseng samples. The results indicated that choosing appropriate underlying classification method is very important for Venn predictors.
Similar to binary case, sensitivity and specificity for classification result of dataset with nine category are defined as: for certain category, sensitivity is the proportion that samples in this category are correctly classified, specificity is one minus proportion that samples in other categories are wrongly classified to this category. The sensitivity and specificity for each category with each method were shown in Table 3.
To further investigate the validity of each method, the cumulative probability value/interval was compared with cumulative correct predictions, as shown in Figure 2. We can see that for all three Venn predictors, as the number of predicted samples increase, the cumulative correct predictions kept increasing between cumulative upper bound and cumulative lower bound of probability intervals, with some exceptions caused by statistical fluctuations. This result indicates that the validity of Venn predictors is guaranteed in spite of underlying methods. However, for Platt’s method and Softmax Regression, as the cumulative numbers of predictions increase, the cumulative probability values deviated gradually from cumulative correct predictions. On the whole, the predicted probability values were underestimated by Platt’s method and overestimated by Softmax Regression.
The average widths of probability interval of VM-SVM and VM-NB were 0.0417 and 0.0396 separately, which demonstrated that the probability intervals by Venn predictors were quite narrow and nearly as precise as single probability value by classical probability prediction methods. The average width of probability interval of VM-SR was 0.1353, which was not as good as VM-SVM and VM-NB. This result indicates that the precision of Venn predictors are greatly dependent on underlying classification methods.
All the results above demonstrated the superior performance of the Venn machine over classical probabilistic prediction methods in the aspect of validity of predictions, and VM-SVM is the most reliable probabilistic prediction method for ginseng samples. Platt’s method achieved a little higher classification rate, but worse validity performance than VM-SVM, so we need to take a balance between them in practical application according to different aims.

4.2. Performance of Venn Predictors in Online Mode

In the initial stage of online prediction, the model performance is poor due to the limited number of training samples. As the number of training samples increases, the model performance improves and gradually becomes stable. Whether the validity of the probabilistic prediction still holds and how the precision of the probabilistic prediction changes in online mode were investigated here. In online mode, three samples of each category were randomly taken from the sample set to constitute the initial training set, and the other samples were treated as the testing set. In every cycle, one sample was taken without replacement from the testing set, and the prediction method model was trained with the training set was applied to predict the sample. Then the sample with correct label was added into the training set, then the model was updated and the next sample predicted. The cycle was repeated until all the samples in the testing set were tested. The correct predictions of VM-SVM, VM-SR and VM-NB were accumulated according to the sequence that the samples were tested and compared with the corresponding cumulative upper and lower bound of the probability intervals, as shown in Figure 3. The validity of probability predictions of Venn predictors still held all the way and in spite of the underlying classification method in online mode. Similar to the performance of Venn predictors in offline mode, VM-SVM performed best in online mode, VM-SR came next, VM-NB performed worst.
The validity of Venn predictors was guaranteed, then we investigated the precision of the probability interval and observed the change of probability interval during online prediction process. As the prediction ability of each method for samples from different categories was different, we separately showed the probability intervals of samples from each category according the sequence in which they were tested. Taking the result of VM-SVM as an example, as shown in Figure 4, as the number of the samples in training set increased, the width of probability intervals gradually decreased and the distribution gradually moved upwards, which means the probabilistic predictions yielded by Venn predictors become more and more precise as the number of samples in training set increased while the validity still held all the way.

5. Conclusions

In this work, a homemade metal-oxide-sensor-based electronic nose was utilized to discriminate a large set of ginseng samples. A flexible machine learning framework, Venn machine, was introduced to make precise and valid probabilistic prediction for ginsengs. Three Venn predictors were developed and compared with three classical probabilistic prediction methods in the aspect of classification rate and especially the validity of probabilistic predictions. The results showed that the validity of Venn predictors held all the way as the samples increased in spite of the underlying classification methods in both offline and online mode. The validity of Platt’s method and Softmax Regression was much worse than that of corresponding Venn predictors. Platt’s method achieved a best classification rate of 88.57% in offline mod, which was 2.22% higher than that of VM-SVM. The classification rates of VM-SR were a little higher than that of corresponding underlying method. We also surprisingly found that VM-NB greatly improved the classification rate of Naïve Bayes from 40% to 60%. The probability intervals output by Venn predictor were quite narrow, and close to single probability value. The results indicated that Venn machine was a flexible tool for the application of electronic noses which output precise and valid probabilistic predictions.

Acknowledgments

The work was supported by the National High Technology Research and Development Program of China (Grant No. 2013AA041201) and the Research Program of the Education Department of Zhejiang Province (Grant No. Y201328881).

Author Contributions

You Wang, Jiacheng Miao, Zhiyuan Luo and Guang Li conceived and designed the experiments; Jiacheng Miao performed the experiments; Jiacheng Miao, Xiaofeng Lyu and Linfeng Liu analyzed the data; Jiacheng Miao, You Wang and Guang Li wrote the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. De Vito, S.; Piga, M.; Martinotto, L.; Di Francia, G. Co, NO 2 and NO x urban pollution monitoring with on-field calibrated electronic nose by automatic bayesian regularization. Sens. Actuators B Chem. 2009, 143, 182–191. [Google Scholar] [CrossRef]
  2. Musatov, V.Y.; Sysoev, V.V.; Sommer, M.; Kiselev, I. Assessment of meat freshness with metal oxide sensor microarray electronic nose: A practical approach. Sens. Actuators B Chem. 2010, 144, 99–103. [Google Scholar] [CrossRef]
  3. Baldwin, E.A.; Bai, J.H.; Plotto, A.; Dea, S. Electronic noses and tongues: Applications for the food and pharmaceutical industries. Sensors 2011, 11, 4744–4766. [Google Scholar] [CrossRef] [PubMed]
  4. Fu, J.; Huang, C.Q.; Xing, J.G.; Zheng, J.B. Pattern classification using an olfactory model with pca feature selection in electronic noses: Study and application. Sensors 2012, 12, 2818–2830. [Google Scholar] [CrossRef] [PubMed]
  5. Schmekel, B.; Winquist, F.; Vikstrom, A. Analysis of breath samples for lung cancer survival. Anal. Chim. Acta 2014, 840, 82–86. [Google Scholar] [CrossRef] [PubMed]
  6. Montuschi, P.; Mores, N.; Trove, A.; Mondino, C.; Barnes, P.J. The electronic nose in respiratory medicine. Respiration 2013, 85, 72–84. [Google Scholar] [CrossRef] [PubMed]
  7. Chatterjee, S.; Castro, M.; Feller, J.F. An e-nose made of carbon nanotube based quantum resistive sensors for the detection of eighteen polar/nonpolar voc biomarkers of lung cancer. J. Mater. Chem. B 2013, 1, 4563–4575. [Google Scholar] [CrossRef]
  8. Moon, J.; Han, H.; Park, S.; Dong, H.; Han, K.-Y.; Kim, H.; Bang, K.-H.; Choi, J.; Noh, B. Discrimination of the origin of commercial red ginseng concentrates using LC-MS/MS and electronic nose analysis based on a mass spectrometer. Food Sci. Biotechnol. 2014, 23, 1433–1440. [Google Scholar] [CrossRef]
  9. Lee, D.Y.; Cho, J.G.; Lee, M.K.; Lee, J.W.; Lee, Y.H.; Yang, D.C.; Baek, N.I. Discrimination of panax ginseng roots cultivated in different areas in Korea using hplc-elsd and principal component analysis. J. Ginseng. Res. 2011, 35, 31–38. [Google Scholar] [CrossRef]
  10. Lin, H.-T.; Lin, C.-J.; Weng, R. A note on platt’s probabilistic outputs for support vector machines. Mach. Learn. 2007, 68, 267–276. [Google Scholar] [CrossRef]
  11. Williams, C.K.I.; Barber, D. Bayesian classification with gaussian processes. IEEE Trans. Pattern Anal. 1998, 20, 1342–1351. [Google Scholar] [CrossRef]
  12. Nouretdinov, I.; Devetyarov, D.; Vovk, V.; Burford, B.; Camuzeaux, S.; Gentry-Maharaj, A.; Tiss, A.; Smith, C.; Luo, Z.Y.; Chervonenkis, A.; et al. Multiprobabilistic prediction in early medical diagnoses. Ann. Math. Artif. Int. 2015, 74, 203–222. [Google Scholar] [CrossRef]
  13. Vork, V.; Gammerman, A.; Shafer, G. Algorithmic Learning in a Random World; Springer: New York, NY, USA, 2005; p. 16. [Google Scholar]
  14. Papadopoulos, H. Reliable probabilistic classification with neural networks. Neurocomputing 2013, 107, 59–68. [Google Scholar] [CrossRef]
  15. Platt, J.C. Probabilistic outputs for support vector machines and comparison to regularized likelihood methods. In Advances in Large Margin Classiers; MIT Press: Cambridge, MA, USA, 2000. [Google Scholar]
  16. Vovk, V.; Petej, I. Venn-abers predictors. In Proceedings of the 30th Conference on Uncertainty in Artificial Intelligence, Quebec City, QC, Canada, 23–27 July 2014.
  17. Miao, J.C.; Zhang, T.L.; Wang, Y.; Li, G. Optimal sensor selection for classifying a set of ginsengs using metal-oxide sensors. Sensors 2015, 15, 16027–16039. [Google Scholar] [CrossRef] [PubMed]
  18. Muezzinoglu, M.K.; Vergara, A.; Huerta, R.; Rulkov, N.; Rabinovich, M.I.; Selverston, A.; Abarbanel, H.D.I. Acceleration of chemo-sensory information processing using transient features. Sens. Actuators B Chem. 2009, 137, 507–512. [Google Scholar] [CrossRef]
  19. Rodriguez-Lujan, I.; Fonollosa, J.; Vergara, A.; Homer, M.; Huerta, R. On the calibration of sensor arrays for pattern recognition using the minimal number of experiments. Chemometr. Intell. Lab. 2014, 130, 123–134. [Google Scholar] [CrossRef]
  20. Fonollosa, J.; Vergara, A.; Huerta, R. Algorithmic mitigation of sensor failure: Is sensor replacement really necessary? Sens. Actuators B Chem. 2013, 183, 211–221. [Google Scholar] [CrossRef]
  21. Chang, C.C.; Lin, C.J. Libsvm: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
Figure 1. Typical responses of 16 sensors to ginseng samples.
Figure 1. Typical responses of 16 sensors to ginseng samples.
Sensors 16 01088 g001
Figure 2. Validity of probabilistic predictions by (a) VM-SVM; (b) VM-SR; (c) VM-NB; (d) Platt’s method; (e) Softmax Regression; and (f) Naïve Bayes in offline mode.
Figure 2. Validity of probabilistic predictions by (a) VM-SVM; (b) VM-SR; (c) VM-NB; (d) Platt’s method; (e) Softmax Regression; and (f) Naïve Bayes in offline mode.
Sensors 16 01088 g002
Figure 3. Validity of probabilistic predictions by (a) VM-SVM (Venn machine based on SVM); (b) VM-SR (Venn machine based on Softmax Regression); (c) VM-NB (Venn machine based on Naïve Bayes) in online mode.
Figure 3. Validity of probabilistic predictions by (a) VM-SVM (Venn machine based on SVM); (b) VM-SR (Venn machine based on Softmax Regression); (c) VM-NB (Venn machine based on Naïve Bayes) in online mode.
Sensors 16 01088 g003
Figure 4. Change of precision of predicted probability intervals for samples from category (19) during online prediction process with VM-SVM.
Figure 4. Change of precision of predicted probability intervals for samples from category (19) during online prediction process with VM-SVM.
Sensors 16 01088 g004
Table 1. Details of the ginseng samples.
Table 1. Details of the ginseng samples.
No.Ginseng SpeciesPlaces of Production
1Chinese red ginsengJi’an
2Chinese red ginsengFusong
3Korean red ginsengJi’an
4Chinese white ginsengJi’an
5Chinese white ginsengFusong
6American ginsengFusong
7American ginsengUSA
8American ginsengCanada
9American ginsengTonghua
Table 2. Classification rates and assessment criteria of validity of probabilistic prediction results for ginseng samples by each method.
Table 2. Classification rates and assessment criteria of validity of probabilistic prediction results for ginseng samples by each method.
MethodsClassification RateAssessment Criteria of Validity
dlndsqd1
VM-SVM86.35%0.38620.34190.0373
Platt’s method88.57%0.38760.34390.1480
VM-SR77.78%0.46900.39380.1085
SR76.19%Inf a0.43760.1853
VM-NB60.32%0.56830.44750.0266
NB40.32%0.58510.45100.0332
a In the prediction result of SR, predicted probability value of certain sample was 1, whereas the prediction was wrong, which lead this criteria to be infinite.
Table 3. Sensitivity and specificity for each category with each method.
Table 3. Sensitivity and specificity for each category with each method.
Method/Category123456789
VM-SVMSensitivity0.97140.885710.85710.828610.68570.82860.7143
Specificity0.98570.98930.99640.98210.978610.95360.98570.9750
Platt’s methodSensitivity0.97140.885710.85710.885710.71430.85710.7714
Specificity0.98570.98930.99640.98930.982110.96070.98570.9786
VM-SRSensitivity0.80000.82860.94290.74290.571410.51430.80000.8000
Specificity0.97140.971410.95360.96430.99640.95710.97500.9607
SRSensitivity0.74290.74290.97140.71430.600010.54290.80000.7429
Specificity0.97140.971410.95360.96430.99640.95710.97500.9607
VM-NBSensitivity0.800000.97140.65710.42860.94290.42860.48570.7143
Specificity0.84640.97860.99290.95000.932110.94290.97140.9393
NBSensitivity00.31430.97140.28570.2860.68570.5710.37140.1429
Specificity10.96070.99290.87500.87860.91070.89290.86790.9500

Share and Cite

MDPI and ACS Style

Wang, Y.; Miao, J.; Lyu, X.; Liu, L.; Luo, Z.; Li, G. Valid Probabilistic Predictions for Ginseng with Venn Machines Using Electronic Nose. Sensors 2016, 16, 1088. https://doi.org/10.3390/s16071088

AMA Style

Wang Y, Miao J, Lyu X, Liu L, Luo Z, Li G. Valid Probabilistic Predictions for Ginseng with Venn Machines Using Electronic Nose. Sensors. 2016; 16(7):1088. https://doi.org/10.3390/s16071088

Chicago/Turabian Style

Wang, You, Jiacheng Miao, Xiaofeng Lyu, Linfeng Liu, Zhiyuan Luo, and Guang Li. 2016. "Valid Probabilistic Predictions for Ginseng with Venn Machines Using Electronic Nose" Sensors 16, no. 7: 1088. https://doi.org/10.3390/s16071088

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop