Next Article in Journal
Wearable Belt With Built-In Textile Electrodes for Cardio—Respiratory Monitoring
Previous Article in Journal
Short-Range Water Temperature Profiling in a Lake with Coastal Acoustic Tomography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Machine Learning Method for the Detection of Brown Core in the Chinese Pear Variety Huangguan Using a MOS-Based E-Nose

1
College of Information Science and Technology, Beijing University of Chemical Technology, Beijing 100029, China
2
Beijing Advanced Innovation Center for Soft Matter Science and Engineering, Beijing University of Chemical Technology, Beijing 100029, China
3
Department of Chemistry, Institute of Inorganic and Analytical Chemistry, Goethe-University, Max-von-Laue-Str. 9, 60438 Frankfurt, Germany
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(16), 4499; https://doi.org/10.3390/s20164499
Submission received: 14 July 2020 / Revised: 5 August 2020 / Accepted: 8 August 2020 / Published: 12 August 2020
(This article belongs to the Section Electronic Sensors)

Abstract

:
The brown core is an internal disorder that significantly affects the palatability and economic value of Chinese pears. In this study, a framework that includes a back-propagation neural network (BPNN) and extreme learning machine (ELM) (BP-ELMNN) was proposed for the detection of brown core in the Chinese pear variety Huangguan. The odor data of pear were collected using a metal oxide semiconductor (MOS) electronic nose (E-nose). Principal component analysis was used to analyze the complexity of the odor emitted by pears with brown cores. The performances of several machine learning algorithms, i.e., radial basis function neural network (RBFNN), BPNN, and ELM, were compared with that of the BP-ELMNN. The experimental results showed that the proposed framework provided the best results for the test samples, with an accuracy of 0.9683, a macro-precision of 0.9688, a macro-recall of 0.9683, and a macro-F1 score of 0.9685. The results demonstrate that the use of machine learning algorithms for the analysis of E-nose data is a feasible and non-destructive method to detect brown core in pears.

1. Introduction

Pears are the second most consumed pome fruit worldwide [1], with consumption by the adult population ranging from 23 to 108 g/day [2]. Pears have many beneficial properties for human health, such as antitussive, anti-inflammatory, antihyperglycemic, and diuretic [3]. As an ancient fruit in temperate regions [4], there are many groups of pears, such as the white pear (Pyrus bretschneideri Rehd.), sand pear (Pyrus pyrifolia (Burm. f.) Nakai), and Ussurian pear (Pyrus ussuriensis Maxim.). The Chinese pear variety Huangguan (Pyrus bretschneideri Rehd.), which is a hybrid between Pyrus bretschneideri Rehd. and Pyrus pyrifolia (Burm. f.) Nakai [5], is widely planted in northern China [6] and is popular among consumers because of its crispy and juicy flesh, good taste, and salubrity [7,8]. However, brown core, which is a physiological disease caused by high concentrations of CO2 [9] and enzymes [10], occurs during the transportation and storage of pears, significantly affecting the palatability and economic value of the fruit [11]. It is difficult to detect browning in pears based on their appearance, and destructive methods are required because browning of pears usually spreads from the core to the flesh [12]. Therefore, it is of practical significance to develop a non-destructive method to detect brown core in the Chinese pear Huangguan.
An electronic nose (E-nose), designed to mimic human olfactory perception and identify volatile gases, is a device which mainly contains three parts: a sampling system, a sensors array unit and a data acquisition and processing system [13]. The principle of E-nose is that the sensor array detects an odor composed of numerous different volatiles in a sample’s headspace and provides a “fingerprint” which is employed to mine potential information about samples based on appropriate algorithm [14]. E-nose devices play an important role in fruit quality detection due to their advantages of non-destructive analysis [15] and portability [16]. Voss et al. [17] obtained odor data from peaches using a metal oxide semiconductor (MOS) E-nose and designed a model to monitor the growth cycle of peaches using the Random Forest method, which achieved an accuracy of 98.08% on sample test step. Feng et al. [18] proposed a method for evaluating the freshness of cherry tomatoes using a MOS-based E-nose and analyzing the data with an extreme learning machine (ELM) and the partial least-squares method, which proved that the combination of the E-nose and the ELM was reliable for evaluating the quality and freshness of cherry tomatoes during the cold storage. Chen et al. [19] demonstrated that using an E-nose and headspace solid-phase micro-extraction combined with gas chromatography-mass spectrometry (HS-SPME/GC-MS) was a fast and accurate method to examine the volatilization of ten varieties of fresh jujube. To the best of our knowledge, most previous studies that utilized E-nose devices have focused on the identification of fruit ripeness, fruit freshness, and fruit flavor; however, few studies investigated the detection of brown core in the Chinese pear variety Huangguan using an E-nose.
Machine learning (ML) is an application of artificial intelligence (AI) and data science [20]; it is an interdisciplinary method that includes probability theory, statistics, approximation theory, and many other disciplines [21]. ML has become one of the fastest-growing technical fields [20]. An artificial neural network (ANN) is considered a flexible mathematical tool that is based on the neurons in the brain and has been used to analyze data in numerous fields [22]. ANNs can describe complex relationships between inputs and outputs. ANNs provide better performance than traditional ML methods and have been widely applied in many fields, such as food science [23], medicine [24], and chemistry [25].
In this study, we combine a back-propagation neural network (BPNN) with an ELM and propose the BP-ELMNN framework to detect the browning of Chinese pears (Huangguan) using a MOS-based E-nose. In the framework, a shallow BPNN is designed to provide sufficient pear odor information for the ELM and improve the classification performance of the ELM. The adaptive moment estimation (Adam) algorithm [26], an efficient stochastic optimization algorithm, is used for weight training instead of the traditional gradient descent method. We use the synthetic minority oversampling technique (SMOTE) [27] to increase the number of minority class samples and avoid unbalanced samples. The performances of three other neural networks (BP, ELM, and radial basis function neural network (RBFNN)) are compared with that of the proposed BP-ELMNN framework.

2. Materials and Methods

2.1. Materials and Preparation

We used 342 Chinese pears (Huangguan variety) (referred to as Chinese pears or pears hereafter) (harvested in Zhao county, Hebei province, China) with similar color and size and without insect pests and mechanical damage. The pears were stored (temperature: 0 °C, relative humidity: 90%) prior to the experiment. We used 240 pears as training samples (after 239 days of storage) and 102 pears as test samples (after 251 days of storage).
As shown in Table 1, the browning levels of the Chinese pears were determined by experts at the National Pear Improvement Center at the Hebei Academy of Agricultural and Forestry Sciences. Class 0 represents healthy pears, and class 1 to class 5 are five classes representing increasing degrees of browning. * denotes whether the browning can be discerned by the appearance of the pears. The number of samples for each class are described in Section 4.

2.2. E-nose Analysis and Data Acquisition

A PEN3 E-nose (Airsense Analytics GmbH, Schwerin, Germany [28]) was used to collect the odor information of the pears. An array of 10 different MOS sensors is the critical component of the PEN3 E-nose; the details of the 10 sensors are shown in Table 2 [28].
All the experiments were conducted in the author’s laboratory (temperature: 25 ± 1 °C, relative humidity: 50 ± 2%). Figure 1 presents the schematic diagram of the collection of odor information using the E-nose. Before using the E-nose to collect data, the pear was placed into a crisper (polypropylene, 150 mm × 150 mm × 180 mm) for 900 s to allow the volatile odor of the pears to fill the crisper. We used the 15 polypropylene crispers to place the sample pears in our experiments. The crisper was cleaned with air filtered by activated carbon after a pear was sampled and not be reused until the remaining 14 crispers were used. The E-nose operation consisted of the collection and cleaning stages. During the collection stage, air flows into the crisper in the direction of arrow-1 and flows out of the E-nose in the direction of arrow-3. During this stage, valve-1 is in the open state and valve-2 is in the closed state. The sample’s volatile gas was drawn into the E-nose at a constant flow rate of 10 mL/s and came into contact with the ten MOS sensors in the chamber. As the sensor surface became saturated, the conductivity increased, and the value stabilized. The collection stage lasted 90 s. During the cleaning stage, air filtered by activated carbon was drawn into the E-nose in the direction of arrow-2 to remove the substances adsorbed on the surface of the sensor. During this stage, valve-2 is in the open state and valve-1 is in the closed state. The cleaning stage also lasted 90 s to ensure that the material adsorbed on the surface of the sensor was removed entirely. The cleaning stage occurred after the completion of the collection stage. According to the characteristics of the MOS sensor, the response curve of the sensor increases rapidly initially and then flattens gradually in the collection stage. The stable value (SV) is a crucial and simpler feature parameter of the E-nose response signal, which can reflect the properties of substances in the volatile gas and be used in pattern recognition algorithms appropriately. We used the SV of the response signal as the input data of the neural network. Since the response signal of the PEN3 E-nose entered the stable state at 30 s, we used the data from 31 s to 90 s as the SVs. An example of the response curves of the E-nose for pears during 90 s of collecting stage was showed in Figure 2. Each curve represents the ratio of G (the conductivity of sensors contact with sample’s volatile gas) to G0 (the conductivity of sensors contact with air filtered by activated carbon) and the feature area was displayed between the two red dotted lines. Therefore, the data we obtained were expressed as a 14,400 × 10 matrix for training, and a 6120 × 10 matrix for testing (the sampling period of the E-nose was 1 s).
After collecting the data using the E-nose, the pears were immediately cut open and labeled by experts (the label result were showed in Table 3). Photos of the Chinese pears in the six classes of browning levels were obtained and shown in Figure 3b–d. It is challenging to determine to brown of the core based on the appearance in class 1 to class 3.

2.3. Adaptive Moment Estimation Algorithm

The selection of an efficient and suitable optimization algorithm is crucial for neural network training. The Adam algorithm combines the advantages of the AdaGrad method [29] and the RMSProp method [30] and is an optimization algorithm that only requires first-order gradients with little memory requirement [26]. The gradient is not used to update the parameters for the optimization of the objective function in the Adam algorithm. The moment estimation, which is defined in Equations (1) and (2), is used to update the parameters:
m t = ε 1 m t 1 + ( 1 ε 1 ) g t
v t = ε 2 v t 1 + ( 1 ε 2 ) g t 2
where t and g t denote the timestep and gradient, respectively. m and v are the first-order moment estimate and second-order moment estimate, respectively. ε 1 and ε 2 are attenuation coefficients. In the initial phase, m t and v t tend to be biased towards the starting values [31]. Therefore, the moment estimation is modified as follows:
m ^ t = m t 1 ε 1 t
v ^ t = v t 1 ε 2 t
The parameter update rule is defined as:
ϑ t = ϑ t 1 σ m ^ t v ^ t + ϵ
where σ is the learning rate. As an extension of the stochastic gradient descent method [32], the Adam algorithm possesses fast speed in the optimization process without falling into the local optimum [33].

2.4. Synthetic Minority Oversampling Technique

Unbalanced sample data adversely affect the training performance in neural networks. This problem can be solved by the oversampling technology to supplement minority samples. The SMOTE is one of the most widely used oversampling methods [34] and generates a new sample x i randomly between a minority class sample x m and a k-nearest neighbor minority class sample x n of x m as follows:
x i = x m + α × ( x m x n )
where α is a random number that is greater than 0 and smaller than 1. After oversampling, the features of each class of data can be learned by the neural networks.

2.5. Back-Propagation Neural Network

The BPNN is one of the most mature and well-known feed-forward neural networks [35]; it is wildly used for prediction and classification [36,37]. A BPNN consists of an input layer, a hidden layer, and an output layer, and all layers are connected. The data in the input layer of the BPNN are processed layer by layer using forward propagation until the output layer is obtained. The data features are extracted using a layer by layer calculation process [38]. The back-propagation algorithm in the BPNN is used to calculate the weights required in the network according to the error between the network output and expected output.

2.6. Extreme Learning Machine

An ELM is a single hidden layer neural network, which has better training performance than traditional algorithms [17,18,39]. One of the most significant features of the ELM is that the weights do not need to be obtained by iterative updating. The optimal solution of the ELM is obtained by one-time learning [40]. The input weights and deviations of the hidden layer of the ELM are generated randomly, and the input weights of the output layer are obtained analytically [40]. Due to the powerful performance of the ELM, it has been widely used in many fields [22].

2.7. Radial Basis Function Neural Network

The RBFNN is a three-layer feed-forward neural network. The input layer of the RBFNN does not process the data but maps the data directly to the hidden layer [41]. Each neuron in the hidden layer represents a radial basis function, and the data is sent to the output layer after the calculation of the radial basis function. A simple linear transformation of the hidden layer output is performed to obtain the output layer [42]. The center and width of the hidden layer neurons in the RBFNN have a significant influence on the training results of the network [43,44].

3. Proposed Method

It is well known that in traditional BPNNs, a large number of hidden layers result in excess features learned by the network, leading to overfitting of the network. However, if there are too few hidden layers, the performance of the BPNN could be affected. The ELM has better generalization performance than the BPNN [22], but the structural characteristics of a single hidden layer may result in an inability of the network to extract sufficient features of the data.
We combined the advantages of the two networks and developed the BP-ELMNN framework (Figure 4) to detect the degree of pear browning. The output of the hidden layer of the BPNN is input into the ELM. This framework contains three dense layers. The activation function of the neurons in the Dense 1 layer and Dense 2 layer is the Relu function, and the neurons in the Dense 3 layer are activated by a Sigmoid function. The Adam algorithm is used to iteratively update the weights and thresholds from the input layer to the Dense 2 layer. The weights and thresholds between the Dense 2 layer and Dense 3 layer are generated randomly and are not iteratively updated during network training. The input weights of the output layer are obtained using Equation (7).
β = O d e n s e 3 1 T
where β is the input weight of the output layer, O d e n s e 3 denotes the output of the Dense 3 layer, and T is the label of the training data.
The loss function is the cross-entropy function, which is defined as:
L o s s 1 = i = 1 n l i log ( y i )
where n is the number of classes, l i is the label of the data, and y i denotes the actual output of the neural network. The L2 regularization term of the weight vectors was used in the loss function to avoid overfitting:
L o s s = L o s s 1 + α r W 2 2
where α r is the regularization coefficient, W denotes the weight vectors, and · 2 is the L2 norm.
The proposed BP-ELMNN Algorithm 1 is summarized as follows:
Algorithm 1. BP-ELMNN
Input: training data.
Output: predicted category.
Begin
Step 1: train the BPNN part on the training data using the Adam algorithm.
Step 2: randomly select the input weights and thresholds of the Dense 3 layer.
Step 3: calculate the input weight β of the output layer using Equation (7).
Step 4: test the BP-ELMNN model on the test data.
Step 5: output the classification results.
End

4. Result and Discussion

4.1. Data Analysis and Sample Supplementation

The training samples and test samples of the Chinese pears were divided into six categories according to the browning levels (see Table 1); the number of samples is shown in Table 3. Since it was not possible to obtain the same number of pears in each category, we supplemented the sample data of the minority class using the SMOTE. In the training data (TR), the number of all minority class samples was increased to obtain 4620 samples. In the test data (TE), the number was increased to 2280. Table 4 shows the changes that were made to the TR and TE using the SMOTE. The new dataset was expressed as a 27,720 × 10 matrix for the new data of TR (NTR) and a 13,680 × 10 matrix for the new data of TE (NTE).

4.2. Principal Component Analysis

Principal component analysis (PCA) is a standard feature extraction method used in pattern recognition [45] that minimizes the correlation between the components by rotating the covariance matrix [46]. We used PCA to extract the principal components of the training samples. The scatter diagram of the PCA result was showed in Figure 4, in which the dimensions were reduced from ten variables to two principal components. The contribution of the principal component 1 (PC1) was 81.72%, and that of principal component 2 (PC2) was 15.53%. In general, the first few principal components whose cumulative contribution exceeds 95% are considered to contain nearly all the information of the original data [45]. Therefore, we believe that PC1 and PC2 are representative of the characteristics of the original data. We selected randomly 18 pears (3 × 6, three pears per category from the training set) to present the PCA results as an example in Figure 5. As shown in Figure 5, the odor information of the six categories of pears has significant overlap, and the categories cannot be visually distinguished in the two-dimensional projection space based on PCA. Accordingly, we concluded that the odor of pears with different levels of browning was rich and strong. Even if the cumulative contribution rate of PC1 and PC2 reaches 97.25%, it was still not easy to distinguish pears with different levels of browning in a low-dimensional space.
The PCA loading plot was shown in Figure 6. Ten vectors represented ten sensors (MOS 1, …, MOS 10) of the PEN3 E-nose. The contribution of sensors is denoted through the direction and length of vector. As shown in Figure 6, MOS 4 has the largest positive coefficient for first principal component so that MOS 4 has the largest contribution to the first principal component during the process of PCA.

4.3. Comparison of the Classification Results of the Four Methods

In this study, the performances of the BPNN and ELM used in the BP-ELMNN were compared with that of the proposed BP-ELMNN model. In addition, the performance of the RBFNN, which has been widely used in many applications [47], was also analyzed. The methods were implemented using a PC (Intel Core i7-8550U processor), and Python and Tensorflow were used for programming.
For the BPNN model, a network structure with three hidden layers (21, 43, and 87 neurons in the hidden layers) was designed to detect the degree of browning of the pears. The activation function of the neurons in the hidden layers was the Relu function, and the neurons in the output layer were activated by the Softmax function. The BPNN model was trained and validated using 10-fold cross-validation on the NTR and achieved a training accuracy of 0.7845. The experimental results of the NTE are shown in Table 5. An accuracy of 0.7623, a macro-precision of 0.6732, a macro-recall of 0.7623, and a macro-F1 score of 0.7150 on NTE were obtained. The BPNN model provided a medium classification accuracy for detecting brown core in the Chinese pears.
The ELM model had 500 neurons in the single hidden layer and was activated by a Sigmoid function. The input weights of the hidden layer and the output layer were obtained randomly and analytically, respectively. As shown in Table 5, the ELM model on the NTE provided the second-best classification result.
In the RBFNN model, a Gaussian function was used as the activation function of the hidden layer neurons. The center of the activation function was obtained via the k-means algorithm. The sum of squared errors (SSE) was calculated as follows:
S S E = i = 1 k d Q i ( d c i ) 2
where Q i is the ith cluster, d is a sample of Q i , c i is the mean of all samples of Q i , k is the number of cluster center. Figure 7 shows the relation curve between the SSE and k; a value of k = 3 was used because the rate of decrease of the SSE was relatively low when k > 3. The width of the activation function was obtained using Equation (11):
w i d t h k = μ m a x k
where μ m a x is the maximum distance between the centers. The Softmax function was utilized to activate the output layer neurons. The classification results of the RBFNN for the NTE is shown in Table 5; this algorithm provided the lowest classification accuracy.
For training the BP-ELMNN, 10-fold cross-validation was used to evaluate the performance of the network on NTR. The curves of validation loss and accuracy during training process are shown in Figure 8. After 8000 epochs, the validation loss did not decrease further, and the accuracy for the NTR was 0.9874, indicating convergence. As shown in Table 5, the BP-ELMNN achieved the highest classification accuracy for the NTE, with an accuracy of 0.9683, macro-precision of 0.9688, macro-recall of 0.9683, and a macro-F1 score of 0.9685.
The experimental results of the four methods are summarized in Table 5; we used four evaluation indicators (accuracy, macro-Precision, macro-recall, and macro-F1 score) to assess the classification performance of the models. The accuracy represents the ratio of the correctly classified samples to the total samples [48]. The precision is the ratio of the true positive results to all positive results [49], and the mean value of the precision of each sample label is the macro-precision. The recall represents the number of positive results that were correctly classified as positive in the experimental result [48], and the mean value of the recall of each sample label is the macro-recall. The macro-F1 score is the harmonic mean of the macro-precision and macro-recall.
The performance of the four neural networks is shown in Figure 9 in a bar chart. The results indicate that the BP-ELMNN had a strong nonlinear fitting ability for the detection of brown core in the Chinese pears.
We also used the four neural networks to classify the data not processed by the SMOTE (the TR and TE) to determine the performance of the BP-ELMNN framework when unbalanced data are used. The classification results for the original data of the TE (OTE) are shown in Table 6. Unbalanced data often results in an inability of the neural network to learn the information of the different types of data and lower classification accuracy. However, Table 6 indicates that the proposed BP-ELMNN framework also provides the best performance for the unbalanced data, with an accuracy of 0.9285, macro-precision of 0.9188, macro-recall of 0.9017, and a macro-F1 score of 0.9089.

5. Conclusions

In this paper, we proposed a multi-layer BP-ELMNN framework to detect brown core in Chinese pears (Huangguan) using a MOS-based E-nose device. PCA was used to extract the principal components, whose cumulative contribution was higher than 95%. However, pears with different degrees of browning could not be distinguished by the PCA. The result of PCA showed that the odors of pears with different degrees of browning were indistinguishable in two-dimensional space (Even if their cumulative contribution rate reached 97.25%) and highly similar. The proposed BP-ELMNN framework obtained the highest classification accuracy among the four neural networks, indicating that the hidden layer structure of the BPNN extracted sufficient data, which were used as input for the ELM. The original data that were not processed by the SMOTE were used in a comparative experiment to demonstrate the robustness of the BP-ELMNN framework to unbalanced data.This study demonstrated that the use of the E-nose and a suitable ANN algorithm provided excellent performance for the non-destructive detection of brown core in the Huangguan Chinese pear. The findings provide insights into the use of ANNs in the food industry, helps manufacturers monitor pear quality, and promotes the practical application of E-nose devices.

Author Contributions

Both authors contributed extensively to the study presented in this manuscript. Y.G. contributed significantly to the conception of the study. H.W. coded the algorithm and performed the experiments. Y.G. supervised the work and contributed with valuable discussions and scientific advice. Both authors contributed to writing this manuscript. All authors have read and agreed to the published version of the manuscript

Funding

The authors would like to thank the Ministry of Science and Technology of the People’s Republic of China (Grant No. 2017YFB1400100) and the National Natural Science Foundation of China (Grant No. 61876059) for their support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Saquet, A. Storage of pears. Sci. Hortic. 2019, 246, 1009–1016. [Google Scholar] [CrossRef]
  2. EFSA. The EFSA Comprehensive European Food Consumption Database; EFSA: Parma, Italy, 2011. [Google Scholar]
  3. Li, X.; Wang, T.; Zhou, B.; Gao, W.-Y.; Cao, J.; Huang, L.-Q. Chemical composition and antioxidant and anti-inflammatory potential of peels and flesh from 10 different pear varieties (Pyrus spp.). Food Chem. 2014, 152, 531–538. [Google Scholar] [CrossRef] [PubMed]
  4. Ulaszewska, M.; Vázquez-Manjarrez, N.; Garcia-Aloy, M.; Llorach, R.; Mattivi, F.; Dragsted, L.O.; Praticò, G.; Manach, C. Food intake biomarkers for apple, pear, and stone fruit. Genes Nutr. 2018, 13, 29. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Dong, Y.; Guan, J.; Ma, S.-J.; Liu, L.-L.; Feng, Y.-X.; Cheng, Y. Calcium content and its correlated distribution with skin browning spot in bagged Huangguan pear. Protoplasma 2015, 252, 165–171. [Google Scholar] [CrossRef]
  6. Wei, C.; Ma, L.; Cheng, Y.; Guan, Y.; Guan, J. Exogenous ethylene alleviates chilling injury of ‘Huangguan’ pear by enhancing the proline content and antioxidant activity. Sci. Hortic. 2019, 257, 108671. [Google Scholar] [CrossRef]
  7. Xu, F.; Dong, S.; Xu, Q.; Liu, S. Control of brown heart in Huangguan pears with 1-methylcyclopropene microbubbles treatment. Sci. Hortic. 2020, 259, 108820. [Google Scholar] [CrossRef]
  8. Xu, F.; Zhang, K.; Liu, S. Evaluation of 1-methylcyclopropene (1-MCP) and low temperature conditioning (LTC) to control brown of Huangguan pears. Sci. Hortic. 2020, 259, 108738. [Google Scholar] [CrossRef]
  9. Cheng, Y.; Liu, L.; Zhao, G.; Shen, C.; Yan, H.; Guan, J.; Yang, K. The effects of modified atmosphere packaging on core browning and the expression patterns of PPO and PAL genes in ‘Yali’ pears during cold storage. LWT Food Sci. Technol. 2015, 60, 1243–1248. [Google Scholar] [CrossRef]
  10. Lin, Y.; Lin, Y.; Lin, H.; Chen, Y.; Wang, H.; Shi, J. Application of propyl gallate alleviates pericarp browning in harvested longan fruit by modulating metabolisms of respiration and energy. Food Chem. 2018, 240, 863–869. [Google Scholar] [CrossRef]
  11. Xu, F.; Liu, S.; Liu, Y.; Xu, Q.; Wang, S. The combined effect of ultraviolet-C irradiation and lysozyme coatings treatment on control of brown heart in Huangguan pears. Sci. Hortic. 2019, 256, 108634. [Google Scholar] [CrossRef]
  12. Li, J.; Bao, X.; Xu, Y.; Zhang, M.; Cai, Q.; Li, L.; Wang, Y. Hypobaric storage reduced core browning of Yali pear fruits. Sci. Hortic. 2017, 225, 547–552. [Google Scholar] [CrossRef]
  13. Gliszczyńska-Świgło, A.; Chmielewski, J. Electronic Nose as a Tool for Monitoring the Authenticity of Food. A Review. Food Anal. Methods 2017, 10, 1800–1816. [Google Scholar] [CrossRef] [Green Version]
  14. Qiu, S.; Wang, J. The prediction of food additives in the fruit juice based on electronic nose with chemometrics. Food Chem. 2017, 230, 208–214. [Google Scholar] [CrossRef] [PubMed]
  15. Ezhilan, M.; Nesakumar, N.; Babu, K.J.; Srinandan, C.S.; Rayappan, J.B.B. An Electronic Nose for Royal Delicious Apple Quality Assessment—A Tri-layer Approach. Food Res. Int. 2018, 109, 44–51. [Google Scholar] [CrossRef]
  16. Li, Q.; Yu, X.; Xu, L.; Gao, J.-M. Novel method for the producing area identification of Zhongning Goji berries by electronic nose. Food Chem. 2017, 221, 1113–1119. [Google Scholar] [CrossRef]
  17. Voss, H.G.J.; Stevan, S.L., Jr.; Ayub, R.A. Peach growth cycle monitoring using an electronic nose. Comput. Electron. Agric. 2019, 163, 104858. [Google Scholar] [CrossRef]
  18. Feng, L.; Zhang, M.; Bhandari, B.; Guo, Z. A novel method using MOS electronic nose and ELM for predicting postharvest quality of cherry tomato fruit treated with high pressure argon. Comput. Electron. Agric. 2018, 154, 411–419. [Google Scholar] [CrossRef]
  19. Chen, Q.; Song, J.; Bi, J.; Meng, X.; Wu, X. Characterization of volatile profile from ten different varieties of Chinese jujubes by HS-SPME/GC–MS coupled with E-nose. Food Res. Int. 2018, 105, 605–615. [Google Scholar] [CrossRef]
  20. Jordan, M.I.; Mitchell, T.M. Machine learning: Trends, perspectives, and prospects. Science 2015, 349, 255–260. [Google Scholar] [CrossRef]
  21. Qu, K.; Guo, F.; Liu, X.; Lin, Y.; Zou, Q. Application of Machine Learning in Microbiology. Front. Microbiol. 2019, 10, 827. [Google Scholar] [CrossRef]
  22. Zhong, H.; Miao, C.; Shen, Z.; Feng, Y. Comparing the learning effectiveness of BP, ELM, I-ELM, and SVM for corporate credit ratings. Neurocomputing 2014, 128, 285–295. [Google Scholar] [CrossRef]
  23. Marini, F. Artificial neural networks in foodstuff analyses: Trends and perspectives A review. Anal. Chim. Acta 2009, 635, 121–131. [Google Scholar] [CrossRef] [PubMed]
  24. Deng, Y.; Sander, A.; Faulstich, L.; Denecke, K. Towards automatic encoding of medical procedures using convolutional neural networks and autoencoders. Artif. Intell. Med. 2018, 93, 29–42. [Google Scholar] [CrossRef]
  25. Hotel, O.; Poli, J.-P.; Mer-Calfati, C.; Scorsone, E.; Saada, S. A review of algorithms for SAW sensors e-nose based volatile compound identification. Sens. Actuators B Chem. 2018, 255, 2472–2482. [Google Scholar] [CrossRef]
  26. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  27. Chawla, N.; Bowyer, K.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  28. Portable Electronic Nose. Available online: https://airsense.com/en/products/portable-electronic-nose (accessed on 28 June 2020).
  29. Duchi, J.; Hazan, E.; Singer, Y. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  30. Tieleman, T.; Hinton, G. Lecture 6.5-RMSProp, COURSERA: Neural Networks for Machine Learning 2012. Available online: https://www.cs.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf (accessed on 30 June 2020).
  31. Yu, Y.; Liu, F. Effective Neural Network Training with a New Weighting Mechanism-Based Optimization Algorithm. IEEE Access 2019, 7, 72403–72410. [Google Scholar] [CrossRef]
  32. Yang, Y.; Duan, J.; Yu, H.; Gao, Z.; Qiu, X. An Image Classification Method Based on Deep Neural Network with Energy Model. CMES-Comput. Model. Eng. Sci. 2018, 117, 555–575. [Google Scholar] [CrossRef] [Green Version]
  33. An, Y.; Wang, X.; Chu, R.; Yue, B.; Wu, L.; Cui, J.; Qu, Z. Event classification for natural gas pipeline safety monitoring based on long short-term memory network and Adam algorithm. Struct. Health Monit. 2019. [Google Scholar] [CrossRef]
  34. Raghuwanshi, B.S.; Shukla, S. SMOTE based class-specific extreme learning machine for imbalanced learning. Knowl.-Based Syst. 2020, 187, 104814. [Google Scholar] [CrossRef]
  35. Yu, D.; Wang, X.; Liu, H.; Gu, Y. A Multitask Learning Framework for Multi-Property Detection of Wine. IEEE Access 2019, 7, 123151–123157. [Google Scholar] [CrossRef]
  36. Sun, X.; Wan, Y.; Guo, R. Chinese wine classification using BPNN through combination of micrographs’ shape and structure features. In Proceedings of the 2009 Fifth International Conference on Natural Computation, Tianjin, China, 14–16 August 2009; Volume 6, pp. 379–384. [Google Scholar]
  37. Cheng, S.; Zhou, X. Network traffic prediction based on BPNN optimized by self-adaptive immune genetic algorithm. In Proceedings of the 2013 International Conference on Mechatronic Sciences, Electric Engineering and Computer, Shengyang, China, 20–22 December 2013; pp. 1030–1033. [Google Scholar]
  38. Tang, H.; Lei, M.; Gong, Q.; Wang, J. A BP Neural Network Recommendation Algorithm Based on Cloud Model. IEEE Access 2019, 7, 35898–35907. [Google Scholar] [CrossRef]
  39. Huang, G.-B.; Zhou, H.; Ding, X.; Zhang, R. Extreme Learning Machine for Regression and Multiclass Classification. IEEE Trans. Syst. Man Cybern. Part B 2011, 42, 513–529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Zhang, N.; Ding, S.; Zhang, J. Multi layer ELM-RBF for multi-label learning. Appl. Soft Comput. 2016, 43, 535–545. [Google Scholar] [CrossRef]
  41. Yu, J.; Zhan, J.; Huang, W. Identification of Wine According to Grape Variety Using Near-Infrared Spectroscopy Based on Radial Basis Function Neural Networks and Least-Squares Support Vector Machines. Food Anal. Methods 2017, 75, 3311–3460. [Google Scholar] [CrossRef]
  42. Yang, X.; Li, Y.; Sun, Y.; Long, T.; Sarkar, T.K. Fast and Robust RBF Neural Network Based on Global K-Means Clustering with Adaptive Selection Radius for Sound Source Angle Estimation. IEEE Trans. Antennas Propag. 2018, 66, 3097–3107. [Google Scholar] [CrossRef]
  43. Liu, Y.; Huang, H.; Huang, T.; Qian, X. An improved maximum spread algorithm with application to complex-valued RBF neural networks. Neurocomputing 2016, 216, 261–267. [Google Scholar] [CrossRef]
  44. Chen, S.; Hong, X.; Harris, C.J.; Hanzo, L. Fully complex-valued radial basis function networks: Orthogonal least squares regression and classification. Neurocomputing 2008, 71, 3421–3433. [Google Scholar] [CrossRef]
  45. Liu, H.; Li, Q.; Yan, B.; Zhang, L.; Gu, Y. Bionic Electronic Nose Based on MOS Sensors Array and Machine Learning Algorithms Used for Wine Properties Detection. Sensors 2018, 19, 45. [Google Scholar] [CrossRef] [Green Version]
  46. Azarbad, M.; Azami, H.; Sanei, S.; Ebrahimzadeh, A. New Neural Network-based Approaches for GPS GDOP Classification based on Neuro-Fuzzy Inference System, Radial Basis Function, and Improved Bee Algorithm. Appl. Soft Comput. 2014, 25, 285–292. [Google Scholar] [CrossRef]
  47. Rouhani, M.; Javan, D.S. Two fast and accurate heuristic RBF learning rules for data classification. Neural Netw. 2016, 75, 150–161. [Google Scholar] [CrossRef] [PubMed]
  48. Lei, Y.; Dogan, U.; Zhou, D.-X.; Kloft, M. Data-Dependent Generalization Bounds for Multi-Class Classification. IEEE Trans. Inf. Theory 2019, 65, 2995–3021. [Google Scholar] [CrossRef]
  49. Patil, T.R.; Sherekar, S.S. Performance Analysis of Naive Bayes and J48 Classification Algorithm for Data Classification. Int. J. Comput. Sci. Appl. 2013, 6, 256–261. [Google Scholar]
Figure 1. Schematic diagram of the sample collection using E-nose.
Figure 1. Schematic diagram of the sample collection using E-nose.
Sensors 20 04499 g001
Figure 2. An example of the response curves of the E-nose for pears during 90 s of collecting stage.
Figure 2. An example of the response curves of the E-nose for pears during 90 s of collecting stage.
Sensors 20 04499 g002
Figure 3. Photos of the Chinese pear (Huangguan) with different browning levels. Photos (af) indicates the levels of pear browning from class 0 to class 5, respectively.
Figure 3. Photos of the Chinese pear (Huangguan) with different browning levels. Photos (af) indicates the levels of pear browning from class 0 to class 5, respectively.
Sensors 20 04499 g003
Figure 4. The proposed BP-ELMNN structure for the discrimination of the categories of pears.
Figure 4. The proposed BP-ELMNN structure for the discrimination of the categories of pears.
Sensors 20 04499 g004
Figure 5. Projections of the two first primary components of the PCA computed on training set to determine the levels of pear browning. The horizontal axis represents the principal component 1 with 81.72% contribution rate and the vertical axis represents the principal component 2 with 15.53%.
Figure 5. Projections of the two first primary components of the PCA computed on training set to determine the levels of pear browning. The horizontal axis represents the principal component 1 with 81.72% contribution rate and the vertical axis represents the principal component 2 with 15.53%.
Sensors 20 04499 g005
Figure 6. PCA loading plot of measurements about different levels of pears browning.
Figure 6. PCA loading plot of measurements about different levels of pears browning.
Sensors 20 04499 g006
Figure 7. The relation curve between the SSE and k.
Figure 7. The relation curve between the SSE and k.
Sensors 20 04499 g007
Figure 8. The loss and cross validation curves of the BP-ELMNN during training process.
Figure 8. The loss and cross validation curves of the BP-ELMNN during training process.
Sensors 20 04499 g008
Figure 9. Bar chart of the performance of the four methods (accuracy, macro-precision, macro-recall, and macro-F1 score) for the NTE.
Figure 9. Bar chart of the performance of the four methods (accuracy, macro-precision, macro-recall, and macro-F1 score) for the NTE.
Sensors 20 04499 g009
Table 1. Browning levels of the Chinese pear Huangguan.
Table 1. Browning levels of the Chinese pear Huangguan.
ClassStandard*
AppearanceCoreFlesh
0GoodGoodGood-
1GoodLight brownGoodNo
2GoodBrownGoodNo
3GoodDark brownGoodNo
4BrownDark brownDark brownYes
5Dark brownDark brownDark brownYes
* denotes whether browning can be discerned by the appearance.
Table 2. The sensor array of the PEN3 electronic nose (E-nose).
Table 2. The sensor array of the PEN3 electronic nose (E-nose).
NumberSensorSubstance Sensitivity
MOS1W1CAroma constituent
MOS2W5SSensitive to nitride oxides
MOS3W3CAmmonia, aroma constituent
MOS4W6SHydrogen
MOS5W5CAlkane, aroma constituent
MOS6W1SSensitive to methane
MOS7W1WSensitive to sulfide
MOS8W2SSensitive to alcohol
MOS9W2WAroma constituent, organic sulfur compounds
MOS10W3SSensitive to alkane
Table 3. Label result and number of samples.
Table 3. Label result and number of samples.
Training Samples.Test Samples
ClassNumberClassNumber
0903
12118
242215
377338
449420
542518
Table 4. The changes that were made to the training data (TR) and test data (TE).
Table 4. The changes that were made to the training data (TR) and test data (TE).
ClassTRTE
The Original Number of SamplesAdded Number of SamplesNew Number of SamplesThe Original Number of SamplesAdded Number of SamplesNew Number of Samples
05404080462018021002280
112603360462048018002280
225202100462090013802280
3462004620228002280
4294016804620120010802280
5252021004620108012002280
Table 5. The accuracy, macro-precision, macro-recall, and macro-F1 score of the four methods for the NTE.
Table 5. The accuracy, macro-precision, macro-recall, and macro-F1 score of the four methods for the NTE.
MethodsAccuracyMacro-PrecisionMacro-RecallMacro-F1 Score
BPNN0.76230.67320.76230.7150
RBFNN0.35040.23470.35040.2811
ELM0.91900.92720.91900.9231
BP-ELMNN0.96830.96880.96830.9685
Table 6. The accuracy, macro-precision, macro-recall, and macro-F1 score of the four methods for the OTE.
Table 6. The accuracy, macro-precision, macro-recall, and macro-F1 score of the four methods for the OTE.
MethodsAccuracyMacro-PrecisionMacro-RecallMacro-F1score
BPNN0.73990.63980.70550.6711
RBFNN0.34930.12550.21530.1586
ELM0.88090.86530.86770.8665
BP-ELMNN0.92850.91880.90170.9089

Share and Cite

MDPI and ACS Style

Wei, H.; Gu, Y. A Machine Learning Method for the Detection of Brown Core in the Chinese Pear Variety Huangguan Using a MOS-Based E-Nose. Sensors 2020, 20, 4499. https://doi.org/10.3390/s20164499

AMA Style

Wei H, Gu Y. A Machine Learning Method for the Detection of Brown Core in the Chinese Pear Variety Huangguan Using a MOS-Based E-Nose. Sensors. 2020; 20(16):4499. https://doi.org/10.3390/s20164499

Chicago/Turabian Style

Wei, Hao, and Yu Gu. 2020. "A Machine Learning Method for the Detection of Brown Core in the Chinese Pear Variety Huangguan Using a MOS-Based E-Nose" Sensors 20, no. 16: 4499. https://doi.org/10.3390/s20164499

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop