Next Article in Journal
Identification of Eight High Yielding Strains via Morpho-Molecular Characterization of Thirty-Three Wild Strains of Calocybe indica
Previous Article in Journal
Assessing the Impact of Roasting Temperatures on Biochemical and Sensory Quality of Macadamia Nuts (Macadamia integrifolia)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EL V.2 Model for Predicting Food Safety Risks at Taiwan Border Using the Voting-Based Ensemble Method

1
Food and Drug Administration, Ministry of Welfare, Taipei 115209, Taiwan
2
Department of Information and Finance Management, National Taipei University of Technology, Taipei 10608, Taiwan
*
Author to whom correspondence should be addressed.
Foods 2023, 12(11), 2118; https://doi.org/10.3390/foods12112118
Submission received: 25 April 2023 / Revised: 17 May 2023 / Accepted: 23 May 2023 / Published: 24 May 2023
(This article belongs to the Section Food Security and Sustainability)

Abstract

:
Border management serves as a crucial control checkpoint for governments to regulate the quality and safety of imported food. In 2020, the first-generation ensemble learning prediction model (EL V.1) was introduced to Taiwan’s border food management. This model primarily assesses the risk of imported food by combining five algorithms to determine whether quality sampling should be performed on imported food at the border. In this study, a second-generation ensemble learning prediction model (EL V.2) was developed based on seven algorithms to enhance the “detection rate of unqualified cases” and improve the robustness of the model. In this study, Elastic Net was used to select the characteristic risk factors. Two algorithms were used to construct the new model: The Bagging-Gradient Boosting Machine and Bagging-Elastic Net. In addition, Fβ was used to flexibly control the sampling rate, improving the predictive performance and robustness of the model. The chi-square test was employed to compare the efficacy of “pre-launch (2019) random sampling inspection” and “post-launch (2020–2022) model prediction sampling inspection”. For cases recommended for inspection by the ensemble learning model and subsequently inspected, the unqualified rates were 5.10%, 6.36%, and 4.39% in 2020, 2021, and 2022, respectively, which were significantly higher (p < 0.001) compared with the random sampling rate of 2.09% in 2019. The prediction indices established by the confusion matrix were used to further evaluate the prediction effects of EL V.1 and EL V.2, and the EL V.2 model exhibited superior predictive performance compared with EL V.1, and both models outperformed random sampling.

1. Introduction

Taiwan’s food supply relies heavily on imports, with a vast array of imported ingredients and products comprising a substantial portion of the population’s dietary consumption. This underscores the importance of managing imported food to protect public health and consumer rights. In Taiwan, the number of inspection applications for imported food has grown annually. Between 2011 and 2022, inspection applications increased from 419,000 batches to 723,000 batches, nearly doubling. Given the substantial volume of food imports, conducting border sampling inspections is of great significance for effectively strengthening control over high-risk products and accurately detecting substandard items.
Food risk management and control at Taiwan’s border employ a food inspection method, which can be primarily classified into two categories: review and inspection. The review is conducted in writing, comparing customs clearance data with product information. The inspection involves sampling selected batches and sending them to authorized inspection laboratories for pesticide, pigment, or heavy metal compound testing. The entire process can be completed in approximately three to seven days. According to Taiwan’s border inspection measures, inspection methods can be classified into general inspection, enhanced inspection, and batch-by-batch inspection. Generally, only 2 to 10% of the products are sampled for random inspection. However, if a single non-compliant item is detected for the same inspection applicant, origin, and product, the next import will be subject to enhanced inspection. Once an inspection application batch is designated for enhanced sampling, the random inspection method is still used but requires 20 to 50% sampling. If violations are detected again, 100% batch-by-batch inspection will be implemented [1].
The maintenance of imported food quality at the border primarily relies on the accurate detection of products that do not satisfy quality standards during sampling inspections, thereby preventing their importation. In 2020, Taiwan developed a first-generation ensemble learning prediction model (hereinafter referred to as EL V.1) for border management to identify high-risk products. Five algorithms were primarily utilized to predict the risk of products for inspection, including Decision Tree C5.0 and CART, Random Forest (RF), Logistic Regression (LR), and Naïve Bayes (NB). The detection rate of unqualified products via sampling inspection was significantly increased.
To further improve the detection rate and enhance the model’s robustness, this study aimed to construct the second-generation ensemble learning prediction model (hereinafter referred to as EL V.2). By refining the screening method for key risk factors and incorporating additional classification algorithms required for the modeling process (including Elastic Net (EN) and Gradient Boosting Machine (GBM)), the robustness of model prediction can be enhanced. With the assistance of Taiwan Food Cloud Big Data and seven machine learning algorithms for ensemble learning, the objective is to further improve the detection rate of unqualified products sampled for inspection, thereby ensuring the food safety of the population.
This study contributed to risk prediction for imported food controls at the border in several ways. First, the machine learning model for predicting the risk of imported food at the border was constructed using over ten years of data from Taiwan’s real-time food cloud while incorporating characteristic risk factors for data screening. Second, the classifier built through ensemble learning combined several classification algorithms, providing a robust prediction method and addressing the classification bias potentially occurring when using a single algorithm. Third, unqualified cases were more easily identified in border inspections when using the proposed model in view of its increased number of algorithms, suitable characteristic risk factors, and flexible adjustment of the sampling rate with Fβ Fourth; it was found that the risk prediction model can help inspectors reduce pressure at work and overcome importer doubts regarding results from human inspections. Finally, EL V. 2 outperformed EL V.1 in predicting unqualified imported food, offering greater assurance for food safety control. (For a detailed comparison table between the abbreviations and full names of this article, please refer to Table A1 in the Appendix A.)

2. Literature Review

This study aims to construct a more robust risk prediction model than EL V.1, namely EV.2. To help achieve this, this section first discusses the collection of “characteristic risk factors” required for modeling. Secondly, the literature on the “selection of algorithms” is reviewed to understand the common algorithms used in the field of food. Furthermore, experts were invited to decide the algorithms to be used in the model by consensus. Finally, insights from the literature on “improvement of ensemble learning methods” were collected as references for the method design in this study.

2.1. Model Characteristic Risk Factors

The international application of big data in the field of food safety encompasses food safety-related monitoring, such as monitoring food additives, animal drug residues, heavy metals, allergens, and foodborne diseases, as well as providing early warnings for production, supply, and sales of products, food adulteration and fraud, and food safety incidents. The collection and integration of data can assist in the risk analysis and management of food, raw materials, and feed [2,3,4,5,6,7,8,9,10]. In 2016, Marvin et al. proposed that the Bayesian network algorithm can handle diverse big data and facilitate the understanding of driving factors related to food safety via systematic analysis, such as the impact of climate change on food quality, economy, and human behavior. Combined with the data, this algorithm can be used to predict possible food safety risk events [11]. In 2015, Bouzembrak et al. used the Rapid Alert System for Food and Feed (RASFF) of the European Union to construct a Bayesian network model to predict the types of food fraud that can occur in imported products of known food product categories and countries of origin. The findings can assist in border risk management and control and serve as an important reference for EU governments in conducting inspections and law enforcement [2,12,13].
The amount of imported food in the United States is increasing year by year. Due to limited inspection capacity, the Food and Drug Administration has divided the control of border imported food into two stages. The first stage is mainly electronic document review, with only 1% of imported food actually inspected each year. The second stage involves using the Predictive Risk-based Evaluation for Dynamic Import Compliance Targeting (PREDICT) system for risk prediction. Big data are employed to collect relevant data from products and manufacturers for evaluation, determining the risk level of imported goods. The risk factors calculated in the PREDICT system include at least four types of data, such as product risk (epidemic outbreak, recall, or adverse event), regulatory risk (specific factors of the manufacturer itself and past compliance with food safety regulations), factory inspection records of the manufacturer within three years, and historical data of the customs broker (quality analysis of data provided by the customs broker or importer within one year, such as reporting status). These data are used to screen factors related to the product itself for risk score calculation and further propose whether to conduct product sampling inspection [14].
The data sources used by the PREDICT system are mainly import alert and import notification data, domestic inspection and product tracking records, foreign factory inspections (such as equipment inspections), and identification system evaluation. Using these data, the PREDICT system can conduct data mining and analysis, enabling it to use artificial intelligence methods to predict the possible risks of imported goods and intercept them in a timely manner. This approach is undoubtedly the best for countries facing massive imports each year, which need to maintain normal export and import while still taking into account the safety and quality of goods.
Regarding the quality sampling inspection of imported food at the border, there are currently the following international experiences: The United States employs machine learning to assist in border inspection operations, while the European Union deploys methods such as Bayesian network analysis to predict factors that may cause border food risks, and then reports back to EU countries to strengthen their attention to import control. These practices demonstrate that big data applications, such as artificial intelligence and machine learning, can provide better operational quality for government border management and ensure the health and safety of the public. Therefore, this study referred to the data sources and practices of the European Union and the United States to collect risk factors and establish prediction model planning.

2.2. Selection of Algorithms

In recent years, ensemble learning has received great attention from researchers and has been widely applied in many fields for various purposes, such as medical diagnosis and disease prediction [15,16,17,18], Improvement of patient quality of life [19], Internet of Things (IoT) security [20,21], fault detection and error prediction for industrial processes [22,23,24], advertising and marketing [25], as well as agricultural monitoring, management, and productivity improvement [6,26]. In the food industry, it has been used for productivity improvement in food manufacturing, quality assessment and monitoring, food ingredient identification, food safety, and the quality of food delivery services (FDS). Parastar [10] developed a handheld near-infrared spectroscopy device based on ensemble learning for measuring and monitoring the authenticity of chicken meat that showed better performance in authenticity testing than common single classification methods such as partial least squares-discriminant analysis (PLS-DA), artificial neural network (ANN) and support vector machine (SVM). Using a combination of deep learning and ensemble learning techniques on milk spectral data, Neto [6] proposed a method for predicting common fraudulent milk adulterations in the dairy industry. Their method outperformed not only common statistical learning methods but also the Fourier transformed infrared spectroscopy (FTIR), which is typically used for identifying the composition of a sample in the dairy industry. Further, Adak [27] constructed a model with customer reviews of FDS using machine learning and deep learning techniques to predict customer sentiment about FDS. Based on previous studies and following consultation with experts, the following algorithms were used for constructing the new model: Decision Tree C5.0 and CART, Random Forest (RF), Logistic Regression (LR), Naïve Bayes (NB), Elastic Net (EN), and Gradient Boosting Machine (GBM). These algorithms offer interpretable approaches that are easy to understand by users, so they were adopted in ensemble learning for EL V.1 and EL V.2. Deep learning has not been included, given its low interpretability, but it may be considered in subsequent studies.
EL V.1 was constructed using five algorithms: Decision Tree C5.0 and CART, Random Forest (RF), Logistic Regression (LR), and Naïve Bayes (NB). These algorithms exhibit great interpretability and explain ability, so they were primarily used for prediction tasks with ensemble-based classification techniques. On the basis of EL V.1, this study intends to construct a model with higher predictive performance and greater computational efficiency. To reduce computation time, Elastic Net (EN) and Gradient Boosting Machine (GBM) were used to control the sampling decision within one minute for each batch of cases. The test results revealed that the computation time could be controlled within the limit using the EN and GBM. Therefore, they were integrated into the construction of EL V.2.

2.3. Improvement of Ensemble Learning Model

The ensemble learning model is jointly established by a group of independent machine learning classifiers, combines their respective prediction results, and implements an integration strategy to reduce the total error and improve the performance of a single classifier [28,29,30]. Each classifier may have different generalization capabilities, i.e., different inference abilities for various samples, similar to the opinions of different experts. Finally, combining the output of these individual classifiers can deliver the final classification results, significantly reducing the probability of classification errors in the results [9,30].
For example, Solano [31] proposed an ensemble voting model for solar radiation prediction based on machine learning algorithms. The results of the study show that the weighted average voting method based on random forest and classification boosting has superior performance and is also better than a single machine learning algorithm and other ensemble models. Chandrasekhar [32] used six algorithms (Random Forest, K-Nearest Neighbors, Logistic Regression, Naive Bayes, Gradient Boosting, and AdaBoost Classifier) for voting ensemble learning, which improved the accuracy of heart disease prediction. Alsulami [33] proposed a data mining model including three traditional algorithms (decision trees, Naive Bays, and random forests) to evaluate student e-learning data to help policy makers make informed and appropriate decisions for their institutions. These methods effectively improve model prediction performance by using three ensemble techniques, including bagging, boosting, and voting. The combination of multiple different classifiers has been proven to improve the classification accuracy of the overall classification system [34,35,36,37].
In this study, four methods proposed by scholars were utilized to enhance the diversity of classification models (or classifiers) within the ensemble learning model, including the use of different training datasets and training of different classification models with different parameter settings, algorithms, and characteristic factors [31,38]. In previous studies, five algorithms were used to construct the ensemble learning model EL V.1. To improve and stabilize the predictive performance of the model, in this study, an attempt was made to construct model EL V.2 by adding “algorithmic classification models”, adjusting the “factor screening method”, and adding “sampling rate control parameters” such that the prediction method of imported food sampling inspection at the border can play a better role. Therefore, in addition to the algorithms used in the first-generation ensemble learning model EL V.1 constructed in previous studies (including Decision Tree C5.0 and CART, Random Forest (RF), Logistic Regression (LR), and Naïve Bayes (NB)), the newly added algorithms in this study were Elastic Net (EN) and Gradient Boosting Machine (GBM). The aforementioned seven algorithms, combined with the classification model constructed by the bagging method, will use the integration method for strategic integration with the “majority decision” approach. After completing the model construction, the prediction of border inspection applications will be conducted.

3. Materials and Methods

To improve the robustness of EL V.2, improvements were made on the basis of EL V.1 with a more refined method for selecting characteristic risk factors and an increased number of algorithms for classification. The details of the research methodology are described in the following sections.

3.1. Data Sources and Analytical Tools

The modeling data for this study were sourced from the food cloud established by the Food and Drug Administration of the Ministry of Health and Welfare of Taiwan. The food cloud is centered around the Food and Drug Administration’s Five Systems, including the Registration Platform of Food Businesses System (RPFBS), the Food Traceability Management System (FTMS), the Inspection Management System (IMS), the Product Management Decision System (PMDS), and the Import Food Information System (IFIS). Additionally, it comprises cross-agency data communication, including financial and tax electronic invoices, customs electronic gate verification data, national business tax registration data, industrial and commercial registration data, indicated chemical substance flow data, domestic industrial oil flow data, imported industrial flow data, waste oil flow data, toxic chemical substance flow data, feed oil flow data, and campus food ingredient login and inspection data [39]. After imported food enters Taiwan, it must be declared and inspected through IFIS. Only after approval can the imported food enter the domestic market. The relevant business data must be registered in RPFBS, national business tax registration data, and business registration data. The flow information generated by domestic and imported products entering the market from the border should be recorded in IFIS and FTMS, as well as in electronic invoices and electronic gate goods import and export verification records. All government-conducted product sampling inspection records should be saved in PMDS, IFIS, and IMS. Information related to the company’s products can also be accessed via RPFBS and FTMS.
The main sources of this study were border inspection application data, food inspection data, food product flow information, and business registration data from Taiwan’s food cloud, as well as international open data databases related to food safety, including gross domestic product (GDP), GDP growth rate, global food security index, corruption perceptions index (CPI), human development index (HDI), legal rights index (LRI), and regional political risk index. A total of 168 factors were included in the analysis. The analytical tools used in the study were R 3.5.3, SPSS 25.0, and Microsoft Excel 2010.

3.2. Research Methodology

In this study, we selected food inspection application data of S-type products that had been sampled and had inspection results as the research scope. The data were divided into training, validation, and testing sets. First, different data types and analysis methods of the training set were considered to establish various models. The optimal model was selected from the prediction results obtained by importing validation set data into the model. The selected optimal model was further imported into the test set for model validation and effectiveness evaluation and confirmation, completing the construction of EL V.2.
The entire modeling process was based on previous studies on the construction of the EL V.1 method, and improvements were made to this method to aid in improving the hit rate of unqualified products detected via sampling inspection. According to the execution order, this study can be divided into four stages: “data collection”, “data integration and pre-processing”, “establishing risk prediction models”, and “evaluating prediction effectiveness”. “Establishing risk prediction models” included three procedures: “characteristic factor extraction”, “data mining and modeling”, and “establishing the optimum prediction model”. Changes were made in the calculation methods of “characteristic factor extraction” and “data mining”, as shown below: (Figure 1).

3.2.1. Data Collection

The data in this study included the border inspection application database, inspection database, flow direction database, and registration database of Taiwan Food Cloud, as well as open information related to international food risk. (as shown in Table 1) A total of 168 factors were used as the main data source for constructing the risk prediction model (as shown in Table 1).

3.2.2. Integration and Data Pre-Processing

In addition to data noise cleaning, the data needed to be subjected to manufacturer name and product name attribution and data string filing to further integrate the data in accordance with six aspects: manufacturer, importer, customs broker, border inspection, product, and country of manufacture. The integration process included data cleaning, error correction, and attribution.

3.2.3. Establishment of Risk Prediction Model

  • Data processing:
This step required data segmentation by year to prepare training, validation, and test sets. The training set was divided into two forms: 2011–2017 and 2016–2017. The validation set was data from 2018, and the test set was data from 2019. To realize accurate model prediction, in this study, we first attempted to model these two data forms and then used the validation set to confirm the most suitable time interval for data modeling.
  • Selection of characteristic risk factors:
This step was to improve the first-generation model of EL V.1. There were two strategies for extracting characteristic factors. First, the “single-factor analysis” and “stepwise regression”, used to extract characteristic factors in EL V.1, were changed to Elastic Net. Specifically, Elastic Net is a combination of Lasso regression (i.e., L1 normalization) and Ridge regression (i.e., L2 normalization). The equations are as follows: (e.g., Equations (1)–(3))
Lasso regression:
m i n i = 1 n V f x i , y i + λ j = 1 p β
Ridge regression:
m i n i = 1 n V f x i , y i + λ j = 1 p β j 2
Elastic Nets:
m i n i = 1 n V f x i , y i + λ 1 j = 1 p β j + λ 2 j = 1 p β j 2
Lasso regression can aid Elastic Net in selecting characteristic factors. When selecting variable factors, Lasso regression retains only one highly collinear variable, making it the best choice. Ridge regression filters the independent variables into separate groups such that highly collinear variables can exist in the model when they have an effect on dependent variables as opposed to retaining only one of them, like in Lasso regression. Ogutu et al. indicated that due to its own characteristics, Elastic Net would try its best to discard variables within the model that have no influence on the independent variables, which can improve the explanatory power and predictive capability of the model. Relatively speaking, if all highly collinear independent variable factors are retained, the prediction performance of the model may not be increased, and the model will become more complex and unstable [40]. In this study, there were many factors. Hence, there were doubts about high collinearity. To avoid the problem of collinearity among factors that may be ignored when using “single-factor analysis and stepwise regression” to select factors in the past, Elastic Net was selected to reduce the possible bias of the prediction model and improve the accuracy of prediction.
The second strategy involved modeling based on inspection data from 2011 to 2017. Monthly data from January to October 2018 were added over time. The model was updated once a month, and the number of characteristic factors used was calculated. With seven algorithms, each factor can be used up to 70 times. The factor that was used more than once was kept and included in the model required for EL V.2 construction. In this study, a total of 68 characteristic risk factors were obtained (as shown in Table 2), which were important characteristic factors that participated in EL V.2 modeling.
  • Data exploration and modeling
In this study, we conducted modeling based on the training set. In addition to the algorithms used in EL V.1 (including Bagging-C5.0, Bagging-CART, Bagging-LR, Bagging-RF, and Bagging-BN), Bagging-EN and Bagging-GBM were also added for “data mining and modeling”. Bagging can train multiple prediction classifiers for the same algorithm with a non-weighted method, which is then aggregated into the model constructed by the computational classifier. In this study, we used seven models established by Bagging-C5.0, Bagging-CART, Bagging-LR, Bagging-RF, Bagging-BN, Bagging-EN, and Bagging-GBM, and then ensembled them via the voting rule of “majority decision” as the final ensemble prediction model (Figure 2).
  • Establish the optimum prediction model
    Training set resampling
    According to historical border inspection application data, the number of unqualified batches accounts for a small proportion of the total number of inspection applications, and modeling based on this data can easily lead to prediction bias. Therefore, in this study, we adopted two resampling methods (the synthesized minority oversampling technique (SMOTE) and proportional amplification) to deal with the data imbalance problem and tried to use the ratios of qualified to unqualified batches of 7:3, 6:4, 5:5, 4:6, and 3:7 for evaluation to find the best proportional parameters and unbalanced data processing method.
    Repeated modeling
    In this study, after the training set was resampled to balance the number of qualified and unqualified cases, the data combination of “time interval (AD)/whether to include the vendor blacklist/data imbalance processing method” was used to reduce the misjudgment due to a single sampling error. There were two types of time intervals (AD): 2016–2017 and 2016–2017. Blacklisted vendors refer to those whose unqualified rate was greater than the average of the overall unqualified rate. The most commonly used methods for handling data imbalance were proportional amplification and SMOTE. Based on this combination, a total of six types A to F were formed, namely, A: 2016–2017/Yes/Proportional Amplification, B: 2016–2017/Yes/SMOTE, C: 2011–2017/Yes/Proportional Amplification, D: 2011–2017/Yes/SMOTE, E: 2011–2017/No/Proportional Amplification, and F: 2011–2017/No/SMOTE. Repeated modeling was conducted ten times, and the average was used to establish the model.
    Selection of the optimal model
    The validation data set was imported into the model to obtain seven classifiers established by seven algorithms. Then seven classifiers were integrated for integrated learning to extract the optimum prediction model from the predicted results.

3.2.4. Evaluation of the Prediction Effectiveness

In this step, the test set was imported into the model, and the confusion matrix (Table 3) output prediction indicators (accuracy rate (ACR), F1, positive predictive value (PPV), Recall, and area under curve (AUC) of receiver operating characteristic (ROC)) were used to evaluate the prediction effect. The purpose was to confirm whether the model can improve the predictive effect of the unqualified rate for border inspection applications.
ACR represents the model’s ability to discriminate among overall samples. However, due to the presence of unbalanced samples in this study and the small number of unqualified samples, ACR may tend to present qualified prediction results due to its strong discriminative power towards qualified predictions. Therefore, in this study, more emphasis was placed on PPV, Recall, and F1 (Equation (4)). Recall represents the proportion of the number of unqualified products correctly identified by the model to the total number of unqualified products (Equation (5)). PPV refers to the proportion of the number of products that are actually unqualified to the number of products identified by the model as unqualified, making it also known as the unqualified rate (Equation (6)). F1 is the harmonic mean of recall and positive predictive value. Assuming that the PPV and F1 thresholds are set to 0.5, i.e., the weights of the two are equal, the performance of F1 is estimated. The larger the numerical value, the more favorable it is for the number of unqualified products TP to increase (Equation (7)).
ACR = (TP + TN)/(TP + TN + FP + FN)
Recall = TP/(FN + TP)
PPV = TP/(TP + FP)
F1 = 2 (PPV × Recall)/(PPV + Recall) = 2TP/(2TP + FP + FN)
The ROC can be plotted as a curve. The larger the area below the curve, the higher the classification accuracy. Performance can be compared between multiple ROC curves. The area under the curve (AUC) refers to the ratio of the area under the ROC curve divided by the total area. AUC can serve as the decision threshold when comparing the changes between the True Positive Rate (TPR) (Equation (8)) and False Positive Rate (FPR) (Equation (9)). The ROC curve is a graphical representation of a binary classification model’s performance that clarifies the trade-off between the True Positive Rate (TPR) and the False Positive Rate (FPR) for various threshold values. When TPR is equivalent to FPR, AUC = 0.5, which indicates that the results of the prediction model sampling inspection are equivalent to those of random sampling inspection, and the prediction model has no classification capability. AUC = 1 indicates that the classifier is perfect; 0.5 < AUC < 1 indicates that the model is superior to random sampling; AUC < 0.5 indicates that the model is inferior to random sampling (Figure 3).
True   Positive   Rate ,   TPR = T P T P + F N
False   Positive   Rate ,   FPR = F P T N + F P
The evaluation index for the effectiveness of model prediction in this study was the confusion matrix. Firstly, the classification prediction results were calculated, and the selection of models with a decision threshold greater than 0.5 for AUC (equivalent to random sampling) was prioritized. Then, a comprehensive evaluation was conducted. This study primarily focused on the unqualified rate to truly reflect the prediction hit rate. Therefore, the main evaluation index was the positive predictive value (PPV), also known as precision, which represented the ratio of the number of samples judged as unqualified by the model to the actual number of unqualified samples. Additionally, there was Recall, which was the ratio of the number of unqualified products correctly identified by the model to the total number of unqualified products. However, the larger the Recall, the higher the sampling rate. Hence, increasing PPV within the tolerable range of the sampling rate was the most important step. This also indicated the importance of realizing a balance between the harmonic mean F1, Recall, and PPV.

3.2.5. Evaluation of the Prediction Effectiveness

In this study, the data from the 2019 test set was used to make predictions through the model and simulated the actual prediction after the model launch for effectiveness evaluation. The evaluation of prediction effectiveness and selection of the optimum prediction model was based on the confusion matrix. The evaluation indicator PPV referred to the proportion of the number of products that were actually unqualified to the number of products identified by the model as unqualified. Recall referred to the accuracy of classification for all unqualified samples. EL V.1 was officially launched to conduct online risk forecasting at the border on 8 April 2020. It was switched to EL V.2 on 3 August 2020 for continuous online real-time forecasting. Therefore, in this study, we compared the unqualified rates in 2020, 2021, and 2022 after the launch with that in 2019 before the launch. The chi-square test was used to evaluate whether there was a significant increase in the unqualified rate with the aid of risk prediction and sampling of EL V.2 constructed in this study, which was used as the final evaluation result of the prediction effectiveness.

4. Results

4.1. Resampling Method and Optimal Ratio

To overcome the problem of the number of unqualified batches being too small, in this study, we tried using proportional amplification and the synthesized minority oversampling technique (SMOTE) for resampling to select the best method to deal with unbalanced data and avoid deviation in model prediction. To explore the proportional parameter of qualified to unqualified batches, tests were conducted using proportional amplification at 7:3 and SMOTE at 7:3, 6:4, 5:5, 4:6, and 3:7. After pairing with Bagging, 10 iterations were conducted to obtain the average result for each of the seven algorithms. Then, the “majority decision” in the ensemble learning method was used to obtain the results. The predictive effect was observed via PPV and F1. Previous studies found that 10 and 100 iterations of modeling exhibited comparable results, but the time required for 100 iterations significantly exceeded that for 10 iterations and was 3–8 times longer. Therefore, 10 iterations were selected for modeling, considering the time limitations.
In this study, we selected the inspection data of S-type food as the training set. After ensemble learning, the research results showed (Table 4) that when the extracted PPV and F1 were the highest, the optimal proportion of imbalanced sample processing was SMOTE 7:3. F1 was 11.03%, PPV was 6.03%, and Recall was 64.91%. Therefore, this study adopted a 7:3 ratio for qualified to unqualified samples. Based on historical experience, a ratio of 7:3 was used for proportional amplification in this study. It was not yet confirmed that SMOTE and proportional amplification were the most suitable methods for processing imbalanced data in this study. Therefore, both will continue to be included in the evaluation project in the future.

4.2. Generation of the Optimum Prediction Model

In this study, the “time interval” and “whether blacklisted manufacturers were included” were used as fixed risk factors in the training set, and the unbalanced data processing method of “SMOTE or proportional amplification” was adopted. Therefore, six data combinations were generated in the study, named A–F. Subsequently, seven algorithms were adopted for modeling, including Bagging-CART, Bagging-C5.0, Bagging-LR, Bagging-NB, Bagging-RF, Bagging-EN, and Bagging-GRM. After that, together with ensemble learning (EL), a total of 42 models and performance indicator evaluation results were generated, as listed in Table 5.
To construct the optimal prediction model in this study, the first step was to examine the effectiveness evaluation index AUC of the model, which should be greater than 50%, to ensure that the probability of unqualified batches being selected was greater than that of random sampling. Secondly, the top three combinations with the highest F1 values were prioritized. Furthermore, 25.0% for D7 random forest and both 23.0% for C8 and D8 ensemble learning indicated better performance. Another important evaluation indicator of PPV was further observed. Among the three aforementioned methods, 22.9% for the C8 ensemble method was the best. Meanwhile, Recall was 29.0%, 23.2%, and 28.3%, respectively, all of which reached the acceptable level. To comply with the requirement in practice that the general sampling rate should be controlled between 2% and 10%, it was important to note that the performance of Recalls was closely related to the sampling rate. When Recall was higher, the sampling rate was also relatively higher. Additionally, in this study, we also focused on the comparison of the number of unqualified pieces in the sampling to avoid situations where the unqualified rate was high while the sampling rate and the number of unqualified pieces were low. In summary, in this study, we selected the “C8 ensemble method” as the optimum prediction model.
In this study, we obtained similar results when examining the robustness of the model’s future prediction and the top three F1 scores of D7, C8, and D8. Therefore, a total of 16 combinations of Group C and Group D were retained for subsequent real-world prediction simulation to determine the appropriateness of the selected optimal prediction model.

4.3. Model Prediction Effectiveness

In this study, we imported the test set data into the best model C8 identified in the previous stage and simultaneously into combinations with similar evaluation results (including C1–7 and D1–8) to observe the predictive performance of the model. The research results showed (Table 6) that the top three models (C8, D7, and D8), which were originally the best choices, output F1 scores of 21.6%, 14.3%, and 15.8% and PPV values of 16.4%, 10.4%, and 12.3%, respectively, after the test set was imported for effectiveness evaluation. This result confirmed that C8 remained the optimum prediction model.
Table 6 demonstrates that the ensemble method for Group C (F1 21.6%, PPV 16.4%) exhibits significant or equivalent predictive results when compared to other single algorithms (C1–7: F1 3.4–22.2%, PPV 4.6–18.8%). The Group D ensemble method (F1 15.1%, PPV 12.3%) also exhibited similar prediction results as Group C when compared to seven algorithms (C1–7: F1 3.4–22.2%, PPV 4.6–18.8%; D1–7: F1 8.6–14.6%, PPV 4.7–13.0%). Therefore, compared to any other algorithm, the ensemble method in this study can have an equivalent or better effect, and it was also more robust.
In 2019, the total number of inspection batches for S-type food was 29,573, and the actual number of randomly selected batches with inspection results was 4154 (excluding annual inspection batches). These 318 batches with sampling results were used as test sets for prediction. The number of batches sampled according to the prediction model recommendation was 318. The recommended sampling rate by the model was 7.66%, the hit rate was 16.35%, and the number of hit batches of the model was 52. The original overall sampling rate was 10.68%, the unqualified rate was 2.09%, and the number of unqualified batches was 618. The hit rate of sampling inspection with model recommendation was 7.82 times that of the original random sampling (Table 7).
In summary, the results of this study showed that the C8 ensemble method was the optimal model choice for this study. After effectiveness evaluation, it was determined that the hit rate of sampling inspection after the model recommendation was greater than that of random sampling.

5. Discussion

To enhance the prediction performance of EL V.2, in this study, we employed several methods that differed from EL V.1. These methods included adjusting the selection approach for characteristic risk factors, incorporating additional algorithms into the model, and utilizing F adjustment to maintain the sampling rate within 2–8% after EL V.2 was launched. Simultaneously, 2% was reserved for random sampling to avoid model overfitting, thereby strengthening the robustness and prediction hit rate of ensemble model prediction results (Table 8).

5.1. Fβ Was Employed to Regulate the Sampling Inspection Rate

In this study, it was discovered that during the operation of EL V.1, the risk score distribution for each model varied (Figure 4). Hence, using the same threshold Fβ to regulate the sampling rate was not advisable. Therefore, the optimal threshold Fβ was set for each model separately through β. The F-value employed in the current evaluation model was the harmonic mean of PPV (unqualified rate in sampling inspection) and Recall (identification rate of unqualified products in sampling inspection). Fβ adjusted the weights of PPV and Recall based on different β values. The larger the β, the greater the weight of Recall (Equation (10)). Then, based on the threshold setting, the unqualified rate and sampling rate were evaluated.
F β = 1 + β 2 × P P V × R e c a l l β 2 × P P V + R e c a l l
In this study, we used Fβ to identify the prediction results of the optimal threshold for each model to maximize the F value with different β values. We reviewed the model thresholds Fβ established via various algorithms to evaluate the sampling unqualified rate and sampling rate of S-type products from 1 May 2020 to 31 May 2020. The final output is listed in the threshold regulation analysis table with different β values, as presented in Table 9.
To control the sampling rate at 7%, using Beta 2.6 as an example, the unqualified rate of sampling was 16.67%, and the sampling rate was 7.23%. When all classification models utilized the same threshold, the unqualified rate of sampling was 15.45%, and the sampling rate was 7.56% (Table 10). This study found that regulating the sampling rate with Beta can increase the unqualified rate of sampling. If the sampling rate was low, the Beta value could be adjusted higher to improve the sampling rate; if the sampling rate was too high, the Beta value could be lowered to reduce the sampling rate. Therefore, the EL V.2 constructed in this study was designed to regulate the Beta value according to the required sampling rate. Through the automated generation of optimal thresholds by the model, the accuracy of each model can be enhanced, and the effectiveness of sampling management can be strengthened.

5.2. Comparison between Single Algorithm and Ensemble Algorithm

Among the 42 prediction models established in the stage of optimal model selection, for each of the six data combinations of A-F, both F1 and PPV of the ensemble learning method ranked in the top three among the eight models when compared to the single algorithm. Moreover, their AUCs were all greater than that of 50% random sampling (Table 4). When further using the test set to simulate actual predictions, the ensemble method in the C and D data combinations (Table 5) remained in the top three (C8 ensemble method F1 21.6%, PPV 16.4%, AUC 69.9% > 50%; D8 ensemble method F1 15.1%, PPV 12.3%, AUC 69.0% > 50%). The results of this study showed that the ensemble method was the most suitable approach for constructing border food prediction models, and its robustness could ensure that high-risk products could be efficiently predicted and detected as unqualified through sampling and inspection. Thus, the occurrence of food safety incidents could be prevented.

5.3. Comparison of Prediction Effectiveness between EL V.2 and EL V.1 Models

In this section, we explored whether the second-generation ensemble learning prediction model (EL V.2) constructed by our research institute (composed of seven algorithms: Bagging-CART, Bagging-C5.0, Bagging-Logistic, Bagging-NB, Bagging-RF, Bagging-EN, and Bagging-GRM) exhibited better predictive performance than the first-generation model (EL V.1) constructed by the previous study using five algorithms: Bagging-CART, Bagging-C5.0, Bagging-Logistic, Bagging-NB, and Bagging-RF. In this study, we selected the time interval in 2020 with ensemble learning for effectiveness evaluation. EL V.1 analysis interval: 8 April 2020 to 2 August 2020; EL V.2 analysis interval: 3 August 2020 to 30 November 2020. After using the prediction index established by the confusion matrix, the results showed that:
  • The AUC of EL V.1 ranged from 53.43% to 69.03%, while the AUC of EL V.2 ranged from 49.40% to 63.39%. After a majority decision, the Bagging-CART model of EL V.2 with AUC less than 50% was considered unsuitable. By adopting a majority decision strategy through ensemble learning, the influence of the Bagging-CART model was diluted by the other six models. Thus, EL V.2 exhibited better robustness than EL V.1. The advantage of ensemble learning was that when a small number of algorithms were not suitable (worse than random sampling), there was a mechanism for eliminating or weakening influence. The performance of AUC showed that EL V.1 and EL V.2 had a greater prediction probability than randomly selecting unqualified cases (Table 11).
  • The predictive evaluation index F1 (8.14%) and PPV (4.38%) of EL V.2 had better results compared to F1 (4.49%) and PPV (2.47%) of EL V.1, indicating that EL V.2 had better predictive effects than EL V.1 (Table 12).
The above results indicated that EL V.2 had better predictive performance than EL V.1, but it should still be noted that the Recall of EL V.2 was about twice that of EL V.1. This suggested that there might be a relative increase in the sampling rate. Therefore, determining how to control the sampling rate within the general sampling rate range (2–10%) while improving the unqualified hit rate was a key consideration after the model’s launch.

5.4. Evaluation of the Effectiveness of the Prediction Model after Its Launch

In this study, we used the ensemble learning method to construct the EL V.1 model, which was launched on 8 April 2020. The S-type food was imported for sampling inspection prediction. On 3 August 2020, EL V.1 was replaced by EL V.2. To understand the effectiveness of the model after its launch, the performance from 2020 to 2022 was compared with that of the random sampling method in 2019. The results showed that from 2020 to 2022, after conducting general sampling inspection predictions using the ensemble learning model, the unqualified rates obtained were 5.10%, 6.36%, and 4.39%, respectively, which were higher than the unqualified rate of 2.09% in 2019. The overall annual sampling rates were 6.07% in 2020, 9.14% in 2021, and 10.9% in 2022, which were all controlled within the range of 2–10% (without rounding below the decimal point) (Table 13 and Table 14). In this study, we further utilized statistical analysis for the chi-square test. The results showed that the ensemble learning method for border food sampling inspection had statistical significance (p value = 0.000 ***) in improving the unqualified rate (Table 14). Therefore, the ensemble learning model EL V.2, constructed by the seven algorithms used in this study and launched on 3 August 2020, can effectively increase the unqualified rate while maintaining the general sampling rate within a reasonable range of 2–10%.
The findings of this study are as follows:
  • EL V.2 is better than random sampling. After the ensemble learning model EL V.2, developed in this study, was launched online, the predicted results from 2020 to 2022 were reviewed. Based on the overall general sampling cases throughout the year, it was determined that the unqualified rate was 3.74% in 2020, 4.16% in 2021, and 3.01% in 2022, all of which were significantly higher than 2.09% in 2019. Further observation showed that the unqualified rates of cases recommended for sampling inspection through ensemble learning in 2020, 2021, and 2022 were 5.10%, 6.36%, and 4.39%, respectively, which were significantly higher than the 2.09% under random sampling inspection in 2019.
  • The ensemble learning model should be periodically re-modeled. Based on Table 12, it can be observed that the unqualified rate showed a growing trend from 2019 to 2021 but a slight decrease in 2022 (Figure 5). The results of the further chi-square test showed that the unqualified rate in 2022 was still significantly higher than that in 2019 (p value = 0.000 *** < 0.001) (Table 14). However, for ensemble learning prediction models constructed using various machine learning algorithms, the factors and data required for modeling often change with factors such as the external environment and policies. Re-modeling was necessary to make the best adjustments to “data drift” or “concept drift” in the real world to prevent model failure. Drift refers to the degradation of predictive performance over time due to hidden external environmental factors. Due to the fact that data changed over time, the model’s capability to make accurate predictions may decrease. Therefore, it was necessary to monitor data drift and conduct timely reviews of modeling factors. When collecting new data, the data predicted by the model should be avoided to prevent the new model from overfitting when making predictions. The goal of this study is to enable the new model to adjust to changes in the external environment, which will be a sustained effort in the future.
  • The trade-off between unqualified batch hit rate and computational efficiency needs to be established. While the rejection rate was improved using the model constructed with seven algorithms (i.e., EL V.2), there were approximately 0.1% of batches where the model took more than one minute to compute. The model was designed to facilitate inspectors at the border to make fast decisions on sampling. Considering computational efficiency and real-time prediction, random sampling would be automatically selected for batches with over 1 min computation time to avoid delay in border inspections due to model failure.

5.5. Research Limitations

When determining the research scope, it was necessary to ensure that each product classification for border inspection applications had unqualified cases and that the number of unqualified cases was not too small. Therefore, for those with an unqualified rate of less than 1% in past sampling and fewer than 10 unqualified cases, the original random sampling mechanism was maintained. The product classification was not included in the scope of this study when it was impossible to find a classification with high product homogeneity and similar inspection items that could be merged. Owing to legal requirements associated with government data, the types, content, and hyperparameters of risk factors cannot be presented in this paper to protect information security and confidentiality.

6. Conclusions

In this study, we constructed a second-generation integrated learning prediction model, EL V.2. The research results showed that EL V.2 exhibited better prediction performance than random sampling and the first-generation integrated learning prediction model, EL V.1. Additionally, the model was composed of seven algorithms. Hence, when the model was inadequate (AUC < 50%), the overall prediction results remained robust when integrated learning was conducted through the majority decision voting method.
The outbreak of the COVID-19 pandemic in late 2020 had a worldwide impact on border control measures as well as economic and trade exchanges. Compared with unqualified rates in 2019, 2020 and 2021 saw increases in unqualified cases in Taiwan, which is likely to be attributed to the great changes in the origin and quantity of imported goods caused by the pandemic. Another reason for the changes in unqualified rates could be the modification of some related regulations and inspection standards. The effects of these aspects on the evaluation of the performance of EL V.1 and EL V.2 still require further observation and analysis in the future. Since 2020, Taiwan’s border management has gradually introduced an intelligent management operation model. Border management powered by artificial intelligence enables Taiwan to strengthen its risk prediction capabilities and quickly adapt to trends in the context of rapid changes in the international environment, thereby ensuring people’s health and safety.

Author Contributions

Conceptualization, L.-Y.W. and F.-M.L.; methodology, L.-Y.W. and S.-S.W.; software, L.-Y.W. and W.-C.L.; validation, L.-Y.W. and F.-M.L.; formal analysis, L.-Y.W.; investigation, L.-Y.W.; resources, F.-M.L. and W.-C.L.; data curation, L.-Y.W. and W.-C.L.; writing—original draft preparation, L.-Y.W.; writing—review and editing, L.-Y.W. and S.-S.W.; visualization, F.-M.L. and S.-S.W.; supervision, F.-M.L.; project administration, L.-Y.W.; funding acquisition, no. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We would like to express our gratitude to the Food and Drug Administration of the Ministry of Health and Welfare of Taiwan for approving the execution of this study on April 2023 (approval document No. 1122001214 and 1122001937).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Comparison table of full and abbreviated names.
Table A1. Comparison table of full and abbreviated names.
Abbreviated NameFull Name
ACRAccuracy rate
AUCArea Under Curve
C5.0Decision Tree C5.0
CARTDecision Tree CART
CPICorruption Perceptions Index
EL V.1First-generation ensemble learning prediction model
EL V.2Second-generation ensemble learning prediction model
ENElastic Net
FTMSFood Traceability Management System
GBMGradient Boosting Machine
GDPGross Domestic Product
HDIHuman Development Index
IFISImport Food Information System
IMSInspection Management System
LRLogistic Regression
LRILegal Rights Index
NBNaïve Bayes
PMDSProduct Management Decision System
PPVPositive Predictive Value
PREDICTPredictive Risk-based Evaluation for Dynamic Import Compliance Targeting System
RASFFRapid Alert System for Food and Feed
RFRandom Forest
ROCReceiver Operating Characteristic
RPFBSRegistration Platform of Food Businesses System
SMOTESynthesized Minority Oversampling Technique

References

  1. Food and Drug Administration, Ministry of Health and Welfare of Taiwan. Analysis of Import Inspection Data of Food and Related Products at the Taiwan Border for the Year 107. Annu. Rep. Food Drug Res. 2019, 10, 404–408. Available online: https://www.fda.gov.tw (accessed on 20 July 2020).
  2. Bouzembrak, Y.; Marvin, H.J.P. Prediction of food fraud type using data from rapid alert system for food and feed (RASFF) and Bayesian network modelling. Food Control 2016, 61, 180–187. [Google Scholar] [CrossRef]
  3. Brandao, M.P.; Neto, M.G.; Anjos, V.C.; Bell, M.J.V. Detection of adulteration of goat milk powder with bovine milk powder by front-face and time resolved fluorescence. Food Control 2017, 81, 168–172. [Google Scholar] [CrossRef]
  4. Lin, Y. Food Safety in the Age of Big Data. Hum. Soc. Sci. Newslett. 2017, 19, 1. [Google Scholar]
  5. Feng, L.; Zhang, Z.; Ma, Y.; Du, Q.; Williams, P.; Drewry, J.; Luck, B. Alfalfa Yield Prediction Using UAV-Based Hyperspectral Imagery and Ensemble Learning. Remote Sens. 2020, 12, 2028. [Google Scholar] [CrossRef]
  6. Neto, H.A.; Tavares, W.L.F.; Ribeiro, D.C.S.Z.; Alves, R.C.O.; Fonseca, L.M.; Campos, S.V.A. On the utilization of deep and ensemble learning to detect milk adulteration. BioData Min. 2019, 12, 13. [Google Scholar] [CrossRef]
  7. Park, M.S.; Kim, H.N.; Bahk, G.J. The analysis of food safety incidents in South Korea, 1998–2016. Food Control 2017, 81, 196–199. [Google Scholar] [CrossRef]
  8. Liu, Y.; Zhou, S.; Han, W.; Li, C.; Liu, W.; Qiu, Z.; Chen, H. Detection of Adulteration in Infant Formula Based on Ensemble Convolutional Neural Network and Near-Infrared Spectroscopy. Foods 2021, 10, 785. [Google Scholar] [CrossRef]
  9. Wang, Z.; Wu, Z.; Zou, M.; Wen, X.; Wang, Z.; Li, Y.; Zhang, Q. A Voting-Based Ensemble Deep Learning Method Focused on Multi-Step Prediction of Food Safety Risk Levels: Applications in Hazard Analysis of Heavy Metals in Grain Processing Products. Foods 2022, 11, 823. [Google Scholar] [CrossRef]
  10. Parastar, H.; Kollenburg, G.V.; Weesepoel, Y.; Doel, A.V.D.; Buydens, L.; Jansen, J. Integration of handheld NIR and machine learning to “Measure & Monitor” chicken meat authenticity. Food Control 2020, 112, 107149. [Google Scholar] [CrossRef]
  11. Marvin, H.J.P.; Bouzembrak, Y.; Janssen, E.M.; van der Fels-Klerx, H.J.; van Asselt, E.D.; Kleter, G.A. A holistic approach to food safety risks: Food fraud as an example. Food Res. Int. 2016, 89, 463–470. [Google Scholar] [CrossRef]
  12. Bouzembrak, Y.; Klüche, M.; Gavai, A.; Marvin, H.J.P. Internet of Things in food safety: Literature review and a bibliometric analysis. Trends Food Sci. Technol. 2019, 94, 54–64. [Google Scholar] [CrossRef]
  13. Marvin, H.J.P.; Janssen, E.M.; Bouzembrak, Y.; Hendriksen, P.J.M.; Staats, M. Big data in food safety: An overview. Crit. Rev. Food Sci. Nutr. 2017, 57, 2286–2295. [Google Scholar] [CrossRef]
  14. U.S. Government Accountability Office. Imported Food Safety: FDA’s Targeting Tool Has Enhanced Screening, but Further Improvements Are Possible; GAO: Washington, DC, USA, 2016. Available online: https://www.gao.gov/products/gao-16-399 (accessed on 1 November 2021).
  15. Alam, T.M.; Shaukat, K.; Khelifi, A.; Aljuaid, H.; Shafqat, M.; Ahmed, U.; Nafees, S.A.; Luo, S. A Fuzzy Inference-Based Decision Support System for Disease Diagnosis. Comput. J. 2022, bxac068. [Google Scholar] [CrossRef]
  16. Athanasiou, M.; Zarkogianni, K.; Karytsas, K.; Nikita, K.S. An LSTM-based Approach Towards Automated Meal Detection from Continuous Glucose Monitoring in Type 1 Diabetes Mellitus. In Proceedings of the 2021 IEEE 21st International Conference on Bioinformatics and Bioengineering (BIBE), Kragujevac, Serbia, 25–27 October 2021. [Google Scholar] [CrossRef]
  17. Bicakci, K.; Tunali, V. Transfer Learning Approach to COVID-19 Prediction from Chest X-Ray Images. In Proceedings of the 2021 Innovations in Intelligent Systems and Applications Conference (ASYU) Intelligent Systems and Applications Conference (ASYU), Elazig, Turkey, 1–5 October 2021. [Google Scholar] [CrossRef]
  18. Nandanwar, H.; Nallamolu, S. Depression Prediction on Twitter using Machine Learning Algorithms. In Proceedings of the 2021 2nd Global Conference for Advancement in Technology (GCAT) Advancement in Technology (GCAT), Bangalore, India, 1–7 October 2021. [Google Scholar] [CrossRef]
  19. Daniels, J.; Herrero, P.; Georgiou, P.A. Deep Learning Framework for Automatic Meal Detection and Estimation in Artificial Pancreas Systems. Sensors 2022, 22, 466. [Google Scholar] [CrossRef]
  20. Hemmer, A.; Abderrahim, M.; Badonnel, R.; Chrisment, I. An Ensemble Learning-Based Architecture for Security Detection in IoT Infrastructures. In Proceedings of the 2021 17th International Conference on Network and Service Management (CNSM), Izmir, Turkey, 25–29 October 2021. [Google Scholar] [CrossRef]
  21. Ennaji, S.; Akkad NEHaddouch, K. A Powerful Ensemble Learning Approach for Improving Network Intrusion Detection System (NIDS). In Proceedings of the 2021 Fifth International Conference On Intelligent Computing in Data Sciences (ICDS), Fez, Morocco, 1–6 October 2021. [Google Scholar] [CrossRef]
  22. Larocque-Villiers, J.; Dumond, P.; Knox, D. Automating Predictive Maintenance Using State-Based Transfer Learning and Ensemble Methods. In Proceedings of the 2021 IEEE International Symposium on Robotic and Sensors Environments (ROSE), Virtual, 1–7 October 2021. [Google Scholar] [CrossRef]
  23. Traore, A.; Chetoui, M.; Landry, F.G.; Akhloufi, M.A. Ensemble Learning Framework to Detect Partial Discharges and Predict Power Line Faults. In Proceedings of the 2021 IEEE Electrical Power and Energy Conference (EPEC), Toronto, ON, Canada, 22–31 October 2021. [Google Scholar] [CrossRef]
  24. Yang, X.; Zhou, D.; Song, W.; She, Y.; Chen, X. A Cable Layout Optimization Method for Electronic Systems Based on Ensemble Learning and Improved Differential Evolution Algorithm. IEEE Trans. Electromagn. Compat. 2021, 63, 1962–1971. [Google Scholar] [CrossRef]
  25. Ranawake, D.; Bandaranayake, S.; Jayasekara, R.; Madhushani, I.; Gamage, M.; Kumari, S. Tievs: Classified Advertising Enhanced Using Machine Learning Techniques. In Proceedings of the 2021 IEEE 12th Annual Information Technology, Electronics and Mobile Communication Conference (IEMCON), Vancouver, BC, Canada, 27–30 October 2021. [Google Scholar] [CrossRef]
  26. Batool, D.; Shahbaz, M.; Shahzad Asif, H.; Shaukat, K.; Alam, T.M.; Hameed, I.A.; Ramzan, Z.; Waheed, A.; Aljuaid, H.; Luo, S. A Hybrid Approach to Tea Crop Yield Prediction Using Simulation Models and Machine Learning. Plants 2022, 11, 1925. [Google Scholar] [CrossRef]
  27. Adak, A.; Pradhan, B.; Shukla, N. Sentiment Analysis of Customer Reviews of Food Delivery Services Using Deep Learning and Explainable Artificial Intelligence: Systematic Review. Foods 2022, 11, 1500. [Google Scholar] [CrossRef]
  28. Dasarathy, B.V.; Sheela, B.V. A composite classifier system design: Concepts and methodology. Proc. IEEE 1979, 67, 708–713. [Google Scholar] [CrossRef]
  29. Hansen, L.K.; Salamon, P. Neural network ensembles. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 993–1001. [Google Scholar] [CrossRef]
  30. Polikar, R. Ensemble Based Systems in Decision Making. Circuits Syst. Mag. IEEE 2006, 6, 21–45. [Google Scholar] [CrossRef]
  31. Solano, E.S.; Affonso, C.M. Solar Irradiation Forecasting Using Ensemble Voting Based on Machine Learning Algorithms. Sustainability 2023, 15, 7943. [Google Scholar] [CrossRef]
  32. Chandrasekhar, N.; Peddakrishna, S. Enhancing Heart Disease Prediction Accuracy through Machine Learning Techniques and Optimization. Processes 2023, 11, 1210. [Google Scholar] [CrossRef]
  33. Alsulami, A.A.; AL-Ghamdi, A.S.A.-M.; Ragab, M. Enhancement of E-Learning Student’s Performance Based on Ensemble Techniques. Electronics 2023, 12, 1508. [Google Scholar] [CrossRef]
  34. Sugsnyadevi, K.; Malmurugan, N.; Sivakumar, R. OF-SMED: An Optimal Foreground Detection Method in Surveillance System for Traffic Monitoring. In Proceedings of the 2012 International Conference on Cyber Security, Cyber Warfare and Digital Forensic, CyberSec, Kuala Lumpur, Malaysia, 26–28 June 2012; pp. 12–17. [Google Scholar] [CrossRef]
  35. Pagano, C.; Granger, E.; Sabourin, R.; Gorodnichy, D.O. Detector Ensembles for Face Recognition in Video Surveillance. In Proceedings of the 2012 International Joint Conference on Neural Networks (IJCNN), Brisbane, QLD, Australia, 10–15 June 2012; pp. 1–8. [Google Scholar] [CrossRef]
  36. Wang, R.; Bunyak, F.; Seetharaman, G.; Palaniappan, K. Static and Moving Object Detection Using Flux Tensor with Split Gaussian Models. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Columbus, OH, USA, 23–28 June 2014; pp. 420–424. [Google Scholar] [CrossRef]
  37. Wang, T.; Sonoussi, H. Detection of Abnormal Visual Events via Global Optical Flow Orientation Histogram. IEEE Trans. Inf. Forensics Secur. 2014, 9, 998. [Google Scholar] [CrossRef]
  38. Tsai, C.J. New feature selection and voting scheme to improve classification accuracy. Methodol. Appl. 2019, 23, 12017–12030. [Google Scholar] [CrossRef]
  39. Food Safety Information Network. Available online: https://www.ey.gov.tw/ofs/A236031D34F78DCF (accessed on 22 July 2022).
  40. Ogutu, J.O.; Schulz-Streeck, T.; Piepho, H.P. Genomic selection using regularized linear regression models: Ridge regression, Lasso, elastic net and their extensions. BMC Proc. 2012, 6, S10. [Google Scholar] [CrossRef]
Figure 1. Modeling process of the second-generation ensemble learning prediction model EL V.2 (Note: The red letter indicates the difference between EL V.2 and EL V.1 modeling processes).
Figure 1. Modeling process of the second-generation ensemble learning prediction model EL V.2 (Note: The red letter indicates the difference between EL V.2 and EL V.1 modeling processes).
Foods 12 02118 g001
Figure 2. Ensemble learning model architecture.
Figure 2. Ensemble learning model architecture.
Foods 12 02118 g002
Figure 3. ROC curve and AUC concept map.
Figure 3. ROC curve and AUC concept map.
Foods 12 02118 g003
Figure 4. Predicted risk score distribution map of inspection application cases with five algorithms in EL V.1 as examples (Data time interval from 8 April 2020 to 7 June 2020).
Figure 4. Predicted risk score distribution map of inspection application cases with five algorithms in EL V.1 as examples (Data time interval from 8 April 2020 to 7 June 2020).
Foods 12 02118 g004
Figure 5. Annual trend chart before and after the introduction of ensemble learning.
Figure 5. Annual trend chart before and after the introduction of ensemble learning.
Foods 12 02118 g005
Table 1. Type and sources of characteristic factors.
Table 1. Type and sources of characteristic factors.
TypeFactorsData Sources
ProductValue, net weight, inspection methods, blacklisted products, packaging methods, validity period, products for which international recall alerts have been issued, manufacturing date, expiry date, etc.Taiwan Food Cloud:
Data on border inspections
Product inspection and testing data
Product alerts

Information on international public opinion and product recall alerts:
United States Food and Drug Administration (US FDA)
https://www.fda.gov
Food Safety and Inspection Service (FSIS) of the U.S. Department of Agriculture (USDA)
https://www.fsis.usda.gov
Rapid Alert System for Food and Feed (RASFF) of the European Union
https://ec.europa.eu/food/safety/rasff_en
Canadian Food Inspection Agency (CFIA)
http://inspection.gc.ca
Food Standards Agency (FSA) of the United Kingdom
https://www.food.gov.uk
Food Safety Authority of Ireland (FSAI)
https://www.fsai.ie
Food Standards Australia New Zealand (FSANZ)
http://www.foodstandards.gov.au
Consumer Affairs Agency (CAA) of Japan
https://www.recall.caa.go.jp
Singapore Food Agency (SFA)
https://www.sfa.gov.sg
China Food and Drug Administration (CFDA)
http://gkml.samr.gov.cn
Foodmate Network of China
http://news.foodmate.net
Centre for Food Safety (CFS) of Hong Kong
http://www.cfs.gov.hk
Border inspection Transportation time, month of inspection, quarter of inspection, year of inspection, method of transportation, agent importation, re-exportation, customs district, etc.Taiwan Food Cloud:
Management data of border inspections
Customs broker Number of declarations filed, number of border inspection cancellations, number of days from the previous importation, rate of change of number of days taken for importation, number of cases of non-conforming labels and external appearances, number of batches forfeited or returned, number of inspections, number of failed inspections, number of failed document reviews, number of product classes, etc.Taiwan Food Cloud:
Food company registration data
Data on border inspections
Business registration data
Importer Capital, years of establishment, number of branches, number of downstream vendors, number of company registration changes, number of late deliveries, sole focus on importation (yes/no), number of lines of businesses, new company (yes/no), district of registration, branch company (yes/no), blacklisted importer (yes/no), county/city, number of preliminary inspections, GHP inspections, HACCP inspections, label inspections, product inspections, number of lines of food businesses, factory registration (yes/no), delayed declaration of goods receipt/delivery (yes/no), interval between importations, variations in the interval between importations, variations in the number of days taken for importation, variations in total net weight, number of declarations filed, number of cases of non-conforming Chinese labels and external appearances, value, net weight, number of non-releases, number of batches detained, forfeited or returned, number of failed inspections, number of inspections, number of failed document reviews, number of border inspection cancellations, number of manufacturers, number of product classes for which declarations have been filed, total number of classes, etc.Taiwan Food Cloud:
Food company registration data
Data on border inspections
Product inspection and testing data
Product flow data
Business registration data
Manufacturer Trademarks, interval between importations, rate of change of interval between importations, internationally alerted manufacturer (yes/no), internationally alerted brand (yes/no), number of cases of non-conforming Chinese labels and external appearances, number of batches detained, forfeited or returned, number of failed inspections, number of inspections, number of failed document reviews, number of declarations filed, number of border inspection cancellations, number of importers, number of product classes, etc.Taiwan Food Cloud:
Food company registration data
Data on border inspections
Product inspection and testing data
Product alerts
Information on international public opinion and product recall alerts:
USFDA https://www.fda.gov
FSIS https://www.fsis.usda.gov
CFIA http://inspection.gc.ca
FSA https://www.food.gov.uk
RASFF https://ec.europa.eu/food/safety/rasff_en
FSAI https://www.fsai.ie
FSANZ http://www.foodstandards.gov.au
CAA https://www.recall.caa.go.jp
SFA https://www.sfa.gov.sg
CFDA http://gkml.samr.gov.cn
Foodmate Network of China http://news.foodmate.net
CFS http://www.cfs.gov.hk
Country of manufactureCountry of manufacture of products subjected to inspectionData on border inspections
GDP, economic growth rate, GFSI, CPI, HDI, LRI, regional PRIInformation on international public opinion and product recall alerts:
https://data.oecd.org/gdp/gross-domestic-product-gdp.htm
https://www.imf.org/en/Publications
https://foodsecurityindex.eiu.com/
https://www.transparency.org/en/cpi/2020/index/nzl
http://hdr.undp.org/en/2020-report
https://data.worldbank.org/indicator/IC.LGL.CRED.XQ
https://www.prsgroup.com/regional-political-risk-index/
Table 2. Characteristic factor usage frequency counting table.
Table 2. Characteristic factor usage frequency counting table.
Characteristic FactorTimesCharacteristic FactorTimesCharacteristic FactorTimesCharacteristic FactorTimes
Country of production70Declaration acceptance unit17Whether there is a trademark5Frequency of business registration changes1
Inspection method64Advance release17Product registration location5Is it an import broker?1
Product classification code64Cumulative sampling number of imports17Regional political risk index5Average price per kilogram1
Blacklist vendor47Human development index17Tax registration data available?5Acceptance month1
Cumulative number of unqualified imports in sampling inspection45Packaging method15Non-punctual declaration rate of delivery5Acceptance season1
Dutiable price in Taiwan dollars44Capital14Input/output mode4Acceptance year1
Total net weight35Import cumulative number of new classifications13Percentage of remaining validity period for acceptance4Overdue delivery1
Global food security indicator29Years of importer establishment10Whether there is factory registration?4Any business registration change?1
Type of Obligatory inspection applicant26Number of companies in the same group10New company established within the past three months3Product classification code1
Legal rights index26Is it a pure input industry?10Number of GHP inspection3Number of GHP inspection failures1
Blacklist product25Customs classification10Accumulated number of unqualified imports3Number of HACCP inspection1
Storage and transportation conditions23County and city level10Transportation time2Number of HACCP inspection1
Packaging materials21Total number of imported product lines8GDP growth rate2Manufacturing date later than effective date1
Cumulative number of reports20Number of downstream manufacturers7Number of branch companies2Valid for more than 5 years1
Total classification number of imports20Is it a branch company?7Number of overdue deliveries2Acceptance date later than manufacturing date1
Corruption perceptions index18Number of non-review inspections7Any intermediary trade?2Acceptance date later than the effective date1
GDP17Rate of non-timely declaration of goods received7Cumulative number of imports not released2Number of business projects in the food industry1
Table 3. Types and definitions of confusion matrices.
Table 3. Types and definitions of confusion matrices.
TypeDefinition
True Positive, TPEach batch of inspection applications was predicted as unqualified by the model, and it was actually unqualified.
False Positive, FPEach batch of inspection applications was predicted as unqualified by the model, but it was actually qualified.
True NegativeEach batch of inspection applications was predicted as qualified by the model, and it was actually qualified.
False NegativeEach batch of inspection applications was predicted as qualified by the model, but it was actually unqualified.
Table 4. Evaluation of imbalanced data sampling ratio.
Table 4. Evaluation of imbalanced data sampling ratio.
Imbalanced Data Processing Methods and Sampling Ratio #Precision
(PPV)
RecallF1
SMOTE 7:36.03%64.91%11.03%
SMOTE 6:45.68%66.15%10.46%
SMOTE 5:55.48%75.16%10.22%
SMOTE 4:64.94%77.33%9.28%
SMOTE 3:74.80%81.68%9.06%
Equal magnification 7:34.62%87.89%8.77%
Note #: The sampling ratio was 7:3 for qualified and unqualified products.
Table 5. Index evaluation for 42 prediction models.
Table 5. Index evaluation for 42 prediction models.
Data SetCombination Number 1AlgorithmACRRecallPPVF1AUCTNFPTPFNSampling RateRejection Rate
Validation setA1Bagging-C5.092.3%2.2%4.1%2.8%60.6%24517031352.754.11
A2Bagging-CART86.6%25.4%12.2%16.5%69.5%22692523510310.7912.20
A3Bagging-EN27.9%90.6%6.2%11.5%68.0%61719041251376.316.16
A4Bagging-GBM84.7%37.7%13.9%20.4%72.3%2200321528614.0313.94
A5Bagging-LR83.0%31.2%10.8%16.0%68.0%2164357439515.0410.75
A6Bagging-NB69.7%60.1%9.9%17.1%73.2%1769752835531.409.94
A7Bagging-RF93.4%0.7%2.6%1.1%71.0%24833811371.472.56
A8EL85.5%28.3%12.0%16.8%72.5%2235286399912.2212.00
B1Bagging-C5.089.7%22.5%15.6%18.4%72.7%2353168311077.4815.58
B2Bagging-CART88.4%19.6%12.0%14.9%68.7%2323198271118.4612.00
B3Bagging-EN7.7%97.8%5.2%9.9%69.7%712450135397.225.22
B4Bagging-GBM90.1%26.8%18.6%22.0%73.1%2359162371017.4818.59
B5Bagging-LR87.9%28.3%14.9%19.5%71.2%229922239999.8214.94
B6Bagging-NB79.6%50.0%12.7%20.3%73.3%2048473696920.3812.73
B7Bagging-RF90.6%24.6%18.9%21.4%75.2%2375146341046.7718.89
B8EL88.2%31.9%16.6%21.8%74.0%230022144949.9716.60
C1Bagging-C5.093.4%11.6%23.2%15.5%67.2%246853161222.5923.19
C2Bagging-CART86.6%39.9%16.8%23.6%69.7%2248273558312.3416.77
C3Bagging-EN81.3%11.6%4.1%6.0%50.1%21453761612214.744.08
C4Bagging-GBM88.9%33.3%18.5%23.8%73.0%231820346929.3618.47
C5Bagging-LR86.1%33.3%14.2%20.0%69.0%2244277469212.1514.24
C6Bagging-NB72.5%58.7%10.7%18.1%73.7%1847674815728.3910.73
C7Bagging-RF94.6%7.2%38.5%12.2%75.4%250516101280.9838.46
C8EL92.0%23.2%22.9%23.0%73.6%2413108321065.2722.86
D1Bagging-C5.092.2%21.7%23.1%22.4%73.2%2421100301084.8923.08
D2Bagging-CART87.3%28.3%14.0%18.8%72.9%2282239399910.4614.03
D3Bagging-EN54.2%50.7%5.7%10.3%52.6%13701151706845.925.73
D4Bagging-GBM91.1%20.3%18.2%19.2%74.2%2395126281105.7918.18
D5Bagging-LR90.1%22.5%16.7%19.1%70.1%2366155311077.0016.67
D6Bagging-NB77.4%52.2%11.9%19.3%73.7%1986535726622.8311.86
D7Bagging-RF91.0%29.0%22.0%25.0%76.4%237914240986.8421.98
D8EL90.2%28.3%19.4%23.0%75.1%235916239997.5619.40
E1Bagging-C5.092.3%9.4%14.1%11.3%66.3%244279131253.4614.13
E2Bagging-CART84.8%33.3%12.9%18.6%68.4%2210311469213.4312.89
E3Bagging-EN88.2%3.6%2.7%3.1%58.1%234118051336.962.70
E4Bagging-GBM88.1%27.5%14.9%19.3%71.5%2304217381009.5914.90
E5Bagging-LR85.9%32.6%13.8%19.4%69.1%2239282459312.3013.76
E6Bagging-NB73.2%58.0%10.9%18.3%73.7%1867654805827.6010.90
E7Bagging-RF94.5%2.9%26.7%5.2%73.1%25101141340.5626.67
E8EL91.1%16.7%15.9%16.3%70.9%2399122231155.4515.86
F1Bagging-C5.092.4%7.2%11.9%9.0%71.1%244774101283.1611.90
F2Bagging-CART89.9%9.4%8.3%8.8%67.2%2377144131255.908.28
F3Bagging-EN62.3%19.6%2.9%5.1%55.6%16298922711134.562.94
F4Bagging-GBM91.1%16.7%16.0%16.3%72.9%2400121231155.4215.97
F5Bagging-LR90.4%18.8%15.3%16.9%68.4%2377144261126.3915.29
F6Bagging-NB79.7%47.8%12.4%19.6%73.6%2053468667220.0812.36
F7Bagging-RF90.9%17.4%15.9%16.6%74.3%2394127241145.6815.89
F8EL91.7%10.1%12.5%11.2%72.1%242398141244.2112.50
Note 1: The data combination representation method was the time interval of data/whether blacklisted vendors are included/unbalanced data processing method. There were a total of 6 combinations, namely, A: 2016–2017/Yes/Proportional Amplification, B: 2016–2017/Yes/SMOTE, C: 2011–2017/Yes/Proportional Amplification, D: 2011–2017/Yes/SMOTE, E: 2011–2017/No/Proportional Amplification, and F: 2011–2017/No/SMOTE.
Table 6. Index evaluation details of the optimal risk prediction model.
Table 6. Index evaluation details of the optimal risk prediction model.
Data SetCombination Number 1AlgorithmACRRecallPPVF1AUCTNFPTPFNSampling RateRejection Rate
Test setC1Bagging-C5.094.7%8.0%15.7%10.6%68.0%392170131502.0015.66
C2Bagging-CART89.4%38.7%15.6%22.2%71.6%3649342631009.7515.56
C3Bagging-EN88.9%9.2%4.6%6.1%53.2%3678313151487.904.57
C4Bagging-GBM87.8%39.9%13.8%20.5%72.0%3584407659811.3613.77
C5Bagging-LR49.7%64.4%4.9%9.1%51.6%196020311055851.424.92
C6Bagging-NB74.3%52.8%8.0%13.9%66.8%2999992867725.957.98
C7Bagging-RF95.8%1.8%18.8%3.4%72.5%39781331600.3918.75
C8EL90.9%31.9%16.4%21.6%69.9%3725266521117.6616.35
D1Bagging-C5.091.2%12.3%8.2%9.8%69.3%3767224201435.87%8.20%
D2Bagging-CART89.5%14.7%7.5%9.9%67.6%3693298241397.75%7.45%
D3Bagging-EN56.7%52.1%4.7%8.6%57.1%22721719857843.43%4.71%
D4Bagging-GBM93.1%13.5%13.0%13.3%71.0%3844147221414.07%13.02%
D5Bagging-LR92.4%16.6%13.0%14.6%65.3%3811180271364.98%13.04%
D6Bagging-NB81.3%39.9%8.7%14.3%66.8%3313678659817.89%8.75%
D7Bagging-RF86.2%33.1%10.4%15.8%68.5%35254665410912.52%10.38%
D8EL91.4%19.6%12.3%15.1%69.0%3763228321316.26%12.31%
Note 1: The representation of data combination was based on the time interval of data in year/whether blacklisted vendors are included/unbalanced data processing method. C: 2011–2017/Yes/proportionally, D: 2011–2017/Yes/SMOTE.
Table 7. Evaluation of the prediction effectiveness of the optimal risk prediction model.
Table 7. Evaluation of the prediction effectiveness of the optimal risk prediction model.
Data
Year
Overall Sampling InspectionEL V.2 Sampling Inspection
Number of Inspection Application BatchesSampling RateRejection RatePrediction Batch Number 1Suggested Number of Inspection BatchesSampling RateNumber of Hit BatchesHit Rate
201929,57310.68%2.09%41543187.66%5216.35%
Note 1: The predicted number of batches referred to the number of batches extracted from the 2019 border inspection application with sampling records and inspection results.
Table 8. Differences between EL V.2 and EL V.1 modeling methods.
Table 8. Differences between EL V.2 and EL V.1 modeling methods.
Model DifferencesEL V.1EL V.2Description
Screening of characteristic risk factorsSingle-factor analysis and stepwise regression were used to screen characteristic factors using simple statistical methods.
  • Elastic Net
  • New data were added monthly to participate in modeling, and then key factors were selected for actual participation.
Prevent factor collinearity. Make the remaining factors more independent and important.
Add algorithms5 algorithms7 algorithmsWhen the prediction effect of multiple models is reduced, the AUC > 50% can still be retained for integration to improve the robustness of the model.
Adjust model parametersFβ regulated the sampling inspection rate.
Five models had consistent values.
Fβ regulated the sampling inspection rate
Seven models were independently adjusted.
The sampling rate was regulated, and the elasticity was set at 2–8%.
Table 9. Analysis table of elastic Fβ threshold regulation for each classification model.
Table 9. Analysis table of elastic Fβ threshold regulation for each classification model.
BetaPPV
(or Sampling Inspection
Rejection Rate)
Recall
(or Hit Rate of Unqualified Products)
Number of Border Inspection Application BatchesNumber of Sampling Inspection BatchesSampling RateFβ Threshold
Bagging-NBBagging-C5.0Bagging-CARTBagging-LRBagging-RF
120.00%33.33%24952.01%0.940.760.480.460.68
1.220.00%33.33%24952.01%0.940.760.480.460.68
1.415.38%66.67%249135.22%0.940.460.480.460.64
1.621.43%100.00%249145.62%0.940.460.480.460.6
1.821.43%100.00%249145.62%0.940.460.480.460.6
221.43%100.00%249145.62%0.310.460.480.460.6
2.221.43%100.00%249145.62%0.310.460.480.460.6
2.417.65%100.00%249176.83%0.310.460.430.460.6
2.616.67%100.00%249187.23%0.310.330.430.460.6
2.816.67%100.00%249187.23%0.310.330.430.460.6
38.82%100.00%2493413.65%0.310.330.280.120.6
3.25.45%100.00%2495522.09%0.310.180.280.120.25
Table 10. Analysis table of all classification models using fixed Fβ threshold regulation.
Table 10. Analysis table of all classification models using fixed Fβ threshold regulation.
Recommended ThresholdPPV
Unqualified Rate of Sampling Inspection
Recall
Identification Rate of Unqualified Sampling
Sampling Rate
0.3913.29%100.00%9.29%
0.4013.29%100.00%9.29%
0.4114.79%100.00%8.41%
0.4215.16%100.00%7.96%
0.4315.45%100.00%7.56%
0.4415.45%100.00%7.56%
0.4520.43%100.00%6.19%
Table 11. AUC comparison between EL V.1 and EL V.2 models.
Table 11. AUC comparison between EL V.1 and EL V.2 models.
Model
Revision
AUC of Algorithm
Bagging-
EN
Bagging-
LR
Bagging-GBMBagging-
BN
Bagging-
RF
Bagging-C5.0Bagging-
CART
EL V.1-69.03%-53.43%57.40%63.20%63.17%
EL V.263.39%63.13%62.67%62.13%61.41%57.72%49.40%
Note: EL is an abbreviation for Ensemble Learning.
Table 12. EL V.1 and EL V.2 model prediction index evaluation table.
Table 12. EL V.1 and EL V.2 model prediction index evaluation table.
Year of AnalysisModel RevisionNumber of AlgorithmsRecallPPVF1
2020EL V.1525.00%2.47%4.49%
EL V.2758.33%4.38%8.14%
Note: EL V.1 analysis interval: 8 April 2020–2 August 2020; EL V.2 analysis interval: 3 August 2020–30 November 2020.
Table 13. Statistical table for border inspection application and sampling over the years.
Table 13. Statistical table for border inspection application and sampling over the years.
General Sampling Inspection Items for Each YearNumber of Inspection Application BatchesOverall Sampling RateEL Sampling RateOverall Rejection RateEL Rejection Rate
202227,07410.90% (2952/27,074)6.48% (1754/27,074)3.01% (89/2952)4.39% (77/1754)
202123,6709.14% (2163/23,670)6.24% (1478/23,670)4.16% (90/2163)6.36% (84/1478)
202026,8236.07% (1629/26,823)2.78% (745/26,823)3.74% (61/1629)5.10% (38/745)
201929,57310.68% (3157/29,573)-2.09% (66/3157)-
Note: On 8 April 2020, the general border sampling inspection for S-type food was adjusted from random sampling inspection to EL V.1 predictive sampling inspection. On 3 August 2020, it was converted to EL V.2 for prediction sampling inspection.
Table 14. Statistical performance evaluation of the ensemble learning prediction model before and after its launch.
Table 14. Statistical performance evaluation of the ensemble learning prediction model before and after its launch.
General Sampling Inspection and Evaluation Items for Each YearAnnual Overall Sampling InspectionEL Sampling Inspection
Annual Rejection Rate
(Number of Unqualified Pieces/Total Number of Sampled Pieces)
p ValueEL Rejection Rate
(Number of EL Unqualified Pieces/Number of EL Sampled Pieces)
p Value
20223.01% (89/2952)0.022 *4.39% (77/1754)0.000 ***
20214.16% (90/2163)0.000 ***6.36% (84/1478)0.000 ***
20203.74% (61/1629)0.001 **5.10% (38/745)0.000 ***
20192.09% (66/3157) --
Note: The chi-square test was used to evaluate whether there was a significant impact on the evaluation results in the years before and after the launch (2019). “*” means p < 0.05; “**” means p < 0.01; “***” means p < 0.001.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, L.-Y.; Liu, F.-M.; Weng, S.-S.; Lin, W.-C. EL V.2 Model for Predicting Food Safety Risks at Taiwan Border Using the Voting-Based Ensemble Method. Foods 2023, 12, 2118. https://doi.org/10.3390/foods12112118

AMA Style

Wu L-Y, Liu F-M, Weng S-S, Lin W-C. EL V.2 Model for Predicting Food Safety Risks at Taiwan Border Using the Voting-Based Ensemble Method. Foods. 2023; 12(11):2118. https://doi.org/10.3390/foods12112118

Chicago/Turabian Style

Wu, Li-Ya, Fang-Ming Liu, Sung-Shun Weng, and Wen-Chou Lin. 2023. "EL V.2 Model for Predicting Food Safety Risks at Taiwan Border Using the Voting-Based Ensemble Method" Foods 12, no. 11: 2118. https://doi.org/10.3390/foods12112118

APA Style

Wu, L. -Y., Liu, F. -M., Weng, S. -S., & Lin, W. -C. (2023). EL V.2 Model for Predicting Food Safety Risks at Taiwan Border Using the Voting-Based Ensemble Method. Foods, 12(11), 2118. https://doi.org/10.3390/foods12112118

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop