Next Article in Journal
Study on the Influence of Clay Content on the Freeze–Thaw Characteristics and Mechanisms of Solidified Low-Liquid-Limit Clay
Previous Article in Journal
Enhanced Privacy-Preserving Architecture for Fundus Disease Diagnosis with Federated Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Ensemble Strategy Based on Surprisingly Popular Algorithm and Classifier Prediction Confidence

School of Computer Science, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(6), 3003; https://doi.org/10.3390/app15063003
Submission received: 19 January 2025 / Revised: 7 March 2025 / Accepted: 7 March 2025 / Published: 10 March 2025

Abstract

:
Traditional ensemble methods rely on majority voting, which may fail to recognize correct answers held by a minority in scenarios requiring specialized knowledge. Therefore, this paper proposes two novel ensemble methods for supervised classification, named Confidence Truth Serum (CTS) and Confidence Truth Serum with Single Regression (CTS-SR). The former is based on the principles of Bayesian Truth Serum (BTS) and introduces classification confidence to calculate the prior and posterior probabilities of events, enabling the recovery of correct judgments provided by a confident minority beyond majority voting. CTS-SR further simplifies the algorithm by constructing a single regression model to reduce computational overhead, making it suitable for large-scale applications. Experiments are conducted on multiple binary classification datasets to evaluate CTS and CTS-SR. Experimental results demonstrate that, compared with existing ensemble methods, both of the proposed methods significantly outperform baseline algorithms in terms of accuracy and F1 scores. Specifically, there is an average improvement of 2–6% in accuracy and an average increase of 2–4% in F1 score. Notably, on the Musk and Hilly datasets, our method achieves a 5% improvement compared to the traditional majority voting approach. Particularly on the Hilly dataset, which generally exhibits the poorest classification performance and poses the greatest prediction challenges, our method demonstrates the best discriminative performance. validating the importance of confidence as a feature in ensemble learning.

1. Introduction

Ensemble learning has gained increasing prominence in machine learning in recent years, emerging as a pivotal tool for enhancing model performance. It refers to methodologies that combine predictions from multiple independent models to improve overall performance. While single models may underperform when addressing complex problems due to overfitting or inherent biases, the core philosophy of ensemble learning lies in leveraging model diversity to overcome performance bottlenecks caused by overfitting or bias in individual models, thereby achieving superior accuracy, generalization capability, and robustness compared to single models [1,2].
With the advancement of deep learning, ensemble learning techniques and their application scenarios have continued to expand and evolve. Traditional majority voting, despite its simplicity, often delivers suboptimal results when handling complex or specialized data and tasks, necessitating more flexible and effective strategies to integrate the strengths of diverse models [3].
Ensemble learning is often regarded as the machine learning counterpart of the “wisdom of crowds” [4]. Conventional methods, such as majority voting and weighted averaging, operate under the assumption that the majority of predictions are correct. Random forest, for instance, exemplifies the effectiveness of majority voting [5]. While these methods perform well in many scenarios, their efficacy may diminish when individual classifiers exhibit significant diversity and minority classifiers demonstrate higher confidence in specific samples. This is particularly evident in cases requiring specialized knowledge to determine ground-truth answers, where the correct judgments of minority classifiers risk being overshadowed by erroneous predictions from the majority. Such limitations can lead to degraded overall performance.
Recent advances in crowd wisdom theory have introduced novel perspectives for ensemble learning. To address the aforementioned constraints, we propose an ensemble strategy that integrates Bayesian Truth Serum (BTS) with the concept of surprise popularity [6,7,8]. BTS leverages two critical pieces of information: participants’ direct answers to questions and their predictions about the group’s responses. By computing the prior probability of events and deriving posterior probabilities for participants, BTS evaluates their credibility based on posterior probabilities and confidence levels. This approach diverges from traditional majority-centric methods by prioritizing participants who provide correct answers while accurately predicting others’ responses, as such individuals are more likely to possess genuine expertise. Recent work by McCoy et al. [9] further extends this theory through a hierarchical Bayesian framework for collective decision-making.
The surprise popularity algorithm shares similarities with BTS, as it also analyzes participants’ answers and their inferences about group responses. For challenging, less-understood problems, this method identifies experts as those who exhibit high confidence in their answers and recognize that most others may fail to provide correct solutions. Under high-difficulty scenarios, the algorithm prioritizes opinions from such experts. Recent studies demonstrate its unique advantages in complex tasks like ranking problems [10].
Our strategy further incorporates classification confidence, which evaluates the reliability of each classifier’s prediction for specific samples. By quantifying confidence levels, we determine which classifiers merit greater trust in their predictions [11]. When classifiers disagree on a sample, confidence metrics enable weighted aggregation of their judgments. In the context of surprise popularity, confidence also adjusts posterior probabilities to identify correct minority predictions, thereby enhancing the ensemble’s ability to extract accurate results from minority classifiers. Experimental results demonstrate that our method outperforms traditional ensemble strategies (e.g., random forest, AdaBoost) and individual supervised learning models across diverse binary datasets, achieving higher accuracy and F1 scores.
This study refines the BTS framework by introducing classification confidence and integrating surprise popularity principles. First, we construct diverse base classifiers and compute their confidence levels for each sample. Using these confidences, we derive prior and posterior probabilities, comparing their discrepancies to determine whether to prioritize correct minority predictions over majority errors. Additionally, we propose a single regression model strategy with reduced spatial overhead. The key contributions of this work include the following:
  • Designing an ensemble strategy that integrates classification confidence with surprise popularity;
  • Proposing a prior regression method based on classification confidence and defining posteriors using confidence metrics;
  • Validating the algorithm on public datasets, with experimental results demonstrating superior performance over conventional ensemble methods. Our research highlights the potential of this approach and provides new directions for future studies and applications in ensemble learning.
The remainder of this paper is organized as follows. Section 2 reviews related work. Section 3 details the proposed ensemble strategy combining classification confidence and surprise popularity. Section 4 presents experimental evaluations. Finally, we conclude the paper.

2. Related Work

The ensemble learning is inspired by people’s decision-making processes in real life, and uses the wisdom of crowds: by integrating the opinions of a large number of individuals, it is often possible to obtain better decision-making solutions than individuals. The goal is to build a collection of individual classifiers. By voting on the decisions of individual classifiers in the ensemble, highly accurate classification decisions can be obtained. Ensemble learning has excellent results in solving classification, regression, prediction, decision-making, and other problems [12]. The following summarizes the relevant research works on ensemble learning from the perspectives of algorithm development and application fields.

2.1. Development of Ensemble Learning

In recent years, with the continuous emergence of various new technologies, decision making has become increasingly difficult. The concept of ensemble learning has evolved continually in research; more and more new ideas have been integrated with it, driving the continuous optimization of model performance based on ensemble learning. The theories and methods of ensemble learning have gradually become more complex and diversified. With integrating different theories and methods, it not only expands the depth of application of ensemble learning but also plays a significant role in improving model generalization ability, enhancing robustness, and addressing complex problems.
Continuing the evolution of boosting-family algorithms, Tianqi Chen [13] proposed the XGBoost algorithm in 2016, which is an improvement on the GBDT (gradient boosting decision trees) algorithm. This algorithm fits the second-order derivative of the loss function from the previous round and adds a regularization term beyond the objective function, achieving significant improvements in accuracy. Upon its release, the model received widespread recognition and application in the academic community. In 2017, Microsoft [14] introduced a new boosting framework model called LightGBM. This algorithm builds on GBDT by incorporating gradient-based one-side sampling (GOSS) and exclusive feature bundling (EFB) techniques. It eliminates most of the data with smaller gradients, reducing the impact of low-gradient parts on information gain. Compared to the XGBoost algorithm, this approach significantly enhances computational speed while maintaining accuracy.
In 2004, Geoffrey et al. [15], based on the hypothesis that the improvement in accuracy of multistrategy ensemble methods is due to the increased diversity among the ensemble members formed, investigated techniques that utilize simple ensemble learning methods to explore the relationship between diversity among members and ensemble error. They developed three multistrategy ensemble learning techniques. The results demonstrated positive effects, reducing the error rate and validating their hypothesis.
Zhang et al. [16] established a semi-supervised support vector machine (S3VM) model composed of base learners, which are based on different interference factors. They also proposed an ensemble method, named EnsembleS3VM, which leverages clustering evaluation techniques. Meanwhile, they combined two multiclass classification strategies to reduce runtime and improve classification accuracy. This method can be used for classification decision making on imbalanced data and data with unknown distributions.
Yu et al. [17] proposed a progressive subspace ensemble learning (PSEL) method that simultaneously considers data sample space and feature space. PSEL first employs random subspace technology to generate a set of subspaces. Then, it introduces a progressive selection process based on a new cost function that incorporates both current and long-term information to sequentially select classifiers. Finally, a weighted voting scheme is used to summarize the predicted labels and obtain the final results. The results of nonparametric tests indicate that PSEL performs well on most real datasets and outperforms many state-of-the-art classifier ensemble methods.
Svargiv et al. [18] studied a novel ensemble learning algorithm based on learning automata called LABEL, which adjusts the contribution of each base learner by dynamically allocating influence coefficients. This adjustment can be made according to different conditions of the problem space, thereby enhancing the performance of the overall ensemble model and improving detection accuracy. This method avoids treating base learners as equal, ignoring the differences and advantages among them. In subsequent research, this method can be applied to problems with nonlinear and unpredictable data. Table 1 demonstrates a comparison of the differences between our study and existing research.

2.2. Applications of Ensemble Learning

As ensemble learning techniques continue to advance and mature, their application scenarios have gradually expanded to various fields, such as medical testing, security monitoring, and more. An increasing number of industries and enterprises are leveraging predictive methods derived from ensemble learning principles to detect and analyze target objects. By integrating the advantages of multiple models, these methods have significantly improved detection accuracy and robustness, especially in target recognition and prediction tasks in complex environments. The widespread application of this technology has not only enhanced system reliability but also provided a more precise basis for decision support and risk management.
Khanh Pham et al. [19] applied ensemble learning to develop a classification model that accurately estimates slope stability. The model constructs an ensemble classifier through two primary techniques: parallel learning and sequential learning. By identifying and learning failure modes of slopes under different conditions, the model is able to assess slope safety, thereby improving the accuracy of disaster warnings.
Luis Gutierrez-Espinoza et al. [20] used ensemble learning to detect fake reviews, integrating random forest, bagging, and AdaBoost, with support vector machines (SVMs) and multilayer perceptrons (MLP) serving as weak classifiers with optimized hyperparameters. The research results indicate that after optimizing the hyperparameters, the accuracy of using MLP as a standalone classifier is 68.2%, while the accuracy of the classifier integrating MLP with AdaBoost reaches 77.3%. This demonstrates that ensemble learning methods can effectively classify fake reviews. The study used a self-collected dataset of fake restaurant reviews to train the classifiers.
Yang et al. [21] proposed an image denoising method based on ensemble learning. This method utilizes different types of image prior knowledge to construct an ensemble denoising model composed of two denoising bases with shrinkage transformations. Through a resampling scheme, multiple denoising bases are sequentially combined multiple times to construct an integrated image denoiser. After performance validation, it can be verified that the ensemble image denoiser operates effectively.
Wang et al. [22] investigated a multikernel ensemble learning method for software defect classification and prediction. By leveraging ensemble learning techniques, they developed a multikernel classifier that combines the advantages of multikernel learning and ensemble learning. Using the widely-used NASA MDP dataset as test data, the results demonstrated that MKEL significantly outperforms several representative and state-of-the-art defect prediction methods.
An et al. [23] proposed a three-layer deep ensemble learning framework named DELearning for the classification and diagnosis of Alzheimer’s disease. This framework aims to integrate multisource data and tap into “expert wisdom” using deep learning algorithms. Neural networks serve as the base classifiers. In the voting layer, two sparse autoencoders are trained for feature learning. In the stacking layer, a nonlinear feature weighting method based on deep belief networks is proposed to rank the base classifiers. In the optimization layer, oversampling and threshold shifting are used to address cost-sensitive issues. An optimized prediction is obtained through an ensemble of probability predictions based on similarity calculations. Based on clinical datasets, the detection accuracy is 4% higher than that of six well-known ensemble classifiers.
An overview of current research both domestically and internationally reveals that ensemble learning and ensemble methods are continuously evolving and improving, tending towards higher accuracy, efficiency, and robustness. They play a crucial role in an increasing number of interdisciplinary fields. However, regarding the core mechanism behind many current ensemble learning methods—the majority voting mechanism—Luo and Liu [8] expressed dissent. They argue that in situations where special knowledge is required to obtain the correct answer, the result selected by the majority voting mechanism may not fully represent the correct answer. In cases where the majority is incorrect, they propose deciding whether to adopt the minority answer by comparing the prior and posterior probabilities of the data points. Table 2 presents the various parameters of different ensemble learning methods.

3. Methodology

3.1. Problem Definition

Let D = ( x i , l i ) i = 1 M be the sample set, D t r D is training set, where ( x i , l i ) is the i -th training sample, x i = X i 1 , X i 2 , X i T R T , l i Φ = ω 1 , ω 2 , ω L , M is the number of training sample, T is the number of features for sample, and L is the number of classes; F = f 1 , f 2 f N represents the set of N base classifiers trained on D t r . For each sample x i , f j independently performs the classification, that is,
x i f j O i j = f j x i
The ensemble learning algorithm, by applying a given strategy 𝓣 , provides the final classification result O i for the sample.
O i = 𝓣 O i 1 , O i 2 , O i N

3.2. Bayesian Truth Serum

The ensemble learning approach in this paper is primarily based on Bayesian Truth Serum (BTS), a mechanism design method grounded in information theory, which incentivizes participants to report truthful information without external validation. Originally proposed by Drazen Prelec in 2004 [6], BTS aims to address certain information asymmetry issues, such as in surveys or prediction markets. The core of this method involves asking participants two questions:
  • The participant’s own answer to the question.
  • The participant’s prediction of how others will answer the question.
Then, Bayesian statistical methods are used to calculate the consistency between each participant’s answer and their prediction of others’ responses. This approach can identify participants who not only provide the correct answer but also accurately predict others’ responses, thereby rewarding them. This process encourages participants to provide more truthful and accurate information.
The basic steps of BTS are as follows:
  • Collect answer: Participants answer a question and provide a probability distribution of how likely they think others will choose each option.
  • Calculate prior probability: Calculate the prior probability for each answer option, typically the average distribution of responses provided by all participants.
P k = i = 1 N I i , k N
Here, I i , k is an indicator function where the i-th participant chooses the k-th option, and N represents the number of participants.
III.
Calculate posterior probability: Calculate each participant’s posterior probability based on their report and the prior probability.
Q i , k = P k y i , k j = 1 N P j y i , j
Here, y i , k is the probability that the i-th participant believes others will choose the k-th option.
IV.
Scoring Mechanism: Calculate a score based on each participant’s posterior probability and their provided confidence; a higher score indicates that their answer is more likely to be close to the truth.

3.3. Confidence Truth Serum

3.3.1. Algorithm Overview

Confidence Truth Serum (CTS) introduces the concept of Bayesian Truth Serum into the ensemble strategy. It uses base classifiers to represent individual decisions within a group and establishes a predictive mechanism that allows the truth to reside in the hands of a knowledgeable minority. This mechanism prevents the wisdom of a single individual within the group from being disregarded due to majority voting, thus recovering the correct answers held by the minority and providing a comprehensive understanding of the uncertainty surrounding the event. Unlike the approach in reference [6], this ensemble strategy considers both the classification results and the confidence level of the base classifiers.
Specifically, for each sample x i , the basic classifier f j classifies independently, yielding the class output result O i j and the prediction probability C i j , that is,
x i f j O i j = f j x i , C i j = P j ( O i j | x i )
Based on comprehensive strategy 𝓣 , the final classification result is provided for the sample: O i .
O i = 𝓣 ( O i 1 , C i 1 ) , ( O i 2 , C i 2 ) , , ( O i N , C i N )
C i j , often referred to as classification confidence in many machine learning models, primarily serves to quantify each classifier’s certainty in a given prediction, thereby optimizing the decision-making process in ensemble learning. The calculation method for confidence is effectively and reasonably defined by us based on the fundamental principles of the classifiers used and the relevant information provided during their prediction processes. By clarifying the calculation methods for the confidence levels of each base learner on the data, we ensure that each classifier has a confidence value for each data point, which is then applied to subsequent computational processes.
Based on the above definitions, we can obtain the output results of all base classifiers on the training set.
C = P 1 O 1 1 | x 1 P 2 O 1 2 | x 1 P N O 1 N | x 1 P 1 O 2 1 | x 2 P 2 O 2 2 | x 2 P N O 2 N | x 2 P 1 O M 1 | x M P 2 O M 2 | x M P N O M N | x M
O = f 1 x 1 f 2 x 1 f N x 1 f 1 x 2 f 2 x 2 f N x 2 f 1 x M f 2 x M f N x M
Similarly, for a test sample x p , the classification results of all base classifiers can be obtained:
O x p = ( f 1 x p , f 2 x p , f N x p ) = O p 1 , O p 2 , O p N
C x p = ( P 1 O 1 1 | x p , P 1 O 1 1 | x p , P 1 O 1 1 | x p ) = C p 1 , C p 2 , C p N

3.3.2. Ensemble Strategy 𝓣

(1)
According to the explanation of the Bayesian Truth Serum algorithm in Section 3.2, we need to calculate the prior probability P and posterior probability Q for each answer option (i.e., the classification result for each sample).
Prior probability and posterior probability are concepts rooted in statistics. The prior refers to the estimated likelihood of an event occurring based on prior knowledge or historical information, before any data are observed. It reflects an initial belief about the occurrence of an event in the absence of new evidence. The posterior probability, on the other hand, is the updated probability of the event after observing the data, representing the belief after taking new evidence into account and making a prediction. Therefore, in this paper, the prior probability P is defined as the result of each base classifier in the ensemble F predicting the classification outcomes of other base classifiers, also referred to as peer prediction information. The posterior probability Q is calculated by aggregating the category prediction probabilities (classification confidence) of all base classifiers for an unknown event.
Taking a supervised binary classification task as an example, the calculation process is as follows:
Based on the matrix O, for the k -th base classifier, calculate the consistency ratio of the other N 1 classifiers agreeing with its classification result according to Equation (11), denoted as y i k (representing the belief score of the k -th classifier for the i -th sample):
y i k = j k ,   O i j = O i k C i j j k C i j
The result y i k is used as the new label for each training sample, and C i k is added as an additional one-dimensional feature to the sample, forming the feature set x i = X 1 , X 2 , X 3 X n , C i k . This forms a new training dataset D k = x i , y i k , which is then divided into D k + and D k according to the original label of the samples, “1” or “0”.
Here, confidence is added as a new feature because traditional regression model training usually utilizes feature vectors by optimizing an objective function, such as using the least squares error to fit the model y ^ = w T + b . However, in ensemble learning, the classification confidence C i k of each base learner for a sample is also an important source of information. Adding classification confidence as a new one-dimensional feature increases the dimensionality of the training samples, allowing the regression model to utilize sample information more comprehensively, leading to a more optimal parameter model and, thus, improving prediction accuracy.
To avoid the phenomenon where directly applying Equation (11) results in the prior and posterior probabilities always being identical, it is necessary to approximate the prior by training a regression model (see reference [6]); therefore, we, respectively, train two regression models for each individual classifier based on the two training sets D k + and D k .
We repeat the above process for each base classifier.
For each test sample x i , l i , we calculate the prior probability of label 1 using the corresponding regression model based on the actual prediction results of each classifier.
p ¯ k x i = p k + x i                     ,   i f   O i k = 1 1 p k x i         , i f   O i k = 0  
After obtaining the predicted regression results from all individual classifiers, we calculate the prior and posterior probabilities for each sample:
P x i , 1 : = k p ¯ k x i K
Q x i , 1 : = k , O i k = 1 C i k k = 1 N C i k
If P x i , 1 < Q x i , 1 , the surprising answer 1 will be considered as the true answer. Otherwise, the final classification result is considered to be 0.
Unlike Bayes’ theorem, the posterior probability in this paper is not an update or replacement of the prior probability but, rather, a comparison with it. Moreover, the two probability values are derived based on classification confidence, providing a more accurate reflection of the reliability of event estimation. Applying the “surprisingly popular” concept, we determine whether to adopt the latest decision result. In this process, estimated probabilities are replaced by the confidence level of individual decisions, placing greater emphasis on those individuals within the minority who possess the correct answer, thereby ensuring more accurate decision making. In fact, the calculation formula in [8] represents the extreme case where confidence equals 1.
(2)
It can be seen that the process of calculating the “surprising” degree of label 1 in the above binary classification task essentially assumes the answer to be 1, which is conceptually similar to hypothesis testing. Similarly, we can assume the answer to be 0 during the algorithm’s calculation process to obtain the corresponding ensemble result. In fact, the above process of calculating the prior and posterior probabilities for label 1 or label 0 does not affect the final classification result. Taking the construction of a multiple linear regression model as an example, we provide a brief proof under the extreme case where confidence is equal to 1:
Proof:
Set
k 1 = k 1 f k x t = 1
k 0   = k 1 f k x t   =   0
K   =   k 1   +   k 0
For the regression model of each basic classifier, we set
p k + = β k + θ k X + ε k
p k   = β k   +   θ k   X   +   ε k
For Label = 1:
if   f x t = 1 , p ¯ k x t = p k +
else   if   f x t = 0 , p ¯ k ( x t ) = 1 p k
  P X t   ,   1   =   k p ¯ k ( x t ) K = k 1 β k + ( k 1 θ k ) X t + k 1 ε k k 0 β k ( k 0 θ k ) X t k 0   ε k + k 0 K = k 0 + k 1 β k + k 1 ε k k 0 β k k 0   ε k + ( k 1 θ k k 0 θ k ) X t K
As the parameters β , θ , and ε are fixed after the regression model is trained, we can set
n = k 1 β k + k 1 ε k k 0 β k k 0   ε k
m = ( k 1 θ k k 0 θ k ) X t
  P X t ,   1   = k 0 + n + m K
Q X t , 1 = k 1 K = 1 k 0 k
For Label = 0:
P X t   ,   0 = k 0 β k + k 0   ε k k β k k 1 θ k + ( k 0 θ k k 1 θ k ) X t + k 1 K
= k 1 n m K
Q X t   , 0 = k 0 K = 1 k 1 K
According to the role for Label = 1:
when   a n s = 1 ,   that   is   P < Q ,   k 0 + n + m K < 1 k 0 K ,
equals   2 k 0 + n + m < K a
According to the role for Label = 0:
when   a n s = 1 ,   that   is   P > Q ,   k 1 n + m K > 1 k 1 K ,
equals   2 k 1 n m > K ;
that   is , 2 k 0 + n + m < K b
This is equivalent to Equation (a), indicating that the classification rules are the same.

3.4. Confidence Truth Serum with Single Regression (CTS-SR)

As previously mentioned, in a binary classification problem, the CTS algorithm requires the construction of a total of 2 N regression models for N base classifiers. This incurs significant memory overhead, especially in the context of ensemble learning where the number of classifiers is large. Additionally, each model is trained only on samples with specific labels, so when the training sample set is small, the regression model has limited data for training. To reduce memory overhead and make full use of the training samples to fit the regression model, this paper proposes a variant of the CTS algorithm, called CTS-SR, which only requires the construction of a single regression model based on all data.
Specifically, in this algorithm, we add the classifier’s classification result, named f k x i , for each sample as a one-dimensional feature, in addition to the confidence named C i k as a new feature, that is, x i = x 1 , x 2 , x 3 x n , C i k , f k x i . Still, we let y i k , whose meaning is as above, be the new label for each training sample (the calculation process for y i k is the same as above). This forms a new training dataset D k = x i , y i k , and each individual classifier trains a regression model p k on D k .
For each testing sample x i , y i T , we calculate the prior and posterior probabilities for the sample using each classifier’s regression model.
P X i : = k p k x i K
Q X i : = k C i k O i k = 1 k = 1 N C i k
The decision rule of this new method differs from CTS because this approach does not incorporate the concept of a predetermined answer. If we still apply P X i < Q X i , then whether the answer is 1 or 0 holds no practical significance.
In fact, the prior probability P in this paper describes the consistency of the classifier ensemble’s decision for a new sample based on past learning experience. If P > Q , this indicates a high level of agreement among the classifiers, in which case we follow the majority voting result. However, when P < Q , it suggests that the decision on this event differs from previous instances, indicating a higher level of surprise. To obtain the result from knowledgeable individuals who hold the correct answer, we still apply the “surprisingly popular” concept. This concept implies that in a survey on an obscure question, those who know the true answer often predict a lower consensus rate. We denote this value as R i k , In other words, these individuals truly know the answer and believe that most of the group does not. We therefore select these individuals’ judgments as the final answer.
As the machine cannot determine whether its answer is correct, we use confidence to express this. We assume that classifiers with high confidence in their classification result and low predicted consensus are more likely to hold the correct answer. Therefore, the decision rule is expressed as follows:
O i = MajorityVote O i 1 , O i 2 , , O i N       ,   i f   P > Q O i k     |   a r g m a x k C i k R i k     ,   i f   P < Q  

3.5. Analysis

The CTS and CTS-SR algorithms proposed in Section 3.3 and Section 3.4 introduce the concept of unexpected popularity based on the Bayesian Truth Serum theory. The former integrates the confidence levels of the classifiers and decides whether to follow the majority or minority results by comparing the prior and posterior probabilities of the event itself. Similarly, CTS-SR also emphasizes confidence as a critical metric, fully considering the answers of expert classifiers with specialized knowledge in scenarios with high decision difficulty, making it more aligned with real-world prediction scenarios. Both methods break through the traditional strategies that focus on majority results, thoroughly considering the expertise and confidence levels of decision makers, and constructing a more comprehensive decision-making process and scenario, thereby enabling a more integrated and detailed judgment of unknown events. However, the proposed algorithms still have certain limitations. As the calculation of relevant parameters during training and prediction mostly relies on the confidence levels of the classifiers, defining and computing confidence poses a challenge and may result in increased time overhead. Additionally, when the classification results of all classifiers are highly consistent, CTS may struggle to recover unexpectedly correct answers, which requires further extensive research.

4. Experiment

This section shows the experimental content and results of the application of the CTS and CTS-SR algorithms to binary classification tasks. By using common binary classifiers to build an ensemble learning set, the two methods are applied to a total of eight processed binary classification task datasets. The HMTS in [6] is compared with the common integrated learning method and the self-recurrence HMTS in important indexes. Experimental results show that our algorithm is significantly better than HMTS and other comparison algorithms on most datasets, and we also carry out corresponding theoretical analysis of the results.

4.1. Experimental Setup

In terms of datasets, we consider six binary data commonly found in UCI and two datasets, magazine and o-beauty, used for the detection of naval reviews. For the two text datasets, magazine and All-Beauty, we used the Bert model to extract feature words and convert them into numerical data. Then principal component analysis was adopted to reduce dimension to a certain extent, and all datasets were normalized. We chose 50% as the training set and 50% as the test set. Detailed information is shown in Table 3.
In order to analyze the effectiveness of the Confidence Truth Serum method in the binary classification task, we selected three common ensemble learning methods (majority voting, random forest, and AdaBoost), and HMTS as the comparison group. These algorithms have greatly improved the prediction effect of the classifier.
Majority vote is a simple ensemble learning method that combines the prediction results of multiple classifiers and determines the final classification results through voting. It is suitable for multiple classifiers with similar performance. By synthesizing the advantages of multiple models, better results can be obtained than a single model. Its classifier settings are consistent with CTS.
Random forest is an ensemble learning method based on decision trees, which makes classification or regression by constructing multiple decision trees and combining their prediction results. It has the characteristics of sample randomization and feature randomization, and has strong anti-overfitting ability; it is especially suitable for processing high-dimensional data. In order to ensure the same number of classifiers, a random forest composed of 15 decision trees is selected.
AdaBoost (adaptive boosting) is an ensemble learning method based on weighted combined weak classifiers. It improves the accuracy of the model by gradually adjusting the weights in the dataset, focusing on correcting misclassified data. In the experiment, 15 weak decision trees are used as the base classifier.
HMTS (Heuristic Machine Truth Serum): The idea is to train a regression model for each classifier to predict the percentage of agreement on predictions from other classifiers for each specific data point. After obtaining the classifier’s prediction label and the predicted peer prediction information, the prior is again approximated using the predicted peer prediction information for each classifier, and the average is calculated and compared to the posterior. Its base classifier setup is consistent with CTS.
In terms of evaluation indicators, we mainly choose the traditional accuracy and F 1 score. T P means that the actual label is positive and the classification is positive; T N means that the actual is negative and the classification is negative, F P means that the actual is negative and the classification is positive; F N means that the actual is positive and the classification is negative. The calculation formula of the evaluation index is as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
R e c a l l = T P T P + F N
P r e c i s i o n = T P T P + F P
F 1 = 2 × P r e c i s I o n × R e c a l l P r e c i s I o n + R e c a l l
In the experiment, we consider five common binary classifiers that can obtain their classification confidence, namely, gradient lift tree (gbt), logistic regression, random forest (RF), support vector machine (SVM), and multilayer perceptron (MLP). In order to enhance the robustness of the model, we selected three different noise rates to flip data labels to improve the prediction ability of the learner in the face of unknown data and enrich the types of learners. Based on this, we obtained a total of 15 different classifiers using our algorithm experiments and other integration methods.
As the idea of this paper is mainly based on confidence, we perform the following calculation for the classification confidence of the above classifier:
For random forests, weighted voting is used to calculate the confidence:
C = ( w i × v i ) w i
where w i represents the weight of the i -th decision tree and the predicted votes of the tree for this category, and v i represents the number of votes that the tree assigns to that category.
For logistic regression, the output probability represents the probability that each sample belongs to a certain class. We output the maximum result in each class of probabilities, indicating confidence:
C = max ( p 1 , p 2 , ⋯⋯ ,   p i )
where p i represents the prediction probability for the i -th class, and the function max() represents the maximum value for all terms in parentheses.
For support vector machine, distance measurement algorithm is combined with sigmoid function. First, decision_function is used to calculate the distance, dis, between the test sample and the support vector, and then we use the function sigmoid to map the distance to a value in the interval [0,1], representing the probability that the sample belongs to a positive class, to represent the confidence degree:
C = 1 1 + e d i s
where d i s is the distance from the test sample to the support vector.
For multilayer perceptrons, we output the maximum result output in each class of probabilities to represent the confidence:
C = max ( p 1 , p 2 , ⋯⋯ ,   p i )
where p i , the prediction probability for the i -th class, is expressed, and the max ( ) function maximizes all terms in parentheses.
For the gradient lifting tree, we sample x i , calculate its distance to the k -th classifier, where d i s i represents the distance from the sample point to the classification hyperplane, i = 1   ,   2   ,   . . . , k , and bring it into the classification function s i g m o i d ( x ) to obtain the confidence before normalization p r o b + i :
p r o b + i = s i g m o i d d i s i = 1 1 + e d i s i
Then the p r o b + i is normalized, that is, the probability of the sample being Class i is obtained, in another word, it is the confidence degree:
C = p r o b + i j = 1 k p r o b + j

4.2. Experimental Results

4.2.1. Experiment 1: Comparison with Baseline Algorithms

To verify the effectiveness and superiority of the two CTS algorithms proposed in this paper across various binary classification tasks, we tested all eight datasets with the CTS algorithms and other baseline algorithms. We recorded the accuracy and F1 scores for each dataset, with Table 4 showing the accuracy of each ensemble algorithm on different datasets and Table 5 displaying their F1 scores. In both tables, bolded values represent the best result on each dataset. It can be seen that our proposed CTS algorithm achieved the highest accuracy on five datasets, while the CTS-SR algorithm achieved the highest accuracy on two datasets. Across both algorithms, they achieved the highest F1 score on six datasets. The results significantly outperform traditional ensemble learning strategies and HMTS, clearly demonstrating the applicability and superiority of our proposed ensemble strategy in various binary classification tasks.
The traditional majority voting and random forest ensemble algorithms heavily rely on the majority results within the group, meaning that the answers affirmed by most base learners significantly influence the final decision. Our proposed CTS algorithm, to some extent, breaks through the limitations of traditional classification, especially by considering and correctly recovering the correct results provided by minority classifiers even when the majority of classifiers agree. As shown in Table 4, CTS improves accuracy by more than 4% compared to majority voting on the musk and Hilly datasets, and by 6% and 9% compared to AdaBoost on the magazine and musk datasets, respectively. This indicates that our method can correctly classify more instances. The data in Table 5 further demonstrate that CTS improves the F1 score by 2–7% compared to traditional random forest and AdaBoost ensemble algorithms. This suggests that our proposed method not only accurately classifies more instances but also maintains a good balance between precision and recall, reflecting the completeness and balance of the approach. This performance improvement is attributed to our introduction of classifier confidence as a crucial factor in the decision-making process, allowing the correction of answers by a few highly confident expert classifiers when the majority provides incorrect results.

4.2.2. Experiment 2: Comparison with Single Classifiers

To further validate the versatility of the two CTS algorithms, we continued to compare them with several common individual classifiers in supervised learning on the same datasets and recorded their performance. Table 6 and Table 7, respectively, present the accuracy and F1 scores of each method across the datasets, with the table settings consistent with Table 4. It is evident that the CTS algorithm achieved the highest accuracy on five datasets, while the CTS-SR algorithm also attained the highest accuracy on two datasets. Together, the two algorithms achieved the highest F1 scores on three datasets. Their performance is significantly superior to that of individual machine learning classifiers. This demonstrates the broad superiority of the strategies we have proposed.

4.2.3. Experiment 3: Friedman–Nemenyi Test

To verify that our algorithm has a significant performance difference compared to the baseline on multiple datasets, we conduct the Friedman–Nemenyi test on various ensemble algorithms. By ranking the performance of each algorithm on each dataset and calculating the average rank of each algorithm across all datasets, we then compute the Friedman statistic χ F 2 using the formula shown in Equation (22). Here, n represents the number of datasets, k represents the number of algorithms, and r m j represents the average rank of the j -th algorithm. Based on the degrees of freedom and the significance level, we calculated the p-value. If the p-value is less than the significance level α , we reject the null hypothesis, indicating that there is a significant difference in algorithm performance. After the Friedman test rejects the null hypothesis, we further use the Nemenyi test to determine which algorithms have significant differences in performance. The critical difference (CD) value is used to judge whether the difference in average ranks between two algorithms is significant. The CD value is calculated using the formula shown in Equation (23), where q α , k is the critical value obtained from the Nemenyi test table. If the difference in average ranks between two algorithms is greater than the CD value, their performance is considered significantly different.
In our experiments, we set n = 8 , k = 6 , degrees of freedom d f = k 1 = 5 , and α = 0.05. From the table, we obtained q α , k = 2.850. Based on these values, we conducted the Friedman–Nemenyi test using accuracy and F1 score, as described below.
χ F 2 = 12 n k k + 1 j = 1 k r m j 2 k k + 1 2 4
C D = q α , k k k + 1 6 n
(1)
Friedman–Nemenyi Test Based on Accuracy
Based on the accuracy values in Table 4, the Friedman statistic was calculated to be 17.125, yielding a p-value of 0.004, which is less than α. This indicates that there are significant differences among the compared ensemble algorithms. Subsequently, the Nemenyi post hoc test was conducted, and the critical difference (CD) was calculated to be 2.666. Figure 1 displays the average ranks of the ensemble algorithms for accuracy. If the two horizontal line segments in the figure do not overlap, it signifies a significant performance difference between the two algorithms. It can be seen that the horizontal lines corresponding to CTS do not intersect with those of RF and AdaBoost. Additionally, the difference calculated from their labeled average rank values is greater than the CD value of 2.666. Moreover, CTS has the smallest average rank value, indicating that our algorithm significantly outperforms the other two, further underscoring the effectiveness of our method.
(2)
Friedman–Nemenyi Test Based on F1 score
Based on the F1 score values in Table 5, the Friedman statistic was calculated to be 14.411, yielding a p-value of 0.0132, which is less than α. This indicates that there are significant differences among the compared ensemble algorithms. Subsequently, the Nemenyi post hoc test was conducted, and the critical difference (CD) was calculated to be 2.666. Similarly, it can be seen from Figure 2 that our algorithm significantly outperforms other ensemble methods in terms of F1 score, further illustrating the reliability of our proposed algorithm.

4.2.4. Experiment 4: Calculation and Analysis of Correlation Coefficient

To demonstrate that confidence, as an inherent attribute of the classifier, can significantly enhance the predictive capability of the regression model when used as a sample feature, we calculated and recorded the correlation coefficients between the confidence-based feature vectors of 15 base classifiers and the label vectors on each dataset. These were then compared with vectors composed of other sample features.
As shown in Figure 3, X i j represents the j -th feature of the i -th sample in the dataset, C i k represents the classification confidence of the k -th classifier for the i -th sample, and y i k is the regression value of the k -th classifier for the i -th sample. X i j i = 1,2 , 3 n is the vector composed of the first feature, C i k i = 1,2 , 3 , n is the vector composed of confidence values, and y i k i = 1,2 , 3 , n is the vector of regression values. The correlation coefficient is calculated as follows:
Let the two vectors be defined as follows:
x = x 1 , x 2 , , x n
y = y 1 , y 2 , , y n
x ¯ = i = 1 n x i n
y ¯ = i = 1 n y i n
The formula for the correlation coefficient ρ is as follows:
ρ = i = 1 n x i x ¯ y i y ¯ i = 1 n x i x ¯ 2 × i = 1 n y i y ¯ 2
The specific experimental results are shown in Figure 4. As we are only concerned with the strength of the correlation between vectors, the ranking here considers the absolute values of the correlation coefficients between the two vectors. It can be seen that the confidence feature vector generally has the highest correlation coefficient with the regression values among all feature vectors, significantly enhancing the predictive capability of the regression model and providing support for prior and posterior probability calculations and classification accuracy.
Readers may notice that the confidence vector of the SVM is often negatively correlated with its regression value vector. This indicates that as the confidence of the SVM increases, the consistency ratio of its peer predictions decreases. This situation aligns precisely with the decision rule in our proposed DTS algorithm when P < Q : namely, to heed the opinions of experts who are highly confident in their decision and believe that most others are unaware of the correct answer. This further reinforces the decision-making concept of CTS-SR when P < Q in our proposed method.

4.2.5. Experiment 5: Data Analysis

To further analyze the effectiveness of the CTS algorithm, we recorded the recovery of minority correct answers by CTS compared to majority voting, shown in Table 8. As seen from the previous results, the accuracy of each ensemble algorithm on the Hilly dataset is lower compared to other datasets. However, Table 8 shows that the proportion of correct answers recoverable by CTS is the highest among all datasets, further demonstrating that our proposed strategy can accurately recover minority answers in challenging decision-making scenarios, reducing the drawbacks of majority voting.
In this section, we evaluated CTS and CTS-SR on common binary classification datasets by measuring accuracy and F1 scores, and compared them with popular ensemble learning algorithms as well as traditional single-classifier models in supervised learning. The experimental results demonstrate that our proposed algorithms achieved significant advantages on most datasets, improving accuracy by 4–6% across diverse test data. Furthermore, the Friedman–Nemenyi test was used to rank the ensemble algorithms and revealed significant performance differences, highlighting the superiority and generality of our proposed methods. Additionally, through Experiment 5, we further demonstrated the ability of CTS to recover correct answers compared to the traditional majority voting strategy, indicating that our proposed method effectively addresses the drawbacks of majority decision making and underscores the importance of confidence in ensemble learning.

5. Conclusions

This paper proposes two new ensemble methods for supervised classification, CTS and CTS-SR, which are based on confidence and incorporate the “surprisingly popular” concept. By using confidence levels instead of direct classification results, we generalize HMTS to a more general case. CTS obtains the prior and posterior probabilities of an event by statistically analyzing the prediction consistency and confidence levels of learners. This allows us to determine whether to recover the correct answers held by a minority group, rather than simply applying majority voting. CTS-SR uses confidence to represent the real-world scenario where the correct answer is known, thereby achieving adherence to the correct decisions of a few experts with lower computational overhead. Experiments demonstrate the applicability of two proposed ensemble methods in various situations, and we explain the reasons behind the results. A key aspect of the CTS algorithm in this paper is treating the classification confidence of base learners—an inherent attribute—as an additional one-dimensional feature of the samples. This enables the regression model to utilize sample information more comprehensively, leading to significant improvements in the following aspects:
(1)
Enhancing feature representation capability: It can capture the uncertainty information of sample classifications under different classifiers, enriching the representation of sample features.
(2)
Improving the model’s robustness: Introducing classification confidence increases model diversity, reduces the risk of overfitting, and enhances robustness.
(3)
Boosting prediction performance: Classification confidence reflects the consistency of samples under different classifiers. By integrating this information, samples can be predicted more accurately.

6. Limitations and Future Work

Although the two CTS algorithms we proposed demonstrate significant superiority in various scenarios, they still have certain limitations. Our algorithm introduces classification confidence as a key variable in the computation process. However, some classifiers do not inherently provide a method for calculating confidence on data, which necessitates us to define the confidence calculation formula based on the principles of the classifiers. This poses certain challenges to the accuracy and rationality of our computations. Additionally, the calculation of confidence somewhat increases the time overhead. Moreover, the current CTS algorithm is only applicable to supervised learning classification tasks, and its universality still needs improvement compared to the HMTS algorithm proposed in [8]. Research on semi-supervised and unsupervised classification algorithms based on the CTS concept, as well as applying our method to more real-life decision-making scenarios, remains an important part of our future work.

Author Contributions

H.S. drafted the main manuscript text, prepared Figure 1, Figure 2, Figure 3, Figure 4, Table 1, Table 2, Table 3, Table 4 and Table 5, provided the detailed proof process, proposed and refined the CTS-SR algorithm, supplied most of the datasets used in the experiments, and designed, implemented, and adjusted the experiments. X.W. proposed the ensemble strategy that combines the “surprisingly popular” algorithm with classification confidence, designed the method details, conducted data analysis, designed and adjusted the experiments, and finalized revisions of the manuscript text. Z.Y. formatted the paper according to the requirements, provided part of the datasets used in the experiments, and wrote the first section of the text. Y.Z. researched the current state of ensemble learning, collected relevant references, and provided the baseline. H.Z. provided the calculation process for confidence, implemented parts of the confidence calculation in the experiments. Y.Z. and H.Z. wrote the second section of the text. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Beijing Natural Science Foundation (4202002).

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Polikar, R. Ensemble based systems in decision making. IEEE Circuits Syst. Mag. 2006, 6, 21–45. [Google Scholar] [CrossRef]
  2. Dietterich, T.G. Ensemble Methods in Machine Learning. In Proceedings of the International Workshop on Multiple Classifier Systems, Cagliari, Italy, 21–23 June 2000. [Google Scholar] [CrossRef]
  3. Ganaie, M.A.; Hu, M.; Malik, A.K.; Tanveer, M.; Suganthan, P.N. Ensemble deep learning: A review. Eng. Appl. Artif. Intell. 2022, 115, 105151. [Google Scholar] [CrossRef]
  4. Sagi, O.; Rokach, L. Ensemble learning: A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2018, 8, e1249. [Google Scholar] [CrossRef]
  5. Ho, T.K.; Hull, J.J.; Srihari, S.N. Decision Combination in Multiple Classifiers Systems. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 66–75. [Google Scholar]
  6. Prelec, D. A Bayesian Truth Serum for Subjective Data. Science 2004, 306, 462–466. [Google Scholar] [CrossRef] [PubMed]
  7. Prelec, D.; Seung, H.S.; Mccoy, J. A solution to the single-question crowd wisdom problem. Nature 2017, 541, 532–535. [Google Scholar] [CrossRef] [PubMed]
  8. Luo, T.; Liu, Y. Machine truth serum: A surprisingly popular approach to improving ensemble methods. Mach. Learn. 2022, 112, 789–815. [Google Scholar] [CrossRef]
  9. Mccoy, J.; Prelec, D. A Bayesian Hierarchical Model of Crowd Wisdom Based on Predicting Opinions of Others. Manag. Sci. 2023, 70, 5931–5948. [Google Scholar] [CrossRef]
  10. Hosseini, H.; Mandal, D.; Shah, N.; Shi, K. Surprisingly Popular Voting Recovers Rankings, Surprisingly! In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, Montreal, QC, Canada, 19–27 August 2021. [Google Scholar] [CrossRef]
  11. Schapire, R.E.; Singer, Y. Improved Boosting Algorithms Using Confidence-rated Predictions. Mach. Learn. 1999, 37, 297–336. [Google Scholar] [CrossRef]
  12. Dietterich, T.G. An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization. Mach. Learn. 2000, 40, 139–157. [Google Scholar] [CrossRef]
  13. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD’16), San Francisco, CA, USA, 13–17 August 2016; ACM: New York, NY, USA, 2016; pp. 785–794. [Google Scholar]
  14. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  15. Webb, G.I.; Zheng, Z. Multistrategy Ensemble Learning: Reducing Error by Combining Ensemble Learning Techniques. IEEE Trans. Knowl. Data Eng. 2004, 16, 981–991. [Google Scholar] [CrossRef]
  16. Zhang, D.; Jiao, L.; Bai, X.; Wang, S.; Hou, B. A robust semi-supervised SVM via ensemble learning. Appl. Soft Comput. 2018, 65, 632–643. [Google Scholar] [CrossRef]
  17. Yu, Z.; Wang, D.; You, J.; Wong, H.-S.; Wu, S.; Zhang, J.; Han, G. Progressive Subspace Ensemble Learning. Pattern Recognit. 2016, 60, 692–705. [Google Scholar] [CrossRef]
  18. Svargiv, M.; Masoumi, B.; Keyvanpour, M.R. A new ensemble learning method based on learning automata. J. Ambient. Intell. Humaniz. Comput. 2020, 13, 3467–3482. [Google Scholar] [CrossRef]
  19. Phama, K.; Kim, D.; Park, S.; Choi, H. Ensemble learning-based classification models for slope stability analysis. Catena 2021, 196, 104886. [Google Scholar] [CrossRef]
  20. Gutierrez-Espinoza, L.; Abri, F.; Namin, A.S.; Jones, K.S.; Sears, D.R.W. Ensemble Learning for Detecting Fake Reviews. In Proceedings of the 2020 IEEE 44th Annual Computers, Software, and Applications Conference (COMPSAC), Madrid, Spain, 13–17 July 2020. [Google Scholar] [CrossRef]
  21. Yang, X.; Xu, Y.; Quan, Y.; Ji, H. Image Denoising via Sequential Ensemble Learning. IEEE Trans. Image Process. 2020, 29, 5038–5049. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, T.; Zhang, Z.; Jing, X.; Zhang, L. Multiple Kernel Ensemble Learning for Softwaredefect prediction. Autom. Softw. Eng. 2016, 23, 569–590. [Google Scholar] [CrossRef]
  23. An, N.; Ding, H.; Yang, J.; Au, R.; Ang, T.F.A. Deep ensemble learning for Alzheimer’s disease classification. J. Biomed. Inform. 2020, 105, 103411. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Friedman–Nemenyi test CD graph based on accuracy.
Figure 1. Friedman–Nemenyi test CD graph based on accuracy.
Applsci 15 03003 g001
Figure 2. Friedman–Nemenyi test CD graph based on F1 score.
Figure 2. Friedman–Nemenyi test CD graph based on F1 score.
Applsci 15 03003 g002
Figure 3. Illustration for correlation coefficient computing.
Figure 3. Illustration for correlation coefficient computing.
Applsci 15 03003 g003
Figure 4. The result of ρ on different datasets. Note: The mark named R above the bars means the rank of confidence vector between all the feature vectors.
Figure 4. The result of ρ on different datasets. Note: The mark named R above the bars means the rank of confidence vector between all the feature vectors.
Applsci 15 03003 g004aApplsci 15 03003 g004b
Table 1. Differences between existing studies and this study.
Table 1. Differences between existing studies and this study.
MethodYearModelDatasetLearning
Paradigms
MTS, HMTS, DMTS [8]2023Perceptron, LR, RF, SVM, MLP.UCI Repository.Supervised, semi-supervised.
Decision-tree
algorithm C4.5 [12]
2000Randomization, bagging, and boosting.UCI Repository.Supervised.
XGBoost [13]2016Tree boosting, C4.5, randomization, sparsity-aware algorithm, weighted quantile sketch.Allstate, Higgs Boson,
Yahoo! LTRC, Criteo.
Supervised.
LightGBM [14]2017GOSS, EFB.Allstate Insurance Claim, Flight Delay, LETOR, KDD10, KDD12.Supervised.
Multistrategy Ensemble Learning [15]2004Bagging, wagging, boosting, SASC.UCI Repository.Supervised.
EnsembleS3VM [16]2018SVM, LapSVM, S3VMs, S4VMs.UCI Repository, PolSAR.Semi-supervised.
PSEL [17]2016Decision tree.UCI Repository, alizadeh—2000—v1, armstrong—2002—v1 and other 16 cancer datasets.Supervised.
LAbEL [18]2020SVM, random forest, naïve Bayes, logistic regression.Heart disease dataset, Breast cancer Wisconsin (Diagnostic) dataset, Yelp, Abalone dataset, Gender voice.Supervised.
Our Study2025Gbt, LR, RF, SVM, MLP.UCI Repository, musk, magazine.Supervised.
Table 2. Different studies in specific fields.
Table 2. Different studies in specific fields.
MethodsFieldYearModelDatasetLearning Paradigms
Parallel learning, Sequential learning [19]Slope
stability
2021Homogeneous ensemble:
DT, SVM, ANN, KNN.
Heterogeneous ensemble: KNN, SVM, SGD, GP, QDA, GNB, DT, ANN.
153 slope cases.Supervised.
Bagging,
AdaBoost [20]
Fake
Reviews
2020DT, RF, SVM, XGBT, MLP.Restaurant Dataset.Supervised.
Sequential learning [21]Image
Denoising
2020Local base denoiser,
Nonlocal base denoiser.
400 cropped images of size 180 × 180 from the Berkeley segmentation dataset, BSD68.Supervised.
MKEL [22]Software
defect
prediction
2016SVM.Datasets from NASA MDP.Supervised.
DELearning [23]Alzheimer’s disease
classification
2020Bayes Network, naïve Bayes,
J48 and so on, DBN, RBM.
UDS.Supervised.
Table 3. The information of datasets.
Table 3. The information of datasets.
Datasets# of Instances# of FeaturesNumeric TypeDescription of Content
Australian69014Integer, FloatCredit card applicant information.
German100024IntegerCredit history characteristics.
spam460157Integer, FloatContent characteristics of the emails.
biodeg105541Integer, FloatChemical characteristics of the substances.
magazine237410DoubleComments of magazine in Amazon.
all-beauty508768DoubleCharacters of Beauty products.
musk476166IntegerStructural characteristics of the molecules.
Hilly1212100DoubleCharacteristics of the terrain.
# indicates that the value represents a quantity.
Table 4. Comparison of accuracy among different ensemble algorithms.
Table 4. Comparison of accuracy among different ensemble algorithms.
Accuracy (%)CTSCTS-SRMajorRFAdaBoostHMTS
Australian86.9587.2486.9584.9288.6983.18
German75.674.274.871.472.672.2
spam94.0889.5693.6192.9590.8392.78
biodeg85.9887.1286.3683.5281.2584.84
magazine81.0480.5380.2878.6075.8277.84
all-beauty85.8286.6185.4382.6781.1085.03
musk84.1381.4380.8380.2375.4483.53
Hilly55.9450.8251.6554.6248.1855.11
Table 5. Comparison of F1 score among different ensemble algorithms.
Table 5. Comparison of F1 score among different ensemble algorithms.
F1 Score (%)CTSCTS-SRMajorRFAdaBoostHMTS
Australian84.9485.2384.8482.3186.5979.16
German49.5837.0746.1545.6942.6748.52
spam92.0884.9991.7190.9688.2492.07
biodeg79.8980.3479.7775.6374.0178.60
magazine87.3282.6587.2885.6384.3984.64
all-beauty82.5282.5282.4678.4375.7580.80
musk81.2776.6976.7475.7373.0280.96
Hilly59.4857.5365.0356.6964.2357.63
Table 6. Accuracy result compared to single classifier.
Table 6. Accuracy result compared to single classifier.
Accuracy (%)CTSCTS-SRDTLightGBMMLPSVM
Australian86.9587.2482.6085.7981.7386.66
German75.674.267.273.668.274.0
spam94.0889.5689.4890.6593.6992.13
biodeg85.9887.1279.1684.8485.6086.93
magazine81.0480.5373.7976.4980.8780.62
all-beauty85.8286.6172.8384.2583.4686.61
musk84.1381.4369.4675.7478.7479.64
Hilly55.9450.8254.7853.4650.1949.17
Table 7. F1 Score compared to single classifier.
Table 7. F1 Score compared to single classifier.
F1 Score (%)CTSCTS-SRDTLightGBMMLPSVM
Australian84.9485.2380.0082.6378.7084.86
German49.5837.0737.9549.7846.3744.91
spam92.0884.9986.7693.5191.9589.69
biodeg79.8980.3468.6479.6677.2680.22
magazine87.3282.6582.5887.3887.7587.39
all-beauty82.5282.5273.9780.7879.4183.16
musk81.2776.6962.6372.4075.1773.64
Hilly59.4857.5358.2957.8863.1763.24
Table 8. Number and proportion of recoveries on each dataset.
Table 8. Number and proportion of recoveries on each dataset.
Datasets# of Recovery/# of Test Proportion
Australian3/3450.9%
German12/5002.4%
spam36/23011.6%
biodeg10/5281.9%
magazine33/11872.8%
all-beauty5/2542.0%
musk19/3345.7%
Hilly84/60613.9%
# indicates that the value represents a quantity.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shi, H.; Yuan, Z.; Zhang, Y.; Zhang, H.; Wang, X. A New Ensemble Strategy Based on Surprisingly Popular Algorithm and Classifier Prediction Confidence. Appl. Sci. 2025, 15, 3003. https://doi.org/10.3390/app15063003

AMA Style

Shi H, Yuan Z, Zhang Y, Zhang H, Wang X. A New Ensemble Strategy Based on Surprisingly Popular Algorithm and Classifier Prediction Confidence. Applied Sciences. 2025; 15(6):3003. https://doi.org/10.3390/app15063003

Chicago/Turabian Style

Shi, Haochen, Zirui Yuan, Yankai Zhang, Haoran Zhang, and Xiujuan Wang. 2025. "A New Ensemble Strategy Based on Surprisingly Popular Algorithm and Classifier Prediction Confidence" Applied Sciences 15, no. 6: 3003. https://doi.org/10.3390/app15063003

APA Style

Shi, H., Yuan, Z., Zhang, Y., Zhang, H., & Wang, X. (2025). A New Ensemble Strategy Based on Surprisingly Popular Algorithm and Classifier Prediction Confidence. Applied Sciences, 15(6), 3003. https://doi.org/10.3390/app15063003

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop