You are currently viewing a new version of our website. To view the old version click .
Entropy
  • Article
  • Open Access

16 September 2017

Attribute Value Weighted Average of One-Dependence Estimators

,
,
and
1
School of Mechanical Engineering and Electronic Information, China University of Geosciences, Wuhan 430074, China
2
School of Mechanical Engineering and Electronic Information, Wuhan University of Engineering Science, Wuhan 430200, China
3
Department of Computer Science, China University of Geosciences, Wuhan 430074, China
4
Hubei Key Laboratory of Intelligent Geo-Information Processing, China University of Geosciences, Wuhan 430074, China

Abstract

Of numerous proposals to improve the accuracy of naive Bayes by weakening its attribute independence assumption, semi-naive Bayesian classifiers which utilize one-dependence estimators (ODEs) have been shown to be able to approximate the ground-truth attribute dependencies; meanwhile, the probability estimation in ODEs is effective, thus leading to excellent performance. In previous studies, ODEs were exploited directly in a simple way. For example, averaged one-dependence estimators (AODE) weaken the attribute independence assumption by directly averaging all of a constrained class of classifiers. However, all one-dependence estimators in AODE have the same weights and are treated equally. In this study, we propose a new paradigm based on a simple, efficient, and effective attribute value weighting approach, called attribute value weighted average of one-dependence estimators (AVWAODE). AVWAODE assigns discriminative weights to different ODEs by computing the correlation between the different root attribute value and the class. Our approach uses two different attribute value weighting measures: the Kullback–Leibler (KL) measure and the information gain (IG) measure, and thus two different versions are created, which are simply denoted by AVWAODE-KL and AVWAODE-IG, respectively. We experimentally tested them using a collection of 36 University of California at Irvine (UCI) datasets and found that they both achieved better performance than some other state-of-the-art Bayesian classifiers used for comparison.

1. Introduction

A Bayesian network (BN) is a graphical model that encodes probabilistic relationships among all variables, where nodes represent attributes, edges represent the relationships between the attributes, and directed arcs can be used to explicitly represent the joint probability distribution. Bayesian networks are often used for classification. Unfortunately, it has been proved that learning the optimal BN structure from an arbitrary BN search space of discrete variables is an non-deterministic polynomial-time hard (NP-hard) problem [].One of the most effective BN classifiers, in the sense that its predictive performance is competitive with state-of-the-art classifiers [], is the so-called naive Bayes (NB). NB is the simplest form of BN classifiers. It runs on labeled training instances and is driven by the conditional independence assumption that all attributes are fully independent of each other given the class. The NB classifier has the simple structure shown in Figure 1, where every attribute A i (every leaf in the network) is independent from the rest of the attributes, given the state of the class variable C (the root in the network). However, it is obvious that the conditional independence assumption in NB is often violated in real-world applications, which will affect its performance with complex attribute dependencies [,].
Figure 1. An example of naive Bayes (NB).
In order to effectively mitigate the assumption of independent attributes for the NB classifier, appropriate structures and approaches are needed to manipulate independence assertions [,,]. Among the numerous Bayesian learning approaches, semi-naive Bayesian classifiers which utilize one-dependence estimators (ODEs) have been shown to be able to approximate the ground-truth attribute dependencies; meanwhile, the probability estimation in ODEs is effective, thus leading to excellent performance. Representative approaches include tree-augmented naive Bayesian (TAN) [], hidden naive Bayes (HNB) [], aggregating one-dependence estimators (AODE) [], weighted average of one-dependence estimators (WAODE) [].
To the best of our knowledge, few studies have focused on attribute value weighting in terms of Bayesian networks []. The attribute value weighting approach is a more fine-grained weighting approach, e.g., when dealing with the pregnancy-related disease classification problem and trying to analyze the importance of gender attributes. It is obvious that the value of the male gender attribute has no effect on the class value (pregnancy-related disease), whereas the value of the female gender attribute has a great effect on the class value. It will be interesting to study whether a better performance can be achieved by combining attribute value weighting with the ODEs. The resulting model which combines attribute value weighting with the ODEs inherits the effectiveness of ODEs; meanwhile, this approach is a new paradigm of weighting approach in classification learning.
In this study, we propose a new paradigm based on a simple, efficient, and effective attribute value weighting approach, called attribute value weighted average of one-dependence estimators (AVWAODE). Our AVWAODE approach is well balanced between the ground-truth dependencies approximation and the effectiveness of probability estimation. We extend the current classification learning utilizing ODEs into the next level by introducing a new set of weight space to the problem. It is a new approach to calculating the discriminative weights for ODEs using the filter approach. We assume that the significance of each ODE can be decomposed, and in the structure of highly predictive ODE, the different root attribute value should be strongly associated with the class. Based on these assumptions, we assign a different weight to each ODE by computing the correlation between the root attribute value and the class. In order to correctly measure the amount of correlation between the root attribute value and the class, our approach uses two different attribute value weighting measures: the Kullback–Leibler (KL) measure and the information gain (IG) measure, and thus two different versions are created, which are simply denoted by AVWAODE-KL and AVWAODE-IG, respectively. We conducted two groups of extensive empirical comparisons with many state-of-the-art Bayesian classifiers using 36 University of California at Irvine (UCI) datasets published on the main website of the WEKA platform [,]. Extensive experiments show that both AVWAODE-KL and AVWAODE-IG can achieve better performance than these state-of-the-art Bayesian classifiers used to compare.
The rest of the paper is organized as follows. In Section 2, we summarize the existing Bayesian networks learning utilizing ODEs. In Section 3, we propose our AVWAODE approach. In Section 4, we describe the experimental setup and results in detail. In Section 5, we draw our conclusions and outline the main directions for our future work.

3. AVWAODE Approach

The remarkable performance of semi-naive Bayesian classifiers utilizing ODEs suggests that ODEs are well balanced between the attribute dependencies assumption and the effectiveness of probability estimation. To the best of our knowledge, few studies have focused on attribute value weighting in terms of Bayesian networks []. Each attribute takes on a number of discrete values, and each discrete value has different importance with respect to the target. Therefore, the attribute value weighting approach is a more fine-grained weighting approach. It is interesting to study whether a better performance can be achieved by exploiting the ODEs while using the attribute value weighting approach. The resulting model which combines attribute value weighting with the ODEs inherits the effectiveness of ODEs; meanwhile, this approach is a new paradigm of weighting approach in Bayesian classification learning.
In this study, we extend the current classification learning utilizing ODEs into the next level by introducing a new set of weight space to the problem. Therefore, this study proposes a new dimension of weighting approach by assigning weights according to attribute values. We propose a new paradigm based on a simple, efficient, and effective attribute value weighting approach, called attribute value weighted average of one-dependence estimators (AVWAODE). Our AVWAODE approach is well balanced between the ground-truth dependencies approximation and the effectiveness of probability estimation. It is a new method for calculating discriminative weights for ODEs using the filter approach. The basic assumption of our AVWAODE approach is that the root attribute value should be strongly associated with the class in the structure of highly predictive ODE, and in ODEs; when a certain root attribute value is observed, it gives a certain amount of information to the class. The more information a root attribute value provides to the class, the more important the root attribute value becomes.
In AVWAODE, a special ODE is built for each attribute. Namely, each attribute is set as the root attribute once and each special ODE is assigned a different weight. Different from the existing WAODE approach which assigns different weights to different ODEs according to the relationship between the root attribute and the class, our AVWAODE approach assigns discriminative weights to different ODEs by computing the correlation between the root attribute value and the class. Since the weight of each special ODE is associated with the root attribute value, AVWAODE uses Equation (5) to classify a test instance x.
c ( x ) = arg   max c C ( i = 1 m W i , a i P ( a i , c ) k = 1 , k i m P ( a k | a i , c ) i = 1 m W i , a i ) ,
where W i , a i is the weight of the ODE when the root attribute A i has the attribute value a i . The base probabilities P ( a i , c ) and P ( a k | a i , c ) are estimated using the m-estimate as follows:
P ( a i , c ) = F ( a i , c ) + 1 . 0 / ( n i n c ) n + 1 . 0 ,
P ( a k | a i , c ) = F ( a k , a i , c ) + 1 . 0 / n k F ( a i , c ) + 1 . 0 ,
where F ( ) is the frequency with which a combination of terms appears in the training data, n is the number of training instances, n i is the number of values of the root attribute A i , n k is the number of values of the leaf attribute A k , and n c is the number of classes.
Now, the only question left to answer is how to use a proper measure which can correctly quantify the amount of correlation between the root attribute value and the class. To address this question, our approach uses two different attribute value weighting measures: the Kullback–Leibler (KL) measure and the information gain (IG) measure, and thus two different versions are created, which are simply denoted by AVWAODE-KL and AVWAODE-IG, respectively. Note that KL and IG are two widely used measures already presented in the existing literature; we just redefine them following our proposed methods.

3.1. AVWAODE-KL

The first candidate that can be used in our approach is the KL measure, which is a widely used method for calculating the correlation between the class variable and an attribute value []. The KL measure calculates the distances of different posterior probabilities from the prior probabilities. The main idea is that the KL measure becomes more reliable as the frequency of a specific attribute value increases. It originally was proposed in the article []. The KL measure calculates the average mutual information between the events c and the attribute value a i with the expectation taken with respect to a posteriori probability distribution of C []. The KL measure uses Equation (8) to quantify the information content of an attribute value a i :
K L ( C | a i ) = c P ( c | a i ) log P ( c | a i ) P ( c ) ,
where K L ( C | a i ) represents the attribute value-class correlation.
It is quite an intuitive argument that if the root attribute value a i gets a higher KL value, the ODE will deserve a higher weight. Therefore, we use the K L ( C | a i ) between the root attribute value a i and the class C to define the weight W i , a i of the ODE as:
W i , a i = K L ( C | a i ) = c P ( c | a i ) log P ( c | a i ) P ( c ) ,
where a i is the attribute value of the root attribute, which is the parent of all other attributes. An example of AVWAODE-KL is shown in Figure 6.
Figure 6. An example of attribute value weighted average of one-dependence estimators (AVWAODE)-KL.
The detailed learning algorithm for our AVWAODE approach using the KL measure (AVWAODE-KL) is described briefly as Algorithm 1.
Algorithm 1: AVWAODE-KL (D, x)
Input: A training dataset D and a test instance x
Output: Class label c ( x ) of x
   1. For each class value c
   2.     Compute P ( c ) from D
   3.     For each attribute value a i
   4.       Compute F ( a i , c ) from D
   5.       Compute P ( a i , c ) by Equation (6)
   6.       For each attribute value a k ( k i )
   7.          Compute F ( a k , a i , c ) from D
   8.          Compute P ( a k | a i , c ) by Equation (7)
   9. For each attribute value a i
   10.     Compute F ( a i ) from D
   11.     Compute W i , a i by Equation (9)
   12. Estimate the class label c ( x ) for x using Equation (5)
   13. Return the class label c ( x ) for x
Compared to the well-known AODE, AVWAODE-KL requires some additional training time to compute the weights of all the attribute values. According to the algorithm given above, the additional training time complexity for computing these weights is only O ( n c m v ) ; where n c is the number of classes, m is the number of attributes, and v is the average number of values for an attribute. Therefore, AVWAODE-KL has only a training time complexity of O ( n m 2 + n c m v ) , where n is the number of training instances. Note that n c v is generally much less than n m in reality. If we only take the highest order term, the training time complexity is still O ( n m 2 ) , which is the same as AODE. In addition, the classification time complexity of AVWAODE-KL is O ( n c m 2 ) , which is also the same as AODE. All of this means that AVWAODE-KL is simple and efficient.

3.2. AVWAODE-IG

The second candidate that can be used in our approach is the IG measure, which is a widely used method for calculating the importance of attributes [,]. The C4.5 approach [] uses the IG measure to construct the decision tree for classifying objects. The information gain used in the C4.5 approach [] is defined as:
H ( C ) - H ( C | A i ) = a i P ( a i ) c P ( c | a i ) l o g P ( c | a i ) - c P ( c ) l o g P ( c ) .
Equation (10) computes the difference between the entropy of a priori distribution and the entropy of a posteriori distribution of the class, and the C4.5 approach uses the difference value as the metric for deciding the branching node. The value of the IG measure can represent the significance of the attribute; the bigger the value is, the greater impact of the attribute when classifying the test instance. The correlation between the attribute A i and the class C can be measured by Equation (10). Since we require discriminative power for an attribute value, we cannot use Equation (10) directly as the measure of the discriminative power for an attribute value. The IG measure calculates the correlation between an attribute value a i and the class variable C by:
I G ( C | a i ) = c P ( c | a i ) l o g P ( c | a i ) - c P ( c ) l o g P ( c ) .
It is quite an intuitive argument that if the root attribute value a i gets a higher IG value, the ODE will deserve a higher weight. Therefore, we use the I G ( C | a i ) between the root attribute value a i and the class C to define the weight W i , a i of the ODE as:
W i , a i = I G ( C | a i ) = c P ( c | a i ) l o g P ( c | a i ) - c P ( c ) l o g P ( c ) ,
where a i is the attribute value of the root attribute, which is the parent of all other attributes. An example of AVWAODE-IG is shown in Figure 7.
Figure 7. An example of AVWAODE-IG.
The training time complexity and the classification time complexity of AVWAODE-IG is the same as AVWAODE-KL. In order to save space, we do not repeat the detailed analysis here.

4. Experiments and Results

In order to validate the classification performance of AVWAODE, we ran our experiments on all the 36 UCI datasets [] published on the main web site of Weka platform [,]. In our experiments, missing values are replaced with the modes and means of the corresponding attribute values from the available data. Numeric attribute values are discretized using the unsupervised ten-bin discretization implemented in Weka platform. Additionally, we manually delete three useless attributes: the attribute “Hospital Number” in the dataset “colic.ORIG”, the attribute “instance name" in the dataset “splice”, and the attribute “animal” in the dataset “zoo”.
We compare the performance of AVWAODE with some state-of-the-art Bayesian classifiers: TAN [], HNB [], AODE [], WAODE [], NB [] and A2DE []. We conduct extensive empirical comparisons in two groups with these state-of-the-art models in terms of the classification accuracy. In the first group, we compared AVWAODE-KL with all of these models, and in the second group, we compared AVWAODE-IG with these competitors.
Table 1 and Table 2 show the detailed results of the comparisons in terms of the classification accuracy. All of the classification accuracy estimates were obtained by averaging the results from 10 separate runs in a stratified ten-fold cross-validation. We then conducted corrected paired two-tailed t-tests at the 95% significance level [] in order to compare our AVWAODE-KL and AVWAODE-IG with each of their competitors: TAN, HNB, AODE, WAODE, NB, and A2DE. The averages and the W i n / T i e / L o s e (W/T/L) values are summarized at the bottom of the tables. Each entry’s W/T/L in the tables implies that, compared to their competitors, AVWAODE-KL and AVWAODE-IG win on W datasets, tie on T datasets, and lose on L datasets. The average (arithmetic mean) of each algorithm across all datasets provides a gross indicator of the relative performance in addition to the other statistics.
Table 1. Accuracy comparisons for AVWAODE-KL versus TAN, HNB, AODE, WAODE, NB, and A2DE.
Table 2. Accuracy comparisons for AVWAODE-IG versus TAN, HNB, AODE, WAODE, NB, and A2DE.
Then, we employ a corrected paired two-tailed t-test with the p = 0 . 05 significance level [] to compare each pair of algorithms. Table 3 and Table 4 show the summary test results with regard to AVWAODE-KL and AVWAODE-IG, respectively. In these tables, for each entry i ( j ) , i is the number of datasets on which the algorithm in the column achieves higher classification accuracy than the algorithm in the corresponding row, and j is the number of datasets on which the algorithm in the column achieves significant wins with the p = 0 . 05 significance level [] with regard to the algorithm in the corresponding row.
Table 3. Summary test results with regard to AVWAODE-KL.
Table 4. Summary test results with regard to AVWAODE-IG.
Table 5 and Table 6 show the ranking test results with regard to AVWAODE-KL and AVWAODE-IG, respectively. In these tables, the first column is the difference between the total number of wins and the total number of losses that the corresponding algorithm achieves compared with all the other algorithms, which is used to generate the ranking. The second and third columns represent the total numbers of wins and losses, respectively.
Table 5. Ranking test results with regard to AVWAODE-KL.
Table 6. Ranking test results with regard to AVWAODE-IG.
According to these comparisons, we can see that both AVWAODE-KL and AVWAODE-IG perform better than TAN, HNB, AODE, WAODE, NB, and A2DE. We summarize the main features of these comparisons as follows.
  • Both AVWAODE-KL and AVWAODE-IG significantly outperform NB with 16 wins and one loss, and they also outperform TAN with 11 wins and zero losses.
  • Both AVWAODE-KL and AVWAODE-IG perform notably better than HNB on six datasets and worse on zero datasets, and also better than A2DE on five datasets and worse on one dataset.
  • Both AVWAODE-KL and AVWAODE-IG are markedly better than AODE with nine wins and zero losses.
  • Both AVWAODE-KL and AVWAODE-IG perform substantially better than WAODE on two datasets and one datset, respectively.
  • Seen from the ranking test results, our proposed algorithms are always the best ones, and NB is the worst one. The overall ranking (descending order) is AVWAODE-KL (AVWAODE-IG), WAODE, A2DE, HNB, AODE, TAN and NB, respectively.
  • The results of these comparisons suggest that assigning discriminative weights to different ODEs by computing the correlation between the root attribute value and the class is highly effective when setting weights for ODEs.
Yet at the same time, based on the accuracy results presented in Table 1 and Table 2, we take advantage of KEEL Data-Mining Software Tool [,] to complete the Wilcoxon signed-ranks test [,] to thoroughly compare each pair of algorithms. The Wilcoxon signed-ranks test is a non-parametric statistical test, which ranks the differences in performances of two algorithms for each dataset, ignoring the signs, and compares the ranks for positive and negative differences. Table 7, Table 8, Table 9 and Table 10 summarize the detailed results of the nonparametric statistical comparisons based on the Wilcoxon test. According to the table, for the exact critical values in the Wilcoxon test at a confidence level of α = 0 . 05 ( α = 0 . 1 ) and with datasets where N = 36 , two algorithms are considered “significantly different" if the smaller of R + and R - is equal to or less than 208 (227), and thus we could reject the null hypothesis. ∘ indicates that the algorithm in the column improves the algorithm in the corresponding row, and • indicates that the algorithm in the row improves the algorithm in the corresponding column. According to these comparisons, we can see that both AVWAODE-KL and AVWAODE-IG performed significantly better than NB, and TAN, and AVWAODE-IG performed even significantly better than A2DE at a confidence level of α = 0 . 1 .
Table 7. Ranks computed by the Wilcoxon test on AVWAODE-KL.
Table 8. Summary of the Wilcoxon test on AVWAODE-KL. ∘: The algorithm in the column improves the algorithm in the corresponding row. •: The algorithm in the row improves the algorithm in the corresponding column. Lower diagonal level of significance α = 0 . 05 ; upper diagonal level of significance α = 0 . 1 .
Table 9. Ranks computed by the Wilcoxon test on AVWAODE-IG.
Table 10. Summary of the Wilcoxon test on AVWAODE-IG. ∘: The algorithm in the column improves the algorithm in the corresponding row. •: The algorithm in the row improves the algorithm in the corresponding column. Lower diagonal level of significance α = 0 . 05 ; upper diagonal level of significance α = 0 . 1 .
Additionally, in our experiments, we have also observed the performance of our proposed AVWAODE in terms of the area under the Receiver Operating Characteristics (ROC) curve (AUC) [,,,]. Table 11 and Table 12 show the detailed comparison results in terms of AUC. From these comparison results, we can find that our proposed AVWAODE is also very promising in terms of the area under the ROC curve. Please note that, in order to save space, we do not present the detailed summary test results, the ranking test results, and the Wilcoxon test results.
Table 11. Area under the ROC curve (AUC) comparisons for AVWAODE-KL versus TAN, HNB, AODE, WAODE, NB, and A2DE.
Table 12. AUC comparisons for AVWAODE-IG versus TAN, HNB, AODE, WAODE, NB, and A2DE.

5. Conclusions and Future Work

Of numerous proposals to improve the accuracy of naive Bayes by weakening its attribute independence assumption, semi-naive Bayesian classifiers which utilize one-dependence estimators (ODEs) have been shown to be able to approximate the ground-truth attribute dependencies; meanwhile, the probability estimation in ODEs is effective. In this study, we propose a new paradigm based on a simple, efficient, and effective attribute value weighting approach, called attribute value weighted average of one-dependence estimators (AVWAODE). Our AVWAODE approach is well balanced between the ground-truth dependencies approximation and the effectiveness of probability estimation. We assign discriminative weights to different ODEs by computing the correlation between the root attribute value and the class. Two different attribute value weighting measures, which are called the Kullback–Leibler (KL) measure and the information gain (IG) measure, are used to quantify the correlation between the root attribute value and the class, and thus two different versions are created, which are simply denoted by AVWAODE-KL and AVWAODE-IG, respectively. Extensive experiments show that both AVWAODE-KL and AVWAODE-IG achieve better performance than some other state-of-the-art ODE models used for comparison.
How to learn the weights is a crucial problem in our proposed AVWAODE approach. An interesting future work will be the exploration of more effective methods to estimate weights to improve our current AVWAODE versions. Furthermore, applying the proposed attribute value weighting approach to improve some other state-of-the-art classification models is another topic for our future work.

Acknowledgments

The work was partially supported by the Program for New Century Excellent Talents in University (NCET-12-0953), the excellent youth team of scientific and technological innovation of Hubei higher education (T201736), and the Open Research Project of Hubei Key Laboratory of Intelligent Geo-Information Processing (KLIGIP201601).

Author Contributions

Liangjun Yu and Liangxiao Jiang conceived the study; Liangjun Yu and Dianhong Wang designed and performed the experiments; Liangjun Yu and Lungan Zhang analyzed the data; Liangjun Yu and Liangxiao Jiang wrote the paper; All authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chickering, D.M. Learning Bayesian networks is NP-Complete. Artif. Intell. Stat. 1996, 112, 121–130. [Google Scholar]
  2. Quinlan, J.R. C4.5: Programs for Machine Learning; Morgan Kaufmann: San Mateo, CA, USA, 1993. [Google Scholar]
  3. Wu, J.; Cai, Z. Attribute Weighting via Differential Evolution Algorithm for Attribute Weighted Naive Bayes (WNB). J. Comput. Inf. Syst. 2011, 7, 1672–1679. [Google Scholar]
  4. Zaidi, N.A.; Cerquides, J.; Carman, M.J.; Webb, G.I. Alleviating Naive Bayes Attribute Independence Assumption by Attribute Weighting. J. Mach. Learn. Res. 2013, 14, 1947–1988. [Google Scholar]
  5. Kohavi, R. Scaling Up the Accuracy of Naive-Bayes Classifiers: A Decision-Tree Hybrid. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, Portland, OR, USA, 2–4 August 1996; AAAI Press: Cambridge, MA, USA, 1996; pp. 202–207. [Google Scholar]
  6. Khalil, E.H. A Noise Tolerant Fine Tuning Algorithm for the Naive Bayesian learning Algorithm. J. King Saud Univ. Comput. Inf. Sci. 2014, 26, 237–246. [Google Scholar]
  7. Hall, M. A decision tree-based attribute weighting filter for naive bayes. Knowl. Based Syst. 2007, 20, 120–126. [Google Scholar] [CrossRef]
  8. Friedman, N.; Geiger, D.; Goldszmidt, M. Bayesian network classifiers. Mach. Learn. 1997, 29, 131–163. [Google Scholar] [CrossRef]
  9. Jiang, L.; Zhang, H.; Cai, Z. A Novel Bayes Model: Hidden Naive Bayes. IEEE Trans. Knowl. Data Eng. 2009, 21, 1361–1371. [Google Scholar] [CrossRef]
  10. Webb, G.; Boughton, J.; Wang, Z. Not So Naive Bayes: Aggregating One-Dependence Estimators. Mach. Learn. 2005, 58, 5–24. [Google Scholar] [CrossRef]
  11. Jiang, L.; Zhang, H.; Cai, Z.; Wang, D. Weighted average of one-dependence estimators. J. Exp. Theor. Artif. Intell. 2012, 24, 219–230. [Google Scholar] [CrossRef]
  12. Lee, C.H. A gradient approach for value weighted classification learning in naive Bayes. Knowl. Based Syst. 2015, 85, 71–79. [Google Scholar] [CrossRef]
  13. Witten, I.H.; Frank, E. Data Mining: Practical Machine Learning Tools and Techniques, 2nd ed.; Morgan Kaufmann: San Francisco, CA, USA, 2005. [Google Scholar]
  14. WEKA The University of Waikato. Available online: http://www.cs.waikato.ac.nz/ml/weka/ (accessed on 15 September 2017).
  15. Qiu, C.; Jiang, L.; Li, C. Not always simple classification: Learning SuperParent for Class Probability Estimation. Expert Syst. Appl. 2015, 42, 5433–5440. [Google Scholar] [CrossRef]
  16. Hall, M. Correlation-based feature selection for discrete and numeric class machine learning. In Proceedings of the 17th International Conference on Machine Learning, Stanford, CA, USA, 29 June–2 July 2000; pp. 359–366. [Google Scholar]
  17. Sahami, M. Learning Limited Dependence Bayesian Classifiers. In Proceedings of the 2th ACM SIGKDD International Conference on Knowledge Disvovery and Data Mining, Portland, OR, USA, 2–4 August 1996; pp. 335–338. [Google Scholar]
  18. Webb, G.; Boughton, J.; Zheng, F. Learning by extrapolation from marginal to full-multivariate probability distributions: Decreasingly naive Bayesian classification. Mach. Learn. 2012, 86, 233–272. [Google Scholar] [CrossRef]
  19. Zheng, F.; Webb, G. Efficient lazy elimination for averaged one-dependence estimators. In Proceedings of the 23rd international conference on Machine learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 1113–1120. [Google Scholar]
  20. Yang, Y.; Webb, G.; Korb, K.; Boughton, J.; Ting, K. To Select or To Weigh: A Comparative Study of Linear Combination Schemes for SuperParent-One-Dependence Estimators. IEEE Trans. Knowl. Data Eng. 2007, 9, 1652–1665. [Google Scholar] [CrossRef]
  21. Li, N.; Yu, Y.; Zhou, Z. Semi-naive Exploitation of One-Dependence Estimators. In Proceedings of the 2009 Ninth IEEE International Conference on Data Mining, Miami, FL, USA, 6–9 December 2009; pp. 278–287. [Google Scholar]
  22. Xiang, Z.; Kang, D. Attribute weighting for averaged one-dependence estimators. Appl. Intell. 2017, 46, 616–629. [Google Scholar] [CrossRef]
  23. Lee, C.H.; Gutierrez, F.; Dou, D. Calculating Feature Weights in Naive Bayes with Kullback-Leibler Measure. In Proceedings of the 11th IEEE International Conference on Data Mining, Vancouver, BC, Canada, 11–14 December 2011; pp. 1146–1151. [Google Scholar]
  24. Kullback, S.; Leibler, R. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  25. Jiang, L. Learning random forests for ranking. Front. Comput. Sci. China 2011, 5, 79–86. [Google Scholar] [CrossRef]
  26. Merz, C.; Murphy, P.; Aha, D. UCI Repository of Machine Learning Databases. In Department of ICS, University of California. 1997. Available online: http://www.ics.uci.edu/mlearn/MLRepository.html (accessed on 15 September 2017).
  27. Nadeau, C.; Bengio, Y. Inference for the generalizaiton error. Mach. Learn. 2003, 52, 239–281. [Google Scholar] [CrossRef]
  28. Alcalá-Fdez, J.; Fernandez, A.; Luengo, J.; Derrac, J.; García, S.; Sánchez, L.; Herrera, F. KEEL Data-Mining Software Tool: Data Set Repository, Integration of Algorithms and Experimental Analysis Framework. J. Mult.-Valued Logic Soft Comput. 2011, 17, 255–287. [Google Scholar]
  29. Knowledge Extraction based on Evolutionary Learning. Available online: http://sci2s.ugr.es/keel/ (accessed on 15 September 2017).
  30. Demsar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  31. Garcia, S.; Herrera, F. An extension on statistical comparisons of classifiers over multiple data sets for all pairwise comparisons. J. Mach. Learn. Res. 2008, 9, 2677–2694. [Google Scholar]
  32. Bradley, P.; Andrew, P. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997, 30, 1145–1159. [Google Scholar] [CrossRef]
  33. Hand, D.; Till, R. A simple generalisation of the area under the roc curve for multiple class classification problems. Mach. Learn. 2001, 45, 171–186. [Google Scholar] [CrossRef]
  34. Ling, C.X.; Huang, J.; Zhang, H. AUC: A statistically consistent and more discriminating measure than accuracy. In Proceedings of the 18th International Joint Conference on Artificial Intelligence, Ribeirao Preto, Brazil, 23–27 October 2006; Morgan Kaufmann: Burlington, MA, USA, 2003; pp. 329–341. [Google Scholar]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.