Next Article in Journal
2D Anisotropic Wavelet Entropy with an Application to Earthquakes in Chile
Next Article in Special Issue
Averaged Extended Tree Augmented Naive Classifier
Previous Article in Journal
The Non-Equilibrium Statistical Distribution Function for Electrons and Holes in Semiconductor Heterostructures in Steady-State Conditions
Previous Article in Special Issue
A Penalized Likelihood Approach to Parameter Estimation with Integral Reliability Constraints
Article Menu

Export Article

Open AccessArticle
Entropy 2015, 17(6), 4134-4154; doi:10.3390/e17064134

General and Local: Averaged k-Dependence Bayesian Classifiers

1
Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, ChangChun 130012, China
2
School of Software, Jilin University, ChangChun 130012, China
*
Author to whom correspondence should be addressed.
Academic Editors: Carlos Alberto de Bragança Pereira and Adriano Polpo
Received: 4 May 2015 / Revised: 2 June 2015 / Accepted: 9 June 2015 / Published: 16 June 2015
(This article belongs to the Special Issue Inductive Statistical Methods)
View Full-Text   |   Download PDF [1339 KB, uploaded 16 June 2015]   |  

Abstract

The inference of a general Bayesian network has been shown to be an NP-hard problem, even for approximate solutions. Although k-dependence Bayesian (KDB) classifier can construct at arbitrary points (values of k) along the attribute dependence spectrum, it cannot identify the changes of interdependencies when attributes take different values. Local KDB, which learns in the framework of KDB, is proposed in this study to describe the local dependencies implicated in each test instance. Based on the analysis of functional dependencies, substitution-elimination resolution, a new type of semi-naive Bayesian operation, is proposed to substitute or eliminate generalization to achieve accurate estimation of conditional probability distribution while reducing computational complexity. The final classifier, averaged k-dependence Bayesian (AKDB) classifiers, will average the output of KDB and local KDB. Experimental results on the repository of machine learning databases from the University of California Irvine (UCI) showed that AKDB has significant advantages in zero-one loss and bias relative to naive Bayes (NB), tree augmented naive Bayes (TAN), Averaged one-dependence estimators (AODE), and KDB. Moreover, KDB and local KDB show mutually complementary characteristics with respect to variance. View Full-Text
Keywords: k-dependence Bayesian classifier; substitution-elimination resolution; functionaldependency rules of probability k-dependence Bayesian classifier; substitution-elimination resolution; functionaldependency rules of probability
Figures

This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. (CC BY 4.0).

Scifeed alert for new publications

Never miss any articles matching your research from any publisher
  • Get alerts for new papers matching your research
  • Find out the new papers from selected authors
  • Updated daily for 49'000+ journals and 6000+ publishers
  • Define your Scifeed now

SciFeed Share & Cite This Article

MDPI and ACS Style

Wang, L.; Zhao, H.; Sun, M.; Ning, Y. General and Local: Averaged k-Dependence Bayesian Classifiers. Entropy 2015, 17, 4134-4154.

Show more citation formats Show less citations formats

Related Articles

Article Metrics

Article Access Statistics

1

Comments

[Return to top]
Entropy EISSN 1099-4300 Published by MDPI AG, Basel, Switzerland RSS E-Mail Table of Contents Alert
Back to Top