RDE: A novel approach to improve the classification performance and expressivity of KDB

PLoS One. 2018 Jul 23;13(7):e0199822. doi: 10.1371/journal.pone.0199822. eCollection 2018.

Abstract

Bayesian network classifiers (BNCs) have demonstrated competitive classification performance in a variety of real-world applications. A highly scalable BNC with high expressivity is extremely desirable. This paper proposes Redundant Dependence Elimination (RDE) for improving the classification performance and expressivity of k-dependence Bayesian classifier (KDB). To demonstrate the unique characteristics of each case, RDE identifies redundant conditional dependencies and then substitute/remove them. The learned personalized k-dependence Bayesian Classifier (PKDB) can achieve high-confidence conditional probabilities, and graphically interpret the dependency relationships between attributes. Two thyroid cancer datasets and four other cancer datasets from the UCI machine learning repository are selected for our experimental study. The experimental results prove the effectiveness of the proposed algorithm in terms of zero-one loss, bias, variance and AUC.

MeSH terms

  • Bayes Theorem
  • Bias
  • Classification / methods*
  • Data Mining / methods*
  • Data Mining / standards
  • Machine Learning*
  • Software*

Grants and funding

The authors received no specific funding for this work.