Author:
Long YuGuang,Wang LiMin,Sun MingHui
Abstract
Due to the simplicity and competitive classification performance of the naive Bayes (NB), researchers have proposed many approaches to improve NB by weakening its attribute independence assumption. Through the theoretical analysis of Kullback–Leibler divergence, the difference between NB and its variations lies in different orders of conditional mutual information represented by these augmenting edges in the tree-shaped network structure. In this paper, we propose to relax the independence assumption by further generalizing tree-augmented naive Bayes (TAN) from 1-dependence Bayesian network classifiers (BNC) to arbitrary k-dependence. Sub-models of TAN that are built to respectively represent specific conditional dependence relationships may “best match” the conditional probability distribution over the training data. Extensive experimental results reveal that the proposed algorithm achieves bias-variance trade-off and substantially better generalization performance than state-of-the-art classifiers such as logistic regression.
Funder
National Science Foundation of China
the Agreement of Science and Technology Development Project, Jilin Province
Subject
General Physics and Astronomy
Reference40 articles.
1. Learning Bayesian Network Classifiers: Searching in a Space of Partially Directed Acyclic Graphs
2. Idiot’s Bayes not so stupid after all?;Hand;Int. Stat. Rev.,2001
3. BAYDA: Software for Bayesian classification and feature selection;Kontkanen,1998
4. Induction of selective Bayesian classifiers;Langley,1994
5. Bounds for the loss in probability of correct classification under model based approximation;Ekdahl;J. Mach. Learn. Res.,2006
Cited by
15 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献