Learning a Flexible K-Dependence Bayesian Classifier from the Chain Rule of Joint Probability Distribution
As one of the most common types of graphical models, the Bayesian classifier has become an extremely popular approach to dealing with uncertainty and complexity. The scoring functions once proposed and widely used for a Bayesian network are not appropriate for a Bayesian classifier, in which class v...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2015-06-01
|
Series: | Entropy |
Subjects: | |
Online Access: | http://www.mdpi.com/1099-4300/17/6/3766 |
id |
doaj-43897a33d59e48eca6d7a574f70b717b |
---|---|
record_format |
Article |
spelling |
doaj-43897a33d59e48eca6d7a574f70b717b2020-11-25T00:30:08ZengMDPI AGEntropy1099-43002015-06-011763766378610.3390/e17063766e17063766Learning a Flexible K-Dependence Bayesian Classifier from the Chain Rule of Joint Probability DistributionLimin Wang0Haoyu Zhao1Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Changchun 130012, ChinaSchool of Software, Jilin University, Changchun 130012, ChinaAs one of the most common types of graphical models, the Bayesian classifier has become an extremely popular approach to dealing with uncertainty and complexity. The scoring functions once proposed and widely used for a Bayesian network are not appropriate for a Bayesian classifier, in which class variable C is considered as a distinguished one. In this paper, we aim to clarify the working mechanism of Bayesian classifiers from the perspective of the chain rule of joint probability distribution. By establishing the mapping relationship between conditional probability distribution and mutual information, a new scoring function, Sum_MI, is derived and applied to evaluate the rationality of the Bayesian classifiers. To achieve global optimization and high dependence representation, the proposed learning algorithm, the flexible K-dependence Bayesian (FKDB) classifier, applies greedy search to extract more information from the K-dependence network structure. Meanwhile, during the learning procedure, the optimal attribute order is determined dynamically, rather than rigidly. In the experimental study, functional dependency analysis is used to improve model interpretability when the structure complexity is restricted.http://www.mdpi.com/1099-4300/17/6/3766Bayesian classifierchain ruleoptimal attribute orderinformation quantity |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Limin Wang Haoyu Zhao |
spellingShingle |
Limin Wang Haoyu Zhao Learning a Flexible K-Dependence Bayesian Classifier from the Chain Rule of Joint Probability Distribution Entropy Bayesian classifier chain rule optimal attribute order information quantity |
author_facet |
Limin Wang Haoyu Zhao |
author_sort |
Limin Wang |
title |
Learning a Flexible K-Dependence Bayesian Classifier from the Chain Rule of Joint Probability Distribution |
title_short |
Learning a Flexible K-Dependence Bayesian Classifier from the Chain Rule of Joint Probability Distribution |
title_full |
Learning a Flexible K-Dependence Bayesian Classifier from the Chain Rule of Joint Probability Distribution |
title_fullStr |
Learning a Flexible K-Dependence Bayesian Classifier from the Chain Rule of Joint Probability Distribution |
title_full_unstemmed |
Learning a Flexible K-Dependence Bayesian Classifier from the Chain Rule of Joint Probability Distribution |
title_sort |
learning a flexible k-dependence bayesian classifier from the chain rule of joint probability distribution |
publisher |
MDPI AG |
series |
Entropy |
issn |
1099-4300 |
publishDate |
2015-06-01 |
description |
As one of the most common types of graphical models, the Bayesian classifier has become an extremely popular approach to dealing with uncertainty and complexity. The scoring functions once proposed and widely used for a Bayesian network are not appropriate for a Bayesian classifier, in which class variable C is considered as a distinguished one. In this paper, we aim to clarify the working mechanism of Bayesian classifiers from the perspective of the chain rule of joint probability distribution. By establishing the mapping relationship between conditional probability distribution and mutual information, a new scoring function, Sum_MI, is derived and applied to evaluate the rationality of the Bayesian classifiers. To achieve global optimization and high dependence representation, the proposed learning algorithm, the flexible K-dependence Bayesian (FKDB) classifier, applies greedy search to extract more information from the K-dependence network structure. Meanwhile, during the learning procedure, the optimal attribute order is determined dynamically, rather than rigidly. In the experimental study, functional dependency analysis is used to improve model interpretability when the structure complexity is restricted. |
topic |
Bayesian classifier chain rule optimal attribute order information quantity |
url |
http://www.mdpi.com/1099-4300/17/6/3766 |
work_keys_str_mv |
AT liminwang learningaflexiblekdependencebayesianclassifierfromthechainruleofjointprobabilitydistribution AT haoyuzhao learningaflexiblekdependencebayesianclassifierfromthechainruleofjointprobabilitydistribution |
_version_ |
1725327711148179456 |