Abstract
Kullback–Leibler divergence KL(p,q) is the standard measure of error when we have a true probability distribution p which is approximate with probability distribution q. Its efficient computation is essential in many tasks, as in approximate computation or as a measure of error when learning a probability. In high dimensional probabilities, as the ones associated with Bayesian networks, a direct computation can be unfeasible. This paper considers the case of efficiently computing the Kullback–Leibler divergence of two probability distributions, each one of them coming from a different Bayesian network, which might have different structures. The paper is based on an auxiliary deletion algorithm to compute the necessary marginal distributions, but using a cache of operations with potentials in order to reuse past computations whenever they are necessary. The algorithms are tested with Bayesian networks from the bnlearn repository. Computer code in Python is provided taking as basis pgmpy, a library for working with probabilistic graphical models.
Subject
General Physics and Astronomy
Reference19 articles.
1. A comparison of structural distance measures for causal Bayesian network models;de Jongh,2009
2. Learning Bayesian networks from big data with greedy search: Computational complexity and efficient implementation;Scutari;Stat. Comput.,2019
3. Learning Bayesian networks with local structure, mixed variables, and exact algorithms;Talvitie;Int. J. Approx. Reason.,2019
4. Consistent learning Bayesian networks with thousands of variables;Natori,2017
5. Information Theory and Statistics;Kullback,1968
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献