Abstract
The entropy of a pair of random variables is commonly depicted using a Venn diagram. This representation is potentially misleading, however, since the multivariate mutual information can be negative. This paper presents new measures of multivariate information content that can be accurately depicted using Venn diagrams for any number of random variables. These measures complement the existing measures of multivariate mutual information and are constructed by considering the algebraic structure of information sharing. It is shown that the distinct ways in which a set of marginal observers can share their information with a non-observing third party corresponds to the elements of a free distributive lattice. The redundancy lattice from partial information decomposition is then subsequently and independently derived by combining the algebraic structures of joint and shared information content.
Funder
Australian Research Council
Subject
General Physics and Astronomy
Reference94 articles.
1. An Introduction to Information Theory, International student edition;Reza,1961
2. On the Amount of Information
3. Information Theory and Coding;Abramson,1963
4. Entropy as a measure
5. Information Theory: Coding Theorems for Discrete Memoryless Systems;Csiszar,1981
Cited by
19 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献