Author:
Wu Tailin,Fischer Ian,Chuang Isaac L.,Tegmark Max
Abstract
The Information Bottleneck (IB) method provides an insightful and principled approach for balancing compression and prediction for representation learning. The IB objective I ( X ; Z ) - β I ( Y ; Z ) employs a Lagrange multiplier β to tune this trade-off. However, in practice, not only is β chosen empirically without theoretical guidance, there is also a lack of theoretical understanding between β , learnability, the intrinsic nature of the dataset and model capacity. In this paper, we show that if β is improperly chosen, learning cannot happen—the trivial representation P ( Z | X ) = P ( Z ) becomes the global minimum of the IB objective. We show how this can be avoided, by identifying a sharp phase transition between the unlearnable and the learnable which arises as β is varied. This phase transition defines the concept of IB-Learnability. We prove several sufficient conditions for IB-Learnability, which provides theoretical guidance for choosing a good β . We further show that IB-learnability is determined by the largest confident, typical and imbalanced subset of the examples (the conspicuous subset), and discuss its relation with model capacity. We give practical algorithms to estimate the minimum β for a given dataset. We also empirically demonstrate our theoretical conditions with analyses of synthetic datasets, MNIST and CIFAR10.
Funder
Casey and Family Foundation, the Foundational Questions Institute and the Rothberg Family Fund for Cognitive Science
Subject
General Physics and Astronomy
Reference36 articles.
1. The information bottleneck method;Tishby;arXiv,2000
2. A Mathematical Theory of Communication
3. Information bottleneck for Gaussian variables;Chechik;J. Mach. Learn. Res.,2005
4. Meta-Gaussian information bottleneck;Rey,2012
5. Deep variational information bottleneck;Alemi;arXiv,2016
Cited by
12 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. A Survey on Information Bottleneck;IEEE Transactions on Pattern Analysis and Machine Intelligence;2024-08
2. Compressed Data Sharing Based On Information Bottleneck Model;ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP);2022-05-23
3. Adversarial Information Bottleneck;IEEE Transactions on Neural Networks and Learning Systems;2022
4. A Provably Convergent Information Bottleneck Solution via ADMM;2021 IEEE International Symposium on Information Theory (ISIT);2021-07-12
5. Optimal prediction with resource constraints using the information bottleneck;PLOS Computational Biology;2021-03-08