Abstract
Properties of data distributions can be assessed at both global and local scales. At a highly localized scale, a fundamental measure is the local intrinsic dimensionality (LID), which assesses growth rates of the cumulative distribution function within a restricted neighborhood and characterizes properties of the geometry of a local neighborhood. In this paper, we explore the connection of LID to other well known measures for complexity assessment and comparison, namely, entropy and statistical distances or divergences. In an asymptotic context, we develop analytical new expressions for these quantities in terms of LID. This reveals the fundamental nature of LID as a building block for characterizing and comparing data distributions, opening the door to new methods for distributional analysis at a local scale.
Funder
Japan Society for the Promotion of Science
Australian Research Council
Subject
General Physics and Astronomy
Reference91 articles.
1. Divergence measures for statistical data processing—An annotated bibliography;Basseville;Signal Process.,2013
2. Houle, M.E. (2017, January 4–6). Local Intrinsic Dimensionality I: An Extreme-Value-Theoretic Foundation for Similarity Applications. Proceedings of the International Conference on Similarity Search and Applications, Munich, Germany.
3. Bailey, J., Houle, M.E., and Ma, X. (October, January 29). Relationships Between Local Intrinsic Dimensionality and Tail Entropy. Proceedings of the Similarity Search and Applications—Proc. of the 14th International Conference, SISAP 2021, Dortmund, Germany.
4. Lee, D.D., Sugiyama, M., von Luxburg, U., Guyon, I., and Garnett, R. (2016). Advances in Neural Information Processing Systems 29 (NIPS 2016), Curran Associates Inc.
5. Reducing multidimensional two-sample data to one-dimensional interpoint comparisons;Maa;Ann. Stat.,1996
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献