Author:
Allegra Michele,Facco Elena,Denti Francesco,Laio Alessandro,Mira Antonietta
Abstract
AbstractOne of the founding paradigms of machine learning is that a small number of variables is often sufficient to describe high-dimensional data. The minimum number of variables required is called the intrinsic dimension (ID) of the data. Contrary to common intuition, there are cases where the ID varies within the same data set. This fact has been highlighted in technical discussions, but seldom exploited to analyze large data sets and obtain insight into their structure. Here we develop a robust approach to discriminate regions with different local IDs and segment the points accordingly. Our approach is computationally efficient and can be proficiently used even on large data sets. We find that many real-world data sets contain regions with widely heterogeneous dimensions. These regions host points differing in core properties: folded versus unfolded configurations in a protein molecular dynamics trajectory, active versus non-active regions in brain imaging data, and firms with different financial risk in company balance sheets. A simple topological feature, the local ID, is thus sufficient to achieve an unsupervised segmentation of high-dimensional data, complementary to the one given by clustering algorithms.
Publisher
Springer Science and Business Media LLC
Reference42 articles.
1. Jolliffe, I. T. (ed) Principal component analysis and factor analysis. In Principal Component Analysis, 115–128 (Springer, Berlin, 1986).
2. Roweis, S. T. & Saul, L. K. Nonlinear dimensionality reduction by locally linear embedding. Science 290, 2323–2326 (2000).
3. Tenenbaum, J. B., De Silva, V. & Langford, J. C. A global geometric framework for nonlinear dimensionality reduction. Science 290, 2319–2323 (2000).
4. Grassberger, P. & Procaccia, I. Measuring the strangeness of strange attractors. In The Theory of Chaotic Attractors (eds Hunt, B. R., Li, T.-Y., Kennedy, J. A., & Nusse, H. E.), 170–189 (Springer, Berlin, 2004).
5. Levina, E. & Bickel, P. J. Maximum likelihood estimation of intrinsic dimension. In Advances in Neural Information Processing Systems 17 (eds Saul L. K., Weiss, Y., & Bottou, L.) (MIT Press, 2005).
Cited by
21 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献