Author:
Brown Chase Alan,Wren Jonathan D.
Abstract
AbstractThis paper provides a novel approach to understanding the nature of innovation and scientific progress by analyzing large-scale datasets of scientific literature. A new measure of novelty potential or disruptiveness for a set of scientific entities is proposed, based in the mathematical formalism of algebraic topology via a method calledpersistent homology. In this framework, understanding where academic ideas depart from the existing body of knowledge tofill knowledge gapsis key to scoring a set of entities and their potential for filling future knowledge gaps. This framework is motivated by the assumption that scientific discovery has underlying regularities that can be modeled and predicted.Our method uses afiltration, which is a type of ranking of hypergraph components along a chosen parameter. In this work two different axes are used, which constructs a growing grid of sub-hypergraphs. The axes of time (scientific knowledge evolution) and normalized point-wise mutual information (network structure) affords the ability to succinctly represent the entire dynamic structure of the scientific literature network. We then find that using very simple and interpretable measures of centrality derived from this crudebifiltrationorvineyardaffords the ability to predict links within the dynamic scientific network.While several different methods of link prediction have been proposed in the past, the method presented hereextendsthe notion of link prediction to a higher dimension, as the boundary of the knowledge gap may be more than just 0-dimensional nodes.The system presented here not only suggests a mathematical basis, consistent with observations in cognitive neurosciences regarding early childhood language acquisition, but additionally provides useful applications for the scientific community in predicting and ranking hypothesis for scientific discovery.
Publisher
Cold Spring Harbor Laboratory
Reference21 articles.
1. Grokking: Generalization beyond overfitting on small algorithmic datasets;arXiv preprint,2022
2. Liu, Z. , Michaud, E. J. & Tegmark, M. Omnigrok: Grokking beyond algorithmic data. in The Eleventh International Conference on Learning Representations (2022).
3. Towards understanding grokking: An effective theory of representation learning;Advances in Neural Information Processing Systems,2022
4. Knowledge gaps in the early growth of semantic feature networks;Nature human behaviour,2018
5. LION LBD: a literature-based discovery system for cancer biology;Bioinformatics,2019