Affiliation:
1. University of Maryland, Baltimore County Department of Information Systems kkeya1@umbc.edu
2. Healx Department of Research and Development yannis.papanikolaou@healx.io
3. University of Maryland, Baltimore County Department of Information Systems jfoulds@umbc.edu
Abstract
Abstract
We propose a method that uses neural embeddings to improve the performance of any given LDA-style topic model. Our method, called neural embedding allocation (NEA), deconstructs topic models (LDA or otherwise) into interpretable vector-space embeddings of words, topics, documents, authors, and so on, by learning neural embeddings to mimic the topic model. We demonstrate that NEA improves coherence scores of the original topic model by smoothing out the noisy topics when the number of topics is large. Furthermore, we show NEA’s effectiveness and generality in deconstructing and smoothing LDA, author-topic models, and the recent mixed membership skip-gram topic model and achieve better performance with the embeddings compared to several state-of-the-art models.
Subject
Artificial Intelligence,Computer Science Applications,Linguistics and Language,Language and Linguistics
Reference64 articles.
1. Big data’s disparate impact;Barocas;California Law Review,2016
2. On the dangers of stochastic parrots: Can language models be too big?;Bender,2021
3. A neural probabilistic language model;Bengio;Journal of Machine Learning Research,2003
4. Latent Dirichlet allocation;Blei;Journal of Machine Learning Research,2003
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献