Abstract
With the recent advances in deep learning, different approaches to improving pre-trained language models (PLMs) have been proposed. PLMs have advanced state-of-the-art (SOTA) performance on various natural language processing (NLP) tasks such as machine translation, text classification, question answering, text summarization, information retrieval, recommendation systems, named entity recognition, etc. In this paper, we provide a comprehensive review of prior embedding models as well as current breakthroughs in the field of PLMs. Then, we analyse and contrast the various models and provide an analysis of the way they have been built (number of parameters, compression techniques, etc.). Finally, we discuss the major issues and future directions for each of the main points.
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference90 articles.
1. Distributional Structure
2. A Neural Probabilistic Language Model;Bengio;J. Mach. Learn. Res.,2003
3. Distributed representations of words and phrases and their compositionality;Mikolov;Proceedings of the 26th International Conference on Neural Information Processing Systems,2013
4. Glove: Global vectors for word representation;Pennington;Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP2014),2014
5. Neural Word Embedding as Implicit Matrix Factorization;Levy,2014
Cited by
39 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献