1. A. Vaswani , N. Shazeer , N. Parmar , J. Uszkoreit , L. Jones , A. N. Gomez , Ł. Kaiser , I. Polosukhin , Attention is all you need. Advances in neural information processing systems 30, (2017).
2. Language models are unsupervised multitask learners;OpenAI blog,2019
3. Bert: Pre-training of deep bidirectional transformers for language understanding;arXiv preprint,2018
4. A text abstraction summary model based on BERT word embedding and reinforcement learning;Applied Sciences,2019
5. in Computational Linguistics: 16th International Conference of the Pacific Association for Computational Linguistics, PACLING 2019, Hanoi, Vietnam, October 11–13, 2019;Revised Selected Papers,2020