1. Deep contextualized word Representations;Peters,2018
2. Universal Language model fine-tuning for text classification;Howard,2018
3. A. Radford, K. Narasimhan, T. Salimans, I. Sutskever, Improving Language Understanding by Generative Pre-training, 2018, unpublished manuscript, https://s3-us-west-2.amazonaws.com/openai-assets/research-covers/language-unsupervised/language_understanding_paper.pdf.
4. BERT: pre-training of deep bidirectional Transformers for language understanding;Devlin,2019
5. Language Models Are Unsupervised Multitask Learners;Radrof,2018