1. J.D.M.-W.C. Kenton, L.K. Toutanova, BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, in: Proceedings of NAACL-HLT, 2019, pp. 4171–4186.
2. Roberta: A robustly optimized bert pretraining approach;Liu,2019
3. Improving language understanding by generative pre-training;Radford,2018
4. B. Li, H. Zhou, J. He, M. Wang, Y. Yang, L. Li, On the Sentence Embeddings from Pre-trained Language Models, in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP, 2020, pp. 9119–9130.
5. W. Zhang, J. Su, C.L. Tan, W.T. Wang, Entity linking leveraging automatically generated annotation, in: Proceedings of the 23rd International Conference on Computational Linguistics (Coling 2010), 2010, pp. 1290–1298.