1. Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171–4186, June 2019
2. Feng, Z., et al.: CodeBERT: A pre-trained model for programming and natural languages. In: Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 1536–1547 (2020)
3. Lecture Notes in Computer Science;E Kacupaj,2020
4. Kapanipathi, P., et al.: Question answering over knowledge bases by leveraging semantic parsing and neuro-symbolic reasoning (2020)
5. Liu, Y., et al.: RoBERTa: a robustly optimized BERT pretraining approach (2020)