1. Devlin, Jacob, Chang, Ming-Wei, Lee, Kenton, Toutanova, Kristina. “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.” North American Chapter of the Association for Computational Linguistics. 2019.
2. Y. Liu , ‘RoBERTa: A Robustly Optimized BERT Pretraining Approach’, ArXiv, vol. abs/1907.11692, 2019.
3. Radford, Alec and Narasimhan, Karthik. “Improving Language Understanding by Generative Pre-Training.” (2018).
4. Walker, Christopher, Strassel, Stephanie, Medero, Julie, Maeda, Kazuaki. “ACE 2005 Multilingual Training Corpus.” Linguistic Data Consortium, 2005.
5. Li, Sha, Heng, Ji, Han, Jiawei. "Document-Level Event Argument Extraction by Conditional Generation." Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, 2021.