1. Huang, C.C. and Lu, Z., Community challenges in biomedical text mining over 10 years: Success, failure and the future, Briefings Bioinf., 2016, vol. 17, no. 1, pp. 132–144.
2. Vaswani, A., Shazeer, N., et al., Attention is all you need, Proc. 31st Int. Conf. Neural Information Processing Systems, 2017, pp. 6000–6010.
3. Devlin, J., Chang, M., et al., BERT: Pre-training of deep bidirectional transformers for language understanding, Proc. Conf. North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019, vol. 1, pp. 4171–4186.
4. Conneau, A. and Lample, G., Cross-lingual language model pretraining, Adv. Neural Inf. Process. Syst., 2019, vol. 32, pp. 7059–7069.
5. Lample, G., Conneau, A., et al., Unsupervised machine translation using monolingual corpora only, Proc. Int. Conf. Learning Representations, 2018.