1. Bahdanau, D., Cho, K., & Bengio, Y. (2015). Neural Machine Translation by Jointly Learning to Align and Translate. In International conference on learning representations.
2. Bhandari, M., Gour, P. N., Ashfaq, A., Liu, P., & Neubig, G. (2020). Re-evaluating Evaluation in Text Summarization. In Proceedings of the 2020 conference on empirical methods in natural language processing (pp. 9347–9359).
3. Casas, N., Fonollosa, J. A. R., & Costa-jussà, M. R. (2018). A differentiable BLEU loss. Analysis and first results. In International conference on learning representations.
4. Chopra, S., Auli, M., & Rush, A. M. (2016). Abstractive Sentence Summarization with Attentive Recurrent Neural Networks. In Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: human language technologies (pp. 93–98).
5. Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies (pp. 4171–4186).