1. Unified language model pre-training for natural language understanding and generation;dong;Advances in neural information processing systems,2019
2. Mass: Masked sequence to sequence pre-training for language generation;song;ArXiv Preprint,2019
3. Dialogue history matters! personalized response selectionin multi-turn retrieval-based chatbots;li;ArXiv Preprint,2021
4. Leveraging Pre-trained Checkpoints for Sequence Generation Tasks
5. Recurrent neural network regularization;zaremba;ArXiv Preprint,2014