1. Attention is all you need;Vaswani,2017
2. BERT: Pre-training of deep bidirectional transformers for language understanding;Devlin,2019
3. End-to-end open-domain question answering with BERTserini;Yang,2019
4. How to fine-tune BERT for text classification?;Sun,2019
5. Text summarization with pretrained encoders;Liu,2019