1. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding;Devlin,2019
2. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer;Raffel;J. Mach. Learn. Res.,2020
3. GPT-4 Technical Report,2023
4. LLaMA: Open and Efficient Foundation Language Models;Touvron,2023
5. Language Models are Few-Shot Learners;Brown,2020