1. On the opportunities and risks of foundation models;Bommasani,2022
2. BERT: pre-training of deep bidirectional transformers for language understanding;Devlin,2019
3. BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension;Lewis,2019
4. Language models are few-shot learners;Brown,2020
5. LLaMA: open and efficient foundation language models;Touvron,2023