1. BERT: pre-training of deep bidirectional transformers for language understanding;Devlin,2019
2. Improving language understanding by generative pre-training;Radford,2018
3. Language models are few-shot learners;Brown,2020
4. Gpt-4 technical report;OpenAI,2023