1. R. Bommasani et al., “On the opportunities and risks of foundation models,”arXiv preprint arXiv:2108.07258, 2021.
2. J. Devlin, M.W. Chang, K. Lee, K. Toutanova, “Bert: Pre-training of deep bidirectional transformers for language understanding,” arXiv preprint arXiv:1810.04805, 2018.
3. H. Touvron et al., “Llama: Open and efficient foundation language models,” arXiv preprint arXiv:2302.13971, 2023.
4. Training language models to follow instructions with human feedback;Ouyang;Adv. Neural Inf. Proces. Syst.,2022
5. Language models are few-shot learners;Brown;Adv. Neural Inf. Proces. Syst.,2020