1. Brown T, Mann B, Ryder N, Language models are few-shot learners[J]. Advances in neural information processing systems, 2020, 33: 1877-1901.
2. Chowdhery A, Narang S, Devlin J, Palm: Scaling language modeling with pathways[J]. arXiv preprint arXiv:2204.02311, 2022.
3. Zhang S, Roller S, Goyal N, Opt: Open pre-trained transformer language models[J]. arXiv preprint arXiv:2205.01068, 2022.
4. Zeng A, Liu X, Du Z, Glm-130b: An open bilingual pre-trained model[J]. arXiv preprint arXiv:2210.02414, 2022.
5. Touvron H, Lavril T, Izacard G, Llama: Open and efficient foundation language models[J]. arXiv preprint arXiv:2302.13971, 2023.