1. Brown, T., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877–1901 (2020)
2. Black, S., et al.: GPT-NeoX-20B: an open-source autoregressive language model. arXiv preprint arXiv:2204.06745 (2022)
3. Zhang, S., et al.: OPT: open pre-trained transformer Language models. arXiv preprint arXiv:2205.01068 (2022)
4. Smith, S., et al.: Using deepspeed and megatron to train megatron-turing NLG 530b, a large-scale generative language model. arXiv preprint arXiv:2201.11990 (2022)
5. OpenAI: GPT-4 Technical report. arXiv abs/2303.08774 (2023)