1. Li J, Tang T, Zhao WX, Wen JR. Pretrained language models for text generation: a survey. Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21). 2021. p. 4492–9.
2. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. Adv Neural Inf Process Syst. 2017. p. 1–11.
3. Achiam J, Adler S, Agarwal S, Ahmad L, Akkaya I, Aleman FL, Almeida D, Altenschmidt J, Altman S, Anadkat S, Avila R. Gpt-4 technical report. arXiv preprint; 2023 Mar 15. arXiv:2303.08774.
4. Sharoff S. Know thy corpus! robust methods for digital curation of web corpora. arXiv preprint; 2020 Mar 13. arXiv:2003.06389.
5. Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, et al. Language models are few-shot learners. In: Adv Neural Inf Process Syst. 2020. p. 1–25.