1. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998–6008
2. Ziegler DM, Stiennon N, Wu J, Brown TB, Radford A, Amodei D, Christiano P, Irving G (2019) Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593
3. Keskar NS, McCann B, Varshney LR, Xiong C, Socher R (2019) Ctrl: a conditional transformer language model for controllable generation. arXiv preprint arXiv:1909.05858
4. Brown TB, Mann B, Ryder N, Subbiah M Kaplan, J Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A, et al. (2020) Language models are few-shot learners. arXiv preprint arXiv:2005.14165
5. Dathathri S, Madotto A, Lan J, Hung J, Frank E, Molino P, Yosinski J, Liu R(2019) Plug and play language models: a simple approach to controlled text generation. arXiv preprint arXiv:1912.02164