1. Adiwardana, D., Luong, M.-T., So, D. R., Hall, J., Fiedel, N., Thoppilan, R., Yang, Z., Kulshreshtha, A., Nemade, G., Lu, Y., and Le, Q. V. (2020). “Towards a Human-like Open-Domain Chatbot.” In arXiv preprint arXiv:2001.09977.
2. Akama, R., Watanabe, K., Yokoi, S., Kobayashi, S., and Inui, K. (2018). “Unsupervised Learning of Style-sensitive Word Vectors.” In Proceedings of ACL, pp. 572–578.
3. Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., McCandlish, S., Radford, A., Sutskever, I., and Amodei, D. (2020). “Language Models are Few-Shot Learners.” In Proceedings of NeurIPS, pp. 1877–1901.
4. Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. (2019). “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.” In Proceedings of NAACL, pp. 4171–4186.
5. Dinan, E., Roller, S., Shuster, K., Fan, A., Auli, M., and Weston, J. (2019). “Wizard of Wikipedia: Knowledge-powered Conversational Agents.” In Proceedings of ICLR, pp. 1–18.