1. Blevins, T., Gonen, H., and Zettlemoyer, L. (2022). “Analyzing the Mono- and Cross-Lingual Pretraining Dynamics of Multilingual Language Models.” In Proceedings of EMNLP, pp. 3575–3590, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
2. Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., McCandlish, S., Radford, A., Sutskever, I., and Amodei, D. (2020). “Language Models are Few-Shot Learners.” In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (Eds.), Advances in Neural Information Processing Systems, Vol. 33, pp. 1877–1901. Curran Associates, Inc.
3. Bybee, J. L. (2013). “49 Usage-based Theory and Exemplar Representations of Constructions.” In The Oxford Handbook of Construction Grammar. Oxford University Press.
4. Chiswick, B. and Miller, P. (2004). “Linguistic Distance: A Quantitative Measure of the Distance Between English and Other Languages.” Tech. rep. 1246, Institute of Labor Economics (IZA).
5. Conneau, A., Khandelwal, K., Goyal, N., Chaudhary, V., Wenzek, G., Guzmán, F., Grave, E., Ott, M., Zettlemoyer, L., and Stoyanov, V. (2020). “Unsupervised Cross-lingual Representation Learning at Scale.” In Proceedings of ACL, pp. 8440–8451, Online. Association for Computational Linguistics.