1. Language Models are Few-Shot Learners;Larochelle;Proceedings of the Advances in Neural Information Processing Systems,2020
2. Exploring the limits of transfer learning with a unified text-to-text transformer;Raffel;J. Mach. Learn. Res.,2020
3. PaLM: Scaling Language Modeling with Pathways;Chowdhery;J. Mach. Learn. Res.,2023
4. Walker, M., Ji, H., and Stent, A. (2018, January 1–6). Newsroom: A Dataset of 1.3 Million Summaries with Diverse Extractive Strategies. Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, LA, USA. (Long Papers).
5. Hermann, K.M., Kočiský, T., Grefenstette, E., Espeholt, L., Kay, W., Suleyman, M., and Blunsom, P. (2015, January 7–12). Teaching Machines to Read and Comprehend. Proceedings of the 28th International Conference on Neural Information Processing Systems—Volume 1, Montreal, QC, Canada.