1. Language models are few-shot learners;brown;Proc Adv Neural Inf Process Syst,2020
2. Attention is all you need;vaswani;Proc 30th Adv Neural Inf Process Syst,0
3. BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension;lewis,2019
4. Zero-shot text-to-image generation;ramesh;Proc Int Conf Mach Learn,0
5. Auto-encoding variational bayes;kingma,2013