1. Language models are few-shot learners;Brown;NeurIPS,2020
2. PaLM: Scaling language modeling with pathways;Chowdhery,2022
3. Megatron-LM: Training multi-billion parameter language models using model parallelism;Shoeybi,2020
4. wav2vec 2.0: A framework for self-supervised learning of speech representations;Baevski;NeurIPS,2020
5. HuBERT: Self-Supervised Speech Representation Learning by Masked Prediction of Hidden Units