1. Wav2vec 2.0: A framework for self-supervised learning of speech representations;baevski;Proceedings of the 34th International Conference on Neural Information Processing Systems,2020
2. Improving Hybrid CTC/Attention End-to-End Speech Recognition with Pretrained Acoustic and Language Models
3. Improved mask-ctc for non-autoregressive end-to-end asr;higuchi,2020
4. BERT: pre-training of deep bidirectional transformers for language understanding;devlin;Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics Human Language Technologies NAACL-HLT 2019 Minneapolis MN USA June 2-7 2019 Volume 1 (Long and Short Papers),2019
5. Non-Autoregressive Transformer for Speech Recognition