1. BERT: Pre-training of deep bidirectional transformers for language understanding;devlin;Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics Human Language Technologies,2019
2. Glove: Global Vectors for Word Representation
3. Adam: A method for stochastic optimization;kingma;Proceedings of International Conference on Learning Representations,2015
4. Multitask learning for poly-phonic piano transcription, a case study;kelz;Proceedings of International Workshop on Multilayer Music Representation and Processing,2019
5. Automated evaluation of non-native English pronunciation quality: combining knowledge- and data-driven features at multiple time scales