1. AiHub (2021). Aihub broadcast content korean speech recognition
data. Retrieved from https://aihub.or.kr/aihubdata/data/view.do?currMenu=115&topMenu=100&aihubDataSe=realm&dataSetSn=463
2. Baevski, A., Zhou, Y., Mohamed, A., & Auli, M. (2020, December).
wav2vec 2.0: A framework for self-supervised learning of speech representations.
Proceedings of the Advances in Neural Information Processing
Systems (pp. 12449-12460). Online Conference.
3. Bang, J. U., Yun, S., Kim, S. H., Choi, M. Y., Lee, M. K., Kim, Y.
J., Kim, D. H., ... Kim, S. H. (2020). KsponSpeech: Korean spontaneous speech
corpus for automatic speech recognition. Applied Sciences,
10(19), 6936. 10.3390/app10196936
4. Chan, W., Jaitly, N., Le, Q., & Vinyals, O. (2016, March).
Listen, attend and spell: A neural network for large vocabulary conversational
speech recognition. Proceedings of the 2016 IEEE International
Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp.
4960-4964). Shanghai, China. 10.1109/ICASSP.2016.7472621
5. Chen, T., Kornblith, S., Norouzi, M., & Hinton, G. (2020, July).
A simple framework for contrastive learning of visual representations.
Proceedings of the 37th International Conference on Machine
Learning (pp. 1597-1607). Online Conference.