Near-Optimal Active Learning for Multilingual Grapheme-to-Phoneme Conversion
-
Published:2023-08-19
Issue:16
Volume:13
Page:9408
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Cao Dezhi12, Zhao Yue23ORCID, Wu Licheng3ORCID
Affiliation:
1. School of Chinese Ethnic Languages and Literature, Minzu University of China, Beijing 100081, China 2. Key Laboratory of Ethnic Language Intelligent Analysis and Security Governance of MOE, Minzu University of China, Beijing 100081, China 3. School of Information Engineering, Minzu University of China, Beijing 100081, China
Abstract
The construction of pronunciation dictionaries relies on high-quality and extensive training data in data-driven way. However, the manual annotation of corpus for this purpose is both costly and time consuming, especially for low-resource languages that lack sufficient data and resources. A multilingual pronunciation dictionary includes some common phonemes or phonetic units, which means that these phonemes or units have similarities in the pronunciation of different languages and can be used in the construction process of pronunciation dictionaries for low-resource languages. By using a multilingual pronunciation dictionary, knowledge can be shared among different languages, thus improving the quality and accuracy of pronunciation dictionaries for low-resource languages. In this paper, we propose using shared articulatory features among multiple languages to construct a universal phoneme set, which is then used to label words for multiple languages. To achieve this, we first developed a grapheme−phoneme (G2P) model based on an encoder−decoder deep neural network. We then adopted a near-optimal active learning method in the process of building the pronunciation dictionary to select informative samples from a large, unlabeled corpus and had them labeled by experts. Our experiments demonstrate that this method selected about 1/5 of the unlabeled data and achieved an even higher conversion accuracy than the results of the large data training method. By selectively labeling samples with a high uncertainty in the model, while avoiding labeling samples that were accurately predicted by the current model, our method greatly enhances the efficiency of pronunciation dictionary construction.
Funder
National Natural Science Foundation of China Minzu University of China
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference36 articles.
1. Taylor, P. (September, January 30). Hidden markov models for grapheme to phoneme conversion. Proceedings of the Annual Conference of the International Speech Communication Association (INTERSPEECH), Brno, Czech Republic. 2. Empirically derived probabilities for grapheme-to-phoneme correspondences in english;Berndt;Behav. Res. Methods Instrum. Comput.,1987 3. Mortensen, D.R., Dalmia, S., and Littell, P. (2018, January 7–12). Epitran: Precision g2p for many languages. Proceedings of the 11th International Conference on Language Resources and Evaluation (LREC), Miyazaki, Japan. 4. Deri, A., and Knight, K. (2016, January 7–12). Grapheme-to-phoneme models for (almost) any language. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (ACL), Berlin, Germany. 5. Sokolov, A., Rohlin, T., and Rastrow, A. (2019, January 15–19). Neural machine translation for multilingual grapheme-to-phoneme conversion. Proceedings of the Annual Conference of the International Speech Communication Association (INTERSPEECH), Graz, Austria.
|
|