A Mongolian-Chinese Neural Machine Translation Model Based on Soft Target Templates and Contextual Knowledge
-
Published:2023-10-30
Issue:21
Volume:13
Page:11845
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Ren Qing-Dao-Er-Ji1ORCID,
Pang Ziyu1ORCID,
Lang Jiajun1
Affiliation:
1. School of Information Engineering, Inner Mongolia University of Technology, Hohhot 010051, China
Abstract
In recent years, Mongolian-Chinese neural machine translation (MCNMT) technology has made substantial progress. However, the establishment of the Mongolian dataset requires a significant amount of financial and material investment, which has become a major obstacle to the performance of MCNMT. Pre-training and fine-tuning technology have also achieved great success in the field of natural language processing, but how to fully exploit the potential of pre-training language models (PLMs) in MCNMT has become an urgent problem to be solved. Therefore, this paper proposes a novel MCNMT model based on the soft target template and contextual knowledge. Firstly, to learn the grammatical structure of target sentences, a selection-based parsing tree is adopted to generate candidate templates that are used as soft target templates. The template information is merged with the encoder-decoder framework, fully utilizing the templates and source text information to guide the translation process. Secondly, the translation model learns the contextual knowledge of sentences from the BERT pre-training model through the dynamic fusion mechanism and knowledge extraction paradigm, so as to improve the model’s utilization rate of language knowledge. Finally, the translation performance of the proposed model is further improved by integrating contextual knowledge and soft target templates by using a scaling factor. The effectiveness of the modified model is verified by a large number of data experiments, and the calculated BLEU (BiLingual Evaluation Understudy) value is increased by 4.032 points compared with the baseline MCNMT model of Transformers.
Funder
National Natural Science Foundation of China
Inner Mongolia Natural Science Foundation
Inner Mongolia Science and Technology Program Project
Young Scientific and Technological Talents in Inner Mongolia Colleges and Universities
Fundamental Research Fund Project
Inner Mongolia Autonomous Region
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference16 articles.
1. Sutskever, I., Vinyals, O., and Le, Q.V. (2014). Sequence to sequence learning with neural networks. arXiv.
2. A neural probabilistic language model;Kandola;Stud. Fuzziness Soft Comput.,2006
3. A Generalized Constraint Approach to Bilingual Dictionary Induction for Low-Resource Language Families;Nasution;ACM Trans. Asian Low-Resour. Lang. Inf. Process. (TALLIP),2017
4. Kitaev, N., Kaiser, Ł., and Levskaya, A. (2020). Reformer: The efficient transformer. arXiv.
5. Wang, S., Li, P., Tan, Z., Tu, Z., Sun, M., and Liu, Y. (2022). A template-based method for constrained neural machine translation. arXiv.