Affiliation:
1. School of Computer Science 8 Technology, Soochow University, China
Abstract
In this article, we show that word translations can be explicitly incorporated into NMT effectively to avoid wrong translations. Specifically, we propose three cross-lingual encoders to explicitly incorporate word translations into NMT: (1)
Factored
encoder, which encodes a word and its translation in a vertical way; (2)
Gated
encoder, which uses a gated mechanism to selectively control the amount of word translations moving forward; and (3)
Mixed
encoder, which stitchingly learns a word and its translation annotations over sequences where words and their translations are alternatively mixed. Besides, we first use a simple word dictionary approach and then a word sense disambiguation (WSD) approach to effectively model the word context for better word translation. Experimentation on Chinese-to-English translation demonstrates that all proposed encoders are able to improve the translation accuracy for both traditional RNN-based NMT and recent self-attention-based NMT (hereafter referred to as
Transformer
). Specifically,
Mixed
encoder yields the most significant improvement of 2.0 in BLEU on the RNN-based NMT, while
Gated
encoder improves 1.2 in BLEU on
Transformer
. This indicates the usefulness of an WSD approach in modeling word context for better word translation. This also indicates the effectiveness of our proposed cross-lingual encoders in explicitly modeling word translations to avoid wrong translations in NMT. Finally, we discuss in depth how word translations benefit different NMT frameworks from several perspectives.
Funder
National Natural Science Foundation of China
Publisher
Association for Computing Machinery (ACM)
Cited by
6 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献