A Survey on Evaluation Metrics for Machine Translation
-
Published:2023-02-16
Issue:4
Volume:11
Page:1006
-
ISSN:2227-7390
-
Container-title:Mathematics
-
language:en
-
Short-container-title:Mathematics
Author:
Lee Seungjun1, Lee Jungseob1, Moon Hyeonseok1ORCID, Park Chanjun12ORCID, Seo Jaehyung1, Eo Sugyeong1, Koo Seonmin1, Lim Heuiseok1
Affiliation:
1. Department of Computer Science and Engineering, Korea University, Seoul 02841, Republic of Korea 2. Upstage, Yongin 16942, Republic of Korea
Abstract
The success of Transformer architecture has seen increased interest in machine translation (MT). The translation quality of neural network-based MT transcends that of translations derived using statistical methods. This growth in MT research has entailed the development of accurate automatic evaluation metrics that allow us to track the performance of MT. However, automatically evaluating and comparing MT systems is a challenging task. Several studies have shown that traditional metrics (e.g., BLEU, TER) show poor performance in capturing semantic similarity between MT outputs and human reference translations. To date, to improve performance, various evaluation metrics have been proposed using the Transformer architecture. However, a systematic and comprehensive literature review on these metrics is still missing. Therefore, it is necessary to survey the existing automatic evaluation metrics of MT to enable both established and new researchers to quickly understand the trend of MT evaluation over the past few years. In this survey, we present the trend of automatic evaluation metrics. To better understand the developments in the field, we provide the taxonomy of the automatic evaluation metrics. Then, we explain the key contributions and shortcomings of the metrics. In addition, we select the representative metrics from the taxonomy, and conduct experiments to analyze related problems. Finally, we discuss the limitation of the current automatic metric studies through the experimentation and our suggestions for further research to improve the automatic evaluation metrics.
Funder
Ministry of Science and ICT, Korea National Research Foundation of Korea
Subject
General Mathematics,Engineering (miscellaneous),Computer Science (miscellaneous)
Reference55 articles.
1. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł., and Polosukhin, I. (2017, January 4–9). Attention is all you need. Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA. 2. Lavie, A. (2011, January 19–23). Evaluating the Output of Machine Translation Systems. Proceedings of the Machine Translation Summit XIII: Tutorial Abstracts, Xiamen, China. 3. White, J.S., and O’Connell, T.A. (1993, January 21–24). Evaluation of machine translation. Proceedings of the Human Language Technology: Proceedings of a Workshop, Plainsboro, NJ, USA. 4. Papineni, K., Roukos, S., Ward, T., and Zhu, W.J. (2002, January 7–12). Bleu: A method for automatic evaluation of machine translation. Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, Philadelphia, PA, USA. 5. Doddington, G. (2002, January 24–27). Automatic evaluation of machine translation quality using n-gram co-occurrence statistics. Proceedings of the Second International Conference on Human Language Technology Research, San Diego, CA, USA.
Cited by
18 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|