Machine Translation of English Language Using the Complexity-Reduced Transformer Model

Author:

Li Qin1ORCID

Affiliation:

1. Hubei University of Education, Wuhan, China

Abstract

Previous translation models like statistical machine translation (SMT), rule-based machine translation (RBMT), hybrid machine translation (HMT), and neural machine translation (NMT) have reached their performance bottleneck. The new Transformer-based machine translation model has become the favorite choice for English language translation. For instance, Google’s BERT translation model organizes the Transformer module into bidirectional encoder representations. It is aware of the users’ search intentions as well as the material that the search engine has indexed. It does not need to evaluate previous searches to comprehend what people mean, unlike RankBrain. BERT comprehends words, sentences, and complete information in the same way that we do. It achieves remarkable translation quality improvement over the other state-of-the-art benchmarks. It demonstrates the great potential of the Transformer model. The Transformer-based translation model mainly improves the performance at the cost of growing model sizes and complexity, usually requiring million-scale parameters. It is hard for the traditional computing systems to cope with the growing memory and computation requirements. However, the latest computers can easily run this model without any lag. The biggest challenge of applying the Transformer model is to deploy these models efficiently onto real-time or embedded devices. In this work, we propose a quantization scheme to reduce the parameter and computation complexity. It is of great importance to promote the usage of the Transformer model. Our experiment results show that the original Transformer model in 32 bit floating-point can be quantized to only 8 bits to 12 bits with only negligible translation quality loss. However, due to the perfect transformation of the block part, this quality loss part can easily be managed by the users. Meanwhile, our algorithm achieves 2.6 × to 4.0 × compression ratio, which is helpful to save the required complexity and energy during the inference phase.

Funder

Hubei University of Education

Publisher

Hindawi Limited

Subject

Computer Networks and Communications,Computer Science Applications

Reference15 articles.

1. Long Short-Term Memory

2. Attention is all you need;A. Vaswani;Advances in Neural Information Processing Systems,2017

3. October. Recurrent continuous translation models;N. Kalchbrenner

4. Stanford neural machine translation systems for spoken language domains;M. T. Luong

5. Sequence to sequence learning with neural networks;I. Sutskever;Advances in Neural Information Processing Systems,2014

Cited by 3 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Translation English to Punjabi: A Concise Review of Significant Approaches;2024 International Conference on Computational Intelligence and Computing Applications (ICCICA);2024-05-23

2. Translation Systems: A Synoptic Survey of Deep Learning Approaches and Techniques;2024 11th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO);2024-03-14

3. Enhancing Neural Text Detector Robustness with μAttacking and RR-Training;Electronics;2023-04-21

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3