Unified Training for Cross-Lingual Abstractive Summarization by Aligning Parallel Machine Translation Pairs

Author:

Cheng Shaohuan1ORCID,Chen Wenyu1,Tang Yujia1,Fu Mingsheng1,Qu Hong1ORCID

Affiliation:

1. School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China

Abstract

Cross-lingual summarization (CLS) is essential for enhancing global communication by facilitating efficient information exchange across different languages. However, owing to the scarcity of CLS data, recent studies have employed multi-task frameworks to combine parallel monolingual summaries. These methods often use independent decoders or models with non-shared parameters because of the mismatch in output languages, which limits the transfer of knowledge between CLS and its parallel data. To address this issue, we propose a unified training method for CLS that combines parallel machine translation (MT) pairs with CLS pairs, jointly training them within a single model. This design ensures consistent input and output languages and promotes knowledge sharing between the two tasks. To further enhance the model’s capability to focus on key information, we introduce two additional loss terms to align the hidden representations and probability distributions between the parallel MT and CLS pairs. Experimental results demonstrate that our method outperforms competitive methods in both full-dataset and low-resource scenarios on two benchmark datasets, Zh2EnSum and En2ZhSum.

Funder

Young Scientists Fund of the Natural Science Foundation of Sichuan Province

Publisher

MDPI AG

Reference36 articles.

1. Muresan, S., Nakov, P., and Villavicencio, A. (2022, January 22–27). A Variational Hierarchical Model for Neural Cross-Lingual Summarization. Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland. Volume 1: Long Papers.

2. Zhu, J., Wang, Q., Wang, Y., Zhou, Y., Zhang, J., Wang, S., and Zong, C. (2019, January 3–7). NCLS: Neural Cross-Lingual Summarization. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China.

3. Cross-lingual c* st* rd: English access to hindi information;Leuski;ACM Trans. Asian Lang. Inf. Process.,2003

4. Wan, X., Li, H., and Xiao, J. (2010, January 11–16). Cross-language document summarization based on machine translation quality prediction. Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, Uppsala, Sweden.

5. A two-stage fine-tuning method for low-resource cross-lingual summarization;Zhang;Math. Biosci. Eng.,2024

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3