Affiliation:
1. School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
2. Department of Automation, Shanghai Jiao Tong University, Shanghai 200240, China
Abstract
Recent cross-lingual summarization research has pursued the use of a unified end-to-end model which has demonstrated a certain level of improvement in performance and effectiveness, but this approach stitches together multiple tasks and makes the computation more complex. Less work has focused on alignment relationships across languages, which has led to persistent problems of summary misordering and loss of key information. For this reason, we first simplify the multitasking by converting the translation task into an equal proportion of cross-lingual summary tasks so that the model can perform only cross-lingual summary tasks when generating cross-lingual summaries. In addition, we splice monolingual and cross-lingual summary sequences as an input so that the model can fully learn the core content of the corpus. Then, we propose a reinforced regularization method based on the model to improve its robustness, and build a targeted ABO mechanism to enhance the semantic relationship alignment and key information retention of the cross-lingual summaries. Ablation experiments are conducted on three datasets of different orders of magnitude to demonstrate the effective enhancement of the model by the optimization approach; they outperform the mainstream approaches on the cross-lingual summarization task and the monolingual summarization task for the full dataset. Finally, we validate the model’s capabilities on a cross-lingual summary dataset of professional domains, and the results demonstrate its superior performance and ability to improve cross-lingual sequencing.
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference35 articles.
1. A survey on cross-lingual summarization;Wang;Trans. Assoc. Comput. Linguist.,2022
2. Mohammadzadeh, A., Sabzalian, M.H., Zhang, C., Castillo, O., Sakthivel, R., and El-Sousy, F.F. (2022). Modern Adaptive Fuzzy Control Systems, Springer Nature.
3. Wan, X. (2011, January 19–24). Using bilingual information for cross-language document summarization. Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Portland, OR, USA.
4. Abstractive cross-language summarization via translation model enhanced predicate argument structure fusing;Zhang;IEEE/ACM Trans. Audio Speech Lang. Process.,2016
5. Pires, T., Schlinger, E., and Garrette, D. (2019). How multilingual is multilingual BERT?. arXiv.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献