Abstract
Large text documents are sometimes challenging to understand and time-consuming to extract vital information from. These issues are addressed by automatic text summarizing techniques, which condense lengthy texts while preserving their key information. Thus, the development of automatic summarization systems capable of fulfilling the ever-increasing demands of textual data becomes of utmost importance. It is even more vital with complex natural languages. This study explores five State-Of-The-Art (SOTA) Arabic deep Transformer-based Language Models (TLMs) in the task of text summarization by adapting various text summarization datasets dedicated to Arabic. A comparison against deep learning and machine learning-based baseline models has also been conducted. Experimental results reveal the superiority of TLMs, specifically the PEAGASUS family, against the baseline approaches, with an average F1-score of 90% on several benchmark datasets.
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference54 articles.
1. Saggion, H., and Poibeau, T. (2013). Multi-Source, Multilingual Information Extraction and Summarization, Springer.
2. Rauniyar, S. (2020, January 26–28). A survey on deep learning based various methods analysis of text summarization. Proceedings of the 2020 International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India.
3. Fejer, H.N., and Omar, N. (2014, January 18–20). Automatic Arabic text summarization using clustering and keyphrase extraction. Proceedings of the 6th International Conference on Information Technology and Multimedia, Putrajaya, Malaysia.
4. A survey of the state-of-the-art models in neural abstractive text summarization;IEEE Access,2021
5. Siragusa, G., and Robaldo, L. (2022). Sentence Graph Attention For Content-Aware Summarization. Appl. Sci., 12.
Cited by
11 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献