Enhancing Abstractive Summarization with Extracted Knowledge Graphs and Multi-Source Transformers
-
Published:2023-06-30
Issue:13
Volume:13
Page:7753
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Chen Tong1ORCID, Wang Xuewei1ORCID, Yue Tianwei1ORCID, Bai Xiaoyu2ORCID, Le Cindy X.3, Wang Wenping1ORCID
Affiliation:
1. Carnegie Mellon University, Pittsburgh, PA 15213, USA 2. Rice University, Houston, TX 77005, USA 3. Columbia University, New York, NY 10027, USA
Abstract
As the popularity of large language models (LLMs) has risen over the course of the last year, led by GPT-3/4 and especially its productization as ChatGPT, we have witnessed the extensive application of LLMs to text summarization. However, LLMs do not intrinsically have the power to verify the correctness of the information they supply and generate. This research introduces a novel approach to abstractive summarization, aiming to address the limitations of LLMs in that they struggle to understand the truth. The proposed method leverages extracted knowledge graph information and structured semantics as a guide for summarization. Building upon BART, one of the state-of-the-art sequence-to-sequence pre-trained LLMs, multi-source transformer modules are developed as an encoder, which are capable of processing textual and graphical inputs. Decoding is performed based on this enriched encoding to enhance the summary quality. The Wiki-Sum dataset, derived from Wikipedia text dumps, is introduced for evaluation purposes. Comparative experiments with baseline models demonstrate the strengths of the proposed approach in generating informative and relevant summaries. We conclude by presenting our insights into utilizing LLMs with graph external information, which will become a powerful aid towards the goal of factually correct and verified LLMs.
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference30 articles.
1. See, A., Liu, P.J., and Manning, C.D. (2017). Get to the point: Summarization with pointer-generator networks. arXiv. 2. Nallapati, R., Zhou, B., dos Santos, C., Gulçehre, Ç., and Xiang, B. (2016, January 7–12). Abstractive text summarization using sequence-to-sequence RNNs and beyond. Proceedings of the Conference on Computational Natural Language Learning, Berlin, Germany. 3. Kenton, J.D.M.W.C., and Toutanova, L.K. (2019, January 2–7). BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Proceedings of the NAACL-HLT, Minneapolis, MN, USA. 4. Lewis, M., Liu, Y., Goyal, N., Ghazvininejad, M., Mohamed, A., Levy, O., Stoyanov, V., and Zettlemoyer, L. (2019). Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv. 5. OpenAI (2023, June 24). GPT-4 Technical Report, Available online: http://xxx.lanl.gov/abs/2303.08774.
Cited by
9 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|