Affiliation:
1. Thu Dau Mot University, Binh Duong, Vietnam
Abstract
Recently, advanced techniques in deep learning such as recurrent neural network (GRU, LSTM and Bi-LSTM) and auto-encoding (attention-based transformer and BERT) have achieved great successes in multiple application domains including text summarization. Recent state-of-the-art encoding-based text summarization models such as BertSum, PreSum and DiscoBert have demonstrated significant improvements on extractive text summarization tasks. However, recent models still encounter common problems related to the language-specific dependency which requires the supports of the external NLP tools. Besides that, recent advanced text representation methods, such as BERT as the sentence-level textual encoder, also fail to fully capture the representation of a full-length document. To address these challenges, in this paper we proposed a novel
s
emantic-ware
e
mbedding approach for
ex
tractive text
sum
marization
, called as:
SE4ExSum.
Our proposed SE4ExSum is an integration between the use of
feature graph-of-words (FGOW)
with BERT-based encoder for effectively learning the word/sentence-level representations of a given document. Then, the
g
raph
c
onvolutional
n
etwork (GCN)
based encoder is applied to learn the global document's representation which is then used to facilitate the text summarization task. Extensive experiments on benchmark datasets show the effectiveness of our proposed model in comparing with recent state-of-the-art text summarization models.
Publisher
Association for Computing Machinery (ACM)
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献