Author:
Li Zhengpeng,Wu Jiansheng,Miao Jiawei,Yu Xinmiao
Abstract
AbstractMost of the news headline generation models that use the sequence-to-sequence model or recurrent network have two shortcomings: the lack of parallel ability of the model and easily repeated generation of words. It is difficult to select the important words in news and reproduce these expressions, resulting in the headline that inaccurately summarizes the news. In this work, we propose a TD-NHG model, which stands for news headline generation based on an improved decoder from the transformer. The TD-NHG uses masked multi-head self-attention to learn the feature information of different representation subspaces of news texts and uses decoding selection strategy of top-k, top-p, and punishment mechanisms (repetition-penalty) in the decoding stage. We conducted a comparative experiment on the LCSTS dataset and CSTS dataset. Rouge-1, Rouge-2, and Rouge-L on the LCSTS dataset and CSTS dataset are 31.28/38.73, 12.68/24.97, and 28.31/37.47, respectively. The experimental results demonstrate that the proposed method can improve the accuracy and diversity of news headlines.
Funder
Science and Technology Innovation Project of University of Science and Technology Liaoning
National Natural Science Foundation of China
Publisher
Springer Science and Business Media LLC
Reference30 articles.
1. Murao, K. et al. A case study on neural headline generation for editing support. In 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL HLT 2019, June 2, 2019–June 7, 2019. 73–82 (Association for Computational Linguistics (ACL)).
2. Song, Y.-Z. et al. Attractive or faithful? Popularity-reinforced learning for inspired headline generation. In 34th AAAI Conference on Artificial Intelligence, AAAI 2020, February 7, 2020–February 12, 2020. 8910–8917 (AAAI press).
3. Gu, X. et al. Generating representative headlines for news stories. In 29th International World Wide Web Conference, WWW 2020, April 20, 2020–April 24, 2020. 1773–1784 (Association for Computing Machinery).
4. Gavrilov, D., Kalaidin, P. & Malykh, V. Self-attentive model for headline generation. In 41st European Conference on Information Retrieval, ECIR 2019, April 14, 2019–April 18, 2019. 87–93 (Springer Verlag).
5. Xu, P., Wu, C.-S., Madotto, A. & Fung, P. Clickbait? Sensational headline generation with auto-tuned reinforcement learning. In 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, November 3, 2019–November 7, 2019. 3065–3075 (Association for Computational Linguistics).
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献