Affiliation:
1. Department of Computer Information Management, Inner Mongolia University of Finance and Economics, Hohhot 010070, China
2. Department of Computer Science and Technology, Northeastern University, Shenyang 110167, China
Abstract
Neural models are widely applied to headline generation. Template-based methods are a promising direction to overcome the shortcomings of the neural headline generation (NHG) model in generating duplicate or extra words. Previous works often retrieve relevant headlines from the training data and adopt them as the soft template to guide the NHG model. However, these works had two drawbacks: reliance on additional retrieval tools, and uncertainty regarding semantic consistency between the retrieved headline and the source article. The NHG model uncertainty can be utilized to generate hypotheses. The hypotheses generated based on a well-trained NHG model not only contain salient information but also exhibit diversity, making them suitable as soft templates. In this study, we use a basic NHG model to generate multiple diverse hypotheses as candidate templates. Then, we propose a novel Multiple-Hypotheses-based NHG (MH-NHG) model. Experiments on English headline generation tasks demonstrate that it outperforms several baseline systems and achieves a comparable performance with the state-of-the-art system. This indicates that MH-NHG can generate more accurate headlines guided by multiple hypotheses.
Funder
Natural Science Foundation of China
Research Program of science and technology at Universities of Inner Mongolia Autonomous Region
MOE (Ministry of Education in China) Humanities and Social Sciences Foundation
Inner Mongolia Autonomous Region Science and Technology Planning Project
Natural Science Foundation of Inner Mongolia
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference49 articles.
1. Rush, A.M., Chopra, S., and Weston, J. (2015, January 17–21). A neural attention model for abstractive sentence summarization. Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal.
2. Gu, J., Lu, Z., Li, H., and Li, V.O. (2016, January 7–12). Incorporating copying mechanism in sequence-to-Sequence learning. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany.
3. Recent Advances on Neural Headline Generation;Ayana;J. Comput. Sci. Technol.,2017
4. Wang, K., Quan, X., and Wang, R. (August, January 28). BiSET: Bi-directional Selective Encoding with Template for Abstractive Summarization. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy.
5. Takase, S., and Kiyono, S. (2021, January 6–11). Rethinking Perturbations in Encoder-Decoders for Fast Training. Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online.