RegRL-KG: Learning an L1 regularized reinforcement agent for keyphrase generation
-
Published:2023-07-20
Issue:4
Volume:27
Page:1003-1021
-
ISSN:1088-467X
-
Container-title:Intelligent Data Analysis
-
language:
-
Short-container-title:IDA
Author:
Yao Yu,Yang Peng,Zhao Guangzhen,Leng Juncheng
Abstract
Keyphrase generation (KG) aims at condensing the content from the source text to the target concise phrases. Though many KG algorithms have been proposed, most of them are tailored into deep learning settings with various specially designed strategies and may fail in solving the bias exposure problem. Reinforcement Learning (RL), a class of control optimization techniques, are well suited to compensate for some of the limitations of deep learning methods. Nevertheless, RL methods typically suffer from four core difficulties in keyphrase generation: environment interaction and effective exploration, complex action control, reward design, and task-specific obstacle. To tackle this difficult but significant task, we present RegRL-KG, including actor-critic based-reinforcement learning control and L1 policy regularization under the first principle of minimizing the maximum likelihood estimation (MLE) criterion by a sequence-to-sequence (Seq2Seq) deep learnining model, for efficient keyphrase generation. The agent utilizes an actor-critic network to control the generated probability distribution and employs L1 policy regularization to solve the bias exposure problem. Extensive experiments show that our method brings improvement in terms of the evaluation metrics on five scientific article benchmark datasets.
Subject
Artificial Intelligence,Computer Vision and Pattern Recognition,Theoretical Computer Science
Reference32 articles.
1. Neuron like adaptive elements that can solve difficult learning control problems;Barto;IEEE Transactions on Systems, Man, and Cybernetics,1983 2. F. Boudin, Y. Gallina and A. Aizawa, Keyphrase generation for scientific document retrieval, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics, ACL 20’, Online, 2016, pp. 1118–1126. 3. K. Bousmalis, G. Trigeorgis, N. Silberman, D. Krishnan and D. Erhan, Domain Separation Networks, in: Proceedings of the Advances in Neural Information Processing Systems, NIPS 16’, Barcelona, Spain, 2016, pp. 343–351. 4. H.P. Chan, W. Chen, L. Wang and I. King, Neural keyphrase generation via reinforcement learning with adaptive rewards, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics, ACL 19’, Florence, Italy, 2019, pp. 2163–2174. 5. W. Chen, H.P. Chan, P.J. Li, L.D. Bing and I. King, An integrated approach for keyphrase generation via exploring the power of retrieval and extraction, in: Proceedings of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 19’, Minneapolis, Minnesota, 2019, pp. 2846–2856.
|
|