Affiliation:
1. School of Mechatronic Engineering and Automation, Shanghai University, Shanghai 200444, People's Republic of China
Abstract
Answer selection is one of the key steps in many question answering (QA) applications. In this paper, a new deep model with two kinds of attention is proposed for answer selection: the double attention recurrent convolution neural network (DARCNN). Double attention means self-attention and cross-attention. The design inspiration of this model came from the transformer in the domain of machine translation. Self-attention can directly calculate dependencies between words regardless of the distance. However, self-attention ignores the distinction between its surrounding words and other words. Thus, we design a decay self-attention that prioritizes local words in a sentence. In addition, cross-attention is established to achieve interaction between question and candidate answer. With the outputs of self-attention and decay self-attention, we can get two kinds of interactive information via cross-attention. Finally, using the feature vectors of the question and answer, elementwise multiplication is used to combine with them and multilayer perceptron is used to predict the matching score. Experimental results on four QA datasets containing Chinese and English show that DARCNN performs better than other answer selection models, thereby demonstrating the effectiveness of self-attention, decay self-attention and cross-attention in answer selection tasks.
Funder
China Postdoctoral Science Foundation
National Natural Science Foundation of China
Reference36 articles.
1. Tang D Qin B Liu T. 2015 Document modeling with gated recurrent neural network for sentiment classification. In Proc. of the 2015 Conf. on EMNLP Lisbon Portugal 17–21 September pp. 1422-1432. Doha Qatar: Association for Computational Linguistics.
2. Bahdanau D Cho K Bengio Y. 2015 Neural machine translation by jointly learning to align and translate. In Int. Conf. on Learning Representations San Diego CA 7-9 May. Computational and Biological Learning Society.
3. Rush AM Chopra S Weston J. 2015 A neural attention model for sentence summarization. In Proc. of the 2014 Conf. on EMNLP Lisbon Portugal 17–21 September pp. 379-389. Doha Qatar: Association for Computational Linguistics.
4. Underactuated robotics: a review;He B;Int. J. Adv. Robot. Syst,2019
5. A robust speech recognition system for communication robots in noisy environments;Ishi CT;IEEE Trans. Robot.,2008
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献