Conserving Semantic Unit Information and Simplifying Syntactic Constituents to Improve Implicit Discourse Relation Recognition
Author:
Fang Zhongyang1, Cong Yue1, Chai Yuhan1, Gao Chengliang1, Chen Ximing1, Qiu Jing1
Affiliation:
1. The Cyberspace Institute of Advanced Technology, Guangzhou University, Guangzhou 510006, China
Abstract
Implicit discourse relation recognition (IDRR) has long been considered a challenging problem in shallow discourse parsing. The absence of connectives makes such relations implicit and requires much more effort to understand the semantics of the text. Thus, it is important to preserve the semantic completeness before any attempt to predict the discourse relation. However, word level embedding, widely used in existing works, may lead to a loss of semantics by splitting some phrases that should be treated as complete semantic units. In this article, we proposed three methods to segment a sentence into complete semantic units: a corpus-based method to serve as the baseline, a constituent parsing tree-based method, and a dependency parsing tree-based method to provide a more flexible and automatic way to divide the sentence. The segmented sentence will then be embedded at the level of semantic units so the embeddings could be fed into the IDRR networks and play the same role as word embeddings. We implemented our methods into one of the recent IDRR models to compare the performance with the original version using word level embeddings. Results show that proper embedding level better conserves the semantic information in the sentence and helps to enhance the performance of IDRR models.
Funder
National Natural Science Foundation of China National Key Research and Development Plan Guangdong Province Key Research and Development Plan Major Key Project of PCL Joint Research Fund of Guangzhou and University Guangdong Higher Education Innovation Group Guangdong Province Universities and Colleges Pearl River Scholar Funded Scheme
Subject
General Physics and Astronomy
Reference56 articles.
1. Survey of Implicit Discourse Relation Recognition Based on Deep Learning;Hu;Comput. Sci.,2020 2. A systematic study of neural discourse models for implicit discourse relation;Rutherford;Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics,2017 3. Xue, N., Ng, H.T., Pradhan, S., Prasad, R., Bryant, C., and Rutherford, A. (2015, January 30–31). The conll-2015 shared task on shallow discourse parsing. Proceedings of the Nineteenth Conference on Computational Natural Language Learning-Shared Task (CoNLL ’15), Beijing, China. 4. Xue, N., Ng, H.T., Pradhan, S., Rutherford, A., Webber, B., Wang, C., and Wang, H. (2016, January 7–12). Conll 2016 shared task on multilingual shallow discourse parsing. Proceedings of the Twentieth Conference on Computational Natural Language Learning-Shared Task (CoNLL ’16), Berlin, Germany. 5. Gerani, S., Mehdad, Y., Carenini, G., Ng, R., and Nejat, B. (2014, January 25–29). Abstractive summarization of product reviews using discourse structure. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP ’14), Doha, Qatar.
|
|