Hierarchical Clause Annotation: Building a Clause-Level Corpus for Semantic Parsing with Complex Sentences
-
Published:2023-08-19
Issue:16
Volume:13
Page:9412
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Fan Yunlong12ORCID, Li Bin12ORCID, Sataer Yikemaiti12, Gao Miao12, Shi Chuanqi12, Cao Siyi3, Gao Zhiqiang12
Affiliation:
1. School of Computer Science and Engineering, Southeast University, Nanjing 211189, China 2. Key Laboratory of Computer Network and Information Integration, Southeast University, Ministry of Education, Nanjing 211189, China 3. School of Foreign Languages, Southeast University, Nanjing 211189, China
Abstract
Most natural-language-processing (NLP) tasks suffer performance degradation when encountering long complex sentences, such as semantic parsing, syntactic parsing, machine translation, and text summarization. Previous works addressed the issue with the intuition of decomposing complex sentences and linking simple ones, such as rhetorical-structure-theory (RST)-style discourse parsing, split-and-rephrase (SPRP), text simplification (TS), simple sentence decomposition (SSD), etc. However, these works are not applicable for semantic parsing such as abstract meaning representation (AMR) parsing and semantic dependency parsing due to misalignments with semantic relations and unavailabilities to preserve the original semantics. Following the same intuition and avoiding the deficiencies of previous works, we propose a novel framework, hierarchical clause annotation (HCA), for capturing clausal structures of complex sentences, based on the linguistic research of clause hierarchy. With the HCA framework, we annotated a large HCA corpus to explore the potentialities of integrating HCA structural features into semantic parsing with complex sentences. Moreover, we decomposed HCA into two subtasks, i.e., clause segmentation and clause parsing, and provide neural baseline models for more-silver annotations. In evaluating the proposed models on our manually annotated HCA dataset, the performances of clause segmentation and parsing resulted in 91.3% F1-scores and 88.5% Parseval scores, respectively. Due to the same model architectures employed, the performance differences of the clause/discourse segmentation and parsing subtasks was reflected in our HCA corpus and compared discourse corpora, where our sentences contained more segment units and fewer interrelations than those in the compared corpora.
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference43 articles.
1. Sataer, Y., Shi, C., Gao, M., Fan, Y., Li, B., and Gao, Z. (2023, January 4–10). Integrating Syntactic and Semantic Knowledge in AMR Parsing with Heterogeneous Graph Attention Network. Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece. 2. Li, B., Gao, M., Fan, Y., Sataer, Y., Gao, Z., and Gui, Y. (2022, January 12–17). DynGL-SDP: Dynamic Graph Learning for Semantic Dependency Parsing. Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea. 3. Tian, Y., Song, Y., Xia, F., and Zhang, T. (2020, January 16–20). Improving Constituency Parsing with Span Attention. Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online. 4. He, L., Lee, K., Lewis, M., and Zettlemoyer, L. (August, January 30). Deep Semantic Role Labeling: What Works and What’s Next. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada. 5. Tang, G., Müller, M., Rios, A., and Sennrich, R. (November, January 31). Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium.
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|