Addressing Long-Distance Dependencies in AMR Parsing with Hierarchical Clause Annotation
-
Published:2023-09-16
Issue:18
Volume:12
Page:3908
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Fan Yunlong12ORCID, Li Bin12ORCID, Sataer Yikemaiti12, Gao Miao12, Shi Chuanqi12, Gao Zhiqiang12
Affiliation:
1. School of Computer Science and Engineering, Southeast University, Nanjing 211189, China 2. Key Laboratory of Computer Network and Information Integration, Ministry of Education, Southeast University, Nanjing 211189, China
Abstract
Most natural language processing (NLP) tasks operationalize an input sentence as a sequence with token-level embeddings and features, despite its clausal structure. Taking abstract meaning representation (AMR) parsing as an example, recent parsers are empowered by transformers and pre-trained language models, but long-distance dependencies (LDDs) introduced by long sequences are still open problems. We argue that LDDs are not actually to blame for the sequence length but are essentially related to the internal clause hierarchy. Typically, non-verb words in a clause cannot depend on words outside of it, and verbs from different but related clauses have much longer dependencies than those in the same clause. With this intuition, we introduce a type of clausal feature, hierarchical clause annotation (HCA), into AMR parsing and propose two HCA-based approaches, HCA-based self-attention (HCA-SA) and HCA-based curriculum learning (HCA-CL), to integrate HCA trees of complex sentences for addressing LDDs. We conduct extensive experiments on two in-distribution (ID) AMR datasets (AMR 2.0 and AMR 3.0) and three out-of-distribution (OOD) ones (TLP, New3, and Bio). Experimental results show that our HCA-based approaches achieve significant and explainable improvements (0.7 Smatch score in both ID datasets; 2.3, 0.7, and 2.6 in three OOD datasets, respectively) against the baseline model and outperform the state-of-the-art (SOTA) model (0.7 Smatch score in the OOD dataset, Bio) when encountering sentences with complex clausal structures that introduce most LDD cases.
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Reference59 articles.
1. Li, Z., Cai, J., He, S., and Zhao, H. (2018, January 20–26). Seq2seq Dependency Parsing. Proceedings of the 27th International Conference on Computational Linguistics, Santa Fe, NM, USA. 2. Tian, Y., Song, Y., Xia, F., and Zhang, T. (2020, January 16–20). Improving Constituency Parsing with Span Attention. Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online. 3. He, L., Lee, K., Lewis, M., and Zettlemoyer, L. (August, January 30). Deep Semantic Role Labeling: What Works and What’s Next. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, BC, Canada. 4. Tang, G., Müller, M., Rios, A., and Sennrich, R. (November, January 31). Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium. 5. Jia, Y., Ye, Y., Feng, Y., Lai, Y., Yan, R., and Zhao, D. (2018, January 15–20). Modeling discourse cohesion for discourse parsing via memory network. Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|