Incorporating Entity Type-Aware and Word–Word Relation-Aware Attention in Generative Named Entity Recognition
-
Published:2024-04-08
Issue:7
Volume:13
Page:1407
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Mo Ying1ORCID, Li Zhoujun1
Affiliation:
1. State Key Lab of Software Development Environment, Beihang University, Beijing 100191, China
Abstract
Named entity recognition (NER) is a critical subtask in natural language processing. It is particularly valuable to gain a deeper understanding of entity boundaries and entity types when addressing the NER problem. Most previous sequential labeling models are task-specific, while recent years have witnessed the rise of generative models due to the advantage of tackling NER tasks in the encoder–decoder framework. Despite achieving promising performance, our pilot studies demonstrate that existing generative models are ineffective at detecting entity boundaries and estimating entity types. In this paper, a multiple attention framework is proposed which introduces the attention of entity-type embedding and word–word relation into the named entity recognition task. To improve the accuracy of entity-type mapping, we adopt an external knowledge base to calculate the prior entity-type distributions and then incorporate the information input to the model via the encoder’s self-attention. To enhance the contextual information, we take the entity types as part of the input. Our method obtains the other attention from the hidden states of entity types and utilizes it in self- and cross-attention mechanisms in the decoder. We transform the entity boundary information in the sequence into word–word relations and extract the corresponding embedding into the cross-attention mechanism. Through word–word relation information, the method can learn and understand more entity boundary information, thereby improving its entity recognition accuracy. We performed experiments on extensive NER benchmarks, including four flat and two long entity benchmarks. Our approach significantly improves or performs similarly to the best generative NER models. The experimental results demonstrate that our method can substantially enhance the capabilities of generative NER models.
Funder
National Natural Science Foundation of China State Key Laboratory of Software Development Environment
Reference73 articles.
1. Cavedon, L., and Zukerman, I. (2006). Proceedings of the ALTA 2006, Australasian Language Technology Association. 2. Li, Q., and Ji, H. (2014, January 22–27). Incremental Joint Extraction of Entity Mentions and Relations. Proceedings of the ACL 2014, Baltimore, MD, USA. 3. Zhong, Z., and Chen, D. (2021, January 6–11). A Frustratingly Easy Approach for Entity and Relation Extraction. Proceedings of the NAACL-HLT 2021, Online. 4. Lample, G., Ballesteros, M., Subramanian, S., Kawakami, K., and Dyer, C. (2016, January 12–17). Neural Architectures for Named Entity Recognition. Proceedings of the NAACL HLT 2016, San Diego, CA, USA. 5. Strubell, E., Verga, P., Belanger, D., and McCallum, A. (2017, January 9–11). Fast and Accurate Entity Recognition with Iterated Dilated Convolutions. Proceedings of the EMNLP 2017, Copenhagen, Denmark.
|
|