Author:
Li Yufei,Zhou Xiangyu,Ma Jie,Ma Xiaoyong,Cheng Pengzhen,Gong Tieliang,Li Chen
Abstract
Abstract
Background
Bio-entity Coreference Resolution (CR) is a vital task in biomedical text mining. An important issue in CR is the differential representation of identical mentions as their similar representations may make the coreference more puzzling. However, when extracting features, existing neural network-based models may bring additional noise to the distinction of identical mentions since they tend to get similar or even identical feature representations.
Methods
We propose a context-aware feature attention model to distinguish similar or identical text units effectively for better resolving coreference. The new model can represent the identical mentions based on different contexts by adaptively exploiting features, which enables the model reduce the text noise and capture the semantic information effectively.
Results
The experimental results show that the proposed model brings significant improvements on most of the baseline for coreference resolution and mention detection on the BioNLP dataset and CRAFT-CR dataset. The empirical studies further demonstrate its superior performance on the differential representation and coreferential link of identical mentions.
Conclusions
Identical mentions impose difficulties on the current methods of Bio-entity coreference resolution. Thus, we propose the context-aware feature attention model to better distinguish identical mentions and achieve superior performance on both coreference resolution and mention detection, which will further improve the performance of the downstream tasks.
Funder
National Natural Science Foundation of China
Innovative Research Group Project of the National Natural Science Foundation of China
Publisher
Springer Science and Business Media LLC
Subject
Health Informatics,Health Policy,Computer Science Applications
Reference23 articles.
1. Dai Z, Fei H, Li P.: Coreference aware representation learning for neural named entity recognition. In: Proceedings of the international joint conference on artificial intelligence (IJCAI), Macau, 2019. p. 4946–53.
2. Luan Y, He L, Ostendorf M, Hajishirzi H. Multi-task identification of entities, relations, and coreference for scientific knowledge graph construction. In: Proceedings of the 2018 conference on empirical methods in natural language processing, 2018; Association for Computational Linguistics, Brussels, Belgium p. 3219–32. https://doi.org/10.18653/v1/D18-1360. https://www.aclweb.org/anthology/D18-1360.
3. Zeng S, Xu R, Chang B, Li L. Double graph based reasoning for document-level relation extraction. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), Association for Computational Linguistics, 2020. p. 1630–40. https://doi.org/10.18653/v1/2020.emnlp-main.127. https://www.aclweb.org/anthology/2020.emnlp-main.127.
4. Hajishirzi H, Zilles L, Weld DS, Zettlemoyer L. Joint coreference resolution and named-entity linking with multi-pass sieves. In: Proceedings of the 2013 conference on empirical methods in natural language processing, 2013. p. 289–99.
5. Weston J, Bordes A, Chopra S, Rush MA, Merrienboer VB, Joulin A. Towards ai-complete question answering: a set of prerequisite toy tasks. In: International conference on learning representations, 2015.