Abstract
Currently, in the field of biomedical named entity recognition, CharCNN (Character-level Convolutional Neural Networks) or CharRNN (Character-level Recurrent Neural Network) is typically used independently to extract character features. However, this approach does not consider the complementary capabilities between them and only concatenates word features, ignoring the feature information during the process of word integration. Based on this, this paper proposes a method of multi-cross attention feature fusion. First, DistilBioBERT and CharCNN and CharLSTM are used to perform cross-attention word-char (word features and character features) fusion separately. Then, the two feature vectors obtained from cross-attention fusion are fused again through cross-attention to obtain the final feature vector. Subsequently, a BiLSTM is introduced with a multi-head attention mechanism to enhance the model’s ability to focus on key information features and further improve model performance. Finally, the output layer is used to output the final result. Experimental results show that the proposed model achieves the best F1 values of 90.76%, 89.79%, 94.98%, 80.27% and 88.84% on NCBI-Disease, BC5CDR-Disease, BC5CDR-Chem, JNLPBA and BC2GM biomedical datasets respectively. This indicates that our model can capture richer semantic features and improve the ability to recognize entities.
Funder
Natural Science Foundation of Heilongjiang Province
Publisher
Public Library of Science (PLoS)
Reference33 articles.
1. Multi-attention deep neural network fusing character and word embedding for clinical and biomedical concept extraction;S Fan;Inf Sci,2022
2. The SWISS-PROT protein knowledgebase and its supplement TrEMBL in 2003;B Boeckmann;Nucleic Acids Res,2003
3. GenBank;EW Sayers;Nucleic Acids Research,2019
4. Improving biomedical information retrieval with neural retrievers;M Luo;Proceedings of the AAAI Conference on Artificial Intelligence,2022