Affiliation:
1. School of Information Technology, Luoyang Normal University, Luoyang 471934, China
2. Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo 315211, China
Abstract
The typical pretrained model’s feature extraction capabilities are insufficient for medical named entity identification, and it is challenging to express word polysemy, resulting in a low recognition accuracy for electronic medical records. In order to solve this problem, this paper proposes a new model that combines the BERT pretraining model and the BilSTM-CRF model. First, word embedding with semantic information is obtained by pretraining the corpus input to the BERT model. Then, the BiLSTM module is utilized to extract further features from the encoded outputs of BERT in order to account for context information and improve the accuracy of semantic coding. Then, CRF is used to modify the results of BiLSTM to screen out the annotation sequence with the largest score. Finally, extensive experimental results show that the performance of the proposed model is effectively improved compared with other models.
Subject
Electrical and Electronic Engineering,General Computer Science,Signal Processing
Reference39 articles.
1. Convolutional Recurrent Deep Learning Model for Sentence Classification
2. Effective use of word order for text categorization with convolutional neural networks;R. Johnson
3. CNN-RNN based intelligent recommendation for online medical pre-diagnosis support;X. Zhou;IEEE/ACM Transactions on Computational Biology and Bioinformatics,2020
4. Character-level neural network for biomedical named entity recognition
5. GRAM-CNN: a deep learning approach with local context for named entity recognition in biomedical text
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献