Affiliation:
1. Hangzhou Normal University
Abstract
Abstract
In the era of information technology, the electronic medical system generates vast and diverse data on a daily basis. Maximizing the utilization of this data will have a profound impact on clinical decision-making and public health. Currently, named entity recognition technology has reached maturity in the field of English. However, due to the scarcity of corpus and the complexity of semantic boundary recognition, Chinese named entity recognition remains an area that warrants further research efforts. Based on the classical model of ALBERT-BiLSTM-Self-Attention-CRF, we propose a named entity recognition model named DWI-Pos which integrates the Pos (position information) of entity words and the features of POS (part of speech), and use a DWI (Dynamic Windows Interception mechanism). Subsequently, two comparative experiments were conducted on the BERT-CRF and LSTM-CRF models, and their effectiveness was validated by the results obtained. Furthermore, the entity word position information designed in this study played a significant role in improving the performance of the proposed model. The dataset employed in this study was derived from the sub-tasks of CCKS2019, which includes named entity recognition of Chinese electronic medical records. This task encompasses two sub-tasks, namely medical named entity recognition and medical entity and attribute extraction. The focus of this research was on medical named entity recognition, where the achieved F1 value was 0.95, surpassing the F1 value of the ELMo-ET-CRF model by 0.09.
Publisher
Research Square Platform LLC
Reference19 articles.
1. Zaremba, Wojciech et al. “Recurrent Neural Network Regularization.” ArXiv abs/1409.2329 (2014): n. pag.
2. “ImageNet classification with deep convolutional neural networks;Krizhevsky Alex;Communications of the ACM,2012
3. “Long Short-Term Memory;Hochreiter Sepp;Neural Computation,1997
4. “ALBERT: A Lite BERT for Self-supervised Learning of Language Representations.” ArXiv abs/1909;Lan Zhenzhong,2020
5. “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding;Devlin Jacob;ArXiv abs,2019