Cross-Lingual Named Entity Recognition Based on Attention and Adversarial Training
-
Published:2023-02-16
Issue:4
Volume:13
Page:2548
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Wang Hao12, Zhou Lekai12, Duan Jianyong12, He Li12
Affiliation:
1. School of Information Science and Technology, North China University of Technology, Beijing 100144, China 2. CNONIX National Standard Application and Promotion Lab, Beijing 100144, China
Abstract
Named entity recognition aims to extract entities with specific meaning from unstructured text. Currently, deep learning methods have been widely used for this task and have achieved remarkable results, but it is often difficult to achieve better results with less labeled data. To address this problem, this paper proposes a method for cross-lingual entity recognition based on an attention mechanism and adversarial training, using resource-rich language annotation data to migrate to low-resource languages for named entity recognition tasks and outputting changing semantic vectors through the attention mechanism to effectively solve the long-sequence semantic dilution problem. To verify the effectiveness of the proposed method, the method in this paper is applied to the English–Chinese cross-lingual named entity recognition task based on the WeiboNER data set and the People-Daily2004 data set. The obtained F1 value of the optimal model is 53.22% (a 6.29% improvement compared to the baseline). The experimental results show that the cross-lingual adversarial named entity recognition method proposed in this paper can significantly improve the results of named entity recognition in low resource languages.
Funder
R&D Program of Beijing Municipal Education Commission National Natural Science Foundation of China Beijing Urban Governance Research Center
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference33 articles.
1. Cross-lingual language model pretraining;Conneau;Advances in Neural Information Processing Systems 32,2019 2. Wu, Q., Lin, Z., Wang, G., Chen, H., Karlsson, B.F., Huang, B., and Lin, C.-Y. (2020, January 7–12). Enhanced meta-learning for cross-lingual named entity recognition with minimal resources. Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA. 3. Patra, B., Moniz, J.R.A., Garg, S., Gormley, M.R., and Neubig, G. (2019). Bilingual lexicon induction with semi-supervision in non-isometric embedding spaces. arXiv. 4. Liu, L., Ding, B., Bing, L., Joty, S., Si, L., and Miao, C. (2021, January 1–6). MulDA: A Multilingual Data Augmentation Framework for Low-Resource Cross-Lingual NER. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Online. Long Papers. 5. Tzeng, E., Hoffman, J., Zhang, N., Saenko, K., and Darrell, T. (2014). Deep domain confusion: Maximizing for domain invariance. arXiv.
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|