EPTQA:a Chinese medical prompt learning method based on entity pair type question answering

Author:

Su Zhitong1,Yu Xinran2,Li Songpu3,Chen Peng1

Affiliation:

1. College of Computer and Information Technology, China Three Gorges University

2. The Second Clinical Medical College, Lanzhou University

3. College of Economics&Management, China Three Gorges University

Abstract

Abstract As a mainstream method in current natural language processing tasks, the “pre-train, fine-tune” method has achieved quite good results in various scenarios. However, the “pre-train, fine-tune” method performs poorly on few-shot learning tasks. Recently, prompt learning has gained popularity. Prompt learning transforms various types of natural language processing tasks into pre-training tasks and shows good results on few-shot learning tasks. The prompt learning method based on entity pair type question answering proposed in this paper creatively applies the prompt learning method successfully to the Chinese medical relationship extraction task. The proposed model shows good results on both full data and and low resource datasets. Background Chinese medical relation extraction is an important step in building a complete medical knowledge system. Although the “pre-train, fine-tune” paradigm has shown good results in the Chinese medical relationship extraction task, the “pre-train, fine-tune” paradigm has slow model convergence, and the results are not satisfactory in the small-sample relationship extraction task. These problems are related to the scarcity of accurately labelled medical text data and the large differences between upstream and downstream models. Results Given the aforementioned problems, we propose a prompt learning method that is based on entity pair type question answering. To start with, we preprocessed the Chinese medical text dataset by transforming the data into a sentence-level relation extraction form, which is more appropriate for prompt learning. The relationship template is then devised by combining entity types, which effectively address the issue of expressing the Chinese medical relationship in an accurate and brief manner. Following the fine-tuning of the pre-trained model, this method can accurately restore the mask and present very competitive outcomes on the full data and low resource data of numerous Chinese medical datasets. Conclusions The method proposed in this paper is superior to the traditional “pre-train, fine-tune” approach, as it can efficiently extract the connections between Chinese medical entities. This method is particularly effective when dealing with small sample datasets. The efficiency of the model can be further improved by using the relationship filtering method which is based on the relevant rules of the Chinese medical entities.

Publisher

Research Square Platform LLC

Reference31 articles.

1. Relation Extraction for Chinese Clinical Records Using Multi-View Graph Learning;Ruan C;in IEEE Access,2020

2. Joint Model of Entity Recognition and Relation Extraction with Self-attention Mechanism;Maofu Liu Y;ACM Trans Asian Low-Resour Lang Inf Process,2020

3. Tao Chen M, Wu H, Li. A general approach for improving deep learning-based medical relation extraction using a pre-trained model and fine-tuning, Database, Volume 2019, 2019, baz116, https://doi.org/10.1093/database/baz116.

4. Xue K, Zhou Y, Ma Z, Ruan T, Zhang H, Ping He. and. 2019. Fine-tuning BERT for Joint Entity and Relation Extraction in Chinese Medical Text.

5. Zhang J, Hao K, Tang Xue-song, Cai X, Yan Xiao, and, Wang T. 2020. A multi-feature fusion model for Chinese relation extraction with entity sense. Knowl.-Based Syst. 206, 106348 (2020), 1–10.

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3