Design and implementation of information extraction system for scientific literature using fine-tuned deep learning models

Author:

Won Kwanghee1,Jang Youngsun1,Choi Hyung-do2,Shin Sung1

Affiliation:

1. South Dakota State University, Brookings, SD

2. Electronics and Telecom Research Institute, Daejeon, South Korea

Abstract

This paper presents an overview of a quality scoring system that utilizes pre-trained deep neural network models. Two types of DL models, a classification and extractive question answering (EQA) models are used to implement components of the system. The abstracts of the scientific literature are classified into two groups, in-vivo and in-vitro, and a question and answering model architecture is constructed for extracting the following types of information (animal type, the number of animals, exposure dose, and signal frequency). The Bidirectional Encoder Representations of Transformers (BERT) model pre-trained with a large text corpus is used as our baseline model for classification and EQA tasks. The models are fine-tuned with 455 EMF-related research papers. In our experiments, the fine-tuned model showed improved performance on EQA tasks for the four-categories of questions compared to the baseline, and it also showed improvements on similar questions that were not used in training. This suggests the importance of retraining of deep learning model specifically in some areas requiring domain expertise such as scientific papers. However, additional research is needed on some implementation issues, in such cases where there are still multiple answers, or where there is no answer given in a context.

Publisher

Association for Computing Machinery (ACM)

Subject

Industrial and Manufacturing Engineering

Reference33 articles.

1. K. Pearce , T. Zhan , A. Komanduri , and J. Zhan . A Comparative Study of Transformer-Based Language Models on Extractive Question Answering . In Computing Research Repository (CoRR) , October 2021 . K. Pearce, T. Zhan, A. Komanduri, and J. Zhan. A Comparative Study of Transformer-Based Language Models on Extractive Question Answering. In Computing Research Repository (CoRR), October 2021.

2. D. Bahdanau , KH. Cho , and Y. Bengio . Neural Machine Translation by Jointly Learning to Align and Translate . In the Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015) , May 2015 , San Diego, CA, USA. D. Bahdanau, KH. Cho, and Y. Bengio. Neural Machine Translation by Jointly Learning to Align and Translate. In the Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), May 2015, San Diego, CA, USA.

3. A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents

4. P. Badjatiya , S. Gupta , M. Gupta , and V. Varma . Deep Learning for Hate Speech Detection in Tweets. In the Proceedings of the 26th International Conference on World Wide Web Companion, (WWW '17 Companion) , April 2017 , Perth, Australia . 759--760. International World Wide Web Conference Committee . P. Badjatiya, S. Gupta, M. Gupta, and V. Varma. Deep Learning for Hate Speech Detection in Tweets. In the Proceedings of the 26th International Conference on World Wide Web Companion, (WWW '17 Companion), April 2017, Perth, Australia. 759--760. International World Wide Web Conference Committee.

5. C. Qu , L. Yang , M. Qiu , W. B. Croft , Y. Zhang , and M. Iyyer . BERT with History Answer Embedding for Conversational Question Answering. In the Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR'19) , July 2019 , New York, NY , United States. 1133--1136. Association for Computing Machinery . C. Qu, L. Yang, M. Qiu, W. B. Croft, Y. Zhang, and M. Iyyer. BERT with History Answer Embedding for Conversational Question Answering. In the Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR'19), July 2019, New York, NY, United States. 1133--1136. Association for Computing Machinery.

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3