Abstract
AbstractAcknowledgments in scientific papers may give an insight into aspects of the scientific community, such as reward systems, collaboration patterns, and hidden research trends. The aim of the paper is to evaluate the performance of different embedding models for the task of automatic extraction and classification of acknowledged entities from the acknowledgment text in scientific papers. We trained and implemented a named entity recognition (NER) task using the flair NLP framework. The training was conducted using three default Flair NER models with four differently-sized corpora and different versions of the flair NLP framework. The Flair Embeddings model trained on the medium corpus with the latest FLAIR version showed the best accuracy of 0.79. Expanding the size of a training corpus from very small to medium size massively increased the accuracy of all training algorithms, but further expansion of the training corpus did not bring further improvement. Moreover, the performance of the model slightly deteriorated. Our model is able to recognize six entity types: funding agency, grant number, individuals, university, corporation, and miscellaneous. The model works more precisely for some entity types than for others; thus, individuals and grant numbers showed a very good F1-Score over 0.9. Most of the previous works on acknowledgment analysis were limited by the manual evaluation of data and therefore by the amount of processed data. This model can be applied for the comprehensive analysis of acknowledgment texts and may potentially make a great contribution to the field of automated acknowledgment analysis.
Funder
GESIS – Leibniz-Institut für Sozialwissenschaften e.V.
Publisher
Springer Science and Business Media LLC
Subject
Library and Information Sciences,Computer Science Applications,General Social Sciences
Reference49 articles.
1. Akbik, A., Bergmann, T., Blythe, D., Rasul, K., Schweter, S., & Vollgraf, R. 2019. FLAIR: An Easy-to-Use Framework for State-of-the-Art NLP. Minneapolis, Minnesota (pp. 54–59). Association for Computational Linguistics.
2. Akbik, A., Blythe, D., & Vollgraf, R. (2018). Contextual string embeddings for sequence labeling. In 2018, 27th International Conference on Computational Linguistics (pp. 1638–1649).
3. Alexandera, D. & Vries, A. P. (2021). This research is funded by...”: Named Entity Recognition of financial information in research papers. In BIR 2021: 11th International Workshop on Bibliometric-enhanced Information Retrieval at ECIR (pp. 102–110).
4. Beltagy, I., Lo, K., & Cohan, A. (2019). SciBERT: A pretrained language model for scientific text. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) (pp. 3613–3618). Association for Computational Linguistics.
5. Borst, T., Mielck, J., Nannt, M., & Riese, W. (2022). Extracting funder information from scientific papers—Experiences with question answering. In Silvello, G., O. Corcho, P. Manghi, G.M. Di Nunzio, K. Golub, N. Ferro, and A. Poggi (Eds.),Linking theory and practice of digital libraries (Vol. 13541, pp. 289–296). Springer International Publishing. Series Title: Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-031-16802-4_24.
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献