Utilizing BERT for Information Retrieval: Survey, Applications, Resources, and Challenges

Author:

Wang Jiajia1ORCID,Huang Jimmy Xiangji2ORCID,Tu Xinhui3ORCID,Wang Junmei4ORCID,Huang Angela Jennifer5ORCID,Laskar Md Tahmid Rahman6ORCID,Bhuiyan Amran2ORCID

Affiliation:

1. School of Sciences, Henan University of Technology, Zhengzhou, China

2. Information Retrieval and Knowledge Management Research Lab, York University, Toronto, Canada

3. School of Computer Science, Central China Normal University, Wuhan, China

4. School of Computer, Hangzhou Dianzi University, Hangzhou, China

5. Lassonde School of Engineering, York University, Toronto, Canada

6. York University & Dialpad Inc., Toronto, Canada

Abstract

Recent years have witnessed a substantial increase in the use of deep learning to solve various natural language processing (NLP) problems. Early deep learning models were constrained by their sequential or unidirectional nature, such that they struggled to capture the contextual relationships across text inputs. The introduction of bidirectional encoder representations from transformers (BERT) leads to a robust encoder for the transformer model that can understand the broader context and deliver state-of-the-art performance across various NLP tasks. This has inspired researchers and practitioners to apply BERT to practical problems, such as information retrieval (IR). A survey that focuses on a comprehensive analysis of prevalent approaches that apply pretrained transformer encoders like BERT to IR can thus be useful for academia and the industry. In light of this, we revisit a variety of BERT-based methods in this survey, cover a wide range of techniques of IR, and group them into six high-level categories: (i) handling long documents, (ii) integrating semantic information, (iii) balancing effectiveness and efficiency, (iv) predicting the weights of terms, (v) query expansion, and (vi) document expansion. We also provide links to resources, including datasets and toolkits, for BERT-based IR systems. Additionally, we highlight the advantages of employing encoder-based BERT models in contrast to recent large language models like ChatGPT, which are decoder-based and demand extensive computational resources. Finally, we summarize the comprehensive outcomes of the survey and suggest directions for future research in the area.

Funder

Natural Science and Engineering Research Council (NSERC) of Canada

York Research Chairs

Ontario Research Fund-Research Excellence

Publisher

Association for Computing Machinery (ACM)

Reference155 articles.

1. A. Abolghasemi, S. Verberne, and L. Azzopardi. 2022. Improving BERT-based query-by-document retrieval with multi-task optimization. In European Conference on Information Retrieval. Springer, Cham.

2. A. Babashzadeh, M. Daoud, and J. X. Huang. 2023. Using semantic-based association rule mining for improving clinical text retrieval. In Health Information Science: Second International Conference, HIS 2013. Springer, Berlin, 2013.

3. Working memory;Baddeley A.;Science,1992

4. Palm: Scaling language modeling with pathways;Chowdhery A.;J. Mach. Learn. Res.,2023

5. Framewise phoneme classification with bidirectional lstm and other neural network architectures;Graves A.;Neural Netw.,2005

Cited by 3 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3