Optimizing large language models in digestive disease: strategies and challenges to improve clinical outcomes

Author:

Giuffrè Mauro12,Kresevic Simone3,Pugliese Nicola45ORCID,You Kisung6,Shung Dennis L.1

Affiliation:

1. Department of Internal Medicine (Digestive Diseases) Yale School of Medicine New Haven Connecticut USA

2. Department of Medical, Surgical, and Health Sciences University of Trieste Trieste Italy

3. Department of Engineering and Architecture University of Trieste Trieste Italy

4. Division of Internal Medicine and Hepatology, Department of Gastroenterology IRCCS Humanitas Research Hospital Rozzano Italy

5. Department of Biomedical Sciences Humanitas University Pieve Emanuele Italy

6. Department of Mathematics Baruch College, City University of New York New York New York USA

Abstract

AbstractLarge Language Models (LLMs) are transformer‐based neural networks with billions of parameters trained on very large text corpora from diverse sources. LLMs have the potential to improve healthcare due to their capability to parse complex concepts and generate context‐based responses. The interest in LLMs has not spared digestive disease academics, who have mainly investigated foundational LLM accuracy, which ranges from 25% to 90% and is influenced by the lack of standardized rules to report methodologies and results for LLM‐oriented research. In addition, a critical issue is the absence of a universally accepted definition of accuracy, varying from binary to scalar interpretations, often tied to grader expertise without reference to clinical guidelines. We address strategies and challenges to increase accuracy. In particular, LLMs can be infused with domain knowledge using Retrieval Augmented Generation (RAG) or Supervised Fine‐Tuning (SFT) with reinforcement learning from human feedback (RLHF). RAG faces challenges with in‐context window limits and accurate information retrieval from the provided context. SFT, a deeper adaptation method, is computationally demanding and requires specialized knowledge. LLMs may increase patient quality of care across the field of digestive diseases, where physicians are often engaged in screening, treatment and surveillance for a broad range of pathologies for which in‐context learning or SFT with RLHF could improve clinical decision‐making and patient outcomes. However, despite their potential, the safe deployment of LLMs in healthcare still needs to overcome hurdles in accuracy, suggesting a need for strategies that integrate human feedback with advanced model training.

Publisher

Wiley

Reference50 articles.

1. NaveedH KhanAU QiuS et al.A comprehensive overview of large language models.2023.

2. NaziZA PengW.Large language models in healthcare and medical domain: a review.arXiv e‐prints2023.

3. Will Generative Artificial Intelligence Deliver on Its Promise in Health Care?

4. Six ways large language models are changing healthcare

5. Evaluating the Utility of a Large Language Model in Answering Common Patients’ Gastrointestinal Health-Related Questions: Are We There Yet?

Cited by 4 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3