Affiliation:
1. Department of Internal Medicine (Digestive Diseases) Yale School of Medicine New Haven Connecticut USA
2. Department of Medical, Surgical and Health Sciences University of Trieste Trieste Italy
3. Department of Engineering and Architecture University of Trieste Trieste Italy
4. Department of Mathematics at Baruch College City University of new York New York New York USA
5. Department of Internal Medicine Yale School of Medicine New Haven Connecticut USA
6. Research & Education Librarian (Clinical) at Harvey Cushing/John Hay Whitney Medical Library Yale University New Haven Connecticut USA
Abstract
SummaryBackgroundInterest in large language models (LLMs), such as OpenAI's ChatGPT, across multiple specialties has grown as a source of patient‐facing medical advice and provider‐facing clinical decision support. The accuracy of LLM responses for gastroenterology and hepatology‐related questions is unknown.AimsTo evaluate the accuracy and potential safety implications for LLMs for the diagnosis, management and treatment of questions related to gastroenterology and hepatology.MethodsWe conducted a systematic literature search including Cochrane Library, Google Scholar, Ovid Embase, Ovid MEDLINE, PubMed, Scopus and the Web of Science Core Collection to identify relevant articles published from inception until January 28, 2024, using a combination of keywords and controlled vocabulary for LLMs and gastroenterology or hepatology. Accuracy was defined as the percentage of entirely correct answers.ResultsAmong the 1671 reports screened, we identified 33 full‐text articles on using LLMs in gastroenterology and hepatology and included 18 in the final analysis. The accuracy of question‐responding varied across different model versions. For example, accuracy ranged from 6.4% to 45.5% with ChatGPT‐3.5 and was between 40% and 91.4% with ChatGPT‐4. In addition, the absence of standardised methodology and reporting metrics for studies involving LLMs places all the studies at a high risk of bias and does not allow for the generalisation of single‐study results.ConclusionsCurrent general‐purpose LLMs have unacceptably low accuracy on clinical gastroenterology and hepatology tasks, which may lead to adverse patient safety events through incorrect information or triage recommendations, which might overburden healthcare systems or delay necessary care.
Reference62 articles.
1. DouglasMR.Large language models.arXiv: 2307.057822023.
2. NaveedH KhanAU QiuS SaqibM AnwarS UsmanM et al.A comprehensive overview of large language models.arXiv2023.
3. ConsensME DufaultC WainbergM ForsterD KarimzadehM GoodarziH et al.To transformers and beyond: large language models for the genome. arXiv: 2311.076212023.
4. LewisP PerezE PiktusA PetroniF KarpukhinV GoyalN et al.Retrieval‐augmented generation for knowledge‐intensive nlp tasks.2020.
5. ChenB ZhangZ LangrenéN ZhuS.Unleashing the potential of prompt engineering in large language models: a comprehensive review.2023.
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献