Navigating the Landscape of Personalized Medicine: The Relevance of ChatGPT, BingChat, and Bard AI in Nephrology Literature Searches
-
Published:2023-09-30
Issue:10
Volume:13
Page:1457
-
ISSN:2075-4426
-
Container-title:Journal of Personalized Medicine
-
language:en
-
Short-container-title:JPM
Author:
Aiumtrakul Noppawit1ORCID, Thongprayoon Charat2, Suppadungsuk Supawadee23ORCID, Krisanapan Pajaree24ORCID, Miao Jing2ORCID, Qureshi Fawad2, Cheungpasitporn Wisit2ORCID
Affiliation:
1. Department of Medicine, John A. Burns School of Medicine, University of Hawaii, Honolulu, HI 96813, USA 2. Division of Nephrology and Hypertension, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA 3. Chakri Naruebodindra Medical Institute, Faculty of Medicine Ramathibodi Hospital, Mahidol University, Samut Prakan 10540, Thailand 4. Department of Internal Medicine, Faculty of Medicine, Thammasat University, Pathum Thani 12120, Thailand
Abstract
Background and Objectives: Literature reviews are foundational to understanding medical evidence. With AI tools like ChatGPT, Bing Chat and Bard AI emerging as potential aids in this domain, this study aimed to individually assess their citation accuracy within Nephrology, comparing their performance in providing precise. Materials and Methods: We generated the prompt to solicit 20 references in Vancouver style in each 12 Nephrology topics, using ChatGPT, Bing Chat and Bard. We verified the existence and accuracy of the provided references using PubMed, Google Scholar, and Web of Science. We categorized the validity of the references from the AI chatbot into (1) incomplete, (2) fabricated, (3) inaccurate, and (4) accurate. Results: A total of 199 (83%), 158 (66%) and 112 (47%) unique references were provided from ChatGPT, Bing Chat and Bard, respectively. ChatGPT provided 76 (38%) accurate, 82 (41%) inaccurate, 32 (16%) fabricated and 9 (5%) incomplete references. Bing Chat provided 47 (30%) accurate, 77 (49%) inaccurate, 21 (13%) fabricated and 13 (8%) incomplete references. In contrast, Bard provided 3 (3%) accurate, 26 (23%) inaccurate, 71 (63%) fabricated and 12 (11%) incomplete references. The most common error type across platforms was incorrect DOIs. Conclusions: In the field of medicine, the necessity for faultless adherence to research integrity is highlighted, asserting that even small errors cannot be tolerated. The outcomes of this investigation draw attention to inconsistent citation accuracy across the different AI tools evaluated. Despite some promising results, the discrepancies identified call for a cautious and rigorous vetting of AI-sourced references in medicine. Such chatbots, before becoming standard tools, need substantial refinements to assure unwavering precision in their outputs.
Subject
Medicine (miscellaneous)
Reference33 articles.
1. Blanco-Gonzalez, A., Cabezon, A., Seco-Gonzalez, A., Conde-Torres, D., Antelo-Riveiro, P., Pineiro, A., and Garcia-Fandino, R. (2023). The Role of AI in Drug Discovery: Challenges, Opportunities, and Strategies. Pharmaceuticals, 16. 2. Living in the digital era: The impact of digital technologies on human health;Salim;Malays. Fam. Physician,2022 3. Progress in evidence-based medicine: A quarter century on;Djulbegovic;Lancet,2017 4. Cooper, C., Booth, A., Varley-Campbell, J., Britten, N., and Garside, R. (2018). Defining the process to literature searching in systematic reviews: A literature review of guidance and supporting studies. BMC Med. Res. Methodol., 18. 5. National Library of Medicine (2023, August 20). PubMed, Available online: https://pubmed.ncbi.nlm.nih.gov/.
Cited by
16 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|