Can large language models provide secondary reliable opinion on treatment options for dermatological diseases?

Author:

Iqbal Usman123ORCID,Lee Leon Tsung-Ju456,Rahmanti Annisa Ristya789,Celi Leo Anthony101112,Li Yu-Chuan Jack781314

Affiliation:

1. School of Population Health, Faculty of Medicine and Health, University of New South Wales (UNSW) , Sydney, NSW 2052, Australia

2. Department of Health , Tasmania 7000, Australia

3. Global Health and Health Security Department, College of Public Health, Taipei Medical University , Taipei 110, Taiwan

4. Graduate Institute of Clinical Medicine, Taipei Medical University , Taipei 110, Taiwan

5. Department of Dermatology, Taipei Medical University Hospital, Taipei Medical University , Taipei 110, Taiwan

6. Department of Dermatology, School of Medicine, College of Medicine, Taipei Medical University , Taipei 110, Taiwan

7. Graduate Institute of Biomedical Informatics, College of Medical Science and Technology, Taipei Medical University , Taipei 110, Taiwan

8. International Center for Health Information and Technology, College of Medical Science and Technology, Taipei Medical University , Taipei 110, Taiwan

9. Department Health Policy and Management, Faculty of Medicine, Public Health and Nursing, Universitas Gadjah Mada , Yogyakarta 55281, Indonesia

10. Laboratory for Computational Physiology, Massachusetts Institute of Technology , Cambridge, MA 02139, United States

11. Division of Pulmonary, Critical Care and Sleep Medicine, Beth Israel Deaconess Medical Center , Boston, MA 02215, United States

12. Department of Biostatistics, Harvard T.H. Chan School of Public Health , Boston, MA 02115, United States

13. Department of Dermatology, Taipei Municipal Wanfang Hospital, Taipei Medical University , Taipei 116, Taiwan

14. The International Medical Informatics Association (IMIA) , Genève CH-1204, Switzerland

Abstract

Abstract Objective To investigate the consistency and reliability of medication recommendations provided by ChatGPT for common dermatological conditions, highlighting the potential for ChatGPT to offer second opinions in patient treatment while also delineating possible limitations. Materials and Methods In this mixed-methods study, we used survey questions in April 2023 for drug recommendations generated by ChatGPT with data from secondary databases, that is, Taiwan’s National Health Insurance Research Database and an US medical center database, and validated by dermatologists. The methodology included preprocessing queries, executing them multiple times, and evaluating ChatGPT responses against the databases and dermatologists. The ChatGPT-generated responses were analyzed statistically in a disease-drug matrix, considering disease-medication associations (Q-value) and expert evaluation. Results ChatGPT achieved a high 98.87% dermatologist approval rate for common dermatological medication recommendations. We evaluated its drug suggestions using the Q-value, showing that human expert validation agreement surpassed Q-value cutoff-based agreement. Varying cutoff values for disease-medication associations, a cutoff of 3 achieved 95.14% accurate prescriptions, 5 yielded 85.42%, and 10 resulted in 72.92%. While ChatGPT offered accurate drug advice, it occasionally included incorrect ATC codes, leading to issues like incorrect drug use and type, nonexistent codes, repeated errors, and incomplete medication codes. Conclusion ChatGPT provides medication recommendations as a second opinion in dermatology treatment, but its reliability and comprehensiveness need refinement for greater accuracy. In the future, integrating a medical domain-specific knowledge base for training and ongoing optimization will enhance the precision of ChatGPT’s results.

Funder

National Science and Technology Council

Higher Education Sprout Project

Ministry of Education

Publisher

Oxford University Press (OUP)

Reference24 articles.

1. Artificial intelligence and machine learning in clinical medicine, 2023;Haug;N Engl J Med,2023

2. ChatGPT: friend or foe?;The Lancet Digital Health;Lancet Digit Health,2023

3. ChatGPT and physicians’ malpractice risk;Mello;JAMA Health Forum,2023

4. ChatGPT: the future of discharge summaries?;Patel;Lancet Digit Health,2023

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3