ChatGPT4’s proficiency in addressing patients’ questions on systemic lupus erythematosus: a blinded comparative study with specialists

Author:

Xu Dan1ORCID,Zhao Jinxia1,Liu Rui1,Dai Yijun2,Sun Kai3,Wong Priscilla4,Ming Samuel Lee Shang5,Wearn Koh Li5,Wang Jiangyuan6,Xie Shasha6,Zeng Lin7,Mu Rong1,Xu Chuanhui58ORCID

Affiliation:

1. Department of Rheumatology and Immunology, Peking University Third Hospital , Beijing, China

2. Department of Rheumatology and Immunology, Fujian Provincial Hospital , Fuzhou, China

3. Department of Medicine, Division of Rheumatology and Immunology, Duke University , Durham, North Carolina, USA

4. Department of Medicine and Therapeutics, Prince of Wales Hospital, The Chinese University of Hong Kong , Hong Kong, China

5. Department of Rheumatology, Allergy and Immunology, Tan Tock Seng Hospital , Singapore, Singapore

6. Beijing Kidney Health Technology Co., Ltd , Beijing, China

7. Research Center of Clinical Epidemiology, Peking University Third Hospital , Beijing, China

8. Lee Kong Chian School of Medicine, Nanyang Technological University , Singapore, Singapore

Abstract

Abstract Objectives The efficacy of artificial intelligence (AI)-driven chatbots like ChatGPT4 in specialized medical consultations, particularly in rheumatology, remains underexplored. This study compares the proficiency of ChatGPT4’ responses with practicing rheumatologists to inquiries from patients with SLE. Methods In this cross-sectional study, we curated 95 frequently asked questions (FAQs), including 55 in Chinese and 40 in English. Responses for FAQs from ChatGPT4 and five rheumatologists were scored separately by a panel of rheumatologists and a group of patients with SLE across six domains (scientific validity, logical consistency, comprehensibility, completeness, satisfaction level and empathy) on a 0–10 scale (a score of 0 indicates entirely incorrect responses, while 10 indicates accurate and comprehensive answers). Results Rheumatologists’ scoring revealed that ChatGPT4-generated responses outperformed those from rheumatologists in satisfaction level and empathy, with mean differences of 0.537 (95% CI, 0.252–0.823; P < 0.01) and 0.460 (95% CI, 0.227–0.693; P < 0.01), respectively. From the SLE patients’ perspective, ChatGPT4-generated responses were comparable to the rheumatologist-provided answers in all six domains. Subgroup analysis revealed ChatGPT4 responses were more logically consistent and complete regardless of language and exhibited greater comprehensibility, satisfaction and empathy in Chinese. However, ChatGPT4 responses were inferior in comprehensibility for English FAQs. Conclusion ChatGPT4 demonstrated comparable, possibly better in certain domains, to address FAQs from patients with SLE, when compared with the answers provided by specialists. This study showed the potential of applying ChatGPT4 to improve consultation in SLE patients.

Funder

National Natural Science Foundation of China

NMRC Clinician-Scientist Individual Research

NHG-LKCMedicine Clinician-Scientist Career Scheme

National Center for Advancing Translational Sciences

National Institutes of Health

American Heart Association COVID-19 Fund to Retain Clinical Scientists

Publisher

Oxford University Press (OUP)

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3