Does ChatGPT Answer Otolaryngology Questions Accurately?

Author:

Maksimoski Matthew12ORCID,Noble Anisha Rhea12ORCID,Smith David F.123ORCID

Affiliation:

1. Division of Pediatric Otolaryngology Cincinnati Children's Hospital Medical Center Cincinnati Ohio U.S.A.

2. Department of Otolaryngology ‐ Head and Neck Surgery University of Cincinnati 231 Albert Sabin Way Cincinnati USA

3. Division of Sleep and Circadian Medicine Cincinnati Children's Hospital Medical Center Cincinnati Ohio U.S.A.

Abstract

ObjectiveInvestigate the accuracy of ChatGPT in the manner of medical questions related to otolaryngology.MethodsA ChatGPT session was opened within which 93 questions were asked related to otolaryngology topics. Questions were drawn from all major domains within otolaryngology and based upon key action statements (KAS) from clinical practice guidelines (CPGs). Twenty‐one “patient‐level” questions were also asked of the program. Answers were graded as either “correct,” “partially correct,” “incorrect,” or “non‐answer.”ResultsCorrect answers were given at a rate of 45.5% (71.4% correct in patient‐level, 37.3% CPG); partially correct answers at 31.8% (28.6% patient‐level, 32.8% CPG); incorrect at 21.6% (0% patient‐level, 28.4% CPG); and 1.1% non‐answers (% patient‐level, 1.5% CPG). There was no difference in the rate of correct answers between CPGs published before or after the period of data collection cited by ChatGPT. CPG‐based questions were less likely to be correct than patient‐level questions (p = 0.003).ConclusionPublicly available artificial intelligence software has become increasingly popular with consumers for everything from story‐telling to data collection. In this study, we examined the accuracy of ChatGPT responses to questions related to otolaryngology over 7 domains and 21 published CPGs. Physicians and patients should understand the limitations of this software as it applies to otolaryngology, and programmers in future iterations should consider giving greater weight to information published by well‐established journals and written by national content experts.Level of EvidenceN/A Laryngoscope, 134:4011–4015, 2024

Publisher

Wiley

Reference40 articles.

1. WeitzmanT.GPT‐4 released: what it means for the future of your business.Forbes2023.

2. AiO.ChatGPT Interface: Open AI.2023.chat.openai.com.

3. MehdiY.blogs.microsoft.com: Microsoft.2023.https://blogs.microsoft.com/blog/2023/02/07/reinventing-search-with-a-new-ai-powered-microsoft-bing-and-edge-your-copilot-for-the-web/.

4. ShahS.The writers strike is taking a stand on AI.Time2023.

Cited by 3 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3