Diagnostic decisions of specialist optometrists exposed to ambiguous deep-learning outputs

Author:

Carmichael Josie1,Costanza Enrico1,Blandford Ann1,Struyven Robbert2,Keane Pearse2,Balaskas Konstantinos2

Affiliation:

1. UCL Interaction Centre, London, United Kingdom

2. NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust and UCL

Abstract

Abstract Artificial intelligence (AI) has great potential in ophthalmology; however, there has been limited clinical integration. Our study investigated how ambiguous outputs from an AI diagnostic support system (AI-DSS) affected diagnostic responses from optometrists when assessing cases of suspected retinal disease. Thirty optometrists at Moorfields Eye Hospital (15 more experienced, 15 less) assessed 30 clinical cases in counterbalanced order. For ten cases, participants saw an optical coherence tomography (OCT) scan, basic clinical information and a retinal photograph (‘no AI’). For another ten, they were also given the AI-generated OCT-based probabilistic diagnosis (‘AI diagnosis’); and for ten, both AI-diagnosis and an AI-generated OCT segmentation (‘AI diagnosis + segmentation’) were provided. Cases were matched across the three types of presentation and were purposely selected to include 40% ambiguous and 20% incorrect AI outputs. Optometrist diagnostic agreement with the predefined reference standard was lowest for the ‘AI diagnosis + segmentation’ presentation (204/300, 68%) compared to both ‘AI diagnosis’ (224/300, 75% p = 0·010), and ‘no Al’ (242/300, 81%, p = < 0·001). Agreement in the ‘AI diagnosis’ presentation was lower (p = 0·049) than in the ‘no AI’. Agreement with AI diagnosis consistent with the reference standard decreased (174/210 vs 199/210, p = 0·003), but participants trusted the AI more (p = 0·029) when segmentations were displayed. There was no significant effect of practitioner experience on diagnostic responses (p = 0·24). More experienced participants were more confident (p = 0·012) and trusted the AI less (p = 0·038). Our findings also highlighted issues around reference standard definition.

Publisher

Research Square Platform LLC

Reference29 articles.

1. Artificial intelligence in medicine: today and tomorrow;Briganti G;Frontiers in medicine,2020

2. Rodriguez-Ruiz A, Lång K, Gubern-Merida A, Broeders M, Gennaro G, Clauser P, et al. Stand-Alone Artificial Intelligence for Breast Cancer Detection in Mammography: Comparison With 101 Radiologists. JNCI: Journal of the National Cancer Institute. 2019;111(9):916 – 22.

3. Development of a computer-aided detection system for colonoscopy and a publicly accessible large colonoscopy video database (with video);Misawa M;Gastrointestinal Endoscopy,2021

4. Human–computer collaboration for skin cancer recognition;Tschandl P;Nature Medicine,2020

5. Clinically applicable deep learning for diagnosis and referral in retinal disease;Fauw J;Nat Med,2018

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3