Quantifying the impact of AI recommendations with explanations on prescription decision making: an interactive vignette study

Author:

Nagendran Myura1,Festor Paul1,Komorowski Matthieu2,Gordon Anthony1ORCID,Faisal Aldo1ORCID

Affiliation:

1. Imperial College London

2. Imperial College

Abstract

Abstract Background: The challenge of responsibly guiding clinicians to incorporate AI recommendations and explanations into their day-to-day practice has thus far neglected the realm of decisions outside of diagnosis (where there is no gold standard to compare against). We assess how clinicians' decisions may be influenced by additional information more broadly, and how this influence can be modified by either the source of the information (human peers or AI) and the presence or absence of an AI explanation (XAI, here using simple feature importance). Methods: We conducted a human-AI interaction study with ICU doctors using a modified between-subjects design. Doctors were presented on a computer for each of 16 trials with a patient case and prompted to prescribe continuous values for IV fluid and vasopressor. We used a multi-factorial experimental design with four arms, where each clinician experienced all four arms on different subsets of our 24 patients. The four arms were (i) baseline (control), (ii) peer human clinician scenario showing what doses had been prescribed by other doctors, (iii) AI suggestion and (iv) XAI suggestion. Findings: Among 86 ICU doctors we had four key findings. First, additional information (peer, AI or XAI) had a strong influence on prescriptions (significantly for AI, not so for peers) but XAI did not have higher influence than AI alone. Second, inter-clinician prescription variability was affected differentially according to whether the recommendation (whether peer, AI or XAI) was higher or lower than what subjects in the baseline arm did. Third, there was no correlation between attitudes to AI or clinical experience on the AI-supported decisions. Fourth, there was no correlation between what doctors self-reported about how useful they found the XAI and whether the XAI actually influenced their prescriptions. Interpretation: Taken together, our findings on a comparatively large clinical expert population raise important questions for the meaning and design of medical XAI systems. Specifically, we show that the marginal impact of XAI as currently formulated is low in a medical population. We also cast doubt on the utility of self-reports as a valid metric for assessing XAI in clinical experts vs our more objective behavioural response paradigm. Further work in this area could look to higher fidelity and more granular markers that assess the natural behaviour of clinicians when they interact with decision support tools.

Publisher

Research Square Platform LLC

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3