Validity evidence supporting clinical skills assessment by artificial intelligence compared with trained clinician raters

Author:

Johnsson Vilma12ORCID,Søndergaard Morten Bo34,Kulasegaram Kulamakan56,Sundberg Karin1,Tiblad Eleonor78,Herling Lotta79,Petersen Olav Bjørn110,Tolsgaard Martin G.1310ORCID

Affiliation:

1. Center for Fetal Medicine, Department of Obstetrics Copenhagen University Hospital, Rigshospitalet Copenhagen Denmark

2. Faculty of Health and Medical Science University of Copenhagen Copenhagen Denmark

3. Copenhagen Academy for Medical Education and Simulation Copenhagen Denmark

4. Department of Computer Science University of Copenhagen Copenhagen Denmark

5. Department of Family and Community Medicine and Scientist Wilson Centre Toronto Ontario Canada

6. Temerty Faculty of Medicine University of Toronto Toronto Ontario Canada

7. Center for Fetal Medicine Karolinska University Hospital Stockholm Sweden

8. Clinical Epidemiology Division, Department of Medicine Solna Karolinska Institutet Stockholm Sweden

9. Department of Clinical Science, Intervention and Technology Karolinska Institutet Stockholm Sweden

10. Department of Clinical Medicine University of Copenhagen Copenhagen Denmark

Abstract

AbstractBackgroundArtificial intelligence (AI) is becoming increasingly used in medical education, but our understanding of the validity of AI‐based assessments (AIBA) as compared with traditional clinical expert‐based assessments (EBA) is limited. In this study, the authors aimed to compare and contrast the validity evidence for the assessment of a complex clinical skill based on scores generated from an AI and trained clinical experts, respectively.MethodsThe study was conducted between September 2020 to October 2022. The authors used Kane's validity framework to prioritise and organise their evidence according to the four inferences: scoring, generalisation, extrapolation and implications. The context of the study was chorionic villus sampling performed within the simulated setting. AIBA and EBA were used to evaluate performances of experts, intermediates and novice based on video recordings. The clinical experts used a scoring instrument developed in a previous international consensus study. The AI used convolutional neural networks for capturing features on video recordings, motion tracking and eye movements to arrive at a final composite score.ResultsA total of 45 individuals participated in the study (22 novices, 12 intermediates and 11 experts). The authors demonstrated validity evidence for scoring, generalisation, extrapolation and implications for both EBA and AIBA. The plausibility of assumptions related to scoring, evidence of reproducibility and relation to different training levels was examined. Issues relating to construct underrepresentation, lack of explainability, and threats to robustness were identified as potential weak links in the AIBA validity argument compared with the EBA validity argument.ConclusionThere were weak links in the use of AIBA compared with EBA, mainly in their representation of the underlying construct but also regarding their explainability and ability to transfer to other datasets. However, combining AI and clinical expert‐based assessments may offer complementary benefits, which is a promising subject for future research.

Funder

Novo Nordisk Fonden

Publisher

Wiley

Subject

Education,General Medicine

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3