Narrowing the gap: expected versus deployment performance

Author:

Zhou Alice X12ORCID,Aczon Melissa D12ORCID,Laksana Eugene12,Ledbetter David R3,Wetzel Randall C124

Affiliation:

1. Department of Anesthesiology and Critical Care Medicine, Children’s Hospital Los Angeles , Los Angeles, California, USA

2. Laura P. and Leland K. Whittier Virtual Pediatric Intensive Care Unit, Children’s Hospital Los Angeles , Los Angeles, California, USA

3. Advanced Analytics for Healthcare, KPMG International Limited , Dallas, Texas, USA

4. Department of Pediatrics and Anesthesiology, University of Southern California Keck School of Medicine , Los Angeles, California, USA

Abstract

Abstract Objectives Successful model development requires both an accurate a priori understanding of future performance and high performance on deployment. Optimistic estimations of model performance that are unrealized in real-world clinical settings can contribute to nonuse of predictive models. This study used 2 tasks, predicting ICU mortality and Bi-Level Positive Airway Pressure failure, to quantify: (1) how well internal test performances derived from different methods of partitioning data into development and test sets estimate future deployment performance of Recurrent Neural Network models and (2) the effects of including older data in the training set on models’ performance. Materials and Methods The cohort consisted of patients admitted between 2010 and 2020 to the Pediatric Intensive Care Unit of a large quaternary children’s hospital. 2010–2018 data were partitioned into different development and test sets to measure internal test performance. Deployable models were trained on 2010–2018 data and assessed on 2019–2020 data, which was conceptualized to represent a real-world deployment scenario. Optimism, defined as the overestimation of the deployed performance by internal test performance, was measured. Performances of deployable models were also compared with each other to quantify the effect of including older data during training. Results, Discussion, and Conclusion Longitudinal partitioning methods, where models are tested on newer data than the development set, yielded the least optimism. Including older years in the training dataset did not degrade deployable model performance. Using all available data for model development fully leveraged longitudinal partitioning by measuring year-to-year performance.

Funder

L.K. Whittier Foundation

Publisher

Oxford University Press (OUP)

Subject

Health Informatics

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Perspectives on implementing models for decision support in clinical care;Journal of the American Medical Informatics Association;2023-08-18

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3