Abstract
This article considers replicability of the performance of predictors across studies. We suggest a general approach to investigating this issue, based on ensembles of prediction models trained on different studies. We quantify how the common practice of training on a single study accounts in part for the observed challenges in replicability of prediction performance. We also investigate whether ensembles of predictors trained on multiple studies can be combined, using unique criteria, to design robust ensemble learners trained upfront to incorporate replicability into different contexts and populations.
Funder
HHS | NIH | National Cancer Institute
Publisher
Proceedings of the National Academy of Sciences
Reference48 articles.
1. Committee on Applied and Theoretical Statistics, Board on Mathematical Sciences and Their Applications, Division on Engineering and Physical Sciences, National Academies of Sciences, Engineering, and Medicine (2016) Statistical Challenges in Assessing and Fostering the Reproducibility of Scientific Results, Summary of a Workshop, ed Schwalbe M (National Academies Press, Washington, DC).
2. Clarifying the terminology that describes scientific reproducibility;Kenett;Nat Methods,2015
3. Estimating the reproducibility of psychological science
4. Deciding whether follow-up studies have replicated findings in a preliminary large-scale omics study
5. Pitfalls in the Use of DNA Microarray Data for Diagnostic and Prognostic Classification
Cited by
34 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献