Affiliation:
1. Department of Cancer Biology, Cancer Institute, University College London, London WC1E 6BT, UK
2. Department of Economics, Reichman University, Herzliya 4610101, Israel
Abstract
The reproducibility of academic research has long been a persistent issue, contradicting one of the fundamental principles of science. Recently, there has been an increasing number of false claims found in academic manuscripts, casting doubt on the validity of reported results. In this paper, we utilize an adapted version of Benford’s law, a statistical phenomenon that describes the distribution of leading digits in naturally occurring datasets, to identify the potential manipulation of results in research manuscripts, solely using the aggregated data presented in those manuscripts rather than the commonly unavailable raw datasets. Our methodology applies the principles of Benford’s law to commonly employed analyses in academic manuscripts, thus reducing the need for the raw data itself. To validate our approach, we employed 100 open-source datasets and successfully predicted 79% of them accurately using our rules. Moreover, we tested the proposed method on known retracted manuscripts, showing that around half (48.6%) can be detected using the proposed method. Additionally, we analyzed 100 manuscripts published in the last two years across ten prominent economic journals, with 10 manuscripts randomly sampled from each journal. Our analysis predicted a 3% occurrence of results manipulation with a 96% confidence level. Our findings show that Benford’s law adapted for aggregated data, can be an initial tool for identifying data manipulation; however, it is not a silver bullet, requiring further investigation for each flagged manuscript due to the relatively low prediction accuracy.
Subject
Information Systems and Management,Computer Science Applications,Information Systems
Reference54 articles.
1. Franzen, M. (2016). Reproducibility, John Wiley & Sons, Ltd.. Chapter 22.
2. Do Pressures to Publish Increase Scientists’ Bias? An Empirical Support from US States Data;Fanelli;PLoS ONE,2010
3. Reproducible and replicable pain research: A critical review;Lee;Pain,2018
4. Outcomes for implementation science: An enhanced systematic review of instruments using evidence-based rating criteria;Lewis;Implement. Sci.,2015
5. Roberts, P., and Priest, H. (2006). Reliability and validity in research. Nurs. Stand., 20.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献