Affiliation:
1. Department of Economics, Virginia Tech, Blacksburg, VA 24061, USA
Abstract
Although large data sets are generally viewed as advantageous for their ability to provide more precise and reliable evidence, it is often overlooked that these benefits are contingent upon certain conditions being met. The primary condition is the approximate validity (statistical adequacy) of the probabilistic assumptions comprising the statistical model Mθ(x) applied to the data. In the case of a statistically adequate Mθ(x) and a given significance level α, as n increases, the power of a test increases, and the p-value decreases due to the inherent trade-off between type I and type II error probabilities in frequentist testing. This trade-off raises concerns about the reliability of declaring ‘statistical significance’ based on conventional significance levels when n is exceptionally large. To address this issue, the author proposes that a principled approach, in the form of post-data severity (SEV) evaluation, be employed. The SEV evaluation represents a post-data error probability that converts unduly data-specific ‘accept/reject H0 results’ into evidence either supporting or contradicting inferential claims regarding the parameters of interest. This approach offers a more nuanced and robust perspective in navigating the challenges posed by the large n problem.
Subject
Statistics and Probability
Reference41 articles.
1. Some difficulties of interpretation encountered in the application of the chi-square test;Berkson;J. Am. Stat.,1938
2. Fisher, R.A. (1935). The Design of Experiments, Oliver and Boyd.
3. Tests of significance considered as evidence;Berkson;J. Am. Assoc.,1942
4. Fisher, R.A. (1925). Statistical Methods for Research Workers, Oliver and Boyd.
5. Note on Dr. Berkson’s criticism of tests of significance;Fisher;J. Am. Stat. Assoc.,1943
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献