Abstract
Testing programs often reuse items due mainly to the difficulty and expense of creating new items. This poses potential problems to test item security if some or all test-takers have knowledge of the items prior to taking the test. Simulated data are used to assess the effect of preknowledge on item response theory true and observed score equating. Root mean square error and bias were used for the recovery of equated scores and linking coefficients for two scaling methods. Results of this study indicated that item preknowledge has a large effect on equated scores and linking coefficients. Furthermore, as the mean ability distribution of the group difference, the number of exposed items, and the number of examinees with item preknowledge increase, the bias and RMSE for equated scores and linking coefficients also increase. Additionally, IRT true score equating results in a larger bias and RMSE than does IRT observed score equating. These findings suggest that item preknowledge has the potential to inflate equated scores, putting the validity of the test scores at risk.
Publisher
Egitimde ve Psikolojide Olcme ve Degerlendirme Dergisi
Subject
Developmental and Educational Psychology,Education
Reference38 articles.
1. American Educational Research Association, American Psychological Association and National Council on Measurement in Education (2014). Standards for educational and psychological testing.
2. Angoff, W. H. (1984). Scales, norms, and equivalent scores. Educational Testing Service.
3. Barri, M. A. (2013). The impact anchor item exposure on mean/sigma linking And IRT true score equating under the neat design [Unpublished master’s thesis]. University of Kansas.
4. Belov, D. I. (2016). Comparing the performance of eight item preknowledge detection statistics. Applied Psychological Measurement, 40(2), 83-97. https://doi.org/10.1177/0146621615603
5. Chen, D. F. (2021). Impact of item parameter drift on IRT linking methods [Unpublished doctoral thesis]. The University of North Carolina.