Abstract
The validity of a standardised test is questioned if an irrelevant construct is accounted for the performance of examinees, which is wrongly modeled as the ability in the construct (test items). A test must ensure precision in the examinee's ability irrespective of their sub-population in any demographic variables. This paper explored the potentials of gender and school location as major covariates on the West African Examinations Council (WAEC) mathematics items among examinees (N=2,866) using Exploratory Structural Equation Modeling (ESEM). The results remarked that the test is multidimensional (six-factors) with compliance fix indices of (χ2 (940)=4882.024, p < 0.05, CFI=0.962, TLI=0.930, RMSEA=0.038, SRMR=0.030, 90 % CI=0.037-0.039, Akaike information criterion (AIC)=147290.577, Bayesian information criterion (BIC)=149585.436 and Sample-size adjusted BIC=148362.154) respectively. Also, there were 10 (20 %) significant DIF items in the WAEC to gender, while 3 (6 %) of the items indicated significant DIF to school location. Observed DIF items acquaint test developers; the existence of DIF may differentially affect the performance of examinees with the same ability level. The implications of the test are severe for the examinees. Hence, accurate and unbiased assessment should be the basic principles for any test item measurement, and test developers need to test the items to be free from biases psychometrically.
Reference50 articles.
1. Finch, W. H. (2015). Detection of Differential Item Functioning for More Than Two Groups: A Monte Carlo Comparison of Methods. Applied Measurement in Education, 29 (1), 30–45. doi: http://doi.org/10.1080/08957347.2015.1102916
2. Abbott, M. L. (2007). A confirmatory approach to differential item functioning on an ESL reading assessment. Language Testing, 24 (1), 7–36. doi: http://doi.org/10.1177/0265532207071510
3. Hemker, B. T., Sijtsma, K., Molenaar, I. W. (1995). Selection of Unidimensional Scales From a Multidimensional Item Bank in the Polytomous Mokken I RT Model. Applied Psychological Measurement, 19 (4), 337–352. doi: http://doi.org/10.1177/014662169501900404
4. Yu, C. H., Popp, S. O., Digangi, S., Jannasch-Pennell, A. (2007). Assessing unidimensionality: A comparison of Rasch modeling, Parallel analysis, and TETRAD. Practical Assessment, Research and Evaluation, 12 (14), 1–19.
5. Ayanwale, M. A., Isaac-oloniyo, F. O., Abayomi, F. R. (2020). Dimensionality assessment of binary response test items: A non-parametric approach of bayesian item response theory measurement. International Journal of Evaluation and Research in Education (IJERE), 9 (2), 385–392. doi: http://doi.org/10.11591/ijere.v9i2.20454
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献