Abstract
AbstractThe purpose of this study was to compare the quality of multiple choice questions (MCQs) developed using automated item generation (AIG) versus traditional methods, as judged by a panel of experts. The quality of MCQs developed using two methods (i.e., AIG or traditional) was evaluated by a panel of content experts in a blinded study. Participants rated a total of 102 MCQs using six quality metrics and made a judgment regarding whether or not each item tested recall or application of knowledge. A Wilcoxon two-sample test evaluated differences in each of the six quality metrics rating scales as well as an overall cognitive domain judgment. No significant differences were found in terms of item quality or cognitive domain assessed when comparing the two item development methods. The vast majority of items (> 90%) developed using both methods were deemed to be assessing higher-order skills. When compared to traditionally developed items, MCQs developed using AIG demonstrated comparable quality. Both modalities can produce items that assess higher-order cognitive skills.
Publisher
Springer Science and Business Media LLC
Subject
Management of Technology and Innovation,Media Technology,Education,Social Psychology
Reference23 articles.
1. Albanese, M., & Case, S. (2016). Progress testing: critical analysis and suggested practices. Advances in Health Science Education, 21(1), 221–234.
2. Coderre SP, Harasym P, Mandin H, Fick G. (2004). The impact of two multiple-choice question formats on the problem-solving strategies used by novices and experts. BMC Medical Education, Nov 5, 4, 23.
3. George D, Mallery P. (2010). SPSS for windows step by step: a simple guide and reference 17.0 update. 10th Edition, Pearson, Boston.
4. Gierl, M. J., & Lai, H. (2013). Evaluating the quality of medical multiple-choice items created with automated processes. Medical Education, 47(7), 726–733.
5. Gierl, M. J., Lai, H., Pugh, D., Touchie, C., Boulais, A. P., & De Champlain, A. (2016). Evaluating the psychometric properties of generated test items. Applied Measurement in Education, 29(3), 196–210.
Cited by
25 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献