Author:
Mee Janet,Pandian Ravi,Wolczynski Justin,Morales Amy,Paniagua Miguel,Harik Polina,Baldwin Peter,Clauser Brian E.
Abstract
AbstractRecent advances in automated scoring technology have made it practical to replace multiple-choice questions (MCQs) with short-answer questions (SAQs) in large-scale, high-stakes assessments. However, most previous research comparing these formats has used small examinee samples testing under low-stakes conditions. Additionally, previous studies have not reported on the time required to respond to the two item types. This study compares the difficulty, discrimination, and time requirements for the two formats when examinees responded as part of a large-scale, high-stakes assessment. Seventy-one MCQs were converted to SAQs. These matched items were randomly assigned to examinees completing a high-stakes assessment of internal medicine. No examinee saw the same item in both formats. Items administered in the SAQ format were generally more difficult than items in the MCQ format. The discrimination index for SAQs was modestly higher than that for MCQs and response times were substantially higher for SAQs. These results support the interchangeability of MCQs and SAQs. When it is important that the examinee generate the response rather than selecting it, SAQs may be preferred. The results relating to difficulty and discrimination reported in this paper are consistent with those of previous studies. The results on the relative time requirements for the two formats suggest that with a fixed testing time fewer SAQs can be administered, this limitation more than makes up for the higher discrimination that has been reported for SAQs. We additionally examine the extent to which increased difficulty may directly impact the discrimination of SAQs.
Publisher
Springer Science and Business Media LLC
Subject
Education,General Medicine
Reference21 articles.
1. Baldwin, P. (2021). A problem with the bookmark procedure’s correction for guessing. Educational Measurement, 40, 7–15.
2. Bridgeman, B. (2012). A simple answer to a simple question on answer changing. Journal of Educational Measurement, 49, 467–468.
3. Clauser, B. E., Margolis, M. J., & Swanson, D. B. (2002). An examination of the contribution of the computer-based case simulations to the USMLE Step 3 examination. Academic Medicine (RIME Supplement), 77(10), S80–S82.
4. Heemskerk, L., Norman, G., Chou, S., Mintz, M., Mandin, H., & McLaughlin, K. (2008). The effect of question format and task difficulty on reasoning strategies and diagnostic performance in internal medicine residents. Advances in Health Sciences Education, 13(4), 453–462.
5. Hift, R. J. (2014). Should essays and other “open-ended”-type questions retain a place in written summative assessment in clinical medicine? BMC Medical Education, 14(1), 1–18.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献