Abstract
Objective structured clinical examinations (OSCEs) are a widely used performance assessment for medical and dental students. A common limitation of OSCEs is that the evaluation results depend on the characteristics of raters and a scoring rubric. To overcome this limitation, item response theory (IRT) models such as the many-facet Rasch model have been proposed to estimate examinee abilities while taking into account the characteristics of raters and evaluation items in a rubric. However, conventional IRT models have two impractical assumptions: constant rater severity across all evaluation items in a rubric and an equal interval rating scale among evaluation items, which can decrease model fitting and ability measurement accuracy. To resolve this problem, we propose a new IRT model that introduces two parameters: (1) a rater–item interaction parameter representing the rater severity for each evaluation item and (2) an item-specific step-difficulty parameter representing the difference in rating scales among evaluation items. We demonstrate the effectiveness of the proposed model by applying it to actual data collected from a medical interview test conducted at Tokyo Medical and Dental University as part of a post-clinical clerkship OSCE. The experimental results showed that the proposed model was well-fitted to our OSCE data and measured ability accurately. Furthermore, it provided abundant information on rater and item characteristics that conventional models cannot, helping us to better understand rater and item properties.
Funder
Japan Society for the Promotion of Science
Publisher
Public Library of Science (PLoS)
Reference61 articles.
1. Temporal stability of objective structured clinical exams: a longitudinal study employing item response theory;LA Baig;BMC Med Educ,2012
2. Quality control of an OSCE using generalizability theory and many-faceted Rasch measurement;C Iramaneerat;Adv Health Sci Educ Theory Pract,2008
3. Should candidate scores be adjusted for interviewer stringency or leniency in the multiple mini-interview?;C Roberts;Med Educ,2010
4. Using the Many-Facet Rasch Model to analyse and evaluate the quality of objective structured clinical examination: a non-experimental cross-sectional design;M Tavakol;BMJ Open,2019
5. Using the multiple mini interview as an assessment strategy within the first year of a health professions curriculum;MD Wolcott;BMC Med Educ,2018