Affiliation:
1. Dipartimento di Scienze della Formazione, Università degli Studi Roma Tre
Abstract
Many methods for measuring agreement among raters have been proposed and applied in many domains in the areas of education, psychology, sociology, and medical research. A brief overview of the most used measures of interrater absolute agreements for ordinal rating scales is provided, and a new index is proposed that has several advantages. In particular, the new index allows to evaluate the agreement between raters for each single case (subject or object), and to obtain also a global measure of the interrater agreement for the whole group of cases evaluated. The possibility of having evaluations of the agreement on the single case is particularly useful, for example, in situations where the rating scale is being tested, and it is necessary to identify any changes to it, or to request the raters for a specific comparison on the single case in which the disagreement occurred. The index is not affected by the possible concentration of ratings on a very small number of levels of the ordinal scale.
Publisher
Led Edizioni Universitarie
Subject
Developmental and Educational Psychology,Education,Social Psychology
Reference15 articles.
1. Bove, G., Nuzzo, E., & Serafini, A. (2018). Measurement of interrater agreement for the assessment of language proficiency. In S. Capecchi, F. Di Iorio, & R. Simone (Eds.), ASMOD 2018. Proceedings of the Advanced Statistical Modelling for Ordinal Data Conference (pp. 61-68). Napoli: Università di Napoli Federico II, FedOAPress.
2. Bove, G., Conti, P. L., & Marella, D. (2020). A measure of interrater absolute agreement for ordinal categorical data. Statistical Methods & Applications.
https://doi.org/10.1007/s10260-020-00551-5
3. Cohen, J. (1960). A coefficient of agreement for nominal scales. Educational and Psychological Measurement, 20, 213-220.
https://doi.org/10.1177/001316446002000104
4. Cohen, J. (1968). Weighted kappa: Nominal scale agreement with provision for scaled disagreement or partial credit. Psychological Bulletin, 70, 213-220.
https://doi.org/10.1037/h0026256
5. Conger, A. J. (1980). Integration and generalization of kappas for multiple raters. Psychological Bulletin, 88, 322-328.
https://doi.org/10.1037/0033-2909.88.2.322