Author:
El Dehaibi N.,MacDonald E. F.
Abstract
AbstractAn important step when designers use machine learning models is annotating user generated content. In this study we investigate inter-rater reliability measures of qualitative annotations for supervised learning. We work with previously annotated product reviews from Amazon where phrases related to sustainability are highlighted. We measure inter-rater reliability of the annotations using four variations of Krippendorff's U-alpha. Based on the results we propose suggestions to designers on measuring reliability of qualitative annotations for machine learning datasets.
Publisher
Cambridge University Press (CUP)
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Image Similarity Using an Ensemble of Context-Sensitive Models;Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining;2024-08-24
2. Invariance Testing and Feature Selection Using Sparse Linear Layers;Proceedings of the 31st ACM International Conference on Information & Knowledge Management;2022-10-17
3. ML4ML: Automated Invariance Testing for Machine Learning Models;2022 IEEE International Conference On Artificial Intelligence Testing (AITest);2022-08