Affiliation:
1. Ubiquitous Knowledge Processing Lab, Department of Computer Science, Technical University of Darmstadt. www.ukp.tu-darmstadt.de
2. School of Informatics,, University of Edinburgh
3. UKP Lab / TU Darmstadt
Abstract
Abstract
Annotated data is an essential ingredient in natural language processing for training and evaluating machine learning models. It is therefore very desirable for the annotations to be of high quality. Recent work, however, has shown that several popular datasets contain a surprising number of annotation errors or inconsistencies. To alleviate this issue, many methods for annotation error detection have been devised over the years. While researchers show that their approaches work well on their newly introduced datasets, they rarely compare their methods to previous work or on the same datasets. This raises strong concerns on methods’ general performance and makes it difficult to assess their strengths and weaknesses. We therefore reimplement 18 methods for detecting potential annotation errors and evaluate them on 9 English datasets for text classification as well as token and span labeling. In addition, we define a uniform evaluation setup including a new formalization of the annotation error detection task, evaluation protocol, and general best practices. To facilitate future research and reproducibility, we release our datasets and implementations in an easy-to-use and open source software package.1
Subject
Artificial Intelligence,Computer Science Applications,Linguistics and Language,Language and Linguistics
Reference100 articles.
1. FLAIR: An easy-to-use framework for state-of-the-art NLP;Akbik,2019
2. TACRED revisited: A thorough evaluation of the TACRED relation extraction task;Alt,2020
3. Error detection for treebank validation;Ambati,2011
4. Spotting spurious data with neural networks;Amiri,2018
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献