Affiliation:
1. University of Pennsylvania
Abstract
High-quality labels are expensive to obtain for many machine learning tasks, such as medical image classification tasks. Therefore, probabilistic (weak) labels produced by weak supervision tools are used to seed a process in which influential samples with weak labels are identified and cleaned by several human annotators to improve the model performance. To lower the overall cost and computational overhead of this process, we propose a solution called CHEF (CHEap and Fast label cleaning), which consists of the following three components. First, to reduce the cost of human annotators, we use INFL, which prioritizes the
most influential
training samples for cleaning and provides cleaned labels to save the cost of one human annotator. Second, to accelerate the sample selector phase and the model constructor phase, we use Increm-INFL to
incrementally
produce influential samples, and DeltaGrad-L to
incrementally
update the model. Third, we redesign the typical label cleaning pipeline so that human annotators iteratively clean smaller batch of samples rather than one big batch of samples. This yields better overall model performance and enables possible early termination when the expected model performance has been achieved. Extensive experiments show that our approach gives good model prediction performance while achieving significant speed-ups.
Subject
General Earth and Planetary Sciences,Water Science and Technology,Geography, Planning and Development
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Data cleaning and machine learning: a systematic literature review;Automated Software Engineering;2024-06-11
2. MisDetect: Iterative Mislabel Detection using Early Loss;Proceedings of the VLDB Endowment;2024-02
3. Label Engineering Methods for ML Systems;Lecture Notes in Networks and Systems;2024
4. How I stopped worrying about training data bugs and started complaining;Proceedings of the Sixth Workshop on Data Management for End-To-End Machine Learning;2022-06-12