Exploiting Class Learnability in Noisy Data
Author
Klawonn, Matthew; Heim, Eric; Hendler, James A.Other Contributors
Date Issued
2019Degree
Terms of Use
Attribution-NonCommercial-NoDerivs 3.0 United StatesFull Citation
Klawonn, Matthew & Heim, Eric & Hendler, James. (2019). Exploiting Class Learnability in Noisy Data. Proceedings of the AAAI Conference on Artificial Intelligence. 33. 4082-4089. 10.1609/aaai.v33i01.33014082.Metadata
Show full item recordURI
http://doi.org/10.1609/aaai.v33i01.33014082; https://www.researchgate.net/publication/335573616_Exploiting_Class_Learnability_in_Noisy_Data; https://hdl.handle.net/20.500.13015/6405Abstract
In many domains, collecting sufficient labeled training data for supervised machine learning requires easily accessible but noisy sources, such as crowdsourcing services or tagged Web data. Noisy labels occur frequently in data sets harvested via these means, sometimes resulting in entire classes of data on which learned classifiers generalize poorly. For real world applications, we argue that it can be beneficial to avoid training on such classes entirely. In this work, we aim to explore the classes in a given data set, and guide supervised training to spend time on a class proportional to its learnability. By focusing the training process, we aim to improve model generalization on classes with a strong signal. To that end, we develop an online algorithm that works in conjunction with classifier and training algorithm, iteratively selecting training data for the classifier based on how well it appears to generalize on each class. Testing our approach on a variety of data sets, we show our algorithm learns to focus on classes for which the model has low generalization error relative to strong baselines, yielding a classifier with good performance on learnable classes.;Department
Publisher
AAAIRelationships
Access
Collections
The following license files are associated with this item: