Similarity Search for Efficient Active Learning and Search of Rare Concepts | AISC
For slides and more information on the paper, visit https://ai.science/e/similarity-search-for-efficient-active-learning-and-search-of-rare-concepts--2020-11-06
Speaker: Cody Coleman; Host: Omar Nada
Motivation:
Many active learning and search approaches are intractable for industrial settings with billions of unlabeled examples. Existing approaches, such as uncertainty sampling or information density, search globally for the optimal examples to label, scaling linearly or even quadratically with the unlabeled data. However, in practice, data is often heavily skewed; only a small fraction of collected data will be relevant for a given learning task. For example, when identifying rare classes,detecting malicious content, or debugging model performance, the ratio of positive to negative examples can be 1 to 1,000 or more. In this work, we exploit this skew in large training datasets to reduce the number of unlabeled examples considered in each selection round by only looking at the nearest neighbors to the labeled examples.