Infinite mixture prototypes for few-shot learning
Author(s)
Allen, Kelsey Rebecca; Shelhamer, Evan; Shin, Hanul; Tenenbaum, Joshua B
DownloadPublished version (348.7Kb)
Publisher Policy
Publisher Policy
Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.
Terms of use
Metadata
Show full item recordAbstract
© 36th International Conference on Machine Learning, ICML 2019. All rights reserved. We propose infinite mixture prototypes to adaptively represent both simple and complex data distributions for few-shot learning. Infinite mixture prototypes combine deep representation learning with Bayesian nonparametrics, representing each class by a set of clusters, unlike existing prototypical methods that represent each class by a single cluster. By inferring the number of clusters, infinite mixture prototypes interpolate between nearest neighbor and prototypical representations in a learned feature space, which improves accuracy and robustness in the few-shot regime. We show the importance of adaptive capacity for capturing complex data distributions such as super-classes (like alphabets in character recognition), with 10-25% absolute accuracy improvements over prototypical networks, while still maintaining or improving accuracy on standard few-shot learning benchmarks. By clustering labeled and unlabeled data with the same rule, infinite mixture prototypes achieve state-of-the-art semi-supcrviscd accuracy, and can perform purely unsupervised clustering, unlike existing fully- and semi-supervised prototypical methods.
Date issued
2019-01Department
Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences; Center for Brains, Minds, and Machines; Massachusetts Institute of Technology. Computer Science and Artificial Intelligence LaboratoryJournal
36th International Conference on Machine Learning, ICML 2019
Citation
Allen, KR, Shelhamer, E, Shin, H and Tenenbaum, JB. 2019. "Infinite mixture prototypes for few-shot learning." 36th International Conference on Machine Learning, ICML 2019, 2019-June.
Version: Final published version