Meta-Sim: Learning to Generate Synthetic Datasets
Author(s)
Kar, Amlan; Prakash, Aayush; Liu, Ming-Yu; Cameracci, Eric; Yuan, Justin; Rusiniak, Matt; Acuna, David; Torralba, Antonio; Fidler, Sanja; ... Show more Show less
DownloadSubmitted version (8.543Mb)
Open Access Policy
Open Access Policy
Creative Commons Attribution-Noncommercial-Share Alike
Terms of use
Metadata
Show full item recordAbstract
© 2019 IEEE. Training models to high-end performance requires availability of large labeled datasets, which are expensive to get. The goal of our work is to automatically synthesize labeled datasets that are relevant for a downstream task. We propose Meta-Sim, which learns a generative model of synthetic scenes, and obtain images as well as its corresponding ground-truth via a graphics engine. We parametrize our dataset generator with a neural network, which learns to modify attributes of scene graphs obtained from probabilistic scene grammars, so as to minimize the distribution gap between its rendered outputs and target data. If the real dataset comes with a small labeled validation set, we additionally aim to optimize a meta-objective, i.e. downstream task performance. Experiments show that the proposed method can greatly improve content generation quality over a human-engineered probabilistic scene grammar, both qualitatively and quantitatively as measured by performance on a downstream task.
Date issued
2020-02Department
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science; Massachusetts Institute of Technology. Computer Science and Artificial Intelligence LaboratoryJournal
2019 IEEE/CVF International Conference on Computer Vision (ICCV)
Publisher
Institute of Electrical and Electronics Engineers (IEEE)
Citation
2020. "Meta-Sim: Learning to Generate Synthetic Datasets." Proceedings of the IEEE International Conference on Computer Vision, 2019-October.
Version: Original manuscript
ISSN
978-1-7281-4803-8
2380-7504