Generating the Future with Adversarial Transformers
Author(s)
Vondrick, Carl Martin; Torralba, Antonio
DownloadAccepted version (4.740Mb)
Terms of use
Metadata
Show full item recordAbstract
We learn models to generate the immediate future in video. This problem has two main challenges. Firstly, since the future is uncertain, models should be multi-modal, which can be difficult to learn. Secondly, since the future is similar to the past, models store low-level details, which complicates learning of high-level semantics. We propose a framework to tackle both of these challenges. We present a model that generates the future by transforming pixels in the past. Our approach explicitly disentangles the model's memory from the prediction, which helps the model learn desirable invariances. Experiments suggest that this model can generate short videos of plausible futures. We believe predictive models have many applications in robotics, health-care, and video understanding. Keywords: predictive models; generators; visualization; network architecture; spatial resolution; semantics; robots
Date issued
2017-11-09Department
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer ScienceJournal
2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
Publisher
IEEE
Citation
Vondrick, Carl, and Antonio Torralba. "Generating the Future with Adversarial Transformers." 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017, Honolulu, Hawaii, USA, IEEE, 2017
Version: Author's final manuscript
ISBN
9781538604571
9781538604588
ISSN
1063-6919