Show simple item record

dc.contributor.authorWu, Jiajun
dc.contributor.authorLu, Erika
dc.contributor.authorKohli, Pushmeet
dc.contributor.authorFreeman, William T
dc.contributor.authorTenenbaum, Joshua B
dc.date.accessioned2021-02-09T21:08:02Z
dc.date.available2021-02-09T21:08:02Z
dc.date.issued2017-12
dc.identifier.urihttps://hdl.handle.net/1721.1/129728
dc.description.abstractWe introduce a paradigm for understanding physical scenes without human annotations. At the core of our system is a physical world representation that is first recovered by a perception module and then utilized by physics and graphics engines. During training, the perception module and the generative models learn by visual de-animation - interpreting and reconstructing the visual information stream. During testing, the system first recovers the physical world state, and then uses the generative models for reasoning and future prediction. Even more so than forward simulation, inverting a physics or graphics engine is a computationally hard problem; we overcome this challenge by using a convolutional inversion network. Our system quickly recognizes the physical world state from appearance and motion cues, and has the flexibility to incorporate both differentiable and non-differentiable physics and graphics engines. We evaluate our system on both synthetic and real datasets involving multiple physical scenes, and demonstrate that our system performs well on both physical state estimation and reasoning problems. We further show that the knowledge learned on the synthetic dataset generalizes to constrained real images.en_US
dc.description.sponsorshipNSF (Grants 1212849,1447476, 1231216)en_US
dc.description.sponsorshipONR MURI (Grant N00014-16-1-2007)en_US
dc.language.isoen
dc.publisherNeural Information Processing Systems Foundation, Incen_US
dc.relation.isversionofhttps://papers.nips.cc/paper/6620-learning-to-see-physics-via-visual-de-animationen_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceNeural Information Processing Systems (NIPS)en_US
dc.titleLearning to see physics via visual de-animationen_US
dc.typeArticleen_US
dc.identifier.citationWu, Jiajun et al. "Learning to see physics via visual de-animation." Advances in Neural Information Processing Systems 30 (NIPS 2017), December 2017, Long Beach, California, Neural Information Processing Systems Foundation, Inc, December 2017. © 2017 Neural Information Processing Systems Foundation, Inc.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratoryen_US
dc.relation.journalAdvances in Neural Information Processing Systems 30 (NIPS 2017)en_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2019-05-28T12:55:56Z
dspace.date.submission2019-05-28T12:55:57Z
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record