dc.contributor.author | Yao, Shunyu | |
dc.contributor.author | Hsu, Tzu Ming | |
dc.contributor.author | Zhu, Jun-Yan | |
dc.contributor.author | Wu, Jiajun | |
dc.contributor.author | Torralba, Antonio | |
dc.contributor.author | Freeman, William T. | |
dc.contributor.author | Tenenbaum, Joshua B. | |
dc.date.accessioned | 2020-04-07T20:28:53Z | |
dc.date.available | 2020-04-07T20:28:53Z | |
dc.date.issued | 2018 | |
dc.identifier.uri | https://hdl.handle.net/1721.1/124516 | |
dc.description.abstract | We aim to obtain an interpretable, expressive, and disentangled scene representation that contains comprehensive structural and textural information for each object. Previous scene representations learned by neural networks are often uninterpretable, limited to a single object, or lacking 3D knowledge. In this work, we propose 3D scene de-rendering networks (3D-SDN) to address the above issues by integrating disentangled representations for semantics, geometry, and appearance into a deep generative model. Our scene encoder performs inverse graphics, translating a scene into a structured object-wise representation. Our decoder has two components: a differentiable shape renderer and a neural texture generator. The disentanglement of semantics, geometry, and appearance supports 3D-aware scene manipulation, e.g., rotating and moving objects freely while keeping the consistent shape and texture, and changing the object appearance without affecting its shape. Experiments demonstrate that our editing scheme based on 3D-SDN is superior to its 2D counterpart. ©2018 Poster presentation at the 32nd annual Conference on Neural Information Processing Systems (NIPS 2018), December 3-5, 2018, Montréal, Québec. | en_US |
dc.description.sponsorship | NSF (no. 1231216) | en_US |
dc.description.sponsorship | NSF (no. 1447476) | en_US |
dc.description.sponsorship | NSF (no. 1524817) | en_US |
dc.description.sponsorship | ONR MURI (no. N00014-16-1-2007) | en_US |
dc.language.iso | en | |
dc.publisher | Neural Information Processing Systems Foundation, Inc. | en_US |
dc.relation.isversionof | https://papers.nips.cc/paper/7459-3d-aware-scene-manipulation-via-inverse-graphics | en_US |
dc.rights | Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use. | en_US |
dc.source | Neural Information Processing Systems (NIPS) | en_US |
dc.title | 3D-aware scene manipulation via inverse graphics | en_US |
dc.type | Article | en_US |
dc.identifier.citation | Yao, Shunyu, et al., "3D-aware scene manipulation via inverse graphics." Advances in Neural Information Processing Systems 31 (2018) url https://papers.nips.cc/book/advances-in-neural-information-processing-systems-31-2018 | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory | en_US |
dc.relation.journal | Advances in Neural Information Processing Systems | en_US |
dc.eprint.version | Final published version | en_US |
dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
dc.date.updated | 2019-05-28T12:47:10Z | |
dspace.date.submission | 2019-05-28T12:47:11Z | |
mit.journal.volume | 31 | en_US |
mit.metadata.status | Complete | |