dc.contributor.author | Soltani, Amir Arsalan | |
dc.contributor.author | Tenenbaum, Joshua B | |
dc.date.accessioned | 2020-08-18T16:00:52Z | |
dc.date.available | 2020-08-18T16:00:52Z | |
dc.date.issued | 2017-07 | |
dc.identifier.issn | 1063-6919 | |
dc.identifier.uri | https://hdl.handle.net/1721.1/126644 | |
dc.description.abstract | We study the problem of learning generative models of 3D shapes. Voxels or 3D parts have been widely used as the underlying representations to build complex 3D shapes; however, voxel-based representations suffer from high memory requirements, and parts-based models require a large collection of cached or richly parametrized parts. We take an alternative approach: learning a generative model over multi-view depth maps or their corresponding silhouettes, and using a deterministic rendering function to produce 3D shapes from these images. A multi-view representation of shapes enables generation of 3D models with fine details, as 2D depth maps and silhouettes can be modeled at a much higher resolution than 3D voxels. Moreover, our approach naturally brings the ability to recover the underlying 3D representation from depth maps of one or a few viewpoints. Experiments show that our framework can generate 3D shapes with variations and details. We also demonstrate that our model has out-of-sample generalization power for real-world tasks with occluded objects. | en_US |
dc.description.sponsorship | United States. Office of Naval Research. Multidisciplinary University Research Initiative (Grant N00014-16-1-2007) | en_US |
dc.description.sponsorship | National Science Foundation (U.S.). Center for Brain, Minds and Machines (NSF STC Award CCF-1231216) | en_US |
dc.language.iso | en | |
dc.publisher | Institute of Electrical and Electronics Engineers (IEEE) | en_US |
dc.relation.isversionof | 10.1109/CVPR.2017.269 | en_US |
dc.rights | Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use. | en_US |
dc.source | Computer Vision Foundation | en_US |
dc.title | Synthesizing 3D Shapes via Modeling Multi-view Depth Maps and Silhouettes with Deep Generative Networks | en_US |
dc.type | Article | en_US |
dc.identifier.citation | Soltani, Amir Arsalan et al. “Synthesizing 3D Shapes via Modeling Multi-view Depth Maps and Silhouettes with Deep Generative Networks.” Paper presented at the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, 21-26 July 2017, IEEE © 2017 The Author(s) | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences | en_US |
dc.relation.journal | 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) | en_US |
dc.eprint.version | Author's final manuscript | en_US |
dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
dc.date.updated | 2019-10-08T14:22:55Z | |
dspace.date.submission | 2019-10-08T14:22:58Z | |
mit.journal.volume | 2017 | en_US |
mit.metadata.status | Complete | |