Show simple item record

dc.contributor.authorSoltani, Amir Arsalan
dc.contributor.authorTenenbaum, Joshua B
dc.date.accessioned2020-08-18T16:00:52Z
dc.date.available2020-08-18T16:00:52Z
dc.date.issued2017-07
dc.identifier.issn1063-6919
dc.identifier.urihttps://hdl.handle.net/1721.1/126644
dc.description.abstractWe study the problem of learning generative models of 3D shapes. Voxels or 3D parts have been widely used as the underlying representations to build complex 3D shapes; however, voxel-based representations suffer from high memory requirements, and parts-based models require a large collection of cached or richly parametrized parts. We take an alternative approach: learning a generative model over multi-view depth maps or their corresponding silhouettes, and using a deterministic rendering function to produce 3D shapes from these images. A multi-view representation of shapes enables generation of 3D models with fine details, as 2D depth maps and silhouettes can be modeled at a much higher resolution than 3D voxels. Moreover, our approach naturally brings the ability to recover the underlying 3D representation from depth maps of one or a few viewpoints. Experiments show that our framework can generate 3D shapes with variations and details. We also demonstrate that our model has out-of-sample generalization power for real-world tasks with occluded objects.en_US
dc.description.sponsorshipUnited States. Office of Naval Research. Multidisciplinary University Research Initiative (Grant N00014-16-1-2007)en_US
dc.description.sponsorshipNational Science Foundation (U.S.). Center for Brain, Minds and Machines (NSF STC Award CCF-1231216)en_US
dc.language.isoen
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.relation.isversionof10.1109/CVPR.2017.269en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceComputer Vision Foundationen_US
dc.titleSynthesizing 3D Shapes via Modeling Multi-view Depth Maps and Silhouettes with Deep Generative Networksen_US
dc.typeArticleen_US
dc.identifier.citationSoltani, Amir Arsalan et al. “Synthesizing 3D Shapes via Modeling Multi-view Depth Maps and Silhouettes with Deep Generative Networks.” Paper presented at the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, 21-26 July 2017, IEEE © 2017 The Author(s)en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Brain and Cognitive Sciencesen_US
dc.relation.journal2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)en_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2019-10-08T14:22:55Z
dspace.date.submission2019-10-08T14:22:58Z
mit.journal.volume2017en_US
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record