Show simple item record

dc.contributor.authorWu, Jiajun
dc.contributor.authorTenenbaum, Joshua B
dc.contributor.authorTorralba, Antonio
dc.contributor.authorFreeman, William T
dc.date.accessioned2020-04-08T15:54:06Z
dc.date.available2020-04-08T15:54:06Z
dc.date.issued2018-09
dc.date.submitted2017-06
dc.identifier.issn1573-1405
dc.identifier.urihttps://hdl.handle.net/1721.1/124536
dc.description.abstractUnderstanding 3D object structure from a single image is an important but challenging task in computer vision, mostly due to the lack of 3D object annotations to real images. Previous research tackled this problem by either searching for a 3D shape that best explains 2D annotations, or training purely on synthetic data with ground truth 3D information. In this work, we propose 3D INterpreter Networks (3D-INN), an end-to-end trainable framework that sequentially estimates 2D keypoint heatmaps and 3D object skeletons and poses. Our system learns from both 2D-annotated real images and synthetic 3D data. This is made possible mainly by two technical innovations. First, heatmaps of 2D keypoints serve as an intermediate representation to connect real and synthetic data. 3D-INN is trained on real images to estimate 2D keypoint heatmaps from an input image; it then predicts 3D object structure from heatmaps using knowledge learned from synthetic 3D shapes. By doing so, 3D-INN benefits from the variation and abundance of synthetic 3D objects, without suffering from the domain difference between real and synthesized images, often due to imperfect rendering. Second, we propose a Projection Layer, mapping estimated 3D structure back to 2D. During training, it ensures 3D-INN to predict 3D structure whose projection is consistent with the 2D annotations to real images. Experiments show that the proposed system performs well on both 2D keypoint estimation and 3D structure recovery. We also demonstrate that the recovered 3D information has wide vision applications, such as image retrieval. ©2018en_US
dc.description.sponsorshipNSF Robust Intelligence (grant: 1212849)en_US
dc.description.sponsorshipNSF Big Data (grant 1447476)en_US
dc.description.sponsorshipNSF Robust Intelligence (grant 1524817)en_US
dc.description.sponsorshipONR MURI (grant N00014-16-1-2007)en_US
dc.language.isoen
dc.publisherSpringer Natureen_US
dc.relation.isversionof10.1007/S11263-018-1074-6en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourcearXiven_US
dc.title3D Interpreter Networks for Viewer-Centered Wireframe Modelingen_US
dc.typeArticleen_US
dc.identifier.citationWu, Jiajun, et al., "3D Interpreter Networks for Viewer-Centered Wireframe Modeling." International Journal of Computer Vision 126 (2018): p. 1009-26 doi 10.1007/S11263-018-1074-6 ©2018 Author(s)en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Brain and Cognitive Sciencesen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.relation.journalInternational Journal of Computer Visionen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2019-05-28T14:48:45Z
dspace.orderedauthorsWu, Jiajun; Xue, Tianfan; Lim, Joseph J.; Tian, Yuandong; Tenenbaum, Joshua B.; Torralba, Antonio; Freeman, William T.en_US
dspace.date.submission2019-05-28T14:48:47Z
mit.journal.volume126en_US
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record