Show simple item record

dc.contributor.authorZeng, Andy
dc.contributor.authorSong, Shuran
dc.contributor.authorSuo, Daniel
dc.contributor.authorWalker, Ed
dc.contributor.authorXiao, Jianxiong
dc.contributor.authorYu, Kuan-Ting
dc.contributor.authorRodriguez Garcia, Alberto
dc.date.accessioned2019-03-29T19:46:15Z
dc.date.available2019-03-29T19:46:15Z
dc.date.issued2017-07
dc.date.submitted2017-06
dc.identifier.isbn978-1-5090-4633-1
dc.identifier.urihttp://hdl.handle.net/1721.1/121121
dc.description.abstractRobot warehouse automation has attracted significant interest in recent years, perhaps most visibly in the Amazon Picking Challenge (APC) [1]. A fully autonomous warehouse pick-and-place system requires robust vision that reliably recognizes and locates objects amid cluttered environments, self-occlusions, sensor noise, and a large variety of objects. In this paper we present an approach that leverages multiview RGB-D data and self-supervised, data-driven learning to overcome those difficulties. The approach was part of the MIT-Princeton Team system that took 3rd- and 4th-place in the stowing and picking tasks, respectively at APC 2016. In the proposed approach, we segment and label multiple views of a scene with a fully convolutional neural network, and then fit pre-scanned 3D object models to the resulting segmentation to get the 6D object pose. Training a deep neural network for segmentation typically requires a large amount of training data. We propose a self-supervised method to generate a large labeled dataset without tedious manual segmentation. We demonstrate that our system can reliably estimate the 6D pose of objects under a variety of scenarios. All code, data, and benchmarks are available at http://apc.cs.princeton.edu/en_US
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.relation.isversionofhttp://dx.doi.org/10.1109/ICRA.2017.7989165en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourcearXiven_US
dc.titleMulti-view self-supervised deep learning for 6D pose estimation in the Amazon Picking Challengeen_US
dc.typeArticleen_US
dc.identifier.citationZeng, Andy, Kuan-Ting Yu, Shuran Song, Daniel Suo, Ed Walker, Alberto Rodriguez, and Jianxiong Xiao. “Multi-View Self-Supervised Deep Learning for 6D Pose Estimation in the Amazon Picking Challenge.” 2017 IEEE International Conference on Robotics and Automation (ICRA), 29 May - 3 July, 2017, Singapore, Singapore, IEEE, 2017. © 2017 IEEEen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Mechanical Engineeringen_US
dc.contributor.mitauthorYu, Kuan-Ting
dc.contributor.mitauthorRodriguez Garcia, Alberto
dc.relation.journal2017 IEEE International Conference on Robotics and Automation (ICRA)en_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2018-12-17T18:13:34Z
dspace.orderedauthorsZeng, Andy; Yu, Kuan-Ting; Song, Shuran; Suo, Daniel; Walker, Ed; Rodriguez, Alberto; Xiao, Jianxiongen_US
dspace.embargo.termsNen_US
dc.identifier.orcidhttps://orcid.org/0000-0002-8954-2310
dc.identifier.orcidhttps://orcid.org/0000-0002-1119-4512
mit.licenseOPEN_ACCESS_POLICYen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record