Show simple item record

dc.contributor.authorLu, Ziqi
dc.contributor.authorZhang, Yihao
dc.contributor.authorDoherty, Kevin
dc.contributor.authorSeverinsen, Odin
dc.contributor.authorYang, Ethan
dc.contributor.authorLeonard, John
dc.date.accessioned2024-03-13T19:43:57Z
dc.date.available2024-03-13T19:43:57Z
dc.date.issued2022-10-23
dc.identifier.urihttps://hdl.handle.net/1721.1/153750
dc.description2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) October 23-27, 2022, Kyoto, Japanen_US
dc.description.abstractRecent progress in object pose prediction provides a promising path for robots to build object-level scene representations during navigation. However, as we deploy a robot in novel environments, the out-of-distribution data can degrade the prediction performance. To mitigate the domain gap, we can potentially perform self-training in the target domain, using predictions on robot-captured images as pseudo labels to fine-tune the object pose estimator. Unfortunately, the pose predictions are typically outlier-corrupted, and it is hard to quantify their uncertainties, which can result in low-quality pseudo-labeled data. To address the problem, we propose a SLAM-supported self-training method, leveraging robot understanding of the 3D scene geometry to enhance the object pose inference performance. Combining the pose predictions with robot odometry, we formulate and solve pose graph optimization to refine the object pose estimates and make pseudo labels more consistent across frames. We incorporate the pose prediction covariances as variables into the optimization to automatically model their uncertainties. This automatic covariance tuning (ACT) process can fit 6D pose prediction noise at the component level, leading to higher-quality pseudo training data. We test our method with the deep object pose estimator (DOPE) on the YCB video dataset and in real robot experiments. It achieves respectively 34.3% and 17.8% accuracy enhancements in pose prediction on the two tests.en_US
dc.language.isoen
dc.publisherIEEEen_US
dc.relation.isversionof10.1109/iros47612.2022.9981145en_US
dc.rightsCreative Commons Attribution-Noncommercial-ShareAlikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourcearxiven_US
dc.titleSLAM-Supported Self-Training for 6D Object Pose Estimationen_US
dc.typeArticleen_US
dc.identifier.citationLu, Ziqi, Zhang, Yihao, Doherty, Kevin, Severinsen, Odin, Yang, Ethan et al. 2022. "SLAM-Supported Self-Training for 6D Object Pose Estimation."
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2024-03-13T19:18:03Z
dspace.orderedauthorsLu, Z; Zhang, Y; Doherty, K; Severinsen, O; Yang, E; Leonard, Jen_US
dspace.date.submission2024-03-13T19:18:05Z
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record