Show simple item record

dc.contributor.authorKaelbling, Leslie P.
dc.contributor.authorLozano-Pérez, Tomás
dc.contributor.authorKim, Beomjoon
dc.date.accessioned2021-11-08T16:46:03Z
dc.date.available2021-11-08T16:46:03Z
dc.date.issued2018
dc.identifier.urihttps://hdl.handle.net/1721.1/137707
dc.description.abstractCopyright © 2018, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. In robotics, it is essential to be able to plan efficiently in high-dimensional continuous state-action spaces for long horizons. For such complex planning problems, unguided uniform sampling of actions until a path to a goal is found is hopelessly inefficient, and gradient-based approaches often fall short when the optimization manifold of a given problem is not smooth. In this paper, we present an approach that guides search in continuous spaces for generic planners by learning an action sampler from past search experience. We use a Generative Adversarial Network (GAN) to represent an action sampler, and address an important issue: search experience consists of a relatively large number of actions that are not on a solution path and a relatively small number of actions that actually are on a solution path. We introduce a new technique, based on an importance-ratio estimation method, for using samples from a non-target distribution to make GAN learning more data-efficient. We provide theoretical guarantees and empirical evaluation in three challenging continuous robot planning problems to illustrate the effectiveness of our algorithm.en_US
dc.language.isoen
dc.relation.isversionofhttps://openreview.net/forum?id=bh-hFWQ7zJyJen_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourceMIT web domainen_US
dc.titleGuiding search in continuous state-action spaces by learning an action sampler from off-target search experienceen_US
dc.typeArticleen_US
dc.identifier.citationKaelbling, Leslie P., Lozano-Pérez, Tomás and Kim, Beomjoon. 2018. "Guiding search in continuous state-action spaces by learning an action sampler from off-target search experience."
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2019-06-04T15:28:25Z
dspace.date.submission2019-06-04T15:28:26Z
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record