Show simple item record

dc.contributor.authorDai, Siyu
dc.contributor.authorXu, Wei
dc.contributor.authorHofmann, Andreas
dc.contributor.authorWilliams, Brian
dc.date.accessioned2023-07-11T17:16:04Z
dc.date.available2023-07-11T17:16:04Z
dc.date.issued2023-02-06
dc.identifier.urihttps://hdl.handle.net/1721.1/151079
dc.description.abstractAbstract In order to provide adaptive and user-friendly solutions to robotic manipulation, it is important that the agent can learn to accomplish tasks even if they are only provided with very sparse instruction signals. To address the issues reinforcement learning algorithms face when task rewards are sparse, this paper proposes an intrinsic motivation approach that can be easily integrated into any standard reinforcement learning algorithm and can allow robotic manipulators to learn useful manipulation skills with only sparse extrinsic rewards. Through integrating and balancing empowerment and curiosity, this approach shows superior performance compared to other state-of-the-art intrinsic exploration approaches during extensive empirical testing. When combined with other strategies for tackling the exploration challenge, e.g. curriculum learning, our approach is able to further improve the exploration efficiency and task success rate. Qualitative analysis also shows that when combined with diversity-driven intrinsic motivations, this approach can help manipulators learn a set of diverse skills which could potentially be applied to other more complicated manipulation tasks and accelerate their learning process.en_US
dc.publisherSpringer USen_US
dc.relation.isversionofhttps://doi.org/10.1007/s10514-023-10087-8en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/en_US
dc.sourceSpringer USen_US
dc.titleAn empowerment-based solution to robotic manipulation tasks with sparse rewardsen_US
dc.typeArticleen_US
dc.identifier.citationDai, Siyu, Xu, Wei, Hofmann, Andreas and Williams, Brian. 2023. "An empowerment-based solution to robotic manipulation tasks with sparse rewards."
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
dc.identifier.mitlicensePUBLISHER_CC
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2023-07-02T03:11:20Z
dc.language.rfc3066en
dc.rights.holderThe Author(s)
dspace.embargo.termsN
dspace.date.submission2023-07-02T03:11:19Z
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record