Show simple item record

dc.contributor.authorHow, Jonathan P.
dc.contributor.authorMichini, Bernard J.
dc.date.accessioned2013-10-23T16:56:46Z
dc.date.available2013-10-23T16:56:46Z
dc.date.issued2012-05
dc.identifier.isbn978-1-4673-1405-3
dc.identifier.isbn978-1-4673-1403-9
dc.identifier.isbn978-1-4673-1578-4
dc.identifier.isbn978-1-4673-1404-6
dc.identifier.urihttp://hdl.handle.net/1721.1/81489
dc.description.abstractInverse reinforcement learning (IRL) is the task of learning the reward function of a Markov Decision Process (MDP) given knowledge of the transition function and a set of expert demonstrations. While many IRL algorithms exist, Bayesian IRL [1] provides a general and principled method of reward learning by casting the problem in the Bayesian inference framework. However, the algorithm as originally presented suffers from several inefficiencies that prohibit its use for even moderate problem sizes. This paper proposes modifications to the original Bayesian IRL algorithm to improve its efficiency and tractability in situations where the state space is large and the expert demonstrations span only a small portion of it. The key insight is that the inference task should be focused on states that are similar to those encountered by the expert, as opposed to making the naive assumption that the expert demonstrations contain enough information to accurately infer the reward function over the entire state space. A modified algorithm is presented and experimental results show substantially faster convergence while maintaining the solution quality of the original method.en_US
dc.description.sponsorshipUnited States. Office of Naval Research (Science of Autonomy Program Contract N000140910625))en_US
dc.language.isoen_US
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.relation.isversionofhttp://dx.doi.org/10.1109/ICRA.2012.6225241en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alike 3.0en_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/3.0/en_US
dc.sourceMIT web domainen_US
dc.titleImproving the efficiency of Bayesian inverse reinforcement learningen_US
dc.typeArticleen_US
dc.identifier.citationMichini, Bernard, and Jonathan P. How. “Improving the efficiency of Bayesian inverse reinforcement learning.” In 2012 IEEE International Conference on Robotics and Automation, 3651-3656. Institute of Electrical and Electronics Engineers, 2012.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Aerospace Controls Laboratoryen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Aeronautics and Astronauticsen_US
dc.contributor.departmentMassachusetts Institute of Technology. Laboratory for Information and Decision Systemsen_US
dc.contributor.mitauthorMichini, Bernard J.en_US
dc.contributor.mitauthorHow, Jonathan P.en_US
dc.relation.journalProceedings of the 2012 IEEE International Conference on Robotics and Automationen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dspace.orderedauthorsMichini, Bernard; How, Jonathan P.en_US
dc.identifier.orcidhttps://orcid.org/0000-0001-8576-1930
mit.licenseOPEN_ACCESS_POLICYen_US
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record