Scalable reward learning from demonstration
Author(s)
Michini, Bernard J.; How, Jonathan P.; Cutler, Mark Johnson
DownloadHow_Scalable reward.pdf (1.500Mb)
OPEN_ACCESS_POLICY
Open Access Policy
Creative Commons Attribution-Noncommercial-Share Alike
Terms of use
Metadata
Show full item recordAbstract
Reward learning from demonstration is the task of inferring the intents or goals of an agent demonstrating a task. Inverse reinforcement learning methods utilize the Markov decision process (MDP) framework to learn rewards, but typically scale poorly since they rely on the calculation of optimal value functions. Several key modifications are made to a previously developed Bayesian nonparametric inverse reinforcement learning algorithm that avoid calculation of an optimal value function and no longer require discretization of the state or action spaces. Experimental results given demonstrate the ability of the resulting algorithm to scale to larger problems and learn in domains with continuous demonstrations.
Date issued
2013-05Department
Massachusetts Institute of Technology. Aerospace Controls Laboratory; Massachusetts Institute of Technology. Department of Aeronautics and AstronauticsJournal
Proceedings of the 2013 IEEE International Conference on Robotics and Automation
Publisher
Institute of Electrical and Electronics Engineers (IEEE)
Citation
Michini, Bernard, Mark Cutler, and Jonathan P. How. “Scalable Reward Learning from Demonstration.” 2013 IEEE International Conference on Robotics and Automation (May 2013).
Version: Author's final manuscript
ISBN
978-1-4673-5643-5
978-1-4673-5641-1
ISSN
1050-4729