Efficient reinforcement learning for robots using informative simulated priors
Author(s)
Cutler, Mark Johnson; How, Jonathan P
DownloadHow_Efficient reinforcement.pdf (679.4Kb)
OPEN_ACCESS_POLICY
Open Access Policy
Creative Commons Attribution-Noncommercial-Share Alike
Terms of use
Metadata
Show full item recordAbstract
Autonomous learning through interaction with the physical world is a promising approach to designing controllers and decision-making policies for robots. Unfortunately, learning on robots is often difficult due to the large number of samples needed for many learning algorithms. Simulators are one way to decrease the samples needed from the robot by incorporating prior knowledge of the dynamics into the learning algorithm. In this paper we present a novel method for transferring data from a simulator to a robot, using simulated data as a prior for real-world learning. A Bayesian nonparametric prior is learned from a potentially black-box simulator. The mean of this function is used as a prior for the Probabilistic Inference for Learning Control (PILCO) algorithm. The simulated prior improves the convergence rate and performance of PILCO by directing the policy search in areas of the state-space that have not yet been observed by the robot. Simulated and hardware results show the benefits of using the prior knowledge in the learning framework.
Date issued
2015-07Department
Massachusetts Institute of Technology. Department of Aeronautics and AstronauticsJournal
2015 IEEE International Conference on Robotics and Automation (ICRA)
Publisher
Institute of Electrical and Electronics Engineers (IEEE)
Citation
Cutler, Mark and How, Jonathan P. “Efficient Reinforcement Learning for Robots Using Informative Simulated Priors.” 2015 IEEE International Conference on Robotics and Automation (ICRA), May 26-30 2015, Seattle, Washington, Institute of Electrical and Electronics Engineers (IEEE), July 2015 © 2015 Institute of Electrical and Electronics Engineers (IEEE)
Version: Author's final manuscript
ISSN
1050-4729