Benchmarking Potential Based Rewards for Learning Humanoid Locomotion
Author(s)
Jeon, Se Hwan; Heim, Steve; Khazoom, Charles; Kim, Sangbae
Download2307.10142.pdf (2.504Mb)
Open Access Policy
Open Access Policy
Creative Commons Attribution-Noncommercial-Share Alike
Terms of use
Metadata
Show full item recordAbstract
The main challenge in developing effective reinforcement learning (RL) pipelines is often the design and tuning the reward functions. Well-designed shaping reward can lead to significantly faster learning. Naively formulated rewards, however, can conflict with the desired behavior and result in overfitting or even erratic performance if not properly tuned. In theory, the broad class of potential based reward shaping (PBRS) can help guide the learning process without affecting the optimal policy. Although several studies have explored the use of potential based reward shaping to accelerate learning convergence, most have been limited to grid-worlds and low-dimensional systems, and RL in robotics has predominantly relied on standard forms of reward shaping. In this paper, we benchmark standard forms of shaping with PBRS for a humanoid robot. We find that in this high-dimensional system, PBRS has only marginal benefits in convergence speed. However, the PBRS reward terms are significantly more robust to scaling than typical reward shaping approaches, and thus easier to tune.
Description
2023 IEEE International Conference on Robotics and Automation (ICRA 2023) May 29 - June 2, 2023. London, UK
Date issued
2023-05-29Department
Massachusetts Institute of Technology. Department of Mechanical EngineeringJournal
2023 IEEE International Conference on Robotics and Automation (ICRA)
Publisher
IEEE
Citation
Jeon, Se Hwan, Heim, Steve, Khazoom, Charles and Kim, Sangbae. 2023. "Benchmarking Potential Based Rewards for Learning Humanoid Locomotion." 2023 IEEE International Conference on Robotics and Automation (ICRA).
Version: Author's final manuscript