Show simple item record

dc.contributor.advisorSolar-Lezama, Armando
dc.contributor.authorLi, Zhening
dc.date.accessioned2025-09-18T14:28:59Z
dc.date.available2025-09-18T14:28:59Z
dc.date.issued2025-05
dc.date.submitted2025-06-23T14:02:51.359Z
dc.identifier.urihttps://hdl.handle.net/1721.1/162720
dc.description.abstractSkills are temporal abstractions that are intended to improve reinforcement learning (RL) performance through hierarchical RL. Despite our intuition about the properties of an environment that make skills useful, there has been little theoretical work aimed to characterize these properties precisely. This work studies the utility of skills in sparse-reward environments with a discrete state space and finite action space. We show, both theoretically and empirically, that RL performance gains from skills are worse in environments where successful trajectories are less compressible. In environments with a highly incompressible distribution of successful trajectories, using unexpressive skills such as macroactions will provably worsen RL performance. We hope our findings can guide research on automatic skill discovery and help RL practitioners better decide when and how to use skills.
dc.publisherMassachusetts Institute of Technology
dc.rightsIn Copyright - Educational Use Permitted
dc.rightsCopyright retained by author(s)
dc.rights.urihttps://rightsstatements.org/page/InC-EDU/1.0/
dc.titleThe Limits of Temporal Abstractions for Reinforcement Learning with Sparse Rewardsen
dc.typeThesis
dc.description.degreeM.Eng.
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
atmire.cua.enabled
mit.thesis.degreeMaster
thesis.degree.nameMaster of Engineering in Electrical Engineering and Computer Science


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record