| dc.contributor.advisor | Solar-Lezama, Armando | |
| dc.contributor.author | Li, Zhening | |
| dc.date.accessioned | 2025-09-18T14:28:59Z | |
| dc.date.available | 2025-09-18T14:28:59Z | |
| dc.date.issued | 2025-05 | |
| dc.date.submitted | 2025-06-23T14:02:51.359Z | |
| dc.identifier.uri | https://hdl.handle.net/1721.1/162720 | |
| dc.description.abstract | Skills are temporal abstractions that are intended to improve reinforcement learning (RL) performance through hierarchical RL. Despite our intuition about the properties of an environment that make skills useful, there has been little theoretical work aimed to characterize these properties precisely. This work studies the utility of skills in sparse-reward environments with a discrete state space and finite action space. We show, both theoretically and empirically, that RL performance gains from skills are worse in environments where successful trajectories are less compressible. In environments with a highly incompressible distribution of successful trajectories, using unexpressive skills such as macroactions will provably worsen RL performance. We hope our findings can guide research on automatic skill discovery and help RL practitioners better decide when and how to use skills. | |
| dc.publisher | Massachusetts Institute of Technology | |
| dc.rights | In Copyright - Educational Use Permitted | |
| dc.rights | Copyright retained by author(s) | |
| dc.rights.uri | https://rightsstatements.org/page/InC-EDU/1.0/ | |
| dc.title | The Limits of Temporal Abstractions for Reinforcement Learning with Sparse Rewards | en |
| dc.type | Thesis | |
| dc.description.degree | M.Eng. | |
| dc.contributor.department | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science | |
| atmire.cua.enabled | | |
| mit.thesis.degree | Master | |
| thesis.degree.name | Master of Engineering in Electrical Engineering and Computer Science | |