Show simple item record

dc.contributor.advisorShah, Devavrat
dc.contributor.authorAlumootil, Varkey
dc.date.accessioned2022-01-14T14:52:30Z
dc.date.available2022-01-14T14:52:30Z
dc.date.issued2021-06
dc.date.submitted2021-06-17T20:12:49.354Z
dc.identifier.urihttps://hdl.handle.net/1721.1/139143
dc.description.abstractPerformance of state-of-the art offline and model-based reinforcement learning (RL) algorithms deteriorates significantly when subjected to severe data scarcity and the presence of heterogeneous agents. In this work, we propose a model-based offline RL method to approach this setting. Using all available data from the various agents, we construct personalized simulators for each individual agent, which are then used to train RL policies. We do so by modeling the transition dynamics of the agents as a low rank tensor decomposition of latent factors associated with agents, states, and actions. We perform experiments on various benchmark environments and demonstrate improvement over existing offline approaches in the scarce data regime.
dc.publisherMassachusetts Institute of Technology
dc.rightsIn Copyright - Educational Use Permitted
dc.rightsCopyright MIT
dc.rights.urihttp://rightsstatements.org/page/InC-EDU/1.0/
dc.titleData-Efficient Offline Reinforcement Learning with Heterogeneous Agents
dc.typeThesis
dc.description.degreeM.Eng.
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
mit.thesis.degreeMaster
thesis.degree.nameMaster of Engineering in Electrical Engineering and Computer Science


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record