Show simple item record

dc.contributor.authorRohanimanesh, Khashayar
dc.contributor.authorRoy, Nicholas
dc.contributor.authorTedrake, Russell Louis
dc.date.accessioned2011-06-15T19:54:57Z
dc.date.available2011-06-15T19:54:57Z
dc.date.issued2009-06
dc.identifier.urihttp://hdl.handle.net/1721.1/64445
dc.descriptionURL to paper listed on conference pageen_US
dc.description.abstractChoosing features for the critic in actor-critic algorithms with function approximation is known to be a challenge. Too few critic features can lead to degeneracy of the actor gradient, and too many features may lead to slower convergence of the learner. In this paper, we show that a wellstudied class of actor policies satisfy the known requirements for convergence when the actor features are selected carefully. We demonstrate that two popular representations for value methods - the barycentric interpolators and the graph Laplacian proto-value functions - can be used to represent the actor in order to satisfy these conditions. A consequence of this work is a generalization of the proto-value function methods to the continuous action actor-critic domain. Finally, we analyze the performance of this approach using a simulation of a torque-limited inverted pendulum.en_US
dc.language.isoen_US
dc.relation.isversionofhttp://www-all.cs.umass.edu/~gdk/arl/papers.htmlen_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alike 3.0en_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/3.0/en_US
dc.sourceMIT web domainen_US
dc.titleTowards feature selection in actor-critic algorithmsen_US
dc.typeArticleen_US
dc.identifier.citationRohanimanesh, Khashayar, Nicholas Roy and Russ Tedrake. "Towards feature selection in actor-critic algorithms." in Proceedings of the ICML/UAI/COLT Workshop on Abstraction in Reinforcement Learning, Montreal, Canada, 2009.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Aeronautics and Astronauticsen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.contributor.approverTedrake, Russell Louis
dc.contributor.mitauthorTedrake, Russell Louis
dc.contributor.mitauthorRoy, Nicholas
dc.relation.journalProceedings of Workshop on Abstraction in Reinforcement Learning, Joint workshop at ICML, UAI, and COLT 2009en_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
dspace.orderedauthorsRohanimanesh, Khashayar; Roy, Nicholas; Tedrake, Russ
dc.identifier.orcidhttps://orcid.org/0000-0002-8712-7092
dc.identifier.orcidhttps://orcid.org/0000-0002-8293-0492
mit.licenseOPEN_ACCESS_POLICYen_US
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record