| dc.contributor.author | Rohanimanesh, Khashayar | |
| dc.contributor.author | Roy, Nicholas | |
| dc.contributor.author | Tedrake, Russell Louis | |
| dc.date.accessioned | 2011-06-15T19:54:57Z | |
| dc.date.available | 2011-06-15T19:54:57Z | |
| dc.date.issued | 2009-06 | |
| dc.identifier.uri | http://hdl.handle.net/1721.1/64445 | |
| dc.description | URL to paper listed on conference page | en_US |
| dc.description.abstract | Choosing features for the critic in actor-critic algorithms
with function approximation is known
to be a challenge. Too few critic features can
lead to degeneracy of the actor gradient, and too
many features may lead to slower convergence
of the learner. In this paper, we show that a wellstudied
class of actor policies satisfy the known
requirements for convergence when the actor features
are selected carefully. We demonstrate that
two popular representations for value methods -
the barycentric interpolators and the graph Laplacian
proto-value functions - can be used to represent
the actor in order to satisfy these conditions.
A consequence of this work is a generalization
of the proto-value function methods to the continuous
action actor-critic domain. Finally, we
analyze the performance of this approach using
a simulation of a torque-limited inverted pendulum. | en_US |
| dc.language.iso | en_US | |
| dc.relation.isversionof | http://www-all.cs.umass.edu/~gdk/arl/papers.html | en_US |
| dc.rights | Creative Commons Attribution-Noncommercial-Share Alike 3.0 | en_US |
| dc.rights.uri | http://creativecommons.org/licenses/by-nc-sa/3.0/ | en_US |
| dc.source | MIT web domain | en_US |
| dc.title | Towards feature selection in actor-critic algorithms | en_US |
| dc.type | Article | en_US |
| dc.identifier.citation | Rohanimanesh, Khashayar, Nicholas Roy and Russ Tedrake. "Towards feature selection in actor-critic algorithms." in Proceedings of the ICML/UAI/COLT Workshop on Abstraction in Reinforcement Learning, Montreal, Canada, 2009. | en_US |
| dc.contributor.department | Massachusetts Institute of Technology. Department of Aeronautics and Astronautics | en_US |
| dc.contributor.department | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science | en_US |
| dc.contributor.approver | Tedrake, Russell Louis | |
| dc.contributor.mitauthor | Tedrake, Russell Louis | |
| dc.contributor.mitauthor | Roy, Nicholas | |
| dc.relation.journal | Proceedings of Workshop on Abstraction in Reinforcement Learning, Joint workshop at ICML, UAI, and COLT 2009 | en_US |
| dc.eprint.version | Author's final manuscript | en_US |
| dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
| dspace.orderedauthors | Rohanimanesh, Khashayar; Roy, Nicholas; Tedrake, Russ | |
| dc.identifier.orcid | https://orcid.org/0000-0002-8712-7092 | |
| dc.identifier.orcid | https://orcid.org/0000-0002-8293-0492 | |
| mit.license | OPEN_ACCESS_POLICY | en_US |
| mit.metadata.status | Complete | |