dc.contributor.author | Branavan, Satchuthanan R. | |
dc.contributor.author | Chen, Harr | |
dc.contributor.author | Zettlemoyer, Luke S. | |
dc.contributor.author | Barzilay, Regina | |
dc.date.accessioned | 2010-10-14T12:46:32Z | |
dc.date.available | 2010-10-14T12:46:32Z | |
dc.date.issued | 2009-08 | |
dc.date.submitted | 2009-08 | |
dc.identifier.isbn | 978-1-932432-45-9 | |
dc.identifier.uri | http://hdl.handle.net/1721.1/59313 | |
dc.description.abstract | In this paper, we present a reinforcement learning approach for mapping natural language instructions to sequences of executable actions. We assume access to a reward function that defines the quality of the executed actions. During training, the learner repeatedly constructs action sequences for a set of documents, executes those actions, and observes the resulting reward. We use a policy gradient algorithm to estimate the parameters of a log-linear model for action selection. We apply our method to interpret instructions in two domains --- Windows troubleshooting guides and game tutorials. Our results demonstrate that this method can rival supervised learning techniques while requiring few or no annotated training examples. | en_US |
dc.description.sponsorship | National Science Foundation (U.S.) (grant IIS-0448168) | en_US |
dc.description.sponsorship | National Science Foundation (U.S.) (grant IIS-0835445) | en_US |
dc.description.sponsorship | United States. Office of Naval Research | en_US |
dc.description.sponsorship | National Science Foundation (U.S.) (grant IIS-0835652) | en_US |
dc.language.iso | en_US | |
dc.publisher | Association for Computational Linguistics | en_US |
dc.relation.isversionof | http://portal.acm.org/citation.cfm?id=1687892 | en_US |
dc.rights | Attribution-Noncommercial-Share Alike 3.0 Unported | en_US |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-sa/3.0/ | en_US |
dc.source | MIT web domain | en_US |
dc.subject | algorithms | en_US |
dc.subject | design | en_US |
dc.subject | experimentation | en_US |
dc.subject | languages | en_US |
dc.subject | measurement | en_US |
dc.subject | performance | en_US |
dc.title | Reinforcement Learning for Mapping Instructions to Actions | en_US |
dc.type | Article | en_US |
dc.identifier.citation | Branavan, S.R.K., Harr Chen, Luke S. Zettlemoyer, and Regina Barzilay (2009). "Reinforcement learning for mapping instructions to actions." Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP (Morristown, N.J.: Association for Computational Linguistics): 82-90. © Association for Computing Machinery. | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory | en_US |
dc.contributor.department | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science | en_US |
dc.contributor.approver | Barzilay, Regina | |
dc.contributor.mitauthor | Branavan, Satchuthanan R. | |
dc.contributor.mitauthor | Chen, Harr | |
dc.contributor.mitauthor | Zettlemoyer, Luke S. | |
dc.contributor.mitauthor | Barzilay, Regina | |
dc.relation.journal | Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP | en_US |
dc.eprint.version | Author's final manuscript | |
dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
eprint.status | http://purl.org/eprint/status/PeerReviewed | en_US |
dspace.orderedauthors | Branavan, S. R. K.; Chen, Harr; Zettlemoyer, Luke S.; Barzilay, Regina | |
dc.identifier.orcid | https://orcid.org/0000-0002-2921-8201 | |
mit.license | OPEN_ACCESS_POLICY | en_US |
mit.metadata.status | Complete | |