Show simple item record

dc.contributor.authorDoshi-Velez, Finale P.
dc.contributor.authorPfau, David
dc.contributor.authorWood, Frank
dc.contributor.authorRoy, Nicholas
dc.date.accessioned2015-05-19T18:28:33Z
dc.date.available2015-05-19T18:28:33Z
dc.date.issued2015-01
dc.date.submitted2013-05
dc.identifier.issn0162-8828
dc.identifier.issn2160-9292
dc.identifier.urihttp://hdl.handle.net/1721.1/97034
dc.description.abstractMaking intelligent decisions from incomplete information is critical in many applications: for example, robots must choose actions based on imperfect sensors, and speech-based interfaces must infer a user’s needs from noisy microphone inputs. What makes these tasks hard is that often we do not have a natural representation with which to model the domain and use for choosing actions; we must learn about the domain’s properties while simultaneously performing the task. Learning a representation also involves trade-offs between modeling the data that we have seen previously and being able to make predictions about new data. This article explores learning representations of stochastic systems using Bayesian nonparametric statistics. Bayesian nonparametric methods allow the sophistication of a representation to scale gracefully with the complexity in the data. Our main contribution is a careful empirical evaluation of how representations learned using Bayesian nonparametric methods compare to other standard learning approaches, especially in support of planning and control. We show that the Bayesian aspects of the methods result in achieving state-of-the-art performance in decision making with relatively few samples, while the nonparametric aspects often result in fewer computations. These results hold across a variety of different techniques for choosing actions given a representation.en_US
dc.language.isoen_US
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.relation.isversionofhttp://dx.doi.org/10.1109/TPAMI.2013.191en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourceOther univ. web domainen_US
dc.titleBayesian Nonparametric Methods for Partially-Observable Reinforcement Learningen_US
dc.typeArticleen_US
dc.identifier.citationDoshi-Velez, Finale, David Pfau, Frank Wood, and Nicholas Roy. “Bayesian Nonparametric Methods for Partially-Observable Reinforcement Learning.” IEEE Trans. Pattern Anal. Mach. Intell. 37, no. 2 (February 2015): 394–407.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratoryen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Aeronautics and Astronauticsen_US
dc.contributor.mitauthorDoshi-Velez, Finale P.en_US
dc.contributor.mitauthorRoy, Nicholasen_US
dc.relation.journalIEEE Transactions on Pattern Analysis and Machine Intelligenceen_US
dc.eprint.versionOriginal manuscripten_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dspace.orderedauthorsDoshi-Velez, Finale; Pfau, David; Wood, Frank; Roy, Nicholasen_US
dc.identifier.orcidhttps://orcid.org/0000-0002-8293-0492
mit.licenseOPEN_ACCESS_POLICYen_US
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record