Show simple item record

dc.contributor.advisorJonathan P. How.en_US
dc.contributor.authorKlein, Robert H. (Robert Henry)en_US
dc.contributor.otherMassachusetts Institute of Technology. Department of Aeronautics and Astronautics.en_US
dc.date.accessioned2014-10-08T15:21:50Z
dc.date.available2014-10-08T15:21:50Z
dc.date.copyright2014en_US
dc.date.issued2014en_US
dc.identifier.urihttp://hdl.handle.net/1721.1/90672
dc.descriptionThesis: S.M., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, 2014.en_US
dc.descriptionCataloged from PDF version of thesis.en_US
dc.descriptionIncludes bibliographical references (pages 111-119).en_US
dc.description.abstractAutonomous agents are increasingly being called upon to perform challenging tasks in complex settings with little information about underlying environment dynamics. To successfully complete such tasks the agent must learn from its interactions with the environment. Many existing techniques make assumptions about problem structure to remain tractable, such as limiting the class of possible models or specifying a fixed model expressive power. Complicating matters, there are many scenarios where the environment exhibits multiple underlying sets of dynamics; in these cases, most existing approaches assume the number of underlying models is known a priori, or ignore the possibility of multiple models altogether. Bayesian nonparametric (BNP) methods provide the flexibility to solve both of these problems, but have high inference complexity that has limited their adoption. This thesis provides several methods to tractably plan under uncertainty using BNPs. The first is Simultaneous Clustering on Representation Expansion (SCORE) for learning Markov Decision Processes (MDPs) that exhibit an underlying multiple-model structure. SCORE addresses the co-dependence between observation clustering and model expansion. The second contribution provides a realtime, non-myopic, risk-aware planning solution for use in camera surveillance scenarios where the number of underlying target behaviors and their parameterization are unknown. A BNP model is used to capture target behaviors, and a solution that reduces uncertainty only as needed to perform a mission is presented for allocating cameras. The final contribution is a reinforcement learning (RL) framework RLPy, a software package to promote collaboration and speed innovation in the RL community. RLPy provides a library of learning agents, function approximators, and problem domains for performing RL experiments. RLPy also provides a suite of tools that help automate tasks throughout the experiment pipeline, from initial prototyping through hyperparameter optimization, parallelization of large-scale experiments, and final publication-ready plotting.en_US
dc.description.statementofresponsibilityby Robert H. Klein.en_US
dc.format.extent119 pagesen_US
dc.language.isoengen_US
dc.publisherMassachusetts Institute of Technologyen_US
dc.rightsM.I.T. theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. See provided URL for inquiries about permission.en_US
dc.rights.urihttp://dspace.mit.edu/handle/1721.1/7582en_US
dc.subjectAeronautics and Astronautics.en_US
dc.titlePlanning under uncertainty with Bayesian nonparametric modelsen_US
dc.typeThesisen_US
dc.description.degreeS.M.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Aeronautics and Astronautics
dc.identifier.oclc890463934en_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record