Show simple item record

dc.contributor.authorLing, Chun Kai
dc.contributor.authorLow, Kian Hsiang
dc.contributor.authorJaillet, Patrick
dc.date.accessioned2017-12-22T14:55:33Z
dc.date.available2017-12-22T14:55:33Z
dc.date.issued2016-02
dc.identifier.urihttp://hdl.handle.net/1721.1/112929
dc.description.abstractThis paper presents a novel nonmyopic adaptive Gaussian process planning (GPP) framework endowed with a general class of Lipschitz continuous reward functions that can unify some active learning/sensing and Bayesian optimization criteria and offer practitioners some flexibility to specify their desired choices for defining new tasks/problems. In particular, it utilizes a principled Bayesian sequential decision problem framework for jointly and naturally optimizing the exploration-exploitation trade-off. In general, the resulting induced GPP policy cannot be derived exactly due to an uncountable set of candidate observations. A key contribution of our work here thus lies in exploiting the Lipschitz continuity of the reward functions to solve for a nonmyopic adaptive ε-optimal GPP (ε-GPP) policy. To plan in real time, we further propose an asymptotically optimal, branch-and-bound anytime variant of ε-GPP with performance guarantee. We empirically demonstrate the effectiveness of our ε-GPP policy and its anytime variant in Bayesian optimization and an energy harvesting task.en_US
dc.description.sponsorshipSingapore-MIT Alliance for Research and Technology (SMART) (52 R-252-000-550-592)en_US
dc.language.isoen_US
dc.publisherAssociation for Computing Machineryen_US
dc.relation.isversionofhttp://dl.acm.org/citation.cfm?id=3016159en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourceMIT Web Domainen_US
dc.titleGaussian Process Planning with Lipschitz Continuous Reward Functionsen_US
dc.typeArticleen_US
dc.identifier.citationLing, Chun Kai, Kian Hsiang Low and Patrick Jaillet. "Gaussian process planning with lipschitz continuous reward functions: towards unifying Bayesian optimization, active learning, and beyond." Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence AAAI'16, 12-17 February, 2016, Phoenix, Arizona, Association for Computing Machinery, 2016.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.contributor.mitauthorJaillet, Patrick
dc.relation.journalProceedings of the Thirtieth AAAI Conference on Artificial Intelligence, AAAI'16en_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dspace.orderedauthorsLing, Chun Kai; Low, Kian Hsiang; Jaillet, Patricken_US
dspace.embargo.termsNen_US
dc.identifier.orcidhttps://orcid.org/0000-0002-8585-6566
mit.licenseOPEN_ACCESS_POLICYen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record