Gaussian Process Planning with Lipschitz Continuous Reward Functions
Author(s)
Ling, Chun Kai; Low, Kian Hsiang; Jaillet, Patrick
DownloadJaillet_Gaussian process planning.pdf (512.8Kb)
OPEN_ACCESS_POLICY
Open Access Policy
Creative Commons Attribution-Noncommercial-Share Alike
Terms of use
Metadata
Show full item recordAbstract
This paper presents a novel nonmyopic adaptive Gaussian process planning (GPP) framework endowed with a general class of Lipschitz continuous reward functions that can unify some active learning/sensing and Bayesian optimization criteria and offer practitioners some flexibility to specify their desired choices for defining new tasks/problems. In particular, it utilizes a principled Bayesian sequential decision problem framework for jointly and naturally optimizing the exploration-exploitation trade-off. In general, the resulting induced GPP policy cannot be derived exactly due to an uncountable set of candidate observations. A key contribution of our work here thus lies in exploiting the Lipschitz continuity of the reward functions to solve for a nonmyopic adaptive ε-optimal GPP (ε-GPP) policy. To plan in real time, we further propose an asymptotically optimal, branch-and-bound anytime variant of ε-GPP with performance guarantee. We empirically demonstrate the effectiveness of our ε-GPP policy and its anytime variant in Bayesian optimization and an energy harvesting task.
Date issued
2016-02Department
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer ScienceJournal
Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, AAAI'16
Publisher
Association for Computing Machinery
Citation
Ling, Chun Kai, Kian Hsiang Low and Patrick Jaillet. "Gaussian process planning with lipschitz continuous reward functions: towards unifying Bayesian optimization, active learning, and beyond." Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence AAAI'16, 12-17 February, 2016, Phoenix, Arizona, Association for Computing Machinery, 2016.
Version: Author's final manuscript