Regret bounds for meta Bayesian optimization with an unknown Gaussian process prior
Author(s)
Kaelbling, Leslie P.; Kim, Beomjoon; Wang, Zi
DownloadPublished version (763.3Kb)
Publisher Policy
Publisher Policy
Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.
Terms of use
Metadata
Show full item recordAbstract
© 2018 Curran Associates Inc.All rights reserved. Bayesian optimization usually assumes that a Bayesian prior is given. However, the strong theoretical guarantees in Bayesian optimization are often regrettably compromised in practice because of unknown parameters in the prior. In this paper, we adopt a variant of empirical Bayes and show that, by estimating the Gaussian process prior from offline data sampled from the same prior and constructing unbiased estimators of the posterior, variants of both GP-UCB and probability of improvement achieve a near-zero regret bound, which decreases to a constant proportional to the observational noise as the number of offline data and the number of online evaluations increase. Empirically, we have verified our approach on challenging simulated robotic problems featuring task and motion planning.
Date issued
2018Department
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence LaboratoryJournal
Advances in Neural Information Processing Systems 31
Publisher
Neural Information Processing Systems
Citation
Kaelbling, Leslie P., Kim, Beomjoon and Wang, Zi. 2018. "Regret bounds for meta Bayesian optimization with an unknown Gaussian process prior."
Version: Final published version