Tight Regret Bounds for Bayesian Optimization in One Dimension
Main:8 Pages
3 Figures
Bibliography:1 Pages
Appendix:9 Pages
Abstract
We consider the problem of Bayesian optimization (BO) in one dimension, under a Gaussian process prior and Gaussian sampling noise. We provide a theoretical analysis showing that, under fairly mild technical assumptions on the kernel, the best possible cumulative regret up to time behaves as and . This gives a tight characterization up to a factor, and includes the first non-trivial lower bound for noisy BO. Our assumptions are satisfied, for example, by the squared exponential and Matérn- kernels, with the latter requiring . Our results certify the near-optimality of existing bounds (Srinivas {\em et al.}, 2009) for the SE kernel, while proving them to be strictly suboptimal for the Matérn kernel with .
View on arXivComments on this paper
