42
0

Improved Regret Bounds for Gaussian Process Upper Confidence Bound in Bayesian Optimization

Main:10 Pages
1 Figures
Bibliography:1 Pages
Appendix:26 Pages
Abstract

This paper addresses the Bayesian optimization problem (also referred to as the Bayesian setting of the Gaussian process bandit), where the learner seeks to minimize the regret under a function drawn from a known Gaussian process (GP). Under a Matérn kernel with a certain degree of smoothness, we show that the Gaussian process upper confidence bound (GP-UCB) algorithm achieves O~(T)\tilde{O}(\sqrt{T}) cumulative regret with high probability. Furthermore, our analysis yields O(Tln4T)O(\sqrt{T \ln^4 T}) regret under a squared exponential kernel. These results fill the gap between the existing regret upper bound for GP-UCB and the best-known bound provided by Scarlett (2018). The key idea in our proof is to capture the concentration behavior of the input sequence realized by GP-UCB, enabling a more refined analysis of the GP's information gain.

View on arXiv
@article{iwazaki2025_2506.01393,
  title={ Improved Regret Bounds for Gaussian Process Upper Confidence Bound in Bayesian Optimization },
  author={ Shogo Iwazaki },
  journal={arXiv preprint arXiv:2506.01393},
  year={ 2025 }
}
Comments on this paper