Timezone: »
Poster
Tight Regret Bounds for Bayesian Optimization in One Dimension
Jonathan Scarlett
We consider the problem of Bayesian optimization (BO) in one dimension, under a Gaussian process prior and Gaussian sampling noise. We provide a theoretical analysis showing that, under fairly mild technical assumptions on the kernel, the best possible cumulative regret up to time $T$ behaves as $\Omega(\sqrt{T})$ and $O(\sqrt{T\log T})$. This gives a tight characterization up to a $\sqrt{\log T}$ factor, and includes the first non-trivial lower bound for noisy BO. Our assumptions are satisfied, for example, by the squared exponential and Mat\'ern-$\nu$ kernels, with the latter requiring $\nu > 2$. Our results certify the near-optimality of existing bounds (Srinivas {\em et al.}, 2009) for the SE kernel, while proving them to be strictly suboptimal for the Mat\'ern kernel with $\nu > 2$.
Author Information
Jonathan Scarlett (National University of Singapore)
Related Events (a corresponding poster, oral, or spotlight)
-
2018 Oral: Tight Regret Bounds for Bayesian Optimization in One Dimension »
Thu Jul 12th 12:20 -- 12:30 PM Room A3
More from the Same Authors
-
2020 Poster: Sample Complexity Bounds for 1-bit Compressive Sensing and Binary Stable Embeddings with Generative Priors »
Zhaoqiang Liu · Selwyn Gomes · Avtansh Tiwari · Jonathan Scarlett