Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Regret Optimality of GP-UCB (2312.01386v1)

Published 3 Dec 2023 in cs.LG and stat.ML

Abstract: Gaussian Process Upper Confidence Bound (GP-UCB) is one of the most popular methods for optimizing black-box functions with noisy observations, due to its simple structure and superior performance. Its empirical successes lead to a natural, yet unresolved question: Is GP-UCB regret optimal? In this paper, we offer the first generally affirmative answer to this important open question in the Bayesian optimization literature. We establish new upper bounds on both the simple and cumulative regret of GP-UCB when the objective function to optimize admits certain smoothness property. These upper bounds match the known minimax lower bounds (up to logarithmic factors independent of the feasible region's dimensionality) for optimizing functions with the same smoothness. Intriguingly, our findings indicate that, with the same level of exploration, GP-UCB can simultaneously achieve optimality in both simple and cumulative regret. The crux of our analysis hinges on a refined uniform error bound for online estimation of functions in reproducing kernel Hilbert spaces. This error bound, which we derive from empirical process theory, is of independent interest, and its potential applications may reach beyond the scope of this study.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (9)
  1. Belkin M (2018) Approximation beats concentration? An approximation view on inference with smooth radial kernels. Proceedings of the 31st Conference on Learning Theory, 1348–1361.
  2. Brown LD, Low MG (1996) A constrained risk inequality with applications to nonparametric functional estimation. Annals of Statistics 24(6):2524–2535.
  3. Letham B, Bakshy E (2019) Bayesian optimization for policy search via online-offline experimentation. Journal of Machine Learning Research 20(145):1–30.
  4. Muleshkov A, Nguyen T (2016) Easy proof of the Jacobian for the n𝑛nitalic_n-dimensional polar coordinates. Pi Mu Epsilon Journal 14(4):269–273.
  5. Pollard D (1990) Empirical Processes: Theory and Applications, volume 2 of NSF-CBMS Regional Conference Series in Probability and Statistics (Institute of Mathematical Statistics).
  6. Rasmussen CE, Williams CKI (2006) Gaussian Processes for Machine Learning (MIT Press).
  7. Santin G, Schaback R (2016) Approximation of eigenfunctions in kernel-based spaces. Advances in Computational Mathematics 42(4):973–993.
  8. van de Geer SA (2000) Empirical Processes in M-estimation (Cambridge University Press).
  9. Wendland H (2004) Scattered Data Approximation (Cambridge University Press).
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Wenjia Wang (68 papers)
  2. Xiaowei Zhang (56 papers)
  3. Lu Zou (9 papers)

Summary

We haven't generated a summary for this paper yet.