[2602.17940] Tighter Regret Lower Bound for Gaussian Process Bandits with Squared Exponential Kernel in Hypersphere

[2602.17940] Tighter Regret Lower Bound for Gaussian Process Bandits with Squared Exponential Kernel in Hypersphere

arXiv - Machine Learning 4 min read Article

Summary

This paper presents a tighter lower bound on cumulative regret for Gaussian process bandits using a squared exponential kernel in a hyperspherical domain, addressing an open question in the field.

Why It Matters

Understanding the regret bounds in Gaussian process bandits is crucial for developing efficient algorithms in machine learning. This research contributes to the theoretical foundation of bandit problems, which are widely applicable in optimization, reinforcement learning, and decision-making scenarios.

Key Takeaways

  • Establishes a worst-case lower bound for Gaussian process bandits.
  • Focuses on the squared exponential kernel in a hyperspherical domain.
  • Addresses the gap in dimension-dependent logarithmic factors between upper and lower bounds.
  • Provides improved upper bounds on maximum information gain for the SE kernel.
  • Confirms the optimality of existing algorithms up to dimension-independent factors.

Computer Science > Machine Learning arXiv:2602.17940 (cs) [Submitted on 20 Feb 2026] Title:Tighter Regret Lower Bound for Gaussian Process Bandits with Squared Exponential Kernel in Hypersphere Authors:Shogo Iwazaki View a PDF of the paper titled Tighter Regret Lower Bound for Gaussian Process Bandits with Squared Exponential Kernel in Hypersphere, by Shogo Iwazaki View PDF HTML (experimental) Abstract:We study an algorithm-independent, worst-case lower bound for the Gaussian process (GP) bandit problem in the frequentist setting, where the reward function is fixed and has a bounded norm in the known reproducing kernel Hilbert space (RKHS). Specifically, we focus on the squared exponential (SE) kernel, one of the most widely used kernel functions in GP bandits. One of the remaining open questions for this problem is the gap in the \emph{dimension-dependent} logarithmic factors between upper and lower bounds. This paper partially resolves this open question under a hyperspherical input domain. We show that any algorithm suffers $\Omega(\sqrt{T (\ln T)^{d} (\ln \ln T)^{-d}})$ cumulative regret, where $T$ and $d$ represent the total number of steps and the dimension of the hyperspherical domain, respectively. Regarding the simple regret, we show that any algorithm requires $\Omega(\epsilon^{-2}(\ln \frac{1}{\epsilon})^d (\ln \ln \frac{1}{\epsilon})^{-d})$ time steps to find an $\epsilon$-optimal point. We also provide the improved $O((\ln T)^{d+1}(\ln \ln T)^{-d})$ upper boun...

Related Articles

Llms

94.42% on BANKING77 Official Test Split — New Strong 2nd Place with Lightweight Embedding + Rerank (no 7B LLM)

94.42% Accuracy on Banking77 Official Test Split BANKING77-77 is deceptively hard: 77 fine-grained banking intents, noisy real-world quer...

Reddit - Artificial Intelligence · 1 min ·
Llms

[D] Tested model routing on financial AI datasets — good savings and curious what benchmarks others use.

Ran a benchmark evaluating whether prompt complexity-based routing delivers meaningful savings. Used public HuggingFace datasets. Here's ...

Reddit - Machine Learning · 1 min ·
Llms

[D] AI research on small language models

i'm doing research on some trending fields in AI, currently working on small language models and would love to meet people who are workin...

Reddit - Machine Learning · 1 min ·
Llms

One of The Worst AI's I've Ever Seen

I'm using Gemini just for they gave us a student-free-pro pack. It can't see the images I sent, most of the time it just rewrites the mes...

Reddit - Artificial Intelligence · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime