Goto

Collaborating Authors

 maximum information gain





Supplementary Material Misspecified GP Bandit Optimization Ilija Bogunovic and Andreas Krause (NeurIPS 2021) A GP bandits: Useful definitions and auxiliary results (Realizable setting)

Neural Information Processing Systems

Such assumptions on the noise variables are frequently used in bandit optimization. Gaussian process with posterior mean and variance that correspond to Eq. (8) and Eq. It also allows us to rewrite Eq. Gaussian Process (supported on D) with the corresponding kernel function. Suppose the learner's hypothesis class is While the first two terms in this bound can be effectively controlled and bounded as in the proof of Theorem 1, the last term, i.e., Such a function can easily be constructed, e.g., via the approach outlined in [36].


Consequences of Kernel Regularity for Bandit Optimization

Lee, Madison, Javidi, Tara

arXiv.org Machine Learning

In this work we investigate the relationship between kernel regularity and algorithmic performance in the bandit optimization of RKHS functions. While reproducing kernel Hilbert space (RKHS) methods traditionally rely on global kernel regressors, it is also common to use a smoothness-based approach that exploits local approximations. We show that these perspectives are deeply connected through the spectral properties of isotropic kernels. In particular, we characterize the Fourier spectra of the Matérn, square-exponential, rational-quadratic, $γ$-exponential, piecewise-polynomial, and Dirichlet kernels, and show that the decay rate determines asymptotic regret from both viewpoints. For kernelized bandit algorithms, spectral decay yields upper bounds on the maximum information gain, governing worst-case regret, while for smoothness-based methods, the same decay rates establish Hölder space embeddings and Besov space norm-equivalences, enabling local continuity analysis. These connections show that kernel-based and locally adaptive algorithms can be analyzed within a unified framework. This allows us to derive explicit regret bounds for each kernel family, obtaining novel results in several cases and providing improved analysis for others. Furthermore, we analyze LP-GP-UCB, an algorithm that combines both approaches, augmenting global Gaussian process surrogates with local polynomial estimators. While the hybrid approach does not uniformly dominate specialized methods, it achieves order-optimality across multiple kernel families.



Supplementary Material Misspecified GP Bandit Optimization Ilija Bogunovic and Andreas Krause (NeurIPS 2021) A GP bandits: Useful definitions and auxiliary results (Realizable setting)

Neural Information Processing Systems

Such assumptions on the noise variables are frequently used in bandit optimization. Gaussian process with posterior mean and variance that correspond to Eq. (8) and Eq. It also allows us to rewrite Eq. Gaussian Process (supported on D) with the corresponding kernel function. Suppose the learner's hypothesis class is While the first two terms in this bound can be effectively controlled and bounded as in the proof of Theorem 1, the last term, i.e., Such a function can easily be constructed, e.g., via the approach outlined in [36].


Graph Neural Network Bandits

Neural Information Processing Systems

The key challenges in this setting are scaling to large domains, and to graphs with many nodes. We resolve these challenges by embedding the permutation invariance into our model.