maximum information gain
- North America > United States (0.27)
- Europe > Switzerland > Zürich > Zürich (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)
- Information Technology > Data Science > Data Mining (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.70)
- North America > Canada > Alberta (0.14)
- Asia > Middle East > Jordan (0.04)
- Europe > Netherlands > South Holland > Delft (0.04)
- (4 more...)
Supplementary Material Misspecified GP Bandit Optimization Ilija Bogunovic and Andreas Krause (NeurIPS 2021) A GP bandits: Useful definitions and auxiliary results (Realizable setting)
Such assumptions on the noise variables are frequently used in bandit optimization. Gaussian process with posterior mean and variance that correspond to Eq. (8) and Eq. It also allows us to rewrite Eq. Gaussian Process (supported on D) with the corresponding kernel function. Suppose the learner's hypothesis class is While the first two terms in this bound can be effectively controlled and bounded as in the proof of Theorem 1, the last term, i.e., Such a function can easily be constructed, e.g., via the approach outlined in [36].
Consequences of Kernel Regularity for Bandit Optimization
In this work we investigate the relationship between kernel regularity and algorithmic performance in the bandit optimization of RKHS functions. While reproducing kernel Hilbert space (RKHS) methods traditionally rely on global kernel regressors, it is also common to use a smoothness-based approach that exploits local approximations. We show that these perspectives are deeply connected through the spectral properties of isotropic kernels. In particular, we characterize the Fourier spectra of the Matérn, square-exponential, rational-quadratic, $γ$-exponential, piecewise-polynomial, and Dirichlet kernels, and show that the decay rate determines asymptotic regret from both viewpoints. For kernelized bandit algorithms, spectral decay yields upper bounds on the maximum information gain, governing worst-case regret, while for smoothness-based methods, the same decay rates establish Hölder space embeddings and Besov space norm-equivalences, enabling local continuity analysis. These connections show that kernel-based and locally adaptive algorithms can be analyzed within a unified framework. This allows us to derive explicit regret bounds for each kernel family, obtaining novel results in several cases and providing improved analysis for others. Furthermore, we analyze LP-GP-UCB, an algorithm that combines both approaches, augmenting global Gaussian process surrogates with local polynomial estimators. While the hybrid approach does not uniformly dominate specialized methods, it achieves order-optimality across multiple kernel families.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Europe > Finland > Uusimaa > Helsinki (0.04)
- North America > United States > Virginia > Arlington County > Arlington (0.04)
- (6 more...)
- North America > Canada > Alberta (0.14)
- Asia > Middle East > Jordan (0.04)
- Europe > Netherlands > South Holland > Delft (0.04)
- (4 more...)
Supplementary Material Misspecified GP Bandit Optimization Ilija Bogunovic and Andreas Krause (NeurIPS 2021) A GP bandits: Useful definitions and auxiliary results (Realizable setting)
Such assumptions on the noise variables are frequently used in bandit optimization. Gaussian process with posterior mean and variance that correspond to Eq. (8) and Eq. It also allows us to rewrite Eq. Gaussian Process (supported on D) with the corresponding kernel function. Suppose the learner's hypothesis class is While the first two terms in this bound can be effectively controlled and bounded as in the proof of Theorem 1, the last term, i.e., Such a function can easily be constructed, e.g., via the approach outlined in [36].
- North America > United States (0.27)
- Europe > Switzerland > Zürich > Zürich (0.04)