Quick-Draw Bandits: Quickly Optimizing in Nonstationary Environments with Extremely Many Arms
Everett, Derek, Lu, Fred, Raff, Edward, Camacho, Fernando, Holt, James
Canonical algorithms for multi-armed bandits typically assume a stationary reward environment where the size of the action space (number of arms) is small. More recently developed methods typically relax only one of these assumptions: existing non-stationary bandit policies are designed for a small number of arms, while Lipschitz, linear, and Gaussian process bandit policies are designed to handle a large (or infinite) number of arms in stationary reward environments under constraints on the reward function. In this manuscript, we propose a novel policy to learn reward environments over a continuous space using Gaussian interpolation. We show that our method efficiently learns continuous Lipschitz reward functions with $\mathcal{O}^*(\sqrt{T})$ cumulative regret. Furthermore, our method naturally extends to non-stationary problems with a simple modification. We finally demonstrate that our method is computationally favorable (100-10000x faster) and experimentally outperforms sliding Gaussian process policies on datasets with non-stationarity and an extremely large number of arms.
Jun-2-2025
- Country:
- Europe > Finland (0.04)
- North America
- Canada > Ontario
- Toronto (0.05)
- United States
- Maryland > Baltimore (0.04)
- New York > New York County
- New York City (0.04)
- Virginia > Fairfax County
- McLean (0.04)
- Canada > Ontario
- Genre:
- Research Report (1.00)
- Technology: