Nearly Tight Bounds for Exploration in Streaming Multi-armed Bandits with Known Optimality Gap

Open in new window