Nearly Tight Bounds for Exploration in Streaming Multi-armed Bandits with Known Optimality Gap