Between Stochastic and Adversarial Online Convex Optimization: Improved Regret Bounds via Smoothness
–Neural Information Processing Systems
Stochastic and adversarial data are two widely studied settings in online learning. But many optimizationtasks are neither i.i.d. In this work we establish novel regret bounds for online convex optimization in a setting that interpolates between stochastic i.i.d. and fully adversarial losses. By exploiting smoothness of the expected losses, these bounds replace a dependence on the maximum gradient length by the variance of the gradients, which was previously known only for linear losses. In addition, they weaken the i.i.d.
Neural Information Processing Systems
Oct-9-2024, 09:52:22 GMT
- Technology: