Online Learning: Stochastic, Constrained, and Smoothed Adversaries
Rakhlin, Alexander, Sridharan, Karthik, Tewari, Ambuj
–Neural Information Processing Systems
Learning theory has largely focused on two main learning scenarios: the classical statistical setting where instances are drawn i.i.d. It can be argued that in the real world neither of these assumptions is reasonable. We define the minimax value of a game where the adversary is restricted in his moves, capturing stochastic and non-stochastic assumptions on data. Building on the sequential symmetrization approach, we define a notion of distribution-dependent Rademacher complexity for the spectrum of problems ranging from i.i.d. to worst-case. The bounds let us immediately deduce variation-type bounds.
Neural Information Processing Systems
Feb-14-2020, 23:13:52 GMT