A Trichotomy for Online Learning
–Neural Information Processing Systems
Qualitatively, we prove a trichotomy, stating that the minimal number of mistakes made by the learner as n grows can take only one of precisely three possible values: n, Θ (log(n)), or Θ(1). Furthermore, this behavior is determined by a combination of the VC dimension and the Littlestone dimension. Quantitatively, we show a variety of bounds relating the number of mistakes to well-known combinatorial dimensions. In particular, (we improve log(d)) the known lower bound on the constant in the Θ(1) case from Ω to Ω(log(d)) where d is the Littlestone dimension. Finally, we extend our results to cover multiclass classification and the agnostic setting.
Neural Information Processing Systems
Feb-6-2025, 21:16:49 GMT