minimax
- North America > United States > California (0.04)
- North America > Canada (0.04)
- Asia > China > Zhejiang Province > Hangzhou (0.04)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Government (0.67)
- North America > United States > California (0.04)
- North America > Canada (0.04)
- Asia > China > Zhejiang Province > Hangzhou (0.04)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Government (0.67)
Multiclass Transductive Online Learning
We consider the problem of multiclass transductive online learning when the number of labels can be unbounded. Previous works by Ben-David et al. [1997] and Hanneke et al. [2024] only consider the case of binary and finite label spaces respectively. The latter work determined that their techniques fail to extend to the case of unbounded label spaces, and they pose the question of characterizing the optimal mistake bound for unbounded label spaces. We answer this question, by showing that a new dimension, termed the Level-constrained Littlestone dimension, characterizes online learnability in this setting. Along the way, we show that the trichotomy of possible minimax rates established by Hanneke et al. [2024] for finite label spaces in the realizable setting continues to hold even when the label space is unbounded.
- Europe > Switzerland > Zürich > Zürich (0.05)
- North America > Canada > Ontario > Toronto (0.04)
- Asia > Middle East > Israel (0.04)
- North America > United States > California (0.04)
- North America > Canada (0.04)
- Asia > China > Zhejiang Province > Hangzhou (0.04)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Government (0.67)
On the Optimality of the Median-of-Means Estimator under Adversarial Contamination
de Juan, Xabier, Mazuelas, Santiago
The Median-of-Means (MoM) is a robust estimator widely used in machine learning that is known to be (minimax) optimal in scenarios where samples are i.i.d. In more grave scenarios, samples are contaminated by an adversary that can inspect and modify the data. Previous work has theoretically shown the suitability of the MoM estimator in certain contaminated settings. However, the (minimax) optimality of MoM and its limitations under adversarial contamination remain unknown beyond the Gaussian case. In this paper, we present upper and lower bounds for the error of MoM under adversarial contamination for multiple classes of distributions. In particular, we show that MoM is (minimax) optimal in the class of distributions with finite variance, as well as in the class of distributions with infinite variance and finite absolute $(1+r)$-th moment. We also provide lower bounds for MoM's error that match the order of the presented upper bounds, and show that MoM is sub-optimal for light-tailed distributions.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- Europe > Spain (0.04)
- North America > United States > California (0.04)
- North America > Canada (0.04)
- Asia > China > Zhejiang Province > Hangzhou (0.04)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Government (0.67)
Transductive and Learning-Augmented Online Regression
Raman, Vinod, Xie, Shenghao, Zhou, Samson
Motivated by the predictable nature of real-life in data streams, we study online regression when the learner has access to predictions about future examples. In the extreme case, called transductive online learning, the sequence of examples is revealed to the learner before the game begins. For this setting, we fully characterize the minimax expected regret in terms of the fat-shattering dimension, establishing a separation between transductive online regression and (adversarial) online regression. Then, we generalize this setting by allowing for noisy or \emph{imperfect} predictions about future examples. Using our results for the transductive online setting, we develop an online learner whose minimax expected regret matches the worst-case regret, improves smoothly with prediction quality, and significantly outperforms the worst-case regret when future example predictions are precise, achieving performance similar to the transductive online learner. This enables learnability for previously unlearnable classes under predictable examples, aligning with the broader learning-augmented model paradigm.
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.14)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > United States > Texas (0.04)
- (2 more...)
- Europe > Switzerland > Zürich > Zürich (0.05)
- North America > Canada > Ontario > Toronto (0.04)
- Asia > Middle East > Israel (0.04)