interpolate
- North America > United States > Wisconsin > Dane County > Madison (0.14)
- Asia > Middle East > Jordan (0.04)
- Europe > Sweden (0.04)
- Research Report > New Finding (0.93)
- Research Report > Experimental Study (0.93)
- Education (0.68)
- Information Technology > Security & Privacy (0.46)
Embedding Space Interpolation Beyond Mini-Batch, Beyond Pairs and Beyond Examples
Mixup refers to interpolation-based data augmentation, originally motivated as a way to go beyond empirical risk minimization (ERM). Its extensions mostly focus on the definition of interpolation and the space (input or embedding) where it takes place, while the augmentation process itself is less studied. In most methods, the number of generated examples is limited to the mini-batch size and the number of examples being interpolated is limited to two (pairs), in the input space. We make progress in this direction by introducing MultiMix, which generates an arbitrarily large number of interpolated examples beyond the mini-batch size, and interpolates the entire mini-batch in the embedding space.
- Europe > Switzerland > Zürich > Zürich (0.14)
- North America > United States > Ohio (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- (5 more...)
- North America > United States > Louisiana > Orleans Parish > New Orleans (0.04)
- Europe > Austria (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- (4 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (0.93)
- Information Technology > Artificial Intelligence > Natural Language (0.88)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.67)
Overfitting or perfect fitting? Risk bounds for classification and regression rules that interpolate
Mikhail Belkin, Daniel J. Hsu, Partha Mitra
Many modern machine learning models are trained to achieve zero or near-zero training error in order to obtain near-optimal (but non-zero) test error. This phenomenon of strong generalization performance for "overfitted" / interpolated classifiers appears to be ubiquitous in high-dimensional data, having been observed in
- Europe > Switzerland > Zürich > Zürich (0.14)
- North America > United States > Ohio (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- (5 more...)
- North America > United States > California > Los Angeles County > Los Angeles (0.28)
- North America > United States > Virginia (0.04)
- North America > United States > California > Santa Clara County > Sunnyvale (0.04)
- (2 more...)
- Information Technology > Security & Privacy (1.00)
- Education (1.00)
- Government > Regional Government > North America Government > United States Government (0.46)