On the Memorization Properties of Contrastive Learning
Sadrtdinov, Ildus, Chirkova, Nadezhda, Lobacheva, Ekaterina
However, data labeling is often time-consuming and costly, as it involves human expertise. Thus, it is common for computer vision to pretrain DNNs vate improvements to DNN training approaches. A pioneer on some large labeled dataset, e. g. ImageNet (Russakovsky work of Zhang et al. (2017) showed that the capacity of et al., 2015), and then to fine-tune the model to a specific modern DNNs is sufficient to fit perfectly even randomly downstream task. The self-supervised learning paradigm labeled data. According to classic learning theory, such a provides a human labeling-free alternative to the supervised huge capacity should lead to catastrophic overfitting, however, pretraining: recently developed contrastive self-supervised recent works (Nakkiran et al., 2020) show that in methods show results, comparable to ImageNet pretraining practice increasing DNN capacity further improves generalization.
Jul-21-2021
- Country:
- North America > Canada (0.28)
- Genre:
- Research Report (1.00)
- Technology:
- Information Technology > Artificial Intelligence > Machine Learning
- Inductive Learning (0.51)
- Memory-Based Learning > Rote Learning (0.49)
- Neural Networks (0.69)
- Statistical Learning (0.46)
- Information Technology > Artificial Intelligence > Machine Learning