A Closer Look at the Robustness of Contrastive Language-Image Pre-Training (CLIP) Weijie T u 1 Weijian Deng
–Neural Information Processing Systems
Contrastive Language-Image Pre-training (CLIP) models have demonstrated remarkable generalization capabilities across multiple challenging distribution shifts. However, there is still much to be explored in terms of their robustness to the variations of specific visual factors.
Neural Information Processing Systems
Feb-9-2026, 12:17:03 GMT
- Genre:
- Research Report > New Finding (0.68)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning
- Neural Networks > Deep Learning (0.93)
- Performance Analysis > Accuracy (0.95)
- Natural Language (1.00)
- Vision (1.00)
- Machine Learning
- Information Technology > Artificial Intelligence