Image-Caption Encoding for Improving Zero-Shot Generalization
Yu, Eric Yang, Liao, Christopher, Ravi, Sathvik, Tsiligkaridis, Theodoros, Kulis, Brian
–arXiv.org Artificial Intelligence
Recent advances in vision-language models have combined contrastive approaches with generative methods to achieve state-of-the-art (SOTA) on downstream inference tasks like zero-shot image classification. However, a persistent issue of these models for image classification is their out-of-distribution (OOD) generalization capabilities. We first show that when an OOD data point is misclassified, the correct class can be typically found in the Top-K predicted classes. In order to steer the model prediction toward the correct class within the top predicted classes, we propose the Image-Caption Encoding (ICE) method, a straightforward approach that directly enforces consistency between the image-conditioned and caption-conditioned predictions at evaluation time only. Intuitively, we take advantage of unique properties of the generated captions to guide our local search for the correct class label within the Top-K predicted classes. We show that our method can be easily combined with other SOTA methods to enhance Top-1 OOD accuracies by 0.5% on average and up to 3% on challenging datasets. Our code: https://github.com/Chris210634/ice
arXiv.org Artificial Intelligence
Feb-4-2024
- Country:
- North America > United States > California (0.14)
- Genre:
- Research Report (1.00)
- Industry:
- Government (0.47)
- Technology: