LOVM: Language-Only Vision Model Selection
–Neural Information Processing Systems
Pre-trained multi-modal vision-language models (VLMs) are becoming increasingly popular due to their exceptional performance on downstream vision applications, particularly in the few-and zero-shot settings. However, selecting the best-performing VLM for some downstream applications is non-trivial, as it is dataset and task-dependent. Meanwhile, the exhaustive evaluation of all available VLMs on a novel application is not only time and computationally demanding but also necessitates the collection of a labeled dataset for evaluation. As the number of open-source VLM variants increases, there is a need for an efficient model selection strategy that does not require access to a curated evaluation dataset. This paper proposes a novel task and benchmark for efficiently evaluating VLMs' zero-shot performance on downstream applications without access to the downstream task dataset.
Neural Information Processing Systems
Dec-25-2025, 20:21:47 GMT
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning (0.83)
- Natural Language > Large Language Model (0.53)
- Vision (0.77)
- Information Technology > Artificial Intelligence