Benchmarking Robustness of Adaptation Methods on Pre-trained Vision-Language Models (Supplementary) Shuo Chen 1,3 Jindong Gu2 Zhen Han 1 Y unpu Ma1,3
–Neural Information Processing Systems
BART had higher robustness against text corruptions on the GQA dataset. This may be due to the different language encoders used in BART and T5.Among all adaptation The top row shows the robustness against image corruptions and the bottom row is results against text corruptions. We also choose different embedding dimensions for adapter-based methods. The relative robustness of adaptation methods based on CLIP-T5 is presented in Table 1. Figure 2
Neural Information Processing Systems
Feb-16-2026, 06:18:36 GMT
- Country:
- Europe
- Germany > Bavaria
- Upper Bavaria > Munich (0.04)
- United Kingdom > England
- Oxfordshire > Oxford (0.04)
- Germany > Bavaria
- Europe
- Technology:
- Information Technology > Artificial Intelligence
- Natural Language (0.64)
- Vision (0.40)
- Information Technology > Artificial Intelligence