Measuring How (Not Just Whether) VLMs Build Common Ground
Imai, Saki, İnan, Mert, Sicilia, Anthony, Alikhani, Malihe
–arXiv.org Artificial Intelligence
Large vision language models (VLMs) increasingly claim reasoning skills, yet current benchmarks evaluate them in single-turn or question answering settings. However, grounding is an interactive process in which people gradually develop shared understanding through ongoing communication. We introduce a four-metric suite (grounding efficiency, content alignment, lexical adaptation, and human-likeness) to systematically evaluate VLM performance in interactive grounding contexts. We deploy the suite on 150 self-play sessions of interactive referential games between three proprietary VLMs and compare them with human dyads. All three models diverge from human patterns on at least three metrics, while GPT4o-mini is the closest overall. We find that (i) task success scores do not indicate successful grounding and (ii) high image-utterance alignment does not necessarily predict task success. Our metric suite and findings offer a framework for future research on VLM grounding.
arXiv.org Artificial Intelligence
Sep-5-2025
- Country:
- Asia
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.14)
- Singapore (0.04)
- Middle East > UAE
- Europe
- Czechia > Prague (0.04)
- Italy > Tuscany
- Florence (0.04)
- Switzerland > Zürich
- Zürich (0.14)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- North America
- Canada > British Columbia (0.04)
- Dominican Republic (0.04)
- United States
- Colorado > Boulder County
- Boulder (0.04)
- Massachusetts > Suffolk County
- Boston (0.04)
- Colorado > Boulder County
- Asia
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Leisure & Entertainment > Games (0.46)
- Technology: