Teaching Vision-Language Models to Ask: Resolving Ambiguity in Visual Questions
Jian, Pu, Yu, Donglei, Yang, Wen, Ren, Shuo, Zhang, Jiajun
–arXiv.org Artificial Intelligence
In visual question answering (VQA) context, users often pose ambiguous questions to visual language models (VLMs) due to varying expression habits. Existing research addresses such ambiguities primarily by rephrasing questions. These approaches neglect the inherently interactive nature of user interactions with VLMs, where ambiguities can be clarified through user feedback. However, research on interactive clarification faces two major challenges: (1) Benchmarks are absent to assess VLMs' capacity for resolving ambiguities through interaction; (2) VLMs are trained to prefer answering rather than asking, preventing them from seeking clarification. To overcome these challenges, we introduce \textbf{ClearVQA} benchmark, which targets three common categories of ambiguity in VQA context, and encompasses various VQA scenarios.
arXiv.org Artificial Intelligence
Sep-17-2025
- Country:
- Asia
- China > Hubei Province
- Wuhan (0.04)
- Singapore (0.04)
- China > Hubei Province
- North America
- Canada > Ontario
- Toronto (0.04)
- United States > Texas
- Travis County > Austin (0.04)
- Canada > Ontario
- Asia
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Education (0.68)
- Technology: