Multi-modal Situated Reasoning in 3D Scenes
–Neural Information Processing Systems
Situation awareness is essential for understanding and reasoning about 3D scenes in embodied AI agents. However, existing datasets and benchmarks for situated understanding suffer from severe limitations in data modality, scope, diversity, and scale. To address these limitations, we propose Multi-modal Situated Question Answering (MSQA), a large-scale multi-modal situated reasoning dataset, scalably collected leveraging 3D scene graphs and vision-language models (VLMs) across a diverse range of real-world 3D scenes. MSQA includes 251K situated questionanswering pairs across 9 distinct question categories, covering complex scenarios and object modalities within 3D scenes. We introduce a novel interleaved multimodal input setting in our benchmark to provide both texts, images, and point clouds for situation and question description, aiming to resolve ambiguity in describing situations with single-modality inputs (e.g., texts).
Neural Information Processing Systems
Jun-2-2025, 09:38:57 GMT
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning (0.40)
- Natural Language (0.66)
- Vision (0.46)
- Information Technology > Artificial Intelligence