Vision Language Models Know Law of Conservation without Understanding More-or-Less
Luo, Dezhi, Lyu, Haiyun, Gao, Qingying, Sun, Haoran, Li, Yijiang, Deng, Hokin
–arXiv.org Artificial Intelligence
Conservation is a critical milestone of cognitive development considered to be supported by both the understanding of quantitative concepts and the reversibility of mental operations. To assess whether this critical component of human intelligence has emerged in Vision Language Models, we have curated the ConserveBench, a battery of 365 cognitive experiments across four dimensions of physical quantities: volume, solid quantity, length, and number. The former two involve only transformational tasks, whereas the latter two involve non-transformational tasks assessing the understanding of quantitative concepts alone. Surprisingly, we find that while Vision Language Models are generally capable of conserving, they tend to fail at non-transformational tasks whose successes are typically considered to be evidence of the ability to conserve. This implies that the law of conservation, at least in concrete domains, may exist without corresponding conceptual understanding of quantity.
arXiv.org Artificial Intelligence
Dec-22-2024
- Country:
- North America > United States (0.68)
- Genre:
- Research Report (0.64)
- Industry:
- Health & Medicine (0.96)
- Technology:
- Information Technology > Artificial Intelligence
- Cognitive Science > Problem Solving (0.35)
- Natural Language (1.00)
- Vision (0.91)
- Information Technology > Artificial Intelligence