More than Correlation: Do Large Language Models Learn Causal Representations of Space?
Chen, Yida, Gan, Yixian, Li, Sijia, Yao, Li, Zhao, Xiaohan
–arXiv.org Artificial Intelligence
Recent work found high mutual information between the learned representations of large language models (LLMs) and the geospatial property of its input, hinting an emergent internal model of space. However, whether this internal space model has any causal effects on the LLMs' behaviors was not answered by that work, led to criticism of these findings as mere statistical correlation. Our study focused on uncovering the causality of the spatial representations in LLMs. In particular, we discovered the potential spatial representations in DeBERTa, GPT-Neo using representational similarity analysis and linear and non-linear probing. Our casual intervention experiments showed that the spatial representations influenced the model's performance on next word prediction and a downstream task that relies on geospatial information. Our experiments suggested that the LLMs learn and use an internal model of space in solving geospatial related tasks.
arXiv.org Artificial Intelligence
Dec-25-2023
- Country:
- Asia
- China > Shanghai
- Shanghai (0.04)
- Japan > Honshū
- Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- Middle East > Iran (0.04)
- South Korea (0.04)
- China > Shanghai
- Europe
- France (0.04)
- United Kingdom > England
- Greater London > London (0.04)
- North America
- Canada > Ontario
- Middlesex County > London (0.14)
- United States
- Massachusetts > Middlesex County
- Cambridge (0.04)
- New York > New York County
- New York City (0.04)
- Massachusetts > Middlesex County
- Canada > Ontario
- South America
- Argentina > Pampas
- Buenos Aires F.D. > Buenos Aires (0.04)
- Brazil > São Paulo (0.04)
- Argentina > Pampas
- Asia
- Genre:
- Research Report
- Experimental Study (0.89)
- New Finding (1.00)
- Research Report
- Industry:
- Health & Medicine (0.46)
- Technology: