Language over Content: Tracing Cultural Understanding in Multilingual Large Language Models
Cho, Seungho, Ko, Changgeon, Hwang, Eui Jun, Lee, Junmyeong, Lee, Huije, Park, Jong C.
–arXiv.org Artificial Intelligence
Large language models (LLMs) are increasingly used across diverse cultural contexts, making accurate cultural understanding essential. Prior evaluations have mostly focused on output-level performance, obscuring the factors that drive differences in responses, while studies using circuit analysis have covered few languages and rarely focused on culture. In this work, we trace LLMs' internal cultural understanding mechanisms by measuring activation path overlaps when answering semantically equivalent questions under two conditions: varying the target country while fixing the question language, and varying the question language while fixing the country. We also use same-language country pairs to disentangle language from cultural aspects. Results show that internal paths overlap more for same-language, cross-country questions than for cross-language, same-country questions, indicating strong language-specific patterns. Notably, the South Korea-North Korea pair exhibits low overlap and high variability, showing that linguistic similarity does not guarantee aligned internal representation.
arXiv.org Artificial Intelligence
Nov-12-2025
- Country:
- Asia
- China (0.05)
- Middle East
- Jordan (0.04)
- UAE > Abu Dhabi Emirate
- Abu Dhabi (0.04)
- North Korea (0.36)
- South Korea
- Thailand > Bangkok
- Bangkok (0.04)
- Europe
- Austria > Vienna (0.15)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Middle East > Malta
- Eastern Region > Northern Harbour District > St. Julian's (0.04)
- Spain (0.05)
- United Kingdom (0.28)
- North America
- Mexico > Mexico City
- Mexico City (0.04)
- United States > Florida
- Miami-Dade County > Miami (0.04)
- Mexico > Mexico City
- South America > Colombia
- Meta Department > Villavicencio (0.04)
- Asia
- Genre:
- Research Report > New Finding (0.88)
- Technology: