Algorithmic Phase Transitions in Language Models: A Mechanistic Case Study of Arithmetic
Sun, Alan, Sun, Ethan, Shepard, Warren
–arXiv.org Artificial Intelligence
Zero-shot capabilities of large language models make them powerful tools for solving a range of tasks without explicit training. It remains unclear, however, how these models achieve such performance, or why they can zero-shot some tasks but not others. In this paper, we shed some light on this phenomenon by defining and investigating algorithmic stability in language models -- changes in problem-solving strategy employed by the model as a result of changes in task specification. We focus on a task where algorithmic stability is needed for generalization: two-operand arithmetic. Surprisingly, we find that Gemma-2-2b employs substantially different computational models on closely related subtasks, i.e. four-digit versus eight-digit addition. Our findings suggest that algorithmic instability may be a contributing factor to language models' poor zero-shot performance across certain logical reasoning tasks, as they struggle to abstract different problem-solving strategies and smoothly transition between them.
arXiv.org Artificial Intelligence
Dec-10-2024
- Country:
- Asia
- Indonesia > Bali (0.04)
- Middle East
- Jordan (0.04)
- UAE > Abu Dhabi Emirate
- Abu Dhabi (0.04)
- Singapore (0.04)
- Europe > Croatia
- Dubrovnik-Neretva County > Dubrovnik (0.04)
- North America
- Canada > Ontario
- Toronto (0.04)
- Dominican Republic (0.04)
- United States > Pennsylvania
- Allegheny County > Pittsburgh (0.04)
- Canada > Ontario
- Asia
- Genre:
- Research Report > New Finding (0.68)
- Technology: