Limits of Transformer Language Models on Learning to Compose Algorithms Jonathan Thomm
–Neural Information Processing Systems
We analyze the capabilities of Transformer language models in learning compositional discrete tasks. To this end, we evaluate training LLaMA models and prompting GPT -4 and Gemini on four tasks demanding to learn a composition of several discrete sub-tasks. In particular, we measure how well these models can reuse primitives observable in the sub-tasks to learn the composition task.
Neural Information Processing Systems
Nov-13-2025, 17:01:25 GMT
- Country:
- Asia
- Middle East > Jordan (0.04)
- Singapore (0.04)
- Europe
- Germany > Baden-Württemberg
- Tübingen Region > Tübingen (0.14)
- Switzerland > Zürich
- Zürich (0.04)
- Germany > Baden-Württemberg
- North America
- Canada > Ontario
- Toronto (0.04)
- United States > Massachusetts
- Suffolk County > Boston (0.04)
- Canada > Ontario
- Asia
- Genre:
- Research Report > Experimental Study (1.00)
- Industry:
- Information Technology (0.68)
- Technology: