Characterizing the Expressivity of Fixed-Precision Transformer Language Models
–arXiv.org Artificial Intelligence
Transformer-based language models (LMs) have achieved widespread empirical success, but their theoretical expressive power remains only partially understood. In this work, we analyze a restricted idealization of fixed-precision transformers with strict future masking, soft attention, and no positional encodings. We establish that this class of models is exactly as expressive as a specific fragment of linear temporal logic that contains only a single temporal operator: the past operator. We further connect this fragment to established classes in formal language theory, automata theory, and algebra, yielding a unified framework for understanding transformer expressivity under this idealization. Finally, we present empirical results that align closely with our theory: transformers trained on languages within their characterized expressive capacity generalize reliably across sequence lengths, while they consistently fail to generalize on languages beyond it.
arXiv.org Artificial Intelligence
Dec-4-2025
- Country:
- Asia
- Middle East > Saudi Arabia
- Asir Province > Abha (0.04)
- Thailand > Bangkok
- Bangkok (0.04)
- Middle East > Saudi Arabia
- Europe > Ireland
- Leinster > County Dublin > Dublin (0.04)
- North America
- Mexico
- Gulf of Mexico (0.04)
- Mexico City > Mexico City (0.04)
- United States
- California > Los Angeles County
- Los Angeles (0.14)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Massachusetts > Middlesex County
- Cambridge (0.04)
- California > Los Angeles County
- Mexico
- Oceania > Australia
- South America > Colombia
- Meta Department > Villavicencio (0.04)
- Asia
- Genre:
- Research Report
- Experimental Study (1.00)
- New Finding (0.92)
- Research Report
- Technology: