Emergence of a High-Dimensional Abstraction Phase in Language Transformers
Cheng, Emily, Doimo, Diego, Kervadec, Corentin, Macocco, Iuri, Yu, Jade, Laio, Alessandro, Baroni, Marco
–arXiv.org Artificial Intelligence
A language model (LM) is a mapping from a linguistic context to an output token. However, much remains to be known about this mapping, including how its geometric properties relate to its function. We take a high-level geometric approach to its analysis, observing, across five pre-trained transformer-based LMs and three input datasets, a distinct phase characterized by high intrinsic dimensionality. During this phase, representations (1) correspond to the first full linguistic abstraction of the input; (2) are the first to viably transfer to downstream tasks; (3) predict each other across different LMs. Moreover, we find that an earlier onset of the phase strongly predicts better language modelling performance. In short, our results suggest that a central high-dimensionality phase underlies core linguistic processing in many common LM architectures.
arXiv.org Artificial Intelligence
May-24-2024
- Country:
- North America
- Canada > Ontario
- Toronto (0.14)
- United States > Oregon (0.14)
- Canada > Ontario
- North America
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Health & Medicine (0.68)
- Technology: