Intra-neuronal attention within language models Relationships between activation and semantics
Pichat, Michael, Pogrund, William, Pichat, Paloma, Gasparian, Armanouche, Demarchi, Samuel, Georgeon, Corbet Alois, Veillet-Guillem, Michael
–arXiv.org Artificial Intelligence
This study investigates the ability of perceptron-type neurons in language models to perform intra-neuronal attention; that is, to identify different homogeneous categorical segments within the synthetic thought category they encode, based on a segmentation of specific activation zones for the tokens to which they are particularly responsive. The objective of this work is therefore to determine to what extent formal neurons can establish a homomorphic relationship between activation-based and categorical segmentations. The results suggest the existence of such a relationship, albeit tenuous, only at the level of tokens with very high activation levels. This intra-neuronal attention subsequently enables categorical restructuring processes at the level of neurons in the following layer, thereby contributing to the progressive formation of high-level categorical abstractions.
arXiv.org Artificial Intelligence
Mar-17-2025
- Country:
- Asia > Singapore (0.04)
- Europe
- France > Auvergne-Rhône-Alpes
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Oxfordshire > Oxford (0.04)
- North America > United States
- California (0.04)
- Illinois (0.04)
- Massachusetts > Middlesex County
- Cambridge (0.04)
- New York (0.04)
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Technology: