The Magnitude of Categories of Texts Enriched by Language Models
Bradley, Tai-Danae, Vigneaux, Juan Pablo
–arXiv.org Artificial Intelligence
The purpose of this article is twofold. Firstly, we use the next-token probabilities given by a language model to explicitly define a $[0,1]$-enrichment of a category of texts in natural language, in the sense of Bradley, Terilla, and Vlassopoulos. We consider explicitly the terminating conditions for text generation and determine when the enrichment itself can be interpreted as a probability over texts. Secondly, we compute the M\"obius function and the magnitude of an associated generalized metric space $\mathcal{M}$ of texts using a combinatorial version of these quantities recently introduced by Vigneaux. The magnitude function $f(t)$ of $\mathcal{M}$ is a sum over texts $x$ (prompts) of the Tsallis $t$-entropies of the next-token probability distributions $p(-|x)$ plus the cardinality of the model's possible outputs. The derivative of $f$ at $t=1$ recovers a sum of Shannon entropies, which justifies seeing magnitude as a partition function. Following Leinster and Schulman, we also express the magnitude function of $\mathcal M$ as an Euler characteristic of magnitude homology and provide an explicit description of the zeroeth and first magnitude homology groups.
arXiv.org Artificial Intelligence
Jan-11-2025
- Country:
- Europe
- Poland > Masovia Province
- Warsaw (0.04)
- Switzerland (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Poland > Masovia Province
- North America > United States
- California
- Los Angeles County
- Pasadena (0.04)
- Santa Clarita (0.04)
- Santa Clara County > Palo Alto (0.04)
- Los Angeles County
- Massachusetts > Middlesex County
- Cambridge (0.04)
- California
- Europe
- Genre:
- Research Report (0.40)
- Technology: