Goto

Collaborating Authors

 linguistic community


On the referential capacity of language models: An internalist rejoinder to Mandelkern & Linzen

Baggio, Giosue, Murphy, Elliot

arXiv.org Artificial Intelligence

In a recent paper, Mandelkern & Linzen (2024) - henceforth M&L - address the question of whether language models' (LMs) words refer. Their argument draws from the externalist tradition in philosophical semantics, which views reference as the capacity of words to "achieve 'word-to-world' connections". In the externalist framework, causally uninterrupted chains of usage, tracing every occurrence of a name back to its bearer, guarantee that, for example, 'Peano' refers to the individual Peano (Kripke 1980). This account is externalist both because words pick out referents 'out there' in the world, and because what determines reference are coordinated linguistic actions by members of a community, and not individual mental states. The "central question to ask", for M&L, is whether LMs too belong to human linguistic communities, such that words by LMs may also trace back causally to their bearers. Their answer is a cautious "yes": inputs to LMs are linguistic "forms with particular histories of referential use"; "those histories ground the referents of those forms"; any occurrence of 'Peano' in LM outputs is as causally connected to the individual Peano as any other occurrence of the same proper name in human speech or text; therefore, occurrences of 'Peano' in LM outputs refer to Peano. In this commentary, we first qualify M&L's claim as applying to a narrow class of natural language expressions. Thus qualified, their claim is valid, and we emphasise an additional motivation for that in Section 2. Next, we discuss the actual scope of their claim, and we suggest that the way they formulate it may lead to unwarranted generalisations about reference in LMs. Our critique is likewise applicable to other externalist accounts of LMs (e.g., Lederman & Mahowald 2024; Mollo & Milliere 2023). Lastly, we conclude with a comment on the status of LMs as members of human linguistic communities.


Do Language Models Refer?

Mandelkern, Matthew, Linzen, Tal

arXiv.org Artificial Intelligence

What do language models (LMs) do with language? Everyone agrees that they produce sequences of (mostly) coherent sentences. But are they saying anything with those strings or simply babbling in a convincing simulacrum of language use? This is a vague question, and there are many ways of making it precise. Here we will address one aspect of the question, namely, whether LMs' words refer: that is, whether the outputs of LMs achieve "word-to-world" connections. There is prima facie reason to think they do not since LMs do not interact with the world in the way that ordinary language users do. Drawing on insights from the externalist tradition in philosophy of language, we argue that appearances are misleading and that there is good reason to think that LMs can refer.


Facebook and Google built a framework to study how AI agents talk to each other

#artificialintelligence

The intricacies of evolutionary linguistics are myriad and underexplored, but new research involving artificial intelligence (AI) might unlock the door to new theories about how dialects develop among users. Their work isn't the first to investigate language with machine learning algorithms -- a paper published by Facebook researchers in June 2017 describes how two agents learned to "negotiate" with each other in chat messages. But they say that it's the first to use "latest-generation deep neural agents" capable of dealing with "rich perceptual input," and that it convincingly demonstrates that language can evolve from simple exchanges. The team deployed groups -- communities -- of agents equipped with the ability to communicate in a simulated environment, with complexities ranging from simple (a set of equations) to relatively complicated (a deep neural network). The "games" the agents were tasked with playing had several key properties: they were symmetric, enabling the agents to act as both "speakers" and "listeners"; they allowed the agents to communicate about something "external" to themselves, such as the sensory experience of something in their environment; and they took place in a world the agents could at least partially observe.


Emergent Linguistic Phenomena in Multi-Agent Communication Games

Graesser, Laura, Cho, Kyunghyun, Kiela, Douwe

arXiv.org Artificial Intelligence

In this work, we propose a computational framework in which agents equipped with communication capabilities simultaneously play a series of referential games, where agents are trained using deep reinforcement learning. We demonstrate that the framework mirrors linguistic phenomena observed in natural language: i) the outcome of contact between communities is a function of inter- and intra-group connectivity; ii) linguistic contact either converges to the majority protocol, or in balanced cases leads to novel creole languages of lower complexity; and iii) a linguistic continuum emerges where neighboring languages are more mutually intelligible than farther removed languages. We conclude that intricate properties of language evolution need not depend on complex evolved linguistic capabilities, but can emerge from simple social exchanges between perceptually-enabled agents playing communication games.