Goto

Collaborating Authors

 topographic similarity





A Datasets A.1 Shapes3d Shapes3d is a dataset (see Burgess and Kim (2018) and the Tensorflow Datasets package) consisting

Neural Information Processing Systems

The obverter dataset (Bogin et al. (2018)) is available at the following address: Each dense layer in the receiver has 64 neurons. The same set of hyperparameters was used for all the experiments. The hyperparameters were chosen on the original obverter dataset available at the repository referenced in Appendix A.2. In the Straight-Through mode (see Jang et al. (2016)), The above implementation of noise is not the only one possible. Each experiment was run on 100 seeds.



Reviews: Ease-of-Teaching and Language Structure from Emergent Communication

Neural Information Processing Systems

Overall, the paper was clearly written and had high experimental standards. However, the setting was simple, and it was unclear if the results would apply in more complex language emergence settings. The results about the population setting raise interesting questions that should be further explored. I do think that this paper is different enough from those works: the listener resetting idea here differs from iterated learning where a listener becomes a speaker, and the agent architectures and communication protocols here follow current neural emergent communication research. One shortcoming of the work is that the space of possible inputs and messages is very simple: inputs are purely symbolic, and there are only two attributes, and two tokens in the messages.


Emergent Communication: Generalization and Overfitting in Lewis Games

Rita, Mathieu, Tallec, Corentin, Michel, Paul, Grill, Jean-Bastien, Pietquin, Olivier, Dupoux, Emmanuel, Strub, Florian

arXiv.org Artificial Intelligence

Lewis signaling games are a class of simple communication games for simulating the emergence of language. In these games, two agents must agree on a communication protocol in order to solve a cooperative task. Previous work has shown that agents trained to play this game with reinforcement learning tend to develop languages that display undesirable properties from a linguistic point of view (lack of generalization, lack of compositionality, etc). In this paper, we aim to provide better understanding of this phenomenon by analytically studying the learning problem in Lewis games. As a core contribution, we demonstrate that the standard objective in Lewis games can be decomposed in two components: a co-adaptation loss and an information loss. This decomposition enables us to surface two potential sources of overfitting, which we show may undermine the emergence of a structured communication protocol. In particular, when we control for overfitting on the co-adaptation loss, we recover desired properties in the emergent languages: they are more compositional and generalize better.


Catalytic Role Of Noise And Necessity Of Inductive Biases In The Emergence Of Compositional Communication

Kuciński, Łukasz, Korbak, Tomasz, Kołodziej, Paweł, Miłoś, Piotr

arXiv.org Artificial Intelligence

Communication is compositional if complex signals can be represented as a combination of simpler subparts. In this paper, we theoretically show that inductive biases on both the training framework and the data are needed to develop a compositional communication. Moreover, we prove that compositionality spontaneously arises in the signaling games, where agents communicate over a noisy channel. We experimentally confirm that a range of noise levels, which depends on the model and the data, indeed promotes compositionality. Finally, we provide a comprehensive study of this dependence and report results in terms of recently studied compositionality metrics: topographical similarity, conflict count, and context independence.


On (Emergent) Systematic Generalisation and Compositionality in Visual Referential Games with Straight-Through Gumbel-Softmax Estimator

Denamganaï, Kevin, Walker, James Alfred

arXiv.org Artificial Intelligence

The drivers of compositionality in artificial languages that emerge when two (or more) agents play a non-visual referential game has been previously investigated using approaches based on the REINFORCE algorithm and the (Neural) Iterated Learning Model. Following the more recent introduction of the \textit{Straight-Through Gumbel-Softmax} (ST-GS) approach, this paper investigates to what extent the drivers of compositionality identified so far in the field apply in the ST-GS context and to what extent do they translate into (emergent) systematic generalisation abilities, when playing a visual referential game. Compositionality and the generalisation abilities of the emergent languages are assessed using topographic similarity and zero-shot compositional tests. Firstly, we provide evidence that the test-train split strategy significantly impacts the zero-shot compositional tests when dealing with visual stimuli, whilst it does not when dealing with symbolic ones. Secondly, empirical evidence shows that using the ST-GS approach with small batch sizes and an overcomplete communication channel improves compositionality in the emerging languages. Nevertheless, while shown robust with symbolic stimuli, the effect of the batch size is not so clear-cut when dealing with visual stimuli. Our results also show that not all overcomplete communication channels are created equal. Indeed, while increasing the maximum sentence length is found to be beneficial to further both compositionality and generalisation abilities, increasing the vocabulary size is found detrimental. Finally, a lack of correlation between the language compositionality at training-time and the agents' generalisation abilities is observed in the context of discriminative referential games with visual stimuli. This is similar to previous observations in the field using the generative variant with symbolic stimuli.


Exploring Structural Inductive Biases in Emergent Communication

Słowik, Agnieszka, Gupta, Abhinav, Hamilton, William L., Jamnik, Mateja, Holden, Sean B., Pal, Christopher

arXiv.org Machine Learning

Human language and thought are characterized by the ability to systematically generate a potentially infinite number of complex structures (e.g., sentences) from a finite set of familiar components (e.g., words). Recent works in emergent communication have discussed the propensity of artificial agents to develop a systematically compositional language through playing co-operative referential games. The degree of structure in the input data was found to affect the compositionality of the emerged communication protocols. Thus, we explore various structural priors in multi-agent communication and propose a novel graph referential game. We compare the effect of structural inductive bias (bag-of-words, sequences and graphs) on the emergence of compositional understanding of the input concepts measured by topographic similarity and generalization to unseen combinations of familiar properties. We empirically show that graph neural networks induce a better compositional language prior and a stronger generalization to out-of-domain data. We further perform ablation studies that show the robustness of the emerged protocol in graph referential games.