Lavechin, Marvin
BabySLM: language-acquisition-friendly benchmark of self-supervised spoken language models
Lavechin, Marvin, Sy, Yaya, Titeux, Hadrien, Blandón, María Andrea Cruz, Räsänen, Okko, Bredin, Hervé, Dupoux, Emmanuel, Cristia, Alejandrina
Self-supervised techniques for learning speech representations have been shown to develop linguistic competence from exposure to speech without the need for human labels. In order to fully realize the potential of these approaches and further our understanding of how infants learn language, simulations must closely emulate real-life situations by training on developmentally plausible corpora and benchmarking against appropriate test sets. To this end, we propose a language-acquisition-friendly benchmark to probe spoken language models at the lexical and syntactic levels, both of which are compatible with the vocabulary typical of children's language experiences. This paper introduces the benchmark and summarizes a range of experiments showing its usefulness. In addition, we highlight two exciting challenges that need to be addressed for further progress: bridging the gap between text and speech and between clean speech and in-the-wild speech.
ProsAudit, a prosodic benchmark for self-supervised speech models
de Seyssel, Maureen, Lavechin, Marvin, Titeux, Hadrien, Thomas, Arthur, Virlet, Gwendal, Revilla, Andrea Santos, Wisniewski, Guillaume, Ludusan, Bogdan, Dupoux, Emmanuel
We present ProsAudit, a benchmark in English to assess structural prosodic knowledge in self-supervised learning (SSL) speech models. It consists of two subtasks, their corresponding metrics, and an evaluation dataset. In the protosyntax task, the model must correctly identify strong versus weak prosodic boundaries. In the lexical task, the model needs to correctly distinguish between pauses inserted between words and within words. We also provide human evaluation scores on this benchmark. We evaluated a series of SSL models and found that they were all able to perform above chance on both tasks, even when evaluated on an unseen language. However, non-native models performed significantly worse than native ones on the lexical task, highlighting the importance of lexical knowledge in this task. We also found a clear effect of size with models trained on more data performing better in the two subtasks.