Benchmarking Training Paradigms, Dataset Composition, and Model Scaling for Child ASR in ESPnet

Ying, Anyu, Shankar, Natarajan Balaji, Lin, Chyi-Jiunn, Shi, Mohan, Wang, Pu, Shim, Hye-jin, Arora, Siddhant, Van hamme, Hugo, Alwan, Abeer, Watanabe, Shinji

arXiv.org Artificial Intelligence 

Despite advancements in ASR, child speech recognition remains challenging due to acoustic variability and limited annotated data. While fine-tuning adult ASR models on child speech is common, comparisons with flat-start training remain underexplored. We compare flat-start training across multiple datasets, SSL representations (WavLM, XEUS), and decoder architectures. Our results show that SSL representations are biased toward adult speech, with flat-start training on child speech mitigating these biases. We also analyze model scaling, finding consistent improvements up to 1B parameters, beyond which performance plateaus. Additionally, age-related ASR and speaker verification analysis highlights the limitations of proprietary models like Whisper, emphasizing the need for open-data models for reliable child speech research. All investigations are conducted using ESPnet, and our publicly available benchmark provides insights into training strategies for robust child speech processing.