Whale: Large-Scale multilingual ASR model with w2v-BERT and E-Branchformer with large speech data

Kashiwagi, Yosuke, Futami, Hayato, Tsunoo, Emiru, Asakawa, Satoshi

arXiv.org Artificial Intelligence 

Whale's architecture integrates w2v-BERT self-supervised model, an encoder-decoder backbone built on E-Branchformer, and a joint CTC-attention decoding strategy. The training corpus comprises varied speech data, of not only public corpora but also in-house data, thereby enhancing the model's robustness to different speaking styles and acoustic conditions. Through evaluations on multiple benchmarks, Whale achieved comparable performance to existing models. In particular, it achieves a word error rate of 2.4% on the Librispeech test-clean set and a character error rate of 3.4% on the CSJ eval3 set, outperforming Whisper large-v3 and OWSM v3.1.