VocalNet: Speech LLM with Multi-Token Prediction for Faster and High-Quality Generation
Wang, Yuhao, Liu, Heyang, Cheng, Ziyang, Wu, Ronghua, Gu, Qunshan, Wang, Yanfeng, Wang, Yu
–arXiv.org Artificial Intelligence
Speech large language models (LLMs) have emerged as a prominent research focus in speech processing. We introduce VocalNet-1B and VocalNet-8B, a series of high-performance, low-latency speech LLMs enabled by a scalable and model-agnostic training framework designed for real-time voice interaction. Central to our contribution is the first application of multi-token prediction (MTP) to speech LLMs. This approach represents a paradigm shift from standard next-token prediction (NTP), offering simultaneous improvements in generation speed and quality. Informed by analysis of MTP's effect on speech generation and experimental comparisons, we designed a straightforward and highly effective MTP implementation. Experiments demonstrate that VocalNet performs on par with mainstream Omni LLMs even with limited training data, and significantly surpasses existing open-source speech LLMs. To foster reproducibility and community advancement, all model weights, inference code, training data, and framework implementations have been made publicly available at https://github.com/SJTU-OmniAgent/VocalNet
arXiv.org Artificial Intelligence
Apr-23-2025
- Country:
- Asia
- China
- Hubei Province > Wuhan (0.04)
- Shanghai > Shanghai (0.04)
- Japan > Honshū
- Kantō > Tokyo Metropolis Prefecture > Tokyo (0.04)
- China
- North America > United States (0.04)
- South America > Chile
- Asia
- Genre:
- Research Report > New Finding (0.88)
- Technology: