MFLA: Monotonic Finite Look-ahead Attention for Streaming Speech Recognition
Xia, Yinfeng, Li, Huiyan, Le, Chenyang, Wang, Manhong, Sun, Yutao, Ma, Xingyang, Qian, Yanmin
–arXiv.org Artificial Intelligence
Applying large pre-trained speech models like Whisper has shown promise in reducing training costs for various speech tasks. However, integrating these models into streaming systems remains a challenge. This paper presents a novel prefix-to-prefix training framework for streaming recognition by fine-tuning the Whisper. We introduce the Continuous Integrate-and-Fire mechanism to establish a quasi-monotonic alignment between continuous speech sequences and discrete text tokens. Additionally, we design Monotonic Finite Look-ahead Attention, allowing each token to attend to infinite left-context and finite right-context from the speech sequences. We also employ the wait-k decoding strategy to simplify the decoding process while ensuring consistency between training and testing. Our theoretical analysis and experiments demonstrate that this approach achieves a controllable trade-off between latency and quality, making it suitable for various streaming applications.
arXiv.org Artificial Intelligence
Jun-5-2025
- Country:
- Asia > China
- Europe > Czechia
- South Moravian Region > Brno (0.04)
- North America > United States (0.04)
- Genre:
- Research Report (0.40)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning (1.00)
- Natural Language (1.00)
- Speech > Speech Recognition (0.33)
- Information Technology > Artificial Intelligence