When End-to-End is Overkill: Rethinking Cascaded Speech-to-Text Translation

Min, Anna, Hu, Chenxu, Ren, Yi, Zhao, Hang

arXiv.org Artificial Intelligence 

Abstract--Though end-to-end speech-to-text translation has been a great success, we argue that the cascaded speech-to-text translation model still has its place, which is usually criticized for the error propagation between automatic speech recognition (ASR) and machine translation (MT) models. In this paper, we explore the benefits of incorporating multiple candidates from ASR and self-supervised speech features into MT. Our analysis reveals that the primary cause of cascading errors stems from the increased divergence between similar samples in the speech domain when mapped to the text domain. By including multiple candidates and self-supervised speech features, our approach allows the machine translation model to choose the right words and ensure precise translation using various speech samples. This strategy minimizes error spread and takes advantage of large ASR and MT datasets, along with pre-trained ASR/MT models, while addressing associated issues. Recent studies [18], [19] have demonstrated the performance In recent years, the academic community has been intrigued improvements achieved by scaling up pre-trained models by the rapid advancement of end-to-end speech-to-text translation for downstream natural language processing tasks.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found