SynParaSpeech: Automated Synthesis of Paralinguistic Datasets for Speech Generation and Understanding
Bai, Bingsong, Lu, Qihang, Yang, Wenbing, Sun, Zihan, Hou, Yueran, Jia, Peilei, Pu, Songbai, Fu, Ruibo, Gao, Yingming, Li, Ya, Gao, Jun
–arXiv.org Artificial Intelligence
ABSTRACT Paralinguistic sounds, like laughter and sighs, are crucial for synthesizing more realistic and engaging speech. However, existing methods typically depend on proprietary datasets, while publicly available resources often suffer from incomplete speech, inaccurate or missing timestamps, and limited real-world relevance. To address these problems, we propose an automated framework for generating large-scale paralinguistic data and apply it to construct the Syn-ParaSpeech dataset. The dataset comprises 6 paralinguistic categories with 118.75 hours of data and precise timestamps, all derived from natural conversational speech. Our contributions lie in introducing the first automated method for constructing large-scale par-alinguistic datasets and releasing the SynParaSpeech corpus, which advances speech generation through more natural paralinguistic synthesis and enhances speech understanding by improving paralinguis-tic event detection.
arXiv.org Artificial Intelligence
Sep-30-2025
- Country:
- Asia > China
- North America > United States
- California (0.04)
- Genre:
- Research Report (0.82)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning (1.00)
- Natural Language (1.00)
- Speech > Speech Recognition (1.00)
- Information Technology > Artificial Intelligence