Robust Target Speaker Diarization and Separation via Augmented Speaker Embedding Sampling
Jalal, Md Asif, Remaggi, Luca, Moschopoulos, Vasileios, Kotsiopoulos, Thanasis, Rajan, Vandana, Saravanan, Karthikeyan, Drosou, Anastasis, Heo, Junho, Oh, Hyuk, Jeong, Seokyeong
–arXiv.org Artificial Intelligence
Traditional speech separation and speaker diarization approaches rely on prior knowledge of target speakers or a predetermined number of participants in audio signals. To address these limitations, recent advances focus on developing enrollment-free methods capable of identifying targets without explicit speaker labeling. This work introduces a new approach to train simultaneous speech separation and diarization using automatic identification of target speaker embeddings, within mixtures. Our proposed model employs a dual-stage training pipeline designed to learn robust speaker representation features that are resilient to background noise interference. Furthermore, we present an overlapping spectral loss function specifically tailored for enhancing diarization accuracy during overlapped speech frames. Experimental results show significant performance gains compared to the current SOT A baseline, achieving 71% relative improvement in DER and 69% in cpWER.
arXiv.org Artificial Intelligence
Aug-11-2025
- Country:
- Asia
- Middle East > Jordan (0.04)
- South Korea (0.04)
- Europe
- Greece (0.04)
- United Kingdom (0.04)
- North America (0.04)
- Asia
- Genre:
- Research Report > New Finding (0.48)
- Technology: