Goto

Collaborating Authors

 interspeech





Discrete Optimal Transport and Voice Conversion

Selitskiy, Anton, Kocharekar, Maitreya

arXiv.org Artificial Intelligence

In this work, we address the voice conversion (VC) task using a vector-based interface. To align audio embeddings between speakers, we employ discrete optimal transport mapping. Our evaluation results demonstrate the high quality and effectiveness of this method. Additionally, we show that applying discrete optimal transport as a post-processing step in audio generation can lead to the incorrect classification of synthetic audio as real.




Neural Analysis and Synthesis: Reconstructing Speech from Self-Supervised Representations

Neural Information Processing Systems

We present a neural analysis and synthesis (NANSY) framework that can manipulate voice, pitch, and speed of an arbitrary speech signal. Most of the previous works have focused on using information bottleneck to disentangle analysis features for controllable synthesis, which usually results in poor reconstruction quality.


Listening to Sounds of Silence for Speech Denoising

Neural Information Processing Systems

When we listen to someone speak, the audio signals we receive are never pure and clean, always contaminated by all kinds of noises--cars passing by, spinning fans in an air conditioner, barking dogs, music from a loudspeaker, and so forth.


MERaLiON-SER: Robust Speech Emotion Recognition Model for English and SEA Languages

Sailor, Hardik B., Ti, Aw Ai, Nancy, Chen Fang Yih, Lay, Chiu Ying, Yang, Ding, Yingxu, He, Ridong, Jiang, Jingtao, Li, Jingyi, Liao, Zhuohan, Liu, Yanfeng, Lu, Yi, Ma, Gupta, Manas, Shahrin, Muhammad Huzaifah Bin Md, Johan, Nabilah Binte Md, Lertcheva, Nattadaporn, Chunlei, Pan, Duc, Pham Minh, Subaidi, Siti Maryam Binte Ahmad, Salleh, Siti Umairah Binte Mohammad, Shuo, Sun, Vangani, Tarun Kumar, Qiongqiong, Wang, Lewis, Won Cheng Yi, Jeremy, Wong Heng Meng, Jinyang, Wu, Huayun, Zhang, Longyin, Zhang, Xunlong, Zou

arXiv.org Artificial Intelligence

We present MERaLiON-SER, a robust speech emotion recognition model designed for English and Southeast Asian languages. The model is trained using a hybrid objective combining weighted categorical cross-entropy and Concordance Correlation Coefficient (CCC) losses for joint discrete and dimensional emotion modelling. This dual approach enables the model to capture both the distinct categories of emotion (like happy or angry) and the fine-grained, such as arousal (intensity), valence (positivity/negativity), and dominance (sense of control), leading to a more comprehensive and robust representation of human affect. Extensive evaluations across multilingual Singaporean languages (English, Chinese, Malay, and Tamil ) and other public benchmarks show that MERaLiON-SER consistently surpasses both open-source speech encoders and large Audio-LLMs. These results underscore the importance of specialised speech-only models for accurate paralinguistic understanding and cross-lingual generalisation. Furthermore, the proposed framework provides a foundation for integrating emotion-aware perception into future agentic audio systems, enabling more empathetic and contextually adaptive multimodal reasoning.


Context-Aware Dynamic Chunking for Streaming Tibetan Speech Recognition

Wang, Chao, Cai, Yuqing, Duojie, Renzeng, Zhang, Jin, Liu, Yutong, Tashi, Nyima

arXiv.org Artificial Intelligence

ABSTRACT In this work, we propose a streaming speech recognition framework for Amdo Tibetan, built upon a hybrid CTC/Atten-tion architecture with a context-aware dynamic chunking mechanism. The proposed strategy adaptively adjusts chunk widths based on encoding states, enabling flexible receptive fields, cross-chunk information exchange, and robust adaptation to varying speaking rates, thereby alleviating the context truncation problem of fixed-chunk methods. To further capture the linguistic characteristics of Tibetan, we construct a lexicon grounded in its orthographic principles, providing linguistically motivated modeling units. During decoding, an external language model is integrated to enhance semantic consistency and improve recognition of long sentences. Experimental results show that the proposed framework achieves a word error rate (WER) of 6.23% on the test set, yielding a 48.15% relative improvement over the fixed-chunk baseline, while significantly reducing recognition latency and maintaining performance close to global decoding.