WavShape: Information-Theoretic Speech Representation Learning for Fair and Privacy-Aware Audio Processing

Baser, Oguzhan, Tanriverdi, Ahmet Ege, Kale, Kaan, Chinchali, Sandeep P., Vishwanath, Sriram

arXiv.org Artificial Intelligence 

Speech embeddings often retain sensitive attributes such as speaker identity, accent, or demographic information, posing risks in biased model training and privacy leakage. We propose WavShape, an information-theoretic speech representation learning framework that optimizes embeddings for fairness and privacy while preserving task-relevant information. We leverage mutual information (MI) estimation using the Donsker-V aradhan formulation to guide an MI-based encoder that systematically filters sensitive attributes while maintaining speech content essential for downstream tasks. Experimental results on three known datasets show that WavShape reduces MI between embeddings and sensitive attributes by up to 81% while retaining 97% of task-relevant information.