A Study on the Data Distribution Gap in Music Emotion Recognition
–arXiv.org Artificial Intelligence
Music Emotion Recognition (MER) is a task deeply connected to human perception, relying heavily on subjective annotations collected from contributors. Prior studies tend to focus on specific musical styles rather than incorporating a diverse range of genres, such as rock and classical, within a single framework. In this paper, we address the task of recognizing emotion from audio content by investigating five datasets with dimensional emotion annotations -- EmoMusic, DEAM, PMEmo, WTC, and WCMED -- which span various musical styles. We demonstrate the problem of out-of-distribution generalization in a systematic experiment. By closely looking at multiple data and feature sets, we provide insight into genre-emotion relationships in existing data and examine potential genre dominance and dataset biases in certain feature representations. Based on these experiments, we arrive at a simple yet effective framework that combines embeddings extracted from the Jukebox model with chroma features and demonstrate how, alongside a combination of several diverse training sets, this permits us to train models with substantially improved cross-dataset generalization capabilities.
arXiv.org Artificial Intelligence
Nov-14-2025
- Country:
- Europe
- Austria > Upper Austria
- Linz (0.04)
- Germany > Saarland (0.04)
- Austria > Upper Austria
- North America > United States
- Florida > Palm Beach County > Boca Raton (0.04)
- Europe
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Leisure & Entertainment (1.00)
- Media > Music (1.00)
- Technology: