Che, Mingjin
MERTech: Instrument Playing Technique Detection Using Self-Supervised Pretrained Model With Multi-Task Finetuning
Li, Dichucheng, Ma, Yinghao, Wei, Weixing, Kong, Qiuqiang, Wu, Yulun, Che, Mingjin, Xia, Fan, Benetos, Emmanouil, Li, Wei
Instrument playing techniques (IPTs) constitute a pivotal component of musical expression. However, the development of automatic IPT detection methods suffers from limited labeled data and inherent class imbalance issues. In this paper, we propose to apply a self-supervised learning model pre-trained on large-scale unlabeled music data and finetune it on IPT detection tasks. This approach addresses data scarcity and class imbalance challenges. Recognizing the significance of pitch in capturing the nuances of IPTs and the importance of onset in locating IPT events, we investigate multi-task finetuning with pitch and onset detection as auxiliary tasks. Additionally, we apply a post-processing approach for event-level prediction, where an IPT activation initiates an event only if the onset output confirms an onset in that frame. Our method outperforms prior approaches in both frame-level and event-level metrics across multiple IPT benchmark datasets. Further experiments demonstrate the efficacy of multi-task finetuning on each IPT class.
Frame-Level Multi-Label Playing Technique Detection Using Multi-Scale Network and Self-Attention Mechanism
Li, Dichucheng, Che, Mingjin, Meng, Wenwu, Wu, Yulun, Yu, Yi, Xia, Fan, Li, Wei
With the advancements in deep learning, deep neural networks have been increasingly used in more recent work [8, 9]. In [10], a Instrument playing technique (IPT) is a key element in enhancing convolutional recurrent neural network (CRNN) based model was the vividness of musical performance. As shown by the Guzheng proposed to classify IPTs in audio sequences concatenated by cello numbered musical notation (a musical notation system widely used notes from 5 sound banks. To alleviate the computational redundancy in China) in Fig.1, a complete automatic music transcription (AMT) caused by the sliding window in [10], Wang et al. [11] proposed system should contain IPT information in addition to pitch and onset a fully convolutional network (FCN) based end-to-end method information. IPT detection aims to classify the types of IPTs and to detect IPTs in segments concatenated by isolated Erhu notes. In locate the associated IPT boundaries in audio. IPT detection and [12], an additional onset detector was used, and its output was fused modeling can be utilized in many applications of music information with IPT prediction in a post-processing step to improve the accuracy retrieval (MIR), like performance analysis [1] and AMT [2]. of IPT detection from monophonic audio sequences concatenated by The research on IPT detection is still in its early stage.