Author:
Mitsufuji Yuki,Fabbro Giorgio,Uhlich Stefan,Stöter Fabian-Robert,Défossez Alexandre,Kim Minseok,Choi Woosung,Yu Chin-Yun,Cheuk Kin-Wai
Abstract
Music source separation has been intensively studied in the last decade and tremendous progress with the advent of deep learning could be observed. Evaluation campaigns such as MIREX or SiSEC connected state-of-the-art models and corresponding papers, which can help researchers integrate the best practices into their models. In recent years, the widely used MUSDB18 dataset played an important role in measuring the performance of music source separation. While the dataset made a considerable contribution to the advancement of the field, it is also subject to several biases resulting from a focus on Western pop music and a limited number of mixing engineers being involved. To address these issues, we designed the Music Demixing Challenge on a crowd-based machine learning competition platform where the task is to separate stereo songs into four instrument stems (Vocals, Drums, Bass, Other). The main differences compared with the past challenges are 1) the competition is designed to more easily allow machine learning practitioners from other disciplines to participate, 2) evaluation is done on a hidden test set created by music professionals dedicated exclusively to the challenge to assure the transparency of the challenge, i.e., the test set is not accessible from anyone except the challenge organizers, and 3) the dataset provides a wider range of music genres and involved a greater number of mixing engineers. In this paper, we provide the details of the datasets, baselines, evaluation metrics, evaluation results, and technical challenges for future competitions.
Reference31 articles.
1. MedleyDB: A Multitrack Dataset for Annotation-Intensive MIR Research;Bittner,2014
2. Dual Path Networks;Chen,2017
3. Investigating U-Nets with Various Intermediate Blocks for Spectrogram-Based Singing Voice Separation;Choi,2020
4. LaSAFT: Latent Source Attentive Frequency Transformation for Conditioned Source Separation;Choi,2021
Cited by
27 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Music source separation via hybrid waveform and spectrogram based generative adversarial network;Multimedia Tools and Applications;2024-08-20
2. MDX-Mixer: Music Demixing by Leveraging Source Signals Separated by Existing Demixing Models;IEICE Transactions on Information and Systems;2024-08-01
3. Toward deep drum source separation;Pattern Recognition Letters;2024-07
4. Why Does Music Source Separation Benefit from Cacophony?;2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW);2024-04-14
5. SingFake: Singing Voice Deepfake Detection;ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP);2024-04-14