Resources for the paper: User-guided one-shot deep model adaptation for music source separation
By G. Cantisani
In this work, we propose to exploit a temporal segmentation provided by the user indicating when each instrument is active, in order to fine-tune a pre-trained deep model for source separation and adapt it to one specific mixture. This paradigm can be referred to as user-guided one-shot deep model adaptation for music source separation, as the adaptation acts on the target song instance only.
The adaptation is possible thanks to a proposed loss function which aims to minimize the energy of the silent sources while at the same time forcing the perfect reconstruction of the mixture.
The results are promising and show that state-of-the-art source separation models have large margins of improvement especially for those instruments which are underrepresented in the training data. Below you can find some audio examples from the MUSDB18 test set.
The paper
Cantisani, G., Ozerov, A., Essid, S., & Richard, G. (2021, October). User-guided one-shot deep model adaptation for music source separation. 2021 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA). https://telecom-paris.hal.science/hal-03219350
Au sein d’ADASP, nous n’aimons pas beaucoup les adieux, mais nous sommes malgré tout très heureux et très fiers de voir nos collègues partir vers un avenir p... ≥≥
At ADASP, we are not especially fond of goodbyes, but we are nevertheless very happy and proud to see our colleagues move on toward such promising futures. T... ≥≥
Nous sommes très heureux d’annoncer que Hi-Audio a reçu un Best Poster Award (short paper) lors de la 9e Web Audio Conference (WAC 2025), organisée à l’IRCAM... ≥≥
We are delighted to share that Hi-Audio received a Best Poster Award (short paper) at the 9th Web Audio Conference (WAC 2025), held at IRCAM in Paris, France... ≥≥