Paper ID | MLSP-23.3 |
Paper Title |
GROOVE2GROOVE: ONE-SHOT MUSIC STYLE TRANSFER WITH SUPERVISION FROM SYNTHETIC DATA |
Authors |
Ondřej Cífka, Télécom Paris, Institut Polytechnique de Paris, France; Umut Şimşekli, INRIA; Télécom Paris, Institut Polytechnique de Paris, France; Gaël Richard, Télécom Paris, Institut Polytechnique de Paris, France |
Session | MLSP-23: Applications in Music and Audio Processing |
Location | Gather.Town |
Session Time: | Wednesday, 09 June, 16:30 - 17:15 |
Presentation Time: | Wednesday, 09 June, 16:30 - 17:15 |
Presentation |
Poster
|
Topic |
Machine Learning for Signal Processing: [MLR-MUSAP] Applications in music and audio processing |
Virtual Presentation |
Click here to watch in the Virtual Conference |
Abstract |
Style transfer is the process of changing the style of an image, video, audio clip or musical piece so as to match the style of a given example. Even though the task has interesting practical applications within the music industry, it has so far received little attention from the audio and music processing community. In this article, we present Groove2Groove, a one-shot style transfer method for symbolic music, focusing on the case of accompaniment styles in popular music and jazz. We propose an encoder-decoder neural network for the task, along with a synthetic data generation scheme to supply it with parallel training examples. This synthetic parallel data allows us to tackle the style transfer problem using end-to-end supervised learning, employing powerful techniques used in natural language processing. We experimentally demonstrate the performance of the model on style transfer using existing and newly proposed metrics, and also explore the possibility of style interpolation. |