Show HN: Cloning a musical instrument from 16 seconds of audio
In 2020, Magenta released DDSP [1], a machine learning algorithm / python library which made it possible to generate good sounding instrument synthesizers from about 6-10 minutes of data. While working with DDSP for a project, we realised how it was actually quite hard to find 6-10 minute of clean recordings of monophonic instruments.
In this project, we have combined the DDSP architecture with a domain adaptation technique from speech synthesis [2]. This domain adaptation technique works by pre-training our model on many different recordings from the Solos dataset [3] first and then fine-tuning parts of the model to the new recording. This allows us to produce decent sounding instrument synthesisers from as little as 16 seconds of target audio instead of 6-10 minutes.
[1] https://arxiv.org/abs/2001.04643
[2] https://arxiv.org/abs/1802.06006
[3] https://arxiv.org/abs/2006.07931
We hope to publish a paper on the topic soon.
Comments URL: https://news.ycombinator.com/item?id=30467328
Points: 19
# Comments: 1
from Hacker News: Front Page https://ift.tt/YlSfHM2
via
Comments
Post a Comment