InverSynth: Deep Estimation of Synthesizer Parameter Configurations from Audio Signals

Oren Barkan, David Tsiris, Ori Katz, Noam Koenigstein

Research output: Contribution to journalArticlepeer-review

Abstract

Sound synthesis is a complex field that requires domain expertise. Manual tuning of synthesizer parameters to match a specific sound can be an exhaustive task, even for experienced sound engineers. In this paper, we introduce InverSynth-an automatic method for synthesizer parameters tuning to match a given input sound. InverSynth is based on strided convolutional neural networks and is capable of inferring the synthesizer parameters configuration from the input spectrogram and even from the raw audio. The effectiveness InverSynth is demonstrated on a subtractive synthesizer with four frequency modulated oscillators, envelope generator and a gater effect. We present extensive quantitative and qualitative results that showcase the superiority InverSynth over several baselines. Furthermore, we show that the network depth is an important factor that contributes to the prediction accuracy.

Original languageEnglish
Article number8854832
Pages (from-to)2385-2396
Number of pages12
JournalIEEE/ACM Transactions on Audio Speech and Language Processing
Volume27
Issue number12
DOIs
StatePublished - Dec 2019
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2014 IEEE.

Keywords

  • Deep synthesizer parameter estimation
  • InverSynth
  • automatic sound synthesis
  • inverse problems

Fingerprint

Dive into the research topics of 'InverSynth: Deep Estimation of Synthesizer Parameter Configurations from Audio Signals'. Together they form a unique fingerprint.

Cite this