A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2019; you can also visit the original URL.
The file type is application/pdf
.
Monoaural Audio Source Separation Using Variational Autoencoders
2018
Interspeech 2018
We introduce a monaural audio source separation framework using a latent generative model. Traditionally, discriminative training for source separation is proposed using deep neural networks or non-negative matrix factorization. In this paper, we propose a principled generative approach using variational autoencoders (VAE) for audio source separation. VAE computes efficient Bayesian inference which leads to a continuous latent representation of the input data(spectrogram). It contains a
doi:10.21437/interspeech.2018-1140
dblp:conf/interspeech/PandeyKN18
fatcat:5js7izrmcjdvbcoxo5px7o2x5i