Semi-Supervised Adversarial Audio Source Separation applied to - - PowerPoint PPT Presentation

semi supervised adversarial audio source separation
SMART_READER_LITE
LIVE PREVIEW

Semi-Supervised Adversarial Audio Source Separation applied to - - PowerPoint PPT Presentation

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Semi-Supervised Adversarial Audio Source Separation applied to Singing Voice Extraction Daniel Stoller 1 , Sebastian Ewert 2 , Simon


slide-1
SLIDE 1

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

Semi-Supervised Adversarial Audio Source Separation applied to Singing Voice Extraction

Daniel Stoller1, Sebastian Ewert2, Simon Dixon1

1Centre for Digital Music

Queen Mary University London

2Spotify

MLSP-L8: Deep Learning III ICASSP 19.04.2018

slide-2
SLIDE 2

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

Audio source separation

Task: Recover sources from mixtures Example: Music instrument separation:

slide-3
SLIDE 3

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

Current state of the art [5, 3, 1]

Training on multitrack datasets Neural network Discriminative, MSE loss

slide-4
SLIDE 4

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

Current state of the art [5, 3, 1]

Training on multitrack datasets (small ⇒ overfitting!) Neural network Discriminative, MSE loss

slide-5
SLIDE 5

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

Our goal

⇒ How to also learn from unpaired mixtures and sources? Random mixing ignores source correlations [4, 2]

slide-6
SLIDE 6

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Theoretical framework

Intuition

Mixture database Accompaniment estimates Magnitude spectrogram Unlabeled mixtures Magnitude spectrogram Vocal estimates Magnitude spectrogram Separator network

slide-7
SLIDE 7

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Theoretical framework

Intuition

Mixture database Accompaniment database Singing voice database Accompaniment estimates Magnitude spectrogram Unlabeled mixtures Magnitude spectrogram Vocal estimates Magnitude spectrogram Unlabeled vocals Magnitude spectrogram Unlabeled accompaniment Magnitude spectrogram Separator network

slide-8
SLIDE 8

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Theoretical framework

Intuition

slide-9
SLIDE 9

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Theoretical framework

Derivation of unsupervised loss

For optimal separator: qφ(sk|m) = p(sk|m)

slide-10
SLIDE 10

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Theoretical framework

Derivation of unsupervised loss

For optimal separator: qφ(sk|m) = p(sk|m) Em∼pdata qφ(sk|m) = Em∼pdata p(sk|m) Overall separator output = Source distribution

slide-11
SLIDE 11

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Theoretical framework

Derivation of unsupervised loss

For optimal separator: qφ(sk|m) = p(sk|m) Em∼pdata qφ(sk|m) = Em∼pdata p(sk|m)

  • utqk

φ

= pk

s

slide-12
SLIDE 12

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Theoretical framework

Derivation of unsupervised loss

For optimal separator: qφ(sk|m) = p(sk|m) Em∼pdata qφ(sk|m) = Em∼pdata p(sk|m)

  • utqk

φ

= pk

s

Necessary condition for optimal separator Loss: Minimise divergence between source outputs: Lu = K

k=1 D[outqk φ||pk s ]

slide-13
SLIDE 13

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Theoretical framework

Overall approach

Supervised loss: MSE between estimate and ground truth

slide-14
SLIDE 14

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Theoretical framework

Overall approach

Supervised loss: MSE between estimate and ground truth Unsupervised loss:

Lu = K

k=1 D[outqk φ||pk s ]

Ladd: MSE between sum of source estimates and mixture

slide-15
SLIDE 15

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Theoretical framework

Overall approach

Supervised loss: MSE between estimate and ground truth Unsupervised loss:

Lu = K

k=1 D[outqk φ||pk s ]

Ladd: MSE between sum of source estimates and mixture

Total loss: L = Ls + αLu + βLadd

slide-16
SLIDE 16

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Implementation using GANs

Divergence minimization with GANs

Discriminator estimates divergence D between generator and real distribution Generator minimises divergence D

slide-17
SLIDE 17

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary Implementation using GANs

Divergence minimization with GANs

Discriminator estimates divergence D between generator and real distribution Generator minimises divergence D Our separator is a conditional generator ⇒ We use one discriminator per source to estimate the Wasserstein distance W [outqk

φ||pk s ]

slide-18
SLIDE 18

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

Experimental setup

Avoids dataset bias Supervised and semi-supervised training with early stopping U-Net as separator, DCGAN as discriminator

slide-19
SLIDE 19

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

Results

Performance

Test set DSD100 MedleyDB CCMixter iKala 6 7 8 9 10 11 12

Mean accompaniment SDR

Baseline Ours Test set DSD100 MedleyDB CCMixter iKala 2 4 6 8 10 12

Mean vocal SDR

Baseline Ours

slide-20
SLIDE 20

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

Results

Qualitative

256 512 768 1024 0.5 1.0 1.5 2.0

f (Hz) t (s)

(a) Separator estimate x

256 512 768 1024 0.5 1.0 1.5 2.0

f (Hz) t (s)

(b) ∇xD(x)

slide-21
SLIDE 21

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

Results

Qualitative

256 512 768 1024 0.5 1.0 1.5 2.0

f (Hz) t (s)

(a) Separator estimate x

256 512 768 1024 0.5 1.0 1.5 2.0

f (Hz) t (s)

(b) ∇xD(x) ⇒ Discriminator appears to work More perceptual loss function?

slide-22
SLIDE 22

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

Summary

Current SotA methods only use multi-track data Our approach also uses solo source recordings Performance improvement in singing voice separation experiment More perceptual loss? (seeks posterior modes, not means)

slide-23
SLIDE 23

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

End

Code available at https://github.com/f90/AdversarialAudioSeparation Thank you for your attention!

slide-24
SLIDE 24

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

  • A. Jansson, E. J. Humphrey, N. Montecchio, R. Bittner,
  • A. Kumar, and T. Weyde.

Singing voice separation with deep U-Net convolutional networks. In Proceedings of the International Society for Music Information Retrieval Conference (ISMIR), pages 323–332, 2017.

  • M. Miron, J. Janer Mestres, and E. G´
  • mez Guti´

errez. Generating data to train convolutional neural networks for classical music source separation. In Proceedings of the 14th Sound and Music Computing

  • Conference. Aalto University, 2017.
  • A. A. Nugraha, A. Liutkus, and E. Vincent.

Multichannel audio source separation with deep neural networks. PhD thesis, Inria, 2015.

slide-25
SLIDE 25

Motivation State of the art Proposed approach Experiment: Singing voice separation Discussion and summary

  • S. Uhlich, F. Giron, and Y. Mitsufuji.

Deep neural network based instrument extraction from music. In 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 2135–2139. IEEE, 2015.

  • S. Uhlich, M. Porcu, F. Giron, M. Enenkl, T. Kemp,
  • N. Takahashi, and Y. Mitsufuji.

Improving music source separation based on deep neural networks through data augmentation and network blending. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 261–265, March 2017.