LUC HENDRIKS RADBOUD UNIVERSITY, NIJMEGEN (NL) - - PowerPoint PPT Presentation

luc hendriks radboud university nijmegen nl
SMART_READER_LITE
LIVE PREVIEW

LUC HENDRIKS RADBOUD UNIVERSITY, NIJMEGEN (NL) - - PowerPoint PPT Presentation

iDark 1 The intelligent dark matter survey VARIATIONAL AUTOENCODERS LUC HENDRIKS RADBOUD UNIVERSITY, NIJMEGEN (NL) VARIATIONAL AUTOENCODERS 2 Conceptual talk about VAEs VAEs as a tool to


slide-1
SLIDE 1

LUC HENDRIKS


RADBOUD UNIVERSITY, NIJMEGEN (NL)

VARIATIONAL AUTOENCODERS

1

iDark

The intelligent dark
 matter survey

slide-2
SLIDE 2

VARIATIONAL AUTOENCODERS

▸ Conceptual talk about VAEs ▸ VAEs as a tool to do: ▸ Anomaly / outlier detection ▸ Noise reduction ▸ Generative modelling ▸ Event generation with a density buffer (Sydney’s talk)



 
 
 
 
 
 


2

slide-3
SLIDE 3

VARIATIONAL AUTOENCODERS

▸ Conceptual talk about VAEs ▸ VAEs as a tool to do: ▸ Anomaly / outlier detection ▸ Noise reduction ▸ Generative modelling ▸ Event generation with a density buffer (Sydney’s talk) ▸ Topics ▸ Normal AEs ▸ The concept of latent spaces ▸ VAEs ▸ β-VAEs

3

slide-4
SLIDE 4

AUTOENCODERS

4

▸ Class of deep 


learning algorithms

▸ Output = input ▸ Unsupervised learning 


(no labels needed)

slide-5
SLIDE 5

AUTOENCODERS

5

▸ Class of deep 


learning algorithms

▸ Output = input ▸ Unsupervised learning 


(no labels needed)

slide-6
SLIDE 6

AUTOENCODERS

6

▸ Class of deep 


learning algorithms

▸ Output = input ▸ Unsupervised learning 


(no labels needed)

slide-7
SLIDE 7

AUTOENCODERS

7

▸ Reconstruction very good —> compression algorithm ▸ Noise reduction ▸ Outlier detection: ▸ Put in something that the AE never saw —> bad

reconstruction

▸ Reconstruction loss = variable for outlier detection

slide-8
SLIDE 8

AUTOENCODERS

8

▸ Outlier: credit card fraud detection

Fraud No fraud Reconstruction loss Reconstruction loss

slide-9
SLIDE 9

AUTOENCODERS

9

▸ Outlier: credit card fraud detection

Fraud No fraud Reconstruction loss Reconstruction loss

▸ Noise reduction: MNIST noisy

slide-10
SLIDE 10

AUTOENCODERS

10

▸ No ordering in 


latent space

Assume 2D
 easy viz.

slide-11
SLIDE 11

AUTOENCODERS

11

▸ No ordering in 


latent space

Assume 2D
 easy viz. Latent dim 1 Latent dim 2

slide-12
SLIDE 12

AUTOENCODERS

12

▸ Input slightly different


than training set —>
 reconstruction loss high, because
 latent space is ill-defined there

▸ Not robust ▸ What is between 


the data points?

? ?

slide-13
SLIDE 13

AUTOENCODERS

13

▸ If only the points could be grouped together… ▸ Unsupervised clustering, interpolation between data

points …

2

slide-14
SLIDE 14

VARIATIONAL AUTOENCODERS

14

slide-15
SLIDE 15

VAE

▸ Force ordering in latent space ▸ During training, you are

minimising some loss function

▸ For regression (normal AE):


MSE(output - input)

15

slide-16
SLIDE 16

VAE

▸ Force ordering in latent space ▸ During training, you are 


minimising some loss function

▸ For regression (normal AE):


MSE(output - input)


▸ Add KL-divergence term: 


Σi KL(𝓞(μi, σi), 𝓞(0,1)) := KL(μ,σ)

▸ So 𝓜 = MSE(output - input) + KL(μ,σ)

16

slide-17
SLIDE 17

VAE

▸ The KL divergence punishes latent space values far away

from the center

▸ Also, every point has a variance that is pushed to 1 ▸ Balance MSE and KL —> group 


similar structures around the 
 center while keeping RL in check

17

slide-18
SLIDE 18

LATENT SPACE

▸ Same example, but now a VAE

18

slide-19
SLIDE 19

VAE

▸ Balancing MSE and KL is tricky ▸ Balance using another hyperparameter β ▸ 𝓜 = (1-β) * MSE(output - input) + β * KL(μ, σ) ▸ β-VAE

19

β Avg var Avg mean 1 1 1.89E-09 5E-01 0.99999905 2.35E-07 5E-02 0.86448085 … 5E-03 0.554529 5E-04 0.3784553 5E-05 0.09676677 5E-06 0.008932933 0.0000442

slide-20
SLIDE 20

VAE

▸ Use the latent space and decoder as generative model\ ▸ Explore the latent space!

20

PCA on the
 latent variables

slide-21
SLIDE 21

PLAYING WITH LATENT SPACES

▸ Train VAE on face images ▸ Change the latent space variables

21

slide-22
SLIDE 22

PLAYING WITH LATENT SPACES

▸ Or 3D objects


22

slide-23
SLIDE 23

PLAYING WITH LATENT SPACES

▸ Or 3D objects



 
 
 


▸ Latent space = abstract representation of your data ▸ Encoder maps input to gaussians in latent space


= Gaussian mixture —> you can do lots of stuff

23

slide-24
SLIDE 24

CONCLUSION

▸ VAEs can be used for ▸ Outlier / anomaly detection ▸ Noise reduction ▸ Generative modelling ▸ Data compression ▸ Exploration of latent space can give very interesting

applications — event generation, hybrid models, density estimation, …

24

Teaser :)