STOCHASTIC PROXIMAL LANGEVIN ALGORITHM Adil Salim Joint work with - - PowerPoint PPT Presentation

stochastic proximal langevin algorithm
SMART_READER_LITE
LIVE PREVIEW

STOCHASTIC PROXIMAL LANGEVIN ALGORITHM Adil Salim Joint work with - - PowerPoint PPT Presentation

STOCHASTIC PROXIMAL LANGEVIN ALGORITHM Adil Salim Joint work with Dmitry Kovalev and Peter Richtrik 1 SAMPLING PROBLEM (d x ) exp( U ( x ))d x , U : d convex where . 2 LANGEVIN MONTE CARLO (LMC) W k Assume


slide-1
SLIDE 1

1

STOCHASTIC PROXIMAL LANGEVIN ALGORITHM

Adil Salim Joint work with Dmitry Kovalev and Peter Richtárik

slide-2
SLIDE 2

2

where . U : ℝd → ℝ convex

SAMPLING PROBLEM

μ⋆(dx) ∝ exp(−U(x))dx,

slide-3
SLIDE 3

Typical non asymptotic result: . KL(μk|μ⋆) = 𝒫(1/

k)

3

LANGEVIN MONTE CARLO (LMC)

Assume smooth, i.i.d standard gaussian and ,

U Wk γ > 0

xk+1 = xk − γ∇U(xk) + 2γWk+1 .

Gaussian noise Gradient descent

slide-4
SLIDE 4

4

FIRST INTUITION FOR LMC

dXt = − ∇U(Xt)dt + 2dWt .

LMC can be seen as a Euler discretization of the Langevin equation: Non asymptotic results using this intuition in [Dalalyan 2017], [Durmus Moulines 2017].

slide-5
SLIDE 5

5

LMC can be seen as an (inexact) Gradient Descent for:

μ⋆ = argmin ∫ Udμ(x) + ∫ μ(x)log(μ(x))dx μ⋆ = argmin KL(μ|μ⋆) .

SECOND INTUITION FOR LMC

Non asymptotic results using this intuition (+ extensions of LMC beyond GD) in [Durmus et al. 2018], [Wibisono 2018], [Bernton 2018].

slide-6
SLIDE 6

6

CONTRIBUTION: STOCHASTIC PROXIMAL LANGEVIN

xk+1 = proxγg(⋅,ξk+1)(xk) +

2γWk+1 .

Stochastic Prox

Case 1: U(x) = Eξ(g(x, ξ))

Nonsmooth

slide-7
SLIDE 7

Case 2:

U(x) = Eξ(f(x, ξ)) + ∑ Eξ(gi(x, ξ))

7

CONTRIBUTION: STOCHASTIC PROXIMAL LANGEVIN

i

Smooth Nonsmooth

See our Poster #161.

slide-8
SLIDE 8

8

STOCHASTIC SUBGRADIENT VS STOCHASTIC PROX

Stochastic subgradients [Durmus et al. 2018] Sampling .

μ⋆(dx) ∝ exp( − |x|)dx

Stochastic proximal [Us]

slide-9
SLIDE 9

9

Thanks for your attention. See us at poster #161.