Conditional gradient algorithms for machine learning Zaid Harchaoui - - PowerPoint PPT Presentation

conditional gradient algorithms for machine learning
SMART_READER_LITE
LIVE PREVIEW

Conditional gradient algorithms for machine learning Zaid Harchaoui - - PowerPoint PPT Presentation

Conditional gradient algorithms for machine learning Zaid Harchaoui LEAR and LJK, INRIA Joint work with A. Juditsky (Grenoble U., France) and A. Nemirovski (GeorgiaTech) and Matthijs Douze, Miro Dudik, Jerome Malick, Mattis Paulin Gargantua


slide-1
SLIDE 1

Conditional gradient algorithms for machine learning

Zaid Harchaoui

LEAR and LJK, INRIA Joint work with A. Juditsky (Grenoble U., France) and A. Nemirovski (GeorgiaTech) and Matthijs Douze, Miro Dudik, Jerome Malick, Mattis Paulin

Gargantua day, Grenoble

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

1 / 42

slide-2
SLIDE 2

The advent of large-scale datasets and “big learning”

From “The Promise and Perils of Benchmark Datasets and Challenges”, D. Forsyth, A. Efros, F.-F. Li, A. Torralba and A. Zisserman, Talk at “Frontiers of Computer Vision”

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

2 / 42

slide-3
SLIDE 3

Large-scale supervised learning

Large-scale supervised learning Let (x1, y1), . . . , (xn, yn) ∈ Rd × Y be i.i.d. labelled training data, and Remp(·) the empirical risk for any W ∈ Rd×k. Constrained formulation minimize Remp(W) subject to Ω(W) ≤ ρ Penalized formulation minimize λΩ(W) + Remp(W) Problem : minimize such objectives in the large-scale setting # examples ≫ 1, # features ≫ 1, # classes ≫ 1

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

3 / 42

slide-4
SLIDE 4

Large-scale supervised learning

Large-scale supervised learning Let (x1, y1), . . . , (xn, yn) ∈ Rd × Y be i.i.d. labelled training data, and Remp(·) the empirical risk for any W ∈ Rd×k. Constrained formulation minimize Remp(W) subject to Ω(W) ≤ ρ Penalized formulation minimize λΩ(W) + Remp(W) Problem : minimize such objectives in the large-scale setting n ≫ 1, d ≫ 1, k ≫ 1

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

4 / 42

slide-5
SLIDE 5

Machine learning cuboid

n d k

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

5 / 42

slide-6
SLIDE 6

Motivating example : multi-class classification with trace-norm penalty

Motivating the trace-norm penalty Embedding assumption : classes may embedded in a low-dimensional subspace of the feature space Computational efficiency : training time and test time efficiency require sparse matrix regularizers Trace-norm The trace-norm, aka nuclear norm, is defined as σ(W)1 =

min(d,k)

  • p=1

σp(W) where σ1(W), . . . , σmin(d,k)(W) denote the singular values of W.

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

6 / 42

slide-7
SLIDE 7

Large-scale supervised learning

Multi-class classification with trace-norm regularization Let (x1, y1), . . . , (xn, yn) ∈ Rd × Y be i.i.d. labelled training data, and Remp(·) the empirical risk for any W ∈ Rd×k. Constrained formulation minimize Remp(W) subject to σ(W)1 ≤ ρ Penalized formulation minimize λ σ(W)1 + Remp(W) Trace-norm reg. penalty (Amit et al., 2007 ; Argyriou et al., 2007) Enforces a low-rank structure of W (sparsity of spectrum σ(W)) Convex problems

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

7 / 42

slide-8
SLIDE 8

About the different formulations

“Alleged” equivalence For a particular set of examples, for any value ρ of the constraint in the constrained formulation, there exists a value of λ in the penalized formulation so that the solutions of resp. the constrained formulation and the penalized formulation coincide. Statistical learning theory theoretical results on penalized estimators and constrained estimators are of different nature → no rigorous comparison possible equivalence frequently called as the rescue depending on the theoretical tools available to jump from one formulation to the other

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

8 / 42

slide-9
SLIDE 9

Summary

In practice Recall that eventually “hyperparameters” (λ, ρ, ε, · · · ) will have to be tuned. Choose the formulation in which you can easily incorporate prior knowledge

Constrained formulation I Minimize

W∈Rd×k

  • 1

n

n

  • i=1

Lossi : σ(W)1 ≤ ρ

  • Penalized formulation

Minimize

W∈Rd×k

  • 1

n

n

  • i=1

Lossi + λ σ(W)1

  • Constrained formulation II

Minimize

W∈Rd×k

  • λ σ(W)1 :
  • 1

n

n

  • i=1

Lossi − Rtarget

emp

  • ≤ ε
  • Zaid Harchaoui (INRIA)

Conditional gradient algorithms

  • Nov. 26th, 2013

9 / 42

slide-10
SLIDE 10

Learning with trace-norm penalty : a convex problem

Supervised learning with trace-norm regularization penalty Let (x1, y1), . . . , (xn, yn) ∈ Rd × Y be a set of i.i.d. labelled training data, with Y = {0, 1}k for multi-class classification Minimize

W∈Rd×k

1 n

n

  • i=1

Lossi + λσ(W)1

  • convex

Penalized formulation Trace-norm reg. penalty (Amit et al., 2007 ; Argyriou et al., 2007) Enforces a low-rank structure of W (sparsity of spectrum σ(W)) Convex, but non-differentiable

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

10 / 42

slide-11
SLIDE 11

Possible approaches

Generic approaches “Blind” approach : subgradient, bundle method → slow convergence rate Other approaches : alternating optimization, iteratively reweighted least-squares, etc. → no finite-time convergence guarantees

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

11 / 42

slide-12
SLIDE 12

Learning with trace-norm penalty : convex but non-smooth

Supervised learning with trace-norm regularization penalty Let (x1, y1), . . . , (xn, yn) ∈ Rd × Y be a set of i.i.d. labelled training data, with Y = {0, 1}k for multi-class classification Minimize

W∈Rd×k

λ σ(W)1

  • nonsmooth

+ 1 n

n

  • i=1

Lossi

  • smooth

where Lossi is e.g. the multinomial logistic loss of i-th example Lossi = log  1 +

  • ℓ∈Y\{yi}

exp

  • wT

ℓ xi − wT y xi

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

12 / 42

slide-13
SLIDE 13

Learning with trace-norm penalty : a convex problem

Supervised learning with trace-norm regularization penalty Let (x1, y1), . . . , (xn, yn) ∈ Rd × Y be a set of i.i.d. labelled training data, with Y = {0, 1}k for multi-class classification Minimize

W∈Rd×k

λσ(W)1 + 1 n

n

  • i=1

Lossi Penalized formulation

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

13 / 42

slide-14
SLIDE 14

Composite minimization for penalized formulation

Strengths of composite minimization (aka proximal-gradient) Attractive algorithms when proximal operator is cheap, as e.g. for vector ℓ1-norm Accurate with medium-accuracy, finite-time accuracy guarantees

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

14 / 42

slide-15
SLIDE 15

Proximal gradient

Algorithm Initialize : W = 0 Iterate : Wt+1 = Proxλ/LΩ(·)

  • Wt − 1

L∇Remp(Wt)

  • with Proxλ/LΩ(·)(U) := min

W

1 2U − W2 + λ LΩ(W)

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

15 / 42

slide-16
SLIDE 16

Composite minimization for penalized formulation

Strengths of composite minimization (aka proximal-gradient) Attractive algorithms when proximal operator is cheap, as e.g. for vector ℓ1-norm Accurate with medium-accuracy, finite-time accuracy guarantees Weaknesses of composite minimization Inappropriate when proximal operator is expensive to compute Too sensitive to conditioning of design matrix (correlated features) Situation with trace-norm, i.e. ProxµΩ(·)(·) with Ω(·) = · σ,1 proximal operator corresponds to singular value thresholding, requiring an SVD running in O(krk(W)2) in time → impractical for large-scale problems

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

16 / 42

slide-17
SLIDE 17

Alternative approach : conditional gradient

We want an algorithm with no SVD, i.e. without any projection or proximal

  • step. Let us get some inspiration from the constrained setting.

Problem Minimize

W∈Rd×k

  • 1

n

n

  • i=1

Lossi : W ∈ ρ · convex hull ({Mt}t≥1)

  • Gauge/atomic decomposition of trace-norm

σ(W)1 = inf

θ

N

  • i=1

θi | ∃N, θi > 0, Mi ∈ M with W =

N

  • i=1

θiMi

  • M = {uvT | u ∈ Rd, v ∈ RY, u2 = v2 = 1}

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

17 / 42

slide-18
SLIDE 18

Conditional gradient descent

Algorithm Initialize : W = 0 Iterate : Find Mt ∈ ρ · convex hull (M) , such that Mt = Arg max

Mℓ∈M

Mℓ, −∇Remp(Wt)

  • linear min. oracle

Perform line-search between Wt and Mt Wt+1 = (1 − δ)Wt + δMt

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

18 / 42

slide-19
SLIDE 19

Conditional gradient descent : example with trace-norm constraint

Algorithm Initialize : W = 0 Iterate : Find Mt ∈ ρ · convex hull (M) such that Mt = Arg max

uℓvT

ℓ , −∇Remp(Wt)

= Arg max

u2=v2=1

uT (−∇Remp(Wt))v i.e. compute top pair of singular vectors of −∇Remp(Wt). Perform line-search between Wt and Mt Wt+1 = (1 − δ)Wt + δMt

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

19 / 42

slide-20
SLIDE 20

Conditional gradient descent

Algorithm Initialize : W = 0 Iterate : Find Mt ∈ ρ · convex hull (M) such that Mt = Arg max

Mℓ∈M

Mℓ, −∇Remp(Wt)

  • easy

Perform line-search between Wt and Mt Wt+1 = (1 − δ)Wt + δMt

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

20 / 42

slide-21
SLIDE 21

Finite-time guarantee

Assumptions (A) [Smoothness] The empirical risk Remp(·) is convex continuously differentiable on D = ρ · conv(M), with Lipschitz constant L w.r.t D Let {Wt} be a sequence generated by the conditional gradient algorithm. Then F(Wt) − F ⋆ ≤ 2L t + 1, t = 1, 2, . . .

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

21 / 42

slide-22
SLIDE 22

Conditional gradient algorithm : review

Conditional gradient for constrained programming aka the Frank-Wolfe algorithm (1956, originally for quadratic programming) convergence results in general Banach spaces in (Demyanov & Rubinov, 1970) finite-time guarantees in (Pshenichnyi, 1975 ; Dunn, 1979) superseded by sequential quadratic programming in the early 80s, and ended up in the “mathematical programming” attic rediscovered several times and revisited with new variants in machine learning ; lately, (Hazan, 2008 ; Jaggi & Sulovsky, 2010 ; Tewari et al., 2011 ; Bach et al., 2012) See (HJN, 2013) and (Jaggi, 2013) for modern proofs.

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

22 / 42

slide-23
SLIDE 23

Conditional gradient algorithms

Question is it possible to design a conditional-gradient-type algorithm for penalized formulations ?

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

23 / 42

slide-24
SLIDE 24

Conditional gradient vs Proximal gradient

Conditional gradient : iteration Wt+1 = (1 − δ)Wt + δMt Mt = Arg max

Mℓ∈M

Mℓ, −∇Remp(Wt)

  • easy

Proximal gradient : iteration Wt+1 = Proxλ/LΩ(·) (Wt − 1/L∇Remp(Wt)) Proxλ/LΩ(·)(U) := min

W

1 2U − W2 + λ LΩ(W)

  • hard

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

24 / 42

slide-25
SLIDE 25

Conditional gradient approach for penalized formulations

Let K ⊂ E a closed convex cone, E a euclidean space, and · a norm on E. Problem Minimize

W∈K

λW + 1 n

n

  • i=1

Lossi(W) Penalized formulation Sketch Augment the variable W by one dimension to handle the regularization penalty Perform a sequence of iterations akin to the conditional gradient iterations and so on...

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

25 / 42

slide-26
SLIDE 26

Turning the problem into a cone constrained problem

Problem Introducing the variable Z := [W, r], we get minimize F(Z) subject to Z ∈ K+ where F(Z) := λr + 1 n

n

  • i=1

Lossi(W) K+ := {[W; r], W ∈ K, W ≤ r} .

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

26 / 42

slide-27
SLIDE 27

linear minimization oracle

First-order information and linear minimization oracle For any W, we can get Remp(W) the empirical risk ∇Remp(W) the gradient of the empirical risk For any g ∈ E∗ we have access to a linear minimization oracle Oracle(g) := Arg max

W∈K1

W, −g . where K1 := {W ∈ K, W ≤ 1} .

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

27 / 42

slide-28
SLIDE 28

Linear minimization oracle

First-order information and linear minimization oracle For any W, we can get Remp(W) the empirical risk ∇Remp(W) the derivative of the empirical risk and any iteration t we have access to a linear minimization oracle Oracle(g) := Arg max

W∈K1

W, −g . where K1 := {W ∈ K, W ≤ 1} .

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

28 / 42

slide-29
SLIDE 29

Conditional gradient for penalized formulation

Algorithm Inputs : instrumental bound D+ on W⋆, first-order oracle, and

  • minim. oracle

Iterate : Compute ∇Remp(Wt) at Zt = (Wt, rt) Call the linear minimization oracle Oracle(∇Remp(Wt)) := Arg max

W∈K1

W, −∇Remp(Wt)

  • linear minimization oracle

. ... The instrumental bound D+ can be loose.

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

29 / 42

slide-30
SLIDE 30

Conditional gradient for penalized formulation

Algorithm Inputs : instrumental bound D+ on W⋆, first-order oracle, and

  • minim. oracle

Iterate : Compute ∇Remp(Wt) at Zt = (Wt, rt) Get ¯ Zt = [Oracle(∇Remp(Wt)), 1] from the linear minimization oracle. Perform line-search to get Zt+1 ∈ argminZ

  • F(Z), Z ∈ Conv{0, Zt, D+ ¯

Zt}

  • .

The instrumental bound D+ can be loose.

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

30 / 42

slide-31
SLIDE 31

Conditional gradient for penalized formulation

Algorithm Inputs : instrumental bound D+ on W⋆, first-order oracle, and

  • minim. oracle

Iterate : Compute ∇Remp(Wt) at Zt = (Wt, rt) Get ¯ Zt = [Oracle(∇Remp(Wt)), 1] from the linear minimization oracle. Perform line-search to get Zt+1 = αt+1 ¯ Zt + βt+1Zt (αt+1, βt+1) = Arg min

α,β

{F(α ¯ Zt + βZt), α + β ≤ 1, α ≥ 0, β ≥ 0 } . Output : WT can be retrieved from ZT = [WT , rT ].

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

31 / 42

slide-32
SLIDE 32

Memory-based extensions : convex-hull

Convex-hull memory-based extension (“restricted simplicial acceleration”) Instead to the 2D line-search, we can perform at each iteration for some M > 0 Zt+1 ∈ Arg min

Z

{F(Z), Z ∈ Ct} . where Ct = Conv{0; D+ ¯ Z0, ..., D+ ¯ Zt}, t ≤ M , Conv{0; Zt−M+1, ..., Zt; D+ ¯ Zt−M+1, ..., D+ ¯ Zt}, t > M . Important computational considerations Line-search sub-problem can be solved with ellipsoid algorithm Maintaining the factorization of W along iterations is essential for speed

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

32 / 42

slide-33
SLIDE 33

Memory-based extensions : conic-hull

Conic-hull memory-based extension Instead to the 2D line-search, we can perform at each iteration for some M > 0 Zt+1 ∈ Arg min

Z

{F(Z), Z ∈ Bt} . where Bt = Conic{ ¯ Z0, ..., ¯ Zt}, t ≤ M , Conic{Zt−M+1, ..., Zt; ¯ Zt−M+1, ..., ¯ Zt}, t > M . M = +∞ : we recover the Atom-Descent algorithm of (DHM, 2012) Important computational considerations Line-search sub-problem can be solved with coordinate-descent Maintaining factorization of W along iterations essential for speed

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

33 / 42

slide-34
SLIDE 34

Finite-time guarantee

Assumptions (A) [Smoothness] The empirical risk Remp(·) is convex continuously differentiable with Lipschitz constant L. (B) [Effective domain] There exists D < 1 such that W ≤ r and r + Remp(W) < Remp(0) imply that r ≤ D Let {Zt} be a sequence generated by the algorithm. Then F(Zt) − F ⋆ ≤ 8LD2 t + 1 , t = 2, 3, . . .

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

34 / 42

slide-35
SLIDE 35

Finite-time guarantee

Finite-time guarantee Let {Zt} be a sequence generated by the algorithm. Then F(Zt) − F ⋆ ≤ 8LD2 t + 1 , t = 2, 3, . . . Important remark The O(1/t) convergence rate depends on D (unknown and not required by the algorithm), but does not depend on D+ ! (known and required by the algorithm).

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

35 / 42

slide-36
SLIDE 36

Finite-time guarantee

Finite-time guarantee Let {Zt} be a sequence generated by the algorithm. Then F(Zt) − F ⋆ ≤ 8LD2 t + 1 , t = 2, 3, . . . Theoretical convergence rate is independent of D+.

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

36 / 42

slide-37
SLIDE 37

Generalization to gauge regularization penalty

Gauge regularization penalty Gauge definition : Ω(W) := inf{t ≥ 0 | W ∈ tB} Unit “ball” : B := conv M Atoms set : M = {Mi ∈ Rd×k : i ∈ I} be a compact set of matrices, called atoms → “overcomplete basis”

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

37 / 42

slide-38
SLIDE 38

Generalization to gauge regularization penalty

Properties Ω(tW) = tΩ(W) for all W and t ≥ 0 Ω(W + W′) ≤ Ω(W) + Ω(W′) for all W and W′. Additional properties Assuming 0 ∈ int B, we also have Ω(W) ≥ 0, with equality if and only if W = 0 {W : Ω(W) ≤ t} = tB for t ≥ 0, i.e., level sets are compact. Polar duality Support function : Ω◦(G) := supM∈BM, G = supM∈MM, G.

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

38 / 42

slide-39
SLIDE 39

Some examples

Examples of gauges with their atomic decomposition

  • i,j

|Wi,j| Mlasso =

  • ±ejeT

ℓ | j ∈ {1, . . . , d}, ℓ ∈ {1, . . . , k}

  • i

Wi,: Mgp-lasso = {ejvT | j ∈ {1, . . . , d}, v ∈ Rk, v2 = 1}

  • p

σp(Wi,:) Mtr-norm = {uvT | u ∈ Rd, v ∈ Rk, u2 = v2 = 1}

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

39 / 42

slide-40
SLIDE 40

Conclusion and perspectives

Large-scale learning conditional gradient algorithm for learning problems with atomic-decomposition-norm regularization efficient and competitive algorithm for large-scale multi-class classification scheme applies to all problems with atomic decomposition norm regularizers (Harchaoui et al., 2011, Chandrasekaran et al., 2012) : nuclear-norm, total-variation norm, overlapping-blocks sparse norm, etc. Extensions non-smooth loss functions ; see (Pierucci et al., ICCOPT 2013)

  • nline/mini-batch extensions

path-following extensions

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

40 / 42

slide-41
SLIDE 41

References

References Atom-descent with smoothing for machine learning with non-smooth loss function, F. Pierucci, Z. Harchaoui, A. Juditsky, A. Nemirovski, ICCOPT 2013 Conditional gradient algorithms for norm-regularized smooth convex

  • ptimization, Z. Harchaoui, A. Juditsky, A. Nemirovski, sub. Math.
  • Prog. A, 2013

Large-scale classification with trace-norm regularization, Z. Harchaoui,

  • M. Douze, M. Paulin, J. Malick, CVPR 2012

Lifted coordinate descent for learning with trace-norm regularization penalty, M. Dudik, Z. Harchaoui, J. Malick, AISTATS 2011 Learning with matrix gauge regularizers, M. Dudik, Z. Harchaoui, J. Malick, NIPS Opt. 2011

Zaid Harchaoui (INRIA) Conditional gradient algorithms

  • Nov. 26th, 2013

41 / 42