Introduction to Sum-of-Squares Ankur Moitra (MIT) Robust Statistics - - PowerPoint PPT Presentation

introduction to sum of squares
SMART_READER_LITE
LIVE PREVIEW

Introduction to Sum-of-Squares Ankur Moitra (MIT) Robust Statistics - - PowerPoint PPT Presentation

Introduction to Sum-of-Squares Ankur Moitra (MIT) Robust Statistics Summer School A CLASSIC HARD PROBLEM: MAXCUT Goal: given a graph : find a cut that maximizes the number of crossing edges A CLASSIC HARD PROBLEM:


slide-1
SLIDE 1

Introduction to Sum-of-Squares

Ankur Moitra (MIT)

Robust Statistics Summer School

slide-2
SLIDE 2

A CLASSIC HARD PROBLEM: MAXCUT

Goal: given a graph : find a cut that maximizes the number of crossing edges

slide-3
SLIDE 3

A CLASSIC HARD PROBLEM: MAXCUT

Goal: given a graph : find a cut that maximizes the number of crossing edges NP-hard to maximize exactly, one of [Karp, ‘72]‘s 21 problems

slide-4
SLIDE 4

A CLASSIC HARD PROBLEM: MAXCUT

Goal: given a graph : find a cut that maximizes the number of crossing edges NP-hard to maximize exactly, one of [Karp, ‘72]‘s 21 problems How well can we approximate MAXCUT?

slide-5
SLIDE 5

A CLASSIC HARD PROBLEM: MAXCUT

Goal: given a graph : find a cut that maximizes the number of crossing edges NP-hard to maximize exactly, one of [Karp, ‘72]‘s 21 problems How well can we approximate MAXCUT? Simple ½-approximation algorithm: Choose U randomly.

slide-6
SLIDE 6

A CLASSIC HARD PROBLEM: MAXCUT

Goal: given a graph : find a cut that maximizes the number of crossing edges NP-hard to maximize exactly, one of [Karp, ‘72]‘s 21 problems How well can we approximate MAXCUT? Simple ½-approximation algorithm: Choose U randomly. But can we do better?

slide-7
SLIDE 7

MAXCUT AS A QUADRATIC PROGRAM

Alternatively we can write

slide-8
SLIDE 8

MAXCUT AS A QUADRATIC PROGRAM

Alternatively we can write xi’s are 0/1 valued

slide-9
SLIDE 9

MAXCUT AS A QUADRATIC PROGRAM

Alternatively we can write counts the number of edges crossing the cut xi’s are 0/1 valued

slide-10
SLIDE 10

MAXCUT AS A QUADRATIC PROGRAM

Alternatively we can write counts the number of edges crossing the cut xi’s are 0/1 valued Now we can leverage the Sum-of-Squares (SOS) Hierarchy…

slide-11
SLIDE 11

MAXCUT AS A QUADRATIC PROGRAM

Alternatively we can write counts the number of edges crossing the cut xi’s are 0/1 valued Now we can leverage the Sum-of-Squares (SOS) Hierarchy… We will utilize an alternative view based on the notion of a pseudo-expectation…

slide-12
SLIDE 12

AN ALTERNATIVE VIEW OF SOS

Pseudo-expectation [informally]: An operator that behaves like an expectation over a distribution on solutions degree ≤ d polynomials in n variables

slide-13
SLIDE 13

AN ALTERNATIVE VIEW OF SOS

Pseudo-expectation [informally]: An operator that behaves like an expectation over a distribution on solutions degree ≤ d polynomials in n variables This formulation is the starting point for state-of-the-art algorithms for quantum separability, tensor completion, tensor PCA, finding a planted sparse vector in a subspace, the best separable state problem, …

slide-14
SLIDE 14

AN ALTERNATIVE VIEW OF SOS

Pseudo-expectation [informally]: An operator that behaves like an expectation over a distribution on solutions degree ≤ d polynomials in n variables Let’s see what it looks like for MAXCUT… This formulation is the starting point for state-of-the-art algorithms for quantum separability, tensor completion, tensor PCA, finding a planted sparse vector in a subspace, the best separable state problem, …

slide-15
SLIDE 15

such that: (1) (2) is linear (3) for all deg(p) ≤ d/2 Degree d relaxation for MAXCUT: (4) for all deg(p) ≤ d-2

slide-16
SLIDE 16

such that: (1) (2) is linear (3) for all deg(p) ≤ d/2 Degree d relaxation for MAXCUT: (1) – (3) are the usual constraints that say Ẽ behaves like it is taking the expectation under some distribution on assignments to the variables (4) for all deg(p) ≤ d-2

slide-17
SLIDE 17

such that: (1) (2) is linear (3) for all deg(p) ≤ d/2 Degree d relaxation for MAXCUT: (1) – (3) are the usual constraints that say Ẽ behaves like it is taking the expectation under some distribution on assignments to the variables (4) for all deg(p) ≤ d-2 (4) is because we want the distribution to be supported on 0/1 valued assignments

slide-18
SLIDE 18

such that: (1) (2) is linear (3) for all deg(p) ≤ d/2 Degree d relaxation for MAXCUT: (4) for all deg(p) ≤ d-2 But why is this a relaxation for MAXCUT?

slide-19
SLIDE 19

such that: (1) (2) is linear (3) for all deg(p) ≤ d/2 Claim: If there is a cut that has at least k edges crossing, there is a feasible solution to (1) – (4) with objective value ≥ k (4) for all deg(p) ≤ d-2 Degree d relaxation for MAXCUT:

slide-20
SLIDE 20

such that: (1) (2) is linear (3) for all deg(p) ≤ d/2 Claim: If there is a cut that has at least k edges crossing, there is a feasible solution to (1) – (4) with objective value ≥ k Proof: if a1, a2, …, an is the indicator vector of the cut U, set (4) for all deg(p) ≤ d-2 Degree d relaxation for MAXCUT:

slide-21
SLIDE 21

Can we efficiently solve this relaxation?

slide-22
SLIDE 22

Can we efficiently solve this relaxation? Theorem: There is an nO(d)-time algorithm for finding such an

  • perator, if it exists
slide-23
SLIDE 23

Can we efficiently solve this relaxation? Theorem: There is an nO(d)-time algorithm for finding such an

  • perator, if it exists

It is a semidefinite program on a nO(d) x nO(d) matrix whose entries are the pseudo-expectation applied to monomials

slide-24
SLIDE 24

Can we efficiently solve this relaxation? Theorem: There is an nO(d)-time algorithm for finding such an

  • perator, if it exists

It is a semidefinite program on a nO(d) x nO(d) matrix whose entries are the pseudo-expectation applied to monomials How well does SOS approximate MAXCUT?

slide-25
SLIDE 25

APPROXIMATION ALGORITHMS FOR MAXCUT

Revolutionary work of [Goemans, Williamson]: Theorem: There is a -approximation algorithm for for MAXCUT

slide-26
SLIDE 26

APPROXIMATION ALGORITHMS FOR MAXCUT

Revolutionary work of [Goemans, Williamson]: Theorem: There is a -approximation algorithm for for MAXCUT We will give an alternate proof by rounding the degree two Sum-of-Squares relaxation

slide-27
SLIDE 27

Main Question: How do you round a pseudo-expectation to find a cut? I.e. if I give you how do you find a cut with at least edges crossing (in expectation)?

slide-28
SLIDE 28

Main Idea: Use a sample from a Gaussian distribution whose moments match the pseudo-moments Main Question: How do you round a pseudo-expectation to find a cut? I.e. if I give you how do you find a cut with at least edges crossing (in expectation)?

slide-29
SLIDE 29

Main Idea: Use a sample from a Gaussian distribution whose moments match the pseudo-moments Main Question: How do you round a pseudo-expectation to find a cut? I.e. if I give you how do you find a cut with at least edges crossing (in expectation)? Aside: Rounding higher degree relaxations is much harder b/c you cannot necc. find a r.v. whose moments match the pseudo-moments

slide-30
SLIDE 30

Claim: Without loss of generality, can assume for all i

slide-31
SLIDE 31

Claim: Without loss of generality, can assume for all i Intuition: You can always change U to V\U without changing the value of the cut, so WLOG xi has probability 1/2 of being in U

slide-32
SLIDE 32

GAUSSIAN ROUNDING

Let y be a Gaussian vector with mean and covariance for and

slide-33
SLIDE 33

GAUSSIAN ROUNDING

Let y be a Gaussian vector with mean and covariance for and Now set if and otherwise

slide-34
SLIDE 34

GAUSSIAN ROUNDING

Let y be a Gaussian vector with mean and covariance for and Now set if and otherwise We will show that for each (i, j) we have which, by linearity of expectation, will complete the proof

slide-35
SLIDE 35

For each edge (i,j), calculate contribution to objective value:

slide-36
SLIDE 36

For each edge (i,j), calculate contribution to objective value:

slide-37
SLIDE 37

For each edge (i,j), calculate contribution to objective value: for

slide-38
SLIDE 38

For each edge (i,j), calculate contribution to objective value: for And its contribution to the expected number of edges crossing:

slide-39
SLIDE 39

For each edge (i,j), calculate contribution to objective value: for And its contribution to the expected number of edges crossing:

slide-40
SLIDE 40

For each edge (i,j), calculate contribution to objective value: for And its contribution to the expected number of edges crossing: and

slide-41
SLIDE 41

For each edge (i,j), calculate contribution to objective value: for And its contribution to the expected number of edges crossing: and Now we can compute: independent std Gaussians

slide-42
SLIDE 42

For each edge (i,j), calculate contribution to objective value: for And its contribution to the expected number of edges crossing: and Now we can compute: independent std Gaussians

slide-43
SLIDE 43

Putting it all together, we have for every edge (i, j): which completes the proof