Amit Chakrabarti Dartmouth College Main result joint with Oded - - PowerPoint PPT Presentation

amit chakrabarti
SMART_READER_LITE
LIVE PREVIEW

Amit Chakrabarti Dartmouth College Main result joint with Oded - - PowerPoint PPT Presentation

Optimal Lower Bound for GHD May 2011 Gap-Hamming-Distance: The Journey to an Optimal Lower Bound Amit Chakrabarti Dartmouth College Main result joint with Oded Regev, Tel Aviv University Sublinear Algorithms Workshop at Bertinoro, May 2011


slide-1
SLIDE 1

Optimal Lower Bound for GHD May 2011

Gap-Hamming-Distance: The Journey to an Optimal Lower Bound

Amit Chakrabarti

Dartmouth College Main result joint with Oded Regev, Tel Aviv University Sublinear Algorithms Workshop at Bertinoro, May 2011

Amit Chakrabarti 1

slide-2
SLIDE 2

Optimal Lower Bound for GHD May 2011

The Gap-Hamming-Distance Problem

Input: Alice gets x ∈ {0, 1}n, Bob gets y ∈ {0, 1}n. Output:

  • ghd(x, y) = 1 if ∆(x, y) > n

2 + √n

  • ghd(x, y) = 0 if ∆(x, y) < n

2 − √n

Want: randomized, constant error protocol Cost: Worst case number of bits communicated

1 x = y = 1 1 1 1 1 1 1 1

n = 12; ∆(x, y) = 3 ∈ [6 − √ 12, 6 + √ 12]

Amit Chakrabarti 2

slide-3
SLIDE 3

Optimal Lower Bound for GHD May 2011

Implications

Data stream lower bounds

  • Distinct elements
  • Frequency moments
  • Norms
  • Entropy
  • General form of bound: ps = Ω(1/ε2)

Distributed functional monitoring lower bounds Connections to differential privacy

Amit Chakrabarti 3

slide-4
SLIDE 4

Optimal Lower Bound for GHD May 2011

The Reductions

E.g., Distinct Elements (Other problems: similar)

( 9 , )

y = 1 1 1 1

( 1 2 , 1 ) ( 1 1 , ) ( 1 , ) ( 1 2 , 1 ) ( 1 1 , ) ( 1 , )

x = 1 1 1 1 1

( 1 , ) ( 3 , ) ( 4 , ) ( 6 , ) ( 8 , 1 ) ( 7 , 1 ) ( 2 , ) ( 5 , ) ( 9 , 1 )

τ : σ :

( 1 , ) ( 3 , ) ( 4 , ) ( 2 , 1 ) ( 5 , 1 ) ( 6 , ) ( 8 , 1 ) ( 7 , 1 )

Alice: x − → σ = (1, x1), (2, x2), . . . , (n, xn) Bob: y − → τ = (1, y1), (2, y2), . . . , (n, yn) Notice: F0(σ ◦ τ) = n + ∆(x, y) =    < 3n

2 − √n, or

> 3n

2 + √n.

Set ε =

1 √n. Amit Chakrabarti 4

slide-5
SLIDE 5

Optimal Lower Bound for GHD May 2011

Ancient History

Amit Chakrabarti 5

slide-6
SLIDE 6

Optimal Lower Bound for GHD May 2011

One-Pass Bounds

Indyk, Woodruff [FOCS 2003]

  • Considered one-pass lower bound for dist-elem
  • Recognized relevance of ghd, difficulty of lower-bounding
  • Defined “related” problem Πℓ2, showed R→(Πℓ2) = Ω(n)
  • Concluded Ω(ε−2) bound for dist-elemm,ε with m =

Ω(1/ε9)

Amit Chakrabarti 6

slide-7
SLIDE 7

Optimal Lower Bound for GHD May 2011

One-Pass Bounds

Indyk, Woodruff [FOCS 2003]

  • Considered one-pass lower bound for dist-elem
  • Recognized relevance of ghd, difficulty of lower-bounding
  • Defined “related” problem Πℓ2, showed R→(Πℓ2) = Ω(n)
  • Concluded Ω(ε−2) bound for dist-elemm,ε with m =

Ω(1/ε9) Woodruff [SODA 2004]

  • Worked with ghd itself, showed R→(ghd) = Ω(n)
  • Very intricate combinatorial proof, with hairy probability estimations
  • Conjectured R(ghd) = Ω(n), implying multi-pass lower bounds

Amit Chakrabarti 6-a

slide-8
SLIDE 8

Optimal Lower Bound for GHD May 2011

The VC-Dimension Technique

  • Consider communication matrix of ghd as set system
  • The system has Ω(n) VC-dimension

1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1

  • Thus, R→(ghd) = Ω(n)

Amit Chakrabarti 7

slide-9
SLIDE 9

Optimal Lower Bound for GHD May 2011

The VC-Dimension Technique

  • Consider communication matrix of ghd as set system
  • The system has Ω(n) VC-dimension

1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 Instance of INDEX

Amit Chakrabarti 7

slide-10
SLIDE 10

Optimal Lower Bound for GHD May 2011

The VC-Dimension Technique

  • Consider communication matrix of ghd as set system
  • The system has Ω(n) VC-dimension

1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 Instance of INDEX

  • Thus, R→(ghd) = Ω(n)

Amit Chakrabarti 7-a

slide-11
SLIDE 11

Optimal Lower Bound for GHD May 2011

The Middle Ages

Amit Chakrabarti 8

slide-12
SLIDE 12

Optimal Lower Bound for GHD May 2011

A Nice Simplification

Jayram, Kumar, Sivakumar [circa 2005]

  • Simpler proof of R→(ghd) = Ω(n)
  • Much simpler: direct reduction from index
  • Geometric intuition:

Alice: x ∈ {0, 1}n − →

  • x ∈
  • 1

√n, − 1 √n

n ∈ Rn Bob: j ∈ [n] − → ej = (0, . . . , 0, 1, 0, . . . , 0) ∈ Rn

  • Observe:

x, ej ≈ 0, and xj determined by sgn x, ej

  • We’ve reduced index to “gap-inner-product”, or gip

Amit Chakrabarti 9

slide-13
SLIDE 13

Optimal Lower Bound for GHD May 2011

Inner Product ↔ Hamming Distance

  • Obviously, ghd → gip:
  • x,

y = 1 − 2∆(x, y) n

  • x,

y ≷ ∓ 2 √n ⇒ ∆(x, y) ≶ n 2 ± √n

  • Also, gip → ghd by “discretization transform”:

Pick random Gaussians r1, . . . , rN, with N = 10n Alice: ¯ x ∈ Rn − → x = (sgn¯ x, r1, . . . , sgn¯ x, rN) ∈ {±1}N Bob: ¯ y ∈ Rn − → y = (sgn¯ y, r1, . . . , sgn¯ y, rN) ∈ {±1}N ¯ x, ¯ y ≷ ∓ 1

√n whp

= ⇒ ∆(x, y) ≶ N

2 ± O(

√ N)

Amit Chakrabarti 10

slide-14
SLIDE 14

Optimal Lower Bound for GHD May 2011

The Renaissance Era

Amit Chakrabarti 11

slide-15
SLIDE 15

Optimal Lower Bound for GHD May 2011

Round Elimination

Brody, Chakrabarti [CCC 2009]

  • Can we at least rule out a two-pass improvement for dist-elem?
  • A cheap first message makes little progress? Then rinse, repeat
  • Tends to decimate problem [Miltersen-Nisan-Safra-Wigderson’98] [Sen’03]

Input: (k rounds)

Padding: (k−1 rounds) Input:

Amit Chakrabarti 12

slide-16
SLIDE 16

Optimal Lower Bound for GHD May 2011

Another VC-Dimension Argument: Subcube Lifting

First message constant on large set:

1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 }

2

points 0.99n

Alice, Bob lift their (n/3)-dim inputs from inner coords to full n-dim space First message now redundant, so eliminate!

Amit Chakrabarti 13

slide-17
SLIDE 17

Optimal Lower Bound for GHD May 2011

Another VC-Dimension Argument: Subcube Lifting

First message constant on large set:

1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 }

2

points 0.99n 1 1 1 1 1 1 1 inner coords, the real input (Rest: outer coords, padding) S:

Alice, Bob lift their (n/3)-dim inputs from inner coords to full n-dim space First message now redundant, so eliminate!

Amit Chakrabarti 13

slide-18
SLIDE 18

Optimal Lower Bound for GHD May 2011

Another VC-Dimension Argument: Subcube Lifting

First message constant on large set:

  • Amit Chakrabarti

13

slide-19
SLIDE 19

Optimal Lower Bound for GHD May 2011

Another VC-Dimension Argument: Subcube Lifting

First message constant on large set:

  • Alice, Bob lift their (n/3)-dim inputs from inner coords to full n-dim space

First message now redundant, so eliminate!

[Brody-C.’09]

Amit Chakrabarti 13-a

slide-20
SLIDE 20

Optimal Lower Bound for GHD May 2011

Better Round Elimination

Brody, Chakrabarti, Regev, Vidick, de Wolf [RANDOM 2010]

  • Previous argument reduced dimension too rapidly
  • Gives Rk(ghd) = n/2O(k2)
  • Can improve to Rk(ghd) = n/O(k2)

Amit Chakrabarti 14

slide-21
SLIDE 21

Optimal Lower Bound for GHD May 2011

Round Elimination V2.0: Geometric Perturbation

First message constant over large set A

x z c

1/2n

y A

ERR

{0,1}n

Amit Chakrabarti 15

slide-22
SLIDE 22

Optimal Lower Bound for GHD May 2011

Round Elimination V2.0: Geometric Perturbation

First message constant over large set A

x z c

1/2n

y A

ERR

{0,1}n

Alice: replace x with z = NearestNeighbour(x, A)

Amit Chakrabarti 15-a

slide-23
SLIDE 23

Optimal Lower Bound for GHD May 2011

Modern History

Amit Chakrabarti 16

slide-24
SLIDE 24

Optimal Lower Bound for GHD May 2011

Main Theorem

Chakrabarti, Regev [STOC 2011] And now, we show: R(ghd) = Ω(n)

Amit Chakrabarti 17

slide-25
SLIDE 25

Optimal Lower Bound for GHD May 2011

The Rectangle Property

Input universe U = {0, 1}n × {0, 1}n Deterministic protocol P, communicating ≤ c bits partitions U into ≤ 2c rectangles Ai × Bi, where Ai, Bi ⊆ {0, 1}n

Bob Alice

If P computes f : U → {0, 1}, then f −1(1) = R1 ∪ R2 ∪ · · · ∪ R2c

Amit Chakrabarti 18

slide-26
SLIDE 26

Optimal Lower Bound for GHD May 2011

The Rectangle Property

Input universe U = {0, 1}n × {0, 1}n Deterministic protocol P, communicating ≤ c bits partitions U into ≤ 2c rectangles Ai × Bi, where Ai, Bi ⊆ {0, 1}n

Bob Alice

If P computes f : U → {0, 1}, then f −1(1) = R1 ∪ R2 ∪ · · · ∪ R2c

Amit Chakrabarti 18

slide-27
SLIDE 27

Optimal Lower Bound for GHD May 2011

The Rectangle Property

Input universe U = {0, 1}n × {0, 1}n Deterministic protocol P, communicating ≤ c bits partitions U into ≤ 2c rectangles Ai × Bi, where Ai, Bi ⊆ {0, 1}n

Bob Alice

If P computes f : U → {0, 1}, then f −1(1) = R1 ∪ R2 ∪ · · · ∪ R2c

Amit Chakrabarti 18

slide-28
SLIDE 28

Optimal Lower Bound for GHD May 2011

The Rectangle Property

Input universe U = {0, 1}n × {0, 1}n Deterministic protocol P, communicating ≤ c bits partitions U into ≤ 2c rectangles Ai × Bi, where Ai, Bi ⊆ {0, 1}n

Bob Alice

If P computes f : U → {0, 1}, then f −1(1) = R1 ∪ R2 ∪ · · · ∪ R2c

Amit Chakrabarti 18

slide-29
SLIDE 29

Optimal Lower Bound for GHD May 2011

The Rectangle Property

Input universe U = {0, 1}n × {0, 1}n Deterministic protocol P, communicating ≤ c bits partitions U into ≤ 2c rectangles Ai × Bi, where Ai, Bi ⊆ {0, 1}n

Bob Alice

If P computes f : U → {0, 1}, then f −1(1) = R1 ∪ R2 ∪ · · · ∪ R2c

Amit Chakrabarti 18

slide-30
SLIDE 30

Optimal Lower Bound for GHD May 2011

The Rectangle Property

Input universe U = {0, 1}n × {0, 1}n Deterministic protocol P, communicating ≤ c bits partitions U into ≤ 2c rectangles Ai × Bi, where Ai, Bi ⊆ {0, 1}n

Bob Alice

If P computes f : U → {0, 1}, then f −1(0) = R1 ∪ R2 ∪ · · · ∪ R2c

Amit Chakrabarti 18

slide-31
SLIDE 31

Optimal Lower Bound for GHD May 2011

The Corruption Technique and a Twist

Deterministic: f −1(0) = R1 ∪ R2 ∪ · · · ∪ R2c Randomized: {P outputs 0} = R1 ∪ R2 ∪ · · · ∪ R2c

  • Partition covers most of f −1(0)
  • Each Ri mostly uncorrupted: contains much fewer 1s than 0s.

Amit Chakrabarti 19

slide-32
SLIDE 32

Optimal Lower Bound for GHD May 2011

The Corruption Technique and a Twist

Deterministic: f −1(0) = R1 ∪ R2 ∪ · · · ∪ R2c Randomized: {P outputs 0} = R1 ∪ R2 ∪ · · · ∪ R2c

  • Partition covers most of f −1(0)
  • Each Ri mostly uncorrupted: contains much fewer 1s than 0s.

For lower bound:

  • Show every large rectangle (size ≥ 20.99n × 20.99n) is corrupted

µ1(R) ≥ α µ0(R)

Amit Chakrabarti 19-a

slide-33
SLIDE 33

Optimal Lower Bound for GHD May 2011

The Corruption Technique and a Twist

Deterministic: f −1(0) = R1 ∪ R2 ∪ · · · ∪ R2c Randomized: {P outputs 0} = R1 ∪ R2 ∪ · · · ∪ R2c

  • Partition covers most of f −1(0)
  • Each Ri mostly uncorrupted: contains much fewer 1s than 0s.

For lower bound:

  • Show every large rectangle (size ≥ 20.99n × 20.99n) is corrupted

µ1(R) ≥ α µ0(R)

Amit Chakrabarti 19-b

slide-34
SLIDE 34

Optimal Lower Bound for GHD May 2011

The Corruption Technique and a Twist

Deterministic: f −1(0) = R1 ∪ R2 ∪ · · · ∪ R2c Randomized: {P outputs 0} = R1 ∪ R2 ∪ · · · ∪ R2c

  • Partition covers most of f −1(0)
  • Each Ri mostly uncorrupted: contains much fewer 1s than 0s.

For lower bound:

  • Show every large rectangle (size ≥ 20.99n × 20.99n) is corrupted

µ1(R) ≥ α µ0(R)

  • Caveat: not true! E.g., {(x, y) : x1:100√n = y1:100√n =

0}

Amit Chakrabarti 19

slide-35
SLIDE 35

Optimal Lower Bound for GHD May 2011

The Corruption Technique and a Twist

Deterministic: f −1(0) = R1 ∪ R2 ∪ · · · ∪ R2c Randomized: {P outputs 0} = R1 ∪ R2 ∪ · · · ∪ R2c

  • Partition covers most of f −1(0)
  • Each Ri mostly uncorrupted: contains much fewer 1s than 0s.

For lower bound:

  • Show every large rectangle (size ≥ 20.99n × 20.99n) is corrupted

µ1(R) ≥ α µ0(R)

  • Caveat: not true! E.g., {(x, y) : x1:100√n = y1:100√n =

0}

  • Show weaker inequality

µ1(R) + β µ⋆(R) ≥ α µ0(R) (α > β)

Amit Chakrabarti 19-a

slide-36
SLIDE 36

Optimal Lower Bound for GHD May 2011

Corruption with Jokers

Pick distribs µ0, µ1 on f −1(0), f −1(1), and another distrib µ⋆ Argue that for all large rectangles R, we have µ1(R) + β µ⋆(R) ≥ α µ0(R) (α > β)

Amit Chakrabarti 20

slide-37
SLIDE 37

Optimal Lower Bound for GHD May 2011

Corruption with Jokers

Pick distribs µ0, µ1 on f −1(0), f −1(1), and another distrib µ⋆ Argue that for all large rectangles R, we have µ1(R) + β µ⋆(R) ≥ α µ0(R) (α > β) Sum over partition {P outputs 0} = 2c

i=1 Ri:

µ1(P −1(0)) + β µ⋆(P −1(0)) ≥ α µ0(P −1(0))

Amit Chakrabarti 20-a

slide-38
SLIDE 38

Optimal Lower Bound for GHD May 2011

Corruption with Jokers

Pick distribs µ0, µ1 on f −1(0), f −1(1), and another distrib µ⋆ Argue that for all large rectangles R, we have µ1(R) + β µ⋆(R) ≥ α µ0(R) (α > β) Sum over partition {P outputs 0} = 2c

i=1 Ri:

µ1(P −1(0)) + β µ⋆(P −1(0)) ≥ α µ0(P −1(0)) ≥ α(1 − ε)

Amit Chakrabarti 20-b

slide-39
SLIDE 39

Optimal Lower Bound for GHD May 2011

Corruption with Jokers

Pick distribs µ0, µ1 on f −1(0), f −1(1), and another distrib µ⋆ Argue that for all large rectangles R, we have µ1(R) + β µ⋆(R) ≥ α µ0(R) (α > β) Sum over partition {P outputs 0} = 2c

i=1 Ri:

ε + β ≥ µ1(P −1(0)) + β µ⋆(P −1(0)) ≥ α µ0(P −1(0)) ≥ α(1 − ε)

Amit Chakrabarti 20-c

slide-40
SLIDE 40

Optimal Lower Bound for GHD May 2011

The Corruption Inequality and Its Proof

Let µ0 = Uniform on {(x, y) : x, y = 0} µ1 = Uniform on {(x, y) : x, y = −10/√n} µ⋆ = Uniform on {(x, y) : x, y = 10/√n} The Key Inequality: For |A|, |B| ≥ 20.99n

1 2(µ1(A × B) + µ⋆(A × B)) ≥ 9 10 µ0(A × B)

“Inner product between large sets not too concentrated around zero”

Amit Chakrabarti 21

slide-41
SLIDE 41

Optimal Lower Bound for GHD May 2011

The Corruption Inequality and Its Proof

Let µ0 = Uniform on {(x, y) : x, y = 0} µ1 = Uniform on {(x, y) : x, y = −10/√n} µ⋆ = Uniform on {(x, y) : x, y = 10/√n} The Key Inequality: For |A|, |B| ≥ 20.99n

1 2(µ1(A × B) + µ⋆(A × B)) ≥ 9 10 µ0(A × B)

“Inner product between large sets not too concentrated around zero” Proof Strategy: For A, B ⊆ Rn with γ(A), γ(B) ≥ 2−0.01n distrib of ˆ x, ˆ y “spread out” like N(0, 1) where γ = n-dim Gaussian, (ˆ x, ˆ y) ← A × B

Amit Chakrabarti 21-a

slide-42
SLIDE 42

Optimal Lower Bound for GHD May 2011

Proof Details

Goal: For A, B ⊆ Rn with γ(A), γ(B) ≥ 2−0.01n distrib of ˆ x, ˆ y “spread out” like N(0, 1) Think A = {directions} Abad = {bad directions in A} = {ˆ x ∈ A : ˆ x, ˆ y not spread out, for ˆ y ← B}

Amit Chakrabarti 22

slide-43
SLIDE 43

Optimal Lower Bound for GHD May 2011

Proof Details

Goal: For A, B ⊆ Rn with γ(A), γ(B) ≥ 2−0.01n distrib of ˆ x, ˆ y “spread out” like N(0, 1) Think A = {directions} Abad = {bad directions in A} = {ˆ x ∈ A : ˆ x, ˆ y not spread out, for ˆ y ← B} For a contradiction, suppose γ(Abad) > 2−0.02n Then (Raz’s Lemma): A contains orthogonal bad dirs ˆ x1, . . . , ˆ xn/2

Amit Chakrabarti 22-a

slide-44
SLIDE 44

Optimal Lower Bound for GHD May 2011

Proof Details

Goal: For A, B ⊆ Rn with γ(A), γ(B) ≥ 2−0.01n distrib of ˆ x, ˆ y “spread out” like N(0, 1) Think A = {directions} Abad = {bad directions in A} = {ˆ x ∈ A : ˆ x, ˆ y not spread out, for ˆ y ← B} For a contradiction, suppose γ(Abad) > 2−0.02n Then (Raz’s Lemma): A contains orthogonal bad dirs ˆ x1, . . . , ˆ xn/2 Therefore (Information Theory): ˆ y ← B can’t have enough entropy Contradicts γ(B) ≥ 2−0.01n

Amit Chakrabarti 22-b

slide-45
SLIDE 45

Optimal Lower Bound for GHD May 2011

Geometric and Info Theoretic Intuition

Large set A

Amit Chakrabarti 23

slide-46
SLIDE 46

Optimal Lower Bound for GHD May 2011

Geometric and Info Theoretic Intuition

Large set of dirs bad for B Large set A

Amit Chakrabarti 23

slide-47
SLIDE 47

Optimal Lower Bound for GHD May 2011

Geometric and Info Theoretic Intuition

Large set of dirs bad for B Large set A

Amit Chakrabarti 23

slide-48
SLIDE 48

Optimal Lower Bound for GHD May 2011

Geometric and Info Theoretic Intuition

Large set of dirs bad for B Large set A

0.99n ≤ H(y) ≤ H(y, x1, . . . , y, xn) = n/2

k=1 H(y, xk | y, x1, . . . , y, xk−1)

+ n

k=n/2+1 H(y, xk | y, x1, . . . , y, xk−1)

≤ n/2

k=1 0.7 + n k=n/2+1 1 = 0.85n Amit Chakrabarti 23

slide-49
SLIDE 49

Optimal Lower Bound for GHD May 2011

The Future

Amit Chakrabarti 24

slide-50
SLIDE 50

Optimal Lower Bound for GHD May 2011

The Future

Two simplifications of our proof [not yet published]

  • Vidick shows following anti-concentration inequality:

E[ x, y2] = Ω(1/n) Avoids “continuous information theory”; just concentration of measure

  • Sherstov: anti-concetration gives corruption-based proof that

R(near-orthogonal) = Ω(n) and reduces near-orthogonal to ghd; thus avoids “jokers”

  • Also, Sherstov proves anti-concentration using Talagrand’s inequality

Amit Chakrabarti 25

slide-51
SLIDE 51

Optimal Lower Bound for GHD May 2011

Conclusions

  • Settled communication complexity of ghd, proving a long-conjectured

Ω(n) bound

  • As a result, understood multi-pass space complexity of a number of

data stream problems

Amit Chakrabarti 26

slide-52
SLIDE 52

Optimal Lower Bound for GHD May 2011

Conclusions

  • Settled communication complexity of ghd, proving a long-conjectured

Ω(n) bound

  • As a result, understood multi-pass space complexity of a number of

data stream problems

Open Problem

Prove that ghd is hard under the uniform distribution

Amit Chakrabarti 26-b