Quantum Chebyshevs Inequality and Applications Yassine Hamoudi, - - PowerPoint PPT Presentation

quantum chebyshev s inequality and applications
SMART_READER_LITE
LIVE PREVIEW

Quantum Chebyshevs Inequality and Applications Yassine Hamoudi, - - PowerPoint PPT Presentation

Quantum Chebyshevs Inequality and Applications Yassine Hamoudi, Frdric Magniez IRIF , Universit Paris Diderot, CNRS QuantAlgo 2018 arXiv: 1807.06456 Buffons needle A needle dropped randomly on a floor with equally spaced parallel


slide-1
SLIDE 1

Quantum Chebyshev’s Inequality and Applications

Yassine Hamoudi, Frédéric Magniez

IRIF , Université Paris Diderot, CNRS QuantAlgo 2018 arXiv: 1807.06456

slide-2
SLIDE 2

Buffon’s needle

Buffon, G., Essai d'arithmétique morale, 1777.

A needle dropped randomly on a floor with equally spaced parallel lines will cross one of the lines with probability 2/π.

  • 2
slide-3
SLIDE 3

Use repeated random sampling and statistical analysis to estimate parameters of interest

Monte Carlo algorithms:

  • 3
slide-4
SLIDE 4

Use repeated random sampling and statistical analysis to estimate parameters of interest

Monte Carlo algorithms: Empirical mean:

2/ Output: (x1 +…+ xn)/n 1/ Repeat the experiment n times: n i.i.d. samples x1, …, xn ~ X

  • 3
slide-5
SLIDE 5

Use repeated random sampling and statistical analysis to estimate parameters of interest

Monte Carlo algorithms: Empirical mean:

2/ Output: (x1 +…+ xn)/n

Law of large numbers: x1 + . . . + xn

n

n→∞ E(X)

1/ Repeat the experiment n times: n i.i.d. samples x1, …, xn ~ X

  • 3
slide-6
SLIDE 6

Empirical mean:

˜ μ = x1 + . . . + xn n

with

x1, . . . , xn ∼ X

How fast does it converge to E(X) ?

  • 4
slide-7
SLIDE 7

Empirical mean:

˜ μ = x1 + . . . + xn n

with

x1, . . . , xn ∼ X

Chebyshev’s Inequality:

How fast does it converge to E(X) ?

Hypothesis:

| ˜ μ − E(X)| ≤ ϵE(X) E(X) ≠ 0 Var(X) = E(X2) − E(X)2 ≠ 0

and finite Objective:

multiplicative error 0 < ε < 1

with high probability

  • 4
slide-8
SLIDE 8

(in fact ) Empirical mean:

˜ μ = x1 + . . . + xn n

with

x1, . . . , xn ∼ X

Chebyshev’s Inequality:

How fast does it converge to E(X) ?

Hypothesis:

| ˜ μ − E(X)| ≤ ϵE(X) E(X) ≠ 0 Var(X) = E(X2) − E(X)2 ≠ 0

and finite Objective:

multiplicative error 0 < ε < 1

with high probability Number of samples needed: O (

E(X2) ϵ2E(X)2 )

O ( Var(X) ϵ2E(X)2 ) = O 1 ϵ2 ( E(X2) E(X)2 − 1)

  • 4
slide-9
SLIDE 9

(in fact ) Empirical mean:

˜ μ = x1 + . . . + xn n

with

x1, . . . , xn ∼ X

Chebyshev’s Inequality:

How fast does it converge to E(X) ?

Hypothesis:

| ˜ μ − E(X)| ≤ ϵE(X) E(X) ≠ 0 Var(X) = E(X2) − E(X)2 ≠ 0

and finite Objective:

multiplicative error 0 < ε < 1

with high probability Number of samples needed: O (

E(X2) ϵ2E(X)2 )

O ( Var(X) ϵ2E(X)2 ) = O 1 ϵ2 ( E(X2) E(X)2 − 1)

  • 4

Relative second moment

slide-10
SLIDE 10

(in fact ) Empirical mean:

˜ μ = x1 + . . . + xn n

with

x1, . . . , xn ∼ X

Chebyshev’s Inequality:

How fast does it converge to E(X) ?

Hypothesis:

| ˜ μ − E(X)| ≤ ϵE(X) E(X) ≠ 0 Var(X) = E(X2) − E(X)2 ≠ 0

and finite Objective:

multiplicative error 0 < ε < 1

with high probability Number of samples needed: O (

E(X2) ϵ2E(X)2 )

O ( Var(X) ϵ2E(X)2 ) = O 1 ϵ2 ( E(X2) E(X)2 − 1)

In practice: given an upper-bound , take samples

Δ2 ≥ E(X2) E(X)2

  • 4

n = Ω ( Δ2 ϵ2 )

Relative second moment

slide-11
SLIDE 11

Data stream model:

Frequency moments, Collision probability [Alon, Matias, Szegedy’99]

[Monemizadeh, Woodruff’] [Andoni et al.’11] [Crouch et al.’16]

Other applications

Testing properties of distributions:

Closeness [Goldreich, Ron’11] [Batu et al.’13] [Chan et al.’14], Conditional independence [Canonne et al.’18]

Estimating graph parameters:

Number of connected components, Minimum spanning tree weight

[Chazelle, Rubinfeld, Trevisan’05], Average distance [Goldreich, Ron’08], Number

  • f triangles [Eden et al. 17]

Counting with Markov chain Monte Carlo methods:

Counting vs. sampling [Jerrum, Sinclair’96] [Štefankovič et al.’09], Volume of convex bodies [Dyer, Frieze'91], Permanent [Jerrum, Sinclair, Vigoda’04]

etc.

  • 5
slide-12
SLIDE 12

Random variable X over sample space Ω ⊂ R+

Classical sample: one value x ∈ Ω, sampled with probability px

  • 6
slide-13
SLIDE 13

Quantum sample: one (controlled-)execution of a quantum sampler or , where

Random variable X over sample space Ω ⊂ R+

Classical sample: one value x ∈ Ω, sampled with probability px

SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩

with ψx = arbitrary garbage state

SX S−1

X

  • 6

( can be replaced with any such that )

|αx|2 = px αx px

slide-14
SLIDE 14

E(X) ≤ H

Δ2 ≥ E(X2) E(X)2 Δ2 ≥ E(X2) E(X)2

Can we use quadratically less samples in the quantum setting?

Sample space Ω ⊂ [0,B]

  • 7

B/(ϵ E(X))

slide-15
SLIDE 15

E(X) ≤ H

Δ2 ≥ E(X2) E(X)2 Δ2 ≥ E(X2) E(X)2

Yes! for additive error approximation Can we use quadratically less samples in the quantum setting?

| ˜ μ − E(X)| ≤ ϵ

[Montanaro’15] Given σ2 ≥ Var(X), σ/ε quantum samples vs σ2/ε2 classical samples

Sample space Ω ⊂ [0,B]

  • 7

B/(ϵ E(X))

slide-16
SLIDE 16

E(X) ≤ H

Δ2 ≥ E(X2) E(X)2 Δ2 ≥ E(X2) E(X)2

Yes! for additive error approximation Can we use quadratically less samples in the quantum setting?

| ˜ μ − E(X)| ≤ ϵ

[Montanaro’15] Given σ2 ≥ Var(X), σ/ε quantum samples vs σ2/ε2 classical samples

??? for multiplicative error approximation | ˜

μ − E(X)| ≤ ϵE(X)

Sample space Ω ⊂ [0,B]

  • 7

B/(ϵ E(X))

slide-17
SLIDE 17

E(X) ≤ H

Δ2 ≥ E(X2) E(X)2 Δ2 ≥ E(X2) E(X)2

Number of samples Conditions

Classical samples (Chebyshev’s inequality)

Δ2/ε2

[Brassard et al.’11] [Wocjan et al.’09] [Montanaro’15] [Montanaro’15] [Li, Wu’17]

Δ2/ε or (Δ/ε)*(H/L) Our result (Δ/ε)*log3(H/E(X))

Yes! for additive error approximation Can we use quadratically less samples in the quantum setting?

| ˜ μ − E(X)| ≤ ϵ

[Montanaro’15] Given σ2 ≥ Var(X), σ/ε quantum samples vs σ2/ε2 classical samples

??? for multiplicative error approximation | ˜

μ − E(X)| ≤ ϵE(X)

Δ2 ≥ E(X2) E(X)2

Sample space Ω ⊂ [0,B]

  • 7

B/(ϵ E(X))

slide-18
SLIDE 18

E(X) ≤ H

Δ2 ≥ E(X2) E(X)2 Δ2 ≥ E(X2) E(X)2

Number of samples Conditions

Classical samples (Chebyshev’s inequality)

Δ2/ε2

[Brassard et al.’11] [Wocjan et al.’09] [Montanaro’15] [Montanaro’15] [Li, Wu’17]

Δ2/ε or (Δ/ε)*(H/L) Our result (Δ/ε)*log3(H/E(X))

Yes! for additive error approximation Can we use quadratically less samples in the quantum setting?

| ˜ μ − E(X)| ≤ ϵ

[Montanaro’15] Given σ2 ≥ Var(X), σ/ε quantum samples vs σ2/ε2 classical samples

??? for multiplicative error approximation | ˜

μ − E(X)| ≤ ϵE(X)

Δ2 ≥ E(X2) E(X)2

Sample space Ω ⊂ [0,B]

  • 7

B/(ϵ E(X))

slide-19
SLIDE 19

E(X) ≤ H

Δ2 ≥ E(X2) E(X)2 Δ2 ≥ E(X2) E(X)2

Number of samples Conditions

Classical samples (Chebyshev’s inequality)

Δ2/ε2

[Brassard et al.’11] [Wocjan et al.’09] [Montanaro’15] [Montanaro’15] [Li, Wu’17]

Δ2/ε or (Δ/ε)*(H/L) Our result (Δ/ε)*log3(H/E(X))

Yes! for additive error approximation Can we use quadratically less samples in the quantum setting?

| ˜ μ − E(X)| ≤ ϵ

[Montanaro’15] Given σ2 ≥ Var(X), σ/ε quantum samples vs σ2/ε2 classical samples

??? for multiplicative error approximation | ˜

μ − E(X)| ≤ ϵE(X)

Δ2 ≥ E(X2) E(X)2

Sample space Ω ⊂ [0,B]

  • 7

B/(ϵ E(X))

slide-20
SLIDE 20

E(X) ≤ H

Δ2 ≥ E(X2) E(X)2 Δ2 ≥ E(X2) E(X)2

Number of samples Conditions

Classical samples (Chebyshev’s inequality)

Δ2/ε2

[Brassard et al.’11] [Wocjan et al.’09] [Montanaro’15] [Montanaro’15] [Li, Wu’17]

Δ2/ε or (Δ/ε)*(H/L) Our result (Δ/ε)*log3(H/E(X))

Yes! for additive error approximation Can we use quadratically less samples in the quantum setting?

| ˜ μ − E(X)| ≤ ϵ

[Montanaro’15] Given σ2 ≥ Var(X), σ/ε quantum samples vs σ2/ε2 classical samples

??? for multiplicative error approximation | ˜

μ − E(X)| ≤ ϵE(X)

Δ2 ≥ E(X2) E(X)2

Sample space Ω ⊂ [0,B]

L ≤ E(X) ≤ H

  • 7

B/(ϵ E(X))

slide-21
SLIDE 21

E(X) ≤ H

Δ2 ≥ E(X2) E(X)2 Δ2 ≥ E(X2) E(X)2

Number of samples Conditions

Classical samples (Chebyshev’s inequality)

Δ2/ε2

[Brassard et al.’11] [Wocjan et al.’09] [Montanaro’15] [Montanaro’15] [Li, Wu’17]

Δ2/ε or (Δ/ε)*(H/L) Our result (Δ/ε)*log3(H/E(X))

Yes! for additive error approximation Can we use quadratically less samples in the quantum setting?

| ˜ μ − E(X)| ≤ ϵ

[Montanaro’15] Given σ2 ≥ Var(X), σ/ε quantum samples vs σ2/ε2 classical samples

??? for multiplicative error approximation | ˜

μ − E(X)| ≤ ϵE(X)

Δ2 ≥ E(X2) E(X)2

Sample space Ω ⊂ [0,B]

L ≤ E(X) ≤ H

  • 7

B/(ϵ E(X))

slide-22
SLIDE 22

Our Approach

slide-23
SLIDE 23

Sampler: Result: O (

B ϵ E(X) ) quantum samples to obtain | ˜

μ − E(X)| ≤ ϵE(X)

  • n sample space Ω ⊂ [0,B]

SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩

  • 9

Subroutine: the Amplitude Estimation algorithm

slide-24
SLIDE 24

Sampler: Result: O (

B ϵ E(X) ) quantum samples to obtain | ˜

μ − E(X)| ≤ ϵE(X)

  • n sample space Ω ⊂ [0,B]

SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩ ∑

x∈Ω

px |ψx⟩|x⟩|0⟩ ∑

x∈Ω

px |ψx⟩|x⟩( 1 − x B |0⟩ + x B |1⟩)

Controlled rotation Reordering

Reduction to a Bernoulli sampler [Brassard et al.’11] [Wocjan et al.’09] [Montanaro’15]:

  • 9

1 − E(X) B |φ0⟩|0⟩ + E(X) B |φ1⟩|1⟩ = SY|0⟩

Subroutine: the Amplitude Estimation algorithm

slide-25
SLIDE 25

Sampler: Result: O (

B ϵ E(X) ) quantum samples to obtain | ˜

μ − E(X)| ≤ ϵE(X) SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩

  • 10

SY|0⟩ = 1 − E(X) B |φ0⟩|0⟩ + E(X) B |φ1⟩|1⟩

Subroutine: the Amplitude Estimation algorithm

Expectation of a Bernoulli sampler [Brassard et al.’02]:

  • n sample space Ω ⊂ [0,B]
slide-26
SLIDE 26

Sampler: Result: O (

B ϵ E(X) ) quantum samples to obtain | ˜

μ − E(X)| ≤ ϵE(X) SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩

  • 10

SY|0⟩ = 1 − E(X) B |φ0⟩|0⟩ + E(X) B |φ1⟩|1⟩

Subroutine: the Amplitude Estimation algorithm

Expectation of a Bernoulli sampler [Brassard et al.’02]: Step 0: the Grover's operator has eigenvalues , where . G = S−1

Y (I − 2|0⟩⟨0|)SY(I − 2I ⊗ |1⟩⟨1|)

e±2iθ θ = sin−1( E(X)/B) Step 2: output as an estimate to E(X)/B. sin2(˜ θ) Step 1: use the Phase Estimation Algorithm on G for steps (i.e. using t quantum samples), to get an estimate of . ˜ θ ±θ t ≥ Ω( B/(ϵ E(X))) (˜ μ = B ⋅ sin2(˜ θ))

  • n sample space Ω ⊂ [0,B]
slide-27
SLIDE 27

If : the number of samples is B ≤ E(X2) E(X) O E(X2) ϵE(X) Result: O (

B ϵ E(X) ) quantum samples to obtain | ˜

μ − E(X)| ≤ ϵE(X)

  • 11

Sampler: SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩

  • n sample space Ω ⊂ [0,B]
slide-28
SLIDE 28

If : the number of samples is B ≤ E(X2) E(X) O E(X2) ϵE(X) If

?

Result: O (

B ϵ E(X) )

B ≫ E(X2) E(X) quantum samples to obtain | ˜ μ − E(X)| ≤ ϵE(X)

  • 11

Sampler: SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩

  • n sample space Ω ⊂ [0,B]
slide-29
SLIDE 29

If : the number of samples is B ≤ E(X2) E(X) O E(X2) ϵE(X) If E(X2) E(X)

?

Result: O (

B ϵ E(X) )

B ≫ E(X2) E(X) quantum samples to obtain | ˜ μ − E(X)| ≤ ϵE(X)

  • 11

: map the outcomes larger than to 0 Sampler: SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩

  • n sample space Ω ⊂ [0,B]
slide-30
SLIDE 30

1

Random variable X

  • 12

B

Largest outcome

px x

slide-31
SLIDE 31

1

Random variable X<b

  • 13

b

New largest outcome

px x

≥ E(X2) E(X)

B

slide-32
SLIDE 32

If : the number of samples is B ≤ E(X2) E(X) O E(X2) ϵE(X) If : map the outcomes larger than to 0 E(X2) E(X) Result: O (

B ϵ E(X) )

B ≫ E(X2) E(X) quantum samples to obtain | ˜ μ − E(X)| ≤ ϵE(X)

  • 14

?

Sampler: SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩

  • n sample space Ω ⊂ [0,B]
slide-33
SLIDE 33

If : the number of samples is B ≤ E(X2) E(X) O E(X2) ϵE(X) If : map the outcomes larger than to 0 E(X2) E(X) Result: O (

B ϵ E(X) )

B ≫ E(X2) E(X) quantum samples to obtain | ˜ μ − E(X)| ≤ ϵE(X)

  • 14

Lemma: If then b ≥ E(X2) ϵE(X) (1 − ϵ)E(X) ≤ E(X<b) ≤ E(X) . Sampler: SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩

  • n sample space Ω ⊂ [0,B]
slide-34
SLIDE 34

If : the number of samples is B ≤ E(X2) E(X) O E(X2) ϵE(X) If : map the outcomes larger than to 0 E(X2) E(X) Result: O (

B ϵ E(X) )

B ≫ E(X2) E(X) quantum samples to obtain | ˜ μ − E(X)| ≤ ϵE(X)

  • 14

Lemma: If then b ≥ E(X2) ϵE(X) (1 − ϵ)E(X) ≤ E(X<b) ≤ E(X) . Problem: given how to find a threshold ? Δ2 ≥ E(X2) E(X)2 b ≈ E(X) ⋅ Δ2 Sampler: SX|0⟩ = ∑

x∈Ω

px |ψx⟩|x⟩

  • n sample space Ω ⊂ [0,B]
slide-35
SLIDE 35

Solution: use the Amplitude Estimation algorithm to do a logarithmic search on b (given an upper-bound H ≥ E(X))

  • 15

Problem: given how to find a threshold ? Δ2 ≥ E(X2) E(X)2 b ≈ E(X) ⋅ Δ2

slide-36
SLIDE 36

Threshold Estimated value Number of samples Estimation

Solution: use the Amplitude Estimation algorithm to do a logarithmic search on b (given an upper-bound H ≥ E(X))

  • 15

Problem: given how to find a threshold ? Δ2 ≥ E(X2) E(X)2 b ≈ E(X) ⋅ Δ2

b0 = HΔ2 b1 = (H/2)Δ2 b2 = (H/4)Δ2 ˜ μ0 …

E(X<b0) b0 E(X<b1) b1 E(X<b2) b2

Δ Δ Δ

˜ μ1 ˜ μ2 … … …

Stopping rule: ˜ μi ≠ 0 Output: bi

slide-37
SLIDE 37

Threshold Estimated value Number of samples Estimation

Solution: use the Amplitude Estimation algorithm to do a logarithmic search on b (given an upper-bound H ≥ E(X))

  • 15

Problem: given how to find a threshold ? Δ2 ≥ E(X2) E(X)2 b ≈ E(X) ⋅ Δ2

b0 = HΔ2 b1 = (H/2)Δ2 b2 = (H/4)Δ2 ˜ μ0 …

E(X<b0) b0 E(X<b1) b1 E(X<b2) b2

Δ Δ Δ

˜ μ1 ˜ μ2 …

Theorem: the first non-zero is obtained w.h.p. when: ˜ μi

2 ⋅ E(X)Δ2 ≤ bi ≤ 104 ⋅ E(X)Δ2

… …

Stopping rule: ˜ μi ≠ 0 Output: bi

slide-38
SLIDE 38

Analysis

  • 16
  • If →

E(X<bi) bi ≈ E(X) bi ≈ 1 Δ2

bi ≈ E(X) ⋅ Δ2

→ Δ samples are enough

slide-39
SLIDE 39

Analysis

  • 16
  • If is very large →

is very small → Δ samples is not enough to distinguish from 0

E(X<bi) bi

  • If →

E(X<bi) bi ≈ E(X) bi ≈ 1 Δ2

bi ≈ E(X) ⋅ Δ2

→ Δ samples are enough

?

E(X<bi) bi

bi

slide-40
SLIDE 40

Analysis

[Brassard et al.’02] The output of the Amplitude-Estimation algorithm is 0 w.h.p. when the estimated value is below the inverse-square of the number of samples

Δ

  • 16

E(X<bi) bi

  • If is very large →

is very small → Δ samples is not enough to distinguish from 0

E(X<bi) bi

  • If →

E(X<bi) bi ≈ E(X) bi ≈ 1 Δ2

bi ≈ E(X) ⋅ Δ2

→ Δ samples are enough

?

E(X<bi) bi

bi

slide-41
SLIDE 41

If then

Analysis

[Brassard et al.’02] The output of the Amplitude-Estimation algorithm is 0 w.h.p. when the estimated value is below the inverse-square of the number of samples

Δ

b ≥ 104 ⋅ E(X)Δ2

  • 16

E(X<bi) bi

  • If is very large →

is very small → Δ samples is not enough to distinguish from 0

E(X<bi) bi

Lemma:

E(X<b) b ≤ 1 104 ⋅ Δ2

  • If →

E(X<bi) bi ≈ E(X) bi ≈ 1 Δ2

bi ≈ E(X) ⋅ Δ2

→ Δ samples are enough

?

E(X<bi) bi

bi

slide-42
SLIDE 42

Step 1: Logarithmic search on b until Amplitude-Estimation(SX<b, Δ) ≠ 0

2 ⋅ E(X)Δ2 ≤ b ≤ 104 ⋅ E(X)Δ2 with high probability

get

Δ ⋅ log3 ( H E(X))

Final algorithm:

  • 17
slide-43
SLIDE 43

Step 1: Logarithmic search on b until Amplitude-Estimation(SX<b, Δ) ≠ 0

2 ⋅ E(X)Δ2 ≤ b ≤ 104 ⋅ E(X)Δ2 with high probability

get Step 2: Set threshold and output

d = b/ϵ

with high probability get | ˜

μ − E(X)| ≤ ϵE(X)

Δ ⋅ log3 ( H E(X))

Δ/ϵ3/2

Final algorithm:

  • 17

Amplitude-Estimation(SX<d, Δ/ϵ3/2) ≠ 0

slide-44
SLIDE 44

Step 1: Logarithmic search on b until Amplitude-Estimation(SX<b, Δ) ≠ 0

2 ⋅ E(X)Δ2 ≤ b ≤ 104 ⋅ E(X)Δ2 with high probability

get Step 2: Set threshold and output

d = b/ϵ

with high probability get | ˜

μ − E(X)| ≤ ϵE(X)

Δ ⋅ log3 ( H E(X))

Δ/ϵ3/2

Final algorithm:

Step 2bis: Slightly refined algorithm, adapted from [Heinrich’01, Montanaro’15]

Δ/ϵ

  • 17

Amplitude-Estimation(SX<d, Δ/ϵ3/2) ≠ 0

slide-45
SLIDE 45

Applications

slide-46
SLIDE 46

Application 1: counting the number of edges in a graph

Estimator X :=

  • 1. Sample a vertex v ∈ V uniformly at random
  • 2. Sample a neighbor w of v uniformly at random
  • 3. If deg(v) < deg(w) (or deg(v) = deg(w) and v <lex w)

Output n*deg(v) Else Output 0

  • 19

λ(v,w)

slide-47
SLIDE 47

Application 1: counting the number of edges in a graph

Estimator X :=

  • 1. Sample a vertex v ∈ V uniformly at random
  • 2. Sample a neighbor w of v uniformly at random
  • 3. If deg(v) < deg(w) (or deg(v) = deg(w) and v <lex w)

Output n*deg(v) Else Output 0

Lemma: E(X) = m and E(X2)/E(X)2 ≤ O(√n).

  • 19

λ(v,w)

(when m ≥ Ω(n))

[Goldreich, Ron’08] [Seshadhri’15]

slide-48
SLIDE 48

Application 1: counting the number of edges in a graph

Estimator X :=

  • 1. Sample a vertex v ∈ V uniformly at random
  • 2. Sample a neighbor w of v uniformly at random
  • 3. If deg(v) < deg(w) (or deg(v) = deg(w) and v <lex w)

Output n*deg(v) Else Output 0

Lemma: E(X) = m and E(X2)/E(X)2 ≤ O(√n).

  • 19

λ(v,w)

(when m ≥ Ω(n))

SX|0⟩ = ∑

v∈V ∑ w∈N(v)

1 n ⋅ deg(v) |v⟩|w⟩|λ(v, w)⟩

[Goldreich, Ron’08] [Seshadhri’15]

slide-49
SLIDE 49

Application 1: counting the number of edges in a graph

Estimator X :=

  • 1. Sample a vertex v ∈ V uniformly at random
  • 2. Sample a neighbor w of v uniformly at random
  • 3. If deg(v) < deg(w) (or deg(v) = deg(w) and v <lex w)

Output n*deg(v) Else Output 0

Lemma: E(X) = m and E(X2)/E(X)2 ≤ O(√n).

  • 19

λ(v,w)

(when m ≥ Ω(n))

SX|0⟩ = ∑

v∈V ∑ w∈N(v)

1 n ⋅ deg(v) |v⟩|w⟩|λ(v, w)⟩ Result: O(n1/4/ε) quantum samples (= quantum queries) to approximate m.

(when m ≥ Ω(n))

[Goldreich, Ron’08] [Seshadhri’15]

slide-50
SLIDE 50

Application 1: counting the number of edges in a graph

Estimator X :=

  • 1. Sample a vertex v ∈ V uniformly at random
  • 2. Sample a neighbor w of v uniformly at random
  • 3. If deg(v) < deg(w) (or deg(v) = deg(w) and v <lex w)

Output n*deg(v) Else Output 0

Lemma: E(X) = m and E(X2)/E(X)2 ≤ O(n/√m), but we don’t know n/√m…

  • 20

λ(v,w) SX|0⟩ = ∑

v∈V ∑ w∈N(v)

1 n ⋅ deg(v) |v⟩|w⟩|λ(v, w)⟩

[Goldreich, Ron’08] [Seshadhri’15]

slide-51
SLIDE 51

Application 1: counting the number of edges in a graph

Estimator X :=

  • 1. Sample a vertex v ∈ V uniformly at random
  • 2. Sample a neighbor w of v uniformly at random
  • 3. If deg(v) < deg(w) (or deg(v) = deg(w) and v <lex w)

Output n*deg(v) Else Output 0

Lemma: E(X) = m and E(X2)/E(X)2 ≤ O(n/√m), but we don’t know n/√m…

  • 20

λ(v,w) SX|0⟩ = ∑

v∈V ∑ w∈N(v)

1 n ⋅ deg(v) |v⟩|w⟩|λ(v, w)⟩ Result: Θ(n1/2/m1/4) quantum samples (= quantum queries) to approximate m.

[Goldreich, Ron’08] [Seshadhri’15]

slide-52
SLIDE 52

Application 2: frequency moments in the streaming model

  • 21

x =

1 2 3 n

Stream of updates to x:

slide-53
SLIDE 53

5

Application 2: frequency moments in the streaming model

  • 21

x =

1 2 3 n

Stream of updates to x: (3,+5)

slide-54
SLIDE 54

5

Application 2: frequency moments in the streaming model

  • 21

x =

1 2 3 n

Stream of updates to x:

  • 6

(3,+5) ; (2,-6)

slide-55
SLIDE 55

4

Application 2: frequency moments in the streaming model

  • 21

x =

1 2 3 n

Stream of updates to x:

  • 6

(3,+5) ; (2,-6) ; (3,-1)

slide-56
SLIDE 56

4

Application 2: frequency moments in the streaming model

  • 21

Fk =

n

i=1

|xi|k

x =

1 2 3 n

Stream of updates to x:

  • 6

Frequency moment of order k ≥ 3:

(3,+5) ; (2,-6) ; (3,-1)

slide-57
SLIDE 57

4

Application 2: frequency moments in the streaming model

  • 21

Fk =

n

i=1

|xi|k

Best P-pass algorithm with space memory M approximating Fk?

x =

1 2 3 n

Stream of updates to x:

  • 6

Frequency moment of order k ≥ 3:

(3,+5) ; (2,-6) ; (3,-1)

slide-58
SLIDE 58

4

Application 2: frequency moments in the streaming model

Classically: PM = Θ(n1-2/k)

  • 21

Fk =

n

i=1

|xi|k

Best P-pass algorithm with space memory M approximating Fk?

x =

1 2 3 n

Stream of updates to x:

  • 6

Frequency moment of order k ≥ 3:

(3,+5) ; (2,-6) ; (3,-1)

[Monemizadeh, Woodruff’10] [Andoni, Krauthgamer, Onak’11]

1 sample from a random variable X with and

E(X2)/E(X)2 ≤ P ⋅ F2

k

1 pass + memory M = n1−2/k

P

| |

E(X) ≈ Fk

slide-59
SLIDE 59

4

Application 2: frequency moments in the streaming model

Classically: PM = Θ(n1-2/k)

  • 21

Fk =

n

i=1

|xi|k

Best P-pass algorithm with space memory M approximating Fk?

x =

1 2 3 n

Stream of updates to x:

  • 6

Frequency moment of order k ≥ 3:

(3,+5) ; (2,-6) ; (3,-1)

Quantumly: P2M = O(n1-2/k)

[Monemizadeh, Woodruff’10] [Andoni, Krauthgamer, Onak’11]

1 sample from a random variable X with and

E(X2)/E(X)2 ≤ P ⋅ F2

k

1 pass + memory M = n1−2/k

P

* can be done in one pass also

S−1

X

| |

1 pass + memory M = n1−2/k

P2

1 quantum sample* SX from a r.v. X with and

| |

E(X) ≈ Fk E(X) ≈ Fk E(X2)/E(X)2 ≤ (P ⋅ Fk)2

slide-60
SLIDE 60

Application 3: counting the number of triangles in a graph

  • 22

More complicated than edges… [Eden, Levi, Ron’15] [Eden, Levi, Ron, Seshadhri’17] Main subroutine: estimator X for the number of triangles adjacent to any vertex v

slide-61
SLIDE 61

Application 3: counting the number of triangles in a graph

  • 22

More complicated than edges… [Eden, Levi, Ron’15] [Eden, Levi, Ron, Seshadhri’17] Main subroutine: estimator X for the number of triangles adjacent to any vertex v

1 classical sample = O(1) queries in expectation but O(√m) in the worst case

slide-62
SLIDE 62

Application 3: counting the number of triangles in a graph

  • 22

More complicated than edges… [Eden, Levi, Ron’15] [Eden, Levi, Ron, Seshadhri’17] Main subroutine: estimator X for the number of triangles adjacent to any vertex v

1 classical sample = O(1) queries in expectation but O(√m) in the worst case

Variable-time Amplitude Estimation: estimate the amplitude when some

“branches” of the computation stop earlier than the others

slide-63
SLIDE 63

Application 3: counting the number of triangles in a graph

  • 22

More complicated than edges… [Eden, Levi, Ron’15] [Eden, Levi, Ron, Seshadhri’17] Main subroutine: estimator X for the number of triangles adjacent to any vertex v

1 classical sample = O(1) queries in expectation but O(√m) in the worst case

Variable-time Amplitude Estimation: estimate the amplitude when some

“branches” of the computation stop earlier than the others

˜ Θ ( n t1/6 + m3/4 t ) quantum queries for triangle counting

Result:

vs. ˜ Θ ( n t1/3 + m3/2 t ) classical queries

slide-64
SLIDE 64

Conclusion

slide-65
SLIDE 65
  • 24

The mean of any quantum sampler SX is estimated with multiplicative error ε using quantum samples, given and .

Δ2 ≥ E(X2) E(X)2

H ≥ E(X)

˜ O ( Δ ϵ ⋅ log3 ( H E(X)))

[Nayak, Wu’99] : corresponding lower bound

Applications:

  • Frequency moments:
  • Edge counting:
  • Triangle counting: ˜

Θ ( n t1/6 + m3/4 t ) ˜ Θ ( n m1/4) P2M = ˜ O(n1−2/k)

Lower bounds with a property testing to communication complexity reduction method (reduction to Disjointness)

[Blais et al’12] [Eden, Rosenbaum’17]

Lower bound: ?

arXiv: 1807.06456

slide-66
SLIDE 66

Extra slides

slide-67
SLIDE 67

No a priori information on E(X2)/E(X)2

  • 26

Result: There is an algorithm that approximates the mean of any quantum sampler SX over Ω ⊂ [0,B] with quantum samples, and no a priori information on X.

O ( B ϵE(X) + E(X2) ϵE(X))

→ straightforward quantization of [Dagum, Karp, Luby, Ross’00]

slide-68
SLIDE 68
  • 27

Lemma: If then b ≥ E(X2) ϵE(X) If then

b ≥ 104 ⋅ E(X)Δ2

Lemma:

E(X<b) b ≤ 1 104 ⋅ Δ2

(1 − ϵ)E(X) ≤ E(X<b) ≤ E(X) .

slide-69
SLIDE 69
  • 27

Lemma: If then b ≥ E(X2) ϵE(X) ∙ E(X<b) = E(X) − E(X≥b) ≥ (1 − ϵ)E(X) If then

b ≥ 104 ⋅ E(X)Δ2

Lemma:

E(X<b) b ≤ 1 104 ⋅ Δ2

Proof:

(1 − ϵ)E(X) ≤ E(X<b) ≤ E(X) . ∙ E(X≥b) ≤ E(X2) b ≤ ϵE(X)

Proof:

E(X<b) b ≤ E(X) 104E(X)Δ2 ≤ 1 104 ⋅ Δ2

slide-70
SLIDE 70

Example

  • 28

1

px x

B

slide-71
SLIDE 71
  • 29

1 b B

E(X<b) b

Example

slide-72
SLIDE 72
  • 29

1 b B

E(X<b) b 1 Δ2 E(X)Δ2

Example

slide-73
SLIDE 73

Lower bound

  • 30

[Nayak, Wu’99] Any algorithm solving the Mean Estimation problem for

parameters 0 < ε <1/6, Δ > 1 on the sample space Ω = {0,1} must use Ω((Δ-1)/ε) quantum samples.

Definition: An algorithm solves the Mean Estimation problem for parameters

ε,Δ if, for any sampler SX satisfying E(X2)/E(X)2 ∈ [Δ,2Δ], it outputs a value satisfying with probability 2/3.

| ˜ μ − E(X)| ≤ ϵE(X) ˜ μ