Discrepancy of Random Set Systems
Rebecca Hoberg and Thomas Rothvoß
Discrepancy of Random Set Systems Rebecca Hoberg and Thomas Rothvo - - PowerPoint PPT Presentation
Discrepancy of Random Set Systems Rebecca Hoberg and Thomas Rothvo Discrepancy theory Set system with m sets, n elements i S b b b Discrepancy theory 1 Set system with m sets, n elements Coloring x { 1 , +1 } n 1
Rebecca Hoberg and Thomas Rothvoß
◮ Set system with m sets, n elements
i S
◮ Set system with m sets, n elements ◮ Coloring x ∈ {−1, +1}n
i S
b b
−1
b
+1 −1
◮ Set system with m sets, n elements ◮ Coloring x ∈ {−1, +1}n ◮ A =
1 1 1 1 1 1 i S
b b
−1
b
+1 −1
◮ Set system with m sets, n elements ◮ Coloring x ∈ {−1, +1}n ◮ A =
1 1 1 1 1 1
◮ disc(A) =
min
x∈{−1,1}n Ax∞
i S
b b
−1
b
+1 −1
◮ Set system with m sets, n elements ◮ Coloring x ∈ {−1, +1}n ◮ A =
1 1 1 1 1 1
◮ disc(A) =
min
x∈{−1,1}n Ax∞
i S
b b
−1
b
+1 −1 Known results:
◮ n sets, n elements: disc(A) = O(√n) [Spencer ’85]
◮ Set system with m sets, n elements ◮ Coloring x ∈ {−1, +1}n ◮ A =
1 1 1 1 1 1
◮ disc(A) =
min
x∈{−1,1}n Ax∞
i S
b b
−1
b
+1 −1 Known results:
◮ n sets, n elements: disc(A) = O(√n) [Spencer ’85] ◮ Every element in ≤ t sets: disc(A) < 2t [Beck & Fiala ’81]
◮ Set system with m sets, n elements ◮ Coloring x ∈ {−1, +1}n ◮ A =
1 1 1 1 1 1
◮ disc(A) =
min
x∈{−1,1}n Ax∞
i S
b b
−1
b
+1 −1 Known results:
◮ n sets, n elements: disc(A) = O(√n) [Spencer ’85] ◮ Every element in ≤ t sets: disc(A) < 2t [Beck & Fiala ’81] ◮ Beck-Fiala Conjecture: disc(A) ≤ O(
√ t)
◮ Set system with m sets, n elements ◮ Coloring x ∈ {−1, +1}n ◮ A =
1 1 1 1 1 1
◮ disc(A) =
min
x∈{−1,1}n Ax∞
i S
b b
−1
b
+1 −1 Known results:
◮ n sets, n elements: disc(A) = O(√n) [Spencer ’85] ◮ Every element in ≤ t sets: disc(A) < 2t [Beck & Fiala ’81] ◮ Beck-Fiala Conjecture: disc(A) ≤ O(
√ t)
Suppose n ≥ ˜ Θ(m2), entries of A ∈ {0, 1}m×n chosen indep. with prob. p. Then with high probability disc(A) ≤ 1.
◮ Suppose X ∈ Zm a random variable, θ ∈ Rm. ◮ We define
ˆ X(θ) = E[e2πiX,θ]
◮ Suppose X ∈ Zm a random variable, θ ∈ Rm. ◮ We define
ˆ X(θ) = E[e2πiX,θ] =
Pr[X = λ]e2πiX,θ
◮ Suppose X ∈ Zm a random variable, θ ∈ Rm. ◮ We define
ˆ X(θ) = E[e2πiX,θ] =
Pr[X = λ]e2πiX,θ
◮ Fourier inversion formula:
For λ ∈ Zm we have Pr[X = λ] =
2, 1 2)m
ˆ X(θ)e−2πiλ,θdθ
◮ Suppose X ∈ Zm a random variable, θ ∈ Rm. ◮ We define
ˆ X(θ) = E[e2πiX,θ] =
Pr[X = λ]e2πiX,θ
◮ Fourier inversion formula:
For λ = 0 we have Pr[X = 0] =
2, 1 2)m
ˆ X(θ)dθ
For fixed A ∈ {0, 1}m×n:
◮ Choose x ∼ {−1, 1}n uniformly, and let D = Ax.
For fixed A ∈ {0, 1}m×n:
◮ Choose x ∼ {−1, 1}n uniformly, and let D = Ax. ◮ Suffices to show Pr[D + R = 0] > 0 where R∞ ≤ ∆
chosen at random
For fixed A ∈ {0, 1}m×n:
◮ Choose x ∼ {−1, 1}n uniformly, and let D = Ax. ◮ Suffices to show Pr[D + R = 0] > 0 where R∞ ≤ ∆
chosen at random
◮ X = D + R ∈ Zm. We have
Pr[X = 0] =
2 , 1 2 )m
ˆ X(θ)dθ
For fixed A ∈ {0, 1}m×n:
◮ Choose x ∼ {−1, 1}n uniformly, and let D = Ax. ◮ Suffices to show Pr[D + R = 0] > 0 where R∞ ≤ ∆
chosen at random
◮ X = D + R ∈ Zm. We have
Pr[X = 0] =
2 , 1 2 )m
ˆ X(θ)dθ =
2 , 1 2 )m
ˆ D(θ) ˆ R(θ)dθ
ˆ D(θ) = E
ˆ D(θ) = E
x∼{−1,1}n
ˆ D(θ) = E
x∼{−1,1}n
=
n
E
xj[e2πixjAj,θ]
where Aj ∈ {0, 1}m the jth column of A.
ˆ D(θ) = E
x∼{−1,1}n
=
n
cos
ˆ D(θ) = E
x∼{−1,1}n
=
n
cos
b b b b b b b b b
θ ∈ [−1
2, 1 2]m
ˆ D(θ) = E
x∼{−1,1}n
=
n
cos
◮ If |Aj, θ| ≤ 1 4 for all j, then ˆ
D(θ) ≥ 0.
b b b b b b b b b
θ ∈ [−1
2, 1 2]m
ˆ D(θ) = E
x∼{−1,1}n
=
n
cos
◮ If |Aj, θ| ≤ 1 4 for all j, then ˆ
D(θ) ≥ 0.
b b b b b b b b b
θ ∈ [−1
2, 1 2]m
t))
ˆ D(θ) = E
x∼{−1,1}n
=
n
cos
◮ If |Aj, θ| ≤ 1 4 for all j, then ˆ
D(θ) ≥ 0.
◮ If Aj, θ ≈ half-integral for all j, then | ˆ
D(θ)| large.
b b b b b b b b b
θ ∈ [−1
2, 1 2]m
ˆ D(θ) = E
x∼{−1,1}n
=
n
cos
◮ If |Aj, θ| ≤ 1 4 for all j, then ˆ
D(θ) ≥ 0.
◮ If Aj, θ ≈ half-integral for all j, then | ˆ
D(θ)| large.
b b b b b b b b b
θ ∈ [−1
2, 1 2]m
ˆ D(θ) = E
x∼{−1,1}n
=
n
cos
◮ If |Aj, θ| ≤ 1 4 for all j, then ˆ
D(θ) ≥ 0.
◮ If Aj, θ ≈ half-integral for all j, then | ˆ
D(θ)| large.
◮ If Aj, θ far from half-integral for many j, then | ˆ
D(θ)| small
b b b b b b b b b
θ ∈ [−1
2, 1 2]m
b b b b b b b b b
θ ∈ [−1
2, 1 2]m
b b b b b b b b b
θ ∈ [−1
2, 1 2]m ◮ With high probability,
√ t
ˆ D(θ)dθ > n−Θ(m)
b b b b b b b b b
θ ∈ [−1
2, 1 2]m ◮ With high probability,
√ t
ˆ D(θ)dθ > n−Θ(m)
◮ With high probability,
2 Zm)>1/
√ t
| ˆ D(θ)| < e−Θ(n/m)
Given A, define R ∈ Zm by Ri = Ai1 even ±1 Ai1 odd (chosen uniformly)
Given A, define R ∈ Zm by Ri = Ai1 even ±1 Ai1 odd (chosen uniformly) Recall D = Ax where x ∈ {−1, 1}n. = ⇒ X = D + R ∈ 2Zm.
Given A, define R ∈ Zm by Ri = Ai1 even ±1 Ai1 odd (chosen uniformly) Recall D = Ax where x ∈ {−1, 1}n. = ⇒ X = D + R ∈ 2Zm. Pr[X = 0] = 2m
4 , 1 4 )m
ˆ X(θ)dθ
Given A, define R ∈ Zm by Ri = Ai1 even ±1 Ai1 odd (chosen uniformly) Recall D = Ax where x ∈ {−1, 1}n. = ⇒ X = D + R ∈ 2Zm. Pr[X = 0] = 2m
4 , 1 4 )m
ˆ X(θ)dθ
b b b b b b b b b
θ ∈ [−1
2, 1 2]m
We can compute ˆ R(θ) = E[e2πiR,θ] =
E[e2πiRiθi] =
cos(2πθi)
We can compute ˆ R(θ) = E[e2πiR,θ] =
E[e2πiRiθi] =
cos(2πθi) and therefore
◮ For θ ∈ [−1 4, 1 4]m, 0 ≤ ˆ
R(θ) ≤ 1
We can compute ˆ R(θ) = E[e2πiR,θ] =
E[e2πiRiθi] =
cos(2πθi) and therefore
◮ For θ ∈ [−1 4, 1 4]m, 0 ≤ ˆ
R(θ) ≤ 1
◮ For θ2 ≤ 1 √n, ˆ
R(θ) ≥ 1
2.
We can compute ˆ R(θ) = E[e2πiR,θ] =
E[e2πiRiθi] =
cos(2πθi) and therefore
◮ For θ ∈ [−1 4, 1 4]m, 0 ≤ ˆ
R(θ) ≤ 1
◮ For θ2 ≤ 1 √n, ˆ
R(θ) ≥ 1
2.
So with high probability (over the choice of A) we have
√ t
ˆ D(θ)dθ > n−Θ(m)
2 Zm)>1/
√ t
| ˆ D(θ)| < e−Θ(n/m)
We can compute ˆ R(θ) = E[e2πiR,θ] =
E[e2πiRiθi] =
cos(2πθi) and therefore
◮ For θ ∈ [−1 4, 1 4]m, 0 ≤ ˆ
R(θ) ≤ 1
◮ For θ2 ≤ 1 √n, ˆ
R(θ) ≥ 1
2.
So with high probability (over the choice of A) we have
√ t
ˆ D(θ) · ˆ R(θ)dθ > n−Θ(m)
2Zm)>1/
√ t
| ˆ D(θ) · ˆ R(θ)| < e−Θ(n/m)
θ ∈ [−1
4, 1 4]m
With high probability (over the choice of A), we have
√ t
ˆ D(θ) · ˆ R(θ)dθ > n−Θ(m)
√ t
| ˆ D(θ) · ˆ R(θ)|dθ < e−Θ(n/m)
θ ∈ [−1
4, 1 4]m
With high probability (over the choice of A), we have
√ t
ˆ D(θ) · ˆ R(θ)dθ > n−Θ(m)
√ t
| ˆ D(θ) · ˆ R(θ)|dθ < e−Θ(n/m) Then for n ≥ Θ(m2 log n), we compute Pr[D+R = 0] = 2m
4 , 1 4 )
ˆ D(θ) ˆ R(θ)dθ
θ ∈ [−1
4, 1 4]m
With high probability (over the choice of A), we have
√ t
ˆ D(θ) · ˆ R(θ)dθ > n−Θ(m)
√ t
| ˆ D(θ) · ˆ R(θ)|dθ < e−Θ(n/m) Then for n ≥ Θ(m2 log n), we compute Pr[D+R = 0] = 2m
4 , 1 4 )
ˆ D(θ) ˆ R(θ)dθ > n−Θ(m)−e−Θ(n/m) > 0
θ ∈ [−1
4, 1 4]m
With high probability (over the choice of A), we have
√ t
ˆ D(θ) · ˆ R(θ)dθ > n−Θ(m)
√ t
| ˆ D(θ) · ˆ R(θ)|dθ < e−Θ(n/m) Then for n ≥ Θ(m2 log n), we compute Pr[D+R = 0] = 2m
4 , 1 4 )
ˆ D(θ) ˆ R(θ)dθ > n−Θ(m)−e−Θ(n/m) > 0 = ⇒ disc(A) ≤ 1.
Recall that ˆ D(θ) = n
j=1 cos(2πAj, θ). ◮ t = pm is the expected column sum. ◮ t ≥ log n =
⇒ whp Aj2 ≤ √ 2t for all j. With high probability |Aj, θ| ≤ 1
8 for all j,
Recall that ˆ D(θ) = n
j=1 cos(2πAj, θ). ◮ t = pm is the expected column sum. ◮ t ≥ log n =
⇒ whp Aj2 ≤ √ 2t for all j. With high probability |Aj, θ| ≤ 1
8 for all j, and so
ˆ D(θ) = exp(−2π2θT AATθ ± O(nt2)θ4
2)
Recall that ˆ D(θ) = n
j=1 cos(2πAj, θ). ◮ t = pm is the expected column sum. ◮ t ≥ log n =
⇒ whp Aj2 ≤ √ 2t for all j. With high probability |Aj, θ| ≤ 1
8 for all j, and so
ˆ D(θ) = exp(−2π2θT AATθ ± O(nt2)θ4
2)
≥ exp(−2π2mnθ2
2)
Recall that ˆ D(θ) = n
j=1 cos(2πAj, θ). ◮ t = pm is the expected column sum. ◮ t ≥ log n =
⇒ whp Aj2 ≤ √ 2t for all j. With high probability |Aj, θ| ≤ 1
8 for all j, and so
ˆ D(θ) = exp(−2π2θT AATθ ± O(nt2)θ4
2)
≥ exp(−2π2mnθ2
2)
= ˆ Y (θ) where Y a gaussian with expectation 0, cov matrix mn · I.
Recall that ˆ D(θ) = n
j=1 cos(2πAj, θ). ◮ t = pm is the expected column sum. ◮ t ≥ log n =
⇒ whp Aj2 ≤ √ 2t for all j. With high probability |Aj, θ| ≤ 1
8 for all j, and so
ˆ D(θ) = exp(−2π2θT AATθ ± O(nt2)θ4
2)
≥ exp(−2π2mnθ2
2)
= ˆ Y (θ) where Y a gaussian with expectation 0, cov matrix mn · I.
1 16 √ t
ˆ D(θ)dθ ≥
√n
ˆ Y (θ)dθ ≥ n−Θ(m)
For c > 0 some constant, θ∞ ≤ 1
4,
E
Aj[| cos(2πAj, θ)|] ≤ 1 − min{1
4pθ2
2, c}
For c > 0 some constant, θ∞ ≤ 1
4, θ2 ≥ 1 16 √ t
E
Aj[| cos(2πAj, θ)|] ≤ 1 − min{1
4pθ2
2, c} ≤ 1 − 1
m
For c > 0 some constant, θ∞ ≤ 1
4, θ2 ≥ 1 16 √ t
E
Aj[| cos(2πAj, θ)|] ≤ 1 − min{1
4pθ2
2, c} ≤ 1 − 1
m Write φk(θ) = k
j=1 | cos(2πAj, θ)|.
For θ2 ≥
1 16 √ t, we have E[φk(θ)|φk−1(θ)] ≤ (1 − 1 m)φk−1(θ).
For c > 0 some constant, θ∞ ≤ 1
4, θ2 ≥ 1 16 √ t
E
Aj[| cos(2πAj, θ)|] ≤ 1 − min{1
4pθ2
2, c} ≤ 1 − 1
m Write φk(θ) = k
j=1 | cos(2πAj, θ)|.
For θ2 ≥
1 16 √ t, we have E[φk(θ)|φk−1(θ)] ≤ (1 − 1 m)φk−1(θ). ◮ Let Φk =
1 16 √ t φk(θ)dθ.
For c > 0 some constant, θ∞ ≤ 1
4, θ2 ≥ 1 16 √ t
E
Aj[| cos(2πAj, θ)|] ≤ 1 − min{1
4pθ2
2, c} ≤ 1 − 1
m Write φk(θ) = k
j=1 | cos(2πAj, θ)|.
For θ2 ≥
1 16 √ t, we have E[φk(θ)|φk−1(θ)] ≤ (1 − 1 m)φk−1(θ). ◮ Let Φk =
1 16 √ t φk(θ)dθ.
◮ E[Φk | Φk−1] ≤ (1 − 1 m) · Φk−1
For c > 0 some constant, θ∞ ≤ 1
4, θ2 ≥ 1 16 √ t
E
Aj[| cos(2πAj, θ)|] ≤ 1 − min{1
4pθ2
2, c} ≤ 1 − 1
m Write φk(θ) = k
j=1 | cos(2πAj, θ)|.
For θ2 ≥
1 16 √ t, we have E[φk(θ)|φk−1(θ)] ≤ (1 − 1 m)φk−1(θ). ◮ Let Φk =
1 16 √ t φk(θ)dθ.
◮ E[Φk | Φk−1] ≤ (1 − 1 m) · Φk−1 ◮ Can use martingale concentration to get that
Φn ≤ e−Θ(n/m) with probability at least 1 − e−Θ(n/m).
For c > 0 some constant, θ∞ ≤ 1
4, θ2 ≥ 1 16 √ t
E
Aj[| cos(2πAj, θ)|] ≤ 1 − min{1
4pθ2
2, c} ≤ 1 − 1
m Write φk(θ) = k
j=1 | cos(2πAj, θ)|.
For θ2 ≥
1 16 √ t, we have E[φk(θ)|φk−1(θ)] ≤ (1 − 1 m)φk−1(θ). ◮ Let Φk =
1 16 √ t φk(θ)dθ.
◮ E[Φk | Φk−1] ≤ (1 − 1 m) · Φk−1 ◮ Can use martingale concentration to get that
Φn ≤ e−Θ(n/m) with probability at least 1 − e−Θ(n/m).
◮ But Φn =
1 16 √ t | ˆ
D(θ)|dθ is exactly the quantity we wanted to bound.
θ ∈ [−1
4, 1 4]m
With high probability (over the choice of A), we have
√ t
ˆ D(θ) · ˆ R(θ)dθ > n−Θ(m)
√ t
| ˆ D(θ) · ˆ R(θ)|dθ < e−Θ(n/m) Then for n ≥ Θ(m2 log n), we compute Pr[D+R = 0] = 2m
4 , 1 4 )
ˆ D(θ) ˆ R(θ)dθ > n−Θ(m)−e−Θ(n/m) > 0 = ⇒ disc(A) ≤ 1.
(1) Can we come up with an algorithm to actually find good colorings?
(1) Can we come up with an algorithm to actually find good colorings? (2) What about random set systems for n < m2?
(1) Can we come up with an algorithm to actually find good colorings? (2) What about random set systems for n < m2? (3) Can we use similar techniques to prove Spencer’s theorem (or Beck-Fiala conjecture?)
(1) Can we come up with an algorithm to actually find good colorings? (2) What about random set systems for n < m2? (3) Can we use similar techniques to prove Spencer’s theorem (or Beck-Fiala conjecture?)
◮ Would need to get stronger decay from R
(1) Can we come up with an algorithm to actually find good colorings? (2) What about random set systems for n < m2? (3) Can we use similar techniques to prove Spencer’s theorem (or Beck-Fiala conjecture?)
◮ Would need to get stronger decay from R ◮ Candidate R: For i = 1, ..., m Ri chosen as a sum of ∆
4.