The interplay of analysis and algorithms (or, Computational Harmonic - - PowerPoint PPT Presentation

the interplay of analysis and algorithms
SMART_READER_LITE
LIVE PREVIEW

The interplay of analysis and algorithms (or, Computational Harmonic - - PowerPoint PPT Presentation

The interplay of analysis and algorithms (or, Computational Harmonic Analysis) Anna Gilbert University of Michigan supported by DARPA-ONR, NSF, and Sloan Foundation Two themes Sparse representation Represent or approximate signal, function


slide-1
SLIDE 1

The interplay of analysis and algorithms

(or, Computational Harmonic Analysis)

Anna Gilbert University of Michigan

supported by DARPA-ONR, NSF, and Sloan Foundation

slide-2
SLIDE 2

Two themes

slide-3
SLIDE 3

Sparse representation

Represent or approximate signal, function by a linear combination of a few atomic elements

slide-4
SLIDE 4

Compressed Sensing

Noisy, sparse signals can be approximately reconstructed from a small number of linear measurements

slide-5
SLIDE 5

Recovery = find significant entries Sparse representation = signal recovery different input models

slide-6
SLIDE 6

How to compute? Analysis and algorithms are both key components

slide-7
SLIDE 7

SPARSE

Signal space: dimension Dictionary: finite collection of unit norm atoms Representation: linear combination of atoms Find best -term representation

d D = {φω : ω ∈ Ω}, |Ω| = N > d s =

  • λ∈Λ

cλφλ

m

slide-8
SLIDE 8

Applications

Approximation theory Signal/Image compression Scientific computing, numerics Data mining, massive data sets Generalized decoding Modern, hyperspectral imaging systems Medical imaging

slide-9
SLIDE 9

SPARSE is NP-HARD SPARSE is NP-COMPLETE

slide-10
SLIDE 10

If dictionary is ONB, then SPARSE is easy (in polynomial time)

slide-11
SLIDE 11

Incoherent dictionaries (a basic result)

  • coherent dictionary, = smallest angle

between vectors = number of terms in sparse representation Algorithm returns -term approx. with error Two-phase greedy pursuit

Joint work with Tropp, Muthukrishnan, and Strauss

µ µ m m < 1 2µ x − am ≤

  • 1 +

2µm2 (1 − 2µm)2 x − aOPT m

slide-12
SLIDE 12

Future for sparse approximation

Hardness of approximation is related to hardness of SET COVER Approximability of SET COVER well-studied (Feige, etc.) Need insight from previous work in TCS Geometry is critical in sparse approximation Need a way to describe better geometry of dictionary and its relation to sparse approximation: VC dimension? Methods for constructing “good” redundant dictionaries (data dependent?) Watch the practitioners!

slide-13
SLIDE 13

Exponential time Polynomial time Linear time Logarithmic time General SPARSE SPARSE, geometry Matrix multiplication FFT AAFFT Chaining, HHS Pursuit Streaming wavelets, etc.

O(d) O(log d)

O(d2)

O(2d)

slide-14
SLIDE 14

Computational Resources

Time Space Randomness Communication

slide-15
SLIDE 15

Models: Sampling

=

m-sparse signal, length d measurements: length N = m log d

slide-16
SLIDE 16

Models: linear measurements

=

m-sparse signal, length d measurements: length N = m log d

slide-17
SLIDE 17

Models: Dictionary

Orthonormal bases Fourier Wavelets Spikes Redundant dictionaries Piecewise constants Wavelet packets Chirps

slide-18
SLIDE 18

Results: Fourier

Theorem: On signal with length , AAFFT builds -term Fourier representation in time using samples with error On each signal, succeed with high probability.

m

r

mpoly(log d/) mpoly(log d/) s − r2 ≤ (1 + )s − sm2 s d

G., Muthukrishnan, and Strauss 2005

slide-19
SLIDE 19

Why sublinear resources?

slide-20
SLIDE 20

Sparsogram

!"#$%&"'()& *+$,-$'./%0"' 11223%456+4)7+6#8%46#59$4:;<=>=?@%+-'%!"#$:A<?B; % % C ; D E F ;A EA >A =A CAA A A<; A<E A<> A<= C !"#$%&"'()& *+$,-$'./%0"' 223G%456+4)7+6#8%46#59$4:CAA@%+-'%!"#$:A<?=E % % C ; D E F ;A EA >A =A CAA A A<; A<E A<> A<= C !"#$%&"'()& *+$,-$'./%0"' 11223%$++)+%"'%456+4)7+6# % % C ; D E F ;A EA >A =A CAA A A<; A<E A<> A<= C C ; D E H%CA

>

!E !; A ; E > IH6#59$%)*%')"4/%"'5-!%4"7'69%)'%$6.J%&"'()& !"#$

slide-21
SLIDE 21

Generalize Fourier sampling algorithm to sublinear algorithm for linear chirps Multi-user detection for wireless comm. Radar detection and identification

Extensions, applications

Calderbank, G., and Strauss 2006 Lepak, Strauss, and G.

slide-22
SLIDE 22

Results: Wavelets

Theorem: On signal with length , streaming algorithm builds -term wavelet representation in time using linear measurements with error On each signal, succeed with high probability.

m

r

s − r2 ≤ (1 + )s − sm2 s d poly(m log d/) poly(m log d/)

G., Guha, Indyk, Kotidis, Muthukrishnan, and Strauss 2001

slide-23
SLIDE 23

Results: Chaining

  • Theorem: With probability at least , the random

measurement matrix has the following property. Suppose that is a d-dimensional signal whose best m-term approximation with respect to norm is . Given the sketch of size and the number m, the Chaining Pursuit algorithm produces a signal with at most O(m) nonzero entries. This signal estimate satisfies The time cost of the algorithm is

v = Φs

1 − d−3

Φ s 1 sm

O(m log2 d)

s − s1 ≤ C log ms − sm1

  • s

O(m log2(m) log2(d))

G., Strauss, Tropp, and Vershynin 2006

slide-24
SLIDE 24

Algorithmic linear dimension reduction in

Theorem: Let be a set of points in endowed with the norm. Assume that each point has at most non-zero coordinates. These points can be linearly embedded in with distortion , using only

  • dimensions. Moreover, we can reconstruct a point

from its low-dimensional sketch in time

G., Strauss, Tropp, and Vershynin 2006

1

1 O(m log2 d) O(log3(m) log2(d)) O(m log2(m) log2(d)) m 1 Y Rd

slide-25
SLIDE 25

Results: HHS

  • Theorem: With probability at least , the random

measurement matrix has the following property. Suppose that is a d-dimensional signal whose m largest entries are given by . Given the sketch of size and the number m, the HHS Pursuit algorithm produces a signal with m nonzero entries. This signal estimate satisfies The time cost of the algorithm is

s − s2 ≤ s − sm2 +

  • √ms − sm1

G., Strauss, Tropp, and Vershynin 2007

v = Φs

1 − d−3

Φ s sm

  • s

mpolylog(d)/2 m2polylog(d)/4

slide-26
SLIDE 26

Desiderata

Uniformity: Sketch works for all signals

simultaneously

Optimal Size: measurements Optimal Speed: Update and output times

are

Must have high quality: answer to query has

near-optimal error mpolylog(d) mpolylog(d)

slide-27
SLIDE 27

less information measure less compute less

slide-28
SLIDE 28

Related Work

Remark: Numerous contributions in area are not strictly comparable

Gilbert et al. 2001, 2005: Cormode-Muthukrishnan 2005; Candes-(Romberg)-Tao 2004, 2005; Donoho 2004, 2005....

Reference Uniform

  • Opt. Storage
  • Sublin. Query

GMS X

  • CM
  • X
  • CRT, Don
  • X

Chaining

  • HHS
slide-29
SLIDE 29

More formally....

slide-30
SLIDE 30

Signal Information Recovery

Golomb-Weinberger 1959

signal space statistic space information space statistic map information map (measurements) recovery algorithm

Φ U A

Ω UΩ ΦΩ

slide-31
SLIDE 31

More Formal Framework...

What signal class are we interested in? What statistic are we trying to compute? How much nonadaptive information is necessary to do so? What type of information? Point samples? Inner products? Deterministic or random information? How much storage does the measurement operator require? How much computation time, space does the algorithm use? How much communication is necessary?

slide-32
SLIDE 32

Computational Harmonic Analysis? Algorithmic Harmonic Analysis = AHA!

slide-33
SLIDE 33

http://www.math.lsa.umich.edu/~annacg annacg@umich.edu

slide-34
SLIDE 34

Isolation = Approximate Group Testing

slide-35
SLIDE 35

Want to find spikes at height , Assign positions into groups by

  • f spikes isolated

groups have groups have single spike and low noise except with probability Union bound over all spike configurations

Approximate group testing

1/m

m d noise1 = 1 n = m log d

Φ

≥ c1m ≤ c2m noise ≥ 1/(2m) ≥ (c1 − c2)m e(−m log d) m