Identifying Undirected Network Structure via Semidefinite Relaxation - - PowerPoint PPT Presentation

identifying undirected network structure via semidefinite
SMART_READER_LITE
LIVE PREVIEW

Identifying Undirected Network Structure via Semidefinite Relaxation - - PowerPoint PPT Presentation

Identifying Undirected Network Structure via Semidefinite Relaxation Rasoul Shafipour, Santiago Segarra , Antonio G. Marques and Gonzalo Mateos Institute for Data, Systems, and Society Massachusetts Institute of Technology segarra@mit.edu


slide-1
SLIDE 1

Identifying Undirected Network Structure via Semidefinite Relaxation

Rasoul Shafipour, Santiago Segarra, Antonio G. Marques and Gonzalo Mateos

Institute for Data, Systems, and Society Massachusetts Institute of Technology segarra@mit.edu http://www.mit.edu/~segarra/

ICASSP, April 20, 2018

Santiago Segarra 1 / 18

slide-2
SLIDE 2

Network Science analytics

Clean energy and grid analy,cs Online social media Internet

◮ Desiderata: Process, analyze and learn from network data [Kolaczyk09]

Santiago Segarra 2 / 18

slide-3
SLIDE 3

Network Science analytics

Clean energy and grid analy,cs Online social media Internet

◮ Desiderata: Process, analyze and learn from network data [Kolaczyk09] ◮ Network as graph G: encode pairwise relationships ◮ Sometimes both G and data at the nodes are available

⇒ Leverage G to process network data ⇒ Graph Signal Processing

Santiago Segarra 2 / 18

slide-4
SLIDE 4

Network Science analytics

Clean energy and grid analy,cs Online social media Internet

◮ Desiderata: Process, analyze and learn from network data [Kolaczyk09] ◮ Network as graph G: encode pairwise relationships ◮ Sometimes both G and data at the nodes are available

⇒ Leverage G to process network data ⇒ Graph Signal Processing

◮ Sometimes we have access to network data but not to G itself

⇒ Leverage the relation between them to infer G from the data

Santiago Segarra 2 / 18

slide-5
SLIDE 5

Graph signal processing (GSP)

◮ Undirected G with adjacency matrix A

⇒ Aij = Proximity between i and j

◮ Define a signal x on top of the graph

⇒ xi = Signal value at node i

2 3 1 4 5 x1 x2 x3 x4 x5

Santiago Segarra 3 / 18

slide-6
SLIDE 6

Graph signal processing (GSP)

◮ Undirected G with adjacency matrix A

⇒ Aij = Proximity between i and j

◮ Define a signal x on top of the graph

⇒ xi = Signal value at node i

2 3 1 4 5 x1 x2 x3 x4 x5 ◮ Associated with G is the graph-shift operator S = VΛVT ∈ MN

⇒ Sij = 0 for i = j and (i, j) ∈ E (local structure in G) ⇒ Ex: adjacency A and Laplacian L = D − A matrices

Santiago Segarra 3 / 18

slide-7
SLIDE 7

Graph signal processing (GSP)

◮ Undirected G with adjacency matrix A

⇒ Aij = Proximity between i and j

◮ Define a signal x on top of the graph

⇒ xi = Signal value at node i

2 3 1 4 5 x1 x2 x3 x4 x5 ◮ Associated with G is the graph-shift operator S = VΛVT ∈ MN

⇒ Sij = 0 for i = j and (i, j) ∈ E (local structure in G) ⇒ Ex: adjacency A and Laplacian L = D − A matrices

◮ Graph filters H : RN → RN are maps between graph signals

⇒ Polynomial in S with coefficients h ∈ RL+1 ⇒ H := L

l=0 hlSl ◮ How to use GSP to infer the graph topology?

Santiago Segarra 3 / 18

slide-8
SLIDE 8

Topology inference: Motivation and context

◮ Network topology inference from nodal observations [Kolaczyk09]

◮ Partial correlations and conditional dependence [Dempster74] ◮ Sparsity [Friedman07] and consistency [Meinshausen06] ◮ [Banerjee08], [Lake10], [Slawski15], [Karanikolas16]

◮ Key in neuroscience [Sporns10]

⇒ Functional net inferred from activity

Santiago Segarra 4 / 18

slide-9
SLIDE 9

Topology inference: Motivation and context

◮ Network topology inference from nodal observations [Kolaczyk09]

◮ Partial correlations and conditional dependence [Dempster74] ◮ Sparsity [Friedman07] and consistency [Meinshausen06] ◮ [Banerjee08], [Lake10], [Slawski15], [Karanikolas16]

◮ Key in neuroscience [Sporns10]

⇒ Functional net inferred from activity

◮ Noteworthy GSP-based approaches

◮ Gaussian graphical models [Egilmez16] ◮ Smooth signals [Dong15], [Kalofolias16] ◮ Stationary signals [Pasdeloup15], [Segarra16] ◮ Directed graphs [Mei15], [Shen16] ◮ Low-rank excitation [Wai18]

◮ Our contribution: topology inference from non-stationary graph signals

Santiago Segarra 4 / 18

slide-10
SLIDE 10

Problem formulation

◮ Underlying graph G with undirected unknown GSO S ◮ Observe signals {yi}K i=1 defined on the unknown graph

Setup y1 y2 y3

Santiago Segarra 5 / 18

slide-11
SLIDE 11

Problem formulation

◮ Underlying graph G with undirected unknown GSO S ◮ Observe signals {yi}K i=1 defined on the unknown graph

Setup y1 y2 y3 Problem statement Given observations {yi}K

i=1, determine the network S knowing that:

{yi}K

i=1 are outputs of a diffusion process on S.

Santiago Segarra 5 / 18

slide-12
SLIDE 12

Problem formulation

◮ Consider an arbitrary linear network process on the GSO S

⇒ Every realization corresponds to a different input xi yi = L

  • l=0

hlSl

  • xi = Hxi,

i = 1, . . . , K

◮ Goal: Recover S from the observation of K signals {yi}K i=1 ◮ Additional unknowns

⇒ The degree of the filter L ⇒ The filter coefficients {hl}L

l=0

⇒ The specific inputs xi; but we know that xi ∼ N(0, Cx)

Santiago Segarra 6 / 18

slide-13
SLIDE 13

Blueprint of our solution

STEP%2:%Find% eigenvalues%via%

  • p5miza5on%

A%priori%info%and% desirable%features%%

%

STEP%1:%Es5mate% the%eigenvectors%of% % %

S

ˆ S

{yi}K

i=1

Santiago Segarra 7 / 18

slide-14
SLIDE 14

Blueprint of our solution

STEP%2:%Find% eigenvalues%via%

  • p5miza5on%

A%priori%info%and% desirable%features%%

%

STEP%1:%Es5mate% the%eigenvectors%of% % %

S

ˆ S

Sparsity%and% GSO%feasibility%

ˆ V :%noisy%

{yi}K

i=1

Santiago Segarra 7 / 18

slide-15
SLIDE 15

Step 1: Estimating the eigenvectors of S

◮ y is the output of a local diffusion process on the graph

y=α0

  • l=1

(I − αlS)x = N−1

  • l=0

hl Sl

  • x := Hx

◮ Whenever the input x is white

⇒ graph stationary process on S [Marques17, Girault15, Perraudin17]

Santiago Segarra 8 / 18

slide-16
SLIDE 16

Step 1: Estimating the eigenvectors of S

◮ y is the output of a local diffusion process on the graph

y=α0

  • l=1

(I − αlS)x = N−1

  • l=0

hl Sl

  • x := Hx

◮ Whenever the input x is white

⇒ graph stationary process on S [Marques17, Girault15, Perraudin17] Stationary case

◮ The covariance Cy of y shares V with S

Cy = H2 = h2

0I + 2h0h1S + h2 1S2 + ... ◮ Estimate covariance from {yi}K i=1 as ˆ

Cy ⇒ Diagonalize ⇒ Obtain ˆ V

Santiago Segarra 8 / 18

slide-17
SLIDE 17

Non-stationary graph signals

◮ Q: What if the signal y = Hx is not stationary (i.e., x colored)?

⇒ Matrices S and Cy no longer simultaneously diagonalizable since Cy = HCxH

Santiago Segarra 9 / 18

slide-18
SLIDE 18

Non-stationary graph signals

◮ Q: What if the signal y = Hx is not stationary (i.e., x colored)?

⇒ Matrices S and Cy no longer simultaneously diagonalizable since Cy = HCxH

◮ Key: still H = L−1 l=0 hlSl diagonalized by the eigenvectors V of S

⇒ Infer V by estimating the unknown diffusion (graph) filter H ⇒ Step 1 boils down to system identification + eigendecomposition

%

System% Iden5fica5on% %

ˆ V

{yi}K

i=1

%

Eigendecomposi5on% %

ˆ H

Santiago Segarra 9 / 18

slide-19
SLIDE 19

System identification

Define Cxyx := C1/2

x

CyC1/2

x

, with eigenvectors Vxyx. If Cx is non- singular then all admissible symmetric filters H are of the form H = C−1/2

x

C1/2

xyx Vxyxdiag(b)VT xyxC−1/2 x

, where b ∈ {−1, 1}N is a binary (signed) vector.

Santiago Segarra 10 / 18

slide-20
SLIDE 20

System identification

Define Cxyx := C1/2

x

CyC1/2

x

, with eigenvectors Vxyx. If Cx is non- singular then all admissible symmetric filters H are of the form H = C−1/2

x

C1/2

xyx Vxyxdiag(b)VT xyxC−1/2 x

, where b ∈ {−1, 1}N is a binary (signed) vector.

◮ Even if we get Cy exactly, H is not identifiable

⇒ Not surprising since we only have second moment info

◮ Consider having access to multiple input distributions {Cx,m}M m=1

Santiago Segarra 10 / 18

slide-21
SLIDE 21

Multiple input processes

◮ Define Am := (C−1/2 x,m Vxyx,m) ⊙ (C−1/2 x,m C1/2 xyx,mVxyx,m)

Ψ :=      A1 −A2 · · · A2 −A3 · · · . . . . . . . . . ... . . . . . . · · · AM−1 −AM     

◮ bm ∈ {−1, 1}N and b = [bT 1 , bT 2 , . . . , bT M]T, then Ψb∗ = 0

Santiago Segarra 11 / 18

slide-22
SLIDE 22

Multiple input processes

◮ Define Am := (C−1/2 x,m Vxyx,m) ⊙ (C−1/2 x,m C1/2 xyx,mVxyx,m)

Ψ :=      A1 −A2 · · · A2 −A3 · · · . . . . . . . . . ... . . . . . . · · · AM−1 −AM     

◮ bm ∈ {−1, 1}N and b = [bT 1 , bT 2 , . . . , bT M]T, then Ψb∗ = 0

Whenever only estimates ˆ Cy,m are available, we can estimate b∗ as ˆ b∗ = argmin

b∈{−1,1}NM bT ˆ

Ψ

T ˆ

Ψb,

  • btaining our estimate for the filter H as

ˆ H = 1 M

M

  • m=1

C−1/2

x,m ˆ

C1/2

xyx,m ˆ

Vxyx,mdiag(ˆ b∗

m)ˆ

VT

xyx,mC−1/2 x,m

Santiago Segarra 11 / 18

slide-23
SLIDE 23

Boolean quadratic program

◮ Our problem then reduces to solving the BQP

ˆ b∗ = argmin

b∈{−1,1}NM bT ˆ

Ψ

T ˆ

Ψb

◮ Define ˆ

W = ˆ Ψ

T ˆ

Ψ and B = bbT min

B0 tr( ˆ

WB)

  • s. to rank(B) = 1, Bii = 1, i = 1, . . . , NM

◮ Drop source of non-convexity to obtain the semi-definite relaxation

B∗ = argmin

B0

tr( ˆ WB)

  • s. to Bii = 1, i = 1, . . . , NM

Santiago Segarra 12 / 18

slide-24
SLIDE 24

Performance guarantee

◮ For l = 1, . . . , L, draw zl ∼ N(0, B∗), round ˜

bl = sign(zl), to obtain l∗ = argmin

l=1,...,L

˜ bT

l ˆ

W˜ bl

Santiago Segarra 13 / 18

slide-25
SLIDE 25

Performance guarantee

◮ For l = 1, . . . , L, draw zl ∼ N(0, B∗), round ˜

bl = sign(zl), to obtain l∗ = argmin

l=1,...,L

˜ bT

l ˆ

W˜ bl Let ˆ b∗ be the true solution of the BQP and let ˜ bl∗ be the output of

  • ur method. Then,

(ˆ b∗)T ˆ Wˆ b∗ ≤ E

bl∗)T ˆ W˜ bl∗

  • ≤ 2

π (ˆ b∗)T ˆ Wˆ b∗ + γ, where γ =

  • 1 − 2

π

  • λmaxNM.

Santiago Segarra 13 / 18

slide-26
SLIDE 26

Blueprint of our solution

STEP%2:%Find% eigenvalues%via%

  • p5miza5on%

A%priori%info%and% desirable%features%%

%

STEP%1:%Es5mate% the%eigenvectors%of% % %

S

ˆ S

Sparsity%and% GSO%feasibility%

ˆ V :%noisy%

{yi}K

i=1

Santiago Segarra 14 / 18

slide-27
SLIDE 27

Step 2: Obtaining the eigenvalues

◮ We can use extra knowledge/assumptions to choose one graph

⇒ Of all graphs, select one that is optimal in some sense ˆ S := argmin

S,λ

f (S, λ)

  • s. to

S =

N

  • k=1

λkvkvT

k , S ∈ S ◮ Set S contains all admissible scaled adjacency matrices

S :={S | Sij ≥ 0, S∈MN , Sii = 0,

j S1j =1}

⇒ Can accommodate Laplacian matrices as well

Santiago Segarra 15 / 18

slide-28
SLIDE 28

Step 2: Obtaining the eigenvalues

◮ We can use extra knowledge/assumptions to choose one graph

⇒ Of all graphs, select one that is optimal in some sense ˆ S := argmin

S,λ

f (S, λ)

  • s. to

S =

N

  • k=1

λkvkvT

k , S ∈ S ◮ Set S contains all admissible scaled adjacency matrices

S :={S | Sij ≥ 0, S∈MN , Sii = 0,

j S1j =1}

⇒ Can accommodate Laplacian matrices as well

◮ Problem is convex if we select a convex objective f (S, λ)

Ex: Sparsity (f (S) = S1), min. energy (f (S) = SF), mixing (f (λ) = −λ2)

◮ Robust recovery from imperfect or incomplete ˆ

V [Segarra16]

Santiago Segarra 15 / 18

slide-29
SLIDE 29

Unveiling urban mobility patterns

◮ Detect mobility patterns in New York City from Uber pickup data ◮ Times and locations (N = 30) from January 1st to June 29th 2015 ◮ M = 2 graph processes: weekday (m = 1) and weekend (m = 2) pickups ◮ Pickups within 6-11am as input signal x and 3-8pm as output y

Santiago Segarra 16 / 18

slide-30
SLIDE 30

Conclusion

STEP%2:%Find% eigenvalues%via%

  • p5miza5on%

A%priori%info%and% desirable%features%%

%

STEP%1:%Es5mate% the%eigenvectors%of% % %

S

ˆ S

Sparsity%and% GSO%feasibility%

ˆ V {yi}K

i=1

Santiago Segarra 17 / 18

slide-31
SLIDE 31

Conclusion

STEP%2:%Find% eigenvalues%via%

  • p5miza5on%

A%priori%info%and% desirable%features%%

%

STEP%1:%Es5mate% the%eigenvectors%of% % %

S

ˆ S

Sparsity%and% GSO%feasibility%

ˆ V

{xi}K

i=1 ∼ N(0, Cx)

{yi}K

i=1

H(S)

Santiago Segarra 17 / 18

slide-32
SLIDE 32

Conclusion

STEP%2:%Find% eigenvalues%via%

  • p5miza5on%

A%priori%info%and% desirable%features%%

ˆ S

Sparsity%and% GSO%feasibility%

ˆ V

{xi}K

i=1 ∼ N(0, Cx) System%ID% Eigendecomposi5on%

ˆ H

{yi}K

i=1

H(S)

Santiago Segarra 17 / 18

slide-33
SLIDE 33

Conclusion

STEP%2:%Find% eigenvalues%via%

  • p5miza5on%

A%priori%info%and% desirable%features%%

ˆ S

Sparsity%and% GSO%feasibility%

ˆ V

{xi}K

i=1 ∼ N(0, Cx)

H(S)

System%ID% Eigendecomposi5on%

ˆ H

{yi}K

i=1

SemiFdefinite%relaxa5on%of% Boolean%quadra5c%program%

Santiago Segarra 17 / 18

slide-34
SLIDE 34

GlobalSIP’18 Symposium on GSP

Symposium on Graph Signal Processing

Topics of interest

· Graph-signal transforms and filters · Distributed and non-linear graph SP · Statistical graph SP · Prediction and learning for graphs · Network topology inference · Recovery of sampled graph signals · Control of network processes · Signals in high-order and multiplex graphs · Neural networks for graph data · Topological data analysis · Graph-based image and video processing · Communications, sensor and power networks · Neuroscience and other medical fields · Web, economic and social networks

Paper submission due: June 17, 2018

2018 6th IEEE Global Conference on Signal and Information Processing

November 26-28, 2018 Anaheim, California, USA http://2018.ieeeglobalsip.org/

Organizers:

Gonzalo Mateos (Univ. of Rochester) Santiago Segarra (MIT) Sundeep Chepuri (TU Delft)

Santiago Segarra 18 / 18