Sparse plus low-rank graphical models of time series
Presented by Rahul Nadkarni Joint work with Nicholas J. Foti, Adrian KC Lee, and Emily B. Fox University of Washington August 14th, 2016
1
Sparse plus low-rank graphical models of time series Presented by - - PowerPoint PPT Presentation
Sparse plus low-rank graphical models of time series Presented by Rahul Nadkarni Joint work with Nicholas J. Foti, Adrian KC Lee, and Emily B. Fox University of Washington August 14 th , 2016 1 Brain Interactions from MEG
Presented by Rahul Nadkarni Joint work with Nicholas J. Foti, Adrian KC Lee, and Emily B. Fox University of Washington August 14th, 2016
1
Magnetoencephalography (MEG) captures weak magnetic field.
Goal: Infer functional connectivity
2
3
edges nodes
No edge (i, j) Xi , Xj conditionallyindependent given rest of variables.
X1 ⊥ ⊥ X2|X3, X4, X5
4
No edge (i,j) time series Xi , Xj conditionallyindependent given entire trajectories of other series.
Natural property for functional connectivity
Examples of existing work: Bach et al. 2004, Songsiri & Vandenberghe 2010, Jung et al. 2015, Tank et al. 2015
latent variables
+
marginalized
variables
latent component
5
Examples of existing work: Chandrasekaran et al. 2012, Jalali & Sanghavi 2012, Liégois et al. 2015
6
7
⇐⇒
Conditional independenceencoded in the precision matrix. Xi , Xj conditionallyindependentgiven rest of variables.
8
X1(t) X2(t) Xp(t)
. . .
lagged covariance:
?
FFT as the Fourier transform of the matrices, Γ(h) = Cov(X(t), X(t + h)):
Spectral density matrix
Lagged covariance matrix
S(λ) =
∞
X
h=−∞
Γ(h)e−iλh
9
10
λ2 λk λT
complex inverse sp spectral densi sity matrices
(Dahlhaus, 2000) For Gaussian stationary time series,
For Gaussian i.i.d. random variables, S(λ)−1 :
11
12
+
i<j
sparsity-inducing penalty
Graphical LASSO (Friedman et al. 2007)
sample covariance matrix inverse covariance matrix
solved with: many existing algorithms
Time Domain Likelihood Frequency Domain Likelihood
Fourier coefficients are asymptotically independent, complex Normal random vectors (Brillinger, 1981)
p(d0, . . . , dT −1|{S(λk)}T −1
k=0 )
p(X(1), . . . , X(T)|[Γ(h)]T −1
h=0 )
Whittle Approximation
k=0 ) ≈
13
T −1
k=0
kS−1 k
dk
Fourier coefficients
k=0
kS−1 k
dk
14
+
Group LASSO penalty
Spectral graphical LASSO (Jung et al. 2015)
sample spectral density matrix inverse spectral density matrix
solved with: ADMM (Jung et al. 2015)
T −1
X
k=0
⇣ − log det Ψ[k] + tr n ˆ S[k]Ψ[k]
X
i<j
v u u t
T −1
X
k=0
|Ψ[k]ij|2
15
activity unrelated to task
activity introduces “point spread”
latent variables
These issues can be addressed by adding a latent component to the model
16
–
⨉ ⨉
low-rank (rank r << p) sparse
p r
S−1 = K = KOO KOH KHO KHH
KO =
hidden-observed hidden-hidden
17
18
⨉ ⨉
sparse penalty: low-rank penalty:
i<j
–
⨉ ⨉
solved with ADMM (Ma et al. 2013) negative log-likelihood:
T −1
X
k=0
tr {L[k]}
Whittle approximation Group LASSO penalty
X
i<j
v u u t
T −1
X
k=0
|Ψ[k]ij|2
19
Used ADMM to solve this convex formulation sparse penalty: low-rank penalty: negative log-likelihood:
T −1
X
k=0
⇣ − log det(Ψ[k] − L[k]) + tr n ˆ S[k](Ψ[k] − L[k])
Multivariate time series data Estimated spectral density
sparse component: low-rank component: graph
time domain frequency domain ADMM
20
21
22
Maintain or Switch attention (Left/Right, High/Low pitch)
sparse component: low-rank component: graph
23