A Rank-Constrained Optimization approach: Application to Factor - - PowerPoint PPT Presentation

a rank constrained optimization approach application to
SMART_READER_LITE
LIVE PREVIEW

A Rank-Constrained Optimization approach: Application to Factor - - PowerPoint PPT Presentation

A Rank-Constrained Optimization approach: Application to Factor Analysis Ramn A. Delgado, Juan C. Agero, Graham C. Goodwin School of Electrical Engineering and Computer Science, The University of Newcastle, Australia. Handling the Rank


slide-1
SLIDE 1

A Rank-Constrained Optimization approach: Application to Factor Analysis

Ramón A. Delgado, Juan C. Agüero, Graham C. Goodwin

School of Electrical Engineering and Computer Science, The University of Newcastle, Australia.

slide-2
SLIDE 2

1

Handling the Rank Constraints

2

Some Existing Representations

3

A new representation of rank constraints

4

Rank-Constrained Optimisation

5

Application to Factor Analysis

6

Conclusions

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 2 / 22

slide-3
SLIDE 3

Motivation

why include rank constraints?

Complexity of a model (e.g rank of a Hankel matrix) Low-Rank Decomposition (Principal Components Analysis) Recent interest on sparse representations.

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 4 / 22

slide-4
SLIDE 4

Some Difficulties with rank constraints rank {X} ≤ r Non-differentiable Nonlinear Combinatorial nature in optimization Find a differentiable representation for the rank constraints, and hopefully reduce the number of nonlinearities.

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 5 / 22

slide-5
SLIDE 5

Some existing representations

By construction

rank {X} ≤ r ⇐ ⇒ X = A Ir

  • B.

rank {X} ≤ r ⇐ ⇒ X = UV. where A ∈ Rm×m, B ∈ Rn×n, U ∈ Rm×r, V ∈ Rr×n. Disadvantage Already assign a structure to X.

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 7 / 22

slide-6
SLIDE 6

Some existing representations

Using the Characteristic Polynomial

Consider that ci(X), for i = 1, . . . , n are the coefficients of the characteristic polynomial of X. X ∈ Sn

+,

[Helmersson 2009] rank {X} ≤ r ⇐ ⇒ cn−r−1(−X) = 0. X ∈ Sn

+,

[d’Aspremont 2003] rank {X} = minv∈Rn n

i=1 vi

s.t. ci(G)(1 − vi) = 0, vi ≥ 0 for i = 1, . . . , n. Disadvantages Only valid for X ∈ Sn

+.

ci(X) is, in general, nonlinear.

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 8 / 22

slide-7
SLIDE 7

Some existing representations used in optimization

Closely related to our representation

  • I. Markovsky rank {X} ≤ r ⇐

⇒ ∃R ∈ R(m−r)×m such that RX = 0m−r×m and R is full row rank X ∈ Rm×n, but a rank constraint is now imposed on an auxiliary matrix.

  • J. Dattorro X ∈ Sn

+,

rank {X} ≤ r ⇐ ⇒ ∃W ∈ Φn,r such that trace(WX) = 0. where Φn,r = {W ∈ Sn, 0 W I, trace(W) = n − r} Only valid for X ∈ Sn

+.

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 9 / 22

slide-8
SLIDE 8

Main result1 Theorem Let X ∈ Rm×n then rank {X} ≤ r ⇐ ⇒ ∃ W ∈ Φn,r, such that GX = 0m×n where Φn,r = {W ∈ Sn, 0 W I, trace(W) = n − r}

1submitted for publication Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 11 / 22

slide-9
SLIDE 9

Advantages Differentiable Freedom to impose a desired structure on X (e.g. Hankel). Generalisation of Dattorro’s result Avoid some difficulties in Markovsky and d’Aspremont’s results. Disadvantage We still have a bilinear condition WX = 0m×n.

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 12 / 22

slide-10
SLIDE 10

Rank-Constrained Optimisation An equivalent representation for rank-constrained optimization problems Prco : min

θ∈Rp f(θ)

s.t. θ ∈ Ω rank {X(θ)} ≤ r ≡ Pbi : min

θ∈Rp min W∈Sn f(θ)

s.t. θ ∈ Ω X(θ)W = 0m×n W ∈ Φn,r

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 14 / 22

slide-11
SLIDE 11

Rank-Constrained Optimisation We have developed a local optimisation method For the case X(θ) ∈ Sn

+, a global optimisation method.

The bilinear constraint can be imposed in several ways: X(θ) ∈ Rm×n X(θ)W = 0 ⇐ ⇒ X(θ)W = 0m×n X(θ) ∈ Sn

+

trace(X(θ)W) = 0 ⇐ ⇒ X(θ)W = 0n×n

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 15 / 22

slide-12
SLIDE 12

Factor Analysis Consider a measured output yk ∈ RN, factors fk ∈ Rr, idiosyncratic noise vk ∈ RN, and a model: yk = Afk + vk (1) where A ∈ RN×n is a tall matrix. fk ∼ N(0, Φ) (2) vk ∼ N(0, Ψ) (3) Then, yk ∼ N(0, Σ), where Σ is given by Σ = AΦA⊤ + Ψ

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 17 / 22

slide-13
SLIDE 13

Sparse Noise Covariance Existing approaches require Ψ diagonal.

◮ Ψ = σ2I, then there is a closed-form solution, e.g. PCA. ◮ Ψ diagonal, e.g. Maximum Likelihood (EM algorithm).

Considering the advances on sparse representations. We propose to assume that Ψ is sparse. Prcofa : min

Ψ∈SN Ψ1

subject to rank {Σ − Ψ} ≤ r Ψ 0 Σ − Ψ 0

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 18 / 22

slide-14
SLIDE 14

Numerical Example

Local Optimization method

Consider r = 3 factors, N = 20 measured outputs and T = 100 Samples, Ψij = (0.7)|i−j|. and the performance index d(Pm) = 1 − trace(APmA⊤) trace(AA⊤) Method d(·) Total Time [s] PCA 0.1992 0.0136 RCO 0.1002 274.6732 EM 0.1330 0.0198

Table: Mean value over Nmc = 100 Monte Carlo simulations of d(·).

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 19 / 22

slide-15
SLIDE 15

Conclusions We have developed a new representation of rank constraints.

◮ Second-order differentiable. ◮ Avoid several nonlinearities, excepting a bilinear constraint.

We have developed two optimization algorithms.

◮ Local Optimisation. ◮ Global optimisation.

We have applied the method:

◮ Factor Analysis with Correlated Errors. ◮ Sparse Control (tomorrow FrA04.2). Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 21 / 22

slide-16
SLIDE 16

Thanks for your attention! Any questions?

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 22 / 22

slide-17
SLIDE 17

Global Optimization Example Consider r = 1 factor, N = 3 measured outputs and T = 100 Samples. d(·)

  • Ψ1

time [s] RCO 0.066 4.46 23.83 RCO-G 0.066 4.46 310.19 d(·) vec( Ψ)1 time [s] RCO 0.165 11.12 43.20 RCO-G 0.218 9.17 1768.31

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 23 / 22

slide-18
SLIDE 18

Solving the Optimisation Problem Proposed method Solve Pbi iteratively. In each iteration, solves a feasibility problem that deals with the bilinear constraint. Given the estimate θm ∈ Ω, at iteration m min

θ∈Rp min W∈Sn X(θ)W

subject to f(θ) ≤ f(θm) − ηm θ ∈ Ω W ∈ Φn,r

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 24 / 22

slide-19
SLIDE 19

Branch and Bound

Global Optimum 1st 2nd

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 25 / 22

slide-20
SLIDE 20

ℓ1-norm Computationally efficient. Unintuitive to chose the sparse parameters Poor handling of group-constraints ℓ0-norm Computationally more expensive. ℓ0-norm constraints have a clear interpretation. Can handle Group constraints. ℓ0-norm equivalence

Pℓ0 : min

θ∈Rp f(θ)

s.t. θ ∈ Ω θ0 ≤ r ≡ Peq : min

θ∈Rp min w∈Rp f(θ)

s.t. θ ∈ Ω θ ◦ w = 0 0 ≤ w ≤ 1 1⊤w = p − r

Delgado, Agüero & Goodwin 2014 (IFAC 2014). Delgado, Agüero & Goodwin submitted to Automatica. Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 26 / 22

slide-21
SLIDE 21

Local Optimization

Ψ (d) True (e) PCA (f) EM (g) RCO

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 27 / 22

slide-22
SLIDE 22

Concave Minimisation Concave minimization may lead to a combinatorial problem. (e.g. matrix-rank function is quasi-concave on the Positive Semidefinite Cone, thus leads to Concave-minimisation and “reverse-convex” constraints)

Delgado, Agüero, Goodwin (UoN) IFAC 2014, Cape Town 28 / 22