0.15in ECE 18-898G: Special Topics in Signal Processing...

27
1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Robust PCA Yuejie Chi Department of Electrical and Computer Engineering Spring 2018

Transcript of 0.15in ECE 18-898G: Special Topics in Signal Processing...

Page 1: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

1/26

ECE 18-898G: Special Topics in Signal Processing:Sparsity, Structure, and Inference

Robust PCA

Yuejie Chi

Department of Electrical and Computer Engineering

Spring 2018

Page 2: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

2/26

Demixing sparse and low-rank matrices

Suppose we are given a matrix

M = L︸︷︷︸low-rank

+ S︸︷︷︸sparse

∈ Rn×n

Question: Can we hope to recover both L and S from M?

Page 3: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

3/26

Principal component analysis (PCA)

• N samples X = [x1,x2, . . . ,xN ] ∈ Rn×N that are centered

• PCA: seeks r directions that explain most variance of data

minimizeL:rank(L)=r ‖X −L‖F

best rank-r approximation of X

Page 4: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

4/26

Sensitivity to corruptions / outliers

What if some samples are corrupted (e.g. due to sensor errors /attacks)?

Classical PCA fails even with a few outliers

Page 5: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

5/26

Video surveillance

Separation of background (low-rank) and foreground (sparse)

Candes, Li, Ma, Wright ’11

Page 6: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

Graph clustering / community recovery

• n nodes, 2 (or more) clusters

• A friendship graph G: for any pair (i, j),

Mi,j =

1, if (i, j) ∈ G0, else

• Edge density within clusters > edge density across clusters

• Goal: recover cluster structure

6/26

Page 7: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

7/26

Graph clustering / community recovery

M = L︸︷︷︸low-rank

+ M −L︸ ︷︷ ︸sparse

• An equivalent goal: recover ground truth matrix

Li,j =

1, if i and j are in same community0, else

• Clustering ⇐⇒ robust PCA

Page 8: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

8/26

When is decomposition possible?

Identifiability issues: a matrix might be simultaneously low-rank andsparse!

1 0 0 · · · 00 0 0 · · · 0...

...... . . . ...

0 0 0 · · · 0

︸ ︷︷ ︸

sparse and low-rank

vs.

1 0 1 · · · 10 1 0 · · · 0...

...... . . . ...

1 0 0 · · · 1

︸ ︷︷ ︸sparse but not low-rank

Nonzero entries of sparse component need to be spread out— assume locations of nonzero entries are random / restrict the

number of nonzeros per row/column

1 1 1 · · · 11 1 1 · · · 1...

......

1 1 1 · · · 1

︸ ︷︷ ︸

low-rank and dense

vs.

1 1 1 · · · 10 0 0 · · · 0...

......

0 0 0 · · · 0

︸ ︷︷ ︸

low-rank but sparse

Low-rank component needs to be incoherent.

Page 9: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

8/26

When is decomposition possible?

Identifiability issues: a matrix might be simultaneously low-rank andsparse!

1 0 0 · · · 00 0 0 · · · 0...

...... . . . ...

0 0 0 · · · 0

︸ ︷︷ ︸

sparse and low-rank

vs.

1 0 1 · · · 10 1 0 · · · 0...

...... . . . ...

1 0 0 · · · 1

︸ ︷︷ ︸sparse but not low-rank

Nonzero entries of sparse component need to be spread out— assume locations of nonzero entries are random / restrict the

number of nonzeros per row/column

1 1 1 · · · 11 1 1 · · · 1...

......

1 1 1 · · · 1

︸ ︷︷ ︸

low-rank and dense

vs.

1 1 1 · · · 10 0 0 · · · 0...

......

0 0 0 · · · 0

︸ ︷︷ ︸

low-rank but sparse

Low-rank component needs to be incoherent.

Page 10: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

9/26

Low-rank component: coherence

Definition 8.1Coherence parameter µ1 of M = UΣV > is smallest quantity s.t.

maxi‖U>ei‖2 ≤

µ1r

nand max

i‖V >ei‖2 ≤

µ1r

n

Page 11: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

10/26

Low-rank component: joint coherence

Definition 8.2 (Joint coherence)Joint coherence parameter µ2 of M = UΣV > is smallest quantitys.t.

‖UV >‖∞ ≤√µ2r

n2

This prevents UV > from being too peaky.• µ1 ≤ µ2 ≤ µ2

1r, since

|(UV >)ij | = |e>i UV >ej | ≤ ‖e>

i U‖ · ‖V >ej‖ ≤µ1r

n

‖UV >‖2∞ ≥

‖UV >ej‖2F

n= ‖V

>ej‖2

n= µ1r

n2 (suppose ‖V >ej‖2 = µ1r

n)

Page 12: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

11/26

Convex relaxation

minimizeL,S rank(L) + λ‖S‖0, s.t. M = L + S (8.1)

minimizeL,S ‖L‖∗ + λ‖S‖1, s.t. M = L + S (8.2)

• ‖ · ‖∗ is nuclear norm; ‖ · ‖1 is entry-wise `1 norm• λ > 0: regularization parameter that balances two terms

Page 13: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

12/26

Theoretical guarantee

Theorem 8.3 (Candes, Li, Ma, Wright ’11)

• rank(L) . nmaxµ1,µ2 log2 n

;

• Nonzero entries of S are randomly located, and ‖S‖0 ≤ ρsn2 forsome constant ρs > 0 (e.g. ρs = 0.2).

Then (8.2) with λ = 1/√n is exact with high prob.

• rank(L) can be quite high (up to n/polylog(n))

• Parameter free: λ = 1/√n

• Ability to correct gross error: ‖S‖0 n2

• Sparse component S can have arbitrary magnitudes / signs!

Page 14: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

13/26

Geometry

Fig. credit: Candes ’14

Page 15: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

14/26

Empirical success rate

rank(L)/n s

1

rank

(L)/

n

s

1

n = 400

Fig. credit: Candes, Li, Ma, Wright ’11

Page 16: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

15/26

Dense error correction

Theorem 8.4 (Ganesh et al. ’10, Chen et al. ’13)

• rank(L) . nmaxµ1,µ2 log2 n

;• Nonzero entries of S are randomly located, have random sign,

and ‖S‖0 = ρsn2.

Then (8.2) with λ √

1−ρsρsn

succeeds with high prob., provided that

1− ρs︸ ︷︷ ︸non-corruption rate

&

√maxµ1, µ2rpolylog(n)

n

• When additive corruptions have random signs, (8.2) works evenwhen a dominant fraction of entries are corrupted

Page 17: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

16/26

Is joint coherence needed?

• Matrix completion: does not need µ2

• Robust PCA: so far we need µ2

Question: can we remove µ2? can we recover L with rank up ton

µ1polylog(n) (rather than nmaxµ1,µ2polylog(n)) with a constant fraction

of outliers?

Answer: no (example: planted clique)

Page 18: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

17/26

Planted clique problem

Setup: a graph G of n nodes generated as follows

1. connect each pair of nodes independently with prob. 0.5

2. pick n0 nodes and make them a clique (fully connected)

Goal: find hidden clique from G

Information theoretically, one can recover a clique if n0 > 2 log2 n

Page 19: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

18/26

Conjecture on computational barrier

Conjecture: ∀ constant ε > 0, if n0 ≤ n0.5−ε, then no tractablealgorithm can find the clique from G with prob. 1− o(1)

— often used as hardness assumption

Lemma 8.5

If there is an algorithm that allows recovery of any L from M withrank(L) ≤ n

µ1polylog(n) , then the above conjecture is violated

Page 20: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

19/26

Proof of Lemma 8.5

Suppose L is true adjacency matrix,

Li,j =

1, if i, j are both in the clique0, else

Let A be adjacency matrix of G, and generate M s.t.

Mi,j =Ai,j , with prob. 2/30, else

Therefore, one can write

M = L + M −L︸ ︷︷ ︸each entry is nonzero w.p. 1/3

Page 21: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

20/26

Proof of Lemma 8.5

Note thatµ1 = n

n0and µ2 = n2

n20

If there is an algorithm that can recover any L of rank nµ1polylog(n)

from M , then

rank(L) = 1 ≤ n

µ1polylog(n) ⇐⇒ n0 ≥ polylog(n)

But this contradicts the conjecture (which claims computationalinfeasibility to recover L unless n0 ≥ n0.5−o(1))

Page 22: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

21/26

Matrix completion with corruptions

What if we have missing data + corruptions?

• Observed entries

Mij = Lij + Sij , (i, j) ∈ Ω

for some observation set Ω, where S = (Sij) is sparse

• A natural extension of RPCA

minimizeL,S ‖L‖∗ + λ‖S‖1 s.t. PΩ(M) = PΩ(L + S)

• Theorems 8.3 - 8.4 easily extend to this setting

Page 23: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

22/26

Efficient algorithm: proximal method

In the presence of noise, one needs to solve

minimizeL,S ‖L‖∗ + λ‖S‖1 + µ

2 ‖M −L− S‖2F

which can be solved efficiently via proximal method

Algorithm 8.1 Iterative soft-thresholdingfor t = 0, 1, · · · :

Lt+1 = T1/µ(M − St

)St+1 = ψλ/µ

(M −Lt+1

)where T is singular-value thresholding operator, and ψ is softthresholding operator

Page 24: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

23/26

Nonconvex approachAlternatively, we can directly solve the nonconvex problem withoutrelaxation with the assumptions• rank(L) ≤ r; if we write the SVD of L = UΣV >, set

X? = UΣ1/2; Y ? = V Σ1/2.

• the non-zero entries of S are “spread out” (no more than αfraction of non-zeros per row/column), but otherwise arbitrary.

Sα =S ∈ Rn×n : ‖Si,:‖0 ≤ αn; ‖S:,j‖0 ≤ αn

minimizeX,Y ,S∈Sα ‖M −XY > − S‖2F︸ ︷︷ ︸quadratic loss

+ 14‖X

>X − Y >Y ‖2F︸ ︷︷ ︸fix scaling ambiguity

where X,Y ∈ Rn×r.

Page 25: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

24/26

Gradient descent and hard thresholding

minimizeX,Y ,S∈Sα F (X,Y ,S)where F (X,Y ,S) := ‖M −XY > − S‖2F + 1

4‖X>X − Y >Y ‖2F.

Algorithm 8.2 Gradient descent + Hard thresholding for RPCAInput: M , r, α, γ, η.Spectral initialization: Set S0 = Hγα(M). Let U0Σ0V 0> be therank-r SVD of M0 := PΩ(M − S); set X0 = U0 (Σ0)1/2 andY 0 = V 0 (Σ0)1/2.for t = 0, 1, 2, . . . , T − 1 do

1 Hard thresholding: St+1 = Hγα(M −XtY t>).2 Gradient updates:

Xt+1 = Xt − η∇XF(Xt,Y t,St+1

),

Y t+1 = Y t − η∇Y F(Xt,Y t,St+1

).

Page 26: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

25/26

Efficient nonconvex recovery

Theorem 8.6 (Yi et al. ’16)

Set γ = 2 and η = 1/(36σmax). Suppose that

α . min

1µ1√κr3

,1

µ1κ2r

.

The nonconvex approach (GD+HT) satisfies

∥∥∥XtY t> −L∥∥∥2

F.(

1− 1288κ

)tµ2

1κr3α2σmax

• O(κ log 1/ε) iterations to reach ε-accuracy.• For adversarial outliers, the optimal fraction of α = O(1/µ1r);

the bound is worse by a factor of√r.

• extendable to partial observation case.

Page 27: 0.15in ECE 18-898G: Special Topics in Signal Processing ...yuejiec/ece18898G_notes/ece18898g_ro… · 1/26 ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and

26/26

Reference

[1] ”Robust principal component analysis?,” E. Candes, X. Li, Y. Ma, andJ. Wright, Journal of ACM, 2011.

[2] ”Rank-sparsity incoherence for matrix decomposition,”V. Chandrasekaran, S. Sanghavi, P. Parrilo, and A. Willsky, SIAMJournal on Optimization, 2011.

[3] ”Incoherence-optimal matrix completion,” Y. Chen, IEEE Transactionson Information Theory, 2015.

[4] “Dense error correction for low-rank matrices via principal componentpursuit,” A. Ganesh, J. Wright, X. Li, E. Candes, Y. Ma, ISIT, 2010.

[5] ”Low-rank matrix recovery from errors and erasures,” Y. Chen, A. Jalali,S. Sanghavi, C. Caramanis, IEEE Transactions on Information Theory,2013.

[6] ”Fast Algorithms for Robust PCA via Gradient Descent,” X. Yi, D. Park,Y. Chen, and C. Caramanis, NIPS, 2016.