Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See...

49
Stability of low-rank matrix recovery and its connections to Banach space geometry J. Alejandro Chávez-Domínguez Department of Mathematics The University of Texas at Austin Joint work with Denka Kutzarova (UIUC) First Brazilian Workshop in Geometry of Banach spaces Maresias, Brazil August 29 th , 2014 http://arxiv.org/abs/1406.6712 J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 1 / 27

Transcript of Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See...

Page 1: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Stability of low-rank matrix recovery and itsconnections to Banach space geometry

J. Alejandro Chávez-Domínguez

Department of MathematicsThe University of Texas at Austin

Joint work with Denka Kutzarova (UIUC)

First Brazilian Workshop in Geometry of Banach spacesMaresias, Brazil August 29th, 2014

http://arxiv.org/abs/1406.6712

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 1 / 27

Page 2: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Practical problems

Inferring quantities of interest from measured information;Compressing information.

x ∈ Rn unknown, measurement y = Ax

whereA : Rn → Rm is a linear map, n > m

We have y, we want to recover x.

Impossible! Too many degrees of freedom.

The magic of compressed sensingIt becomes possible if x is sparse, i.e. has few nonzero coordinates.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 2 / 27

Page 3: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Practical problems

Inferring quantities of interest from measured information;Compressing information.

x ∈ Rn unknown, measurement y = Ax

whereA : Rn → Rm is a linear map, n > m

We have y, we want to recover x.

Impossible! Too many degrees of freedom.

The magic of compressed sensingIt becomes possible if x is sparse, i.e. has few nonzero coordinates.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 2 / 27

Page 4: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Practical problems

Inferring quantities of interest from measured information;Compressing information.

x ∈ Rn unknown, measurement y = Ax

whereA : Rn → Rm is a linear map, n > m

We have y, we want to recover x.

Impossible! Too many degrees of freedom.

The magic of compressed sensingIt becomes possible if x is sparse, i.e. has few nonzero coordinates.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 2 / 27

Page 5: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Sparse recovery problem

Given the measurement y, what’s a reasonable guess for x?

IDEA: Find the sparsest vector that is consistent with themeasurement.

minimize ‖z‖0 subject to Az = y,

where‖z‖0 = number of nonzero coordinates of z.

This is NP-hard! [Natarajan 1995].

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 3 / 27

Page 6: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Sparse recovery problem

Given the measurement y, what’s a reasonable guess for x?

IDEA: Find the sparsest vector that is consistent with themeasurement.

minimize ‖z‖0 subject to Az = y,

where‖z‖0 = number of nonzero coordinates of z.

This is NP-hard! [Natarajan 1995].

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 3 / 27

Page 7: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Sparse recovery problem

Given the measurement y, what’s a reasonable guess for x?

IDEA: Find the sparsest vector that is consistent with themeasurement.

minimize ‖z‖0 subject to Az = y,

where‖z‖0 = number of nonzero coordinates of z.

This is NP-hard! [Natarajan 1995].

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 3 / 27

Page 8: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Norm minimization

Instead ofminimize ‖z‖0 subject to Az = y,

we would like to consider

minimize ‖z‖`1subject to Az = y.

This is a convex problem that can be solved efficiently.

MoralIf the original vector x is sparse enough, both problems have x assolution (as long as we choose the linear map A wisely).

(Rough) Example of a ResultFor a fixed k-sparse vector x and a random Gaussian matrix A,`1-minimization exactly recovers x with high probability ifm > 2k ln(n/k).

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 4 / 27

Page 9: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Norm minimization

Instead ofminimize ‖z‖0 subject to Az = y,

we would like to consider

minimize ‖z‖`1subject to Az = y.

This is a convex problem that can be solved efficiently.

MoralIf the original vector x is sparse enough, both problems have x assolution (as long as we choose the linear map A wisely).

(Rough) Example of a ResultFor a fixed k-sparse vector x and a random Gaussian matrix A,`1-minimization exactly recovers x with high probability ifm > 2k ln(n/k).

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 4 / 27

Page 10: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Stability

True sparsity: only inidealized situations.

More realistic: the unknownvector is close to sparsevectors.

Examples: Image or soundcompression.

2 1 An Invitation to Compressive Sensing

Fig. 1.1 Antonella, Niels, and Paulina. Top: Original Image. Bottom: Reconstruction using 1% ofthe largest absolute wavelet coefficients, i.e., 99 % of the coefficients are set to zero

Thus, it came as a surprise that under certain assumptions it is actually possible toreconstruct signals when the number m of available measurements is smaller thanthe signal length N . Even more surprisingly, efficient algorithms do exist for thereconstruction. The underlying assumption which makes all this possible is sparsity.The research area associated to this phenomenon has become known as compressivesensing, compressed sensing, compressive sampling, or sparse recovery. This wholebook is devoted to the mathematics underlying this field.

Sparsity. A signal is called sparse if most of its components are zero. As empiri-cally observed, many real-world signals are compressible in the sense that they arewell approximated by sparse signals—often after an appropriate change of basis.This explains why compression techniques such as JPEG, MPEG, or MP3 work sowell in practice. For instance, JPEG relies on the sparsity of images in the discretecosine basis or wavelet basis and achieves compression by only storing the largest

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 5 / 27

Page 11: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Stability

True sparsity: only inidealized situations.

More realistic: the unknownvector is close to sparsevectors.

Examples: Image or soundcompression.

2 1 An Invitation to Compressive Sensing

Fig. 1.1 Antonella, Niels, and Paulina. Top: Original Image. Bottom: Reconstruction using 1% ofthe largest absolute wavelet coefficients, i.e., 99 % of the coefficients are set to zero

Thus, it came as a surprise that under certain assumptions it is actually possible toreconstruct signals when the number m of available measurements is smaller thanthe signal length N . Even more surprisingly, efficient algorithms do exist for thereconstruction. The underlying assumption which makes all this possible is sparsity.The research area associated to this phenomenon has become known as compressivesensing, compressed sensing, compressive sampling, or sparse recovery. This wholebook is devoted to the mathematics underlying this field.

Sparsity. A signal is called sparse if most of its components are zero. As empiri-cally observed, many real-world signals are compressible in the sense that they arewell approximated by sparse signals—often after an appropriate change of basis.This explains why compression techniques such as JPEG, MPEG, or MP3 work sowell in practice. For instance, JPEG relies on the sparsity of images in the discretecosine basis or wavelet basis and achieves compression by only storing the largest

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 5 / 27

Page 12: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Unpaid advertisement

Applied and Numerical Harmonic Analysis

Simon FoucartHolger Rauhut

A Mathematical Introduction to Compressive Sensing

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 6 / 27

Page 13: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Stability

We would like to recover a vector x with an error controlled by itsdistance to k-sparse vectors.

ρk(x)`p := inf{∥∥x− x′

∥∥`p

:∥∥x′∥∥

0 ≤ k}.

Denote by ∆p(y) a solution to

minimize ‖z‖`psubject to Az = y,

(Rough) Example of a ResultThere are linear maps A : Rn → Rm such that for every vector x

‖x−∆1(Ax)‖`1≤ 6ρk(x)`1 .

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 7 / 27

Page 14: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Stability and geometry

Theorem (Kashin-Temlyakov 2007)Linear map A : Rn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all x ∈ Rn,

‖x−∆1(Ax)‖`2≤ Ck−1/2ρk(x)`1

2 For all x ∈ ker(A),‖x‖`2

≤ Ck−1/2 ‖x‖`1

RemarkThe latter is a statement about the norm of the identity map `n

1 → `n2

restricted to the subspace ker(A).

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 8 / 27

Page 15: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Stability and geometry

Theorem (Kashin-Temlyakov 2007)Linear map A : Rn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all x ∈ Rn,

‖x−∆1(Ax)‖`2≤ Ck−1/2ρk(x)`1

2 For all x ∈ ker(A),‖x‖`2

≤ Ck−1/2 ‖x‖`1

RemarkThe latter is a statement about the norm of the identity map `n

1 → `n2

restricted to the subspace ker(A).

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 8 / 27

Page 16: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Gelfand numbers

Gelfand m-number of a linear map T : X → Y

cm(T) := inf{∥∥T

∣∣L

∥∥ : L subspace of X with codim(L) < m}

Punchline: How small can the norm of the operator be if werestrict ourselves to subspaces of a given codimension?

The sequence(cm(T)

)∞m=1 is a measure of the compactness of T.

Kashin-Temlyakov:Results of [Kashin 1977; Garnaev-Gluskin 1984] on Gelfandnumbers immediately imply the existence of good measurementmaps for sparse recovery through `1-minimization.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 9 / 27

Page 17: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Gelfand: numbers vs. widths

In approximation theory/compressed sensing they normally useGelfand widths instead of Gelfand numbers.

See Pietsch’s book History of Banach spaces and linear operatorsfor an argument of why the numbers “won” in Banach spacetheory.

In the cases we are considering, both concepts coincide.

For geometric conditions guaranteeing the coincidence in moregeneral situations, see [Edmunds-Lang 2013].

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 10 / 27

Page 18: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Extra structure: matrices

Sometimes the space of unknown vectors has an extra matricialstructure.

Matrix completion (Netflix problem)TASK: fill in missing entries of a matrix.

Online store sells products indexed by the rows, consumersindexed by the columns rate some of these products.

For purposes of individualized advertisement, the store isinterested in predicting the whole matrix of consumer ratings.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 11 / 27

Page 19: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Low-rank recovery

To stand a chance of success, the unknown matrix should havelow rank (corresponding to sparsity).

Low-rank recovery problemSolve

minimize rank(Z) subject to AZ = y,

where A : Mn → Rm is a linear map, n2 > m.

This is also NP-hard.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 12 / 27

Page 20: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Norm minimization

Can some sort of norm minimization help?

Natural candidate: Schatten p-norm.

‖X‖Sp:= `p-norm of the vector of singular values of X

Snp = (Mn, ‖·‖Sp

) is considered a noncommutative version of `np.

Instead ofminimize rank(Z) subject to AZ = y,

we would like to consider

minimize ‖Z‖Spsubject to AZ = y.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 13 / 27

Page 21: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Plenty of recent work

Fazel 2002.Candès-Recht 2009.Candès-Tao 2010.Recht-Fazel-Parrilo 2010.Dvijotham-Fazel 2010Candès-Plan 2011.Fornasier-Rauhut-Ward 2011.Recht-Hu-Hassibi 2011.Oymak-Mohan-Fazel-Hassibi 2011.Kong-Xiu 2013.Cai-Zhang 2014.And more...

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 14 / 27

Page 22: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Kashin-Temlyakov TheoremLinear map A : Rn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all x ∈ Rn,

‖x−∆1(Ax)‖`2≤ Ck−1/2ρk(x)`1

2 For all x ∈ ker(A),‖x‖`2

≤ Ck−1/2 ‖x‖`1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 23: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Noncommutative Kashin-Temlyakov TheoremLinear map A : Rn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all x ∈ Rn,

‖x−∆1(Ax)‖`2≤ Ck−1/2ρk(x)`1

2 For all x ∈ ker(A),‖x‖`2

≤ Ck−1/2 ‖x‖`1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 24: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Noncommutative Kashin-Temlyakov TheoremLinear map A : Rn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all x ∈ Rn,

‖x−∆1(Ax)‖`2≤ Ck−1/2ρk(x)`1

2 For all x ∈ ker(A),‖x‖`2

≤ Ck−1/2 ‖x‖`1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 25: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Noncommutative Kashin-Temlyakov TheoremLinear map A : Mn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all x ∈ Rn,

‖x−∆1(Ax)‖`2≤ Ck−1/2ρk(x)`1

2 For all x ∈ ker(A),‖x‖`2

≤ Ck−1/2 ‖x‖`1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 26: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Noncommutative Kashin-Temlyakov TheoremLinear map A : Mn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all x ∈ Rn,

‖x−∆1(Ax)‖`2≤ Ck−1/2ρk(x)`1

2 For all x ∈ ker(A),‖x‖`2

≤ Ck−1/2 ‖x‖`1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 27: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Noncommutative Kashin-Temlyakov TheoremLinear map A : Mn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all X ∈ Mn,

‖X −∆1(AX)‖`2≤ Ck−1/2ρk(X)`1

2 For all X ∈ ker(A),‖X‖`2

≤ Ck−1/2 ‖X‖`1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 28: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Noncommutative Kashin-Temlyakov TheoremLinear map A : Mn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all X ∈ Mn,

‖X −∆1(AX)‖`2≤ Ck−1/2ρk(X)`1

2 For all X ∈ ker(A),‖X‖`2

≤ Ck−1/2 ‖X‖`1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 29: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Noncommutative Kashin-Temlyakov TheoremLinear map A : Mn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all X ∈ Mn,

‖X −∆1(AX)‖S2≤ Ck−1/2ρk(X)S1

2 For all X ∈ ker(A),‖X‖S2

≤ Ck−1/2 ‖X‖S1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 30: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Noncommutative Kashin-Temlyakov TheoremLinear map A : Mn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all X ∈ Mn,

‖X −∆1(AX)‖S2≤ Ck−1/2ρk(X)S1

2 For all X ∈ ker(A),‖X‖S2

≤ Ck−1/2 ‖X‖S1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 31: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Noncommutative Kashin-Temlyakov TheoremLinear map A : Mn → Rm, k � n/log(m/n). TFAE (up to a change in theconstant):

1 For all X ∈ Mn,

‖X −∆1(AX)‖S2≤ Ck−1/2ρk(X)S1

2 For all X ∈ ker(A),‖X‖S2

≤ Ck−1/2 ‖X‖S1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 32: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal: Noncommutative version of theKashin-Temlyakov theorem

Noncommutative Kashin-Temlyakov TheoremLinear map A : Mn → Rm, k �?????. TFAE (up to a change in theconstant):

1 For all X ∈ Mn,

‖X −∆1(AX)‖S2≤ Ck−1/2ρk(X)S1

2 For all X ∈ ker(A),‖X‖S2

≤ Ck−1/2 ‖X‖S1

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 15 / 27

Page 33: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Known results

Theorem (Oymak-Mohan-Fazel-Hassibi 2011)Let C be a constant. Then the following are equivalent:

(i) For any X and X′ with ‖X′‖S1≤ ‖X‖S1

and AX = AX′,∥∥X − X′∥∥

S2≤ Ck−1/2ρk(X)S1

(ii) For any Y ∈ ker(A) we have

∥∥Y − Y[k]∥∥

S1−∥∥Y[k]

∥∥S1≥ 2√

kC‖Y‖S1

Where Y[k] is the k-spectral truncation of Y (i.e. keeping just the klargest singular values).

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 16 / 27

Page 34: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Known results II

Theorem (Oymak-Mohan-Fazel-Hassibi 2011)Suppose that for all X ∈ ker(A) we have

‖X‖S2≤ 1√

D‖X‖S1

Then for any X and X′ with ‖X′‖S1≤ ‖X‖S1

and AX = AX′, it holds that

∥∥X − X′∥∥

S1≤ 2

1− 2√

k/Dρk(X)S1 .

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 17 / 27

Page 35: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Noncommutative Kashin-Temlyakov

Theorem (CD-K)Linear map A : Mn → Rm, k � n/m. TFAE (up to a change in theconstant):

1 For all X ∈ Mn,

‖X −∆1(AX)‖S2≤ Ck−1/2ρk(X)S1

2 For all X ∈ ker(A),

‖X‖S2≤ Ck−1/2 ‖X‖S1

.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 18 / 27

Page 36: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Turning the tables

Use compressed sensing ideas to prove results about Gelfandnumbers.

Theorem (Foucart-Pajor-Rauhut-Ullrich 2010)For 0 < p ≤ 1 and p < q ≤ 2, if m < n, then

cm(Id : `np → `n

q) � min{

1,ln(n/m) + 1

m

}1/p−1/q

.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 19 / 27

Page 37: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal, Part II

Noncommutative version of the Foucart-Pajor-Rauhut-Ullrichtheorem: calculate the Gelfand numbers of identity maps

Id : Snp → Sn

q

for 0 < p ≤ 1 and p < q ≤ 2.

Known result: [Carl-Defant 1997]

cm(Id : Sn1 → Sn

2) � min{

1,nm

}1/2.

Their approach is based on tensor product arguments plus a deepinequality of Pajor and Tomczak-Jaegermann.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 20 / 27

Page 38: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Goal, Part II

Noncommutative version of the Foucart-Pajor-Rauhut-Ullrichtheorem: calculate the Gelfand numbers of identity maps

Id : Snp → Sn

q

for 0 < p ≤ 1 and p < q ≤ 2.

Known result: [Carl-Defant 1997]

cm(Id : Sn1 → Sn

2) � min{

1,nm

}1/2.

Their approach is based on tensor product arguments plus a deepinequality of Pajor and Tomczak-Jaegermann.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 20 / 27

Page 39: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Main result

We are able to calculate Gelfand numbers in a noncommutativeversion of the Foucart-Pajor-Rauhut-Ullrich result.

Theorem (CD-K)For 0 < p ≤ 1 and p < q ≤ 2, if 1 ≤ m < n2, then

cm(Id : Snp → Sn

q) � min{

1,nm

}1/p−1/q.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 21 / 27

Page 40: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Technical tool: Restricted Isometry Property

A condition on the measuring map that guarantees stable recovery.

RIP [Candès-Tao 2005]A : Rn → Rm has the RIP of order k with constant δk > 0 if for all vectorsz ∈ Rn of sparsity at most k,

(1− δk)1/2 ‖z‖`2

≤ ‖Az‖`2≤ (1 + δk)

1/2 ‖z‖`2

Note: There are strong connections between the RIP and theJohnson-Lindenstrauss lemma [Baraniuk-Davenport-DeVore-Wakin2008, Krahmer-Ward 2011].

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 22 / 27

Page 41: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Technical tool: Restricted Isometry Property

A condition on the measuring map that guarantees stable recovery.

RIP [Candès-Tao 2005]A : Rn → Rm has the RIP of order k with constant δk > 0 if for all vectorsz ∈ Rn of sparsity at most k,

(1− δk)1/2 ‖z‖`2

≤ ‖Az‖`2≤ (1 + δk)

1/2 ‖z‖`2

Note: There are strong connections between the RIP and theJohnson-Lindenstrauss lemma [Baraniuk-Davenport-DeVore-Wakin2008, Krahmer-Ward 2011].

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 22 / 27

Page 42: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Technical tool: Restricted Isometry Property

A condition on the measuring map that guarantees stable recovery.

RIP [Candès-Tao 2005]A : Rn → Rm has the RIP of order k with constant δk > 0 if for all vectorsz ∈ Rn of sparsity at most k,

(1− δk)1/2 ‖z‖`2

≤ ‖Az‖`2≤ (1 + δk)

1/2 ‖z‖`2

Note: There are strong connections between the RIP and theJohnson-Lindenstrauss lemma [Baraniuk-Davenport-DeVore-Wakin2008, Krahmer-Ward 2011].

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 22 / 27

Page 43: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Noncommutative Restricted Isometry Property

Matrix RIP [Recht-Fazel-Parrilo 2010]A : Mn → Rm has the RIP of order k with constant δk > 0 if for allmatrices Z ∈ Mn of rank at most k,

(1− δk)1/2 ‖Z‖S2

≤ ‖AZ‖`2≤ (1 + δk)

1/2 ‖Z‖S2

We use a modified version inspired by [Foucart-Lai 2009].

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 23 / 27

Page 44: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Noncommutative Restricted Isometry Property

Matrix RIP [Recht-Fazel-Parrilo 2010]A : Mn → Rm has the RIP of order k with constant δk > 0 if for allmatrices Z ∈ Mn of rank at most k,

(1− δk)1/2 ‖Z‖S2

≤ ‖AZ‖`2≤ (1 + δk)

1/2 ‖Z‖S2

We use a modified version inspired by [Foucart-Lai 2009].

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 23 / 27

Page 45: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Stability and Robustness: technical result

Theorem (CD-K)Given 0 < p ≤ 1, if for integers t ≥ k we have

2δ2t

1− δ2t< 4(√

2− 1)( t

k

)1/p−1/2,

then a solution X∗ of

minimize ‖Z‖Spsubject to ‖AZ − y‖2 ≤ β2k · θ.

approximates the original matrix X with errors

‖X − X∗‖Sp≤ C1ρk(X)Sp + D1 · k1/p−1/2 · θ,

‖X − X∗‖S2≤ C2

ρk(X)Sp

t1/p−1/2 + D2 · θ.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 24 / 27

Page 46: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Idea of the proof

It follows the general strategy of [Foucart-Lai 2009].

BIG ISSUE: if P : Rn → Rn is a projection over a subset ofcoordinates, then for any vector x ∈ Rn,

‖x‖p`p

= ‖Px‖p`p

+ ‖x− Px‖p`p.

This is not the case for matrices.

We overcome this difficulty by modifying some matrixdecompositions from [Kong-Xiu 2013].

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 25 / 27

Page 47: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Comparison to related results

A recent result similar to ourshas appeared in[Liu-Huang-Chen 2014].Their conclusions have thesame form, but ourhypothesis are weaker.

[MalekMohammadi-BabaieZadeh-Skoglund]have independently obtainedthe same stability result asus.

Neither one of those paperscalculate Gelfand widths.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 26 / 27

Page 48: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

Comparison to related results

A recent result similar to ourshas appeared in[Liu-Huang-Chen 2014].Their conclusions have thesame form, but ourhypothesis are weaker.

[MalekMohammadi-BabaieZadeh-Skoglund]have independently obtainedthe same stability result asus.

Neither one of those paperscalculate Gelfand widths.

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 26 / 27

Page 49: Stability of low-rank matrix recovery and its connections ...banach/bwb2014/slides/... · See Pietsch’s book History of Banach spaces and linear operators for an argument of why

THANKS!

J.A. Chávez-Domínguez (UT Austin) Low-rank matrix recovery 08/29/2014 27 / 27