An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD...

10
An Optimization Framework with Flexible Inexact Inner Iterations for Nonconvex and Nonsmooth Programming Yiyang Wang School of Mathematical Sciences Dalian University of Technology [email protected] Risheng Liu School of Software Technology Dalian University of Technology [email protected] Xiaoliang Song School of Mathematical Sciences Dalian University of Technology [email protected] Zhixun Su School of Mathematical Sciences Dalian University of Technology [email protected] Abstract In recent years, numerous vision and learning tasks have been (re)formulated as nonconvex and nonsmooth program- mings (NNPs). Although some algorithms have been pro- posed for particular problems, designing fast and flexible optimization schemes with theoretical guarantee is a chal- lenging task for general NNPs. It has been investigated that performing inexact inner iterations often benefit to special applications case by case, but their convergence behaviors are still unclear. Motivated by these practical experiences, this paper designs a novel algorithmic framework, named inexact proximal alternating direction method (IPAD) for solving general NNPs. We demonstrate that any numer- ical algorithms can be incorporated into IPAD for solv- ing subproblems and the convergence of the resulting hy- brid schemes can be consistently guaranteed by a series of simple error conditions. Beyond the guarantee in theory, numerical experiments on both synthesized and real-world data further demonstrate the superiority and flexibility of our IPAD framework for practical use. 1. Introduction Nonconvex and nonsmooth programmings (NNPs) have received wide attentions in recent years [27, 22, 33, 37, 5]. Many problems in vision and learning societies, such as sparse coding and dictionary learning [5], matrix factor- ization [22], image restoration [37], and image classifica- tion [40] can be (re)formulated as specific NNPs. In this work, we consider a general NNP in following formulation: min X:=(x1,...,x K ) Ψ(X) := K X i=1 f i (x i )+ H(X), (1) where x i s are vectors or matrices throughout the paper. This general NNP covers a variety of techniques in im- age processing and machine learning. For example, princi- pal component analysis with sparse constraints, like lasso constraint [26] and elastic-net regularization [40] can be written in problem (1). In non-negative matrix factoriza- tion problem [22], it is common to adopt H as a 2 distance on describing the ability of restoration and restrict f i s to be non-negative on each component. Sparse dictionary learn- ing (SDL) can also be formulated in problem (1). Many lit- eratures [18, 27, 5] have posted the superiority on restricting dictionaries with normalized bases while at the same time constraining sparsity for codes with various nonconvex and nonsmooth sparse penalties. In the past few years, there have been literatures [2, 8, 33, 34, 7] in optimization and numerical analysis on de- signing converged algorithms in view of the general NNP (1). These algorithms have been applied to various prob- lems, such as non-negative matrix factorization [8], SDL with 0 penalty [5] and non-negative Tucker decomposition [34]; at the same time, abundant experimental analyses have demonstrated the efficiency and convergence properties of these algorithms. However, in pursuit of convergence, most of the existing algorithms are designed with fixed iteration schemes; those inflexible schemes are tightly constrained and fail to take the model structures of specific problems into consideration. While on the other hand, specific solvers designed for practical use are far more flexible than the converged algo- 1 arXiv:1702.08627v3 [cs.CV] 30 Jun 2017

Transcript of An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD...

Page 1: An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure

An Optimization Framework with Flexible Inexact Inner Iterationsfor Nonconvex and Nonsmooth Programming

Yiyang WangSchool of Mathematical SciencesDalian University of Technology

[email protected]

Risheng LiuSchool of Software TechnologyDalian University of Technology

[email protected]

Xiaoliang SongSchool of Mathematical SciencesDalian University of Technology

[email protected]

Zhixun SuSchool of Mathematical SciencesDalian University of Technology

[email protected]

Abstract

In recent years, numerous vision and learning tasks havebeen (re)formulated as nonconvex and nonsmooth program-mings (NNPs). Although some algorithms have been pro-posed for particular problems, designing fast and flexibleoptimization schemes with theoretical guarantee is a chal-lenging task for general NNPs. It has been investigated thatperforming inexact inner iterations often benefit to specialapplications case by case, but their convergence behaviorsare still unclear. Motivated by these practical experiences,this paper designs a novel algorithmic framework, namedinexact proximal alternating direction method (IPAD) forsolving general NNPs. We demonstrate that any numer-ical algorithms can be incorporated into IPAD for solv-ing subproblems and the convergence of the resulting hy-brid schemes can be consistently guaranteed by a series ofsimple error conditions. Beyond the guarantee in theory,numerical experiments on both synthesized and real-worlddata further demonstrate the superiority and flexibility ofour IPAD framework for practical use.

1. Introduction

Nonconvex and nonsmooth programmings (NNPs) havereceived wide attentions in recent years [27, 22, 33, 37, 5].Many problems in vision and learning societies, such assparse coding and dictionary learning [5], matrix factor-ization [22], image restoration [37], and image classifica-tion [40] can be (re)formulated as specific NNPs. In this

work, we consider a general NNP in following formulation:

minX:=(x1,...,xK)

Ψ(X) :=

K∑i=1

fi(xi) +H(X), (1)

where xis are vectors or matrices throughout the paper.This general NNP covers a variety of techniques in im-

age processing and machine learning. For example, princi-pal component analysis with sparse constraints, like lassoconstraint [26] and elastic-net regularization [40] can bewritten in problem (1). In non-negative matrix factoriza-tion problem [22], it is common to adopt H as a `2 distanceon describing the ability of restoration and restrict fis to benon-negative on each component. Sparse dictionary learn-ing (SDL) can also be formulated in problem (1). Many lit-eratures [18, 27, 5] have posted the superiority on restrictingdictionaries with normalized bases while at the same timeconstraining sparsity for codes with various nonconvex andnonsmooth sparse penalties.

In the past few years, there have been literatures [2, 8,33, 34, 7] in optimization and numerical analysis on de-signing converged algorithms in view of the general NNP(1). These algorithms have been applied to various prob-lems, such as non-negative matrix factorization [8], SDLwith `0 penalty [5] and non-negative Tucker decomposition[34]; at the same time, abundant experimental analyses havedemonstrated the efficiency and convergence properties ofthese algorithms. However, in pursuit of convergence, mostof the existing algorithms are designed with fixed iterationschemes; those inflexible schemes are tightly constrainedand fail to take the model structures of specific problemsinto consideration.

While on the other hand, specific solvers designed forpractical use are far more flexible than the converged algo-

1

arX

iv:1

702.

0862

7v3

[cs

.CV

] 3

0 Ju

n 20

17

Page 2: An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure

rithms proposed for the general problem (1). Those specificsolvers always take advantages of the problem structuresand then employ effective numerical methods to solve spe-cific sub-problems. Moreover, we have noticed from pre-vious work [10, 30, 19, 11, 25] that solving sub-problemswith inner iterations is a frequently used strategy in numer-ous efficient solvers. Though few of the inexact solvers aredesigned with rigid theoretical support and analyses, theirefficiencies and convergence properties have been verifiedfrom experiments under certain conditions.

1.1. Contributions

Motivated from various inexact solvers, we in this paperpropose an unified and flexible algorithm framework namedinexact proximal alternating direction method (IPAD). OurIPAD is designed for solving the general NNP problem (1),at the same time, keeping the flexibility when dealing withspecific problems. Different from existing solvers in prac-tice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure the convergence, thusis more rigid and robust for practical use. The theoreticalsupport provided in this paper can be regarded as a guid-ance for designing inexact algorithms for solving specificproblems in a concise framework. As far as we know, weare the first to incorporate various numerical methods intoa general algorithm framework and at the same time giverigorous convergence analyses for NNPs. In summary, ourcontributions are three folds:

1. Different from most existing numerical algorithms forNNPs, which always fix their updating schemes duringiterations, we provide a novel perspective to incorpo-rate different optimization strategies into a unified andflexible proximal optimization framework for (1).

2. Even with inexact subproblems and flexible inner iter-ation schemes, we prove that the convergence prop-erty of the resulting hybrid optimization frameworkcan still be guaranteed. Indeed, our theoretical resultsare the best we can ask for, unless further assumptionsare made on the general NNPs in (1).

3. As an application example, we show implementa-tions of applying IPAD with different inner algorithmsto solve the widely concerned `0 regularized SDLmodel. Numerical evaluations and abundant exper-imental comparisons demonstrate promising experi-mental results of the proposed algorithms, which ver-ify the flexibility and efficiency of our optimizationframework.

2. Related Work2.1. Existing Optimization Strategies for NNPs

In past several years, accompanied with the rising pop-ularity of investigating sparsity and low-rankness (naturallywith nonconvex and nonsmooth properties) for vision andlearning tasks [22, 37, 32, 1, 23, 39], developing numeri-cal solvers for different types of NNP models have attractedconsiderable research interests from computer vision, ma-chine learning and optimization societies.

As a typical nonconvex and nonsmooth measurement, `0norm has been widely applied for image processing prob-lems [37, 32, 1]; and various methods have been designedfor solving its optimization. In [37], the authors proposea proximal method for finding desirable solution to `0TVproblem, but they only prove a weak convergence result un-der mild conditions. The rank function [23, 39] has alsobeen well studied: the work in [23] optimizes a low-rankminimization by an iteratively reweighted algorithm. How-ever, they can only prove that any limit point is a stationarypoint, which cannot guarantee the correct convergence ofthe iterations. Another typical problem, non-negative ma-trix factorization is a powerful technique for various appli-cations; algorithms like ANLS [22] have been specially de-signed for solving this specific problem.

On the other hand, some algorithms with rigid theoret-ical analyses have been proposed for solving the generalNNP problem (1) [2, 8, 31]. The authors in [2] propose analgorithm named GIST for solving a special class of NNP(1) and apply it to logistic regression problem. A proxi-mal alternating linearized method is proposed in [8], how-ever, it is time-consuming for solving special problems. Xuet al. in [31] propose a MM method for the general NNP,but their algorithm is too complicated to be implemented inpractice. The authors in [2] provide their convergence resultby assuming subproblems are exactly solved, however, thiscondition is unattainable in most situations.

In summary, most optimization schemes for specificNNPs are efficient in practice, however, their convergenceguarantees are relatively weak. While, algorithms designedfor the general NNP are well-converged, but they alwayshave strict conditions and are inefficient in practice.

2.2. Inexact Inner Iterations in Application Fields

We have observed that in some application fields, rela-tively good performances are often obtained by “improper”numerical algorithms. In previous literatures [41, 36, 25],the “improper” iteration skills have been frequently used insolving sparse coding, blind kernel estimation, and medicalimaging problems in practice.

One of those inexact schemes is employing numericalmethods as inner iteration methods for solving special sub-problems. For instance, the authors in [41] employ half-

Page 3: An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure

quadratic splitting and Lagrangian dual method as the in-ner iteration schemes for estimating clear images in blinddeconvolution problem. While in a text image deblurringpaper [25], the authors update variables though alternat-ing direction method, meanwhile, applying half-quadraticsplitting for solving subproblems. Although these inexactsolvers are effective in practice, they are totally designed infreestyle and are uncontrolled in the performances.

Another kind of inexact schemes is adopting neural net-works for approximating exact solutions during iterations.For example, time-unfolded recurrent neural networks canbe used to produce the best possible approximations forsparse coding problem [18, 28]. On the other hand, fusingvarious neural networks in ADMM framework [13, 36] hasgreat performances for approximately optimizing problemsfor image restoration and compressed sensing. The efficien-cies of these inexact solvers and their convergence perfor-mances have been verified from experiments, however, fewof them are designed with rigid theoretical support.

3. The Proposed Optimization Framework

To simplify subsequent derivations, we propose an IPADand analyze its properties for problem (1) with 2 variables1:

minz:=(x,y)

Ψ(z) := f(x) + g(y) +H(x,y). (2)

Though the whole properties are analyzed for problem (2),it is straightforward to extend them to the general one. Thenwe give assumptions on the objective function:(1) f and g are proper, lower semi-continuous functions;(2) H is a C1 function and its gradient is Lipschitz contin-uous on a bounded set;(3) Ψ is a coercive, Kurdyka-Łojasiewicz (KŁ) function2.

Remark 1 It should be mentioned that the most frequentlyused `2-norm H(x,y) is a KŁ function which also satisfiesthe second assumption. On the other hand, regularizers like`0 penalty, `p norm, SCAD [16], MCP [38] and indicatorfunctions are all KŁ functions and at the same time satisfythe first assumption. Since the finite sums of KŁ functionsare also KŁ functions [8], thus not a few models in imageprocessing and machine learning satisfy these assumptions.

3.1. Inexact Proximal Reformulation

For solving problem (2), it is natural and common to ap-ply a proximal alternating direction method (PAD) [35, 11]

1For simplicity, we replace the notations in problem (1) with the newones. We hope this will not cause confusions.

2To be self-contained, the definition of KŁ function is given in the sup-plemental material due to space limit.

Algorithm 1 Inexact Proximal Alternating Direction1: Setting parameters: Cx, Cy , {ηt1}t∈N, {ηt2}t∈N.2: Initializing variables: x0, y0.3: while not converged do4: xt = ψ(x;xt−1,yt−1, f,H, ηt1, Cx).

(i.e., Perform Alg. 2 for x subproblem)5: yt = ψ(y;yt−1,xt, g,H, ηt2, Cy).

(i.e., Perform Alg. 2 for y subproblem)6: end while

Algorithm 2 ut = ψ(u;ut−1,v, h,H, η, Cu).1: With parameters Cx, Cy , {ηt1}t∈N, {ηt2}t∈N in Alg. 1.2: Denote A as inner iteration schemes for optimizingh(u) +H(u,v) + η

2‖u− ut−1‖2.3: Let ut,0 = ut−1.4: while ‖et,iu ‖ > Cu‖ut,i − ut−1‖ do5: ut,i = A(ut,i−1).6: In theory, use Eq. (5) to analyze convergence.7: In practice, use Eq. (13) for judging criterion.8: end while9: ut = ut,i.

for alternatively updating variables in a cyclic order:

xt∗ = arg minxf(x) +H(x,yt−1) +

ηt−11

2‖x− xt−1‖2,

yt∗ = arg minyg(y) +H(xt,y) +

ηt−12

2‖y − yt−1‖2,

(3)where ηt−1

1 and ηt−12 are proximal parameters added on the

subproblems respectively; the notations xt∗ and yt∗ are usedto represent exact solutions of the subproblems in Eq. (3).It should be pointed out that adding proximal terms on sub-problems is a common skill to improve the stability of algo-rithms on solving nonconvex problems [2].

As claimed before, in most cases, it is either impos-sible or extremely hard for calculating exact solutions ofsubproblems. Thus not a few work employ inner iterationschemes to compute approximations to the exact solutions,which means, the inexact solutions xt and yt calculated byinner methods are approximations to xt∗ and yt∗:

xt ≈ xt∗, yt ≈ yt∗. (4)

We can see from the above that our IPAD is in a generalframework: it alternatively updates variables by approxi-mately solving the subproblems of PAD. While on the otherhand, it does not restrict specific formulas for solving thesubproblems, hence IPAD has quite flexible inner iterationschemes that can fuse efficient numerical methods into.

Page 4: An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure

3.2. Flexible Inner Iteration Schemes

Our IPAD framework is highly flexible since it allowsfusing any efficient algorithms, to compute inexact solu-tions for specific subproblems. This flexibility is especiallywelcomed in practice: since the constraints added on vari-ables are always quite different from one another, efficientalgorithms for solving specific subproblems should be care-fully chosen. For example, when facing the SDL prob-lem with `1 regularization [15, 24], both subproblems areconvex. Then various numerical methods designed underconvex case like homotopy method [14], FISTA [6] andADMM [9] can be applied to solving these subproblems.

On the other hand, although many previously used in-exact solvers can be presented under our IPAD framework[21, 30, 11], almost all the inner iteration schemes are ex-perimentally designed. For example, a 2-step inner loop issupposed to be “good-enough” in [30]; the authors of [21]stop the inner iterations at fixed steps to achieve acceptablesolutions. Different from the previous work, we in the fol-lowing Criterion 1 provide theoretical conditions for stop-ping the inner iterations.

Since calculating the inexact solutions brings errors etxand ety in the first-order optimality conditions:

etx = gtx +∇xH(xt,yt−1) + ηt−11 (xt − xt−1),

ety = gty +∇yH(xt,yt) + ηt−12 (yt − yt−1),

(5)

with gtx ∈ ∂f(xt) and gty ∈ ∂g(yt), we in the followingCriterion 1 show that these errors should be bounded to acertain extent at every iteration.

Criterion 1 The errors etx and ety in Eq. (5) must satisfy

‖etx‖ ≤ Cx‖xt − xt−1‖, ‖ety‖ ≤ Cy‖yt − yt−1‖, (6)

where parameters Cx and Cy are two positive integers de-fined before the iteration starts.

With the stopping criteria (6) for inner iteration schemes,we summarize the whole algorithm in Alg. 1. However, thisinexact framework is still on conceptual progress: the stop-ping criteria can not be directly calculated from (6) and animplementation should be put forward for carrying IPADfor practical use. Before providing a practical implementa-tion in Sec. 3.4, we first analyze the convergence propertiesof IPAD with the help of the well-designed Criterion 1.

3.3. Convergence Analyses

In this section, we provide the theoretical support forIPAD method3. The strategic point on analyzing the con-vergence properties of IPAD is regarding xt+1 and yt+1 as

3Due to space limit, all the related proofs in this section will be de-tailedly given in the supplemental material.

the exact solutions on solving the following subproblems:

minxf(x) +H(x,yt) +

ηt12‖x− xt‖2 − (et+1

x )>x,

minyg(y) +H(xt+1,y) +

ηt22‖y − yt‖2 − (et+1

y )>y.

(7)

This equivalent conversion is rigid since the first-order op-timality conditions of (7) are exactly the same with Eq. (5).However, it should be emphasized that xt+1 and yt+1 arenot computed by directly minimizing (7); this equivalentconversion is nothing but assisting in theoretical analyses.

Before proposing the main theorem, we give the require-ments on {ηt1}t∈N and {ηt2}t∈N.

Assumption 1 Parameters {ηt1}t∈N and {ηt2}t∈N satisfy

ηt1 > 2Cx, ηt2 > 2Cy, for all t ∈ N, (8)

to ensure the whole IPAD algorithm converges.

Then with the help of Assumption 1, we can obtain themain theorem in Theorem 1: our proposed IPAD has thebest convergence property, that is, the global convergenceproperty for general NNP: our IPAD generates a Cauchysequence that converges to a critical point of the problem.

Theorem 1 Under the Assumption 1 and suppose the se-quence {(xt,yt)}t∈N generated by IPAD is bounded. Then{(xt,yt)}t∈N is a Cauchy sequence that converges to a crit-ical point (x∗,y∗) of Ψ.

For proving this main convergence theorem, the two as-sertions in the following key lemma are the cornerstones.From the two assertions, obviously, the objective functionΨ is sufficiently descent (Eq. (9) in Lemma 1) during it-erations, which together with the second assertion ensuresthat there is a subsequence of {zt}t∈N that converges to acritical point of the problem. By further combining the KŁproperty, we have the global convergence property for ourproposed IPAD method as claimed in Theorem 1.

Lemma 1 Suppose that the sequence {(xt,yt)}t∈N gener-ated by IPAD is bounded. Then the following two assertionshold under the Assumption 1:

Ψ(zt)−Ψ(zt+1) ≥ a(‖xt+1 − xt‖2 + ‖yt+1 − yt‖2), (9)

dist(0, ∂Ψ(zt)) ≤ b(‖xt − xt−1‖+ ‖yt − yt−1‖), (10)

where constants a = mint∈N{ηt1

4 −(Cx)2

ηt1,ηt24 −

(Cy)2

ηt2} and

b = maxt∈N{ηt−11 + Cx,M + ηt−1

2 + Cy} with Lipschitzconstant M of∇H on bounded sets.

For the convergence rate, our IPAD method shares thesame result with PALM [8] when the desingularising func-tion φ(s) = C

θ sθ : [0, µ) → R+ of Ψ is satisfied with

Page 5: An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure

positive constant C and θ ∈ [0, 1). Specifically, IPAD con-verges in a finite number of steps when θ equals to 1. Forθ ∈ [0, 1

2 ) and θ ∈ ( 12 , 1), IPAD converges with a sublinear

rate and a linear rate respectively. Though the convergencerate of IPAD is the same with PALM in theory (the conver-gence rate is not affected by the algorithm but the objectivefunction Ψ of NNP [17]), the experimental results in Sec-tion 4 verify the efficiency of our algorithm.

Remark 2 It is natural to extend IPAD to solving the gen-eral multi-block NNP (1). Furthermore, all the conver-gence analyses conducted on problem (2) can be straight-forwardly extended to the multi-block case. Since not a fewproblems can be (re)formulated as the general formulation(1), thus our IPAD method can be applied for solving awider range of problems in applications.

We can see from the above contents that, arbitrarily in-corporating inner numerical methods into IPAD still ensurethe global convergence property of the whole algorithm aslong as the stopping criteria (6) and the parameter con-ditions (8) are satisfied. From this perspective, IPAD isnot a single algorithm, but a general and flexible algorithmframework with rigid convergence properties. In addition,through blending other numerical methods into our IPADframework, some previous inexact methods applied to ap-plications [24, 21, 11] can also be proved to achieve theglobal convergence property.

3.4. Implementable Error Bound

Though we have given criteria of inexactness in the Cri-terion 1, there appears another problem: the errors etx andety can not be directly calculated. Since gtx and gty in Eq. (5)are unattainable in practice, etx and ety can not be directlycalculated by the equalities in Eq. (5). Thus the aforemen-tioned IPAD algorithm, i.e., Alg. 1 is only in the conceptualprogress; the criteria (6) are not implementable in practice.

Instead, we in this section provide a truly implementa-tion for carrying IPAD for practical use. After calculatingthe inexact solutions xt and yt of Eq. (4) by some numer-ical methods, we first compute two intermediate variablesxt and yt as the solutions of specific proximal mappings4

xt = prox1f (vtx), yt = prox1

g(vty), (11)

where variables vtx and vty are computed as follows:

vtx = xt −∇xH(xt,yt−1)− ηt−11 (xt − xt−1),

vty = yt −∇yH(xt,yt)− ηt−12 (yt − yt−1).

(12)

4proxτσ(x) ∈ arg minz σ(z)+ τ2‖z−x‖2 denotes proximal mapping

to proper, lower semi-continuous function σ.

After getting the intermediate variables xt and yt from xt

and yt, the errors etx and ety are calculated by:

etx = (1− ηt−11 )(xt − xt) +∇xH(xt,yt−1)−∇xH(xt,yt−1),

ety = (1− ηt−12 )(yt − yt) +∇yH(xt,yt)−∇yH(xt, yt).

(13)We give a proposition as follows to show that these two

errors are exactly the implementable versions of Eq. (5). Inorder to make discussions smoother, we provide the proofof Proposition 1 at the end of this paper in Appendix A.

Proposition 1 The errors, etx and ety calculated though theEq. (13) are equivalent to the ones in Eq. (5).

Thus, the errors in (13) are used for checking whether thestopping criteria (6) are satisfied; once the criteria are sat-isfied, we re-assign xt to xt and yt to yt. Then xt and yt

become the final solutions of their corresponding subprob-lems at t-th step. For clarity, we give the detailed inexactprocess of inner iteration schemes in Alg. 2.

4. Experimental ResultsThough our IPAD algorithm framework can be directly

applied to numerous problems in computer vision and ma-chine learning [40, 21, 27, 11], we in this paper just considerto utilize the widely concerned `0-regularized SDL model[5] as an example to verify the flexibility and efficiency ofour IPAD framework. All the compared algorithms are im-plemented by Matlab R2013b and tested on a PC with 8 GBof RAM and Intel Core i5-4200M CPU.

4.1. SDL with `0 Penalty

The SDL problem with `0 penalty is formulated as:

minD,W

1

2‖I−DW>‖2+λ‖W‖0+XW(W)+XD(D), (14)

where ‖ · ‖0 denotes the `0 penalty that counts the numberof non-zero elements of W. Here the indicator function Xacts on the set

D = {D = {di}mi=1 ⊂ Rn×m : ‖di‖ = 1,∀i}.

As for another setW , we make it empty for synthetic databut define it as

W = {W = {wi}mi=1 ⊂ Rp×m : ‖wi‖∞ ≤ Ub,∀i},

for real-world data to enhance the stability of the model [5].Moreover, we denote S(W) = λ‖W‖0 +XW(W) to sim-plify the deduction.

It is observed that problem (14) is a special case of prob-lem (2). Thus, we can apply IPAD for inexactly solvingthe following subproblems since it is extremely hard to getexact solutions of the subproblems.

minWS(W) +

1

2‖I−DtW>‖2 +

ηt12‖W −Wt‖2, (15)

Page 6: An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure

0 10 20 30 40 50 60 70 80 90 100

0.5

1

1.5

2

2.5

3

3.5

4

4.5

x 10−3

Iteration Number

‖Wt+

1−W

t ‖/‖W

t ‖

PALMINVPITHADMMP2A

(a) ‖Wt+1 −Wt‖/‖Wt‖

0 10 20 30 40 50 60 70 80 90 100

0.5

1

1.5

2

2.5

3

3.5

4

4.5

x 10−3

Iteration Number

‖Dt+

1−D

t ‖/‖D

t ‖

PALMINVPITHADMMP2A

(b) ‖Dt+1 −Dt‖/‖Dt‖

0 10 20 30 40 50 60 70 80 90 100

0.5

1

1.5

2

2.5

3

3.5

4

4.5

x 10−3

Iteration Number

‖Ψt+

1−Ψ

t ‖/‖Ψ

t ‖

PALMINVPITHADMMP2A

(c) ‖Ψt+1 −Ψt‖/‖Ψt‖

0 5 10 15 20 25 30 35 40 45 50

2

4

6

8

10

12

14

16

18

20

Iteration Number

Inne

r Ite

ratio

n N

umbe

r

PITH (C)ADMM (D)P2A (D)

(d) Inner Iteration

0 5 10 15 20 25 30 35 40 45 50 55

1

2

3

4

5

6

7

8

9

x 10−3

Iteration Number

‖Wt+

1−W

t ‖/‖W

t ‖

PALMINVADMMP2A

(e) ‖Wt+1 −Wt‖/‖Wt‖

0 5 10 15 20 25 30 35 40 45 50 55

0.5

1

1.5

2

2.5

3

3.5

4

4.5

x 10−3

Iteration Number

‖Dt+

1−D

t ‖/‖D

t ‖

PALMINVADMMP2A

(f) ‖Dt+1 −Dt‖/‖Dt‖

0 5 10 15 20 25 30 35 40 45 50 55

0.5

1

1.5

2

2.5

3

3.5

4

4.5

x 10−3

Iteration Number

‖Ψt+

1−Ψ

t ‖/‖Ψ

t ‖

PALMINVADMMP2A

(g) ‖Ψt+1 −Ψt‖/‖Ψt‖

0 2 4 6 8 10 12 14 16 18 20 221

1.5

2

2.5

3

3.5

4

Iteration Number

Inne

r Ite

ratio

n N

umbe

r

ADMM (D)P2A (D)

(h) Inner Iteration

0 5 10 15 20 25 30 35

1

2

3

4

5

6

7

8

9

10x 10

−3

Iteration Number

‖Wt+

1−W

t ‖/‖W

t ‖

PALMINVADMMP2A

(i) ‖Wt+1 −Wt‖/‖Wt‖

0 5 10 15 20 25 30 35

0.5

1

1.5

2

2.5

3

3.5

4

4.5

x 10−3

Iteration Number

‖Dt+

1−D

t ‖/‖D

t ‖

PALMINVADMMP2A

(j) ‖Dt+1 −Dt‖/‖Dt‖

0 5 10 15 20 25 30 35

0.5

1

1.5

2

2.5

3

3.5

4

4.5

x 10−3

Iteration Number

‖Ψt+

1−Ψ

t ‖/‖Ψ

t ‖

PALMINVADMMP2A

(k) ‖Ψt+1 −Ψt‖/‖Ψt‖

0 2 4 6 8 10 12 14 16 181

1.5

2

2.5

3

3.5

4

4.5

5

Iteration Number

Inne

r Ite

ratio

n N

umbe

r

ADMM (D)P2A (D)

(l) Inner Iteration

Figure 1. The convergence properties of using PALM [8], INV [5], IPAD-PITH (PITH for short in the figures), IPAD-ADMM (ADMM forshort) and IPAD-P2A (P2A for short) for SDL problem with `0 penalty on synthetic data. The convergence results in the first row belongsto n = 64; the second row belongs to n = 144 and the last row belongs to n = 256.

minDXD(D) +

1

2‖I−D(Wt+1)>‖2 +

ηt22‖D−Dt‖2. (16)

Moreover, PALM can also be applied to solve (14), but itrequires computing Lipschitz constants at every iteration.

Notice that solving the nonconvex subproblem (15) isnot a trivial task. Here we apply a proximal iterative hard-thresholding (PITH) algorithm [3, 20] to solve this subprob-lem. On the other hand, we apply ADMM [9] for solvingsubproblem (16).

4.2. Synthetic Data

We generate synthetic data with different sizes to helpanalyze the property of IPAD (see Table 1). All the algo-rithms for the synthetic data stop when satisfying:

max

{‖Dt+1 −Dt‖‖Dt‖

,‖Wt+1 −Wt‖‖Wt‖

,‖Ψt+1 −Ψt‖‖Ψt‖

}< 1e−4,

(17)where Ψt is the objective value at step t.

4.2.1 Efficiency of Inexact Strategy

To show the respective effects of using inexact strategies ondifferent subproblems, we propose IPAD-PITH which ob-tains Wt+1 by PITH but keeps D-subproblem the same asPALM5. We also design IPAD-ADMM that computes Dt+1

by ADMM but remains W-subproblem the same as PALM.The comparisons in Table 1 among PALM, IPAD-PITH

and IPAD-ADMM show that inexact strategies help reducethe iteration steps: both the IPAD-PITH and IPAD-ADMMconverges with less iterations than PALM. However, theperformances of IPAD-PITH and IPAD-ADMM are quitedifferent in inner iterations. We can see from Figure 1(d)that IPAD-ADMM uses few inner steps during iterations.However, IPAD-PITH reaches the maximum inner steps(set as 20) at almost every iteration. This from one sideshows that ADMM is suitable for solving (16) but PITH

5Due to the space limit, we give detailed algorithm implementations,further theoretical analyses and more results in supplemental material.

Page 7: An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure

Data n = 64,m = 600, p = 4000 n = 144,m = 900, p = 10000 n = 256,m = 1600, p = 16000

Alg. PALM INV PITH ADMM P2A PALM INV ADMM P2A PALM INV ADMM P2AOut-iter 104 23 51 22 14 56 33 22 15 31 38 18 12Time (s) 52.82 7.81 253.56 8.72 8.31 96.08 45.93 35.58 38.94 319.97 253.17 150.25 158.42

Table 1. The number of outer iterations and the iteration time (s) of PALM [8], INV [5], our IPAD-PITH (PITH for short in this table), ourIPAD-ADMM (ADMM for short) and our IPAD-P2A (P2A for short) for synthetic data. The results are the averages of multiple tests.

Image / σ / λ PALM [8] mPALM [4] INV [5] IPAD-ADMM

“Peppers512” / 30 / 5500PSNR Iter Time PSNR Iter Time PSNR Iter Time PSNR Iter Time28.64 4 4.11 30.11 10 325.53 30.14 61 48.63 30.21 25 23.29

“Lena512” / 25 / 4500PSNR Iter Time PSNR Iter Time PSNR Iter Time PSNR Iter Time28.85 4 3.98 31.04 14 459.10 31.11 62 50.20 31.13 31 37.78

“Barbara512” / 20 / 3500PSNR Iter Time PSNR Iter Time PSNR Iter Time PSNR Iter Time28.73 4 4.13 30.06 11 399.19 30.09 58 45.09 30.22 18 16.54

“Hill512” / 15 / 2500PSNR Iter Time PSNR Iter Time PSNR Iter Time PSNR Iter Time29.84 4 4.03 31.20 26 85.34 31.31 84 69.73 31.32 19 17.81

Table 2. The PSNR scores of the recovered images, number of outer iterations and the whole iteration time (s) of PALM [8], mPALM [4],INV [5] and our IPAD-ADMM for real-world data.

Method σ = 30 σ = 25 σ = 20 σ = 15

KSVD[15]

PSNR / Time PSNR / Time PSNR / Time PSNR / Time29.21 / 184.88 30.09 / 235.22 31.14 / 318.06 32.49 / 481.48

IPAD-ADMM

PSNR / Time PSNR / Time PSNR / Time PSNR / Time28.98 / 20.04 29.85 / 26.75 30.89 / 22.72 32.28 / 22.65

Table 3. Quantitative denoising results of KSVD [15] and IPAD-ADMM on 7 widely-used example images.

is less efficient for solving (14). On the other side it iscaused by the problems themselves: (16) has unique solu-tion but problem (15) is a challenging NP hard problem andonly sub-optimal solution can be found in polynomial time.Since PITH converges with unexpected time, thus we onlytest it on the data with relatively low dimension (n = 64).

We also apply inexact strategies to both subproblems ofD and W. Since PITH is time-consuming for solving (15),thus we reduce the maximum inner step to 2 for PITH. Wename this algorithm as IPAD-P2A. Then we can see fromthe Figure 1 and Table 1 that IPAD-P2A uses less iterationsteps and sometimes converges faster to an optimal solution.

Remark 3 Though Theorem 1 is proved for applying theinexact strategy to both subproblems, the two hybrid formsof IPAD, i.e. IPAD-PITH and IPAD-ADMM, are also con-verged. Furthermore, a hybrid IPAD optionally combinesPALM, PAM and IPAD can be proved to have global con-vergence property for solving problem (2)

Remark 4 It can be observed that IPAD-P2A can not beseen as a special case of the hybrid IPAD since it containsone inexact strategy for D and two-steps prox-linear itera-tions for W. But fortunately, the experimental results ver-ify the convergence of IPAD-P2A and we can also prove theconvergence from theoretical analyses.

By comparing the algorithms, all the inexact strategiesof our IPAD method perform better than PALM and areverified to be practicable, converged and efficient. How-ever, a less efficient numerical method for solving subprob-lem indeed reduce the efficiency of the whole algorithm.Therefore, we suggest carefully choosing effective numeri-cal methods for solving subproblems.

4.2.2 Other Comparisons

At last, we compare INV that solves Wt+1 in the same wayas PALM but treats Dt+1 as the solution of a linear systemfirst and then project the solution on set D. Though thisstrategy seems to be efficient in practice [5], it lacks the-oretical guarantee. Firstly, Dt+1 calculated by INV is notan exact solution of (16). Secondly, it is computed with-out measuring the inexactness. So applying INV sometimescreates oscillations during iterations (Figure 1 (f)) and theperformances of INV are unstable especially in real-worldapplications (see the experimental results in Section 4.3).Thus we do not recommend using it in practice.

4.3. Real-world Data

We apply IPAD to real-world data on image denoisingproblem [15, 12] and compare PALM [8], mPALM [4], INV[5] with our IPAD-ADMM for solving this real-world appli-

Page 8: An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure

0 10 20 30 40 50 60 70 80

0.01

0.02

0.03

0.04

0.05

0.06

0.07

Iteration Number

‖Dt+

1−D

t ‖/‖D

t ‖

PALMmPALMINVADMM

(a) “Hill512”, σ = 15

0 10 20 30 40 50 600.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

0.09

0.1

Iteration Number

‖Dt+

1−D

t ‖/‖D

t ‖

PALMmPALMINVADMM

(b) “Child512”, σ = 30

Figure 2. Comparing the convergence performances with variousalgorithms on two images.

0 5 10 15 200.5

0.6

0.7

0.8

0.9

1

1.1

1.2

1.3

1.4

1.5

Iteration Number

Inne

r Ite

ratio

n N

umbe

r

ADMM (D)

(a) “Couple512”, σ = 20

0 5 10 15 20 25 30

1

2

3

4

5

6

Iteration Number

Inne

r Ite

ratio

n N

umbe

r

ADMM (D)

(b) “Lena512”, σ = 25

Figure 3. Inner iterations of IPAD-ADMM on two examples.

cation. All the algorithms for comparison terminate whenreaching ‖Dt+1−Dt‖/‖Dt‖ < 1e−2 and are demonstratedon 7 widely-used images. The patches in each image, ofsize 8 × 8, are regularly sampled in an overlapping man-ner. The noisy images are obtained by adding Gaussianrandomly noises with level σ = 15, 20, 25, 30.

As shown in Table 2, PALM seems to converge quicklybut get bad recovered results. However, the truth is: thelarge upper bound ‖(Dt+1)>Dt+1‖2 of the Lipschitz con-stant emphasizes the function of the proximal term. So itcauses tiny differences between Dt+1 and Dt. Thus, PALMdoes not converge when reaching the stopping criterion; onthe contrary, it converges quite slow. For the failure of us-ing PALM, we adopt the strategy used in [4], which regardsthe problem (14) as a multi-block problem: solving {di}mi=1

separately by PALM. We name their algorithm mPALM andshow the results in Table 2. This time, mPALM convergesto an acceptable result when reaching the stopping criterion.

We in Fig. 2 present comparison results on convergenceperformances: the vibration of INV [5] causes more itera-tions than IPAD-ADMM. We also select two examples inFig. 3 to present the inner iteration numbers of our inexactalgorithm. It can be seen from the figure that the inner iter-ation behaviors may be in totally different style; it dependson the input data and algorithm parameters. In addition, wealso some visual comparisons in Fig. 4 and 4 to show thatour IPAD-ADMM performs more stable and efficient forthe real-world applications.

(a) mPALM [4] (b) INV [5] (c) IPAD-ADMMFigure 4. The dictionaries learned by mPALM [4], INV [5] andour IPAD-ADMM on “Barbara512” with σ = 20.

(a) Noisy image (b) Recovered image

Figure 5. Illustrating image denoising performance on an exampleimage. (a) noisy image with σ = 30, (b) image recovered byIPAD-ADMM.

Finally, we give comparisons between the state-of-the-art KSVD technique [15] and our proposed IPAD-ADMMin Table 3. First we want to point it out that KSVD is de-signed for solving a quasi-`0 (not the exact `0 penalty) prob-lem with the usage of OMP [29]. We can see that thoughour PSNR values are slightly less than KSVD, we use lessthan a tenth of the time of KSVD. Thus, our IPAD is a fastalgorithm with flexible inner iterations for solving the prob-lem SDL with `0 penalty in practice.

5. Conclusions

This paper provided a fast optimization framework tosolve the challenging nonconvex and nonsmooth program-mings (NNPs) in vision and learning societies. Differentfrom most existing solvers, which always fix their updat-ing schemes during iterations, we showed that under somemild conditions, any numerical algorithms can be incorpo-rated into our general algorithmic framework and the con-vergence of the hybrid iterations can always be guaranteed.We conjectured that our theoretical results are the best wecan ask for unless further assumptions are made on the gen-eral NNPs in (1). Numerical evaluations on both syntheticdata and real images demonstrated promising experimentalresults of the proposed algorithms.

Page 9: An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure

6. Appendix A: proof of Proposition 1Proof From the Alg. 2, at the t-th iteration, suppose that aninner iteration scheme is conducted by K time, and denoteits current iterative solution as xt,K . Then from the calcu-lations in Eq. (11), we can deduce the following equalities.

xt,K =prox1f (xt,K −∇xH(xt,K ,yt−1)− ηt−1

1 (xt,K − xt−1))

=prox1f (xt,K −∇xH(xt,K ,yt−1)− ηt−1

1 (xt,K − xt−1)

+(1− ηt−11 )(xt,K − xt,K)

+∇xH(xt,K ,yt−1)−∇xH(xt,K ,yt−1)).(18)

Once the error et,Kx satisfies the inexact condition Crit 1,then xt,K will be regarded as the solution of the t-th itera-tion. That is, by substituting the notation xt,K with xt inEq. (18), thus we get

xt = prox1f (xt−∇xH(xt,yt−1)−ηt−1

1 (xt−xt−1)+etx).

The above deductions can be similarly extended to the caseof yt. Together with Eq. (12), we have

xt = prox1f (vtx + etx), yt = prox1

g(vty + ety). (19)

From the definition of prox mapping, Eq. (19) is equal to

etx = gtx +∇xH(xt,yt−1) + ηt−11 (xt − xt−1),

ety = gty +∇yH(xt,yt) + ηt−12 (yt − yt−1).

(20)

where gt+1x ∈ ∂f(xt+1) and gt+1

y ∈ ∂g(yt+1). The aboveequalities are exactly the first-order optimality conditions of(4) by regarding etx and ety as the inexactness. �

References[1] M. Afonso and S. J. Miguel. Blind inpainting us-

ing l0 and total variation regularization. IEEE TIP,24(7):2239–53, 2015. 2

[2] H. Attouch, J. Bolte, P. Redont, and A. Soubeyran.Proximal alternating minimization and projectionmethods for nonconvex problems: an approach basedon the kurdyka-łojasiewicz inequality. Mathematics ofOperations Research, 35:438–457, 2010. 1, 2, 3

[3] F. Bach, R. Jenatton, J. Mairal, and G. Obozinski.Optimization with sparsity-inducing penalties. Foun-dations & Trendsrin Machine Learning, 4(1):1–106,2011. 6

[4] C. Bao, H. Ji, Y. Quan, and Z. Shen. `0 norm baseddictionary learning by proximal methods with globalconvergence. In CVPR, 2014. 7, 8

[5] C. Bao, H. Ji, Y. Quan, and Z. Shen. Dictionary learn-ing for sparse coding: Algorithms and convergenceanalysis. IEEE TPAMI, 38(7):1356–1369, 2016. 1,5, 6, 7, 8

[6] A. Beck and M. Teboulle. A fast iterative shrinkage-thresholding algorithm for linear inverse problems.Siam Journal on Imaging Sciences, 2(1):183–202,2009. 4

[7] J. Bolte and E. Pauwels. Majorization-minimizationprocedures and convergence of sqp methods for semi-algebraic and tame programs. 41(2), 2015. 1

[8] J. Bolte, S. Sabach, and M. Teboulle. Proximal al-ternating linearized minimization for nonconvex andnonsmooth problems. Mathematical Programming,146:459–494, 2014. 1, 2, 3, 4, 6, 7

[9] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eck-stein. Distributed optimization and statistical learn-ing via the alternating direction method of multipli-ers. Foundations & Trendsrin Machine Learning,3:1–122, 2011. 4, 6

[10] X. Bresson. A short note for nonlocal tv minimization.Technical report, 2009. 2

[11] C. Chen, M. N. Do, and J. Wang. Robust image andvideo dehazing with visual artifact suppression viagradient residual minimization. In ECCV, 2016. 2,3, 4, 5

[12] P. Y. Chen and I. W. Selesnick. Group-sparse signaldenoising: Non-convex regularization, convex opti-mization. IEEE TSP, 62(13):3464–3478, 2013. 7

[13] Y. Chen and T. Pock. Trainable nonlinear reaction dif-fusion: A flexible framework for fast and effective im-age restoration. IEEE TPAMI, pages 1–1, 2015. 3

[14] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani.Least angle regression. Mathematics, 32(2):2004,2004. 4

[15] M. Elad and M. Aharon. Image denoising via sparseand redundant representations over learned dictionar-ies. IEEE TIP, 15(12):3736–3745, 2006. 4, 7, 8

[16] J. Fan and R. Li. Variable selection via nonconcave pe-nalized likelihood and its oracle properties. Journal ofthe American statistical Association, 96(456):1348–1360, 2001. 3

[17] P. Frankel, G. Garrigos, and J. Peypouquet.Splitting methods with variable metric forkurdyka−łojasiewicz functions and general con-vergence rates. Journal of Optimization Theory &Applications, 165(3):874–900, 2015. 5

[18] K. Gregor and Y. Lecun. Learning fast approximationsof sparse coding. In Proc. International Conference onMachine Learning, 2010. 1, 3

[19] X. Guo, X. Cao, and Y. Ma. Robust separation of re-flection from multiple images. In CVPR, 2014. 2

[20] K. K. Herrity, A. C. Gilbert, and J. A. Tropp. Sparseapproximation via iterative thresholding. In ICASSP,2006. 6

Page 10: An Optimization Framework with Flexible Inexact Inner ...arxiv.org/pdf/1702.08627v3.pdftice, IPAD theoretically give rigid conditions on parame-ters and stopping criteria to ensure

[21] K. Huang, N. D. Sidiropoulos, and A. P. Liavas. Aflexible and efficient algorithmic framework for con-strained matrix and tensor factorization. IEEE TSP,64(19):5052–5065, 2015. 4, 5

[22] J. Kim and H. Park. Fast nonnegative matrix factor-ization: An active-set-like method and comparisons.Siam J. Scientific Computing, 33(6):3261–3281, 2011.1, 2

[23] C. Lu, J. Tang, S. Yan, and Z. Lin. Noncon-vex nonsmooth low-rank minimization via iterativelyreweighted nuclear norm. IEEE TIP, 25(2):829–839,2015. 2

[24] J. Mairal, F. Bach, J. Ponce, and G. Sapiro. Onlinelearning for matrix factorization and sparse coding.Journal of Machine Learning Research, 11(1):19–60,2010. 4, 5

[25] J. Pan, Z. Hu, Z. Su, and M. H. Yang. l0-regularizedintensity and gradient prior for deblurring text imagesand beyond. IEEE TPAMI, 39(2):342, 2017. 2, 3

[26] H. Shen and J. Z. Huang. Sparse principal componentanalysis via regularized low rank matrix approxima-tion. Journal of Multivariate Analysis, 99(6):1015–1034, 2008. 1

[27] J. Shi, X. Ren, G. Dai, and J. Wang. A non-convexrelaxation approach to sparse dictionary learning. InCVPR, 2011. 1, 5

[28] P. Sprechmann, A. M. Bronstein, and G. Sapiro.Learning efficient sparse and low rank models. IEEETPAMI, 37(9):1821–33, 2012. 3

[29] J. Tropp and A. C. Gilbert. Signal recovery from ran-dom measurements via orthogonal matching pursuit.IEEE Transactions on Information Theory, 53:4655–4666, 2007. 8

[30] Y. Wang, R. Liu, X. Song, and Z. Su. A nonlocall0 model with regression predictor for saliency detec-tion and extension. The Visual Computer, pages 1–16,2016. 2, 4

[31] C. Xu, Z. Lin, Z. Zhao, and H. Zha. Relaxedmajorization-minimization for non-smooth and non-convex optimization. 2016. 2

[32] L. Xu, S. Zheng, and J. Jia. Unnatural l0 sparse rep-resentation for natural image deblurring. In CVPR,2013. 2

[33] Y. Xu and W. Yin. A block coordinate descentmethod for regularized multiconvex optimization withapplications to nonnegative tensor factorization andcompletion. SIAM Journal on imaging sciences,6(3):1758–1789, 2013. 1

[34] Y. Xu and W. Yin. A globally convergent algorithmfor nonconvex optimization based on block coordinateupdate. arXiv preprint, 2014. 1

[35] Y. Xu, W. Yin, Z. Wen, and Y. Zhang. An alternatingdirection algorithm for matrix completion with non-negative factors. technical report, Shanghai JiaotongUniversity, 2012. 3

[36] Y. Yang, J. Sun, H. B. Li, and Z. B. Xu. Deep admm-net for compressive sensing mri. In NIPS, 2016. 2,3

[37] G. Yuan and B. Ghanem. `0tv: A new method forimage restoration in the presence of impulse noise. InCVPR, 2013. 1, 2

[38] C. Zhang. Nearly unbiased variable selection underminimax concave penalty. The Annals of Statistics,pages 894–942, 2010. 3

[39] T. Zhang, S. Liu, N. Ahuja, M. H. Yang, andB. Ghanem. Robust visual tracking via consistent low-rank sparse learning. IJCV, 111(2):171–190, 2015. 2

[40] H. Zou, T. Hastie, and R. Tibshirani. Sparse princi-pal component analysis. Journal of Computational &Graphical Statistics, 2007:1–30, 2012. 1, 5

[41] W. Zuo, D. Ren, D. Zhang, S. Gu, and L. Zhang.Learning iteration-wise generalized shrinkage-thresholding operators for blind deconvolution. IEEETIP, 25(4):1751–1764, 2016. 2