a stochastic majorize-minimize subspace algorithm for online … · 2016-01-05 · the classical...

12
1 A Stochastic Majorize-Minimize Subspace Algorithm for Online Penalized Least Squares Estimation Emilie Chouzenoux, Member, IEEE, and Jean-Christophe Pesquet, Fellow, IEEE Abstract—Stochastic approximation techniques play an impor- tant role in solving many problems encountered in machine learning or adaptive signal processing. In these contexts, the statistics of the data are often unknown a priori or their direct computation is too intensive, and they have thus to be estimated online from the observed signals. For batch optimization of an objective function being the sum of a data fidelity term and a penalization (e.g. a sparsity promoting function), Majorize- Minimize (MM) methods have recently attracted much interest since they are fast, highly flexible, and effective in ensuring convergence. The goal of this paper is to show how these methods can be successfully extended to the case when the data fidelity term corresponds to a least squares criterion and the cost function is replaced by a sequence of stochastic approximations of it. In this context, we propose an online version of an MM subspace algorithm and we study its convergence by using suitable probabilistic tools. Simulation results illustrate the good practical performance of the proposed algorithm associated with a memory gradient subspace, when applied to both non-adaptive and adaptive filter identification problems. Keywords: stochastic approximation, optimization, subspace al- gorithms, memory gradient methods, descent methods, recursive algorithms, majorization-minimization, filter identification, Newton method, sparsity, machine learning, adaptive filtering. I. I NTRODUCTION A classical problem in data sciences consists of inferring the structure of a linear model linking some observed random variables (X n ) n1 in R N×Q to some other observed random variables (y n ) n1 in R Q . Unless otherwise specified, we will assume in this work that the following wide-sense stationarity properties hold: (n N ) E(y n 2 )= (1) E(X n y n )= r (2) E(X n X n )= R, (3) where ]0, +[, r R N , R R N×N is a symmetric positive semi-definite matrix, E(·) denotes the mathematical expectation, and ‖·‖ is the Euclidean norm. We will then be interested in the following optimization formulation: minimize hR N F (h), (4) E. Chouzenoux and J.-C. Pesquet (corresponding author) are with the Laboratoire d’Informatique Gaspard Monge, UMR CNRS 8049, Uni- versit´ e Paris-Est, 77454 Marne la Vall´ ee Cedex 2, France. E-mail: [email protected]. This work was supported by the CNRS Imag’in project under grant 2015 OPTIMISME. Part of it was presented at the EUSIPCO 2014 conference [1]. with (h R N ) F (h)= 1 2 E ( y n X n h2 ) + Ψ(h), (5) where Ψ is a function from R N to R, playing the role of a regularization function. In particular, this function may be useful to incorporate some prior knowledge about the sought parameter vector h, e.g. some sparsity requirement, possibly in some transformed domain. Problem (4) is encountered in numerous applications such as system identification, channel equalization, linear prediction or interpolation, echo cancel- lation, interference removal, and supervised classification. In the latter area, (X n ) n1 are vectors (Q =1) which may correspond to features obtained through some nonlinear map- ping of the data to be classified in a given training sequence, and (y n ) n1 may be the associated (discrete-valued) class index vector. Although some other measures (e.g. the logistic regression function) are often more effective in this context, the use of a least squares criterion may still be competitive for simplicity reasons, while the regularization term serves here to avoid overfitting which could arise when the number of extracted features is large. Signal reconstruction constitutes another application field of interest. Then, the vector h corre- sponds to an unknown signal related to some measurements (y n ) n1 obtained through products with matrices (X n ) n1 , and additionally corrupted by some noise process. Each matrix X n with n N corresponds to Q lines of the full acquisition matrix and it is here considered as random. Under suitable stationarity assumptions, the classical least squares data fi- delity term can be modeled as E ( y n X n h2 ) /2, whereas due to the ill-posedness of the great majority of such inverse problems, a regularization term Ψ needs to be introduced so as to obtain reliable estimates. Many optimization algorithms can be devised to solve Problem (4) depending on the assumptions made on Ψ [2]– [5]. In this work, we will be interested in Majorize-Minimize (MM) algorithms [6], [7]. In such approaches, the iterates result from successive minimizations of simple surrogates (e.g. quadratic surrogates) majorizing the cost-function. MM algo- rithms are very flexible and benefit from good theoretical and practical convergence properties. However, the computation load resulting from the minimization of the majorant function may be prohibitive in the context of large scale problems. The strategy we will adopt in this work is to account for subspace acceleration [8], i.e., to constrain the inner minimization step to a subspace of low dimension, typically restricted to the

Upload: others

Post on 12-Jul-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

1

A Stochastic Majorize-Minimize SubspaceAlgorithm for Online Penalized Least Squares

EstimationEmilie Chouzenoux,Member, IEEE, and Jean-Christophe Pesquet,Fellow, IEEE

Abstract—Stochastic approximation techniques play an impor-tant role in solving many problems encountered in machinelearning or adaptive signal processing. In these contexts, thestatistics of the data are often unknown a priori or their directcomputation is too intensive, and they have thus to be estimatedonline from the observed signals. For batch optimization of anobjective function being the sum of a data fidelity term anda penalization (e.g. a sparsity promoting function), Majorize-Minimize (MM) methods have recently attracted much interestsince they are fast, highly flexible, and effective in ensuringconvergence. The goal of this paper is to show how these methodscan be successfully extended to the case when the data fidelityterm corresponds to a least squares criterion and the costfunction is replaced by a sequence of stochastic approximationsof it. In this context, we propose an online version of anMM subspace algorithm and we study its convergence by usingsuitable probabilistic tools. Simulation results illustrate the goodpractical performance of the proposed algorithm associated witha memory gradient subspace, when applied to both non-adaptiveand adaptive filter identification problems.

Keywords: stochastic approximation, optimization, subspace al-gorithms, memory gradient methods, descent methods, recursivealgorithms, majorization-minimization, filter identification, Newtonmethod, sparsity, machine learning, adaptive filtering.

I. I NTRODUCTION

A classical problem in data sciences consists of inferringthe structure of a linear model linking some observed randomvariables(Xn)n>1 in R

N×Q to some other observed randomvariables(yn)n>1 in R

Q. Unless otherwise specified, we willassume in this work that the following wide-sense stationarityproperties hold:

(∀n ∈ N∗) E(‖yn‖2) = (1)

E(Xnyn) = r (2)

E(XnX⊤n ) = R, (3)

where ∈]0,+∞[, r ∈ RN , R ∈ R

N×N is a symmetricpositive semi-definite matrix,E(·) denotes the mathematicalexpectation, and‖ · ‖ is the Euclidean norm.We will then beinterested in the following optimization formulation:

minimizeh∈R

N

F (h), (4)

E. Chouzenoux and J.-C. Pesquet (corresponding author) arewith theLaboratoire d’Informatique Gaspard Monge, UMR CNRS 8049, Uni-versite Paris-Est, 77454 Marne la Vallee Cedex 2, France. E-mail:[email protected].

This work was supported by the CNRS Imag’in project under grant 2015OPTIMISME.

Part of it was presented at the EUSIPCO 2014 conference [1].

with

(∀h ∈ RN ) F (h) =

1

2E(‖yn −X⊤

nh‖2)+Ψ(h), (5)

where Ψ is a function fromRN to R, playing the role of

a regularization function. In particular, this function may beuseful to incorporate some prior knowledge about the soughtparameter vectorh, e.g. some sparsity requirement, possiblyin some transformed domain. Problem (4) is encountered innumerous applications such as system identification, channelequalization, linear prediction or interpolation, echo cancel-lation, interference removal, and supervised classification. Inthe latter area,(Xn)n>1 are vectors (Q = 1) which maycorrespond to features obtained through some nonlinear map-ping of the data to be classified in a given training sequence,and (yn)n>1 may be the associated (discrete-valued) classindex vector. Although some other measures (e.g. the logisticregression function) are often more effective in this context,the use of a least squares criterion may still be competitivefor simplicity reasons, while the regularization term serveshere to avoid overfitting which could arise when the numberof extracted features is large. Signal reconstruction constitutesanother application field of interest. Then, the vectorh corre-sponds to an unknown signalrelated to some measurements(yn)n>1 obtained through products withmatrices(X⊤

n )n>1,and additionallycorrupted by some noise process. EachmatrixX⊤

n with n ∈ N∗ corresponds toQ linesof thefull acquisition

matrix and it is here considered as random. Under suitablestationarity assumptions, the classical least squares data fi-delity term can be modeled asE

(‖yn −X⊤

nh‖2)/2, whereas

due to the ill-posedness of the great majority of such inverseproblems, a regularization termΨ needs to be introduced soas to obtain reliable estimates.

Many optimization algorithms can be devised to solveProblem (4) depending on the assumptions made onΨ [2]–[5]. In this work, we will be interested in Majorize-Minimize(MM) algorithms [6], [7]. In such approaches, the iteratesresult from successive minimizations of simple surrogates(e.g.quadratic surrogates) majorizing the cost-function. MM algo-rithms are very flexible and benefit from good theoretical andpractical convergence properties. However, the computationload resulting from the minimization of the majorant functionmay be prohibitive in the context of large scale problems. Thestrategy we will adopt in this work is to account for subspaceacceleration [8], i.e., to constrain the inner minimization stepto a subspace of low dimension, typically restricted to the

Page 2: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

2

gradient computed at the current iterate and to a memory part(e.g. the difference between the current iterate and a previousone). In a number of recent works [9]–[11], MM subspacealgorithms have shown to provide fast numerical solutions tooptimization problems involving smooth functions, in partic-ular in the case of large-scale problems. Note that, althoughour approach willrequirethat Ψ is a differentiable function,it has been shown that tight approximations of nonsmoothpenalizations such asℓ1 (resp.ℓ0) functions, namelyℓ2 − ℓ1(resp. ℓ2 − ℓ0) functions, can be employed and are oftenquite effective in practice [10], [11]. Another advantage ofthe class of optimization methods under investigation is thattheir convergence can be established under some technicalassumptions, even in the case whenΨ is a nonconvex function(see [10] for more details).

One of the difficulties encountered in machine learning oradaptive processing is that Problem (4) cannot be directlysolved since the second-order statistical moments, r andR are often unknown a priori or their direct computationis too intensive, and they have thus to be estimated onlinefrom the related time series. In the simple case whenΨ = 0,the classical Recursive Least Squares (RLS) algorithm can beused for this purpose [12]. WhenΨ is nonzero, stochasticapproximation algorithms have been developed such as thecelebrated stochastic gradient descent (SGD) algorithm [13]–[16] and some of its proximal extensions [17]–[20]. Thesealgorithms have been at the origin of a tremendous amount ofworks. SGD is known to be robust and easy to implement,but its convergence speed may be relatively slow. Variousextensions of this algorithm have been developed to alleviatethis problem (see [21]–[24] and the references therein). Manyefforts have also been devoted to developing adaptive variantsof this algorithm [25], [26], in particular when identifyingfilters having sparse impulse responses (see e.g. [27]–[33]).In addition, in [34], a set theoretic approach is adopted foronline sparse estimation based on projections onto weightedℓ1 balls, which is extended in [35] by making use of gen-eralized thresholding mappings. It is worth noting that asparse RLS algorithm was proposed in [36] for complex-valued signals in the case whenΨ is anℓ1 norm. This methodtheoretically requires an inner loop implementing an iterativesoft-thresholding technique. An online variant of the RLSalgorithm corresponding to a time weighted LASSO estimatorwas also designed in [37] which relies on a coordinate descentapproach. A similar problem was also addressed in [38] byadopting a novel Bayes variational approach, for which weaktheoretical convergence guarantees however exist. If we except[39] where an adaptive primal-dual splitting is employed todeal with a total variation penalization, in almost all the workson sparse adaptive filtering, the sparsity is directly imposed onthe filter coefficients, without introducing any linear transformof them.

Designing Majorize-Minimize optimization algorithms in astochastic context constitutes a challenging task since mostof the existing works concerning these methods have beenfocused on batch optimization procedures, and the relatedconvergence proofs usually rely on deterministic tools. Wecan however mention a few recent works [40]–[42] where

stochastic MM algorithms are investigated for general lossfunctions under specific assumptions (e.g. the independenceof the involved random variables [40], [41]), but without in-troducing any search subspace. Works which are more closelyrelated to ours are those based on Newton or quasi-Newtonstochastic algorithms [43]–[46], in particular the approaches in[45], [46] provide extensions of BFGS algorithm, but provingthe convergence of these algorithms requires some specificassumptions. Like BFGSapproaches, MM subspace methodsuse a memory of previous estimates so as to accelerate theconvergence.

In Section II, we show how Problem (4) can be reformulatedin a learning context. The MM strategy which is proposedin this work is described in Section III-A. In Section III-B,we give the form of the resulting recursive algorithm and,in Section III-C, we evaluate its computational complexity.A convergence analysis of the proposed stochastic Majorize-Minimize subspace algorithm is performed in Section IV. InSection V, two simulation examples in the context of filteridentification show the good performance of our algorithmwhen a memory gradient subspace is employed. Some con-clusions are drawn in Section VII.

II. PROBLEM FORMULATION

In a learning context, functionF can be replaced by asequence(Fn)n>1 of stochastic approximations of it, whichare defined as follows: for everyn ∈ N

∗,

(∀h ∈ RN ) Fn(h) =

1

2ϑn

n∑

k=1

ϑn−k‖yk −X⊤k h‖2 +Ψ(h)

=1

2ρn − r⊤nh+

1

2h⊤Rnh+Ψ(h),

(6)

whereϑ ∈]0, 1],

ϑn =n−1∑

k=0

ϑk =

n if ϑ = 11− ϑn

1− ϑif ϑ ∈]0, 1[,

(7)

andρn, rn, andRn are given by

ρn =1

ϑn

n∑

k=1

ϑn−k‖yk‖2 (8)

rn =1

ϑn

n∑

k=1

ϑn−kXkyk (9)

Rn =1

ϑn

n∑

k=1

ϑn−kXkX⊤k . (10)

In the case whenϑ = 1, we retrieve the classical sampleestimates of , r, andR. Whenϑ ∈]0, 1[, it can be interpretedas an exponential forgetting factor [12] which may be usefulin adaptive processing scenarios (see Section VI).

Hereafter, we will assume that the regularization functionΨ has the following form:

(∀h ∈ RN ) Ψ(h) =

1

2h⊤V0h−v⊤

0 h+S∑

s=1

ψs(‖Vsh−vs‖)

(11)

Page 3: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

3

wherev0 ∈ RN , V0 ∈ R

N×N is a symmetric positive semi-definite matrix, and, for everys ∈ 1, . . . , S, vs ∈ R

Ps ,Vs ∈ R

Ps×N , and ψs : R → R is a smooth function. Thefirst termh 7→ 1

2h⊤V0h − v⊤

0 h can be viewed as an elasticnet penalty [47], while various choices can be made for thesecond one. As shown in Table I, in addition to quadraticregularization functions (obtained whenS = 1 andψ1 = 0),ℓ2 − ℓ1 functions and smoothedℓ2 − ℓ0 functions constitutestandard choices. The matrices(Vs)16s6S may be set toidentity or they may serve to model possible transforms ordiscrete differentiation operators, and vectors(vs)16s6S maybe used to define reference values.

Note that the regularization strategy adopted in [37] amountsto replacing Ψ in (6) by λnΨ where Ψ is a (possiblyweighted) ℓ1 norm andλn ∈ [0,+∞[. Consistency resultscan then be established under the assumption thatϑ = 1 andlimn→+∞ λn = 0. Our approach here is different, not onlybecause we are interested in a wide class of regularizationfunctions, but also in the sense that we are looking for asolution to the fully regularized problem (4) instead of asolution to the mean square criterion.

Our objective in the next section will be to propose anefficient recursive method for minimizing functions(Fn)n>1.

III. PROPOSED METHOD

A. Majorization property

At each iterationn ∈ N∗, we propose to replaceFn by a

surrogate functionΘn(·,hn) based on the current estimatehn

(computed at the previous iteration). More precisely, a tangentmajorant function is chosen such that

(∀h ∈ RN ) Fn(h) 6 Θn(h,hn) (12)

Fn(hn) = Θn(hn,hn). (13)

For the so-defined MM strategy to be worthwhile, the surro-gate function has to be built in such a way that its minimizationis simple. For this purpose, the following assumptions willbemade on the regularization functionΨ defined in (11):

Assumption 1.(i) For every s ∈ 1, . . . , S, ψs is an even lower-

bounded function, which is continuously differentiable,and limt→0

t 6=0

ψs(t)/t ∈ R, whereψs denotes the deriva-

tive ofψs.(ii) For every s ∈ 1, . . . , S, ψs(

√.) is concave on

[0,+∞[.(iii) There existsν ∈ [0,+∞[ such that(∀s ∈ 1, . . . , S)

(∀t ∈ [0,+∞[) 0 6 νs(t) 6 ν, whereνs(t) = ψs(t)/t.1

These assumptions are satisfied by a wide class of functionsΨ, in particular those corresponding to the choices of thepotential functions(ψs)16s6S listed in Table I.

Note that it follows from (6), (11), and the above definitionof functions (νs)16s6S that, for everyn ∈ N

∗, the gradientof Fn is given by

(∀h ∈ RN ) ∇Fn(h) = An(h)h− cn(h), (14)

1The function is extended by continuity whent = 0.

where

An(h) = Rn + V0 + V ⊤Diag(b(h)

)V ∈ R

N×N (15)

cn(h) = rn + v0 + V ⊤Diag(b(h)

)v ∈ R

N (16)

V = [V ⊤1 . . .V ⊤

S ]⊤ ∈ RP×N (17)

v = [v⊤1 . . .v

⊤S ]⊤ ∈ R

P (18)

with P = P1 + · · ·+ PS , andb(h) =(bi(h)

)16i6P

∈ RP is

such that(∀s ∈ 1, . . . , S) (∀p ∈ 1, . . . , Ps)bP1+···+Ps−1+p(h) = νs(‖Vsh− vs‖). (19)

We have then the following result [48]:

Proposition 1. Under Assumptions 1(i)-1(iii), for everyn ∈N

∗ andh ∈ RN , a tangent majorant ofFn at h is

(∀h′ ∈ RN ) Θn(h

′,h) = Fn(h) +∇Fn(h)⊤(h′ − h)

+1

2(h′ − h)⊤An(h)(h

′ − h),

(20)

whereAn(h) is given by(15).

The proposed MM subspace algorithm consists of definingthe following sequence of random vectors(hn)n>1:

(∀n ∈ N∗) hn+1 ∈ argmin

h∈spanDn

Θn(h,hn), (21)

where spanDn is the vector subspace delineated by thecolumns of matrixDn ∈ R

N×Mn , andh1 has to be set to aninitial value. A common assumption for subspace algorithmswhich will be adopted subsequently is that∇Fn(hn) belongsto spanDn. If, for every n ∈ N

∗, rank(Dn) = N , thealgorithm is similar to a half-quadratic one [48]. Half-quadraticalgorithms are known to be effective batch optimization meth-ods, but the use of such method requires the inversion ofmatrix An(hn) at each iterationn, which may have a highcomputational cost. On the other hand, if for everyn ∈ N

∗,Dn reduces to[−∇Fn(hn),hn], then

hn+1 = un,2hn − un,1∇Fn(hn), (22)

where (un,1, un,2) is a pair of real-valued random variables.In the special case whenun,2 = 1, we recover the formof a SGD-like algorithm with step-sizeun,1. In the machinelearning literature, various forms of the step-size for SGDhave been proposed [23], which often require to tune upsome parameters (e.g. a multiplicative factor) so as to getthe best convergence profile on the available dataset. Theadvantage of the proposed approach is that(un,1, un,2) isautomatically adjusted at each iteration following the MM rule.An intermediate size subspace matrix is obtained by choosing,for everyn ∈ N

∗,

Dn =

[−∇Fn(hn),hn,hn − hn−1] if n > 1

[−∇Fn(h1),h1] if n = 1.(23)

This particular choice for the subspace yields the S3MGalgorithm that will be shown in the next sections to haveboth good convergence properties and a low computationalcomplexity. Note that a similar subspace choice can be foundin batch optimization algorithms such as TWIST [49] orFISTA [50].

Page 4: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

4

TABLE ISMOOTH PENALTY FUNCTIONSψs AND THEIR ASSOCIATED WEIGHTING FUNCTIONSνs . ALL EXPRESSIONS ARE VALID FORt ∈ R, (λs, δs) ∈]0,+∞[2

AND κs ∈ [1, 2].

λ−1s ψs(t) λ−1

s νs(t) Type Name

|t| − δs log(|t|/δs + 1) (|t|+ δs)−1 ℓ2 − ℓ1

t2 if |t| < δs2δs|t| − δ2s otherwise

2 if |t| < δs2δs/|t| otherwise

ℓ2 − ℓ1 Huber

Con

vex

log(cosh(t))

tanh(t)/t if t 6= 0

1 otherwiseℓ2 − ℓ1 Green

(1 + t2/δ2s)κs/2 − 1 κsδ

−2s (1 + t2/δ2s)

κs/2−1 ℓ2 − ℓκs

1− exp(−t2/(2δ2s)) δ−2s exp(−t2/(2δ2s)) ℓ2 − ℓ0 Welsch

t2/(2δ2s + t2) 4δ2s/(2δ2s + t2) ℓ2 − ℓ0 Geman

-McClure

1− (1− t2/(6δ2s))3 if |t| 6

√6δs

1 otherwise

δ−2s (1− t2/(6δ2s))

2 if |t| 6√6δs

0 otherwiseℓ2 − ℓ0 Tukey biweight

Non

conv

ex

tanh(t2/(2δ2s)) δ−2s (cosh(t2/(2δ2s))

−2 ℓ2 − ℓ0 Hyberbolictangent

log(1 + t2/δ2s) 2/(t2 + δ2s) ℓ2 − log Cauchy1− exp(1− (1 + t2/(2δ2s))

κs/2) (κs/(2δ2s))(1 + t2/(2δ2s))κs/2−1 exp(1− (1 + t2/(2δ2s))

κs/2) ℓ2 − ℓκs− ℓ0 Chouzenoux

B. Recursive MM strategy

Since, for everyn ∈ N∗, hn+1 ∈ spanDn, let us set

hn+1 = Dnun, (24)

whereun is anRMn -valued random vector. We deduce from(14), (20), and (21) that, for everyn ∈ N

∗,

un = B†nD

⊤n

(An(hn)hn −∇Fn(hn)

)

= B†nD

⊤n cn(hn), (25)

whereBn = D⊤

nAn(hn)Dn (26)

and(·)† is the pseudo-inverse operation. It is important to notethat, asBn is of dimensionMn × Mn whereMn is small(typically Mn = 3 for the choice of the subspace in (23)whenn > 1), this pseudo-inversion is light. This constitutesthe key advantage of the proposed approach.

By using (7), (9) and (10), the following recursive updatesof (rn)n>1 and (Rn)n>1, can be performed

(∀n ∈ N∗) rn = rn−1 +

1

ϑn(Xnyn − rn−1) (27)

Rn = Rn−1 +1

ϑn(XnX

⊤n −Rn−1), (28)

where we have setr0 = 0 andR0 = ON and we have usedthe identity:ϑϑn−1/ϑn = 1− ϑ

−1

n .Let us now introduce the intermediate variables:

(∀n ∈ N∗) DR

n = RnDn ∈ RN×Mn (29)

DV0

n = V0Dn ∈ RN×Mn (30)

DV

n = V Dn ∈ RP×Mn . (31)

It follows from (15) and (26) that

(∀n ∈ N∗) Bn = D⊤

n

(DR

n +DV0

n

)

+(DV

n

)⊤Diag

(b(hn)

)DV

n . (32)

Without loss of generality, it can be assumed that the algorithmis initialized with h1 = D0u0, whereD0 ∈ R

N×M0 andu0 ∈ R

M0 . Then, (14) and (24) yield

(∀n ∈ N∗) ∇Fn(hn) = DA

n−1un−1 − cn(hn), (33)

where we have set

(∀n ∈ N) DA

n = An+1(hn+1)Dn ∈ RN×Mn . (34)

By using (15) and (28)-(31), the latter variable can be reex-pressed as

DA

n = Rn+1Dn +DV0

n + V ⊤Diag(b(hn+1)

)DV

n

= (1− 1

ϑn+1

)DR

n +1

ϑn+1

Xn+1(X⊤n+1Dn) +DV0

n

+ V ⊤Diag(b(hn+1)

)DV

n . (35)

The resulting relations are summarized in Algorithm 1.

Algorithm 1: Stochastic MM subspace method

r0 = 0,R0 = ON

Initialize D0,u0

h1 = D0u0,DR0 = ON×Mn

,DV0

0 = V0D0,DV0 = V D0

for n = 1, . . . do1rn = rn−1 +

1

ϑn

(Xnyn − rn−1)

2cn(hn) = rn + v0 + V ⊤Diag(b(hn)

)v

3DAn−1 = (1− 1

ϑn

)DRn−1 +

1

ϑn

Xn(X⊤nDn−1)

+DV0

n−1 + V ⊤Diag(b(hn)

)DV

n−1

4∇Fn(hn) = DAn−1un−1 − cn(hn)

5Rn = Rn−1 +1

ϑn

(XnX⊤n −Rn−1)

6SetDn using∇Fn(hn)7DR

n = RnDn,DV0

n = V0Dn,DVn = V Dn

8Bn = D⊤n

(DR

n +DV0

n

)+(DV

n

)⊤Diag

(b(hn)

)DV

n

9un = B†nD

⊤n

(cn(hn)

)

10hn+1 = Dnun

end

Page 5: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

5

TABLE IICOMPLEXITY IN TERMS OF MULTIPLICATIONS FOR ITERATIONn OF

ALGORITHM 1.

Step Complexity Complexityfor V ∈ R

P×N arbitrary whenV = IN

1 N(Q+ 1)2 (N + 1)P N3 Mn−1

(

N(2Q+ P + 1) + P +Q)

Mn−1

(

N(2Q+ 1) +Q)

4 NMn−1

5 N(N + 1)Q/27 NMn(2N + P ) 2N2Mn

8 Mn(

(Mn + 1)(N + P )/2 + P)

NMn(Mn + 3)/29 O(M3

n) +Mn(N +Mn)10 NMn

C. Complexity

As shown in Table II, provided that the subspace dimensions(Mn)n∈N are small, the proposed algorithm has a low com-plexity. Indeed, the global complexity of a direct implementa-tion of the algorithm, evaluated in terms of multiplications atiterationn, is of the order of

N(P (Mn +Mn−1 + 1) +N(4Mn +Q)/2

),

if we assume thatN ≫ maxMn,Mn−1, Q. The first termNP (Mn +Mn−1 +1) corresponds to an upper bound on thecomplexity induced by the use of matrices(Vs)16s6S withinthe regularization term. These matrices often have a sparsestructure (in particular when discrete derivative operators areemployed) which leads to a much lower computational cost.WhenV = IN , the identity matrix ofRN , which is a scenariofrequently encountered in adaptive filtering, this term merelyvanishes in the evaluation of the global complexity.

The computational complexity can also be reduced bytaking advantage of the specific form of matrices(Dn)n>1.For example, if the subspace is chosen according to (23),

(∀n > 1) DV

n = [−V ∇Fn(hn),V hn,V hn − V hn−1].(36)

On the other hand, for everyn > 1,

V hn = V Dn−1un−1 = DV

n−1un−1, (37)

which shows that a recursive formula holds to compute thelast two components ofDV

n in (36). The initial complexityof 3NP multiplications is thus reduced toN(P +3). Similarrecursive procedures can be employed to compute(DV0

n )n>1

allowing the complexity to be reduced toN(N+3) from 3N2.In addition, we have, for everyn > 1,

DR

n = [−Rn∇Fn(hn),hR

n ,hR

n −Rnhn−1], (38)

where, by using (28),

hR

n = Rnhn = (1− 1

ϑn)Rn−1hn +

1

ϑnXnX

⊤nhn

= (1− 1

ϑn)DR

n−1un−1 +1

ϑnXnX

⊤nhn (39)

Rnhn−1 = (1− 1

ϑn)hR

n−1 +1

ϑnXnX

⊤nhn−1. (40)

It can be further observed that last term(ϑn)−1XnX⊤nhn−1

has already been computed in Step 3 of Algorithm 1. There-fore, instead of3N2 multiplications, we have now to perform

N(N + 2Q+ 4) ones. With these simplifications, in the casewhenV0 andV are null matrices, the global complexity ofthe algorithm is equal toN2(Q + 2)/2. WhenQ = 1, wethus recover the order of complexity of the classical RLSalgorithm. Since the objective function then reduces to aquadratic function, Sherman-Morrison-Woodbury formula canbe invoked to compute iteratively the minimizer on the wholespace in an efficient manner.

Note finally that the computation ofXnX⊤n with n ∈

N∗, which needs to be performed in Step 5, remains a

main source of complexity. However, if(∀n > Q) Xn =[xn−Q+1, . . . ,xn] wherexn ∈ R

N (as it is the case in affineprojection based algorithms for adaptive processing [51]), thena recursive computation ofXnX

⊤n only requiresxnx

⊤n to

be computed at each iterationn > Q. If we further assumethat the model is a one-dimensional convolutive one, i.e.xn

corresponds to shifted samples of a signal(x(n)

)n>1

, then(∀n > N) xn = [x(n−N +1), . . . , x(n)]⊤ andxnx

⊤n can be

itself computed recursively with a complexity ofN operations.Such ideas have been deeply investigated in the literature onfast RLS algorithms [52].

IV. CONVERGENCE STUDY

Throughout this section and the related appendices, it isassumed thatϑ = 1 and the underlying probability space isdenoted by(Ω,F,P). We will say in short that a property isP-a.s. satisfied if this property holds almost surely.

A. Assumptions

For everyn ∈ N∗, let Xn = σ

((Xk,yk)16k6n

)be the sub-

sigma algebra ofF generated by(Xk,yk)16k6n. In order togive a proof of convergence of the proposed stochastic MMsubspace algorithm, we will make the following additionalassumption:

Assumption 2.

(i) R+ V0 is a positive definite matrix.(ii)

((Xn,yn)

)n>1

is a stationary ergodic sequence and, foreveryn ∈ N

∗, the elements ofXn and the componentsof yn have finite fourth-order moments.

(iii) For everyn ∈ N∗,

E(‖yn+1‖2 |Xn) = (41)

E(Xn+1yn+1 |Xn) = r (42)

E(Xn+1X⊤n+1 |Xn) = R. (43)

(iv) For everyn ∈ N∗, ∇Fn(hn),hn ⊂ spanDn.

(v) h1 is X1-measurable and, for everyn ∈ N∗, Dn is

Xn-measurable.

The following asymptotic results will then be useful in therest of our developments.

Lemma 1. Under Assumptions 2(ii) and 2(iii), the followingproperties hold:

(i) (ρn)n>1, (Rn)n>1, and (rn)n>1 convergeP-a.s. to ,R and r, respectively

Page 6: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

6

(ii)+∞∑

n=1

n−1|ρn − | < +∞ P-a.s.

+∞∑

n=1

n−1‖rn − r‖ < +∞ P-a.s.

+∞∑

n=1

n−1|||Rn −R||| < +∞ P-a.s.,

where||| · ||| denotes the spectral matrix norm.

Proof: See Appendix A.

Remark 1.(i) Assumptions 2(ii) and 2(iii) are more general than

assuming that((Xn,yn)

)n>1

is an independent identi-cally distributed (i.i.d.) sequence and, for everyn ∈ N

∗,the elements ofXn and the components ofyn have finitefourth-order moments.

(ii) Assumptions 2(iv) and 2(v) are satisfied for the choiceof subspace given by(23) if h1 is X1-measurable (e.g.h1 is deterministic).

B. Almost sure convergence

Let us give the following preliminary property:

Lemma 2. Under Assumptions 1 and 2(ii)-2(iii),(hn)n>1 isP-a.s.bounded.2

Proof: See Appendix B.

Combining the previous lemma with classical results on theasymptotic behaviour of almost supermartingales, the conver-gence of the sequence

(Fn(hn)

)n>1

can be established:

Lemma 3. Under Assumptions 1 and 2,(Fn(hn)

)n>1

is P-a.s. convergent and((hn+1 − hn)

⊤An(hn)(hn+1 −hn))n>1

is P-a.s.summable.

Proof: See Appendix C.

Lemma 3 allows us to deduce the following result on thesequence of gradients computed at each iteration of the al-gorithm:

Lemma 4. Under Assumptions 1 and 2,(‖∇Fn(hn)‖)n>1 isP-a.s.square-summable.

Proof: See Appendix D.

By gathering all the previous results, our main convergenceresults can now be stated:

Proposition 2. Assume that Assumptions 1 and 2 hold. Then,the following hold:

(i) The set of cluster points of(hn)n>1 is almost surely anonempty compact connected set.

(ii) Any element of this set is almost surely a critical pointof F .

(iii) If the functions(ψs)16s6S are convex, then(hn)n>1

convergesP-a.s.to the unique (global) minimizer ofF .

2We say that a sequence of random vectors is almost surely bounded whenthe norms of all these vectors can be bounded by some random variable withprobability 1.

Proof: See Appendix E.

It can be noticed that the conclusion of Proposition 2(iii) is stillvalid if the functions(ψs)16s6S are nonconvex, they are twicecontinuously differentiable, and the regularization constants(λs)16s6S as defined in Table I are small enough so that thefunctionF is strongly convex.

V. A PPLICATION TO 2D FILTER IDENTIFICATION

A. Problem statement

We first demonstrate the efficiency of the proposed stochas-tic algorithm in a 2D filter identification problem. We considerthe following observation model:

y = S(h)x+w, (44)

wherex ∈ RL andy ∈ R

L represent the original and degradedversions of a given image,h ∈ R

N is the vectorized versionof an unknown two-dimensional blur kernel,S is the linearoperator which maps the kernel to its associated Hankel-blockHankel matrix form, andw ∈ R

L represents a realization ofan additive noise. When the imagesx andy are of very largesize, finding an estimateh ∈ R

N of the blur kernel can bequite memory consuming, but one can expect good estimationperformance by learning the blur kernel through a sweep ofblocks in the dataset.

Let us denote byX ∈ RL×N the matrix such thatS(h)x =

Xh. Then, we propose to defineh as a solution to (4), where,for everyn ∈ N

∗, yn ∈ RQ andX⊤

n ∈ RQ×N , are subparts

of y andX, respectively, corresponding toQ ∈ 1, . . . , Llines of this vector/matrix. For the regularization termΨ, weconsider, for everys ∈ 1, . . . , N (S = N ), an isotropicpenalization on the gradient between neighboring coefficients

of the blur kernel, i.e.,Ps = 2 andVs =[∆h

s ∆vs

]⊤, where

∆hs ∈ R

N (resp.∆vs ∈ R

N ) is the horizontal (resp. vertical)gradient operator applied at pixels. The smoothness ofh isthen enforced by choosing, for everys ∈ 1, . . . , S andu ∈R, ψs(u) = λ

√1 + u2/δ2 with (λ, δ) ∈]0,+∞[2. Finally, in

order to guarantee the existence of a unique minimizer, thestrong convexity ofF is imposed by takingv0 = 0 andV0 =τIN , whereτ is a small positive value (typicallyτ = 10−10).

B. Simulation results

The original image, presented in Figure 1(a), is a satelliteimage, of size4096 × 4096 pixels. The original blur kernelh with size 21 × 21, and the resulting blurred image, whichhas been corrupted with a zero-mean white Gaussan noisewith standard deviationσ = 0.03 (the blurred signal-to-noise ratio equals 25.7 dB), are displayed in Figures 1(b)(c).Figure 1(d) presents the estimated kernel, using Algorithm1,with the subspace given by (23), leading to the so-calledS3MG algorithm. Parameters(λ, δ) were adjusted so as tominimize the normalized root mean square estimation error,here equal to0.064. Figure 2 illustrates the variations of thisestimation error with respect to the computation time for theproposed algorithm, the SGD algorithm with a decreasingstepsize proportional ton−1/2, the regularized dual averaging

Page 7: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

7

(RDA) method with a constant stepsize from [40], and theaccelerated stochastic gradient averaging Finito method with aconstant stepsize from [53] when running tests on an Intel(R)Xeon(R) E5-2630 @ 2.6GHz using a Matlab 7 implemen-tation. Note that for the latter three algorithms, the stepsizeparameter was optimized manually so as to obtain the bestperformance in terms of convergence speed. Finally, note thatall tested algorithms were observed to provide asymptoticallythe same estimation quality, whatever the size of the blocks.In this example, as illustrated in Figure 3, the best trade-offin terms of convergence speed is obtained forQ = 256×256.

(a) (b)

(c) (d)

Fig. 1. (a) Original image. (b) Blurred and noisy image. (c) Original blurkernel. (d) Estimated blur kernel, with relative error0.064.

0 500 1000 1500 2000 2500 3000 3500

10−1

100

101

102

Time (s.)

Rel

ativ

e er

ror

Fig. 2. Comparison of S3MG algorithm (solid black line), SGD algorithmwith decreasing stepsize∝ n−1/2 (dashed-dotted red line), RDA algorithmwith constant stepsize (dashed blue line) and Finito algorithm with constantstepsize (turquoise thin line).

Time (s.)0 500 1000 1500 2000 2500 3000 3500

Rel

ativ

e er

ror

10-1

100

101

102

Q = 8x8Q = 16x16Q = 32x32Q = 64x64Q = 128x128Q = 256x256Q = 512x512Q = 1024x1024Q = 2048x2048

Fig. 3. Effect of the block sizeQ on the convergence speed of S3MG.

VI. A PPLICATION TO SPARSE ADAPTIVE FILTERING

A. Problem statement

As emphasized in Sections II and III, one of the advantagesof Algorithm 1 compared with some other online optimizationalgorithms is that it is able to deal with adaptive data process-ing problems. In this section, we apply the S3MG algorithmto the identification of a sparse time-varying system. Givenareal-valued discrete-time input signal

(x(n)

)n∈Z

, the outputof the system at timen > 1 is defined as

yn = X⊤n hn + wn, (45)

whereXn = [x(n − N + 1), . . . , x(n)]⊤, wn models somemeasurement noise, andhn ∈ R

N gathers the unknown filtertaps at timen. Then, the objective is to provide an estimateof the vectorhn at each time by solving Problem (4) wherethe regularization functionΨ is chosen in order to promotethe sparsity of the impulse response of the time-varying filter.

B. Simulation results

We generate data according to Model (45) where the inputsignal

(x(n)

)n∈Z

consists of identically and independentrandom binary values−1,+1. The measurement noise(wn)n∈Z is white Gaussian with zero mean and variance0.05.In order to evaluate the tracking capability of the proposedS3MG method, the following time-varying linear system isconsidered:

hn =

h1 if n 6 L/2,

hL/2+1 if n > L/2 + 1.(46)

The filter lengthN is equal to 200 and the output of the systemis observed at every timen ∈ 1, . . . , L with L = 5000.The sparse impulse responses corresponding to vectorsh1 andhL/2+1 are represented in Figure 4.

We compute, for everyn ∈ 1, . . . , L, the Euclidean normof the error between the current estimatehn and the true filtercoefficient vectorhn. The minimal estimation error is obtainedfor the nonconvex Welsch penalty function (see Table I) anda smoothedℓ2 − ℓ0 regularization function is thus employedby settingS = N , v0 = 0, V0 = ON , and, for everys ∈

Page 8: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

8

1, . . . , N, Ps = 1, vs = 0, while Vs ∈ R1×N is the s-th

vector of the canonical basis ofRN .We present the results generated by S3MG in Figure 5 for

two values of the forgetting factorϑ, namelyϑ = 1 whichcorresponds to a non adaptive strategy, andϑ = 0.995 whichappears to be the best choice in terms of tracking propertiesfor this example.

We also show the results obtained with several state-of-the-art approaches in the context of sparse adaptive filtering,namely SPAL [34], RLMS [54], RZAAPA [32] and SM-PAPA [55]. Note that, for each tested method, the involvedparameters (stepsize, regularization weight, blocksize)havebeen tuned manually in order to optimize the performancein terms of error decay.

0 20 40 60 80 100 120 140 160 180 200−1

−0.5

0

0.5

1

1.5

0 20 40 60 80 100 120 140 160 180 200−1

−0.5

0

0.5

1

1.5

Fig. 4. Values of the coefficients of the considered sparse filtersh1 (top)andhL/2+1 (bottom).

0 500 1000 1500 2000 2500 3000 3500 4000 4500 5000−50

−45

−40

−35

−30

−25

−20

−15

−10

n

Err

or in

dB

S3MG (ϑ = 0.995)S3MG (ϑ = 1)SPALRLMSRZAAPASM−PAPA

Fig. 5. Quadratic estimation error on the filter coefficients as a function oftime indexn for various adaptive algorithms.

VII. C ONCLUSION

In this work, we have proposed a stochastic MM subspacealgorithm for online penalized least squares estimation prob-

lems. The method makes it possible to use large-size datasetsthe second-order moments of which are not known a priori.We have shown that the proposed algorithm is of the sameorder of complexity as the classical RLS algorithm and thatits computational cost can be reduced by taking advantage ofspecific forms of the search subspace. The choice of a memorygradient subspace led to the S3MG algorithm whose goodnumerical performance has been demonstrated in the contextof 2D filter identification for large scale image processingproblems. In the context of sparse adaptive filtering, S3MGhas also been shown to be competitive with respect to recentmethods. Although an analysis of the convergence of theproposed method has been carried out, it would be interestingto extend the obtained results to weaker assumptions. Inaddition, in a nonstationary context, a theoretical study ofthe tracking abilities of the algorithm should be conducted.Finally, a detailed analysis of the convergence rate of theproposed method will be undertaken in a forthcoming paper.

APPENDIX APROOF OFLEMMA 1

Property (i) is a consequence of the ergodic theorem [56,Theorem 13.12]. In addition, the law of the iterated logarithmfor martingale difference sequences [57] ensures that

lim supn→+∞

|∑nk=1(‖yk‖2 − )|

(n log(log n)

)1/2 < +∞ P-a.s. (47)

lim supn→+∞

‖∑nk=1(Xkyk − r)‖

(n log(log n)

)1/2 < +∞ P-a.s. (48)

lim supn→+∞

|||∑nk=1(XkX

⊤k −R)|||

(n log(log n)

)1/2 < +∞ P-a.s. (49)

that is

lim supn→+∞

n1/2|ρn − |(log(log n)

)1/2 < +∞ P-a.s. (50)

lim supn→+∞

n1/2‖rn − r‖(log(log n)

)1/2 < +∞ P-a.s. (51)

lim supn→+∞

n1/2|||Rn −R|||(log(log n)

)1/2 < +∞ P-a.s. (52)

Consequently, for everyn0 ∈ N with n0 > 2,

+∞∑

n=n0

n−1|ρn − |

6 supn>n0

(n1/2|ρn − |(log(log n)

)1/2

)( +∞∑

n=n0

n−3/2| log(log n)|1/2).

(53)

Since∑+∞

n=2 n−3/2| log(log n)|1/2 < +∞, it follows from

(50) that∑+∞

n=n0n−1|ρn − | convergesP-a.s. to 0 asn0 →

+∞, which means that the first line in Property (ii) is satisfied.By proceeding similarly, (51) and (52) allow us to establishthe remaining two assertions in Property (ii).

Page 9: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

9

APPENDIX BPROOF OFLEMMA 2

For everyn ∈ N∗, minimizing Θn(·,hn) is equivalent to

minimizing the function

(∀h ∈ RN ) Θn(h,hn) =

1

2h⊤An(hn)h− cn(hn)

⊤h.

(54)It follows from Assumption 2(ii)-2(iii) and Lemma 1(i) thatthere existsΛ ∈ F such thatP(Λ) = 1 and, for everyω ∈ Λ,

limn→+∞

rn(ω) = r (55)

limn→+∞

Rn(ω) = R. (56)

Let ω ∈ Λ. According to Assumption 1(iii) and Eq. (19),b(h)is bounded as a function ofh. It is then deduced from (16)and (55) that

(cn(hn)(ω)

)n>1

is bounded, i.e. there existsη ∈ [0,+∞[ such that

(∀n ∈ N∗) ‖cn(hn)(ω)‖ 6 η. (57)

In addition, as a consequence of (19) and Assumption 1(iii),for every n ∈ N

∗, Diag(b(hn)

)is a positive semidefinite

matrix. Hence, because of (15), Assumptions 1(iii) and 2(i),and (56), there existsǫ ∈]0,+∞[ andn0 ∈ N

∗ such that

(∀n > n0) An(hn)(ω) R− ǫIN + V0 ≻ ON . (58)

(It suffices to chooseǫ lower than the minimum eigenvalueof R + V0). As a consequence of (54), (57), (58), and theCauchy-Schwarz inequality, we have

(∀n > n0)(∀h ∈ RN )

1

2h⊤(R− ǫIN + V0)h− η‖h‖ 6 Θn(h,hn). (59)

SinceR − ǫIN + V0 is a positive definite matrix, the lowerbound corresponds to a coercive function with respect toh.There thus existsζ ∈]0,+∞[ such that, for everyh ∈ R

N ,

‖h‖ > ζ ⇒ (∀n > n0) Θn(h,hn)(ω) > 0. (60)

On the other hand, since0 ∈ span(Dn(ω)

), we have

Θn(hn+1,hn)(ω) 6 Θn(0,hn)(ω) = 0. (61)

The last two inequalities allow us to conclude that

(∀n > n0) ‖hn+1(ω)‖ 6 ζ. (62)

APPENDIX CPROOF OFLEMMA 3

According to Assumption 2(iv), the proposed algorithm isactually equivalent to

(∀n ∈ N∗) hn+1 = hn +Dnun (63)

un = argminu∈RM

Θn(hn +Dnu,hn). (64)

By using (20) and cancelling the derivative of the functionu 7→ Θn(hn +Dnu,hn),

D⊤n∇Fn(hn) +D⊤

nAn(hn)Dnun = 0. (65)

Hence,

Θ(hn+1,hn)

= Fn(hn)−1

2u⊤nD

⊤nAn(hn)Dnun

= Fn(hn)−1

2(hn+1 − hn)

⊤An(hn)(hn+1 − hn). (66)

In view of (12) and Proposition 1, this yields

(∀n ∈ N∗) Fn(hn+1)+

1

2(hn+1−hn)

⊤An(hn)(hn+1−hn)

6 Fn(hn). (67)

In addition, the following recursive relation holds

(∀h ∈ RN ) Fn+1(h) = Fn(h) +

1

2(ρn+1 − ρn)

− (rn+1 − rn)⊤h+

1

2h⊤(Rn+1 −Rn)h.

(68)

As a consequence of Assumption 2(v), for everyn ∈ N∗, hn+1

is Xn-measurable. It can thus be deduced from (67) and theprevious two relations that

E(Fn+1(hn+1) |Xn)+1

2(hn+1−hn)

⊤An(hn)(hn+1−hn)

6 Fn(hn) + χn (69)

where

χn =1

2E(ρn − ρn+1 |Xn)− E(rn − rn+1 |Xn)

⊤hn+1

+1

2h⊤n+1E(Rn −Rn+1 |Xn)hn+1. (70)

By using (8)-(10) withϑ = 1 and Assumption 2(iii), we have

χn =1

2(n+ 1)

(ρn − E(‖yn+1‖2 |Xn)

)

− 1

n+ 1

(rn − E(Xn+1yn+1 |Xn)

)⊤hn+1

+1

2(n+ 1)h⊤n+1

(Rn − E(Xn+1X

⊤n+1 |Xn)

)hn+1

=1

2(n+ 1)

(ρn −

)− 1

n+ 1

(rn − r

)⊤hn+1

+1

2(n+ 1)h⊤n+1

(Rn −R

)hn+1 (71)

which yields

|χn| 61

2(n+ 1)|ρn − |+ 1

n+ 1‖rn − r‖‖hn+1‖

+1

2(n+ 1)|||Rn −R||| ‖hn+1‖2. (72)

According to Lemma 2,(hn)n>1 is P-a.s. bounded, andAssumptions 2(ii)-2(iii) and Lemma 1(ii) thus guarantee that

+∞∑

n=1

|χn| < +∞ P-a.s. (73)

Page 10: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

10

Assumption 1(i) entails that, for everyn ∈ N∗, Fn is lower

bounded byinf Ψ > −∞. Furthermore, (69) leads to

E(Fn+1(hn+1)− inf Ψ |Xn)

+1

2(hn+1 − hn)

⊤An(hn)(hn+1 − hn)

6 Fn(hn)− inf Ψ + |χn|. (74)

Since, for everyn ∈ N∗, Fn(hn) − inf Ψ and (hn+1 −

hn)⊤An(hn)(hn+1 − hn) are nonnegative,(Fn(hn) −

inf Ψ)n>1 is a nonnegative almost supermartingale [58]. Byinvoking now Siegmund-Robbins lemma [59], it can be de-duced from (73) that the desired convergence results hold.

APPENDIX DPROOF OFLEMMA 4

According to (20), we have, for everyφ ∈ R andn ∈ N∗,

Θn

(hn − φ∇Fn(hn),hn

)= Fn(hn)− φ‖∇Fn(hn)‖2

+φ2

2

(∇Fn(hn)

)⊤An(hn)∇Fn(hn). (75)

Let

Φn ∈ Argminφ∈R

Θn

(hn − φ∇Fn(hn),hn

). (76)

The following optimality condition holds:

(∇Fn(hn)

)⊤An(hn)∇Fn(hn) Φn = ‖∇Fn(hn)‖2. (77)

As a consequence of Assumption 2(iv),(∀φ ∈ R) hn −φ∇Fn(hn) ∈ spanDn. It then follows from (21) and (77)that

Θn

(hn+1,hn

)6 Θn

(hn − Φn∇Fn(hn),hn

)

6 Fn(hn)−Φn

2‖∇Fn(hn)‖2 (78)

which, by using (66), leads to

Φn‖∇Fn(hn)‖2 6 (hn+1−hn)⊤An(hn)(hn+1−hn). (79)

Let ǫ > 0. Assumption 1(iii) and (15) yield, for everyn ∈N

∗,

An(hn) (|||Rn + V0|||+ ν|||V |||2)IN . (80)

Therefore, according to Assumptions 2(i) and 2(ii), andLemma 1(i), there existsΛ ∈ F such thatP(Λ) = 1 and,for everyω ∈ Λ,

(∃n0 ∈ N∗)(∀n > n0) ON ≺ An(hn)(ω) α−1

ǫ IN (81)

where

αǫ = (|||R+ V0|||+ ν|||V |||2 + ǫ)−1 > 0. (82)

Let ω ∈ Λ. By using now (77), it can be deduced from (81)that, if n > n0 and∇Fn(hn)(ω) 6= 0, then

Φn(ω) > αǫ. (83)

Then, it follows from (79) that

αǫ

+∞∑

n=n0

‖∇Fn(hn)(ω)‖2

6

+∞∑

n=n0

(hn+1(ω)− hn(ω)

)⊤An(hn)(ω)

(hn+1(ω)− hn(ω)

).

(84)

By invoking Lemma 3, we can conclude that(‖∇Fn(hn)‖2)n>1 is P-a.s. summable.

APPENDIX EPROOF OFPROPOSITION2

It follows from Lemma 3 that((hn+1 −

hn)⊤An(hn)(hn+1 − hn)

)n>1

converges P-a.s. to 0.In addition, we have seen in the proof of Lemma 2 that thereexistsΛ ∈ F such thatP(Λ) = 1 and, for everyω ∈ Λ, (58)holds with ǫ ∈]0,+∞[ and n0 ∈ N

∗. This implies that, foreveryn > n0,

|||R− ǫIN + V0||| ‖hn+1(ω)− hn(ω)‖2

6(hn+1(ω)− hn(ω)

)⊤An(hn)(ω)

(hn+1(ω)− hn(ω)

)

(85)

where|||R−ǫIN +V0||| > 0. Consequently,(hn+1−hn)n>1

convergesP-a.s. to0. In addition, according to Lemma 2,(hn)n>1 belongs almost surely to a compact set. The result isthen obtained by invoking Ostrowski’s theorem [60, Theorem26.1].(ii) By using (14)-(16), we have

(∀n ∈ N∗) ∇Fn(hn)−∇F (hn) = (Rn −R)hn − rn + r.

(86)Since (hn)n>1 is almost surely bounded, it follows fromLemma 1(i) that

(∇Fn(hn)−∇F (hn)

)n>1

convergesP-a.s.to 0. Since Lemma 4 ensures that

(∇Fn(hn)

)n>1

convergesP-a.s. to0,

(∇F (hn)

)n>1

also convergesP-a.s. to0. Therethus existsΛ ∈ F such thatP(Λ) = 1 and, for everyω ∈ Λ, ∇F

(hn(ω)

)→ 0. Let h be a cluster point of(

hn(ω))n>1

. There exists a subsequence(hkn

(ω))n>1

such

that hkn(ω) → h. As we have assumed that the regulariza-

tion functions(ψs)16s6S are continuously differentiable (seeAssumption 1(i)),F is also continuously differentiable, and

∇F (h) = limn→+∞

∇F(hkn

(ω))= 0. (87)

This means thath is a critical point ofF .(iii) Because of Assumption 2(i), when the functions(ψs)16s6S are convex,F is a strongly convex function. Itthus possesses a unique critical pointh, which is the globalminimizer ofF . It follows from (i) and (ii) that, almost surely,the unique cluster point of(hn)n>1 is h, which shows thathn → h P-a.s.

Page 11: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

11

ACKNOWLEDGEMENTS

The authors would like to thank Professors Markus V. S.Lima and Paulo S. R. Diniz from Federal University of Riode Janeiro for kindly making us accessible some of their codes.We would also like to thank Professor Anisia Florescu fromDunarea de Jos University of Galati for providing the initialmotivation for this work.

REFERENCES

[1] E. Chouzenoux, J.-C. Pesquet, and A. Florescu, “A stochastic 3MGalgorithm with application to 2D filter identification,” inProc. 22ndEuropean Signal Process. Conf. (EUSIPCO 2014), Lisboa, Portugal, 1-5 Sep. 2014, pp. 1587–1591.

[2] J. Nocedal and S. J. Wright,Numerical Optimization, Springer-Verlag,New York, 1999.

[3] P. L. Combettes and J.-C. Pesquet, “Proximal splitting methods insignal processing,” inFixed-Point Algorithms for Inverse Problems inScience and Engineering, H. H. Bauschke, R. Burachik, P. L. Combettes,V. Elser, D. R. Luke, and H. Wolkowicz, Eds., pp. 185–212. Springer-Verlag, New York, 2010.

[4] M. V. Afonso, J. M. Bioucas-Dias, and M. A. T. Figueiredo,“An aug-mented Lagrangian aproach to the constrained optimization formulationof imaging inverse problems,”IEEE Trans. Image Process., vol. 20, no.3, pp. 681–695, 2011.

[5] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein, “Distributedoptimization and statistical learning via the alternating direction methodof multipliers,” Found. Trends Machine Learn., vol. 8, no. 1, pp. 1–122,2011.

[6] D. R. Hunter and K. Lange, “A tutorial on MM algorithms,”Amer. Stat,vol. 58, no. 1, pp. 30–37, Feb. 2004.

[7] Z. Zhang, J. T. Kwok, and D.-Y. Yeung, “Surrogate maximiza-tion/minimization algorithms and extensions,”Mach. Learn., vol. 69,pp. 1–33, 2007.

[8] M. Zibulevsky and M. Elad, “ℓ2 − ℓ1 optimization in signal and imageprocessing,”IEEE Signal Process. Mag., vol. 27, pp. 76–88, May 2010.

[9] E. Chouzenoux, J. Idier, and S. Moussaoui, “A majorize-minimize sub-space strategy for subspace optimization applied to image restoration,”IEEE Trans. Image Process., vol. 20, no. 18, pp. 1517–1528, Jun. 2011.

[10] E. Chouzenoux, A. Jezierska, J.-C. Pesquet, and H. Talbot, “A majorize-minimize subspace approach forℓ2-ℓ0 image regularization,”SIAM J.Imag. Sci., vol. 6, no. 1, pp. 563–591, 2013.

[11] A. Florescu, E. Chouzenoux, J.-C. Pesquet, P. Ciuciu, and S. Ciochina,“A majorize-minimize memory gradient method for complex-valuedinverse problem,” Signal Process., vol. 103, pp. 285–295, Oct. 2014,Special issue on Image Restoration and Enhancement: Recent Advancesand Applications.

[12] S. O. Haykin,Adaptive Filter Theory, Prentice Hall, New Jersey, USA,4th edition, 2002.

[13] H. Robbins and S. Monro, “A stochastic approximation method,” Ann.Math. Stat., vol. 22, no. 3, pp. 400–407, 1951.

[14] Ju. M. Ermoliev and Z. V. Nekrylova, “The method of stochasticgradients and its application,” inSeminar: Theory of Optimal Solutions.No. 1 (Russian), pp. 24–47. Akad. Nauk Ukrain. SSR, Kiev, 1967.

[15] O. V. Guseva, “The rate of convergence of the method of generalizedstochastic gradients,”Kibernetika (Kiev), , no. 4, pp. 143–145, 1971.

[16] D. P. Bertsekas and J. N. Tsitsiklis, “Gradient convergence in gradientmethods with errors,” SIAM J. Optim., vol. 10, no. 3, pp. 627–642,2000.

[17] Y. F. Atchade, G. Fort, and E. Moulines, “On stochastic proximalgradient algorithms,” 2014, http://arxiv.org/abs/1402.2365.

[18] L. Rosasco, S. Villa, and B. C. Vu, “Convergence of stochastic proximalgradient algorithm,” 2014, http://arxiv.org/abs/1403.5074.

[19] P. L. Combettes and J.-C. Pesquet, “Stochastic approximations and per-turbations in forward-backward splitting for monotone operators,” PureAppl. Funct. Anal., 2016, to appear, http://arxiv.org/abs/1507.07095.

[20] J. Konecny, J. Liu, P. Richtarik, and M. Takac, “mS2GD: Mini-batch semi-stochastic gradient descent in the proximal setting,” 2014,http://arxiv.org/pdf/1410.4744v1.pdf.

[21] B. T. Polyak and A. B. Juditsky, “Acceleration of stochastic approxima-tion by averaging,”SIAM J. Control Optim., vol. 30, no. 4, pp. 838–855,1992.

[22] A. Nemirovski, A. Juditsky, G. Lan, and A. Shapiro, “Robust stochasticapproximation approach to stochastic programming,”SIAM J. Optim.,vol. 19, no. 4, pp. 1574–1609, 2008.

[23] F. Bach and E. Moulines, “Non-asymptotic analysis of stochasticapproximation algorithms for machine learning,” inProc. Ann. Conf.Neur. Inform. Proc. Syst., Granada, Spain, Dec. 12 - 17 2011, pp. x–x+8.

[24] M. Pereyra, P. Schniter, E. Chouzenoux, J.-C. Pesquet,J.-Y. Tourneret,A. O Hero, and S. McLaughlin, “A survey of stochastic simula-tion and optimization methods in signal processing,”IEEE Jour-nal of Selected Topics in Signal Processing, 2016, to appear,http://arxiv.org/abs/1505.00273.

[25] B. Widrow and S.D. Stearns,Adaptive Signal Processing, Prentice Hall,New Jersey, 1985.

[26] O. Macchi, Adaptive Processing: The Least Mean Squares ApproachWith Applications in Transmission, Wiley, Chichester, UK, 1995.

[27] O. Hoshuyama, R. A. Goubran, and A. Sugiyama, “A generalizedproportionate variable step-size algorithm for fast changing acousticenvironments,” inProc. Int. Conf. Acoust., Speech Signal Process.(ICASSP 2004), Montreal, Canada, May 17-21 2004, vol. 4, pp. 161–164.

[28] A. W. H. Khong and P. A Naylor, “Efficient use of sparse adaptivefilters,” in Proc. Asilomar Conf. Signal, Systems and Computers, PacificGrove, CA, Oct. 29-Nov. 1 2006, pp. 1375–1379.

[29] Y. Chen, Y. Gu, and A. O. Hero, “Sparse LMS for system identification,”in Proc. Int. Conf. Acoust., Speech Signal Process. (ICASSP 2009),Taipei, Taiwan, Apr. 19-24 2009, pp. 3125–3128.

[30] C. Paleologu, J. Benesty, and S. Ciochina, Sparse adaptive filters forecho cancellation, Synthesis Lectures on Speech and Audio Processing.Morgan and Claypool, San Rafael, USA, 2010.

[31] Y. Murakami, M. Yamagishi, M. Yukawa, and I. Yamada, “A sparseadaptive filtering using time-varying soft-thresholding techniques,” inProc. Int. Conf. Acoust., Speech Signal Process. (ICASSP 2010), Dallas,Texas, Mar. 14-19 2010, pp. 3734–3737.

[32] R. Meng, R. C. De Lamare, and V. H. Nascimento, “Sparsity-awareaffine projection adaptive algorithms for system identification,” in Proc.Sensor Signal Process. Defence, London, U.K., Sept. 27-29 2011, pp.1–5.

[33] L. V. S. Markus, W. A. Martins, and P. S. R. Diniz, “Affine projectionalgorithms for sparse system identification,” inProc. Int. Conf. Acoust.,Speech Signal Process. (ICASSP 2013), Vancouver, Canada, May 26-312013, pp. 5666–5670.

[34] Y. Kopsinis, K. Slavakis, and S. Theodoridis, “Online sparse systemidentification and signal reconstruction using projections onto weightedℓ1 balls,” IEEE Trans. Signal Process., vol. 59, no. 3, pp. 936–952,Mar. 2011.

[35] K. Slavakis, Y. Kopsinis, S. Theodoridis, and S. McLaughlin, “Gen-eralized thresholding and online sparsity-aware learningin a union ofsubspaces,”IEEE Trans. Signal Process., vol. 61, no. 15, pp. 3760–3773,Aug. 2013.

[36] B. Babadi, N. Kalouptsidis, and V. Tarokh, “SPARLS: Thesparse RLSalgorithm,” IEEE Trans. Signal Process., vol. 58, no. 8, pp. 4013–4025,Aug. 2010.

[37] D. Angelosante, J. A. Bazerque, and G. B. Giannakis, “Online adaptiveestimation of sparse signals: where RLS meets theℓ1 -norm,” IEEETrans. Signal Process., vol. 58, no. 7, pp. 3436–3447, Jul. 2010.

[38] K. E. Themelis, A. A. Rontogiannis, and K. D. Koutroumbas,“Avariational bayes framework for sparse adaptive estimation,” IEEETrans. Signal Process., vol. 62, no. 18, pp. 4723–4736, 2014.

[39] S. Ono, M. Yamagishi, and I. Yamada, “A sparse system identificationby using adaptively-weighted total variation via a primal-dual splittingapproach,” inProc. Int. Conf. Acoust., Speech Signal Process. (ICASSP2013), Vancouver, Canada, 26-31 May 2013, pp. 6029–6033.

[40] L. Xiao, “Dual averaging methods for regularized stochastic learningand online optimization,”J. Mach. Learn. Res., vol. 11, pp. 2543–2596,Oct. 2010.

[41] J. Mairal, “Stochastic Majorization-Minimization algorithms for large-scale optimization,” inProc. Adv. Conf. Neur. Inform. Proc. Syst., LakeTahoe, Nevada, Dec. 5-8 2013, pp. x–x+8.

[42] D. Carlson, Y.-P. Hsieh, E. Collins, L. Carin, and V. Cevher, “Stochas-tic spectral descent for discrete graphical models,”IEEE Jour-nal of Selected Topics in Signal Processing, 2016, to appear,http://people.ee.duke.edu/˜lcarin/SSDfor belief nets.pdf.

[43] J. R. Birge, X. Chen, L. Qi, and Z. Wei, “A stochastic Newton methodfor stochastic quadratic programs with recourse,” 1995, technical report,http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.49.4279.

Page 12: A Stochastic Majorize-Minimize Subspace Algorithm for Online … · 2016-01-05 · the classical Recursive Least Squares (RLS) algorithm can be used for this purpose [12]. When Ψ

12

[44] A. Bordes, L. Bottou, and P. Gallinari, “SGD-QN: Careful quasi-Newtonstochastic gradient descent,”J. Mach. Learn. Res., vol. 10, pp. 1737–1754, Jul. 2009.

[45] J. Yu, S. V. N. Vishwanathan, S. Gunter, and N. N. Schraudolph, “Aquasi-Newton approach to nonsmooth convex optimization problems inmachine learning,”J. Mach. Learn. Res., vol. 11, pp. 1145–1200, Mar.2010.

[46] R. H. Byrd, S. L. Hansen, J. Nocedal, and Y. Singer, “A stochasticquasi-Newton method for large-scale optimization,” 2014, technicalreport, http://arxiv.org/abs/1401.7020.

[47] H. Zou and T. Hastie, “Regularization and variable selection via theelastic net,”J. R. Statist. Soc. B, vol. 67, no. 2, pp. 301–320, 2005.

[48] M. Allain, J. Idier, and Y. Goussard, “On global and local convergenceof half-quadratic algorithms,”IEEE Trans. Image Process., vol. 15, no.5, pp. 1130–1142, May 2006.

[49] J. M. Bioucas-Dias and M. A. T. Figueiredo, “A new TwIST:two-stepiterative shrinkage/thresholding algorithms for image restoration,” IEEETrans. Image Process., vol. 16, no. 12, pp. 2992–3004, Dec. 2007.

[50] A. Beck and M. Teboulle, “A fast iterative shrinkage-thresholdingalgorithm for linear inverse problems,”SIAM J. Imaging Sci., vol. 2,no. 1, pp. 183–202, 2009.

[51] S. L. Gay and S. Tavathia, “The fast affine projection algorithm,” inProc. Int. Conf. Acoust., Speech Signal Process., Detroit, MI, May 9-121995, vol. 5, pp. 3023–3026.

[52] D. G. Manolakis, V. K. Ingle, and S. M. Kogon,Statistical and AdaptiveSignal Processing, Artech House, Norwood, MA, 2005.

[53] A. Defazio, J. Domke, and T. Cartano, “Finito: A faster, permutableincremental gradient method for big data problems,” inProc. 31st Int.Conf. Mach. Learn., Beijing, China, Jun. 21-26 2014, pp. 1125–1133.

[54] Y. Chen, Y. Gu, and A. O. Hero, “Regularized least-mean-squarealgorithms,” pp. x–x+7, 2010, http://arxiv.org/pdf/1012.5066.pdf.

[55] S. Werner, J.A. Apolinario, and P. S. R. Diniz, “Set-membershipproportionate affine projection algorithms,”EURASIP Journal on Audio,Speech, and Music Processing, vol. 2007, no. 1, pp. 034242, 2007.

[56] J. Davidson, Stochastic Limit Theory, Oxford University Press, NewYork, 1994.

[57] W. F. Stout, “The Hartman-Wintner law of the iterated logarithm formartingales,”Ann. Math. Stat., vol. 41, no. 6, pp. 2158–2160, 1970.

[58] T. L. Lai, “Martingales in sequential analysis and time series, 1945-1985,” J. Electron. Hist. Probab. Stat., vol. 5, no. 1, pp. x–x+30, June2009.

[59] H. Robbins and D. Siegmund, “A convergence theorem for non negativealmost supermartingales and some applications,” inOptimizing Methodsin Statistics, J. S. Rustagi, Ed., pp. 233–257. Academic Press, New York,1971.

[60] A. M. Ostrowski, Solution of Equations in Euclidean and BanachSpaces, Academic Press, London, 1973.