Skip to main content
Springer logoLink to Springer
. 2018 Mar 6;70(3):841–863. doi: 10.1007/s10589-018-9989-y

A convergent relaxation of the Douglas–Rachford algorithm

Nguyen Hieu Thao 1,2,
PMCID: PMC6445491  PMID: 31007390

Abstract

This paper proposes an algorithm for solving structured optimization problems, which covers both the backward–backward and the Douglas–Rachford algorithms as special cases, and analyzes its convergence. The set of fixed points of the corresponding operator is characterized in several cases. Convergence criteria of the algorithm in terms of general fixed point iterations are established. When applied to nonconvex feasibility including potentially inconsistent problems, we prove local linear convergence results under mild assumptions on regularity of individual sets and of the collection of sets. In this special case, we refine known linear convergence criteria for the Douglas–Rachford (DR) algorithm. As a consequence, for feasibility problem with one of the sets being affine, we establish criteria for linear and sublinear convergence of convex combinations of the alternating projection and the DR methods. These results seem to be new. We also demonstrate the seemingly improved numerical performance of this algorithm compared to the RAAR algorithm for both consistent and inconsistent sparse feasibility problems.

Keywords: Almost averagedness, Picard iteration, Alternating projection method, Douglas–Rachford method, RAAR algorithm, Krasnoselski–Mann relaxation, Metric subregularity, Transversality, Collection of sets

Introduction

Convergence analysis has been one of the central and very active applications of variational analysis and mathematical optimization. Examples of recent contributions to the theory of the field that have initiated efficient programs of analysis are [1, 2, 7, 38]. It is the common recipe emphasized in these and many other works that there are two key ingredients required in order to derive convergence of a numerical method (1) regularity of individual functions or sets such as convexity and averaging property, and (2) regularity of collections of functions or sets at their critical points such as transversality, Kurdyka-Łojasiewicz property and metric subregularity. As a result, the question about convergence of a solving method can often be reduced to checking whether certain regularity properties of the problem data are satisfied. There have been a considerable number of papers studying these two ingredients of convergence analysis in order to establish sharper convergence criteria in various circumstances, especially those applicable to algorithms for solving nonconvex problems [5, 12, 13, 19, 26, 27, 3133, 38, 42, 45].

This paper suggests an algorithm called Tλ, which covers both the backward-backward and the DR algorithms as special cases of choosing the parameter λ[0,1], and analyzes its convergence. When applied to feasibility problem for two sets one of which is affine, Tλ is a convex combination of the alternating projection and the DR methods. On the other hand, Tλ can be viewed as a relaxation of the DR algorithm. Motivation for relaxing the DR algorithm comes from the lack of stability of this algorithm when applied to inconsistent problems. This phenomenon has been observed for the Fourier phase retrieval problem which is essentially inconsistent due to the reciprocal relationship between the spatial and frequency variables of the Fourier transform [35, 36]. To address this issue, a relaxation of the DR algorithm, often known as the RAAR algorithm, was proposed and applied to phase retrieval problems by Luke in the aforementioned papers. In the framework of feasibility, the RAAR algorithm is described as a convex combination of the basic DR operator and one of the projectors. Our preliminary numerical experiments have revealed a promising performance of algorithm Tλ in comparison with the RAAR method. This observation has motivated the study of convergence analysis of algorithm Tλ in this paper.

After introducing the notation and proving preliminary results in Sect. 2, we introduce Tλ as a general fixed point operator, characterize the set of fixed points of Tλ (Proposition 1), and establish abstract convergence criteria for iterations generated by Tλ (Theorem 2) in Sect. 3. We discuss algorithm Tλ in the framework of feasibility problems in Sect. 4. The set of fixed points of Tλ is characterized for convex inconsistent feasibility (Proposition 3). For consistent feasibility we show that almost averagedness of Tλ (Proposition 4) and metric subregularity of Tλ-Id (Lemma 3) can be obtained from regular properties of the individual sets and of the collection of sets, respectively. As a result, the two regularity notions are combined to yield local linear convergence of iterations generated by Tλ (Theorem 4). Section 5 is devoted to demonstrate the improved numerical performance of algorithm Tλ compared to the RAAR algorithm for both consistent and inconsistent feasibility problems. In this section, we study the feasibility approach for solving the sparse optimization problem. Our linear convergence result established in Sect. 4 for iterations generated by Tλ is also illustrated in this application (Theorem 5).

Notation and preliminary results

Our notation is standard, c.f. [11, 40, 46]. The setting throughout this paper is a finite dimensional Euclidean space E. The norm · denotes the Euclidean norm. The open unit ball in a Euclidean space is denoted B, and Bδ(x) stands for the open ball with radius δ>0 and center x. The distance to a set AE with respect to the bivariate function dist(·,·) is defined by

dist(·,A):ER+:xinfyAdist(x,y).

We use the convention that the distance to the empty set is +. The set-valued mapping

PA:EE:xyAdist(x,y)=dist(x,A)

is the projector on A. An element yPA(x) is called a projection. This exists for any closed set AE. Note that the projector is not, in general, single-valued. Closely related to the projector is the prox mapping corresponding to a function f and a stepsize τ>0 [41]

proxτ,f(x):=argminyEf(y)+12τy-x2.

When f=ιA is the indicator function of A, that is ιA(x)=0 if xA and ιA(x)=+ otherwise, then proxτ,ιA=PA for all τ>0. The inverse of the projector, PA-1, is defined by

PA-1(a):=xEaPA(x).

The proximal normal cone to A at x¯ is the set, which need not be either closed or convex,

NAprox(x¯):=conePA-1(x¯)-x¯. 1

If x¯A, then NAprox(x¯) is defined to be empty. Normal cones are central to characterizations both of the regularity of individual sets and of the regularity of collections of sets. For a refined numerical analysis of projection methods, one also defines the Λ-proximal normal cone to A at x¯ by

NA|Λprox(x¯):=conePA-1(x¯)Λ-x¯.

When Λ=E, it coincides with the proximal normal cone (1).

For ε0 and δ>0, a set A is (ε,δ)-regular relative to Λ at x¯A [13, Definition 2.9] if for all xBδ(x¯), aABδ(x¯) and vNA|Λprox(a),

x-a,vεx-av.

When Λ=E, the quantifier “relative to” is dropped.

For a set-valued operator T:EE, its fixed point set is defined by FixT:=xExTx. For a number λ[0,1], we denote the λ-reflector of T by RT,λ:=(1+λ)T-λId. A frequently used example in this paper corresponds to T being a projector.

In the context of convergence analysis of Picard iterations, the following generalization of the Fejér monotonicity of sequences appears frequently, see, for example, the book [4] or the paper [39] for the terminology.

Definition 1

(Linear monotonicity) The sequence (xk) is linearly monotone with respect to a set SE with rate c[0,1] if

dist(xk+1,S)cdist(xk,S)kN.

Our analysis follows the abstract analysis program proposed in [38] which requires the two key components of the convergence: almost averagedness and metric subregularity.

Definition 2

(Almost nonexpansive/averaging mappings) [38] Let T:EE and UE.

  • (i)
    T is pointwise almost nonexpansive at y on U with violation ε0 if for all xU, x+Tx and y+Ty,
    x+-y+1+εx-y.
  • (ii)
    T is pointwise almost averaging at y on U with violation ε0 and averaging constant α>0 if for all xU, x+Tx and y+Ty,
    x+-y+21+εx-y2-1-αα(x+-x)-(y+-y)2. 2

When a property holds at all yU on U, we simply say that the property holds on U.

From Definition 2, almost nonexpansiveness is actually the almost averaging property with the same violation and averaging constant α=1.

Remark 1

(the range of quantitative constants) In the context of Definition 2, it is natural to consider violation ε0 and averaging constant α(0,1]. Mathematically, it also makes sense to consider ε<0 and α>1 provided that the required estimate (2) holds true. Simple examples for the later case are linear contraction mappings. In this paper, averaging constant α>1 will frequently be involved implicitly in intermediate steps of our analysis without any contradiction or confusion. This is the reason why in Definition 2 (ii) we considered α>0 instead of α(0,1] as in [38, Definition 2.2].

It is worth noting that if the iteration xk+1Txk is linearly monotone with respect to FixT with rate c(0,1) and T is almost averaging on some neighborhood of FixT with averaging constant α(0,1], then (xk) converges R-linearly to a fixed point of T [39, Proposition 3.5].

We next prove a fundamental preliminary result for our analysis regarding almost averaging mappings.

Lemma 1

Let T:EE, UE, λ[0,1], ε0 and α>0. The following two statements are equivalent.

  • (i)

    T is almost averaging on U with violation ε and averaging constant α.

  • (ii)

    The λ-reflector of T, RT,λ=(1+λ)T-λId, is almost averaging on U with violation (1+λ)ε and averaging constant (1+λ)α.

Proof

Take any x,yU, x+Tx, y+Ty, x~=(1+λ)x+-λxRT,λx and y~=(1+λ)y+-λyRT,λy. We have by definition of RT,λ and [4, Corollary 2.14] that

x~-y~2=(1+λ)(x+-y+)-λ(x-y)2=(1+λ)x+-y+2-λx-y2+λ(1+λ)(x+-x)-(y+-y)2. 3

We also note that

(x~-x)-(y~-y)=(1+λ)(x+-x)-(y+-y). 4

(i) (ii). Suppose that T is almost averaging on U with violation ε and averaging constant α. Substituting (2) into (3) and using (4), we obtain that

x~-y~2(1+(1+λ)ε)x-y2-(1+λ)1-αα-λ(x+-x)-(y+-y)2=(1+(1+λ)ε)x-y2-1-αα-λ1+λ(x~-x)-(y~-y)2=(1+(1+λ)ε)x-y2-1-(1+λ)α(1+λ)α(x~-x)-(y~-y)2, 5

which means that RT,λ is almost averaging on U with violation (1+λ)ε and averaging constant (1+λ)α.

(ii) (i). Suppose that RT,λ is almost averaging on U with violation (1+λ)ε and averaging constant (1+λ)α, that is, the inequality (5) is satisfied. Substituting (3) into (5) and using (4), we obtain

(1+λ)x+-y+2-λx-y2+λ(1+λ)(x+-x)-(y+-y)2(1+(1+λ)ε)x-y2-1+λ1-αα-λ(x+-x)-(y+-y)2.

Equivalently,

x+-y+2(1+ε)x-y2-1-αα(x+-x)-(y+-y)2.

Hence T is almost averaging on U with violation ε and averaging constant α and the proof is complete.

Lemma 1 generalizes [13, Lemma 2.4] where the result was proved for α=1/2 and λ=1.

The next lemma recalls facts regarding the almost averagedness of projectors and reflectors associated with regular sets.

Lemma 2

Let AE be closed and (ε,δ)-regular at x¯A and define

U:={xEPAxBδ(x¯)}.
  • (i)

    The projector PA is pointwise almost nonexpansive on U at every point zABδ(x¯) with violation 2ε+ε2.

  • (ii)

    The projector PA is pointwise almost averaging on U at every point zABδ(x¯) with violation 2ε+2ε2 and averaging constant 1 / 2.

  • (iii)

    The λ-reflector RPA,λ is pointwise almost averaging on U at every point zABδ(x¯) with violation (1+λ)(2ε+2ε2) and averaging constant 1+λ2.

Proof

Statements (i) and (ii) can be found in [13, Theorem 2.14] or [38, Theorem 3.1 (i) & (iii)]. Statement (iii) follows from (ii) and Lemma 1 applied to T=PA and α=1/2.

The following concept of metric subregularity with functional modulus has played a central role, explicitly or implicitly, in the convergence analysis of Picard iterations [1, 13, 38, 39]. Recall that a function μ:[0,)[0,) is a gauge function if μ is continuous and strictly increasing and μ(0)=0.

Definition 3

(Metric subregularity with functional modulus) A mapping F:EE is metrically subregular with gauge μ on UE for y relative to ΛE if

μdistx,F-1(y)Λdisty,F(x)xUΛ.

When μ is a linear function, that is μ(t)=κt,t[0,), one says “with constant κ” instead of “with gauge μ=κId”. When Λ=E, the quantifier “relative to” is dropped.

Metric subregularity has many important applications in variational analysis and mathematical optimization, see the monographs and papers [11, 1518, 20, 21, 25, 40, 44]. For the discussion of metric subregularity in connection with subtransversality of collections of sets, we refer the reader to [23, 24, 29, 30].

The next theorem serves as the basic template for the quantitative convergence analysis of fixed point iterations. By the notation T:ΛΛ where Λ is a subset of E, we mean that T:EE and TxΛ for all xΛ. This simplification of notation should not lead to any confusion if one keeps in mind that there may exist fixed points of T that are not in Λ. For the importance of the use of Λ in isolating the desirable fixed point, we refer the reader to [1, Example 1.8]. In the following, riΛ denotes the relative interior of Λ.

Theorem 1

[38, Theorem 2.1] Let T:ΛΛ for ΛE and let SriΛ be closed and nonempty such that TyFixTS for all yS. Let O be a neighborhood of S such that OΛriΛ. Suppose that

  1. T is pointwise almost averaging at all points yS with violation ε and averaging constant α(0,1) on OΛ, and

  2. there exists a neighborhood V of FixTS and a constant κ>0 such that for all yS, y+Ty and all x+Tx the estimate
    κdist(x,S)x-x+-y-y+ 6
    holds whenever xOΛ\VΛ.

Then for all x+Tx

distx+,FixTS1+ε-(1-α)κ2αdist(x,S)

whenever xOΛ\VΛ.

In particular, if κ>εα1-α, then for any initial point x0OΛ the iteration xk+1Txk satisfies

distxk+1,FixTSckdist(x0,S)

with c:=1+ε-(1-α)κ2α<1 for all k such that xjOΛ\VΛ for j=1,2,,k.

Remark 2

[38, p. 13] In the case of S=FixT condition (6) reduces to metric subregularity of the mapping F:=T-Id for 0 on the annular set OΛ\VΛ, that is

κdist(x,F-1(0))dist(0,F(x))xOΛ\VΛ.

The inequality κ>εα1-α then states that the constant of metric subregularity κ is sufficiently large relative to the violation of the averaging property of T to guarantee linear progression of the iterates through that annular region.

For a comprehensive discussion on the roles of S and Λ in the analysis program of Theorem 1, we would like to refer the reader to the paper [38].

For the sake of simplification in terms of presentation, we have chosen to reduce the number of technical constants appearing in the analysis. It would be obviously analogous to formulate more theoretically general results by using more technical constants in appropriate places.

Tλ as a fixed point operator

We consider the problem of finding a fixed point of the operator

Tλ:=T1(1+λ)T2-λId-λT2-Id, 7

where λ[0,1] and Ti:EE (i=1,2) are assumed to be easily computed.

Examples of Tλ include the backward-backward and the DR algorithms [8, 10, 34, 36, 43] for solving the structured optimization problem

minimizexEf1(x)+f2(x)

under different assumptions on the functions fi (i=1,2). Indeed, when Ti are the prox mappings of fi with parameters τi>0, then Tλ with λ=0 and 1 takes the form Tλ=proxτ1,f1proxτ2,f2, and Tλ=proxτ1,f12proxτ2,f2-Id-proxτ2,f2+Id, respectively.

We first characterize the set of fixed points of Tλ via those of the constituent operators Ti (i=1,2).

Proposition 1

Let T1,T2:EE, λ[0,1] and consider Tλ defined at (7). The following statements hold true.

  • (i)

    (1+λ)Tλ-λId=(1+λ)T1-λId(1+λ)T2-λId.

    As a consequence,
    FixTλ=Fix(1+λ)T1-λId(1+λ)T2-λId.
  • (ii)
    Suppose that T1=PA is the projector on an affine set A and T2 is single-valued. Then
    FixTλ={xEPAx=λT2x+(1-λ)x}{xEPAx=PAT2x}. 8

Proof

(i). We have by the construction of Tλ that

(1+λ)Tλ-λId=(1+λ)T1(1+λ)T2-λId-λ(T2-Id)-λId=(1+λ)T1(1+λ)T2-λId-λ(1+λ)T2-λId=(1+λ)T1-λId(1+λ)T2-λId.

(ii). We first take an arbitrary xFixTλ and prove that

PAx=PAT2x=λT2x+(1-λ)x.

Indeed, from x=Tλx, we get

x=PA(1+λ)T2x-λx-λ(T2x-x)λT2x+(1-λ)x=PA(1+λ)T2x-λx. 9

In particular, λT2x+(1-λ)xA. Thus by equality (9) and the assumption that PA is affine, we have

PAλT2x+(1-λ)x=PA(1+λ)T2x-λxλPAT2x+(1-λ)PAx=(1+λ)PAT2x-λPAxPAx=PAT2x. 10

Substituting (10) into (9) also yields

λT2x+(1-λ)x=(1+λ)PAT2x-λPAx=(1+λ)PAx-λPAx=PAx.

Finally, let us take an arbitrary x satisfying PAx=λT2x+(1-λ)x and prove that xFixTλ. Indeed, we note that λT2x+(1-λ)xA. Since PA is affine, one can easily check (10) and then (9), which is equivalent to xFixTλ. The proof is complete.

The inclusion (8) in Proposition 1 can be strict as shown in the next example.

Example 1

Let us consider E=R2, the set A=(x1,x2)R2x1=0 and the two operators T1=PA and T2x=12x (xR2). Then for any point x=(x1,0) with x10, we have PAx=PAT2x=(0,0) but PAx=(0,0)(1-λ/2)x=λT2x+(1-λ)x, that is xFixTλ.

The next proposition shows that the almost averagedness of Tλ naturally inherits from that of T1 and T2 via Krasnoselski–Mann relaxations.

Proposition 2

(Almost averagedness of Tλ) Let λ[0,1], Ti be almost averaging on UiE with violation εi0 and averaging constant αi>0 (i=1,2) and define the set

U:={xU2RT2,λxU1}.

Then Tλ is almost averaging on U with violation ε=ε1+ε2+(1+λ)ε1ε2 and averaging constant α=2max{α1,α2}1+(1+λ)max{α1,α2}.

Proof

By the implication (i) (ii) of Lemma 1, the operators RTi,λ=(1+λ)Ti-λId are almost averaging on Ui with violation (1+λ)εi and averaging constant (1+λ)αi (i=1,2). Then thanks to [38, Proposition 2.4 (iii)], the operator T:=RT1,λRT2,λ is almost averaging on U with violation (1+λ)ε1+ε2+(1+λ)ε1ε2 and averaging constant 2(1+λ)max{α1,α2}1+(1+λ)max{α1,α2}. Note that Tλ=(1+λ)T-λId by Proposition 1. We have by the implication (ii) (i) of Lemma 1 that Tλ is almost averaging on U with violation ε=ε1+ε2+(1+λ)ε1ε2 and averaging constant α=2max{α1,α2}1+(1+λ)max{α1,α2} as claimed.

We next discuss convergence of Tλ based on the abstract results established in [38]. Our agenda is to verify the assumptions of Theorem 1. To simplify the exposure in terms of presentation, we have chosen to state the results corresponding to S=FixTλ and Λ=E in Theorem 1. In the sequel, we will denote, for a nonnegative real ρ,

Sρ:=FixTλ+ρB.

Theorem 2

(Convergence of algorithm Tλ with metric subregularity) Let Tλ be defined at (7), δ>0 and γ(0,1). Suppose that for each nN, the following conditions are satisfied.

  • (i)

    T2 is almost averaging on Sγnδ with violation ε2,n0 and averaging constant α2,n(0,1), and T1 is almost averaging on the set SγnδRT2,λSγnδ with violation ε1,n0 and averaging constant α1,n(0,1).

  • (ii)
    The mapping Tλ-Id is metrically subregular on Dn:=Sγnδ\Sγn+1δ for 0 with gauge μn satisfying
    infxDnμndistx,FixTλdistx,FixTλκn>αnεn1-αn, 11
    where εn:=ε1,n+ε2,n+(1+λ)ε1,nε2,n and αn:=2max{α1,n,α2,n}1+(1+λ)max{α1,n,α2,n}.

Then all iterations xk+1Tλxk starting in Sδ satisfy

distxk,FixTλ0 12

and

distxk+1,FixTλcndistxk,FixTλxkDn, 13

where cn:=1+εn-(1-αn)κn2αn<1.

In particular, if (1-αn)κn2αn-εn is bounded from below by some τ>0 for all n sufficiently large, then the convergence (12) is R-linear with rate at most 1-τ.

Proof

For each nN, we verify the assumptions of Theorem 1 for O=Sγnδ, V=Sγn+1δ and Dn=O\V=Sγnδ\Sγn+1δ. Under assumption (i) of Theorem 2, Proposition 2 ensures that Tλ is almost averaging on Sγnδ with violation εn and averaging constant αn. In other words, condition (a) of Theorem 1 is satisfied with ε=εn and α=αn. Assumption (ii) of Theorem 2 also fulfills condition (b) of Theorem 1 with κ=κn in view of Remark 2. Theorem 1 then yields the conclusion of Theorem 2 after a straightforward care of the involving quantitative constants.

The first inequality in (11) essentially says that the gauge function μn can be bounded from below by a linear function on the reference interval.

Remark 3

In Theorem 2, the fundamental goal of formulating assumption (i) on the set Sγnδ and assumption (ii) on the set Dn is that one can characterize sublinear convergence of an iteration on Sδ via linear progression of its iterates through each of the annular set Dn. This idea is based on the fact that for larger n, the almost averaging property of Tλ on Sγnδ is always improved but the metric subregularity on Dn may get worse, however, if the corresponding quantitative constants still satisfy condition (11), then convergence is guaranteed. For an illustrative example, we refer the reader to [38, Example 2.4].

Application to feasibility

We consider algorithm Tλ for solving feasibility problem involving two closed sets A,BE,

x+Tλx=PA(1+λ)PBx-λx-λPBx-x=PARPB,λ(x)-λPBx-x. 14

Note that Tλ with λ=0 and 1 corresponds to the alternating projections PAPB and the DR method 12(RARB+Id), respectively.

It is worth recalling that feasibility problem for m2 sets can be reformulated as a feasibility problem for two constructed sets on the product space Em with one of the later sets is a linear subspace, and the regularity properties in terms of both individual sets and collections of sets of the later sets are inherited from those of the former ones [3, 32].

When A is an affine set, then the projector PA is affine and Tλ is a convex combination of the alternating projection and the DR methods since

Tλx=PA(1-λ)PBx+λ(2PBx-x)-λPBx-x=(1-λ)PAPBx+λx+PA(2PBx-x)-PBx=(1-λ)T0(x)+λT1(x).

In this case, we establish convergence results for all convex combinations of the alternating projection and the DR methods. To our best awareness, this kind of results seems to be new.

Recall that when applied to inconsistent feasibility problems the DR operator has no fixed points. We next show that the set of fixed points of Tλ with λ[0,1) for convex inconsistent feasibility problems is nonempty. This result follows the lines of [36, Lemma 2.1] where the fixed point set of the RAAR operator is characterized.

Proposition 3

(Fixed points of Tλ for convex inconsistent feasibility) For closed convex sets A,BE, let G=B-A¯, g=PG0, E=A(B-g) and F=(A+g)B. Then

FixTλ=E-λ1-λgλ[0,1).

Proof

We first show that E-λ1-λgFixTλ. Pick any eE and denote f=e+gF as definitions of E and F. We are checking that

x:=e-λ1-λgFixTλ.

Since x=f-11-λg and -gNB(f), we get PBx=f.

Analogously, since gNA(e) and

(1+λ)PBx-λx=(1+λ)f-λx=e+11-λg,

we have PA((1+λ)PBx-λx)=e.

Hence,

x-Tλx=x-PA(1+λ)PBx-λx+λPBx-x=x-e+λf-x=0.

That is xFixTλ.

We next show that FixTλE-λ1-λg. Pick any xFixTλ. Let f=PBx and y=x-f. Thanks to xFixTλ and the definition of Tλ,

PA((1+λ)PBx-λx)=λ(PBx-x)+x=-λy+y+f=f+(1-λ)y. 15

Now, for any aA, since A is closed and convex, we have

0a-PA((1+λ)PBx-λx),(1+λ)PBx-λx-PA((1+λ)PBx-λx)=a-(f+(1-λ)y),(1+λ)f-λx-(f+(1-λ)y)=a-f-(1-λ)y,-y=-a+f,y+(1-λ)y2.

On the other hand, for any bB, since B is closed and convex, we have

b-f,y=b-f,x-f=b-PBx,x-PBx0.

Combining the last two inequalities yields

b-a,y-(1-λ)y20aA,bB.

Take a sequence (an) in A and a sequence (bn) in B such that gn:=bn-ang. Then

gn,y-(1-λ)y20n. 16

Taking the limit and using the Cauchy–Schwarz inequality yields

y11-λg.

Conversely, by (15) with noting that fB and PA((1+λ)PBx-λx)A,

y=11-λf-PA((1+λ)PBx-λx)11-λg.

Hence y=11-λg, and taking the limit in (16), which yields y=-11-λg. Since fB and f-g=f+(1-λ)y=PA((1+λ)PBx-λx)A, we have f-gA(B-g)=E and, therefore,

x=f+y=f-11-λg=f-g-λ1-λgE-λ1-λg.

We next discuss the two key ingredients for convergence of algorithm Tλ applied to feasibility problems: 1) almost averagedness of Tλ, and 2) metric subregularity of Tλ-Id. The two properties will be deduced from the (ε,δ)-regularity of the individual sets and the transversality of the collection of sets, respectively.

The next proposition shows averagedness of Tλ applied to feasibility problems involving (ε,δ)-regular sets.

Proposition 4

Let A and B be (ε,δ)-regular at x¯AB and define the set

U:={xEPBxBδ(x¯)andPARPB,λxBδ(x¯)}. 17

Then Tλ is pointwise almost averaging on U at every point zS:=ABBδ(x¯) with averaging constant 23+λ and violation

ε~:=2(2ε+2ε2)+(1+λ)(2ε+2ε2)2. 18

Proof

Let us define the two sets

UA:={yEPAyBδ(x¯)},UB:={xEPBxBδ(x¯)}

and note that xU if and only if xUB and RPB,λxUA. Thanks to Lemma 2 (iii), RPA,λ and RPB,λ are pointwise almost averaging at every point zS with violation (1+λ)(2ε+2ε2) and averaging constant 1+λ2 on UA and UB, respectively. Then due to [38, Proposition 2.4 (iii)], the operator T:=RPA,λRPB,λ is pointwise almost averaging on U at every point zS with averaging constant 2(1+λ)3+λ and violation (1+λ)ε~, where ε~ is given by (18). Note that Tλ=(1+λ)T-λId by Proposition 1. Thanks to Lemma 1, Tλ is pointwise almost averaging on U at every point zS with violation ε~ and averaging constant 23+λ as claimed.

Remark 4

It follows from Lemma 2 (i) & (iii) that the set U defined by (17) contains at least the ball Bδ(x¯), where

δ:=δ2(1+ε)1+(1+λ)(2ε+2ε2)>0.

We next integrate Proposition 4 into Theorem 2 to obtain convergence of algorithm Tλ for solving consistent feasibility problems involving (ε,δ)-regular sets.

Corollary 1

(Convergence of algorithm Tλ for feasibility) Consider the algorithm Tλ defined at (14) and suppose that FixTλ=AB. Denote Sρ=FixTλ+ρB for a nonnegative real ρ. Suppose that there are δ>0, ε0 and γ(0,1) such that A and B are (ε,δ)-regular at avery point zAB, where

δ:=2δ(1+ε)1+(1+λ)(2ε+2ε2),

and for each nN, the mapping Tλ-Id is metrically subregular on Dn:=Sγnδ\Sγn+1δ for 0 with gauge μn satisfying

infxDnμndistx,ABdistx,ABκn>2ε~1+λ,

where ε~ is given at (18).

Then all iterations xk+1Tλxk starting in Sδ satisfy (12) and (13) with cn:=1+ε~-(1+λ)κn22<1.

In particular, if (κn) is bounded from below by some κ>2ε~1+λ for all n sufficiently large, then (xk) eventually converges R-linearly to a point in AB with rate at most 1+ε~-(1+λ)κ22<1.

Proof

Let any xDn, for some nN, x+Tλx and x¯PABx. A combination of Proposition 4 and Remark 4 implies that Tλ is pointwise almost averaging on Bδ(x¯) at every point zABBδ(x¯) with violation ε~ given by (18) and averaging constant 23+λ. In other words, condition (a) of Theorem 1 is satisfied. Condition (b) of Theorem 1 is also fulfilled by the same argument as the one used in Theorem 2. The desired conclusion now follows from Theorem 1.

In practice, the metric subregularity assumption is often more challenging to be verified than the averaging property. In the concrete example of consistent alternating projections PAPB, that metric subregularity condition holds true if and only if the collection of sets is subtransversal. We next show that the metric subregularity of Tλ-Id can be deduced from the transversality of the collection of sets {A,B}. As a result, if the sets are also sufficiently regular, then local linear convergence of the iteration xk+1Tλxk is guaranteed.

We first describe the concept of relative transversality of collections of sets. In the sequel, we set Λ:=aff(AB), the smallest affine set in E containing both A and B.

Assumption 3

The collection {A,B} is transversal at x¯AB relative to Λ with constant θ¯<1, that is, for any θ(θ¯,1), there exists δ>0 such that

u,v-θu·v

holds for all aABδ(x¯), bBBδ(x¯), uNA|Λprox(a) and vNB|Λprox(b).

Thanks to [22, Theorem 1] and [28, Theorem 1], Assumption 3 also ensures subtransversality of {A,B} at x¯ relative to Λ with constant at least 1-θ2 on the neighborhood Bδ(x¯), that is

1-θ2dist(x,AB)max{dist(x,A),dist(x,B)}xΛBδ(x¯). 19

The next lemma is at the heart of our subsequent discussion.

Lemma 3

Suppose that Assumption 3 is satisfied. Then for any θ(θ¯,1), there exists a number δ>0 such that for all xBδ(x¯) and x+Tλx,

κdist(x,AB)x-x+, 20

where κ is defined by

κ:=(1-θ)1+θ2max1,λ+1-θ2>0. 21

Proof

For any θ(θ¯,1), there is a number δ>0 satisfying the property described in Assumption 3. Let us set δ=δ/6 and show that condition (20) is fulfilled with δ.

Indeed, let us consider any xBδ(x¯), bPBx, y=(1+λ)b-λx, aPAy and x+=a-λ(b-x)Tλx. From the choice of δ, it is clear that a,bBδ(x¯). Since x-bNB|Λprox(b) and y-aNA|Λprox(a), Assumption 3 yields that

x-b,y-a-θx-b·y-a. 22

By the definition of Tλ, we have

x-x+2=x-b+y-a2=x-b2+y-a2+2x-b,y-ax-b2+y-a2-2θx-b·y-a1-θ2x-b2=1-θ2dist2(x,B), 23

where the first inequality follows from (22).

We will take care of the two possible cases regarding dist(x,A) as follows.

Case 1 dist(x,A)λ+1-θ2dist(x,B). Thanks to (23) we get

x-x+21-θ2λ+1-θ22dist2(x,A). 24

Case 2 dist(x,A)>λ+1-θ2dist(x,B). By the triangle inequality and the construction of Tλ, we get

x-x+x-a-a-x+=x-a-λx-bdist(x,A)-λdist(x,B)1-λλ+1-θ2dist(x,A). 25

Since

1-θ2λ+1-θ22=1-λλ+1-θ22,

we always have from (24) and (25) that

x-x+21-θ2λ+1-θ22dist2(x,A). 26

Combining (23), (26) and (19), we obtain

x-x+21-θ2max1,λ+1-θ22maxdist2(x,A),dist2(x,B)(1-θ2)(1-θ)2max1,λ+1-θ22dist2(x,AB),

which yields (20) as claimed.

In the special case that λ=1, Lemma 3 refines [13, Lemma 3.14] and [45, Lemma 4.2] where the result was proved for the DR operator with an additional assumption on regularity of the sets.

The next result is the final preparation for our linear convergence result.

Lemma 4

[45, Proposition 2.11] Let T:EE, SE be closed and x¯S. Suppose that there are δ>0 and c[0,1) such that for all xBδ(x¯), x+Tx and zPSx,

x+-zcx-z. 27

Then every iteration xk+1Txk starting sufficiently close to x¯ converges R-linearly to a point x~SBδ(x¯). In particular,

xk-x~x0-x¯(1+c)1-cck.

We are now ready to prove local linear convergence for algorithm Tλ which generalizes the corresponding results established in [13, 45] for the DR method.

Theorem 4

(Linear convergence of algorithm Tλ for feasibility) In addition to Assumption 3, suppose that A and B are (ε,δ)-regular at x¯ with ε~<(1+λ)κ22, where ε~ and κ are given by (18) and (21), respectively. Then every iteration xk+1Tλxk starting sufficiently close to x¯ converges R-linearly to a point in AB.

Proof

Assumption 3 ensures the existence of δ1>0 such that Lemma 3 holds true. In view of Proposition 4 and Remark 4, one can find a number δ2>0 such that Tλ is pointwise almost averaging on Bδ2(x¯) at every point zABBδ2(x¯) with violation ε~ given by (18) and averaging constant 23+λ. Define δ=min{δ1,δ2}>0.

Now let us consider any xBδ/2(x¯), x+Tλx and zPABx. It is clear that zBδ(x¯). Proposition 4 and Lemma 3 then respectively yield

x+-z2(1+ε~)x-z2-1+λ2x-x+2, 28
x-x+2κ2dist2(x,AB)=κ2x-z2, 29

where κ is given by (21).

Substituting (29) into (28), we get

x+-z21+ε~-(1+λ)κ22x-z2,

which yields condition (27) of Lemma 4 and the desired conclusion now follows from this lemma.

Application to sparse optimization

Our goal in this section is twofold: 1) to illustrate the linear convergence of algorithm Tλ formulated in Theorem 4 via the sparse optimization problem, and 2) to demonstrate a promising performance of the algorithm Tλ in comparison with the RAAR algorithm for this applied problem.

Sparse optimization

We consider the sparse optimization problem

minxRnx0subjecttoMx=b, 30

where MRm×n (m<n) is a full rank matrix, b is a given vector in Rm, and x0 is the number of nonzero entries of the vector x. The sparse optimization problem with complex variable is defined analogously by replacing R by C everywhere in the above model.

Many strategies for solving (30) have been proposed. We refer the reader to the famous paper by Candès and Tao [9] for solving this problem by using convex relaxations. On the other hand, assuming to have a good guess on the sparsity of the solutions to (30), one can tackle this problem by solving the sparse feasibility problem [14] of finding

x¯AsB, 31

where As:={xRnx0s} and B:={xRnMx=b}.

It is worth mentioning that the initial guess s of the true sparsity is not numerically sensitive with respect to various projection methods, that is, for a relatively wide range of values of s above the true sparsity, projection algorithms perform very much in the same nature. Note also that the approach via sparse feasibility does not require convex relaxations of (30) and thus can avoid the likely expensive increase of dimensionality.

We run the two algorithms Tλ and RAAR to solve (31) and compare their numerical performances. By taking s smaller than the true sparsity, we can also compare their performances for inconsistent feasibility.

Since B is affine, there is the closed algebraic form for the projector PB,

PBx=x-M(Mx-b)xRn,

where M:=MT(MMT)-1 is the Moore–Penrose inverse of M. We have denoted MT the transpose matrix of M and taken into account that M is full rank. There is also a closed form for PAs [6]. For each xRn, let us denote Is(x) the set of all s-tubles of indices of s largest in absolute value entries of x. The set Is(x) can contain multiple such s-tubles. The projector PAs can be described as

PAsx=zRnIIs(x)suchthatz(k)=x(k)ifkI,0else.

For convenience, we recall the two algorithms in this specific setting

RAARβ=βPAs(2PB-Id)+(1-2β)PB+βId,Tλ=PAs(1+λ)PB-λId-λ(PB-Id).

Convergence analysis

We analyze the convergence of algorithm Tλ for the sparse feasibility problem (31). The next theorem establishes local linear convergence of algorithm Tλ for solving sparse feasibility problems.

Theorem 5

(Linear convergence of algorithm Tλ for sparse feasibility) Let x¯=(x¯i)AsB and suppose that s is the sparsity of the solutions to the problem (30). Then any iteration xk+1Tλxk starting sufficiently close to x¯ converges R-linearly to x¯.

Proof

We first show that x¯ is an isolated point of AsB. Since s is the sparsity of the solutions to (30), we have that x¯0=s and the set Is(x¯) contains a unique element, denoted Ix¯. Note that Ex¯:=span{ei:iIx¯} is the unique s-dimensional space component of As containing x¯, where {ei:1in} is the canonical basic of Rn. Let us denote

δ:=miniIx¯|x¯i|>0.

We claim that

AsBδ(x¯)=Ex¯Bδ(x¯), 32
Ex¯B={x¯}. 33

Indeed, for any x=(xi)AsBδ(x¯), we have by definition of δ that xi0 for all iIx¯. Hence x0=s and xEx¯Bδ(x¯). This proves (32).

For (33), it suffices to show the singleton of Ex¯B since we already know that x¯Ex¯B. Suppose otherwise that there exists x=(xi)Ex¯B with xjx¯j for some index j. Since both Ex¯ and B are affine, the intersection Ex¯B contains the line {x+t(x¯-x):tR} passing x and x¯. In particular, it contains the point z:=x+xjxj-x¯j(x¯-x). Then we have that zB and z0s-1 as zj=0. This contradicts to the assumption that s is the sparsity of the solutions to (30), and hence (33) is proved.

A combination of (32) and (33) then yields

AsBBδ(x¯)=Ex¯BBδ(x¯)={x¯}. 34

This means that x¯ is an isolated point of AsB as claimed. Moreover, the equalities in (34) imply that

PAsx=PEx¯xxBδ/2(x¯).

Therefore, for any starting point x0Bδ/2(x¯), the iteration xk+1Tλxk for solving (31) is identical to that for solving the feasibility problem for the two sets Ex¯ and B. Since Ex¯ and B are two affine subspaces intersecting at the unique point x¯ by (33), the collection of sets {Ex¯,B} is transversal at x¯ relative to the affine hull aff(Ex¯B). Theorem 4 now can be applied to conclude that the iteration xk+1Tλxk converges R-linearly to x¯. The proof is complete.

It is worth mentioning that the convergence analysis in Theorem 5 is also valid for the RAAR algorithm.

Numerical experiment

We now set up a toy example as in [9, 14] which involves an unknown true object x¯R2562 with x¯0=328 (the sparsity rate is .005). Let b be 1 / 8 of the measurements of F(x¯), the Fourier transform of x¯, with the sample indices denoted J. The Poisson noise was added when calculating the measurement b. Note that since x¯ is real, F(x¯) is conjugate symmetric, we indeed have nearly a double number of measurements. In this setting, we have

B={xC2562F(x)(k)=b(k),kJ},

and the two prox operators, respectively, take the forms

PAsx=zRnIIs(x)suchthatz(k)=Rex(k)ifkI,0else,PBx=F-1(x^),wherex^(k)=b(k)ifkJ,F(x)(k)else,

where Re(x(k)) denotes the real part of the complex number x(k), and F-1 is the inverse Fourier transform.

The initial point was chosen randomly, and a warm-up procedure with 10 DR iterates was performed before running the two algorithms. The stopping criterion x-x+<10-10 was used. We have used the Matlab ProxToolbox [37] to run this numerical experiment. The parameters were chosen in such a way that the performance is seemingly optimal for both algorithms. We chose β=.65 for the RAAR algorithm and λ=.45 for algorithm Tλ in the case of consistent feasibility problem corresponding to s=340, and β=.6 for the RAAR algorithm and λ=.4 for algorithm Tλ in the case of inconsistent feasibility problem corresponding to s=310.

The change of distances between two consecutive iterates is of interest. When linear convergence appears to be the case, it can yield useful information of the convergence rate. Under the assumption that the iterates will remain in the convergence area, one can obtain error bounds for the distance from the current iterate to a nearest solution. We also pay attention to the gaps in iterates that in a sense measure the infeasibility at the iterates. If we think feasibility problem as the problem of minimizing the sum of the squares of the distance functions to the sets, then gaps in iterates are the values of that function evaluated at the iterates. For the two algorithms under consideration, the iterates are themselves not informative but their shadows, by which we mean the projections of the iterates on one of the sets. Hence, the gaps in iterates are calculated for the iterate shadows instead of the iterates themselves.

Figure 1 summarizes the performances of the two algorithms for both consistent and inconsistent sparse feasibility problems. We first emphasize that the algorithms appear to be convergent in both cases of feasibility. For the consistent case, algorithm Tλ appears to perform better than the RAAR algorithm in terms of both the iterate changes and gaps. Also, the CPU time of algorithm Tλ is around 10% less than that of the RAAR algorithm. For the inconsistent case, we have a similar observation except that the iterate gaps for the RAAR algorithm are slightly better (smaller) than those for algorithm Tλ. Extensive numerical experiments in imaging problems illustrating the empirical performance of algorithm Tλ will be the future work.

Fig. 1.

Fig. 1

Performances of the RAAR and Tλ algorithms for sparse feasibility problem: iterate changes in consistent case (top-left), iterate gaps in consistent case (top-right), iterate changes in inconsistent case (bottom-left) and iterate gaps in inconsistent case (bottom-right)

Acknowledgements

The author would like to thank Prof. Dr. Russell Luke and Prof. Dr. Alexander Kruger for their encouragement and valuable suggestions during the preparation of this work. He also would like to thank the anonymous referees for their very helpful and constructive comments on the manuscript version of the paper.

Footnotes

The research leading to these results has received funding from the German-Israeli Foundation Grant G-1253-304.6 and the European Research Council under the European Union’s Seventh Framework Programme (FP7/2007-2013)/ERC Grant Agreement No. 339681.

References

  • 1.Aspelmeier T, Charitha C, Luke DR. Local linear convergence of the ADMM/Douglas–Rachford algorithms without strong convexity and application to statistical imaging. SIAM J. Imaging Sci. 2016;9(2):842–868. doi: 10.1137/15M103580X. [DOI] [Google Scholar]
  • 2.Attouch H, Bolte J, Redont P, Soubeyran A. Proximal alternating minimization and projection methods for nonconvex problems: an approach based on the Kurdyka–Łojasiewicz inequality. Math. Oper. Res. 2010;35(2):438–457. doi: 10.1287/moor.1100.0449. [DOI] [Google Scholar]
  • 3.Bauschke HH, Borwein JM. On projection algorithms for solving convex feasibility problems. SIAM Rev. 1996;38(3):367–426. doi: 10.1137/S0036144593251710. [DOI] [Google Scholar]
  • 4.Bauschke HH, Combettes PL. Convex Analysis and Monotone Operator Theory in Hilbert Spaces. New York: Springer; 2011. [Google Scholar]
  • 5.Bauschke HH, Luke DR, Phan HM, Wang X. Restricted normal cones and the method of alternating projections: applications. Set-Valued Var. Anal. 2013;21:475–501. doi: 10.1007/s11228-013-0238-3. [DOI] [Google Scholar]
  • 6.Bauschke HH, Luke DR, Phan HM, Wang X. Restricted normal cones and sparsity optimization with affine constraints. Found. Comput. Math. 2014;14:63–83. doi: 10.1007/s10208-013-9161-0. [DOI] [Google Scholar]
  • 7.Bolte J, Sabach S, Teboulle M. Proximal alternating linearized minimization for nonconvex and nonsmooth problems. Math. Program. 2014;146(1–2):459–494. doi: 10.1007/s10107-013-0701-9. [DOI] [Google Scholar]
  • 8.Borwein JM, Tam MK. The cyclic Douglas–Rachford method for inconsistent feasibility problems. J. Nonlinear Convex Anal. 2015;16(4):537–584. [Google Scholar]
  • 9.Candés E, Tao T. Decoding by linear programming. IEEE Trans. Inf. Theory. 2005;51(12):4203–4215. doi: 10.1109/TIT.2005.858979. [DOI] [Google Scholar]
  • 10.Combettes, P.L., Pesquet, J.-C.: Proximal splitting methods in signal processing. In: Fixed-Point Algorithms for Inverse Problems in Science and Engineering, vol. 49. Springer, Berlin, pp. 185–212 (2011)
  • 11.Dontchev AL, Rockafellar RT. Implicit Functions and Solution Mapppings. New York: Srpinger; 2014. [Google Scholar]
  • 12.Drusvyatskiy D, Ioffe AD, Lewis AS. Transversality and alternating projections for nonconvex sets. Found. Comput. Math. 2015;15(6):1637–1651. doi: 10.1007/s10208-015-9279-3. [DOI] [Google Scholar]
  • 13.Hesse R, Luke DR. Nonconvex notions of regularity and convergence of fundamental algorithms for feasibility problems. SIAM J. Optim. 2013;23(4):2397–2419. doi: 10.1137/120902653. [DOI] [Google Scholar]
  • 14.Hesse R, Luke DR, Neumann P. Alternating projections and Douglas–Rachford for sparse affine feasibility. IEEE Trans. Signal. Process. 2014;62(18):4868–4881. doi: 10.1109/TSP.2014.2339801. [DOI] [Google Scholar]
  • 15.Ioffe AD. Metric regularity and subdifferential calculus. Russian Math. Surv. 2000;55(3):501–558. doi: 10.1070/RM2000v055n03ABEH000292. [DOI] [Google Scholar]
  • 16.Ioffe AD. Regularity on a fixed set. SIAM J. Optim. 2011;21(4):1345–1370. doi: 10.1137/110820981. [DOI] [Google Scholar]
  • 17.Ioffe AD. Nonlinear regularity models. Math. Program. 2013;139(1–2):223–242. doi: 10.1007/s10107-013-0670-z. [DOI] [Google Scholar]
  • 18.Ioffe AD. Metric regularity: a survey. Part I. Theory. J. Aust. Math. Soc. 2016;101(2):188–243. doi: 10.1017/S1446788715000701. [DOI] [Google Scholar]
  • 19.Khanh Phan Q, Kruger AY, Thao Nguyen H. An induction theorem and nonlinear regularity models. SIAM J. Optim. 2015;25(4):2561–2588. doi: 10.1137/140991157. [DOI] [Google Scholar]
  • 20.Klatte D, Kummer B. Nonsmooth Equations in Optimization. Dordrecht: Kluwer; 2002. [Google Scholar]
  • 21.Klatte D, Kummer B. Optimization methods and stability of inclusions in Banach spaces. Math. Program. 2009;117(1–2):305–330. doi: 10.1007/s10107-007-0174-9. [DOI] [Google Scholar]
  • 22.Kruger AY. Stationarity and regularity of set systems. Pac. J. Optim. 2005;1(1):101–126. [Google Scholar]
  • 23.Kruger AY. About regularity of collections of sets. Set-Valued Anal. 2006;14:187–206. doi: 10.1007/s11228-006-0014-8. [DOI] [Google Scholar]
  • 24.Kruger AY. About stationarity and regularity in variational analysis. Taiwan. J. Math. 2009;13(6A):1737–1785. doi: 10.11650/twjm/1500405612. [DOI] [Google Scholar]
  • 25.Kruger AY. Error bounds and metric subregularity. Optimization. 2015;64(1):49–79. doi: 10.1080/02331934.2014.938074. [DOI] [Google Scholar]
  • 26.Kruger, A.Y., Luke, D.R., Thao, Nguyen H.: Set regularities and feasibility problems. Math. Program. B. 10.1007/s10107-016-1039-x
  • 27.Kruger AY, Luke DR, Thao Nguyen H. About subtransversality of collections of sets. Set-Valued Var. Anal. 2017;25(4):701–729. doi: 10.1007/s11228-017-0436-5. [DOI] [Google Scholar]
  • 28.Kruger AY, Thao Nguyen H. About uniform regularity of collections of sets. Serdica Math. J. 2013;39:287–312. [Google Scholar]
  • 29.Kruger AY, Thao Nguyen H. About [q]-regularity properties of collections of sets. J. Math. Anal. Appl. 2014;416(2):471–496. doi: 10.1016/j.jmaa.2014.02.028. [DOI] [Google Scholar]
  • 30.Kruger AY, Thao Nguyen H. Quantitative characterizations of regularity properties of collections of sets. J. Optim. Theory Appl. 2015;164:41–67. doi: 10.1007/s10957-014-0556-0. [DOI] [Google Scholar]
  • 31.Kruger AY, Thao Nguyen H. Regularity of collections of sets and convergence of inexact alternating projections. J. Convex Anal. 2016;23(3):823–847. [Google Scholar]
  • 32.Lewis AS, Luke DR, Malick J. Local linear convergence of alternating and averaged projections. Found. Comput. Math. 2009;9(4):485–513. doi: 10.1007/s10208-008-9036-y. [DOI] [Google Scholar]
  • 33.Lewis AS, Malick J. Alternating projections on manifolds. Math. Oper. Res. 2008;33:216–234. doi: 10.1287/moor.1070.0291. [DOI] [Google Scholar]
  • 34.Li G, Pong TK. Douglas–Rachford splitting for nonconvex feasibility problems. Math. Program. 2016;159(1):371–401. doi: 10.1007/s10107-015-0963-5. [DOI] [Google Scholar]
  • 35.Luke DR. Relaxed averaged alternating reflections for diffraction imaging. Inverse Problems. 2005;21:37–50. doi: 10.1088/0266-5611/21/1/004. [DOI] [Google Scholar]
  • 36.Luke DR. Finding best approximation pairs relative to a convex and a prox-regular set in Hilbert space. SIAM J. Optim. 2008;19(2):714–739. doi: 10.1137/070681399. [DOI] [Google Scholar]
  • 37.Luke, D.R.: ProxToolbox. http://num.math.uni-goettingen.de/proxtoolbox (2017). Accessed Aug 2017
  • 38.Luke, D.R., Thao, Nguyen H., Tam, M.K.: Quantitative convergence analysis of iterated expansive, set-valued mappings. Math. Oper. Res. 10.1287/moor.2017.0898
  • 39.Luke, D.R., Thao, Nguyen H., Teboulle, M.: Necessary conditions for linear convergence of Picard iterations and application to alternating projections. https://arxiv.org/pdf/1704.08926.pdf (2017)
  • 40.Mordukhovich BS. Variational Analysis and Generalized Differentiation. I: Basic Theory. Berlin: Springer; 2006. [Google Scholar]
  • 41.Moreau J-J. Fonctions convexes duales et points proximaux dans un espace Hilbertien. Comptes Rendus de l’Académie des Sciences de Paris. 1962;255:2897–2899. [Google Scholar]
  • 42.Noll D, Rondepierre A. On local convergence of the method of alternating projections. Found. Comput. Math. 2016;16(2):425–455. doi: 10.1007/s10208-015-9253-0. [DOI] [Google Scholar]
  • 43.Patrinos, P., Stella, L., Bemporad, A.: Douglas-Rachford splitting: Complexity estimates and accelerated variants. In: 53rd IEEE Conference on Decision and Control, pp. 4234–4239 (2014)
  • 44.Penot J-P. Calculus Without Derivatives. New York: Springer; 2013. [Google Scholar]
  • 45.Phan HM. Linear convergence of the Douglas–Rachford method for two closed sets. Optimization. 2016;65:369–385. doi: 10.1080/02331934.2015.1051532. [DOI] [Google Scholar]
  • 46.Rockafellar, R.T., Wets, R.J.-B.: Variational Analysis. Grundlehren Math. Wiss. Springer, Berlin (1998)

Articles from Computational Optimization and Applications are provided here courtesy of Springer

RESOURCES