Skip to main content
Springer logoLink to Springer
. 2017 Jun 12;2017(1):135. doi: 10.1186/s13660-017-1409-9

New generalized variable stepsizes of the CQ algorithm for solving the split feasibility problem

Peiyuan Wang 1,2,, Jianjun Zhou 1, Risheng Wang 1, Jie Chen 1
PMCID: PMC5488062  PMID: 28680238

Abstract

Variable stepsize methods are effective for various modified CQ algorithms to solve the split feasibility problem (SFP). The purpose of this paper is first to introduce two new simpler variable stepsizes of the CQ algorithm. Then two new generalized variable stepsizes which can cover the former ones are also proposed in real Hilbert spaces. And then, two more general KM (Krasnosel’skii-Mann)-CQ algorithms are also presented. Several weak and strong convergence properties are established. Moreover, some numerical experiments have been taken to illustrate the performance of the proposed stepsizes and algorithms.

Keywords: split feasibility problem, CQ algorithm, variable stepsize, generalized variable stepsize

Introduction

Since the CQ algorithm for solving the split feasibility problem (SFP) was proposed [1] in order to get better convergence speed, much attention has been paid to improve the variable stepsize of CQ algorithm.

Let H1 and H2 be real Hilbert spaces, C and Q be nonempty closed convex subsets of H1 and H2, respectively, and A:H1H2 be a bounded linear operator. In this setting, the SFP [2] is formulated as finding a point with the requirement

xˆCandAxˆQ. 1.1

Denote the set of solutions for the SFP by Γ=CA1(Q), and define f:H1R by

f(x)=12(IPQ)Ax2,

we see the function f(x) is convex and continuously differentiable.

In [1, 3], Byrne proposed his CQ algorithm; it generates a sequence {xn} via the recursion

xn+1=PC(IτnA(IPQ)Axn),n0, 1.2

where initial x0H1 and τn(0,2/A2), PC and PQ are the orthogonal projections onto C and Q, respectively.

Considering that the projection onto the nonempty closed convex sets C and Q might be hard to be implemented, Yang [4] and Xu [5] showed that if C and Q are level sets of convex functions, it just needs to project onto half-spaces Cn and Qn, thus the so-called relaxed CQ algorithm is reduced to the following formula:

xn+1=PCn(IτnA(IPQn)Axn),n0. 1.3

In (1.3), the projections PCn and PQn have closed-form expressions, thus they are easily to be computed. Define fn:H1R by

fn(x)=12(IPQn)Ax2,

then the convex objective fn is also differentiable and has a Lipschitz gradient given by

fn(x)=A(IPQn)Ax.

Noting that when applying (1.2) and (1.3) to solve the practical problems such as signal processing and image reconstruction, which can be covered by the SFP, it is hard to avoid that a fixed stepsize related to the norm of A sometimes affects convergence of the algorithms. Therefore, in order not to compute the matrix inverse and the largest eigenvalue of the matrix ATA and have a sufficient decrease of the objective function at each iteration, people have invented various algorithms with a variable or self-adaptive stepsize. Since Qu [6] presented a searching method by adopting Armijo-like search, many similar methods have been proposed, such as [712] etc. However, through these methods, self-adaptive stepsize at each iteration can be achieved, most formats of them are becoming more complex, it is difficult to apply them to some practical problems, and this needs considerable time complexity, especially for the large-scale setting and sparse problem.

On the other hand, another way to construct the variable stepsize without calculating matrix norm was proposed by Yang in [13], that is,

τn:=ρnf(xn), 1.4

where {ρn} is a sequence of positive real numbers such that

n=0ρn=,n=0ρn2<.

at the same time, the additional conditions that Q is a bounded subset and A is a full column rank matrix are required. Wang et al. [14] applied (1.4) to solve the SFP. Afterwards, in order to remove the two additional conditions, López et al. [15] introduced another choice of the stepsize sequence {τn} as follows:

τn:=ρnfn(xn)fn(xn)2, 1.5

where {ρn} is chosen in the interval (0,4). Furthermore, in paper [16] we were aware that if fn(xn)=0 in (1.5) for some n1, the corresponding algorithms in [15] have to be terminated. In this case, xn may not be in C and is not necessarily a solution of SFP (1.1). With this observation, we have introduced a stepsize sequence {τn} as follows:

τn:=ρnfn(xn)(fn(xn)+ωn)2, 1.6

where {ρn} is also chosen in the interval (0,4) and {ωn} is a sequence of positive numbers in (0,1). This choice of the stepsize sequence makes the associated algorithms never terminate unless the solution of SFP has been found. However, there exists inconvenience dealing with the choice of the parameter ωn. Only when it is a small number, a similar convergence speed as adopting (1.5) can be guaranteed. After observing many experiments, the order of magnitude for ωn usually about less than 10−5 can satisfy that, it closely relates with the bit of computer and the float precision of the calculation software.

In order to improve these and avoid the calculations of fn(x) or fn(xn), in this paper, we firstly propose two simpler choices of a stepsize as follows:

τn:=ρnxnx¯n2AxnAx¯n2 1.7

and

τn:=ρnx¯n2Ax¯n2, 1.8

where {ρn} is chosen in the interval (0,2), xnx¯n, and x¯n0. The advantages of our choices (1.7) and (1.8) lie in facts that they not only possess simpler formats easily to be calculated and implemented in practice, but also have significantly faster convergence speed, especially in a large-scale setting and sparse problem. Secondly, we present two generalized variable stepsizes as follows:

τn:=ρnr(xn)2Ar(xn)2 1.9

or

τn:=ρnp(xn)2ATp(xn)2, 1.10

where {ρn} is chosen in the interval (0,2), and r(xn)0 and p(xn)0. Consequently, stepsizes (1.5)-(1.8) are the special cases of (1.9) or (1.10), and many similar stepsize formats can be obtained from them.

Recently, Yao et al. [17] have applied (1.5) to an improved CQ algorithm with a generalized Halpern iteration. In paper [18], we have modified the relative parameters with satisfactory conditions. Then, in this paper, we combine the iterations in [17, 18] with the KM-CQ iterations in [19, 20]. We propose two more general KM-CQ algorithms with the generalized variable stepsize (1.9) or (1.10), and they can be used to approach the minimum norm solution of the SFP that solves special variational inequalities.

The rest of this paper is organized as follows. Section 2 reviews some propositions and known lemmas. Section 3 gives two modified CQ algorithms with simpler variable stepsizes and shows the weak convergence. Section 4 presents two general KM-CQ algorithms with the generalized variable stepsizes and proves the strong convergence. In Section 5, we include numerical experiments to testify the better performance of the proposed stepsizes and algorithms with typical problems of signal processing and image restoration. Finally, Section 6 gives some conclusions and further research aim.

Preliminaries

Let H be a real Hilbert space with the inner product , and the norm , respectively. Let C be a nonempty closed convex subset of H. We will use the following notations:

  • ‘→’ stands for strong convergence;

  • ‘⇀’ stands for weak convergence;

  • I stands for the identity mapping on H.

Recall that a mapping T:CH is nonexpansive iff TxTyxy for all x,yC.

A mapping ψ:CH is said to be δ-contractive iff there exists a constant δ[0,1) such that

ψ(x)ψ(y)δxy

for all x,yC.

Recall also that the nearest point projection from H onto C, denoted by PC, assigns to each xH the unique point PCxC with the property xPCxxy, yC. We collect the basic properties of PC as follows.

Proposition 2.1

[6, 21]

(p1)

xPCx,yPCx0 for all xH, yC;

(p2)

PCxPCy2PCxPCy,xy for every x,yH;

(p3)

PCxPCy2xy2(IPC)x(IPC)y2 for all x,yH;

(p4)

(IPC)x(IPC)y,xy(IPC)x(IPC)y2 for all x,yH;

(p5)

PCxz2xz2PCxx2 for all xH, zC.

In a Hilbert space H, the next following facts are well known.

Proposition 2.2

x,yH, tR,

  • (i)

    x±y2=x2±2x,y+y2;

  • (ii)

    tx+(1t)y2=tx2+(1t)y2t(1t)xy2.

For the SFP, we assume that the following conditions are satisfied in a Hilbert space [5]:

(i) The solution set of the SFP is nonempty.

(ii) The set C is given by

C={xH1|c(x)0},

where c:H1R is a convex function and C is nonempty.

The set Q is given by

Q={yH2|q(y)0},

where q:H2R is a convex function and Q is nonempty.

(iii) Assume that both c and q are subdifferentiable on H1 and H2, respectively. The subdifferentials are defined as follows:

c(x)={ξH1|c(z)c(x)+ξ,zx for all zH1},

for all xC, and

q(y)={ηH2|q(u)q(y)+η,uy for all uH2},

for all yQ.

(iv) We also assume that ∂c and ∂q are bounded on bounded sets. The set Cn is given by

Cn={xH1|c(xn)+ξn,xxn0},

where ξnc(xn). The set Qn is constructed by

Qn={yH2|q(Axn)+ηn,yAxn0},

where ηnq(Axn).

It is easily seen that CCn and QQn for all n.

The relaxed CQ algorithm [4] can be seen as a special case of the classical gradient projection method (GPM). To see this, we can consider the following convex minimization problem:

minxCnfn(x). 2.1

It is well known that xˆCn is a solution of problem (2.1) if and only if

fn(xˆ),xxˆ0,xCn. 2.2

Also, we know that (2.2) holds true if and only if

xˆ=PCn(Iτfn)xˆ,τ>0.

Therefore, we use the GPM to solve the SFP, for any x0H1,

xn+1=PCn[xnτnfn(xn)],n0, 2.3

where τn(0,2/L), while L is the Lipschitz constant of fn. Noting that L=A2, we see that (2.3) is exactly the relaxed CQ algorithm (1.3) .

Proposition 2.3

see [22]

Let {αn} be a sequence of nonnegative real numbers such that

αn+1(1tn)αn+tnbn,n1,

where {tn} is a sequence in (0,1) and bnR such that

  • (i)

    n=1tn=;

  • (ii)

    limnbn0 or n=1|tnbn|<.

Then αn0.

Proposition 2.4

[23, 24]

For some countable index set Λ, we denote by p=p(Λ), 1p. Soft-thresholding leads to the unique minimizer of a functional combining 2 and l1-norms,

Sμ(a)=argminx2(Λ)(xa2+2μx1), 2.4

where μ is a certain positive number and Sμ is the soft-thresholding operation defined by Sμ(a)i=Sμ(ai), iΛ, with

Sμ(z)={zμ,z>μ,0,|z|μ,z+μ,z<μ.

For the 1-ball BR={x2:x1R} with R:=x1, where x2 is the solution of problem (2.4), we replace the thresholding with the projection PBR, and with a slight abuse of notation, we denote PBR by PR; then we introduce two properties of 2-projections onto 1-balls [25].

Lemma 2.1

For any a2 and for μ>0, Sμ(a)1 is a piecewise linear, continuous, decreasing function of μ; moreover, if a1, then S0(a)1=a1 and Sμ(a)1=0 for μmaxi|ai|.

Lemma 2.2

If a1>R, then the 2 projection of a on the 1-ball with radius R is given by PR(a)=Sμ(a), where μ (depending on a and R) is chosen such that Sμ(a)1=R. If a1R, then PR(a)=S0(a)=a.

Next, we discuss a method to compute μ.

Proposition 2.5

[25]

For any aΩ2, dim(Ω)=n, sort the absolute values of the components of a by descending order, obtaining the rearranged sequence (ai)i=1,,n. Then we perform a search to find k such that

Sak(a)1=i=1k1(aiak)R<i=1k(aiak+1)=Sak+1(a)1

or equivalently,

Sak(a)1=i=1k1i(aiai+1)R<i=1ki(aiai+1)=Sak+1(a)1.

Set ν:=k1(RSak(a)1), and μ:=ak+ν, then

Sμ(a)1=iΩmax(|ai|μ,0)=i=1k(aiμ)=i=1k1(aiak)+kν=Sak(a)1+kν=R.

CQ algorithms with two simpler variable stepsizes

In this section, two simpler variable stepsizes are proposed below. The advantages of the two stepsizes, comparing with (1.5) and (1.6), are that neither prior information about the matrix norm A nor any other conditions on Q and A are required.

A simpler variable stepsize for CQ algorithm

We propose a new and simpler variable stepsize method for solving the feasibility problem. The algorithm is presented as follows.

Algorithm 3.1

For any initial data x0H1, uH1 and u0, assume that the nth iterate xn has been constructed; then we compute the (n+1) th iterate xn+1 via the formula

x¯n=PCn(tnu+(1tn)xn), 3.1
xn+1=PCn(xnτnfn(xn)), 3.2

where

τn:=ρnxnx¯n2AxnAx¯n2, 3.3

with {ρn}(0,2) and {tn}(0,1). If xn+1=xn or Axn=Ax¯n for some n0, then xn is a solution of SFP (1.1) and the iteration stops; otherwise, we set n:=n+1 and go to (3.1) to compute the next iterate xn+2.

Remark 3.1

We can easily approximate the upper bound λ of the eigenvalue interval to the symmetric matrix ATA from [1, 3], thus for any xn0, we can obtain τn(0,2/λ)(0,2/L), where L is the largest eigenvalue of ATA.

Now we prove the convergence property of Algorithm 3.1.

Theorem 3.1

If Γ and lim_nτn(2λτn)σ>0, the sequence {xn} generated by Algorithm  3.1 converges weakly to a solution of SFP (1.1).

Proof

Let x be a solution of SFP, since CCn, QQn, thus x=PC(x)=PCn(x) and Ax=PQ(Ax)=PQn(Ax). It shows that xCn and fn(x)=0 for all n=0,1, , using (3.2) and (p4), we have

xn+1x2=PCn(xnτnfn(xn))x2(xnx)τnfn(xn)2=xnx2+τn2fn(xn)22τnxnx,fn(xn)=xnx2+τn2fn(xn)22τnxnx,fn(xn)fn(x)xnx2+τn2L(IPQn)Axn22τnAxnAx,(IPQn)Axn(IPQn)Axxnx2+τn2L(IPQn)Axn22τn(IPQn)Axn(IPQn)Ax2=xnx2+(τn2L2τn)(IPQn)Axn2. 3.4

Combining with Remark 3.1, we know τn2L2τn<0, thus it implies that the sequence {xnx2} is monotonically decreasing and hence {xn} is bounded. Consequently, from (3.4) we get

limn(IPQn)Axn2=0. 3.5

Assume that is an accumulation point of {xn} and {xni}x¯, {xni} is a subsequence of {xn}. Then from (3.5) it follows

limni(IPQni)Axni2=0. 3.6

Next we show x¯Γ.

Firstly, we show x¯C. We prove it from two cases.

Case 1. limnixni+1x¯. Without loss of generality, we may assume limnixni+1=x˜x¯. Set zni=xniτnifni(xni), so xni+1=PCni(zni) is the solution of the following programming:

min12zzni2s.t. c(xni)+ξni,zxni0.

By the Kuhn-Tucker condition, there exists a nonnegative number νni such that

xni+1zni+νniξni=0, 3.7
νni(c(xni)+ξni,xni+1xni)=0. 3.8

If there exist infinite ni such that νni=0 or ξni=0, from (3.7) and (3.8) it leads to x˜=x¯, so the contradiction happens. Therefore, νni>0 and ξni0 for sufficiently large ni. We go on to divide the discussion into two cases.

(1) If inf{ξni}>0, we may assume ξniξ¯0. Then ξ¯c(x¯) by the lower semicontinuity of c(x). From (3.7) we have νniξ¯,x˜x¯/ξ¯2=Δν¯. Thus we obtain from (3.7) and (3.8)

x˜x¯+ν¯ξ¯=0, 3.9
ν¯(c(x¯)+ξ¯,x˜x¯)=0. 3.10

They mean that x˜=PC(x¯)(x¯).

Therefore, x¯PC(x¯)(x¯) by assumption. From (p5) we have

x˜x2x¯x2x¯x˜2<x¯x2.

Since {xnx2} is decreasing, both and are the accumulation points of {xn}, then x˜x2=x¯x2. It is a contradiction, so this case will not occur.

(2) If inf{ξni}=0, then we get the lower semicontinuity of c(x). Since c(x) is convex, then is a minimizer of c(x) over H1. Since c(x)0, then c(x¯)c(x)0. So x¯C.

Case 2. limnixni+1=x¯. As in Case 1, one has (3.7) and (3.8) by the Kuhn-Tucker condition. If there exist infinite ni s.t. νni=0 or ξni=0, then we have PCni(zni)=zni, so c(xni)+ξni,znixni0. Since

limni(znixni)=limniτniAT(PQniI)Axni=0,

then c(x¯)0. Therefore, we have x¯C.

Assume νni>0 and ξni0 for sufficiently large ni. If inf{ξni}=0, such as above, it follows x¯C. If inf{ξni}>0, similar to Case 1(1), it leads to x¯=PC(x¯)(x¯), which implies c(x¯)+ξ¯,x¯x¯0. So x¯C.

In summary, we can conclude x¯C.

Secondly, we need to show Ax¯Q. From (3.5) we have

limni(IPQni)Axni2=0. 3.11

Since PQni(Axni)Qni, we have

q(Axni)+ηni,PQni(Axni)Axni0.

Moreover, limiting the inequality and taking account of (3.11), we obtain that

q(Ax¯)0,

that is, Ax¯Q.

Therefore is a solution of SFP. Thus we may replace x in (3.4) with , and get {xnx¯} is convergent. Because there exists a subsequence {xnix¯} convergent to 0, then xnx¯ as n. □

The other simpler variable stepsize for CQ algorithm

In this part, we introduce the other simpler choice of the stepsize τn, which also is a variable stepsize to CQ algorithm. Either combining with the relaxed CQ algorithm [4], we have the next algorithm.

Algorithm 3.2

Choose the initial data x0H1, for uH1 and u0. Assume that the nth iterate xn has been constructed; then we compute the (n+1) th iterate xn+1 via the formula

x¯n=PCn(tnu+(1tn)xn), 3.12
xn+1=PCn(xnτnfn(xn)), 3.13

where

τn:=ρnx¯n2Ax¯n2, 3.14

with {ρn}(0,2) and {tn}(0,1). If xn+1=xn, then stop and xn is a solution of SFP (1.1); otherwise, back to (3.12) and continue to compute xn+2.

Obviously, (3.14) is also consistent with Remark 3.1. Thus, similar to the proof of Theorem 3.1, we can deduce that Algorithm 3.2 converges weakly to a solution of SFP (1.1).

Two general KM-CQ algorithms with generalized variable stepsize

In this section, we integrate the variable stepsizes from (1.5) to (1.8) and obtain a variable stepsize that can cover them. After that, we apply it to improve the algorithms presented in [17] and [18] and construct two algorithms for approximating some solution of (1.1).

Let ψ:CH1 be a δ-contraction with δ(0,1), let r:H1H1Θ and q:H1H2Θ be nonzero operators, where Θ denotes the zero point.

A generalized variable stepsize for a general KM-CQ algorithm

The next recursion not only possesses a more generalized adaptive descent step, but it also can be implemented easily by the relaxed method.

Algorithm 4.1

Choose the initial data x0H1 arbitrarily. Assume that the nth iterate xn has been constructed; then we compute the (n+1) th iterate xn+1 via the formula

xn+1=(1βn)xn+βnPCn[αnψ(xn)+(1αn)Unxn], 4.1

where Unxn=(IτnAT(IPQn)A)xn, τn:=ρnr(xn)2Ar(xn)2 or τn:=ρnp(xn)2ATp(xn)2, {ρn}(0,2), {tn}(0,1), {αn} and {βn} are two real sequences in [0,1]. If xn+1=xn for some n0, then xn is a solution of SFP (1.1) and the iteration stops; otherwise, continue to compute xn+2.

Theorem 4.1

Suppose that the SFP is consistent, that is, Γ=CA1(Q), lim_nτn(2λτn)σ>0. Assume that the sequences {αn} and {βn} satisfy the following conditions:

(C1)

limnαn=0 and n=1αn=;

(C2)

0<lim_nβn.

Then {xn} defined by (4.1) converges strongly to x=PΓψx, which solves the following variational inequality:

(ψI)x,yx0,yΓ. 4.2

Proof

Since PΓ:H1ΓC is nonexpansive and ψ:CH1 is δ-contractive, therefore, we have PΓψ:CC is a contraction with δ(0,1). By the Banach contractive mapping principle, there exists a unique xC such that x=PΓψx. By virtue of (p1), we see that (4.2) holds true.

By virtue of x being a solution of the SFP, xCA1(Q), and CCn, QQn, then x=PC(x)=PCn(x) and Ax=PQ(Ax)=PQn(Ax). From (p4) we have

Unxnx2=xnτnfn(xn)x2=xnx2+τnfn(xn)22τnxnx,fn(xn)=xnx2+τn2fn(xn)22τnAxnAx,(IPQn)Axn(IPQn)Axxnx2+τn2L(IPQn)Axn22τn(IPQn)Axn2=xnx2+(τn2L2τn)(IPQn)Axn2. 4.3

Obviously, τn(0,2/L) from Remark 3.1, then τn2L2τn<0, in particular, we obtain

Unxnxxnx. 4.4

At this point, we can establish the boundedness of {xn}. To see this, using (4.1) we have

xn+1x=(1βn)xn+βnPCn[αnψ(xn)+(1αn)Unxn]x(1βn)xnx+βnPCn[αnψ(xn)+(1αn)Unxn]x(1βn)xnx+βnαnψ(xn)+(1αn)Unxnx=(1βn)xnx+βnαn(ψ(xn)ψ(x))+(1αn)(Unxnx)+αnψ(x)αnx(1βn)xnx+αnβnψ(xn)ψ(x)+(1αn)βnUnxnx+αnβnψ(x)x(1(1δ)αnβn)xnx+αnβnψ(x)x=(1(1δ)αnβn)xnx+(1δ)αnβnψ(x)x1δmax{x0x,ψ(x)x1δ}=M,

for all n0, which indicates {xn} is bounded. Set zn=PCn[αnψ(xn)+(1αn)Unxn], thus {zn} is also bounded.

Next, we prove xnx (n). By virtue of (4.1), Proposition 2.1(p2) and (4.4), we have

xn+1x2=(1βn)xn+βnPCn[αnψ(xn)+(1αn)Unxn]x2(1βn)xnx2+βnPCn[αnψ(xn)+(1αn)Unxn]x2, 4.5

where we set

PCn[αnψ(xn)+(1αn)Unxn]x2=PCn[wn]x2=PCn[wn]x,PCn[wn]x=PCn[wn]wn,PCn[wn]x+wnx,PCn[wn]x,

since PCn[wn]wn,PCn[wn]x0, we have

PCn[wn]x2wnx,PCn[wn]x=αnψ(xn)+(1αn)Unxnx,PCn[wn]x=αn(ψ(xn)ψ(x))+(1αn)(Unxnx)+αn(ψ(x)x),PCn[wn]x(αnψ(xn)ψ(x)+(1αn)Unxnx)PCn[wn]x+αnψ(x)x,PCn[wn]x(1(1δ)αn)xnxPCn[wn]x+αnψ(x)x,PCn[wn]x1(1δ)αn2xnx2+12PCn[wn]x2+αnψ(x)x,PCn[wn]x.

Therefore,

PCn[wn]x2(1(1δ)αn)xnx2+2αnψ(x)x,PCn[wn]x. 4.6

Substituting (4.6) into (4.5) can yield

xn+1x2(1βn)xnx2+(1(1δ)αn)βnxnx2+2αnβnψ(x)x,PCn[wn]x(1(1δ)αnβn)xnx2+(1δ)αnβn21δψ(x)x,PCn[wn]x. 4.7

Since xCCn, PCn:H1CCn and ψ:CCnH1, then PCnψ:CnCn, x=PCnψx.

Due to the property of the projection (p1) in Proposition 2.1,

lim supnψ(x)x,PCn[wn]x=maxPCn[wn]Cnψ(x)PCnψ(x),PCn[wn]PCnψ(x)0. 4.8

Applying (4.8) and Proposition 2.3 to (4.7), we deduce that xnx.

Assume that is an accumulation point of {xn} and xnixˆ, where {xni}i=1 is a subsequence of {xn}. Next we will prove that is a solution of SFP.

As xnx, we know xnx, that is,

limnxn+1xn=0. 4.9

Therefore, limit (4.4), we can obtain

limnUn(xn)x=0.

Then, limit (4.3), we get

limn(IPQn)Axn=0. 4.10

On the one hand, we show xˆC.

Notice that xnixˆ and xni+1xni0 (i). Since xni+1Cni, then by virtue of the definition of Cni, we have

c(xni)+ξni,xni+1xni0,i=1,2,,

taking the limit and using (4.9), we obtain that

c(xˆ)0.

Hence, we get xˆC.

On the other hand, we need to show AxˆQ.

Since PQni(Axni)Qni, we have

q(Axni)+ηni,PQni(Axni)Axni0,

taking ni, by virtue of (4.10), we deduce that

q(Axˆ)0,

that is, AxˆQ.

Therefore, is a solution of SFP.

Thus we may replace x in (4.7) with and get {xnxˆ} is convergent. Because there exists a subsequence {xnixˆ} convergent to 0, then xnxˆ as n. □

The other extended algorithm

Let h:CH1 be a κ-contraction. Let B:H1H1 be a self-adjoint strongly positive bounded linear operator with coefficient λ(0,1), for xH1, there exists Bx,xλx2. Take a constant σ such that 0<σκ<λ.

As B is self-adjoint, we haveB=supx=1Bx,x. IB is also self-adjoint, then

IB=supx=1(IB)x,x=supx=1{x2Bx,x}supx=1{(1λ)x2}1λ.

In (4.1), we set ψ(x)=σh(x)+(IB)Unx, thus

ψ(x)ψ(y)σκxy+IBxy(σκ+1λ)xy,

for x,yH1, we know that σκ+1λ(0,1), ψ:CH1 is still a contraction. Accordingly, we have the following extended algorithm that is a special case of Algorithm 4.1.

Algorithm 4.2

Choose the initial data x0H1 arbitrarily. Assume that the nth iterate xn has been constructed; then we compute the (n+1) th iterate xn+1 via the formula

xn+1=(1βn)xn+βnPCn[αnσh(xn)+(IαnB)Unxn], 4.11

where Unxn=(IτnAT(IPQn)A)xn, τn:=ρnr(xn)2Ar(xn)2 or τn:=ρnp(xn)2ATp(xn)2, τn(0,2/L), {ρn}(0,2), {tn}(0,1), {αn} and {βn} are two real sequences in [0,1]. If xn+1=xn, then stop and xn is a solution of SFP (1.1); otherwise, continue to compute xn+2.

Theorem 4.2

Suppose that the SFP is consistent, that is, Γ=CA1(Q), lim_nτn(2λτn)σ>0, assume that the sequences {αn} and {βn} satisfy the following conditions:

(C1)

limnαn=0 and n=1αn=;

(C2)

0<lim_nβn.

Then {xn} defined by (4.11) converges strongly to x=PΓ[σh(x)+(IB)Unx], which solves the following variational inequality:

σh(x)B(x),yx0,yΓ.

Numerical experiments and results

This section considers two numerical experiments to illustrate the performance of the above proposed variable stepsizes in CQ algorithm. Firstly, we see that a great amount of problems in signal and image processing can be seen as estimating xRN from the linear observation model

y=Ax+ε, 5.1

where yRM is the observed or measured data with noisy ε. A:RNRM denotes the bounded linear observation or measurement operator. Sometimes, the range of A may not be closed in most inverse problems, therefore, if A is ill-conditioned, the problem will be ill-posed.

If x is a sparse expansion, finding the solutions of (5.1) can be seen as finding a solution to the least-square problem

minxRN12yAx2subject to x1<t 5.2

for any real number t>0.

Problem (5.2) is a particular case of SFP (1.1) where C={xRN:x1t} and Q={y}, i.e., find x1t such that Ax=y. Therefore, CQ algorithm can be applied to solve (5.2). From Propositions 2.4 and 2.5 the projection onto C can be easily computed [15], while Lemmas 2.1 and 2.2 show the special situation of Proposition 2.4.

Next, following the experiments in [15, 26], we choose two particular problems, i.e., the compressed sensing and the image deconvolution, which are covered by (5.1). The experiments compare the performances of the proposed stepsizes of the CQ algorithm in this paper with the stepsizes in [15] and [16].

Compressed sensing

We consider a typical compressed sensing model, where a sparse signal recovery problem with a signal xRN, and N=212. This original signal x contains only m=50 spikes with amplitude ±1, and the spikes are located at random, see the top of Figure 1. x is being reconstructed from M=210 measurements, thus A is a M×N matrix randomly obtained with independent samples of a orthonormalized standard Gaussian distribution, and the noisy ε is with variance σε2=104. To (5.2), we also set t=50.

Figure 1.

Figure 1

Compressed sensing problem, from top to bottom: original signal, results of CQ algorithm with stepsizes ( 1.5 ), ( 1.6 ), ( 3.3 ) and ( 3.14 ), the last is KM-CQ algorithm with stepsize ( 3.3 ).

For stepsizes (1.5) and (1.6) in [15] and [16], respectively, we still consider the stepsize with constant ρ=2, ωn=(n+2)5, while for stepsizes (3.3) and (3.14), we set ρ=1. For (3.1) and (3.12), we set t=0.1 and u=rand(N,1)(0,1). For (4.1) we use (3.1) and (3.3), and set αn=(n+2)1, βn=1(n+2)1 and ψ0 as its special case. All the processes are started with the initial signal x0=0 and finished with the stop rule

xn+1xn/xn<103.

We also calculated the mean squared error (MSE) for the results

MSE=(1/N)xx,

where x is an estimated signal of x.

The simulated results of different algorithms with different steps can be seen in Figure 1 and Table 1. Algorithms 3.1 and 3.2 have less iteration steps and smaller MSE, especially for Algorithm 3.1. Thus, we see that stepsizes (3.3) and (3.14) not only have simper formats than before, but also can make CQ algorithms have faster iteration and better restored precision.

Table 1.

Results of different stepsizes and algorithms for Figure  1

Algorithms MSE n CPU time (s)
CQ with (1.5) 2.8021e−005 66 0.8504
CQ with (1.6) 2.844e−005 66 0.8579
3.1 5.1714e−006 25 0.6858
3.2 1.5253e−006 64 1.5685

Image deconvolution

In this subsection, we apply the CQ algorithms in the paper to image deconvolution. The observation model can also be described as (5.1), we wish to estimate an original image x from an observation y, while matrix A represents the observation operator, and ε is a sample of a zero-mean white Gaussian field of variance σ. For the 2D image deconvolution problem, A is a block-circulant matrix with circulant blocks [27]. We stress that the goal of these experiments is not to assess the restored precision of the algorithms, but to apply the algorithms in paper to solve this particular SFP, then compare the iterative speed and restored precision of the proposed stepsizes against the CQ algorithms.

According to papers [26, 27], we also take the well-known Cameraman image. In the experiments, we employ Haar wavelets, and the blur point spread functions are uniform blur with size 9×9, hij=(1+i2+j2)1, for i,j=4,,4 and for i,j=7,,7. The noise variance is σ2=0.308, 2 and 8, respectively. We have N=M=2562, then the block-circulant matrix A can be constructed by the blur point spread functions, and A may be very ill-conditioned. Set all the threshold values μ=0.25, t is the sum of all the pixel values in the original image. Moreover, we use y=IFFT(FFT(A).FFT(x))+ε to obtain the observation, where FFT is the fast Fourier transform, IFFT is the inverse fast Fourier transform. Other settings in the above stepsizes and algorithms are the same as in 4.1. We set the initial image x0=0 and also follow the stop rule xn+1xn/xn<103.

The results of iteration steps, CPU time and the SNR improvements are presented in Table 2. It also testifies that the proposed stepsizes and algorithms in this paper can give better performance.

Table 2.

Results for the restorations of different stepsizes and algorithms

Blur kernel σ2 Algorithms SNR (dB) n CPU time (s)
9 × 9 uniform 0.308 CQ with (1.5) 16.1802 52 3.3883
CQ with (1.6) 16.1722 48 3.1636
3.1 14.5266 19 1.3847
3.2 14.5265 19 1.3199
4.1 with (3.3) 14.2464 34 2.4280
hij=(1+i2+j2)1 for i,j = −4,…,4 2 CQ with (1.6) 22.6184 19 1.3007
CQ with (1.5) 22.4329 17 1.1718
3.1 19.8401 14 1.0746
3.2 19.8401 14 1.0266
4.1 with (3.3) 19.4912 33 2.3306
hij=(1+i2+j2)1 for i,j = −7,…,7 8 CQ with (1.6) 12.7305 39 2.6739
CQ with (1.5) 14.6363 42 2.8604
3.1 19.3561 18 1.3390
3.2 19.3560 18 1.3057
4.1 with (3.3) 18.6140 33 2.3628

Conclusions and discussion

In this paper, we have proposed two simpler variable stepsizes for the CQ algorithm. Compared with the other related variable stepsizes, they also need not to compute the largest eigenvalue of A and can be calculated easily. Furthermore, we also presented a more general KM-CQ algorithm with generalized variable stepsizes. As a special case, we deduced another general format. Obviously, both the general algorithms with the generalized variable stepsizes can solve the SFP and some special variational inequality problem better. The corresponding weak and strong convergence properties have also been established. In the experiments, through the compressed sensing and image deconvolution models, we compare the proposed stepsizes with the former ones, the results obtained from the proposed stepsizes and algorithms appear to be significantly better.

We should notice that the values of parameter ρn are fixed in the above experiments. Actually, a different value of ρn can also affect the convergence speed of the algorithms. Therefore, our future work is to find the method to choose a self-adaptive sequence {ρn}.

Acknowledgements

The authors would like to thank the associate editor and the referees for their comments and suggestions. The research was supported by Professor Haiyun Zhou.

Footnotes

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

All authors took equal roles in deriving results and writing of this paper. All authors have read and approved the final manuscript.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Byrne C. Iterative oblique projection onto convex sets and the split feasibility problem. Inverse Probl. 2002;18:441–453. doi: 10.1088/0266-5611/18/2/310. [DOI] [Google Scholar]
  • 2.Censor Y, Elfving T. A multiprojection algorithm using Bregman projections in a product space. Numer. Algorithms. 1994;8:221–239. doi: 10.1007/BF02142692. [DOI] [Google Scholar]
  • 3.Byrne CL. A unified treatment of some iterative algorithms in signal processing and image reconstruction. Inverse Probl. 2004;20:103–120. doi: 10.1088/0266-5611/20/1/006. [DOI] [Google Scholar]
  • 4.Yang QZ. The relaxed CQ algorithm solving the split feasibility problem. Inverse Probl. 2004;20:1261–1266. doi: 10.1088/0266-5611/20/4/014. [DOI] [Google Scholar]
  • 5.Xu HK. Iterative methods for the split feasibility problem in infinite-dimensional Hilbert spaces. Inverse Probl. 2010;26:1–17. doi: 10.1088/0266-5611/26/10/105018. [DOI] [Google Scholar]
  • 6.Qu B, Xiu N. A note on the CQ algorithm for the split feasibility problem. Inverse Probl. 2005;21:1655–1665. doi: 10.1088/0266-5611/21/5/009. [DOI] [Google Scholar]
  • 7.Qu B, Xiu N. A new halfspace-relaxation projection method for the split feasibility problem. Linear Algebra Appl. 2008;428:1218–1229. doi: 10.1016/j.laa.2007.03.002. [DOI] [Google Scholar]
  • 8.Wang W, Gao Y. A modified algorithm for solving the split feasibility problem. Int. Math. Forum. 2009;4:1389–1396. [Google Scholar]
  • 9.Wang Z, Yang Q. The relaxed inexact projection methods for the split feasibility problem. Appl. Math. Comput. 2011;217:5347–5359. [Google Scholar]
  • 10.Li M. Improved relaxed cq methods for solving the split feasibility problem, advanced modeling and optimization. Appl. Math. Comput. 2011;13:305–317. [Google Scholar]
  • 11.Abdellah B, Muhammad AN. On descent-projection method for solving the split feasibility problems. J. Glob. Optim. 2012;54:627–639. doi: 10.1007/s10898-011-9782-2. [DOI] [Google Scholar]
  • 12.Dong Q, Yao Y, He S. Weak convergence theorems of the modified relaxed projection algorithms for the split feasibility problem in Hilbert spaces. Optim. Lett. 2014;8:1031–1046. doi: 10.1007/s11590-013-0619-4. [DOI] [Google Scholar]
  • 13.Yang Q. On variable-step relaxed projection algorithm for variational inequalities. J. Math. Anal. Appl. 2005;302:166–179. doi: 10.1016/j.jmaa.2004.07.048. [DOI] [Google Scholar]
  • 14.Wang F, Xu H, Su M. Choices of variable of the steps of the cq algorithm for the split feasibility problem. Fixed Point Theory. 2011;12:489–496. [Google Scholar]
  • 15.López G, Martín-Máquez V, Wang F. Solving the split feasibility problem without prior knowledge of matrix norms. Inverse Probl. 2012;28:1–18. doi: 10.1088/0266-5611/28/8/085004. [DOI] [Google Scholar]
  • 16.Zhou HY, Wang PY. Adaptively relaxed algorithms for solving the split feasibility problem with a new stepsize. J. Inequal. Appl. 2014;2014 doi: 10.1186/1029-242X-2014-448. [DOI] [Google Scholar]
  • 17.Yao Y, Postolache M, Liou Y. Strong convergence of a self adaptive method for the split feasibility problem. Fixed Point Theory Appl. 2013;2013 doi: 10.1186/1687-1812-2013-201. [DOI] [Google Scholar]
  • 18.Zhou HY, Wang PY. Some remarks on the paper “strong convergence of a self-adaptive method for the split feasibility problem”. Numer. Algorithms. 2015;70:333–339. doi: 10.1007/s11075-014-9949-2. [DOI] [Google Scholar]
  • 19.Dang Y, Gao Y. The strong convergence of a km-cq-like algorithm for a split feasibility problem. Inverse Probl. 2011;27 doi: 10.1088/0266-5611/27/1/015007. [DOI] [Google Scholar]
  • 20.Yu X, Shahzad N, Yao Y. Implicit and explicit algorithms for solving the split feasibility problem. Optim. Lett. 2012;6:1447–1462. doi: 10.1007/s11590-011-0340-0. [DOI] [Google Scholar]
  • 21.Byrne C. A unified treatment of some iterative algorithms in signal processing and image reconstruction. Inverse Probl. 2004;20:103–120. doi: 10.1088/0266-5611/20/1/006. [DOI] [Google Scholar]
  • 22.Xu H. Iterative algorithms for nonlinear operators. J. Lond. Math. Soc. 2002;66:240–256. doi: 10.1112/S0024610702003332. [DOI] [Google Scholar]
  • 23.Deubechies I, Defrise M, DeMol C. An iterative thresholding algorithm for linear inverse problems. Commun. Pure Appl. Math. 2004;57:1413–1457. doi: 10.1002/cpa.20042. [DOI] [Google Scholar]
  • 24.Chambolle A, DeVore R, Lee N. Nonlinear wavelet image processing: variational problems, compression, and noise removal through wavelet shrinkage. IEEE Trans. Image Process. 1998;7:319–335. doi: 10.1109/83.661182. [DOI] [PubMed] [Google Scholar]
  • 25.Daubechies I, Fornasier M, Loris I. Accelerated projected gradient method for linear inverse problems with sparsity constraints. J. Fourier Anal. Appl. 2008;14:764–792. doi: 10.1007/s00041-008-9039-8. [DOI] [Google Scholar]
  • 26.Figueiredo M, Nowak R, Wright S. A gradient projection for sparse reconstruction: application to compressed sensing and other inverse problems. IEEE J. Sel. Top. Signal Process. 2007;1:586–598. doi: 10.1109/JSTSP.2007.910281. [DOI] [Google Scholar]
  • 27.Figueiredo M, Nowak R. IEEE International Conference on Image Processing - ICIP’2005. 2005. A bound optimization approach to wavelet-based image deconvolution; pp. 1–4. [Google Scholar]

Articles from Journal of Inequalities and Applications are provided here courtesy of Springer

RESOURCES