Abstract
The split feasibility problem (SFP) is finding a point such that , where C and Q are nonempty closed convex subsets of Hilbert spaces and , and is a bounded linear operator. Byrne’s CQ algorithm is an effective algorithm to solve the SFP, but it needs to compute , and sometimes is difficult to work out. López introduced a choice of stepsize , , . However, he only obtained weak convergence theorems. In order to overcome the drawbacks, in this paper, we first provide a regularized CQ algorithm without computing to find the minimum-norm solution of the SFP and then obtain a strong convergence theorem.
Keywords: split feasibility problem, regularized CQ algorithm, minimum-norm solution, strong convergence, operator norm
Introduction
Let and be real Hilbert spaces and let C and Q be nonempty closed convex subsets of and , and let be a bounded linear operator. Let and denote the sets of positive integers and real numbers.
In 1994, Censor and Elfving [1] came up with the split feasibility problem (SFP) in finite-dimensional Hilbert spaces. In infinite-dimensional Hilbert spaces, it can be formulated as
| 1.1 |
where C and Q are nonempty closed convex subsets of and , and is a bounded linear operator. Suppose that SFP (1.1) is solvable, and let S denote its solution set. The SFP is widely applied to signal processing, image reconstruction and biomedical engineering [2–4].
So far, some authors have studied SFP (1.1) [5–17]. Others have also found a lot of algorithms to study the split equality fixed point problem and the minimization problem [18–20]. Byrne’s CQ algorithm is an effective method to solve SFP (1.1). A sequence , generated by the formula
| 1.2 |
where the parameters , , and , is a set of orthogonal projections.
As is well-known, Cencor and Elfving’s algorithm needs to compute , and Byrne’s CQ algorithm needs to compute . However, they are difficult to calculate.
Consider the following convex minimization problem:
| 1.3 |
where
| 1.4 |
| 1.5 |
is differentiable and the gradient ∇f is L-Lipschitz with .
The gradient-projection algorithm [21] is the most effective method to solve (1.3). A sequence is generated by the recursive formula
| 1.6 |
where the parameter . Then we know that Byrne’s CQ algorithm is a special case of the gradient-projection algorithm.
In Byrne’s CQ algorithm, depends on the operator norm . However, it is difficult to compute. In 2005, Yang [22] considered as follows:
where and satisfies
In 2012, López [23] introduced as follows:
where . However, López’s algorithm only has weak convergence.
In 2013, Yao [24] introduced a self-adaptive method for the SFP and obtained a strong convergence theorem. However, the algorithm is difficult to work out.
In general, there are two types of algorithms to solve SFPs. One is the algorithm which depends on the norm of the operator. The other is the algorithm without a priori knowledge of the operator norm. The first type of algorithm needs to calculate , but is not easy to work out. The second type of algorithm also has a drawback. It always has weak convergence. If we want to obtain strong convergence, we have to use the composited iterative method, but then the algorithm is difficult to calculate. In order to overcome the drawbacks, we propose a new regularized CQ algorithm without a priori knowledge of the operator norm to solve the SFP and we obtain a strong convergence theorem.
Consider the following regularized minimization problem:
| 1.7 |
where the regularization parameter . A sequence is generated by the formula
| 1.8 |
where , , and , . Then, under suitable conditions, the sequence generated by (1.8) converges strongly to a point , where is the minimum-norm solution of SFP (1.1).
Preliminaries
In this part, we introduce some lemmas and some properties that are used in the rest of the paper. Throughout this paper, let and be real Hilbert spaces, be a bounded linear operator and I be the identity operator on or . If is a differentiable functional, then the gradient of f is denoted by ∇f. We use the sign ‘→’ to denote strong convergence and use the sign ‘⇀’ to denote weak convergence.
Definition 2.1
See [25]
Let D be a nonempty subset of H, and let . Then T is firmly nonexpansive if
Lemma 2.2
See [26]
Let be an operator. Then the following are equivalent:
-
(i)
T is firmly nonexpansive,
-
(ii)
is firmly nonexpansive,
-
(iii)
is nonexpansive,
-
(iv)
, ,
-
(v)
.
Recall is an orthogonal projection, where C is a nonempty closed convex subset of H. Then to each point , the unique point satisfies the following property:
also has the following characteristics.
Lemma 2.3
See [27]
For a given ,
-
(i)
, ,
-
(ii)
, ,
-
(iii)
, .
Lemma 2.4
See [28]
Let f be given by (1.4). Then
-
(i)
f is convex and differential,
-
(ii)
, ,
-
(iii)
f is w-lsc on H,
-
(iv)
∇f is -Lipschitz: , .
Lemma 2.5
See [29]
Let be a sequence of nonnegative real numbers such that
where is a sequence in and is a sequence in such that
-
(i)
,
-
(ii)
or .
Then .
Lemma 2.6
See [30]
Let be a sequence of real numbers such that there exists a subsequence of such that for all . Then there exists a nondecreasing sequence of such that and the following properties are satisfied by all (sufficiently large) numbers :
In fact, is the largest number n in the set such that the condition
holds.
Main results
In this paper, we always assume that is a real-valued convex function, where , the gradient , C and Q are nonempty closed convex subsets of real Hilbert spaces and , and is a bounded linear operator.
Algorithm 3.1
Choose an initial guess arbitrarily. Assume that the nth iterate has been constructed and . Then we calculate the th iterate via the formula
| 3.1 |
where is chosen as follows:
with . If , then is a solution of SFP (1.1) and the iterative process stops. Otherwise, we set and go to (3.1) to evaluate the next iterate .
Theorem 3.1
Suppose that and the parameters and satisfy the following conditions:
-
(i)
, , ,
-
(ii)
for some small enough.
Then the sequence generated by Algorithm 3.1 converges strongly to , where .
Proof
Let . Since minimization is an exactly fixed point of its projection mapping, we have and .
By (3.1) and the nonexpansivity of , we derive
| 3.2 |
Since is firmly nonexpansive, from Lemma 2.2, we deduce that is also firmly nonexpansive. Hence, we have
| 3.3 |
Note that . From (3.3), we obtain
| 3.4 |
By condition (ii), without loss of generality, we assume that for all . Thus from (3.2) and (3.4), we obtain
| 3.5 |
Hence, is bounded.
Let . From (3.5), we deduce
| 3.6 |
We consider the following two cases.
Case 1. One has for every large enough.
In this case, exists as finite and hence
| 3.7 |
This, together with (3.6), implies that
Since (where is a constant), we get
Noting that is bounded, we deduce immediately that
| 3.8 |
Next, we prove that
| 3.9 |
Since is bounded, there exists a subsequence satisfying and
By the lower semicontinuity of f, we get
So
That is, ẑ is a minimizer of f, and . Therefore
| 3.10 |
Then we have
Note that is bounded, and that . Thus by (3.8). From Lemma 2.5, we deduce that
Case 2. There exists a subsequence of such that
By Lemma 2.6, there exists a strictly nondecreasing sequence of positive integers such that and the following properties are satisfied by all numbers :
| 3.11 |
We have
Consequently,
Hence,
| 3.12 |
By a similar argument to that of Case 1, we prove that
| 3.13 |
where
In particular, from (3.13), we get
| 3.14 |
Since , we deduce that
Then, from (3.14), we have
Then
| 3.15 |
Then, from (3.12), we deduce that
| 3.16 |
Thus, from (3.11) and (3.16), we conclude that
Therefore, . This completes the proof. □
Conclusion
Recently, the SFP has been studied extensively by many authors. However, some algorithms need to compute , and this is not an easy thing to work out. Others do not need to compute , but the algorithms always have weak convergence. If we want to obtain strong convergence theorems, the algorithms are complex and difficult to calculate. We try to get over the drawbacks. In this article, we use the regularized CQ algorithm without computing to find the minimum-norm solution of the SFP, where , . Then, under suitable conditions, the explicit strong convergence theorem is obtained.
Acknowledgements
The authors thank the referees for their helping comments, which notably improved the presentation of this paper. This work was supported by the Fundamental Research Funds for the Central Universities [grant number 3122017072]. MT was supported by the Foundation of Tianjin Key Laboratory for Advanced Signal Processing. H-FZ was supported in part by the Technology Innovation Funds of Civil Aviation University of China for Graduate (Y17-39).
Footnotes
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
All the authors read and approved the final manuscript.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Ming Tian, Email: tianming1963@126.com.
Hui-Fang Zhang, Email: huifangzhang109@126.com.
References
- 1.Censor Y, Elfving T. A multiprojection algorithm using Bregman projections in a product space. Numer. Algorithms. 1994;8:221–239. doi: 10.1007/BF02142692. [DOI] [Google Scholar]
- 2.Censor Y, Bortfeld T, Martin B, Trofimov A. A unified approach for inversion problems in intensity-modulated radiation therapy. Phys. Med. Biol. 2003;51:2353–2365. doi: 10.1088/0031-9155/51/10/001. [DOI] [PubMed] [Google Scholar]
- 3.Censor Y, Elfving T, Kopf N, Bortfeld T. The multiple-sets split feasibility problem and its applications for inverse problem. Inverse Probl. 2005;21:2071–2084. doi: 10.1088/0266-5611/21/6/017. [DOI] [Google Scholar]
- 4.López G, Martín-Márquez V, Xu HK. Perturbation techniques for nonexpansive mappings with applications. Nonlinear Anal., Real World Appl. 2009;10:2369–2383. doi: 10.1016/j.nonrwa.2008.04.020. [DOI] [Google Scholar]
- 5.López G, Martín-Márquez V, Xu HK. Iterative algorithms for the multiple-sets split feasibility problem. In: Censor Y, Jiang M, Wang G, editors. Biomedical Mathematics: Promising Directions in Imaging, Therapy Planning and Inverse Problems. Madison: Medical Physics Publishing; 2010. pp. 243–279. [Google Scholar]
- 6.Xu HK. Iterative methods for the split feasibility problem in infinite-dimensional Hilbert space. Inverse Probl. 2010;26 doi: 10.1088/0266-5611/26/10/105018. [DOI] [Google Scholar]
- 7.Dang Y, Gao Y. The strong convergence of a KM-CQ-like algorithm for a split feasibility problem. Inverse Probl. 2011;27 doi: 10.1088/0266-5611/27/1/015007. [DOI] [Google Scholar]
- 8.Qu B, Xiu N. A note on the CQ algorithm for the split feasibility problem. Inverse Probl. 2005;21:1655–1665. doi: 10.1088/0266-5611/21/5/009. [DOI] [Google Scholar]
- 9.Wang F, Xu HK. Approximating curve and strong convergence of the CQ algorithm for the split feasibility problem. J. Inequal. Appl. 2010;2010 [Google Scholar]
- 10.Xu HK. A variable Krasnosel’skii-Mann algorithm and the multiple-set split feasibility problem. Inverse Probl. 2006;22:2021–2034. doi: 10.1088/0266-5611/22/6/007. [DOI] [Google Scholar]
- 11.Yang Q. The relaxed CQ algorithm for solving the split feasibility problem. Inverse Probl. 2004;20:1261–1266. doi: 10.1088/0266-5611/20/4/014. [DOI] [Google Scholar]
- 12.Byrne C. Iterative oblique projection onto convex sets and the split feasibility problem. Inverse Probl. 2002;18:441–453. doi: 10.1088/0266-5611/18/2/310. [DOI] [Google Scholar]
- 13.Byrne C. A unified treatment of some iterative algorithms in signal processing and image reconstruction. Inverse Probl. 2004;20:103–120. doi: 10.1088/0266-5611/20/1/006. [DOI] [Google Scholar]
- 14.Yao YH, Wu JG, Liou YC. Regularized methods for the split feasibility problem. Abstr. Appl. Anal. 2012;2012 [Google Scholar]
- 15.Yao YH, Liou YC, Yao JC. Iterative algorithms for the split variational inequality and fixed point problems under nonlinear transformations. J. Nonlinear Sci. Appl. 2017;10:843–854. doi: 10.22436/jnsa.010.02.43. [DOI] [Google Scholar]
- 16.Yao YH, Agarwal RP, Postolache M, Liou YC. Algorithms with strong convergence for the split common solution of the feasibility problem and fixed point problem. Fixed Point Theory Appl. 2014;2014 doi: 10.1186/1687-1812-2014-183. [DOI] [Google Scholar]
- 17.Latif A, Qin X. A regularization algorithm for a splitting feasibility problem in Hilbert spaces. J. Nonlinear Sci. Appl. 2017;10:3856–3862. doi: 10.22436/jnsa.010.07.40. [DOI] [Google Scholar]
- 18.Chang SS, Wang L, Zhao YH. On a class of split equality fixed point problems in Hilbert spaces. J. Nonlinear Var. Anal. 2017;1:201–212. [Google Scholar]
- 19.Tang JF, Chang SS, Dong J. Split equality fixed point problem for two quasi-asymptotically pseudocontractive mappings. J. Nonlinear Funct. Anal. 2017;2017 [Google Scholar]
- 20.Shi LY, Ansari QH, Wen CF, Yao JC. Incremental gradient projection algorithm for constrained composite minimization problems. J. Nonlinear Var. Anal. 2017;1:253–264. [Google Scholar]
- 21.Xu HK. Averaged mappings and the gradient-projection algorithm. J. Optim. Theory Appl. 2011;150:360–378. doi: 10.1007/s10957-011-9837-z. [DOI] [Google Scholar]
- 22.Yang Q. On variable-step relaxed projection algorithm for variational inequalities. J. Math. Anal. Appl. 2005;302:166–179. doi: 10.1016/j.jmaa.2004.07.048. [DOI] [Google Scholar]
- 23.López G, Martín-Márquez V, Wang FH, Xu HK. Solving the split feasibility problem without prior knowledge of matrix norms. Inverse Probl. 2012 [Google Scholar]
- 24.Yao YH, Postolache M, Liou YC. Strong convergence of a self-adaptive method for the split feasibility problem. Fixed Point Theory Appl. 2013;2013 doi: 10.1186/1687-1812-2013-201. [DOI] [Google Scholar]
- 25.Bauschke HH, Combettes PL. Convex Analysis and Monotone Operator Theory in Hilbert Spaces. Berlin: Springer; 2011. [Google Scholar]
- 26.Goebel K, Kirk WA. Topics on Metric Fixed Point Theory. Cambridge: Cambridge University Press; 1990. [Google Scholar]
- 27.Takahashi W. Nonlinear Functional Analysis. Yokohama: Yokohama Publishers; 2000. [Google Scholar]
- 28.Aubin JP. Optima and Equilibria: An Introduction to Nonlinear Analysis. Berlin: Springer; 1993. [Google Scholar]
- 29.Xu HK. Viscosity approximation methods for nonexpansive mappings. J. Math. Anal. Appl. 2004;298:279–291. doi: 10.1016/j.jmaa.2004.04.059. [DOI] [Google Scholar]
- 30.Maingé PE. Strong convergence of projected subgradient methods for non-smooth and non-strictly convex minimization. Set-Valued Anal. 2008;17(7-8):899–912. doi: 10.1007/s11228-008-0102-z. [DOI] [Google Scholar]
