Abstract
In this paper, we present two iterative algorithms for approximating a solution of the split feasibility problem on zeros of a sum of monotone operators and fixed points of a finite family of nonexpansive mappings. Weak and strong convergence theorems are proved in the framework of Hilbert spaces under some mild conditions. We apply the obtained main result for the problem of finding a common zero of the sum of inverse strongly monotone operators and maximal monotone operators, for finding a common zero of a finite family of maximal monotone operators, for finding a solution of multiple sets split common null point problem, and for finding a solution of multiple sets split convex feasibility problem. Some applications of the main results are also provided.
Keywords: Maximal monotone operator, Inverse strongly monotone operator, Resolvent operator, Convex feasibility problems
Introduction
A very common problem in different areas of mathematics and physical sciences consists of finding a point in the intersection of convex sets and is formulated as finding a point satisfying the property
where , , are nonempty, closed, and convex subsets of a Hilbert space H. This problem is called the convex feasibility problem (CFP). There are various applications of CFP in many applied disciplines as diverse as applied mathematics, approximation theory, image recovery and signal processing, control theory, biomedical engineering, communications, and geophysics (see [1–7] and the references therein).
The problem of finding such that and is called the split feasibility problem (SFP), where C and D are nonempty, closed, and convex subsets of real Hilbert spaces and , respectively, and is a bounded linear operator. Let , then the SFP can be viewed as a special case of the CFP since it can be rewritten as . However, the methodologies for studying the SFP are actually different from those for the CFP; see [8–14].
The theory of monotone operators has appeared as a powerful and effective tool for studying a wide class of problems arising in different branches of social, engineering, and pure sciences in a unified and general framework. There is a notion about monotone operators and it is one of generalized sums of two monotone operators; see [15, 16] and the references therein. In recent years, monotone operators have received a lot of attention for treating zero points of monotone operators and fixed point of mappings which are Lipschitz continuous; see [17–22] and the references therein. The first algorithm for approximating the zero points of the maximal monotone operator was introduced by Martinet [23]. He considered the proximal point algorithm for finding zero points of a maximal monotone operator. Then, Passty [24] introduced a forward-backward algorithm method for finding zero points of the sum of two operators. There are various applications of the problem of finding zero points of the sum of two operators; see [25–29] for example and the references therein.
Therefore, there are some generalizations of the CFP, which can be formulated in various ways such as: finding a common fixed point of nonexpansive operators, finding a common minimum of convex functionals, finding a common zero of maximal monotone operators, solving a system of variational inequalities, and solving a system of convex inequalities. Surveys of methods for solving such problems can be found in [2, 4].
Recently, some authors introduced and studied algorithms to get a common solution to inclusion problems and fixed point problems in the framework of Hilbert spaces; see [30–32]. Cho et al. [30] considered the problem of finding a common solution to the zero point problems involving two monotone operators and fixed point problems involving asymptotically strictly pseudocontractive mappings based on a one-step iterative method and proved the weak convergence theorems in the framework of Hilbert spaces.
In this paper, motivated and inspired by the above literature, we consider an iterative algorithm for finding a solution of split feasibility problem for a point in zeros of a finite sum of α-inverse strongly monotone operators and maximal monotone operators and fixed points of nonexpansive mappings. That is, we are going to consider the following problem: Let and be real Hilbert spaces. Let , , be -inverse strongly monotone operators and , , be maximal monotone operators, , , be nonexpansive mappings, be a bounded linear operator. We are interested in considering the problem of finding a solution such that
| 1.1 |
where . Weak and strong convergence theorems will be provided under some mild conditions.
The paper is organized as follows. Section 2 gathers some definitions and lemmas of geometry of Hilbert spaces and monotone operators, which will be needed in the remaining sections. In Sect. 3, we prepare an iterative algorithm and prove the weak and strong convergence theorems. Finally, in Sect. 4, the results of Sect. 3 are applied to solve CFP, multiple-set null point problems, variational inequality problems, fixed point problems, and equilibrium problems.
Preliminaries
Throughout this paper, H will be a Hilbert space with norm and inner product , respectively. We now provide some basic concepts, definitions, and lemmas which will be used in the sequel. We write to indicate that the sequence converges strongly to x and to indicate that converges weakly to x.
Let be a mapping. We say that T is a Lipschitz mapping if there exists such that
The number L, associated with T, is called a Lipschitz constant. If , we say that T is a nonexpansive mapping, that is,
We will say that T is firmly nonexpansive if
The set of fixed points of T will be denoted by , that is, . It is well known that if T is nonexpansive, then is closed and convex. Moreover, every nonexpansive operator satisfies the following inequality:
Therefore, for all and ,
| 2.1 |
Lemma 2.1
([35])
Let H be a real Hilbert space and be a nonexpansive mapping with . Then the mapping is demiclosed at zero, that is, if is a sequence in H such that and , then .
A mapping is called α-averaged if there exists such that , where S is a nonexpansive mapping of H into H. It should be observed that firmly nonexpansive mappings are -averaged mappings.
We now recall the concepts and facts on the class of monotone operators, for both single and multi-valued operators.
An operator is called α-inverse strongly monotone (α-ism) for a positive number α if
Lemma 2.2
([21])
Let C be a nonempty, closed, and convex subset of a real Hilbert space H. Let the mapping be α-inverse strongly monotone and be a constant. Then we have
for all . In particular, if , then is nonexpansive.
Lemma 2.3
We have
The composite of finitely many averaged mappings is averaged. In particular, if is -averaged, where for , then the composite is α-averaged, where .
If A is β-ism and , then is firmly nonexpansive.
A multifunction is called a monotone operator if, for every ,
A monotone operator is said to be maximal monotone, when its graph is not properly included in the graph of any other monotone operators on the same space. For a maximal monotone operator B on H, and , we define the single-valued resolvent by . It is well known that is firmly nonexpansive, and .
Next, we collect some useful facts on monotone operators that will be used in our proof.
Lemma 2.4
([38])
Let C be a nonempty, closed, and convex subset of a real Hilbert space H and be an operator. If is a maximal monotone operator, then .
Lemma 2.5
([39])
Let be a maximal monotone operator. For , , and ,
For each sequence , we put
The following lemma plays an important role in concluding our results.
Lemma 2.6
([37])
Let C be a nonempty closed convex subset of a real Hilbert space H. Let be a sequence in H satisfying the properties:
-
(i)
exists for each ;
-
(ii)
.
Then converges weakly to a point in C.
Parallel algorithm
Let and be real Hilbert spaces. Let , , be -inverse strongly monotone operators and , , be maximal monotone operators, , , be nonexpansive mappings, be a bounded linear operator. We will denote by the adjoint operator of L. Let and be sequences of positive real numbers. For , we introduce the following parallel algorithm:
| 3.1 |
We start by some lemmas.
Lemma 3.1
Let . If
-
(i)
and
-
(ii)
for some ,
then the sequences and generated by (3.1) are bounded.
Proof
Let . We have
| 3.2 |
By (2.1), we get
| 3.3 |
It follows from (3.2) and (3.3) that
| 3.4 |
Hence, from Lemma 2.2, Lemma 2.4, and the control conditions on and , we have
This means that is a nonincreasing sequence of nonnegative real numbers, so it follows that it is a convergent sequence. Also, from the above inequality, we have and converge to the same limit point. These imply that the sequences and are bounded, and the proof is completed. □
Lemma 3.2
If , then .
Proof
By (3.4) we have
and hence,
Therefore, from (3.1), we get
| 3.5 |
for each , which implies that
| 3.6 |
From (2.1), we have
| 3.7 |
for each . Thus, by (3.6) and the assumption of , we have
| 3.8 |
for each . From Lemma 2.1, we obtain for each . This completes the proof. □
Lemma 3.3
Let . If . Then, for each , we have .
Proof
Since and are firmly nonexpansive, they are both -averaged and hence is -averaged by Lemma 2.3. Thus, for each and , we can write
where is a nonexpansive mapping and for each and . Then we can rewrite as
| 3.9 |
Let , we have
and hence,
Then
From (3.9),
| 3.10 |
By (3.5), we get
| 3.11 |
Now, from (3.1), (3.10), and (3.11), we obtain
| 3.12 |
□
Lemma 3.4
Assume that for some positive real number β. Then, for each , we have , .
Proof
Set , so . By Lemma 2.5, we have
| 3.13 |
On the other hand, we have
Since is inverse strongly monotone, is bounded, (3.11) and (3.12) we know that is bounded. It follows from and (3.13) that
| 3.14 |
We also have
| 3.15 |
It follows form (3.12), (3.14), and (3.15) that
for each . This completes the proof of the lemma. □
Now, the weak convergence of algorithm (3.1) is given by the following theorem.
Theorem 3.5
Let and be real Hilbert spaces. Let , , be nonexpansive mappings, be a bounded linear operator, , , be -inverse strongly monotone operators, and , , be maximal monotone operators such that . Let , for each and , then the sequence generated by (3.1) converges weakly to a point .
Proof
In Lemma 3.1, we show that exists for each . From Lemmas 3.2 and 3.4 we imply that . Then it follows from Lemma 2.6 that converges weakly to a point . □
Recall that for a subset C of H, a mapping is said to be semi-compact if for any bounded sequence such that (), there exists a subsequence of such that converges strongly to .
Strong convergence of algorithm (3.1), under the concept of semi-compact assumption, is given by the following theorem.
Theorem 3.6
Let and be real Hilbert spaces. Let , , be nonexpansive mappings, be a bounded linear operator, , , be -inverse strongly monotone operators, and , , be maximal monotone operators such that . Let , for each and . If at least one of the maps is semi-compact, then the sequence generated by (3.1) converges strongly to a point .
Proof
Let be semi-compact for some fixed . Since by (4.7), there exists a subsequence of such that it converges strongly to q. Since converges weakly to p, we get . On the other hand, exists and , which show that converges strongly to . This completes the proof of the theorem. □
Deduced results of parallel algorithm
One can obtain some results from Theorem 3.5. We give some of them in the following.
If we take , we have the following corollary.
Corollary 3.7
Let and be real Hilbert spaces. Let be a nonexpansive mapping, be a bounded linear operator, be an α-inverse strongly monotone operator, and be a maximal monotone operator such that . Suppose that the sequence is defined by the following algorithm:
where , , and for each . Then the sequence converges weakly to a point . If T be semi-compact, then the convergence is strong.
From Theorem 3.5, we have the following corollary for the problem of finding a common zero of the sum of α-inverse strongly monotone operators and maximal monotone operators.
Corollary 3.8
Let H be a real Hilbert space, , , be -inverse strongly monotone operators, and , , be maximal monotone operators such that and . Suppose that the sequence is defined by the following algorithm:
where and for each . Then the sequence converges weakly to a point .
In the following corollary, we have a result for finding a common zero of a finite family of maximal monotone operators.
Corollary 3.9
Let H be a real Hilbert space, , , be maximal monotone operators such that . Suppose that the sequence is defined by the following algorithm:
where and for each . Then the sequence converges weakly to a point .
Corollary 3.10
Let H be a real Hilbert space, , , be -inverse strongly monotone operators such that and . Suppose that the sequence is defined by the following algorithm:
where and for each . Then the sequence converges weakly to a point .
Corollary 3.11
Let and be real Hilbert spaces and , , be nonexpansive mappings and be a bounded linear operator such that . Suppose that the sequence is defined by the following algorithm:
where and . Then the sequence converges weakly to a point . If is semi-compact for some , then the convergence is strong.
Parallel hybrid algorithm
Notice that, in order to guarantee the strong convergence theorem of the introduced algorithm (3.1), we proposed an additional assumption to one of the operators , as a semi-compact assumption (see Theorem 3.6). Next, we propose the following hybrid algorithm to obtain a strong convergence theorem for finding a point in zeros of a finite family of sums of α-inverse strongly monotone operators and maximal monotone operators and nonexpansive mappings. Of course, the strong convergence theorems of the following algorithm will be guaranteed without any additional assumptions on the considered operators. To do this, we recall some necessary concepts and facts: let C be a closed and convex subset of a Hilbert space H. The operator is called a metric projection operator if it assigns to each its nearest point such that
An element y is called the metric projection of H onto C and is denoted by . It exists and is unique at any point of the Hilbert space. It is known that the metric projection operator is a firmly nonexpansive mapping. Also, the following characterization is very useful in our proof.
Lemma 4.1
Let H be a Hilbert space and C be a nonempty, closed, and convex subset of H. Then, for all , the element if and only if
Now we are in a position to introduce the aforementioned algorithm: Let and be a sequence generated by the following algorithm:
| 4.1 |
Theorem 4.2
Let and be real Hilbert spaces. Let , , be nonexpansive mappings, be a bounded linear operator, , , be -inverse strongly monotone operators, and , , be maximal monotone operators such that . Let , for each and . Then the sequence generated by (4.1) converges strongly to .
Proof
We prove that the sequence generated by (4.1) is well defined. We first show that is closed and convex for each . is closed and convex and suppose that is closed and convex for some . Set
then . For each , we obtain
This implies that is closed and convex. In a similar manner, is closed and convex and so is . By the induction, is closed and convex for each .
We show that for each . Let . From Lemmas 2.2 and 2.4 and (4.1), we have
This together with (3.4) implies that . Then is well defined.
Since is nonempty, closed, and convex, there exists a unique element such that . From , we get
| 4.2 |
Since again and , we get
| 4.3 |
Thus, the sequence is a bounded above and nondecreasing sequence, so exists, and the sequence is bounded. By (3.4) the sequence is bounded too.
We show that , , and . From , , and Lemma 4.1, we obtain
Then we get
and hence,
By and the definition of , we obtain
and then
which implies that
| 4.4 |
Also, we have
therefore,
| 4.5 |
| 4.6 |
Now, we show that . From (3.5), (3.7), and (4.4), we get
| 4.7 |
for each . It follows from Lemma 2.1 that . By arguing similarly to the proof of Lemma 3.4, (4.4), and (4.6), we conclude . Therefore,
| 4.8 |
Finally, we show that the sequence generated by (4.1) converges strongly to . Since and , we get
| 4.9 |
Let be an arbitrary subsequence of converging weakly to . Then by (4.8) and hence it follows from the lower semi-continuity of the norm that
Thus, we obtain that . Using the Kadec–Klee property of , we get that . Since is an arbitrary weakly convergent subsequence of and exists, we can imply that converges strongly to q. This completes the proof. □
Deduced results of the parallel hybrid algorithm
One can obtain some results from Theorem 4.2. We give some of them in the following.
If we take , we have the following corollary.
Corollary 4.3
Let and be real Hilbert spaces. Let be a nonexpansive mapping, be a bounded linear operator, be an -inverse strongly monotone operator, and be a maximal monotone operator such that . Suppose that the sequence is defined by the following algorithm:
where , , and for each . Then the sequence converges strongly to .
From Theorem 4.2, we have the following corollary for the problem of finding a common zero of the sum of α-inverse strongly monotone operators and maximal monotone operators.
Corollary 4.4
Let H be a real Hilbert space, , , be -inverse strongly monotone operators, and , , be maximal monotone operators such that and . Suppose that the sequence is defined by the following algorithm:
where and for each . Then the sequence converges strongly to .
Applications
Zeros of maximal monotone operators
In this section, we discuss some applications of the main theorems. Let , , be maximal monotone operators. Set , where and . We know that is nonexpansive and for each . By applying Theorem 3.5, we can get the following results.
Theorem 5.1
Let and be real Hilbert spaces, , , be -inverse strongly monotone operators, , , and , , be maximal monotone operators, and be a bounded linear operator such that . Let and the sequence be generated by the following algorithm:
If , , and for each , then converges weakly to a point .
By Theorem 5.1, we have the following corollary for multiple sets split null point problems.
Corollary 5.2
Let and be real Hilbert spaces, , , , , be maximal monotone operators, and be a bounded linear operator such that . Let and the sequence be generated by the following algorithm:
If and for each , then converges weakly to a point .
By applying Theorem 4.2, we have the following theorem.
Theorem 5.3
Let and be real Hilbert spaces, , , be -inverse strongly monotone operators, , , and , , be maximal monotone operators, and be a bounded linear operator such that . Let and the sequence be generated by the following algorithm:
| 5.1 |
If , , and for each , then converges strongly to .
Multiple set split convex feasibility problems
Let be a proper, convex, and lower semi-continuous function. It is well known that the subdifferential , which is defined as
is a maximal monotone operator. In particular, let C be a nonempty, closed, and convex subset of a real Hilbert space H. Let us consider the indicator function of C, denoted by , which is defined as
We know that is a proper, convex, and lower semi-continuous function on H, and it follows that the subdifferential of is a maximal monotone operator. Furthermore, we get if and only if , where and for each . Using these facts, by Theorems 3.5 and 4.2, we have the following corollaries for the multiple set split convex feasibility problem in Hilbert spaces.
Corollary 5.4
Let and be real Hilbert spaces, , , , , be nonempty, closed, and convex, and be a bounded linear operator such that . Let and the sequence be generated by the following algorithm:
If for each , then converges weakly to a point .
Corollary 5.5
Let and be real Hilbert spaces, , , , , be nonempty, closed, and convex, and be a bounded linear operator such that . Let and the sequence be generated by the following algorithm:
If for each , then converges strongly to .
Multiple sets split equilibrium problems
Now, we apply Theorem 3.5 for getting a common solution of multiple sets split equilibrium problems. In this respect, let C be a nonempty closed convex subset of a Hilbert space and be a bifunction. The equilibrium problem for bifunction F is the problem of finding a point such that
| 5.2 |
The set of solutions of equilibrium problem (5.2) is denoted by . The bifunction is called monotone if for all . For finding a solution of equilibrium problem (5.2), we assume that F satisfies the following properties:
for all ;
F is monotone;
for each , ;
for each , is convex and lower semi-continuous.
Then we have the following lemma which can be found in [40, 41].
Lemma 5.6
Let C be a nonempty closed convex subset of a Hilbert space and be a bifunction satisfying properties (A1)–(A4). Let r be a positive real number and . Then there exists such that
Further, define
for all and . Then the following hold:
is single-valued;
- is firmly nonexpansive; that is,
;
is closed and convex.
Let , , and , , be nonempty, closed, and convex subsets of real Hilbert spaces and , respectively, , , and , , be bifunctions which satisfy properties (A1)–(A4), and be a bounded linear operator. From Lemma 5.6 there exist the sequences of and of satisfying
| 5.3 |
Therefore, by applying Theorem 3.5, we have the following theorem for multiple sets split equilibrium problem.
Theorem 5.7
Let , , and , , be nonempty, closed, and convex subsets of real Hilbert spaces and , respectively, , , and , , be bifunctions which satisfy properties (A1)–(A4). Suppose that is a bounded linear operator such that . If , for each and r is a positive real number, then the sequence generated by (5.3) converges weakly to a solution of multiple sets split equilibrium problem.
We also have the following strong convergence theorem for finding a solution of multiple sets split equilibrium problem.
Theorem 5.8
Let , , and , , be nonempty, closed, and convex subsets of real Hilbert spaces and , respectively, , , and , , be bifunctions which satisfy properties (A1)–(A4). Suppose that is a bounded linear operator such that . Suppose that and the sequence is generated by the following algorithm:
| 5.4 |
If , for each and r is a positive real number, then the sequence converges strongly to .
Numerical experiments
In this section, we show some numerical examples and discuss the possible good choices of step size parameters and , which satisfy the control conditions in Theorem 3.5.
Let and be equipped with the Euclidean norm. Let , , and be fixed in , and and be scalars. Set and , where and are the following closed convex ice-cream cones in :
We will consider 1-ism operators and , where and are defined by the above settings.
Next, for each , we are also concerned with the following two norms:
Consider a function , which is defined by
We know that f is a convex function and subdifferential of f is
Moreover, since f is a convex function, we know that must be a maximal monotone operator, and for each , we have
where is denoted for the signum function.
On the other hand, let , , and be three fixed vectors in . We consider a nonempty convex subset of , where , , and . We notice that .
Now, let us consider a matrix . We see that L is a bounded linear operator on into with .
Based on the above settings, we will present some numerical experiments to show the efficiency of the constructed algorithm (3.1). That is, we are going to show that algorithm (3.1) converges to a point such that
| 6.1 |
and in this experiment, we consider the stopping criterion by .
We will consider the following cases of the step size parameters and with the initial vectors , , , , and in :
- Case 1.
, .
- Case 2.
, .
- Case 3.
, .
- Case 4.
, .
- Case 5.
, .
- Case 6.
, .
- Case 7.
, .
- Case 8.
, .
- Case 9.
, .
From Tables 1, 2, and 3, we may suggest that, for each initial point, the step size of the parameters provides a faster convergence rate than other cases. While the step size parameters seem to have less impact on the speed of algorithm (3.1) to a solution set (6.1).
Table 1.
Influence of the step size parameters and (cases 1–3) of algorithm (3.1) for different initial points
| Case → | Case 1 | Case 2 | Case 3 | ||||||
|---|---|---|---|---|---|---|---|---|---|
| #Initial point ↓ | Iters | Time (s) | Sol | Iters | Time (s) | Sol | Iters | Time (s) | Sol |
| 1647 | 0.644764 | 145 | 0.210611 | 110 | 0.172755 | ||||
| 790 | 0.393530 | 51 | 0.117471 | 27 | 0.098625 | ||||
| 195 | 0.231496 | 49 | 0.123486 | 36 | 0.127907 | ||||
| 1069 | 0.486436 | 150 | 0.207209 | 113 | 0.181702 | ||||
| 2121 | 0.847208 | 449 | 0.313106 | 361 | 0.284821 | ||||
Table 2.
Influence of the step size parameters and (cases 4–6) of algorithm (3.1) for different initial points
| Case → | Case 4 | Case 5 | Case 6 | ||||||
|---|---|---|---|---|---|---|---|---|---|
| #Initial point ↓ | Iters | Time (s) | Sol | Iters | Time (s) | Sol | Iters | Time (s) | Sol |
| 1647 | 0.650587 | 106 | 0.176374 | 56 | 0.124235 | ||||
| 790 | 0.398679 | 51 | 0.122999 | 27 | 0.098005 | ||||
| 3 | 0.078350 | 3 | 0.079696 | 3 | 0.083422 | ||||
| 1032 | 0.500529 | 61 | 0.133658 | 31 | 0.108214 | ||||
| 1658 | 0.689241 | 107 | 0.180100 | 57 | 0.129912 | ||||
Table 3.
Influence of the step size parameters and (cases 7–9) of algorithm (3.1) for different initial points
| Case → | Case 7 | Case 8 | Case 9 | ||||||
|---|---|---|---|---|---|---|---|---|---|
| #Initial point ↓ | Iters | Time (s) | Sol | Iters | Time (s) | Sol | Iters | Time (s) | Sol |
| 1647 | 0.644395 | 106 | 0.167910 | 56 | 0.122966 | ||||
| 790 | 0.403824 | 51 | 0.118171 | 27 | 0.095997 | ||||
| 3 | 0.080739 | 3 | 0.080157 | 3 | 0.080880 | ||||
| 1032 | 0.463895 | 61 | 0.133494 | 31 | 0.104363 | ||||
| 1658 | 0.646397 | 107 | 0.173753 | 57 | 0.127317 | ||||
Conclusions
In this paper, we present two iterative algorithms, (3.1) and (4.1), for approximating a solution of the split feasibility problem on zeros of a finite sum of monotone operators and fixed points of a finite family of nonexpansive mappings. Under some mild conditions, we show the convergence theorems of the mentioned algorithms. Subsequently, some corollaries and applications of those main results are provided. We point out that the construction of algorithm (3.1) seems to be less complicated than that of (4.1). However, algorithm (3.1) requires some additional assumptions in order to guarantee the strong convergence theorem, while algorithm (4.1) does not need them (see Theorem 3.6 and Theorem 4.2). This observation may lead to the future works that are to analyze and discuss the rate of convergence of these suggested algorithms.
Acknowledgements
The authors thank the anonymous referees for their remarkable comments and suggestions to improve this paper.
Authors’ contributions
All authors contributed equally to the writing of this paper. All authors read and approved the final manuscript.
Funding
This work is partially supported by the Thailand Research Fund under the project RSA5880028.
Competing interests
The authors declare that they have no competing interests.
Footnotes
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Narin Petrot, Email: narinp@nu.ac.th.
Montira Suwannaprapa, Email: montira.sw@gmail.com.
Vahid Dadashi, Email: vahid.dadashi@iausari.ac.ir.
References
- 1.Bauschke H.H., Borwein J.M. On projection algorithms for solving convex feasibility problems. SIAM Rev. 1996;38(3):367–426. doi: 10.1137/S0036144593251710. [DOI] [Google Scholar]
- 2.Censor Y. Iterative methods for the convex feasibility problem. North-Holl. Math. Stud. 1984;87:83–91. doi: 10.1016/S0304-0208(08)72812-3. [DOI] [Google Scholar]
- 3.Combettes P.L. The convex feasibility problem in image recovery. In: Hawkes P., editor. Advances in Imaging and Electron Physics. New York: Academic Press; 1996. pp. 155–270. [Google Scholar]
- 4.Combettes P.L. The foundations of set theoretic estimation. Proc. IEEE. 1993;81(2):182–208. doi: 10.1109/5.214546. [DOI] [Google Scholar]
- 5.Deutsch F. The method of alternating orthogonal projections. In: Singh S.P., editor. Approximation Theory, Spline Functions and Applications. The Netherlands: Kluwer Academic; 1992. pp. 105–121. [Google Scholar]
- 6.Rockafellar R.T. Maximal monotone operators and proximal point algorithm. SIAM J. Control Optim. 1976;14(5):877–898. doi: 10.1137/0314056. [DOI] [Google Scholar]
- 7.Stark H., editor. Image Recovery Theory and Applications. Orlando: Academic Press; 1987. [Google Scholar]
- 8. Yao, Y., Liou, Y.C., Postolache, M.: Self-adaptive algorithms for the split problem of the demicontractive operators. Optimization 10.1080/02331934.2017.1390747 [DOI]
- 9. Yao, Y., Postolache, M., Qin, X., Yao, J.C.: Iterative algorithms for the proximal split feasibility problem. U. Politeh. Buch. Ser. A. (in printing)
- 10.Yao Y., Leng L., Postolache M., Zheng X. Mann-type iteration method for solving the split common fixed point problem. J. Nonlinear Convex Anal. 2017;18(5):875–882. [Google Scholar]
- 11.Yao Y., Agarwal R.P., Postolache M., Liu Y.C. Algorithms with strong convergence for the split common solution of the feasibility problem and fixed point problem. Fixed Point Theory Appl. 2014;2014:183. doi: 10.1186/1687-1812-2014-183. [DOI] [Google Scholar]
- 12.Yao Y., Postolache M., Liou Y.C. Strong convergence of a self-adaptive method for the split feasibility problem. Fixed Point Theory Appl. 2013;2013:201. doi: 10.1186/1687-1812-2013-201. [DOI] [Google Scholar]
- 13.Ansari Q.H., Nimana N., Petrot N. Split hierarchical variational inequality problems and related problems. Fixed Point Theory Appl. 2014;2014:208. doi: 10.1186/1687-1812-2014-208. [DOI] [Google Scholar]
- 14.Suwannaprapa M., Petrot N., Suantai S. Weak convergence theorems for split feasibility problems on zeros of the sum of monotone operators and fixed point sets in Hilbert spaces. Fixed Point Theory Appl. 2017;2017:6. doi: 10.1186/s13663-017-0599-7. [DOI] [Google Scholar]
- 15.Moudafi A. On the regularization of the sum of two maximal monotone operators. Nonlinear Anal., Theory Methods Appl. 2000;42(7):1203–1208. doi: 10.1016/S0362-546X(99)00136-4. [DOI] [Google Scholar]
- 16.Moudafi A., Oliny M. Convergence of a splitting inertial proximal method for monotone operators. J. Comput. Appl. Math. 2003;155(2):447–454. doi: 10.1016/S0377-0427(02)00906-8. [DOI] [Google Scholar]
- 17.Chang S.S., Lee H.J., Chan C.K. A new method for solving equilibrium problem fixed point problem and variational inequality problem with application to optimization. Nonlinear Anal., Theory Methods Appl. 2009;70(9):3307–3319. doi: 10.1016/j.na.2008.04.035. [DOI] [Google Scholar]
- 18.Dadashi V. Shrinking projection algorithms for the split common null point problem. Bull. Aust. Math. Soc. 2017;96:299–306. doi: 10.1017/S000497271700017X. [DOI] [Google Scholar]
- 19.Kang S., Cho S., Liu Z. Convergence of iterative sequences for generalized equilibrium problems involving inverse-strongly monotone mappings. J. Inequal. Appl. 2010;2010(1):827082. [Google Scholar]
- 20.Lv S. Generalized systems of variational inclusions involving (A, η)-monotone mappings. Adv. Fixed Point Theory. 2011;1(1):15. [Google Scholar]
- 21.Nadezhkina N., Takahashi W. Weak convergence theorem by an extragradient method for nonexpansive mappings and monotone mappings. J. Optim. Theory Appl. 2006;128(1):191–201. doi: 10.1007/s10957-005-7564-z. [DOI] [Google Scholar]
- 22.Qin X., Cho Y.J., Kang S.M. Convergence theorems of common elements for equilibrium problems and fixed point problems in Banach spaces. J. Comput. Appl. Math. 2009;225(1):20–30. doi: 10.1016/j.cam.2008.06.011. [DOI] [Google Scholar]
- 23.Martinet B. Régularisation d’inéquations variationnelles par approximations successives. Rev. Fr. Inform. Rech. Oper. 1970;3:154–158. [Google Scholar]
- 24.Passty G.B. Ergodic convergence to a zero of the sum of monotone operators in Hilbert space. J. Math. Anal. Appl. 1979;72(2):383–390. doi: 10.1016/0022-247X(79)90234-8. [DOI] [Google Scholar]
- 25.Dadashi V., Khatibzadeh H. On the weak and strong convergence of the proximal point algorithm in reflexive Banach spaces. Optimization. 2017;66(9):1487–1494. doi: 10.1080/02331934.2017.1337764. [DOI] [Google Scholar]
- 26.Dadashi V., Postolache M. Hybrid proximal point algorithm and applications to equilibrium problems and convex programming. J. Optim. Theory Appl. 2017;174:518–529. doi: 10.1007/s10957-017-1117-0. [DOI] [Google Scholar]
- 27.Moudafi A., Thera M. Finding a zero of the sum of two maximal monotone operators. J. Optim. Theory Appl. 1997;94(2):425–448. doi: 10.1023/A:1022643914538. [DOI] [Google Scholar]
- 28.Qin X., Cho S.Y., Wang L. A regularization method for treating zero points of the sum of two monotone operators. Fixed Point Theory Appl. 2014;2014:75. doi: 10.1186/1687-1812-2014-75. [DOI] [Google Scholar]
- 29.Tseng P. A modified forward-backward splitting method for maximal monotone mappings. SIAM J. Control Optim. 2000;38:431–446. doi: 10.1137/S0363012998338806. [DOI] [Google Scholar]
- 30.Cho S.Y., Li W., Kang S.M. Convergence analysis of an iterative algorithm for monotone operators. J. Inequal. Appl. 2013;2013(1):199. doi: 10.1186/1029-242X-2013-199. [DOI] [Google Scholar]
- 31.Wu C., Liu A. Strong convergence of a hybrid projection iterative algorithm for common solutions of operator equations and of inclusion problems. Fixed Point Theory Appl. 2012;2012(1):90. doi: 10.1186/1687-1812-2012-90. [DOI] [Google Scholar]
- 32.Zhang M. Iterative algorithms for common elements in fixed point sets and zero point sets with applications. Fixed Point Theory Appl. 2012;2012(1):21. doi: 10.1186/1687-1812-2012-21. [DOI] [Google Scholar]
- 33.Shimoji K., Takahashi W. Strong convergence to common fixed points of infinite nonexpansive mappings and applications. Taiwan. J. Math. 2001;5(2):387–404. doi: 10.11650/twjm/1500407345. [DOI] [Google Scholar]
- 34.Suzuki T. Strong convergence theorems for an infinite family of nonexpansive mappings in general Banach spaces. Fixed Point Theory Appl. 2005;1:103–123. [Google Scholar]
- 35.Goebel K., Kirk W.A. Topics in Metric Fixed Point Theory. Cambridge: Cambridge University Press; 1990. [Google Scholar]
- 36.Boikanyo O.A. The viscosity approximation forward-backward splitting method for zeros of the sum of monotone operators. Abstr. Appl. Anal. 2016;2016:2371857. doi: 10.1155/2016/2371857. [DOI] [Google Scholar]
- 37.Xu H.K. Averaged mappings and the gradient-projection algorithm. J. Optim. Theory Appl. 2011;150:360–378. doi: 10.1007/s10957-011-9837-z. [DOI] [Google Scholar]
- 38.Aoyama K., Kimura Y., Takahashi W., Toyoda M. On a strongly nonexpansive sequence in Hilbert spaces. J. Nonlinear Convex Anal. 2007;8:471–489. [Google Scholar]
- 39.Bruck R.E., Passty G.B. Almost convergence of the infinite product of resolvents in Banach spaces. Nonlinear Anal. 1979;3:279–282. doi: 10.1016/0362-546X(79)90083-X. [DOI] [Google Scholar]
- 40.Blum E., Oettli W. From optimization and variational inequalities to equilibrium problems. Math. Stud. 1994;63:123–145. [Google Scholar]
- 41.Combettes P.L., Hirstoaga S.A. Equilibrium programming in Hilbert spaces. J. Nonlinear Convex Anal. 2005;6:117–136. [Google Scholar]
