Abstract
In this paper, we propose a viscosity approximation method to solve the split common fixed point problem and consider the bounded perturbation resilience of the proposed method in general Hilbert spaces. Under some mild conditions, we prove that our algorithms strongly converge to a solution of the split common fixed point problem, which is also the unique solution of the variational inequality problem. Finally, we show the convergence and effectiveness of the algorithms by two numerical examples.
Keywords:
strong convergence; viscosity iterative algorithm; split common fixed point problem; nonexpansive mapping; averaged mapping; bounded perturbation resilience; variational inequality problem MSC:
47H05; 47H09; 47H10; 47J25
1. Introduction
Let and be two real Hilbert spaces with the inner product and the induced norm .
The split feasibility problem (SFP for short) is as follows:
where C and Q are nonempty closed convex subsets of and , respectively, and A is a bounded linear operator of into .
If the set of solutions of the problem (1) is nonempty, then solving problem (1) is equivalent to
where and denotes the metric projection of onto C and is the corresponding adjoint operator of A.
Recently, many problems in engineering and technology can be modeled by problem (1) and many authors have shown that the SFP has many applications in our real life such as image reconstruction, signal processing and intensity-modulated radiation therapy (see [1,2,3]).
In 1994, Censor and Elfving [4] used their algorithm to solve the SFP in the finite-dimensional Euclidean space. In 2002, Byrne [5] improved the algorithm of Censor and Elfving and presented a new method called the algorithm for solving the SFP (1) as follows:
The split common fixed point problem (shortly, SCFPP) is formulated as follows:
where and are nonlinear mappings; here, denotes the set of fixed points of the mapping U. We use S to denote the solution set of problem (4).
Note that, since every closed convex subset of a Hilbert space is the fixed point set of its associating projection if and , the SFP becomes a special case of the SCFPP.
In 2007, Censor and Segal [6] first studied the SCFPP and, to solve the SCFPP, they proposed the following iterative algorithm:
where is a properly chosen stepsize. Algorithm (5) was originally designed to solve the problem (4) for directed mappings.
In 2010, Moudafi [7] proposed an iterative method to solve the SCFPP for quasi-nonexpansive mappings. In 2014, combining the Moudafi method with the Halpern iterative method, Kraikaew and Saejung [8] proposed a new iterative algorithm which does not involve the projection operator to solve the split common fixed point problem. More specifically, their algorithm generates a sequence via the recursions:
where is a fixed element, U and T are quasi-nonexpansive operators.
Recently, many authors have studied the SCFPP, the generalized SCFPP and some relative problems (see, for instance, refs. [3,4,5,9,10,11,12,13] and they have also proposed a lot of algorithms to solve the SCFPP (see [14,15,16,17] and the references therein).
On the other hand, the bounded perturbation resilience and superiorization of iterative methods have been studied by some authors (see [18,19,20,21,22,23]). These problems have received much attention because of their applications in convex feasibility problems [24], image reconstruction [25] and inverse problems of radiation therapy [26] and so on.
Let denote an algorithm operator. If the iteration is replaced by
where is a sequence of nonnegative real numbers and is a sequence in H such that
Then, the algorithm is still convergent and so the algorithm is the bounded perturbation resilient [19].
In 2016, Jin, Censor and Jiang [21] introduced the projected scaled gradient method (PSG for short) with bounded perturbations for solving the following minimization problem:
where f is a continuous differentiable, convex function. The method PSG generates a sequence defined by
where is a diagonal scaling matrix and denotes the sequence of outer perturbations satisfying .
Recently, Xu [22] projected the superiorization techniques for the relaxed PSG as follows:
where is a sequence in
Recently, for solving minimization problem of the combination of two convex functions , Guo and Cui [20] considered the modified proximal gradient method:
and, under suitable conditions, they proved some strong convergence theorems of the method. The definition of proximal operator is as follows.
Definition 1
(see [27]). Let be the space of functions on a real Hilbert space H that are proper, lower semicontinuous and convex. The proximal operator of is defined by
The proximal operator of φ of order is defined as the proximal operator of , that is,
Now, we propose a viscosity method for the problem (4) as follows:
If we treat the above algorithm as the basic algorithm , the bounded perturbation of it is a sequence generated by the iterative process:
In this paper, mainly based on the above works [6,20,22], we prove that our main iterative method (11) is the bounded perturbation resilient and, under some mild conditions, our algorithms strongly converge to a solution of the split common fixed point problem, which is also the unique solution of the variational inequality problem (13). Finally, we give two numerical examples to demonstrate the effectiveness of our iterative schemes.
2. Preliminaries
Let be a sequence in the real Hilbert space H. We adopt the following notations:
- (1)
- Denote converging weakly to x by and converging strongly to x by .
- (2)
- Denote the weak -limit set of by .
Definition 2.
A mapping is said to be:
- (i)
- Lipschitz if there exists a positive constant L such thatIn particular, if , then we say that F is nonexpansive, namely,If , then we say F is contractive.
- (ii)
- α-averaged mapping (shortly, α-av) ifwhere and is nonexpansive.
Definition 3.
A mapping is said to be:
- (i)
- monotone if
- (ii)
- η-strongly monotone if there exists a positive constant η such that
- (iii)
- α-inverse strongly monotone (shortly, α-ism) if there exists a positive constant α such thatIn particular, if , then we say B is firmly nonexpansive, namely,
Using the Cauchy–Schwartz inequality, it is easy to deduce that B is Lipschitz if it is -ism.
Now, we give the following lemmas and propositions needed in the proof of the main results.
Lemma 1
([28]). Let H be a real Hilbert space. Then, the following inequality holds:
Lemma 2
([29]). Let be a ρ-contraction with and be a nonexpansive mapping. Then,
- (i)
- is -strongly monotone, that is,
- (ii)
- is monotone, that is,
Proposition 1
([30]).
- (i)
- If are averaged mappings, then we can get that is averaged. In particular, if is -av for each , where , then is -av.
- (ii)
- If the mappings are averaged and have a common fixed point, then we have
- (iii)
- A mapping T is nonexpansive if and only if is ism.
- (iv)
- If T is ν-ism, then, for any , is -ism.
- (v)
- T is averaged if and only if is ν-ism for some . Indeed, for any , T is averaged if and only if is -ism.
Lemma 3
([31]). Let H be a real Hilbert space and be a nonexpansive mapping with . If is a sequence in H weakly converging to x and converges strongly to y, then . In particular, if , then
Lemma 4
([32] or [33]). Assume that is a sequence of nonnegative real numbers such that
where is a sequence in , is a sequence of nonnegative real numbers, and are two sequences in such that
- (i)
- ;
- (ii)
- ;
- (iii)
- implies for any subsequence .
Then, .
Lemma 5.
Assume that is a bounded linear operator and is the corresponding adjoint operator of A. Let be a nonexpansive mapping. If there exists a point such that , then
Proof.
It is clear that implies for all .
To see the converse, let such that . Take . Since T is nonexpansive, we have
and
Combine the above two formulas, we have
This completes the proof. □
3. The Main Results
In 2000, Moudafi [34] proposed the viscosity approximation method:
which converges strongly to a fixed point of the nonexpansive mapping N (see [35,36]). In 2004, Xu [29] further proved that is also the unique solution of the following variational inequality problem:
where is a -contraction. By Lemma 2, we get is strongly monotone, hence the solution of problem (13) is unique.
In this section, we present a viscosity iterative method for solving problem (4). Meanwhile, the algorithm approximates the unique fixed point of variational inequality problem (13).
Putting we can rewrite the iteration (11) as follows:
where
Since U is nonexpansive and h is contractive, it is easy to get
Theorem 1.
Let , be two real Hilbert spaces and be a bounded linear operator with , where is the adjoint of A. Suppose that and are two averaged mappings with the coefficients and , respectively. Assume that the problem is consistent (i.e., ). Let be a ρ-contraction with . For any , define the sequence by . If the following conditions are satisfied:
- (i)
- and ;
- (ii)
- ;
- (iii)
- .
Then, the sequence converges strongly to a point , which is also the unique solution of the variational inequality problem .
Proof.
Set . Then, by Proposition 1, it follows that is - as .
Step 1. Show that is bounded. For any , we have
Note that the condition (iii) and (15) imply that and, from the conditions (i), (iii) and , it is easy to show that is bounded. Therefore, there exists , such that . Thus, since the induction argument shows that
it turns out that the sequence is bounded and so are , and .
Step 2. Show that, for any sequence if then First, if , then we have
where
Second, we can rewrite as
where and is nonexpansive. By the condition (ii), we get . Thus, it follows from (14), (17) and (18) that
Furthermore, set
Using the condition (i), it is easy to get and . In order to complete the proof, from Lemma 4, it suffices to verify that as , which implies that
for any subsequence . Indeed, as implies that as from the condition (iii). Thus, from (18), it follows that
Step 3. Show that
where is the set of all weak cluster points of . To see (21), we prove the following:
Take and assume that is a subsequence of weakly converging to . Without loss of generality, we still use to denote . Assume . Then, we have . Setting , we deduce that
Since as , it follows immediately from (22) that
as . Thus, we have
Using Lemma 3, we get . Since both U and T are averaged, it follows from Proposition 1 (ii) that
Then, by Lemma 5, we obtain immediately. Meanwhile, we have
In addition, since is the unique solution of the variational inequality problem (13), we have
together with (20) and hence . This completes the proof. □
Next, we consider the bounded perturbation of (14) generated by the following iterative process:
Theorem 2.
Assume that the sequences and satisfy the condition . Let , be two real Hilbert spaces and be a bounded linear operator with , where is the adjoint of A. Suppose that and are two averaged mappings with the coefficients and , respectively. Assume that problem is consistent (i.e., ). Let be a ρ-contraction with . For any , define the sequence by . If the following conditions are satisfied:
- (i)
- and ;
- (ii)
- ;
- (iii)
- .
Then, the sequence converges strongly to , where is a solution of the problem , which is also the unique solution of the variational inequality problem .
Proof.
Now, put
Then, Equation (25) can be rewritten as follows:
In fact, by Proposition 1 (iii) and the nonexpansiveness of T, it is not hard to show that is Lipschitz. Thus, we have
From the condition (iii) and condition (6), we have . Consequently, using Theorem 1, it follows that the algorithm (14) is bounded perturbation resilient. This completes the proof. □
4. Numerical Results
In this section, we consider the following numerical examples to present the effectiveness, realization and convergence of Theorems 1 and 2:
Example 1.
Let . Suppose and
Take and , where C and Q are defined as follows:
and
where denotes the element of y.
We can compute the solution set
Take the experiment parameters and in the following iterative algorithms and the stopping criteria is According to the iterative process of Theorem 1, the sequence is generated by
As , we have . Then, taking the random initial guess and using MATLAB software (MATLAB R2012a, MathWorks, Natick, MA, USA), we obtain the numerical experiment results in Table 1.
Table 1.
, results without bounded perturbation.
Next, we consider the algorithm with bounded perturbation resilience. Choose the the bounded sequence and the summarable nonnegative real sequence as follows:
and
for some , where the indicator function
and
is the normal cone to C. The point is taken from . Setting the numerical results can be seen in Table 2.
Table 2.
, results with bounded perturbation.
As we have seen above, the accuracy of the solution is improved with the decrease of the stop criteria. In addition, the sequence converges to the point , which is a solution of the numerical example. Of course, it is also the unique solution of the variational inequality
In addition, we contrast the approximate value of solution of Example 1 under the same parameter conditions, the same iterative numbers and the same initial value. The numerical results are reported in Table 3 and Table 4, where and denote the iterative sequences generated by the algorithm (14) in this paper and Theorem 3.2 in Ref. [8], respectively.
Table 3.
, results without bounded perturbation.
Table 4.
, results with bounded perturbation.
Example 2.
Let . Suppose and
Define by
It is obvious that T is and the set of fixed points is nonempty. Let and Then, we use the iterative algorithm of Theorem 1 to approximate a point such that .
Take the experiment parameters and in the following iterative algorithms. Let and the stopping criteria is
Then, taking the random initial guess and using MATLAB software, we obtain the numerical experiment results in Table 5.
Table 5.
, results without bounded perturbation.
Next, we consider the bounded perturbation. The definitions of and are similar to the Example 1. Setting the numerical results can be seen in Table 4.
As we have seen in Table 5 and Table 6, the sequence approximates to the point , which is a solution of the numerical example. Of course, it is also the unique solution of the variational inequality
Table 6.
, results with bounded perturbation.
5. Conclusions
The SCFPP is an inverse problem that consists in finding a point in a fixed point set such that its image under a bounded linear operator belongs to another fixed point set. Many iterative algorithms have been developed to solve these kinds of problems. In this paper, we have introduced a viscosity iterative sequence and obtained the strong convergence. We prove the main result using the weaker conditions than many existing similar methods—for example, Xu’s algorithm [37] for the SFP. More specifically, his algorithm generates a sequence via the following recursions:
where u is a a fixed element and satisfies the assumptions:
- (i)
- and ,
- (ii)
- either or
The second condition is not necessary in our theorems. We also consider the bounded perturbation resilience of the proposed method and get theoretical convergence results. Finally, numerical experiments have been presented to illustrate the effectiveness of the proposed algorithms.
Author Contributions
All authors read and approved the final manuscript. Conceptualization, P.D.; Data Curation, P.D. and X.Z.; Formal Analysis, P.D. and J.Z.
Funding
This work was supported of the scientific research project of the Tianjin Municipal Education Commission (2018KJ253) and the Fundamental Research Funds for the Central Universities (3122017072).
Acknowledgments
The authors would like to thank the referee for valuable suggestions to improve the manuscript.
Conflicts of Interest
The authors declare that they have no competing interests.
References
- Byrne, C. A united treatment of some iterative algorithms in signal processing and image reconstruction. Inverse Probl. 2004, 20, 103–120. [Google Scholar] [CrossRef]
- Censor, Y.; Elfving, T.; Kopf, N.; Bortfeld, T. The multiple-sets split feasibility problem and its applications for inverse problems. Inverse Probl. 2005, 21, 2071–2084. [Google Scholar] [CrossRef]
- Censor, Y.; Bortfeld, T.; Martin, B.; Trofimov, A. A unified approach for inversion problems in intensity modulated radiation therapy. Phys. Med. Biol. 2006, 51, 2353–2365. [Google Scholar] [CrossRef] [PubMed]
- Censor, Y.; Elfving, T. A multiprojection algorithm using Bregman projections in product space. Numer. Algorithms 1994, 8, 221–239. [Google Scholar] [CrossRef]
- Byrne, C. Iterative oblique projection onto convex sets and the split feasibility problem. Inverse Probl. 2002, 18, 441–453. [Google Scholar] [CrossRef]
- Censor, Y.; Segal, A. The split common fixed point problem for directed operators. J. Convex Anal. 2009, 16, 587–600. [Google Scholar]
- Moudafi, A. A note on the split common fixed-point problem for quasi-nonexpansive operators. Nonlinear Anal. 2011, 74, 4083–4087. [Google Scholar] [CrossRef]
- Kraikaew, R.; Saejung, S. On split common fixed point problems. J. Math. Anal. Appl. 2014, 415, 513–524. [Google Scholar] [CrossRef]
- Dong, Q.L.; He, S.N.; Zhao, J. Solving the split equality problem without prior knowledge of operator norms. Optimization 2015, 64, 1887–1906. [Google Scholar] [CrossRef]
- He, S.N.; Tian, H.; Xu, H.K. The selective projection method for convex feasibility and split feasibility problems. J. Nonlinear Convex Anal. 2018, 19, 1199–1215. [Google Scholar]
- Padcharoen, A.; Kumam, P.; Cho, Y.J. Split common fixed point problems for demicontractive operators. Numer. Algorithms 2018. [Google Scholar] [CrossRef]
- Zhao, J. Solving split equality fixed-point problem of quasi-nonexpansive mappings without prior knowledge of operators norms. Optimization 2015, 64, 2619–2630. [Google Scholar] [CrossRef]
- Zhao, J.; Hou, D.F. A self-adaptive iterative algorithm for the split common fixed point problems. Numer. Algorithms 2018. [Google Scholar] [CrossRef]
- Byrne, C.; Censor, Y.; Gibali, A.; Reich, S. Weak and strong convergence of algorithms for the split common null point problem. J. Nonlinear Convex Anal. 2012, 13, 759–775. [Google Scholar]
- Combettes, P.L. Solving monotone inclusions via compositions of nonexpansive averaged operators. Optimization 2004, 53, 475–504. [Google Scholar] [CrossRef]
- Chuang, C.S.; Lin, I.J. New strong convergence theorems for split variational inclusion problems in Hilbert spaces. J. Inequal. Appl. 2015, 176, 1–20. [Google Scholar] [CrossRef]
- Dong, Q.L.; Yuan, H.B.; Cho, Y.J.; Rassias, T.M. Modified inertial Mann algorithm and inertial CQ-algorithm for nonexpansive mappings. Optim. Lett. 2018, 12, 87–102. [Google Scholar] [CrossRef]
- Censor, Y.; Motova, A.; Segal, A. Perturbed projections and subgradient projections for the multiple-sets split feasibility problem. J. Math. Anal. Appl. 2007, 327, 1244–1256. [Google Scholar] [CrossRef]
- Censor, Y.; Davidi, R.; Herman, G.T. Perturbation resilience and superiorization of iterative algorithms. Inverse Probl. 2010, 26, 65008. [Google Scholar] [CrossRef]
- Guo, Y.N.; Cui, W. Strong convergence and bounded perturbation resilience of a modified proximal gradient algorithm. J. Inequal. Appl. 2018, 103, 1–15. [Google Scholar] [CrossRef]
- Jin, W.; Censor, Y.; Jiang, M. Bounded perturbation resilience of projected scaled gradient methods. Comput. Optim. Appl. 2016, 63, 365–392. [Google Scholar] [CrossRef]
- Xu, H.K. Bounded perturbation resilience and superiorization techniques for the projected scaled gradient method. Inverse Probl. 2017, 33, 044008. [Google Scholar] [CrossRef]
- Dong, Q.L.; Zhao, J.; He, S.N. Bounded perturbation resilience of the viscosity algorithm. J. Inequal. Appl. 2016, 299, 1–12. [Google Scholar] [CrossRef]
- Censor, Y.; Chen, W.; Combettes, P.L.; Davidi, R.; Herman, G.T. On the effectiveness of projection methods for convex feasibility problem with linear inequality constrains. Comput. Optim. Appl. 2012, 51, 1065–1088. [Google Scholar] [CrossRef]
- Davidi, R.; Herman, G.T.; Censor, Y. Perturbation-resilient block-iterative projection methods with application to image reconstruction from projections. Int. Trans. Oper. Res. 2009, 16, 505–524. [Google Scholar] [CrossRef] [PubMed]
- Davidi, R.; Censor, Y.; Schulte, R.W.; Geneser, S.; Xing, L. Feasibility-seeking and superiorization algorithm applied to inverse treatment plannning in rediation therapy. Contemp. Math. 2015, 636, 83–92. [Google Scholar]
- Moreau, J.J. Proprietes des applications ‘prox’. C. R. Acad. Sci. Paris Ser. A Math. 1963, 256, 1069–1071. [Google Scholar]
- Marino, G.; Scardamaglia, B.; Karapinar, E. Strong convergence theorem for strict pseudo-contractions in Hilbert spaces. J. Inequal. Appl. 2016, 134, 1–12. [Google Scholar] [CrossRef]
- Xu, H.K. Viscosity approximation methods for nonexpansive mappings. J. Math. Anal. Appl. 2004, 298, 279–291. [Google Scholar] [CrossRef]
- Xu, H.K. Averaged mappings and the gradient-projection algorithm. J. Optim. Theory Appl. 2011, 150, 360–378. [Google Scholar] [CrossRef]
- Geobel, K.; Kirk, W.A. Topics in Metric Fixed Point Theory; Cambridge Studies in Advanced Mathematics 28; Cambridge University Press: Cambridge, UK, 1990. [Google Scholar]
- He, S.N.; Yang, C.P. Solving the variational inequality problem defined on intersection of finite level sets. Abstr. Appl. Anal. 2013, 2013, 942315. [Google Scholar] [CrossRef]
- Yang, C.P.; He, S.N. General alterative regularization methods for nonexpansive mappings in Hilbert spaces. Fixed Point Theory Appl. 2014, 2014, 203. [Google Scholar] [CrossRef]
- Moudafi, A. Viscosity approximation methods for fixed point problems. J. Math. Anal. Appl. 2000, 241, 46–55. [Google Scholar] [CrossRef]
- Cho, Y.J.; Qin, X. Viscosity approximation methods for a finite family of m-accretive mappings in reflexive Banach spaces. Positivity 2008, 12, 483–494. [Google Scholar] [CrossRef]
- Qin, X.; Cho, Y.J.; Kang, S.M. Viscosity approximation methods for generalized equilibrium problems and fixed point problems with applications. Nonlinear Anal. 2010, 72, 99–112. [Google Scholar] [CrossRef]
- Xu, H.K. A variable Krasnosel’skii-Mann algorithm and the multiple-set split feasibility problem. Inverse Probl. 2006, 22, 2021–2034. [Google Scholar] [CrossRef]
© 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).