Abstract
In this paper, we consider the solving of an equilibrium problem over the common fixed set of cutter mappings in a real Hilbert space. To this end, we present a subgradient-type extrapolation cyclic method. The proposed method is generated based on the ideas of a subgradient method and an extrapolated cyclic cutter method. We prove a strong convergence of the method provided that some suitable assumptions of step-size sequences are assumed. We finally show the numerical behavior of the proposed method.
1. Introduction
Let be a real Hilbert space with inner product and its induced norm . In this paper, we present an iterative method to the equilibrium problem over the intersection of fixed-point set:
Problem 1 (BEP).
Let , , be cutters with , and let be a bifunction satisfying for all . Then, our objective is to find a point such that
where denotes the set of fixed points of .
The equilibrium problem, which was first introduced by Fan [1], includes many problems as particular cases, for example, the fixed-point problem, the variational inequality, the optimization problem, the saddle point problem, the Nash equilibrium problem in non-cooperative games, and others; see, for instance, [2,3,4,5] and references therein.
The equilibrium problems over the fixed-point set have been considered in many articles; see, for instance, [6,7,8,9,10] and references therein. The computational algorithms for solving these kinds of problems have been studied and developed by using the idea of the methods for equilibrium problems and the iterative schemes for fixed-point problems. In particular, Iiduka and Yamada [6] considered the equilibrium problems over the fixed-point set of a firmly nonexpansive mapping and presented a subgradient-type method for solving the considered problems. They showed the convergence of their method and applied the method to the Nash equilibrium problems. After that, the equilibrium problems over the common fixed-point of nonexpansive mappings were considered by Duc and Muu in [7]. They proposed the splitting algorithm, which was updated based on the idea of the classical gradient method and the Krasnosel’skii–Mann method and presented the strong convergence of the presented algorithm. Recently, Thuy and Hai [8] considered the bilevel equilibrium problems and proposed the projected subgradient algorithm to solve the considered problem. They exhibited the strong convergence of the proposed method and applied it to the equilibrium problems over the fixed-point set of a nonexpansive mapping. We notice that the aforementioned literature is considered in the case of the equilibrium problems over the fixed-point set of nonexpansive mappings.
Let us focus on the constrained set of BEP. Now, let , , be cutter operators. The common fixed-point problem is to find a point
The well-known methods of finding a point that belongs to the intersection of fixed-point sets are initially motivated by the cyclic projection method, which was introduced by Kaczmarz [11]. After that, the convergence of cyclic projection-type methods are investigated in several directions and their convergence results are guaranteed under the operators’ assumptions, such as cutters or nonexpansive operators, see [12,13,14,15,16] and references therein. In particular, Bauschke and Combettes [17] proposed the cyclic cutter method and showed a weak convergence of the proposed method. In [18], Cegielski and Censor presented the extrapolated cyclic cutter method, which is an acceleration of the cyclic cutter method by imposing an appropriate step-size function to the method. Indeed, let , , be cutters with , we define the step-size function as follows:
where the operator T, and are defined as
It was shown that the extrapolated cyclic cutter method weakly converges provided that the cutter operators satisfied the demi-closedness principle for all . Note that, for some practical problem, the value of the extrapolation function may be huge, which lead to some numerical instabilities. To keep away from these instabilities, Cegielski and Nimana [19] proposed the modified extrapolated subgradient projection method for solving the convex feasibility problem, which is a particular situation of common fixed-point problem, and established the convergence of the proposed method as well as demonstrated the performance of the method by the numerical results. After that, the authors in [20] also utilized the idea of the extrapolated cyclic cutter method for dealing with the variational inequality problem with common fixed-point constraints. It can be noted that from [19,20], the iterative methods with extrapolated cyclic cutter terms achieve not only some numerical superiorities to utilizing the classical cyclic cutter scheme but also guarantee the boundedness of the generated sequence, see [20] for further discussions.
In this paper, we propose an iterative algorithm called the Subgradient-type extrapolation cyclic method for solving the equilibrium problems over the intersection of fixed-point sets of cutter operators. The proposed algorithm can be considered as a combination of the subgradient iterative scheme for equilibrium problems in [8] and the extrapolated cyclic cutter method for the intersection of fixed-point sets of cutter operators in [18]. Using the cutter operators and assumptions, we investigate the convergence of the presented algorithm. Moreover, we also present a numerical result of our presented method to illustrate the efficiency of the method.
This paper is organized as follows. In Section 2, we recall some definitions and tools which are needed for our convergence work. In Section 3, we present the subgradient-type extrapolation cyclic method for finding the solution of BEP. We subsequently present the convergence result in this section. In Section 4, the efficacy of the subgradient-type extrapolation cyclic method is illustrated by numerical experiments of the solving equilibrium problem governed by the positive definite symmetric matrices over the common fixed-point set. Finally, we give some concluding remarks in Section 5.
2. Preliminaries
In section, we collect some basic definitions, properties, and useful tools for our work. The readers can consult the books [16,21] for more details.
We denote by the identity operator on a real Hilbert space . For a sequence , the strong and weak convergences of a sequence to a point are defined by the expression and , respectively.
In what follows, we recall some definitions and properties of the operator that will be referred to in our analysis.
Definition 1
([16]). Let be an operator having a fixed point. The operator T is called
- (i)
- quasi-nonexpansive, iffor all and
- (ii)
- η-strongly quasi-nonexpansive, if there exists ,for all and ,
- (iii)
- cutter, iffor all and
Lemma 1
[16] (Remark 2.1.31 and Theorem 2.1.39). Let be an operator having a fixed point. Then the following statements are equivalent:
- (i)
- T is cutter.
- (i)
- for all and for all .
- (ii)
- T is 1-strongly quasi-nonexpansive.
Definition 2
[16] (Definition 3.2.6). Let be an operator having a fixed point. The operator T is said to satisfy the demi-closedness (DC) principle if for every sequence , and , we have .
Definition 3
[16] (Definition 2.1.2). Let be an operator and be given. We define the relaxation of the operator T by
and we call λ a relaxation parameter.
Next, we recall the definition of a generalization of relaxation of an operator.
Definition 4
[16] (Definition 2.4.1). Let be an operator, and . We define the operator by
This operator T is called a generalized relaxation of the operator T, the value λ is called a relaxation parameter and the function σ is called a step-size function. The operator is called an extrapolation of if the function for every .
We notice that, if , for every , then . Note that . Then, for every , the following relations hold
and for any , we have
Next, we provide an important lemma of the step-size function for proving the convergence result.
Lemma 2
[16] (Section 4.10). Let be cutters with and denote the operator T, and as in (2). Let the function be given by (1). Then the following statements are true:
- (i)
- For every , it is true that
- (ii)
- The operator is a cutter.
Now, we recall a notion and some properties of a diagonal subdifferential which will be used in this work.
A function is said to be subdifferentiable at if there exists a vector such that
The vector w is called a subgradient of the function h at . The collection of all such vectors constitute the subdifferential of h at and is denoted by
Let be a bifunction which is convex in the second argument, that is, the function is convex at x, for all Then, the set of all subgradient of at x is called the diagonal subdifferential and is denoted by . The reader can find the notion of the diagonal subdifferential in [22], for more detail.
We end this section by recalling some technical lemmas that are important tools in proving our convergence result.
Lemma 3
[23] (Lemma 3.1). Let and be sequences of nonnegative real numbers such that
If , then exists.
Lemma 4
[24] (Lemma 3.1). Let be a sequence of real numbers such that there exists a subsequence of with for all . If, for all , we define
then the following hold:
- (i)
- is non-decreasing.
- (ii)
- .
- (iii)
- and for every .
3. Algorithm and Its Convergence Result
In this section, we firstly propose the subgradient-type extrapolation cyclic method for solving BEP. Subsequently, we present useful lemmas and prove the main convergence theorem.
Remark 1.
- (i)
- When the number and , Algorithm 1 becomes Algorithm 2 considered in [8]. Moreover, it is worth noting that the class of operator considered in this work is different from [8]. In fact, we consider the cutter property of , whereas the nonexpansiveness of T is assumed in [8].
- (ii)
- If the function , Algorithm 1 is reduced towhere for all . In the case when for all , this scheme is related to the extrapolated cyclic cutter proposed by [18]. Moreover, this scheme is also related to the work of Cegielski and Nimana [19] for solving a convex feasibility problem when the operator is omitted in their paper.
The following assumption relating to the convergence of Algorithm 1 is assumed throughout this work.
| Algorithm 1: Subgradient-type extrapolation cyclic method |
Initialization: Given the positive real sequences and . Choose and arbitrarily. Step 1. For given , compute the step size as
Step 2. Update the next iterate as
Put and go to Step 1. |
Assumption A1.
Assume that
- (A1)
- The bifunction f is ρ-strongly monotone on , that is, there exists a constant satisfying
- (A2)
- For each , the function is convex, subdifferentiable and lower semicontinuous on .
- (A3)
- The function is bounded on a bounded subset of .
- (A4)
- The sequences and satisfy , , , and .
Remark 2.
- (i)
- If the whole space is finite dimensional, the assumption that, for all , the function is subdifferentiable and weakly lower semicontinuous in (A2) can be omitted. This is because, in the finite dimensional setting, the convexity implies the continuity of a function.
- (ii)
- The convexity of the function implies that the lower semicontinuity is equivalent to the weak lower semicontinuity of the function for all .
- (iii)
- If the whole space is finite dimensional, by invoking the assumption (A2), we have the diagonal subdifferential is nonempty for all . Moreover, in this case, the assumption (A3) can be omitted, see [21] (Proposition 16.20).
- (iv)
- An example of the corresponding step-size sequences in (A4) is the positive real sequences and given bywhere and with and . In fact, since and , we have and then . Furthermore, since , we have . We note that . Moreover, we have that .
The following lemma states the important relation of the generated iterates.
Lemma 5.
Let be the sequence generated by Algorithm 1. Then, for every and , it holds that
Proof.
Let be fixed. Now, let us note that
By using the properties of in (3), Lemma 1 and Lemma 2, we note that
Finally, by utilizing the fact that , we obtain
as desired. □
The following lemma guarantees the boundedness of the constructed sequence .
Lemma 6.
The sequence generated by Algorithm 1 is bounded.
Proof.
Let and be fixed. Let us notice that
which together with Lemma 5 yields
Now, we set for all . Thus, the relation (4) can be rewritten as
To show that the sequence is bounded, we will consider the proof in two cases:
Case I: Suppose that there exists such that the sequence is nonincreasing for all . Then, we obtain that for all , which means that the sequence is bounded and, subsequently, is also a bounded sequence.
Case II: Suppose that there exists a subsequence of such that for all , and let be given in Lemma 4. This yields, for every , that
and
Invoking the relation (6) in the inequaluty (5) and the positivity of the sequences , and , we obtain that
On the other hand, by using the definition of and the fact that , we get
This together with the inequality (8) yields that
Now, it follows from the -strong monotonicity of f that
This means that the sequence is bounded. Now, since
it follows that is bounded above. Thus, by using (7), we get that is bounded and hence is also bounded. This completes the proof. □
The following lemma provides some important boundedness properties of the sequences and .
Lemma 7.
The sequences and are bounded.
Proof.
Let and be fixed. Now, let note that
where the first inequality holds true by (3) and the last one holds true by the fact that is a cutter and consequently a quasi-nonexpansive operator.
As , we have from Assumption (A3) and the boundedness of that the sequence is bounded which implies the sequence is also bounded. Consequently, from the definition of the sequence , it can be seen that is also bounded. □
Now, we are in a position to present our main theorem.
Theorem 1.
Let be the sequence generated by Algorithm 1. Suppose that Assumption A1 is satisfied and the operators , , satisfy the DC principle. Then the sequence converges strongly to the unique solution ofBEP.
Proof.
Let be the unique solution of BEP. Firstly, we note from Lemma 5 with replacing that
For simplicity, we denote for all . Then the inequality (10) is nothing else than
To obtain the strong convergence of the generated sequence, we investigate the proof in two cases based on the behavior of the sequence .
Case I: Suppose that there is such that for every . Thus, by using the definition of , we note that
By utilizing Lemma 3 and the fact that , we obtain that the sequence is convergent and
Now, as , we get that
As the sequence is bounded, we have that, for all ,
On the other hand, we note from Lemma 5 and the fact that , for all , that
By summing up this relation and the condition that , we obtain
Now, since the sequence is bounded, there is a real number such that for all . This together with the assumption implies that
Next, we show that Suppose to the contrary that there exist and such that for all . Then,
which leads to a contradiction. Thus, we obtain
From the -strongly monotonicity of f, it follows that
Then,
By taking the inferior limit, we have
Combining this and the inequality (15), we have
Since the sequence is bounded, there exist a weakly cluster point and a subsequence of such that . We note from (12) that
Thus, by using the DC principle of , we have that Furthermore, since and it holds that
which imply that Moreover, we note that
By utilizing the DC principle of , we have
By processing the similar argument as above, we acquire that for all and hence .
In virtue of the weak lower semicontinuity of , we obtain
Case II: Suppose that there exists a subsequence of such that for all . By Lemma 4, there exists a sequence of indices such that, for all ,
and
Then,
By using the definition of and the fact that , we get
which implies that
Now, using the -strongly monotonicity of f that
and for a fixed such that
we obtain
By using the boundedness of and , we obtain
This implies that
By means of the fact that in (20), it follows that
Combining this and the above inequality, we obtain
By taking the superior limit, we have
As the sequence is bounded, there exist a weakly cluster point and a subsequence of such that . By following the argument as used in Case I together with the fact (22) and the DC principle of each , we obtain that, for any subsequence of , .
By using the weak lower semicontinuity of , we obtain
It follows from the inequality (23) that
Then, we obtain
Note that from the definition of and using the fact that , we have
Combining this and using the triangle inequality, we have
By using the inequality (24) and the fact that , we obtain
Next, using the inequality (19) and the fact that , we have
Finally, by using the inequality (25), and the fact , we obtain
This completes the proof. □
Remark 3.
The DC principle assumption, which is assumed in the Theorem 1 holds true when the operators , , are nonexpansive. Actually, the metric projection onto closed convex sets and the subgradient projections of a continuous convex function, which is Lipschitz continuous on every bounded subset also satisfy the DC principle, see [16] further details.
Remark 4.
It can be noted that the convergence result obtained in Theorem 1 holds true without any boundedness assumption of the generated sequence as in the previous works, for instance [20]. This underlines the convergence improvements accomplished in this work.
4. A Numerical Example
In this section, we present a numerical example for solving the equilibrium problem over a finite number of half-space constraints. Let be matrices, , and be given for all , we consider the following equilibrium problem: find a point such that
where the constrained set is
We consider the operator in two cases. In the first case, we put to be the subgradient projection defined by
where with the distance function is given by . In the second case, we put , the metric projection onto , for all . Note that it is known that the operators and are cutters and satisfy the DC principle with . We consider positive definite symmetric matrices A and B defined by , , where the matrices are randomly generated in , and is the identity matrix. Note that the bifunction is strongly monotone on , and for fixed , we have is convex on . Moreover, we note that the diagonal subdifferential , and we also know that the function is bounded on a bounded subset of . These mean that the assumptions (A1)–(A3) are now satisfied. In this case, the problem (27) is the particular case of Problem 1 so that the sequence generated by Algorithm 1 can be applied to solve the problem.
We consider behavior of the sequence generated by Algorithm 1 for various positive real sequences and in the forms of Remark 3. We choose , and generate a vector in by uniformly distributed random generating between and a scalar , for all . We choose the starting point of Algorithm 1 to be a vector whose coordinates are one. We terminate Algorithm 1 by the stopping criterions
In the first experiment, we fix the parameters and . We perform 10 independent tests for any collections of parameters , and and , and when utilizing the operator and and the results are presented respectively in Table 1 and Table 2, where the average number of iterations and the average computational runtime for any collection of and are presented.
Table 1.
Influence of parameters and when using the subgradient projection operator where and .
Table 2.
Influence of parameters and when using the metric projection where and .
In Table 1, we presented the number of iterations (k) (Iter), the computational time (Time) in seconds when the stopping criteria of Algorithm 1 was met. Note that the larger requires a larger number of iterations and computational runtime. Furthermore, the best choice of the involved parameters for both cases is and .
In a similar fashion with Table 1, we also presented in Table 2 the number of iterations (k) (Iter), the computational time (Time) in seconds, when the stopping criterions of Algorithm 1 when using the operator was met. The experimented results are in the same direction with Table 1 where the best choice of the involved parameters for both cases is and .
In the next experiment, we consider the influence of parameters a and b by fixing the best parameters and . We performed 10 independent tests for any collections of parameters , , , , , to and , , , , , , , and when utilizing the operator and and the results of the average number of iterations and the average computational runtime for any collection of a and b are presented in Table 3 and Table 4, respectively. We omit the combinations that do not satisfy the assumption in Theorem 1 and label it by -.
Table 3.
Influence of parameters a and b when using the subgradient projection where and .
Table 4.
Influence of parameters a and b when using the metric projection where and .
In Table 3, we see that the numbers of iterations as well as computational running time decrease when the value a increases. The the best result is obtained for the combination of and .
In the same direction as the results in Table 3, it can be seen from Table 4 that the numbers of iterations and the computational running time is decreases when the values a grow up. The the best result is acquired for the combination of and .
From these all above experiment, we observe that the choice of corresponding parameters , , and yields the best performance of both considered cases.
In the next experiment, we consider the behavior of Algorithm 1 for various values n and m by fixing the corresponding parameters as the above best choice. We also terminate Algorithm 1 when the error tolerance was met, and the results are presented in Table 5.
Table 5.
Comparisons between the using of the subgradient projection and the metric projection for different sizes of n and m.
It is observed from Table 5 that for the values , and 500, the using of the subgradient projection is more efficient than using the metric projection in the sense that the first one requires less computation than the second one in the average number of iterations for all values m. In the case of , we observe that there is no difference on these two cases. One notable behavior is that for each value n, we observe that even if the value m increases, the average numbers of iterations are almost the same, whereas the average computational runtime is increasing.
Finally, we present the comparison of the use of the subgradient projection and the metric projection for various optimality tolerances . We set and and choose the corresponding parameters in the same manner as above, the average numbers of iterations with respect to the optimality tolerances are presented in Figure 1.
Figure 1.
Comparison between the using of the subgradient projection and the metric projection with different errors of tolerance.
The plots in Figure 1 show that using the subgradient projection is more efficient than the metric projection for all the optimality tolerances. This emphasizes the superiority of using the subgradient projection when performing Algorithm 1.
5. Conclusions
In this work, we consider the solving of the bilevel equilibrium problem governed by a strongly monotone bifunction over the intersection of fixed-point sets of cutter operators. We associated with it the so-called subgradient-type extrapolation cyclic method. We present that the generated sequence generated by the proposed method converges to the unique solution to the problem. Our numerical experiment showed that using appropriate operators can yield a better convergence behavior to the proposed method.
It can be seen that the proposed subgradient-type extrapolation cyclic method (Algorithm 1) allows us to compute the operator , sequentially. The main advantage of our method is that the computing machine is not necessary to store information while computing. Notwithstanding, the nature of the cyclic method, it is well-known that to compute , the cyclic method needs to have the estimate in hand. This means that there has a waiting process while performing the method. In this case, one may consider the simultaneous extrapolation method [25] when dealing with the common fixed-point constrained of BEP.
Author Contributions
Conceptualization, P.P. and N.N.; methodology, P.P. and N.N.; software, N.N.; validation, P.P. and N.N.; convergence analysis, P.P. and N.N.; investigation, P.P. and N.N.; writing—original draft preparation, P.P. and N.N.; writing—review and editing, P.P. and N.N.; visualization, P.P. and N.N.; supervision, N.N.; project administration, N.N.; funding acquisition, N.N. All authors have read and agreed to the published version of the manuscript.
Funding
This research was supported by the Fundamental Fund of Khon Kaen University. This research has received funding support from the National Science, Research and Innovation Fund or NSRF.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The data presented in this study are available on request from the corresponding author.
Acknowledgments
The authors are thankful to the Editor and three anonymous referees for comments and remarks which improved the quality and presentation of the paper.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Fan, K. A minimax inequality and applications. In Inequality III; Shisha, O., Ed.; Academic Press: New York, NY, USA, 1972; pp. 103–113. [Google Scholar]
- Blum, E.; Oettli, W. From optimization and variational inequalities to equilibrium problems. Math. Student 1994, 63, 123–145. [Google Scholar]
- Muu, L.D.; Oettli, W. Convergence of an adaptive penalty scheme for finding constrained equilibria. Nonlinear Anal. 1992, 18, 1159–1166. [Google Scholar] [CrossRef]
- Mastroeni, G. On auxiliary principle for equilibrium problems. In Equilibrium Problems and Variational Models; Daniele, P., Giannessi, F., Maugeri, A., Eds.; Kluwer Academic: Dordrecht, The Netherlands, 2003; pp. 289–298. [Google Scholar]
- Nguyen, T.T.V.; Strodiot, J.J.; Nguyen, V.H. The interior proximal extragradient method for solving equilibrium problems. J. Glob. Optim. 2009, 44, 175–192. [Google Scholar] [CrossRef]
- Iiduka, H.; Yamada, I. A subgradient-type method for the equilibrium problem over the fixed point set and its applications. Optimization 2009, 58, 251–261. [Google Scholar] [CrossRef]
- Duc, P.M.; Muu, L.D. A splitting algorithm for a class of bilevel equilibrium problems involving nonexpansive mappings. Optimization 2016, 65, 1855–1866. [Google Scholar] [CrossRef]
- Thuy, L.Q.; Hai, T.N. A Projected Subgradient Algorithm for Bilevel Equilibrium Problems and Applications. J. Optim. Theory Appl. 2017, 175, 411–431. [Google Scholar] [CrossRef]
- Anh, P.N.; Van Hong, N. New projection methods for equilibrium problems over fixed point sets. Optim. Lett. 2021, 15, 627–648. [Google Scholar] [CrossRef]
- Kim, J.K.; Anh, P.N.; Hai, T.N. The Bruck’s ergodic iteration method for the Ky Fan inequality over the fixed point set. Int. J. Comput. Math. 2017, 94, 2466–2480. [Google Scholar]
- Kaczmarz, S. Angenäherte Auflösung von Systemen linearer Gleichungen. Bull. Int. 1937, A35, 355–357, reprinted in Int. J. Control 1993, 57, 1269–1271. [Google Scholar]
- Opial, Z. Weak convergence of the sequence of successive approximations for nonexpansive mappings. Bull. Am. Math. Soc. 1967, 73, 591–597. [Google Scholar] [CrossRef]
- Cegielski, A.; Censor, Y. Opial-Type Theorems and the Common Fixed Point Problem. In Fixed-Point Algorithms for Inverse Problems in Science and Engineering; Bauschke, H., Burachik, R., Combettes, P., Elser, V., Luke, D., Wolkowicz, H., Eds.; Springer Optimization and Its Applications; Springer: New York, NY, USA, 2011; Volume 49. [Google Scholar]
- Liu, C. An acceleration scheme for row projection methods. J. Comput. Appl. 1995, 57, 363–391. [Google Scholar] [CrossRef][Green Version]
- Cegielski, A. Generalized relaxations of nonexpansive operators and convex feasibility problems. Contemp. Math. 2010, 513, 111–123. [Google Scholar]
- Cegielski, A. Iterative Methods for Fixed Point Problems in Hilbert Spaces; Lecture Notes in Mathematics, 2057; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
- Bauschke, H.H.; Combettes, P.L. A weak-to-strong convergence principle for Fejér-monotone methods in Hilbert spaces. Math. Oper. Res. 2001, 26, 248–264. [Google Scholar] [CrossRef]
- Cegielski, A.; Censor, Y. Extrapolation and local acceleration of an iterative process for common fixed point problems. J. Math. Anal. Appl. 2012, 394, 809–818. [Google Scholar] [CrossRef]
- Cegielski, A.; Nimana, N. Extrapolated cyclic subgradient projection methods for the convex feasibility problems and their numerical behaviour. Optimization 2019, 68, 145–161. [Google Scholar] [CrossRef]
- Prangprakhon, M.; Nimana, N. xtrapolated sequential constraint method for variational inequality over the intersection of fixed-point sets. Numer. Algorithms 2021, 88, 1051–1075. [Google Scholar] [CrossRef]
- Bauschke, H.H.; Combettes, P.L. Convex Analysis and Monotone Operator Theory in Hilbert Spaces, 2nd ed.; CMS Books in Mathematics; Springer: New York, NY, USA, 2017. [Google Scholar]
- Iusem, A.N. On the maximal monotonicity of diagonal subdifferential operators. J. Convex Anal. 2011, 18, 489–503. [Google Scholar]
- Combettes, P.L. Quasi-Fejérian analysis of some optimization algorithms. In Inherently Parallel Algorithms for Feasibility and Optimization; Butnariu, D., Censor, Y., Reich, S., Eds.; Elsevier: New York, NY, USA, 2001; pp. 115–152. [Google Scholar]
- Maingé, P.E. Strong convergence of projected subgradient methods for nonsmooth and nonstrictly convex minimization. Set-Valued Anal. 2008, 16, 899–912. [Google Scholar] [CrossRef]
- Butnariu, D.; Censor, Y.; Gurfil, P.; Hadar, E. On the behavior of subgradient projections methods for convex feasibility problems in Euclidean spaces. SIAM J. Optim. 2008, 19, 786–807. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
