Abstract
Convergence results of the subgradient algorithm for equilibrium problems were mainly obtained using a Lipschitz continuity assumption on the given bifunctions. In this paper, we first provide a complexity result for monotone equilibrium problems without assuming Lipschitz continuity. Moreover, we give a convergence result of the value of the averaged sequence of iterates beyond Lipschitz continuity. Next, we derive a rate convergence in terms of the distance to the solution set relying on a growth condition. Applications to convex minimization and min–max problems are also stated. These ideas and results deserve to be developed and further refined.
MSC:
49J53; 65K10
1. Introduction
The subgradient algorithm has recently enjoyed regained popularity, see for example [1,2,3,4]. However, the Lipschitz continuity assumption can be stringent even in a convex setting. It is well known that not only SVM (Support Vector Machine) but also Feed-Forward and Recurrent Neural Networks do not satisfy the Lipschitz continuity assumption, see [5]. Indeed, most -regularized convex learning problems lack this property, while regularization and weight decay are ubiquitous in the learning field. The proposed equilibrium approach will allow us not only to generalize some very recent results in convex minimization [1] and present them for min–max problems in a unified way but also to yield new insights into equilibrium problems. In order to go to the essential to share, we took the same paper outline as in [1] and we assume the reader has some basic knowledge of variational and convex analysis as can be found, for example, in [6,7].
The subgradient algorithm is a powerful tool for constructing algorithms to approximate solutions of optimization and equilibrium problems, the latter being the problem of finding such that
where C is a given nonempty closed convex subset of and is a bifunction. The solution set of will be denoted by . Such a problem is also known as the equilibrium problem in the sense of Blum and Oettli [8]. It is worth mentioning that variational inequalities, Nash equilibrium, the saddle-point problem, optimization, and many problems arising in applied nonlinear analysis are special cases of equilibrium problems. In this paper, we will be concerned with the convergence analysis of a subgradient method for solving problem in which F is not assumed to be Lipschitz continuous. For more simplicity and clarity, we suppose that F verifies the usual conditions:
;
for all ;
;
for each is convex and lower semicontinuous.
2. The Main Results
Let us state the subgradient method which generates a sequence by Algorithm 1:
| Algorithm 1 (SGM) |
| Step 0: Choose and set . |
| Step 1: Let and compute , where stands for the convex subdifferential of F with respect to the second variable. |
| If , then stop. |
| Step 2: Update , where , for all with . |
Now we are in a position to provide a complexity result.
Theorem 1.
Suppose F verifies conditions to and note by T the total number of iterations. Set , , and and let be a solution of . Then, for with . Moreover, we have
where is the local Lipschitz continuity constant of on the smallest open convex set such that the closed ball .
Proof.
For all , we have
On the other hand, implies that , which combined with the monotonicity of F ensures that . Consequently,
Setting , we have
because together with . In other words, the sequence is quasi-Fejér monotone to the solution set .
From the latter inequality, we infer that
Relation (4) assures that .
In light of the consequence of [6]—Proposition 9.13, we have that is locally Lipschitz continuous with constant on which, for , in turn gives that . This combined with (2) leads to
From which we derive
or equivalently
Using the convexity of the function , we finally obtain
This completes the proof. □
A convergence result of the values of the averaged sequence of iterates generated by (SGM) is provided in the next Theorem.
Theorem 2.
Suppose that the bifunction F verifies conditions to . Set , , and and let be a solution of . Then, the sequence where and is such that . Moreover,
where is the local Lipschitz continuity constant of on the smallest open convex set such that .
Proof.
We again have that
Therefore,
This ensures that . Following the same lines as in the proof of Theorem 1, we obtain
This together with the convexity of the function yields to
which leads to the announced result. □
Remark 1.
We can obtain more than the convergence of the averaged sequence of the iterates. Actually, we have
and the whole sequence converges to a solution of .
Indeed, for all , the inequality
leads to
Since , this implies that .
Now, with the sequence being quasi-Fejér convergent to the set Γ, namely verifying (3) with , this implies its boundedeness. Further, in view of (9), which is still valid for all together with both lower semicontinuity and upper hemi-continuity assumptions, we obtain that any cluster point of belongs to Γ. Consequently, the whole sequence converges to , see for example [9], and we retrieve the main results in [10].
A convergence result based on a growth property.
Corollary 1.
Suppose in addition to hypotheses to that F verifies the following growth property:
Consider the sequence given by for all . Then, for all , , and we have
where is the local Lipschitz continuity constant of on the smallest open convex set such that .
Proof.
The beginning of Theorem 2 ensures that lies in and assures again that . Remember that (7) reads as
By taking , we obtain
This implies
or equivalently
Following the same lines as in the proof of [1]—Corollary 1, for all , we further have
Summing the last inequality for to k leads to
From which we derive
□
3. Applications
In the convex minimization case, (SGM) coincides with the classical subgradient method, and we recover the results obtained in the convex setting in [1]. Indeed, just take , f being a proper convex lower semicontinuous convex function, clearly , and we retrieve
Theorem 1 reduces to
where is a minimizer of f and is the local Lipschitz continuity constant of f on the smallest open convex set such that . Theorem 2, in turn, leads to the fact that the sequence , where and , is such that . Moreover, for all , we obtain the following convergence result in terms of the suboptimality error:
were defined in Theorem 2, and and are minimizers of f.
If we assume, in addition, for all that
then
Likewise, by taking with , and L being a proper closed convex–concave function defined on , with closed convex sets of and , respectively, then clearly
we recover the subgradient algorithm for the saddle function considered in [11] and more recently in [12] and we extend some results in [1] to the convex–concave case. More precisely, (SGM) reduces to
Theorem 1 reads, in this case, as
where and are a saddle point of L, namely verifying for all ( and being the local Lipschitz continuity constant of on the smallest open convex set such that .
Theorem 2, in turn, leads to the sequence , where and is such that . Moreover, for all , we obtain the following convergence result in terms of a merit function:
were defined in Theorem 2 and and are saddle points of L.
Now, if in addition, we assume for all that
then
4. Conclusions
To conclude, the equilibrium approach allowed us not only to generalize some very recent results obtained by X. Li, L. Zhao, D. Zhu, and A. M-Ch.; use them in convex minimization [3]; and present them for min–max problems in a unified way but also to yield new insights into equilibrium problems. More precisely, a complexity result for monotone equilibrium problems was provided without assuming Lipschitz continuity. Moreover, a convergence result of the value of the averaged sequence of iterates beyond Lipschitz continuity was also given. Finally, a rate convergence in terms of the distance to the solution set relying on a growth condition was derived. The proposed results were declined in the convex minimization context as well as in the convex–concave min–max setting.
Funding
This research received no external funding.
Data Availability Statement
No new data were created or analyzed in this study.
Acknowledgments
I would like to thank Zoubida Marzaq for her invitation to Saint-Sébastien and wish to thank the “Image & Modèle” team and also the Computer Science System Laboratory (L.I.S.) at Aix Marseille University for their support.
Conflicts of Interest
The author declares no conflicts of interest.
References
- Li, X.; Zhao, L.; Zhu, D.; So, A.M.C. Revisiting subgradient method: Complexity and Convergence Beyond Lipschitz Continuity. arXiv 2023, arXiv:2305.14161. [Google Scholar]
- Wang, J.; Xu, H.K. Stepsize Choice for Korpelevich’s and Popov’s extragradient Algorithms for Convex-Concave Minimax Problems. Carpathian J. Math. 2023, 39, 315–323. [Google Scholar] [CrossRef]
- Abass, H.A.; Narain, O.K.; Onifade, O.M. Inertial extrapolation method for solving systems of monotone variational inclusion and fixed point problems using Bregman distance approach. Nonlinear Funct. Anal. Appl. 2023, 28, 497–520. [Google Scholar]
- Kheawborisut, A.; Khuangsatung, W. A modified Krasnoselskii-type subgradient extragradient algorithm with inertial effects for solving variational inequality problems and fixed point problem. Nonlinear Funct. Anal. Appl. 2024, 29, 393–418. [Google Scholar]
- Patel, V.; Berahas, A.S. Gradient Descent in the Absence of Global Lipschitz Continuity of the Gradients. arXiv 2023, arXiv:2210.02418v2. [Google Scholar]
- Rockafellar, R.T.; Wets, R.J.-B. Variational Analysis; Springer: Berlin/Heidelberg, Germany, 1998. [Google Scholar]
- Bauschke, H.H.; Combettes, P.L. Convex Analysis and Monotone Operator Theory in Hilbert Spaces, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
- Blum, E.; Oettli, W. From optimization and variational inequalities to equilibrium problems. Math. Stud. 1994, 63, 123–145. [Google Scholar]
- Iusem, A.N.; Svaiter, B.F.; Teboulle, M. Entropy-like proximal methods in convex programming. Math. Oper. Res. 1994, 9, 790–814. [Google Scholar] [CrossRef]
- Neto, J.X.C.; Santos, P.S.M.; Scheimberg, S.; Souza, S.S. A Subgradient Projected Algorithm for Nonconvex Equilibrium Problems; CLAIO, SBPO: Rio de Janeiro, Brazil, 2012; pp. 3522–3529. [Google Scholar]
- Nedic, A.; Ozdaglar, A. Subgradient Methods for Saddle-Point Problems. J. Optim. Theory Appl. 2009, 142, 205–228. [Google Scholar] [CrossRef]
- Yen, L.H.; Muu, L.D. A subgradient method for equilibrium problems involving quasiconvex bifunctions. Oper. Res. Lett. 2020, 48, 579–583. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).