Abstract
In this work, we consider the problem of minimizing a quasi-convex function over a nonempty closed convex constrained set. In order to approximate a solution of the considered problem, we propose delayed star subgradient methods. The main feature of the proposed methods is that it allows us to use the stale star subgradients when updating the next iteration rather than computing the new star subgradient in every iteration. We subsequently investigate the convergence results of sequences generated by the proposed methods. Finally, we present some numerical experiments on the Cobb–Douglas production efficiency problem to illustrate the effectiveness of the proposed method.
1. Introduction
In this work, we consider the quasi-convex optimization problem:
where the objective function is quasi-convex continuous (possibly) and nondifferentiable, and the constrained set X is nonempty closed and convex. Throughout this work, we will denote the set of all optimal solutions and the optimal value of the problem (1) as and , respectively. We also assume that .
The problem (1) not only forms a fundamental setting for optimization as every convex function is quasi-convex but also generalizes to several practical situations in many fields such as economics, engineering, and others. An important special setting is the problem of minimizing the ratios of two functions [1] (Lemma 4), [2] (Theorem 2.3.8), for example, optimizing the ratio of outputs and inputs for the highest-quality productivity assessments in economics, such as debt-to-equity ratio analysis in corporate financial planning, inventory-to-sales analysis in production planning, and nurse–patient ratio analysis in hospital planning. This kind of problem is known as fractional programming [3,4,5]. As mentioned above, various kinds of iterative methods have been developed to solve the quasi-convex optimization problem (1); see [1,6,7,8,9,10,11,12,13,14] and references therein.
If the function f is convex, it is well-known that the classical method to deal with the constrained convex optimization problem is the projected subgradient method: for an initial point compute
where is the Fenchel subgradient of f at Because of the simplicity, this method has been continuously developed in the literature, see for instance [8,15]. However, for the quasi-convex case, it should be noted that the Fenchel subgradient may not exist, in general. In this situation, we need a more general subgradient concept, which is the so-called star subgradient proposed in [16,17]. By utilizing the notion of a star subgradient, Kiwiel [1] proposed the star subgradient method to solve the constrained quasi-convex optimization problem (1): for an initial point
where is the nonzero star subgradient of f at . After that, Hu et al. [6] proposed an inexact version of the star-subgradient method by considering the approximated star subgradient with the presence of noises. By imposing the Hölder condition, they presented the convergence results in various aspects for both constant and diminishing step sizes. Subsequently, several methods have been investigated for solving more general problem settings, for instance, a conditional star subgradient method [8], a stochastic star subgradient method [7], and an incremental star subgradient method [9].
As we can notice, the above-mentioned methods require computing the star subgradient in every single iteration; however, there are some practical situations such as large-scale problems in which the exact star subgradients may be computationally expensive. This shortcoming can be overcome by using the idea of delayed subgradients in which we can retrieve stale subgradients instead of spending time computing a new subgradient in every iteration. This approach is not only useful for dealing with excessive computation but also for handling communication delays in networked systems. Note that various convex methods with delayed gradient or even subgradient updates are developed and studied, for instance, stochastic gradient descent [18,19], incremental type methods [20,21,22,23], distributed type methods [24,25,26], and references therein.
To the best of our knowledge, there is no report about the star subgradient method with delayed updates to solve the problem (1). To fill this gap, in this work, we propose two delayed star subgradient methods (in short, DSSMs) based on the ideas of the star subgradient method (2) and the delayed subgradient updates, called DSSM-I and DSSM-II, respectively. We investigate the convergence properties of the proposed methods in both objective values and iterations, including proving finite convergence.
This paper is organized as follows: Section 2 provides the essential notations, definitions, and facts utilized in this work. Section 3 contains the proposed methods, including their convergence results. This section is divided into two subsections. In Section 4, we then apply DSSM-I to the Cobb–Douglas production efficiency problem. Finally, we provide the summary in Section 5.
2. Preliminaries
This section contains important symbols, definitions, and tools utilized in this work. Throughout this paper, let be the set of all real numbers, the set of all non-negative integers, and the set of all positive integers. We denote with a Euclidean space with an inner product and an induced norm . We denote the unit sphere by the set and the open ball centered at the with radius by the set .
Let be a nonempty set. The symbol is the closure of X. The distance (function) between and X is the function defined by
The metric projection of onto X is the point such that for all Note that the metric projection exists and is unique for all whenever X is nonempty, closed, and convex [27] (Theorem 1.2.3). The metric projection onto a nonempty closed convex set X has a nonexpansive property, i.e., for all [27] (Theorem 2.2.21).
Let be a function, and let . The strictly sublevel and sublevel set of f corresponding to are defined by
respectively. The function f is called upper semicontinuous on if is an open set for all The function f is called lower semicontinuous on if is a closed set for all The function f is continuous on if it is both upper semicontinuous and lower semicontinuous on . The function f is called quasi-convex on if is convex for all
For any the star subdifferential [16,17] of f at is the set
The element is called the star subgradient of f at , and it is denoted by It is obvious that for all . The following fact provides the basic properties of the star subgradient, which can be found in [16] (Proposition 30) and [6] (Lemma 2.1).
Fact 1.
Let be a quasi-convex, and let Then the following statements hold:
- (i)
- is a nonempty set;
- (ii)
- is a closed, convex cone.
Next, we will recall the definition and properties of the quasi-Fejer monotone sequence. Let X be a nonempty subset of . The sequence is said to be quasi-Fejer monotone with respect to X if for any there exists a sequence such that
Fact 2
([28]). (Theorem 5.33) Let be a nonempty set and be a sequence in . If the sequence is quasi-Fejer monotone with respect to X, then is bounded and exists for all
Finally, we provide the lemma, which is used in the next section.
Lemma 1
([29]). (Lemma 2.1) Let be a scalar sequence, and let be a positive real number sequence such that If , then .
3. Algorithms and Convergence Results
In this section, we introduce two delayed star subgradient methods (DSSM-I and DSSM-II) to solve Problem (1) and subsequently investigate the convergence properties of the proposed methods. To deal with delayed-type methods, we need the following assumption of the boundedness of the time-varying delays, as stated below.
Assumption 1.
The sequence of time-varying delays is bounded, that is, there exists a non-negative integer τ such that for all
3.1. Delayed Star Subgradient Method I
Now, we are ready to propose the delayed star subgradient method of the first kind which is defined as the following Algorithm 1.
| Algorithm 1 Delayed Star Subgradient Method I (DSSM-I) |
| Initialization: Given a stepsize the delays , and initial points . Iterative Step: For a current point , we compute Update . |
Throughout this work, to simplify the analysis, we denote
Remark 1.
- (i)
- Since the function f is quasi-convex, we note that DSSM-I is well defined. In fact, using Fact 1, we have for any , . Hence, we are able to select any point and subsequently set . Moreover, since it is clear that for all .
- (ii)
- If for all , DSSM-I coincides with the star subgradient method (SSM), which was proposed by Kiwiel [1].
Remark 2.
Note that methods with time-varying delays allow us to use stale information, which is very helpful when the star subgradients are not easily computed. Assumption 1 is typically assumed for analyzing the convergence results of the delayed-type methods, see [18,19,20,24,25,26] and references therein. The presence of the delayed bound τ ensures that we need to update the unit star subgradient of f at in at least every τ iteration. Some examples of the delayed sequences are as follows:
- (i)
- Constant delay [18], that is, for all .
- (ii)
- Cyclic delay [19,20,24,26] and references therein. The typical form of this type is for all . In this case, the delays () are chosen with deterministic order from the set . This means that we will use stale information over a consistent period of length and update the new star subgradient at every iteration.
- (iii)
- Random delay [25], that is, the delays () are randomly chosen in the set .
Assumption 2.
The function f satisfies the Hölder condition with order and modulus on X, i.e., for all .
The Hölder condition assumption is typically assumed when investigating the convergence results of methods for solving quasi-convex optimization problems [6,7,9,10,11,12]. Note that, for the special case of , the Hölder condition is nothing else than the Lipschitz condition.
The following lemma provides a useful property of the star subgradient and will play an important role in the convergence analysis. The formal proof is due to Konnov [10] (Proposition 2.1).
Lemma 2.
Suppose that Assumption 2 holds. Let be a sequence generated by DSSM-I. For any and , if then
Next, we will provide some lemma, which are important tools for proving the convergence results of DSSM-I. We first provide the basic inequality relating to the sufficient decreasing property of the generated sequence.
Lemma 3.
Suppose that Assumption 2 holds. Let be a sequence generated by DSSM-I. For any and if then
Proof.
Let and be fixed. Suppose that We note from the definition in DSSM-I that
where the first inequality follows from the nonexpansiveness of metric projection, and the second one follows from the Cauchy–Schwarz inequality and the fact that for all , and the last one follows from Lemma 2. This completes the proof. □
Lemma 4.
Suppose that Assumption 1 holds. Let be a sequence generated by DSSM-I, and let be a nonincreasing sequence with and . Then for all and
Proof.
Now, we are ready to prove the first convergence result in the sense of the inferior limit of the function values of the generated sequence is equal to the optimal value .
Theorem 1.
Suppose that Assumptions 1 and 2 hold. Let be a sequence generated by DSSM-I and be a nonincreasing sequence with and . Then, we have .
Proof.
It is obvious that since for all so that it is sufficient to prove that Now, assume to the contrary that Then there exists and such that for all This yields that
On the other hand, since and (from Lemma 4), there exists such that
Let and let be such that . Then, we obtain from Lemma 3 together with the relations (6) and (7) that for any
which yields
Since , we have and it follows that there exists such that
Thus, for any we obtain from (8) and (9) that
which is a contradiction. Therefore, we can conclude that and hence , as desired. □
Corollary 1.
Suppose that Assumptions 1 and 2 hold. Let be a sequence generated by DSSM-I, and let be a nonincreasing sequence with and . If the sequence is bounded, then there exists a subsequence of that converges to an optimal solution in .
Proof.
We first note from the fact that in Theorem 1 together with the boundedness of the sequence and the continuity of f that there exists a subsequence of such that
Since the sequence is also bounded, there exists a subsequence of such that Again, the continuity of f and using (10) yield
Note that since , the closedness of X yields which therefore implies that is an element in . □
By imposing either the coercivity of the objective function f or the compactness of the constraint X, we can obtain an approximate convergence behavior and some nice convergence results in comparison to the one obtained in Theorem 1.
Theorem 2.
Suppose that Assumptions 1 and 2 hold. Let be a sequence generated by DSSM-I, and let be a nonincreasing sequence with and . If the function f is coercive or the constrained set X is compact, then the following statements hold:
- (i)
- For any there exists such thatwhere ;
- (ii)
- , and
Proof.
Let be arbitrary and denote the notation
Since the function f is coercive or the constrained set X is compact, we ensure that which together with yield that Moreover, the continuity and quasi-convexity of f, respectively, imply that is closed and convex, and so the set is also closed and convex. Again, with either the compactness of X or the coercivity of f, the intersection is bounded, and hence, is a compact set, which ensures that
Let be given. Since we also have Consequently, there exists such that and
This together with the relation that which is obtained in Lemma 4 yield that
Now, we will show that
Let be fixed. We divide the proof into two cases according to the behavior of .
Case 1: Suppose that Then we have Applying Lemma 3 and the nonincreasing property of the stepsize , we get
Substituting the inequalities (11) and (12) in (14) gives
By putting we obtain
Case 2: Suppose that Then we have and We note from the definition of the nonexpansiveness of and the nonincreasing property of that
which together with the inequality (11) and (12) yield
By putting and invoking we obtain
Thus, we proved the relation (13). By using the same analogy of Theorem 1, we obtain that . Consequently, there exists such that and Therefore, we obtain from Case 2 that
Put . Suppose that for all Invoking (13), we obtain . Hence, we conclude by the strong induction that
For each we denote Observe that is nonincreasing and Subsequently, is also nonincreasing and . Hence, we obtain that .
Now, let be arbitrary. Since there exists such that
Since , there exists such that
Hence, we conclude that . Furthermore, by Assumption 2, we therefore obtain that □
Remark 3.
Under the setting of Theorem 2, we observe that if an optimal solution is unique, that is, we obtain that which means that the generated sequence converges to an unique solution
3.2. Delayed Star Subgradient Method II
In order to obtain the convergence of the generated sequence, we need a modified version of DDSM-I which is defined as the following Algorithm 2.
| Algorithm 2 Delayed Star Subgradient Method II (DSSM-II) |
| Initialization: Given a stepsize the delays , and initial points . Iterative Step: For a current point , if we set Update: . |
Remark 4.
- (i)
- If the delays for all , DSSM-II is relating to the method proposed by Hu et al. [9] (Algorithm 1) with the special setting of .
- (ii)
- Note that DSSM-II involves evaluating the function value at the current iterate and deciding to update the next iteration . If the function value equals the optimal value , then no additional calculations are performed, and DSSM-II terminates so that an optimal solution is obtained at the point .
Next, we will derive some fundamental properties for the convergence of DSSM-II.
Lemma 5.
Suppose that Assumption 2 holds. Let be a sequence generated by DSSM-II. For any and , if then
Proof.
Let and . Suppose that . We have from the definition of DSSM-II that . Moreover, since we also have Thus, by following the proving lines of Lemma 3, we obtain
as desired. □
Theorem 3.
Suppose that Assumptions 1 and 2 hold. Let be a sequence generated by DSSM-II, and let be a nonincreasing sequence with and . Then, we have for all , , and
Proof.
By invoking Lemma 5, we can prove the same analog of Lemma 4 and Theorem 1 to obtain □
Theorem 4.
Suppose that Assumptions 1 and 2 hold. Let be a sequence generated by DSSM-II, and let be a nonincreasing sequence with and Then the sequence converges to an optimal solution in and, moreover,
Proof.
If there exists such that , we have that is an optimal solution and for all Consequently, we obtain and as required.
On the other hand, we suppose that for all Let and let We note from Lemma 5 that
Let us admit from the inequalities (3) and (4) that and for all Thus, for a fixed , we have
which together with the nonincreasing property of implies
By approaching in (15) and using the assumption that , we obtain that , and so
Therefore, the sequence is quasi-Fejer monotone. By applying Fact 2, we obtain that the sequence is bounded and exists.
Since the sequence is bounded and (Theorem 3), there exists a subsequence of such that
Since is also bounded, there exists a subsequence of such that . Thus, the continuity of f yields
Moreover, the closeness of X implies , and hence we obtain that Since exists and , we conclude that . Moreover, invoking the continuity of f, we also have that The proof is complete. □
We end the convergence analysis of DSSM-II by investigating the finite convergence of the generated sequence as the following theorem.
Theorem 5.
Suppose that Assumptions 1 and 2 hold and for some Let be a sequence generated by DSSM-II. If one of the following statements holds:
- (i)
- for all
- (ii)
- is nonincreasing with and ,
then there exists such that .
Proof.
We suppose by contradiction that for all . This implies for all . Note that , which implies that Since , we have Thus, by the definition of the star subgradient of f at , we obtain that
which yields
On the other hand, since we have from the definition of DSSM-II that , and the nonexpansive property of implies that
which combining (16), leads to
Let N be a fixed non-negative integer. By summing up (17) from to , we have
which implies
Based on the conditions (i) and (ii), we will divide our consideration into two cases.
Case 1: Assume that for all
Case 2: Assume that is a nonincreasing sequence with and . We obtain from Theorem 3 that . Thus, we finally obtain from the fact that and and with Lemma 1 that
which also contradicts to the inequality (18). □
4. Numerical Experiments
In this section, we present experimental results when we apply DSSM-I to solve the Cobb–Douglas production efficiency problem. We performed all experiments on MATLAB (R2023b) on a MacBook Air 13.3-inch with Apple M1 chip processor and 8GB memory.
The Cobb–Douglas production efficiency problem [3,6,12] aims to maximize the ratio of the total profit represented by the Cobb–Douglas product function and the total cost represented by the linear function on the product factor subject to the funding level limitations constraint. More precisely, let and let for all and and the Cobb–Douglas production efficiency problem is defined by
where Here, when is the j-th product factor , and for each , is the profit to be obtained from project i, and is the support of product factor j to project i to achieve . According to the literature [6,12], we know that the Cobb–Douglas production efficiency problem (19) is a quasi-convex optimization problem. Note that the function f is continuous and quasi-concave [4] (Theorem 2.5.1), satisfying the Hölder condition on [12] (Appendix B), and the constraint set is the nonempty closed convex set, where is a vector in . Therefore, we can apply DSSM-I to solve the problem (19). However, since the exact optimal solution to problem (19) is not known, we are unable to apply DSSM-II to solve it. To ensure the boundedness of the constrained set, we set the constrained set to be the set which is a subset of C. Thus, the established convergence result from Theorem 2 guarantees that It is worth noting that the computational procedure of the metric projection onto the constrained set D is not an easy task since there is no explicit form of in general. To deal with the situation, we utilized the classical Halpern iteration by performing the following inner loop: for arbitrary initial point and a sequence such that we compute
where , for all Then the sequence converges to the unique point , see [28] (Theorem 30.1) for further details. In our experiment, we put the inner initial point as the vector whose all coordinates are 1 and set for all We use the stopping criterion for the inner loop.
To perform the numerical experiments, we randomly set the parameter We choose the stepsize and time-varying delays to be and for all respectively. We put the initial point as the vector whose all coordinates are 1. To explain the influence of the delay bound , we perform Algorithm DSSM-I for delay bounds and 10 for various sizes of the problem n and m. Note that DSSM-I with is nothing else than the star subgradient method (SSM) proposed by Kiwiel [1]. Figure 1 illustrates the number of subgradient calculations performed by SSM and DSSM-I with the delay strategies for all .
Figure 1.
Behaviors of a number of subgradint calculations for various delays bounds versus number of iterations.
Figure 2.
Behaviors of objective function values for various delays bounds and problem sizes n and m.
Figure 3.
Behavior of relative errors for various delay bounds and problem sizes n and m.
It can be observed from Figure 2 that, overall, all the results show a similar pattern as they approach the (approximated) optimal value as supported by Theorem 2. For each subfigure, we notice that the result of seems stable and decreases faster than other .
Next, we plot the relative errors of the objective functions values in the same setting as above in Figure 3.
It can be seen from Figure 3 that all results seem to decrease and stabilize. More precisely, the graph exhibits fluctuations that are alternating decreases and increases in small periods. These fluctuations gradually diminish and vanish to stabilize. Again, this behavior satisfies the convergence result of DSSM-I, and we notice that the results with seems to decrease rapidly to 0 compared to other values of .
5. Conclusions
In this work, we proposed the so-called projected star subgradient methods with delayed subgradient updates, namely DSSM-I and DSSM-II, for solving the constrained quasi-convex optimization problem. For DSSM-I, we proved that there exists a subsequence of the generated sequence that converges to an optimal solution provided that the boundedness of the generated sequence is imposed. Furthermore, under the compactness or the coercive assumption with the diminishing step size, we proved that the distance between the sequence and the optimal solution set converges to zero, resulting in the convergence of the objective function values to the optimal value. For the modified version as DSSM-II and without requiring the compactness or coercive assumptions, we proved that the generated sequence converges to an optimal solution. We also established the finite convergence provided that the interior of the optimal solution set is nonempty. Finally, we applied DSSM-I to solve the Cobb–Douglas production efficiency problem. It is worth noting that the proposed methods can reduce the computational costs of star subgradients by allowing the method to retrieve stale data instead. It is noted that since we could not indicate the precise upper bound on the number of times the star subgradient could be reused, the open problem with this topic still remains. Moreover, one can notice that we require the closed-form expression of the metric projection onto the constraint set X and overcome this obstacle by solving the sub-problem to approximate at each iteration. This also motivated us to consider a future direction that involves a method that does not require solving a sub-problem.
Author Contributions
Conceptualization, O.P. and N.N.; methodology, O.P. and N.N.; software, O.P.; validation, O.P. and N.N.; formal analysis, O.P. and N.N.; investigation, O.P. and N.N.; writing—original draft preparation, O.P.; writing—review and editing, O.P. and N.N.; visualization, O.P.; supervision, N.N.; project administration, N.N.; funding acquisition, N.N. All authors have read and agreed to the published version of the manuscript.
Funding
This research was supported by the Fundamental Fund of Khon Kaen University. This research has received funding support from the National Science, Research, and Innovation Fund or NSRF. O. Pankoon was supported by the Development and Promotion of Science and Technology Talents Project (DPST).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Dataset available on request from the authors.
Acknowledgments
The authors are thankful to the editor and two anonymous referees for comments and remarks which improved the quality and presentation of this paper.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Kiwiel, K.C. Convergence and efficiency of subgradient methods for quasiconvex minimization. Math. Program. 2001, 90, 1–25. [Google Scholar] [CrossRef]
- Cambini, A.; Martein, L. Generalized Convexity and Optimization; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
- Bradley, S.P.; Frey, S.C. Fractional programming with homogeneous functions. Oper. Res. 1974, 22, 350–357. [Google Scholar] [CrossRef]
- Stancu-Minasian, I.M. Fractional Programming; Kluwer Academic Publishers: Dordrecht, The Netherlands, 1997. [Google Scholar]
- Schaible, S.; Shi, J. Fractional programming: The sum-of-ratios case. Optim. Methods Softw. 2003, 18, 219–229. [Google Scholar] [CrossRef]
- Hu, Y.; Yang, X.; Sim, C.-K. Inexact subgradient methods for quasi-convex optimization problems. Eur. J. Oper. Res. 2015, 240, 315–327. [Google Scholar] [CrossRef]
- Hu, Y.; Yu, C.K.W.; Li, C. Stochastic subgradient method for quasi-convex optimization problems. J. Nonlinear Convex Anal. 2016, 17, 711–724. [Google Scholar]
- Hu, Y.; Yu, C.K.W.; Li, C.; Yang, X. Conditional subgradient methods for constrained quasi-convex optimization problems. J. Nonlinear Convex Anal. 2016, 17, 2143–2158. [Google Scholar]
- Hu, Y.; Yu, C.K.W.; Yang, X. Incremental quasi-subgradient methods for minimizing the sum of quasi-convex functions. J. Glob. Optim. 2019, 75, 1003–1028. [Google Scholar] [CrossRef]
- Konnov, I.V. On convergence properties of a subgradient method. Optim. Methods Softw. 2003, 18, 53–62. [Google Scholar] [CrossRef]
- Konnov, I.V. On properties of supporting and quasi-supporting vectors. J. Math. Sci. 1994, 71, 2760–2763. [Google Scholar] [CrossRef]
- Hishinuma, K.; Iiduka, H. Fixed point quasiconvex subgradient method. Eur. J. Oper. Res. 2020, 282, 428–437. [Google Scholar] [CrossRef]
- Choque, J.; Lara, F.; Marcavillaca, R.T. A subgradient projection method for quasiconvex minimization. Positivity 2024, 28, 64. [Google Scholar] [CrossRef]
- Zhao, X.; Köbis, M.A.; Yao, Y. A projected subgradient method for nondifferentiable quasiconvex multiobjective optimization problems. J. Optim. Theory Appl. 2021, 190, 82–107. [Google Scholar] [CrossRef]
- Ermol’ev, Y.M. Methods of solution of nonlinear extremal problems. Cybern 1966, 2, 1–14. [Google Scholar] [CrossRef]
- Penot, J.-P. Are generalized derivatives useful for generalized convex functions? In Generalized Convexity, Generalized Monotonicity: Recent Results; Crouzeix, J.-P., Martinez-Legaz, J.-E., Volle, M., Eds.; Springer: Boston, MA, USA, 1998; pp. 3–59. [Google Scholar]
- Penot, J.-P.; Zălinescu, C. Elements of quasiconvex subdifferential calculus. J. Convex Anal. 2000, 7, 243–269. [Google Scholar]
- Arjevani, Y.; Shamir, O.; Srebro, N. A tight convergence analysis for stochastic gradient descent with delayed updates. In Proceedings of the 31st International Conference on Algorithmic Learning Theory, San Diego, CA, USA, 8–11 February 2020; Kontorovich, A., Neu, G., Eds.; Proceedings of Machine Learning Research, PMLR: Cambridge, MA, USA, 2020; Volume 117, pp. 111–132. [Google Scholar]
- Stich, S.U.; Karimireddy, S.P. The error-feedback framework: Better rates for sgd with delayed gradients and compressed updates. J. Mach. Learn. Res. 2020, 21, 9613–9648. [Google Scholar]
- Gürbüzbalaban, M.; Ozdaglar, A.; Parrilo, P.A. On the convergence rate of incremental aggregated gradient algorithms. SIAM J. Optim. 2017, 27, 1035–1048. [Google Scholar] [CrossRef]
- Tseng, P.; Yun, S. Incrementally updated gradient methods for constrained and regularized optimization. J. Optim. Theory Appl. 2014, 160, 832–853. [Google Scholar] [CrossRef]
- Vanli, N.D.; Gürbüzbalaban, M.; Ozdaglar, A. Global convergence rate of proximal incremental aggregated gradient methods. SIAM J. Optim. 2008, 28, 1282–1300. [Google Scholar] [CrossRef]
- Butnariu, D.; Censor, Y.; Reich, S. Distributed asynchronous incremental subgradient methods. In Inherently Parallel Algorithms in Feasibility and Optimization and Their Applications; Brezinski, C., Wuytack, L., Reich, S., Eds.; Elsevier Science B.V.: Amsterdam, The Netherlands, 2001; Volume 8, p. 381. [Google Scholar]
- Namsak, S.; Petrot, N.; Nimana, N. A distributed proximal gradient method with time-varying delays for solving additive convex optimizations. Results Appl. Math. 2023, 18, 100370. [Google Scholar] [CrossRef]
- Deng, X.; Shen, L.; Li, S.; Sun, T.; Li, D.; Tao, D. Towards understanding the generalizability of delayed stochastic gradient descent. arXiv 2023, arXiv:2308.09430. [Google Scholar] [CrossRef]
- Arunrat, T.; Namsak, S.; Nimana, N. An asynchronous subgradient-proximal method for solving additive convex optimization problems. J. Appl. Math. Comput. 2023, 69, 3911–3936. [Google Scholar] [CrossRef]
- Cegielski, A. Iterative Methods for Fixed Point Problems in Hilbert Spaces; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
- Bauschke, H.H.; Combettes, P.L. Convex Analysis and Monotone Operator Theory in Hilbert Spaces, 2nd ed.; Springer: Cham, Switzerland, 2017. [Google Scholar]
- Kiwiel, K.C. Convergence of Approximate and Incremental Subgradient Methods for Convex Optimization. SIAM J. Optim. 2004, 14, 807–840. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).