Abstract
Biconvex programming (or inequality constrained biconvex optimization) is an important model in solving many engineering optimization problems in areas like machine learning and signal and information processing. In this paper, the partial exactness of the partial optimum for the penalty function of biconvex programming is studied. The penalty function is partially exact if the partial Karush–Kuhn–Tucker (KKT) condition is true. The sufficient and necessary partially local stability condition used to determine whether the penalty function is partially exact for a partial optimum solution is also proven. Based on the penalty function, an algorithm is presented for finding a partial optimum solution to an inequality constrained biconvex optimization, and its convergence is proven under some conditions.
1. Introduction
Multi-convex programming is a non-convex optimization problem [1,2], where biconvex programming is a special case. It is ubiquitous nowadays in fields such as control [3,4,5], machine learning [6,7], signal and information processing [8,9], communication [10,11], and also NP-hard problems [12]. The existing research on multi-convex programming mainly solves some very special models [12,13,14,15,16,17,18,19]. These studies all give specific methods for each special model. When problems change, new methods need to be studied. Biconvex programming is a simple case of multi-convex programming, and it is also studied in research [14] as a special case. Therefore, it is of great significance to study the algorithm of biconvex programming for solving practical engineering optimization problems.
Bilinear programming is the simplest case of biconvex programming and is the earliest and most studied as per [14]. There are two main ways to solve bilinear programming. One is the simplex algorithm based on sub-problems, and the other is the alternating direction method. For example, Liang and Bai [15] and Hajinezhad and Shi [16] both proposed the alternating direction method of multipliers (ADMM) algorithm for two special bilinear programming problems, where the extended Lagrangian penalty function uses a square penalty. Furthermore, Charkhgard et al. [17] presented a multi-linear programming algorithm using the linear programming algorithm.
In 2007, Gorski et al. [14] reviewed the development of the theory and algorithms of biconvex optimization. The biconvex functions have some good properties that are similar to convex functions, such as the biconvex separation theorem and an equivalence between local optimal solutions and stationary points. The biconvex programming algorithm is based on the idea that the solution to alternating sub-problems can converge to the stationary point of the original problem. For example, in 2015, Li et al. [18] studied an alternating convex search method to solve a stationary point problem of biconvex programming. In 2016, Shah et al. [19] presented an alternating search method with a square penalty function to solve biconvex programming, which had great effect in image recognition. Al-Khayyaltt and Falk [20] discussed an algorithm of biconvex programming using the idea of branch-and-bound.
In short, based on the above, the mainstream method used to solve biconvex programming is an alternative to subproblem solving, because the bi-convexity of biconvex programming guarantees the convergence of alternative search methods and effectually cuts down the scale of problem solving via decomposition calculation. Hence, in view of the large scale of biconvex programming(BCP), alternative subproblem solving will be the main method adopted in further research regarding biconvex programming.
In this paper we consider biconvex programming with inequality constraints as follows:
where are biconvex if and are convex in for every fixed and in for every fixed . Let and . The feasible set of (BCP) is denoted by:
Let
When is fixed, define the suboptimal problem:
Let
When is fixed, define the suboptimal problem:
Many practical engineering optimization problems can be transformed into biconvex programming, for example the two-cardinality sparse convex optimization problem and the nonconvex quadratic programming problem.
It is well known that an effective method to solve constrained optimization is by penalty function [21]. Its main idea is to transform a constrained optimization problem into a sequence of unconstrained optimization subproblems that are easier to solve. When the penalty function is not exact, many calculations are needed to solve unconstrained optimization subproblems so as to obtain an approximate solution to inequality constrained optimization. For example, it is proven that the exact penalty function method for solving constrained optimization is very efficient, as first proposed by Zangwill (1967) in [22]. In theory, if the penalty function is exact, to obtain an optimal solution to a constrained optimization problem only one unconstrained optimization subproblem is solved. Hence, the exact penalty function algorithm takes less time than the inexact penalty function algorithm. Additionally, this approach has recently received great attention from both theoretical and practical arenas. Many studies were later presented based on the exact penalty function algorithm, such as Rosenberg (1986) [23] and Di Pillo (1986) [24]. In addition, it is essential to determine the exactness of a penalty function under the stability condition [25,26]. Hence, this paper mainly focuses on the relationship between the partial exactness of the penalty function and the partial stability of biconvex programming. On the other hand, there exist other approaches to reduce constrained optimization problems to unconstrained ones; for example, the index method presented in the monograph [27] or the method for computable boundaries presented in [28]. This shows the significance of the study of the partial exact penalty function.
In order to ensure exactness, we propose the following penalty function:
where penalty parameter and . By the definition of exactness in [29], a penalty function is exact at , but is not exact at , such that . Hence, the exactness of an optimization problem depends on the structure of the problem. In this paper, we will study the exactness of a more extensive penalty function for biconvex programming than the one presented in [30].
The remainder of the paper is organized as follows. In Section 2, for a partial optimum solution, the partial exactness of the penalty function is proven under the partial Karush–Kuhn–Tucker (KKT) condition or the partial stableness condition. An algorithm is presented to find out a partial optimum solution to (BCP) with convergence.
2. Partial Exactness and a Penalty Function for (BCP)
According to Gorski et al. [14], defining the partial optimum of (BCP) is very meaningful. The concept of the partial optimum of (BCP) is given as follows:
Definition 1.
Let . If:
then is called a partial optimum of (BCP). A partial optimum of (BCP) means that is an optimal solution to (BCP)() and is an optimal solution to (BCP)().
Next, let us give the equivalence of a partial optimum of (BCP) to a partial KKT point under some conditions.
Let be biconvex and differentiable. Let . If there are such that:
then is a KKT point of (BCP).
Let . If there are such that:
then is a partial KKT point of (BCP).
Let . The constraint of (BCP) is called a partial Slater constraint qualification at , if there is such that:
In fact, if the Slater constraint qualification is satisfied for a convex programming, an optimal solution of the convex programming is equal to a KKT condition. For biconvex programming, we have the results in [28] as follows.
Theorem 1.
Let . If (BCP) is satisfied with partial Slater constraint qualification at , then is a partial optimum of (BCP) if and only if is a partial KKT point of (BCP).
Corollary 1.
Let be a partial optimum of (BCP). If (BCP) is satisfied with partial Slater constraint qualification at , then is a KKT point of (BCP) if and only if (5), (6), and (7) hold with .
Example 1.
Let the biconvex programming:
where . For , it is clear that is the partial KKT point and the partial optimum of (BCP1). If , then . We have as . Hence, a local optimal solution to (BCP1) cannot be solved, such as in .
Example 1 means that if there is no optimal solution to biconvex programming, there may exist a partial KKT point or a partial optimum. It is obvious that an optimal solution to biconvex programming is the partial optimum. For biconvex programming, there is a partial optimum even if an optimal solution to biconvex programming is not obtained. Example 1 further indicates that the partial optimum is very important to biconvex programming.
Let be given. Consider the following optimization problem:
where is a decision variable when is fixed, and
where is a decision variable when is fixed.
Definition 2.
Let . If:
then is called a partial optimum of .
Next, we define the concept of a partially exact penalty function for biconvex programming as follows.
Definition 3.
(1) Let be a partial optimum of (BCP). If there is a such that is a partial optimum of for , then is called a partially exact penalty function.
(2) Let be a partial optimum of . If of is a partial optimum of (BCP), then ρ is called a partially exact value of the penalty parameter.
Example 2 below shows the partial exactness of a penalty function under different p values.
Example 2.
Let the biconvex programming:
It is clear that is a partial optimum of (BCP2). The penalty function for (BCP2) is defined by:
When , a partial optimum of is . Letting , we have . Hence, is not a partially exact penalty function. When , we have:
A partial optimum of is for . Hence, is a partially exact penalty function for . Example 2 means that the partial exactness of penalty function depends on the parameter p.
Example 3.
Let the biconvex programming:
It is easy to verify that is an optimal solution to (BCP3) for any . Thus, is a partial optimum of (BCP3) too. The example illustrates that all partial optimums of (BCP) may be optimal solutions.
is a partial optimum of (BCP3). The square penalty function for (BCP3) is defined by:
It is easy to check that is a partial optimum of for and . It is easy to check that is not partially exact.
For , a penalty function for (BCP3) is defined by:
We have that is a partial optimum of for . It is easy to check that is partially exact for . We easily check that is a KKT point.
Example 3 illustrates that the partially exact penalty function for a partial optimum can be as good as traditional exact penalty functions.
We prove the similarity of the partially exact penalty result to [28].
Theorem 2.
Let . Let be a partial optimum of (BCP). If is a KKT partial point of (BCP), i.e., there are such that (5), (6), and (7) are true, then is a partially exact penalty function for , where:
Proof.
If is a partial KKT point of (BCP), there are and such that (5), (6) and (7) are true. Because are biconvex functions for and , we have:
By (5), (7), (10), and (11), we have:
Let , and . Let us take with If there is a , then . We have:
i.e.,
Otherwise, if for , from (14) we have:
Hence, for , we have:
Similarly, for , we have:
By Definition 2, we have that is a partially exact penalty function for .
□
Note. If there is a number A such that for any , then the conclusion of Theorem 2 holds too. If we let , we have on . It is clear that the problem
is equal to the problem (BCP). Particularly, when , we have the following conclusion.
Theorem 3.
Let be a partial optimum of (BCP). If is a KKT partial point of (BCP), i.e., there are and such that (5), (6), and (7) are true, then is a partially exact penalty function with for , where:
Theorem 3 is consistent with Theorem 2 in [28].
Similar to that for a constrained penalty function presented in [25,26], the concept of stability for a penalty function of (BCP) is defined. Let and:
When is fixed, define a perturbed problem:
Let and:
When is fixed, define a perturbed problem:
Definition 4.
Let be a partial optimum of (BCP), and and be optimal solutions to (BCP)() and (BCP)(), respectively, for any . If there is a such that:
where and , then is partially stable at . Furthermore, if there are a and a such that (16) and (17) hold for and , then is partially locally stable at .
Theorem 4.
Let be a partial optimum of (BCP). If is partially stable, then is a partially exact penalty function at .
Proof.
Let us prove that is a partially exact penalty function when is partially stable at . Suppose that is not a partially exact penalty function. According to the definition of partial stability, for any , we obtain that there is a satisfying that:
Then, there always exists some such that is not a partial optimum of , i.e., there is some such that:
Thus,
Suppose that . We have:
This implies that and , which shows that is not a partial optimum of (BCP). A contradiction occurs. Hence, and do not hold, and or .
Let with and with , , and and be optimal solutions to (BCP)() and (BCP)(), respectively. Then, and . Thus,
Therefore,
and:
which shows that:
where and . These inequalities contradict (18) and (19). Hence, that is not partially stable yields a contradiction with the assumption, which proves that is a partially exact penalty function. □
Theorem 5.
Let be a partial optimum of (BCP). If is a partially exact penalty function at , then is partially locally stable. In particular, for , is partially stable.
Proof.
Let us prove that is partially locally stable when is a partially exact penalty function. According to the definition of a partially exact penalty function, if is a partial optimum of (BCP), there always exist some such that:
Let and be optimal solutions to (BCP)() and (BCP)(), respectively, for any . By (20), and (21), we have:
It is clear that for , we have:
It follows from the definition that is partially locally stable. □
Theorem 4 and Theorem 5 mean that the stability condition is sufficient for the partial exact penalty function, but the necessary condition of the partial exact penalty function is partially locally stable, as shown in the following example.
Example 4.
Let the perturbed problem (BCP3), and its (BCP3()) and (BCP3()):
It is clear that is a partial optimum of (MCP3) and its objective function value is 1. and are optimal solutions to (BCP3)() and (BCP3)(), respectively, for any . Let and with . For ,
then is partially locally stable. When , is partially stable.
3. Partial Optimum Penalty Function Algorithm for (BCP)
Now, we present an algorithm to solve a partial optimum of (BCP) by solving the penalty function problem BCP( as follows.
Based on the above results, the Partial Optimum Penalty Function Algorithm (Algorithm 1) was designed to compute a partial optimum of (BCP).
We prove the convergence of the Algorithm 1 in Theorem 6. Let:
be a level set. If for any given , is bounded, then is called bounded.
| Algorithm 1: POPFA Algorithm |
| Step 1: Choose and . Step 2: Solve to be a partial optimum of Otherwise, and go to Step 2. |
Theorem 6.
Let be the sequence generated by the Algorithm 1 and be continuous.
(i) If is a finite sequence (i.e., the Algorithm 1 stops at the -th iteration), then is a partial optimum of (BCP).
(ii) Let be an infinite sequence, sequence be bounded and the level set be bounded. Then is bounded and any limit point of it is a partial optimum of (BCP).
Proof.
(i) The conclusion is clear.
(ii) By the Algorithm 1, since is bounded as , there must be some such that:
Since h is continuous and the level set is bounded and closed, and are bounded. Without loss of generality, suppose . Hence, there is an such that . From (25), we have:
We have as . Hence, is a feasible solution to (BCP).
Let any and . Since is a partial optimum of
we have:
Let , and the above inequations are:
Hence, is a partial optimum of (BCP). □
Theorem 6 means that the Algorithm 1 has good convergence in theory. From Theorem 4, when the penalty function is partially stable, the Algorithm 1 solves a single unconstrained optimization problem for the smaller penalty parameter . Since the penalty function is nonsmooth for , it is necessary to smooth the constrained nonsmooth term to design an effective algorithm. Therefore, the smoothing algorithm of a partial exact penalty function is worthy of further study. In fact, for , we have published a paper regarding a smoothing partially exact penalty function algorithm [30], where two numerical examples show the proposed algorithm is effective for biconvex programming.
4. Conclusions
In this paper, we studied the partial optimum solution to biconvex programming using the penalty function, which is partially exact. The form of this penalty function is more general than that in [28]. We proved that the partial exactness of the penalty function for biconvex programming is equivalent to the partial KKT condition, and we proved that the partial exactness of the penalty function for biconvex programming is equivalent to partially local stability. Based on the penalty function, the Algorithm 1 was theoretically presented to solve a partial optimum solution to biconvex programming. The convergence of the algorithm was also proven. The Algorithm 1 may solve a partial optimum solution to biconvex programming under the smaller penalty parameter. In the future, we may study the smoothing problem of the partial exact penalty function and its algorithm.
Author Contributions
Formal analysis and methodology, M.J.; investigation and methodology, Z.M.; methodology, R.S. All authors have read and agreed to the published version of the manuscript.
Funding
This work is supported by the National Natural Science Foundation of China (grant No.11871434) and the Natural Science Foundation of Zhejiang Province (grant No.LY18A010031).
Acknowledgments
This work is supported by the National Natural Science Foundation of China (grant No.11871434) and the Natural Science Foundation of Zhejiang Province (grant No.LY18A010031). The authors would like to express their gratitudes to anonymous referees’ detailed comments and remarks that help us improve our presentation of this paper considerably.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Shen, X.; Diamond, S.; Udell, M.; Gu, Y. Disciplined multi-convex programming. In Proceedings of the Control And Decision Conference (CCDC), Chongqing, China, 28–30 May 2017. [Google Scholar]
- Grant, M.; Boyd, S.; Ye, Y. Disciplined convex programming. In Global Optimization: From Theory to Implementation, Nonconvex Optimization and its Applications; Liberti, L., Maculan, N., Eds.; Springer: New York, NY, USA, 2006; pp. 155–210. [Google Scholar]
- Chiu, W.Y. Method of reduction of variables for bilinear matrix inequality problems in system and control designs. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 1241–1256. [Google Scholar] [CrossRef]
- Ichihara, H.; Nobuyama, E. Difference of multiconvex relaxation of parameterized LMIs: Control applications. In Proceedings of the SICE 2003 Annual Conference, Fukui, Japan, 4–6 August 2003. [Google Scholar]
- Hours, J.; Jones, C. A parametric multiconvex splitting technique with application to real-time NMPC. In Proceedings of the 53rd IEEE Conference on Decision and Control, Los Angeles, CA, USA, 15–17 December 2014. [Google Scholar]
- Suh, S.; Shin, S.; Lee, J.; Reddy, C.K.; Choo, J. Localized user-driven topic discovery via boosted ensemble of nonnegative matrix factorization. Knowl. Inf. Syst. 2018, 56, 503–531. [Google Scholar] [CrossRef]
- Udell, M.; Horn, C.; Zadeh, R.; Boyd, S. Generalized low rank models. Found. Trends Mach. Learn. 2016, 9, 1–118. [Google Scholar] [CrossRef]
- Fu, X.; Huang, K.; Sidiropoulos, N.D. On identifiability of nonnegative matrix factorization. IEEE Signal Process. Lett. 2018, 25, 328–332. [Google Scholar] [CrossRef]
- Hong, M.; Luo, Z. On the linear convergence of the alternating direction method of multipliers. Math. Program. 2017, 162, 165–199. [Google Scholar] [CrossRef]
- Serbetli, S.; Yener, A. Transceiver optimization for multiuser MIMO systems. IEEE Trans. Signal Process. 2004, 52, 214–226. [Google Scholar] [CrossRef]
- Al-Shatri, H.; Li, X.; Ganesa, R.S.S. Maximizing the Sum Rate in Cellular Networks Using Multiconvex Optimization. IEEE Trans. Wirel. Commun. 2016, 15, 3199–3211. [Google Scholar] [CrossRef]
- Wen, Z.; Yin, W.; Zhang, Y. Solving a low-rank factorization model for matrix completion by a nonlinear successive over-relaxation algorithm. Math. Program. Comput. 2012, 4, 333–361. [Google Scholar] [CrossRef]
- Xu, Y.; Yin, W. A block coordinate descent method for regularized multiconvex optimization with applications to nonnegative tensor factorization and completion. Siam J. Imaging Sci. 2013, 6, 1758–1789. [Google Scholar] [CrossRef]
- Gorski, J.; Pfeuffer, F.; Klamroth, K. Biconvex sets and optimization with biconvex functions: A survey and extensions. Math. Methods Oper. Res. 2007, 66, 373–407. [Google Scholar] [CrossRef]
- Liang, X.; Bai, J. Preconditioned ADMM for a class of bilinear programming problems. Math. Probl. Eng. 2018, 2018, 1–9. [Google Scholar] [CrossRef]
- Hajinezhad, D.; Shi, Q.J. Alternating direction method of multipliers for a class of nonconvex bilinear optimization: Convergence analysis and applications. J. Glob. Optim. 2018, 70, 261–288. [Google Scholar] [CrossRef]
- Charkhgard, H.; Savelsbergh, M.; Talebian, M. A linear programming based algorithm to solve a class of optimization problems with a multi-linear objective function and affine constraints. Comput. Oper. Res. 2018, 89, 17–30. [Google Scholar] [CrossRef]
- Li, G.; Wen, C.; Zheng, W.X.; Zhao, G. Iterative identification of block-oriented nonlinear systems based on biconvex optimization. Syst. Control. Lett. 2015, 79, 68–75. [Google Scholar] [CrossRef]
- Shah, S.; Yadav, A.K.; Castillo, C.D.; Jacobs, D.W.; Studer, C.; Goldstein, T. Biconvex Relaxation for Semidefinite Programming in Computer Vision. In Proceedings of the European Conference on Computer Vision: Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016. [Google Scholar]
- Al-Khayyaltt, F.A.; Falk, J.E. Jointly constrained biconvex programming. Math. Oper. Res. 1983, 8, 273–286. [Google Scholar] [CrossRef]
- Pardalos, P.M.; Resende, M.G.C. Handbook of Applied optimization; Oxford University Press: Oxford, UK, 2002. [Google Scholar]
- Zangwill, W.I. Nonlinear programming via penalty function. Manag. Sci. 1967, 13, 334–358. [Google Scholar] [CrossRef]
- Rosenberg, E. Globally convergent algorithms for convex programming. Math. Oper. Res. 1981, 6, 437–443. [Google Scholar] [CrossRef]
- Di Pillo, G.; Grippo, L. An exact penalty function method with global conergence properties for nonlinear programming problems. Math. Program. 1981, 36, 1–18. [Google Scholar] [CrossRef]
- Clarke, F.H. Optimization and Nonsmooth Analysis; Wiley: New York, NY, USA, 1983. [Google Scholar]
- Burke, J.V. Calmness and exact penalization. Siam J. Control. Optim. 1991, 29, 493–497. [Google Scholar] [CrossRef]
- Strongin, R.G.; Sergeyev, Y.D. Global Optimization with Non-Convex Constraints; Science & Business: Dordrecht, The Netherlands, 2000. [Google Scholar]
- Grishagin, V.; Israfilov, R. Multidimensional Constrained Global Optimization in Domains with Computable Boundaries. Available online: https://elar.urfu.ru/bitstream/10995/36002/1/ural_pdc-2015-09.pdf (accessed on 16 January 2021).
- Yang, X.Q.; Meng, Z.Q. Lagrange Multipliers and Calmness Conditions of Order p. Math. Oper. Res. 2007, 32, 95–101. [Google Scholar] [CrossRef]
- Shen, R.; Meng, Z.Q.; Jiang, M. Smoothing Partially Exact Penalty Function of Biconvex Programming. Asia-Pac. J. Oper. Res. 2020, 37, 27. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).