Next Article in Journal
CTOA: Toward a Chaotic-Based Tumbleweed Optimization Algorithm
Previous Article in Journal
A Differential Relation of Metric Properties for Orientable Smooth Surfaces in ℝ3
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Systems of Linear Equations with Non-Negativity Constraints: Hyper-Rectangle Cover Theory and Its Applications

Department of Electrical and Computer Engineering, McMaster University, 1280 Main Street West, Hamilton, ON L8S 4L8, Canada
*
Author to whom correspondence should be addressed.
Deceased.
Mathematics 2023, 11(10), 2338; https://doi.org/10.3390/math11102338
Submission received: 6 March 2023 / Revised: 8 May 2023 / Accepted: 12 May 2023 / Published: 17 May 2023

Abstract

:
In this paper, a novel hyper-rectangle cover theory is developed. Two important concepts, the cover order and the cover length, are introduced. We construct a specific échelon form of the matrix in the same manner as that employed to determine the rank of the matrix to obtain the cover order of any given matrix. Using the properties of the cover order, we obtain the necessary and sufficient conditions for the existence and uniqueness of the solutions for linear equations system with non-negativity constraints on variables for both homogeneous and nonhomogeneous cases. In addition, we apply the cover theory to analyze some typical problems in linear algebra and optimization with non-negativity constraints on variables, including linear programming (LP) problems and non-negative least squares (NNLS) problems. For LP problems, the three possible behaviours of the solutions are studied through cover theory. On the other hand, we develop a method to obtain the cover length of the covered variable. In this process, we discover the relationship between the cover length determination problem and the NNLS problem. This enables us to obtain an analytical optimal value for the NNLS problem.

1. Introduction

The problems with non-negativity constraints on variables play a prominent role in engineering, physics, chemistry, computer science, and economics. These problems with non-negative constraints often appear as (1) finding solutions for systems of linear equations, (2) solving LP problems, and (3) finding solutions for NNLS problems [1].
The analysis of systems of linear equations is a fundamental part of linear algebra, and forms the core of mathematical modelling of many different branches of science and engineering such as, to name but a few, electric circuits, communications, radars, optics, controls, etc. [2,3,4,5,6]. Lately, it has also been used to model the outbreak of COVID-19 and calcium diffusion [7,8,9]. Thus, methods for finding the solutions of linear equations system also play an important role in various applications [10]. Some mature methods have been developed to analyze and solve systems of linear equations without non-negativity constraints on the variables [11,12,13]. However, with non-negativity constraints added to the variables, the analysis of the solutions to the linear equations becomes harder [14]. For such problems, the classical analysis for the existence of non-negative solutions is mainly based on Farkas’ lemma [15]. In terms of uniqueness, there is no direct characterization in the general case. It is also noted that the analysis of non-negative solutions to the system of homogeneous or nonhomogeneous linear equations is mostly concerned with investigating other associated problems rather than addressing the problem in a direct way. Thus, a new approach is needed for the analysis of systems of linear equations with non-negativity constraints on the variables.
LP problems arise in many applications [16]. Many problems can be reformulated as linear programs both in theory and in practice so that fast algorithms can be applied [17,18,19,20]. Dantzig developed the simplex method in 1947, which was the first efficient method for solving LP problems and has been widely accepted as a computational tool [21,22,23]. Geometrically, the procedure of the simplex method involves moving one feasible solution to another and, for each step, the value of the objective function improves. This continues until the optimum objective is reached. It would thus be desirable if we can determine the optimal solution directly rather than moving through the feasible solutions for matrices with some specific structures. In this paper, we propose a new systematic procedure for solving the LP problem by applying cover theory using a transformed objective function.
The problem of NNLS is a type of least squares problem with non-negativity constraints on variables, which arises in applications throughout science and engineering. Various methods have been proposed to solve this kind of problems, and they can normally be divided into three classes: active-set methods [24,25,26], iterative methods [27,28], and other methods [29,30,31,32]. The first technique to solve the NNLS problem was proposed by Lawson and Hanson in [33], which is a typical active-set method, and the corresponding algorithm is named lsqnonneg in Matlab. This commonly used algorithm always converges and terminates in finite steps; however, there is no upper limit on the possible number of iterations that the algorithm might need to reach the point of the optimum value. In contrast to active-set methods, iterative methods enable one to incorporate multiple active constraints at each iteration. Since most existing algorithms for solving the NNLS problem are based on numerical analysis, we are motivated to propose a method to solve it from the matrix perspective by applying the techniques we developed in cover theory. More specifically, we solve the problem by investigating the structure of the matrix itself so as to obtain the analytical optimal value of NNLS problem.

Overview of the Paper

In this paper, we establish the novel hyper-rectangle cover theory for which we obtain the necessary and sufficient conditions that guarantee the existence and the uniqueness of the solution for a system of linear equations with non-negativity constraints on variables. A specific échelon form of the matrix is introduced and based on this form, and an efficient method is developed to determine the cover order for any given matrix. Moreover, we investigate in detail the structures of the échelon form in various cases leading to the development of feasible solutions for the system of linear equations with non-negativity constraints on variables. Parallel investigations are carried out for the LP problems. Based on the échelon form and the corresponding results on the system of linear equations with non-negative constraints, we also analyze the various possibilities of the solution for an LP problem. Finally, we develop a method to determine the cover length of the covered variables, establishing their strong relationship with the optimal objective value of NNLS problems. Based on this relationship, a new method is derived to obtain the analytical optimal value of NNLS problems.
Notation: Most notations used throughout this paper are standard: column vectors and matrices are denoted by boldface lowercase and uppercase characters, respectively; the matrix transpose is denoted by  ( · ) T R + N  denotes the set of all the  N × 1  vectors with all entries being non-negative.

2. Concept of Hyper-Rectangle Cover

In this section, we formally give the definition of hyper-rectangle cover [34,35].
Definition 1. 
Given a matrix  A R M × N  and  x R + N x n  is the n-th element in  x  for index  n 1 , 2 , , N . Let  c n ( x n )  denote the smallest number  c n  such that
x R + N : x T A T A x 1 x R + N : x n c n
We say that  x n  is a covered variable if  c n ( x n )  is finite and we refer to  c n ( x n )  as the cover length of the covered variable  x n . The cover order of  A , denoted by  R c ( A ) , is the number of indices  n { 1 , 2 , , N }  for which  c n ( x n ) < . We say that  A  has full cover if  R c ( A ) = N  and has zero cover if  R c ( A ) = 0 .
The following nontrivial examples may serve as illustrations of the definitions of cover order and cover length.
Example 1. 
Suppose that  A = 1 2 2 1 , and we have  A T A = 5 4 4 5 . For this matrix,  x R + 2 : x T A T A x τ 2  is an ellipse in the whole plane, which is shown in Figure 1, and the part which is located in the non-negative domain is fully covered by a rectangle. Thus,  x 1  and  x 2  are both covered in this example.
Example 2. 
Consider  A = 1 1  and  A T A = 1 1 1 1 ; for this case, the feasible set determined by  x T A T A x = x 1 x 2 2 τ 2  is shown in Figure 2 and is open and unbounded with respect to both  x 1  and  x 2 . Hence,  R c ( A ) = 0 .
From the above examples, we can find that the cover order  R c ( A )  and the cover length  c i ( x i )  represent the maximal dimension and minimal side lengths of the hyper-rectangle that covers  x : x R + N , x T A T A x τ 2  respectively.

3. Systems of Linear Equations with Non-Negativity Constraints on Solutions

3.1. Homogeneous Systems of Linear Equations

In this section, we present an important result in cover theory, allowing us to determine if a column vector in  A  or the corresponding variable  x i  in  A x  is covered or not. Furthermore, this result provides us with a method of investigating the non-negative solution to a system of linear equations. Let us first introduce the definition of a convex cone [36].
Definition 2. 
A set  C R M  is a convex cone if  α x + β y C  for all  x , y C  and  α , β > 0 .
A cone  C  is polyhedral if it is the conic combination of finitely many vectors, i.e., there is a set of vectors  { a 1 , , a N } , so that  C = { a 1 u 1 + a 2 u 2 + a N u N | a i R M , u i R + } . The polyhedral cone  C  is a closed convex cone.
With the above definition, we are now able to obtain the following result:
Theorem 1. 
Let  A  be an  M × N  real matrix. Then, the i-th column of  A , or the i-th variable  x i  associated with the i-th column vector  a i  in  A x , is covered if and only if  A x 0  for any  x R + N  with  x i > 0 .
Proof. 
Necessity: Here, by assuming that  x i  is covered in  A x , we need to show that for any  x R + N  with  x i > 0 , we have  A x 0 . Suppose that this statement was not true. Then, there exists  x 0 R + N  with  x 0 , i > 0  such that  A x 0 = 0 , where  x 0 , i  is the i-th element in  x 0 . As a consequence, for any positive number  p > 0 , we would also have  A ( p x 0 ) = 0 , implying that  p x 0 , i  is not bounded if for any given  τ > 0 0 = ( p x 0 ) T A T A ( p x 0 ) τ . This contradicts the assumption that  x i  is covered in  A x . Therefore, the necessary condition is true.
Sufficiency: For  x i > 0 , the quadratic form  x T A T A x  can be rewritten as:  x T A T A x = A ¯ i x ¯ i + a i x i 2 = x i 2 A ¯ i u + a i 2 , where  u = x ¯ i x i A ¯ i  is the  M × ( N 1 )  sub-matrix formed by deleting the i-th column from  A  and  u 0 . Consider the set  A ¯ i u : u R + N 1 . It is a closed convex cone according to Definition 2, and the function  A ¯ i u + a i 2  is convex; thus, the minimum of  A ¯ i u + a i 2  exists, i.e., there exists a  u 0 0  such that  A ¯ i u 0 + a i 2 A ¯ i u + a i 2 , for any  u R + N 1 . In fact,  A ¯ i u 0 + a i 0 . Otherwise, if  x 0 , i = 1 , x 0 , k = u 0 , k , for  k = 1 , 2 , , i 1 , and  x 0 , k = u 0 , k 1 , for  k = i + 1 , , N , then, we have  A ¯ i u 0 + a i = A x 0 = 0 , which contradicts the assumption. Now for any given positive real value  τ , if we let  x T A T A x τ 2 , then, we have  x i 2 A ¯ i u 0 + a i 2 x T A T A x τ 2 . Hence, we obtain  0 < x i < τ A ¯ i u 0 + a i , i.e.,  x i  is covered in  A x .
Thus, the proof of Theorem 1 is complete.    □
  From Theorem 1, the following results can be obtained.
Corollary 1. 
Let  A  be an  M × N  real matrix, and let  A ¯ j  be the  M × ( N 1 )  sub-matrix formed by deleting the j-th column from  A . Then, the following statements are true:
1. 
A system of homogeneous linear equations,  A x = 0 , has a nonzero solution in  R + N  if and only if  A  does not have full cover.
2. 
Let the j-th column of  A  be covered. Then, any column of  A ¯ j  is covered in  A ¯ j  if and only if it, as a column of  A , is also covered in  A .
3. 
If the i-th column of  A  is covered, then it is also covered in  A ¯ j  for  j i .
4. 
A full column rank matrix  A  always has full cover.
For a homogeneous system of linear equations with non-negative constraints on solutions, it is important to determine the necessary and sufficient condition which guarantees the existence of nonzero solutions. The direct determination of whether the system has nonzero solution is not simple [14,37]. Here, Theorem 1 makes a statement paralleled to the first statement of Corollary 1, providing us with the condition for the existence of nonzero solutions for the system.

3.2. Nonhomegeneous Systems of Linear Equations with Non-Negativity Constraints on Solutions

Nonhomogeneous systems of linear equations with non-negativity constraints on solutions are frequently encountered in the field of signal and image processing, multispectral data handling, fibre optics, etc. [38,39,40]. The classical way for determining the existence of non-negative solution is based on Farkas’ lemma [15]. According to Farkas’ lemma, given a problem of linear equations with non-negativity constraints on the variables, there exists another problem associated with it such that the original problem has a solution in the required domain if and only if the associated problem has no solution. Thus, this lemma provides an indirect way to check the existence of non-negative solutions to a nonhomogeneous system of linear equations.
In the following, based on the cover theory, we will derive the direct necessary and sufficient conditions for the existence and the uniqueness of non-negative solutions of the aforementioned system.
Existence of non-negative solutions:
Theorem 2. 
Let  A R M × N  and  b R M . Then, there exists an  x R + N  such that  A x = b  if and only if the cover order of the augmented matrix  A ˜ = A , b  is less than or equal to that of  A .
Proof. 
Let us rewrite the linear equations  A x = b  as  A ˜ x ˜ = 0 , where  A ˜ = A , b  and  x ˜ = x , x ˜ N + 1 . First we prove the sufficiency: Under the assumption  R c ( A ˜ ) R c ( A ) , by Statement 2 of Corollary 1, we can claim that  x ˜ N + 1  is not covered in  A ˜ x ˜ . Since  A ˜  does not have full cover, then by Theorem 1, there exists an  x ˜ 0 = ( x 0 , x ˜ 0 , N + 1 ) R + N + 1  with  x ˜ 0 , N + 1 > 0 , where  x ˜ 0 , N + 1  is the  ( N + 1 ) -th element of  x ˜ 0 , such that  A ˜ x ˜ 0 = 0 . Hence, we have  A x 0 = b x ˜ 0 , N + 1 , implying that  x 0 / x ˜ 0 , N + 1  is a solution of  A x = b . Therefore, the sufficient condition is true.
To prove the necessary condition, we assume that the system of linear equations  A x = b  has a solution in  R + N , i.e., there exists an  x 0 R + N  such that  A x 0 = b . Then, by Theorem 1, the  ( N + 1 ) -th column vector of  A ˜  is not covered. In addition, by Statement 3) of Corollary 1, we know that if  a ˜ i , the i-th column vector of  A ˜ i { 1 , , N } , is covered in  A ˜ , then  a ˜ i , being a column vector of  A , is also covered in  A . Therefore, the inequality  R c ( A ˜ ) R c ( A )  holds. This completes the proof of Theorem 2.    □
Theorem 2 provides us with the regularization condition for non-negative solutions for a systems of linear algebraic equations. It is interesting to point out that in some application of systems of ordinary differential equations, there are parallel regularizations that provide non-negative solutions [41,42].
Uniqueness of non-negative solutions:
With the definition of convex cone in Definition 2, let us first introduce Carathéordory’s theorem [15]:
Lemma 1 (Carathéordory’s theorem). 
Let  S R M  be a finite set, and let  y R M . If  y cone S , then there exists a linearly independent set  T  such that  y cone T .
By applying Carathéordory’s theorem, we are able to show the necessary and sufficient condition for the uniqueness of non-negative solution to a nonhomogeneous system of linear equations. This is stated in the following theorem:
Theorem 3. 
Let  A R M × N  and  b R M . Then,  A x = b  has a unique solution  x  in  R + N , if and only if  R c ( A ˜ ) R c ( A )  and  R c ( A ˜ ) + R r ( A ¯ ) = N , where  R r ( A ¯ )  is the rank of  A ¯ , and  A ¯ = { a i } i N ¯ , with  N ¯  being a set consisting of all the column indices of  A  not covered in  A ˜ .
Proof. 
Sufficiency: Let  R c ¯ ( A ˜ )  be the number of uncovered column vectors in  A ˜ . Thus, we have  R c ( A ˜ ) + R c ¯ ( A ˜ ) = N + 1 , and under the assumption  R c ( A ˜ ) + R r ( A ¯ ) = N , we have  R c ¯ ( A ˜ ) = R r ( A ¯ ) + 1 . In addition, since  R c ( A ˜ ) R c ( A ) , we can obtain  R c ¯ ( A ˜ ) = | N ¯ | + 1  and, as a result,  R r ( A ¯ ) = | N ¯ | . Therefore, all the column vectors  { a i } i N ¯  are linearly independent in  R M . As a consequence,  A x = b  has a unique solution in  R + N .
Necessity: According to the property of  R c ( A ˜ )  and the definition of  A ¯ , we have  R c ( A ˜ ) + R r ( A ¯ ) N . Considering  R c ( A ˜ ) + R r ( A ¯ ) < N , in this case, we have  R r ( A ¯ ) < | N | ¯ . Then, for all  i N ¯ , there are  μ i R  (not all 0) such that  i N ¯ μ i a i = 0  and there are  λ i 0  such that  i N ¯ λ i a i = b . In addition, there is a real number  α  such that  λ i + α μ i > 0  for all  i N ¯ . Let us assume that  x 0 R + N  is the unique solution of  A x = b , where  x 0 , i = λ i + α μ i , for  i N ¯  and  x 0 , i = 0 , for  i { 1 , 2 , , N } N ¯ , then we have  A x 0 = b . As Carathéordory’s theorem states, we are able to find a linearly independent subset  { a i } i N ˜  of  { a i } i N ¯ . Let  N ˜  be the set consisting of the linearly independent column indices of  { a i } i N ¯ , and  | N ˜ | < | N ¯ | . Then there is another solution  x 1 R + N  such that  A x 1 = b  and  x 1 , i = 0 , for  i { 1 , 2 , , N } N ˜ , where the number of zero element in  x 1  is larger than that in  x 0 . This contradicts the assumption that  x 0  is the unique solution of  A x = b . Therefore,  R c ( A ˜ ) + R r ( A ¯ ) = N  must be satisfied to guarantee the uniqueness of the solution. Thus, the proof of Theorem 3 is complete.    □

4. Cover Order

In this section, we develop a specific échelon form of a matrix that can be used to determine the cover order of any given matrix. Some special properties of cover order are also explored.

4.1. Cover Order Determination

Let  R + + N  denote the set of  N × 1  vectors with all entries being positive. The vector  x , of which the elements are all positive, is called a positive vector. Similarly, let  R + N R N , and  R N , respectively, denote a non-negative, a negative, and a nonpositive set of  N × 1  vectors. We first present some related results in the following lemmas which are useful for our derivation of the cover order of a real matrix.
Lemma 2 
([13]). Let  S  be a subspace of  R N  and  S  be the orthogonal complementary subspace of  S . Then,
1. 
S R + N =  if and only if  S R + + N .
2. 
S R + + N =  if and only if  S R + N .
Denoting the row space of  A  by  S A  and the orthogonal complement to this row space by  S A , and using Lemma 2, we have the following [34]:
Lemma 3. 
Let  R + N ( K )  denote the set of all the non-negative vectors with K positive entries. Specifically,  R + N ( 0 )  denotes the set  { 0 N × 1 } . For any  A R M × N , the cover order of  A  is equal to  max S A R + N ( K ) K .
Lemma 3 shows us the necessary and sufficient condition for determining the cover order of a matrix. Now, we show an important property of cover order in the following:
Theorem 4. 
Given an  M × N  real matrix  A  and given an invertible  M × M  real matrix  T , let  B = TA , then we have  R c ( B ) = R c ( A ) .
Proof. 
Let us consider  x T B T B x , which is equivalent to  x T ( A T T T T A ) x  according to the assumption. Suppose that  λ min  and  λ max  are the minimum eigenvalue and the maximum eigenvalue of  T T T , respectively, since  T T T  is a real symmetric matrix, then by Rayleigh–Ritz theorem [43], we have  x R n , and the inequalities  λ min · x T x x T T T T x λ max · x T x  hold. Letting  y = Ax , we have  λ min · y T y y T T T T y λ max · y T y , i.e.,
λ min · x T A T A x x T A T T T T A x λ max · x T A T A x
Then, using the left-hand side inequality  λ min · x T A T A x x T B T B x  in Equation (2) and the definition of cover order in Definition 1, we have:
x R + N : x T B T B x 1 x R + N : x T A T A x 1 λ m i n x R + N : x k i c k i λ m i n , i = 1 , , R c ( A )
where  k i { 1 , 2 , , N }  and  c k i  are positive real numbers. Then, by the definition of cover order in Definition 1, we know that at least  R c ( A )  variables in  x  associated with the column vectors in  B x  are covered. Thus, we have  R c ( A ) R c ( B ) .
According to the right-hand side inequality in Equation (2), we have:
x R + N : x T A T A x 1 λ m a x x R + N : x T B T B x 1 x : 0 x k i c k i , i = 1 , , R c ( B )
Similarly, at least  R c ( B )  variables in  x  associated with the column vectors in  A x  are covered. Thus, we have  R c ( A ) R c ( B ) .
Hence, we can conclude that, if det ( T ) 0  and  B = TA , then  R c ( A ) = R c ( B ) .    □
Theorem 4 carries important implications. It states that the cover order of a matrix is invariant under any row transformation. From Lemma 3, we know that if we are able to find non-negative vectors in  S A , then the cover order of  A  is equal to the largest number of the positive entries of these vectors. Thus, Theorem 4 together with Lemma 3 implicitly suggests that we can perform a series of linear elementary row transformations and column permutations to determine the cover order of the matrix. This indeed leads us to the development of a straightforward procedure transforming  A  into an échelon form for the evaluation of its cover order.

4.2. Procedure of the Échelon Transformation

An échelon form of a rectangular matrix [1] has the following structures:
Definition 3 (échelon form). 
A rectangular matrix is in échelon form (or row échelon form) if it has the following three properties:
(a)
All nonzero rows are above any rows of all zeros.
(b)
Each leading entry of a row is in a column to the right of the leading entry of the row above it.
(c)
All entries in a column below a leading entry are zeros.  
Our procedure of échelon transformation can now be laid out as follows:
1.
The Échelon Form of  A . Given an  M × N  real matrix  A , we can find matrices  E 0  and  P 0  such that [1]:
E 0 A P 0 = I B 0 0 0
where  I R R r × R r  and  B 0 R R r × ( N R r )  with  R r  being the rank of the matrix  A . Here,  E 0  and  P 0  are, respectively, the elementary transformation and the permutation matrices, either of which may be made up of a product of simpler elementary and permutation matrices. The right side of Equation (5) conforms with the description of échelon form; thus, Equation (5) is an échelon transformation of  A . Note that the échelon transformation of  A  is not unique; we can choose different  E 0  and  P 0 , arriving at different values of  B 0 .
2.
The Cover Order. Without loss of generality, we can assume that  A  has full rank. In particular, from Theorem 1 and Lemma 3, if the initial échelon transformation of  A  in Equation (5) results in every entry in some row of  B 0  being positive, then  R c ( A ) = N , i.e.,  A , has full cover. On the other hand, if every entry in some column of  B 0  is negative, then  R c ( A ) = 0 . However, if the cover order of  A  is not immediately obvious from the structure of  B 0  resulted from the initial échelon transformation, we need the following steps of structural arrangement to determine the cover order.
(1)
Structure Arrangement. Search for all non-negative rows in  B 0  and select the one which has the greatest number of positive elements. Move this selected row to the first row and assume that it contains  N 1  positive entries. By performing the row and column permutation, we can always ensure the identity matrix structure ahead and let the following statements hold:
b 11 , b 12 , , b 1 N 1 > 0 , b 1 ( N 1 + 1 ) , , b 1 ( N M ) = 0
where  b 1 i i = 1 , , N M  are the elements in the first row of the new structure of  B 0 . Ignoring the above first  N 1  columns in new  B 0 , we find all non-negative rows in the remaining part of it and choose the row with the largest number of positive elements. Moving this row to the second row and assuming that it contains  N 2  positive entries in the remaining  N M N 1  columns, we have:
b 2 ( N 1 + 1 ) , b 2 ( N 1 + 2 ) , , b 2 ( N 1 + N 2 ) > 0 , b 2 ( N 1 + N 2 + 1 ) , , b 2 ( N M ) = 0
where  b 2 i i = N 1 + 1 , , N M , are the elements in the second row of the new form of  B 0  after the above steps. By arranging the following rows similarly, after s times, we obtain:
b 11 , b 12 , , b 1 N 1 > 0 , b 1 ( N 1 + 1 ) , , b 1 ( N M ) = 0 b 2 ( N 1 + 1 ) , b 2 ( N 1 + 2 ) , , b 2 ( N 1 + N 2 ) > 0 , b 2 ( N 1 + N 2 + 1 ) , , b 2 ( N M ) = 0 b s ( N 1 + N 2 + + N s 1 + 1 ) , , b s ( N 1 + + N s ) > 0 , b s ( N 1 + N 2 + + N s + 1 ) , , b s ( N M ) = 0
where  b i j  in Equation (8),  i = 1 , 2 , , s j = 1 , 2 , , N M , and  s M , are the elements in the first s rows of the structure of  B 0  after s times transformation. The procedure ends when one of the following two cases happens:
(a)
i = 1 s N i = N M , in which case  A  has full cover.
(b)
There is no non-negative row vector in the row space of the  B 0  after s times of transformation.
Let:
B ¯ = b s + 1 , N 1 + + N s + 1 b s + 1 , N M b M , N 1 + + N s + 1 b M , N M
(2)
Cover Order. At the end of the above structural arrangement, we arrive at the conclusion that the cover order of  A  is  R c ( A ) = i = 1 s N i + s  and  s M .
The next theorem states the property of the final échelon form of the matrix from which the cover order of  A  can be deduced.
Theorem 5. 
For any  M × N  real matrix  A , there exists an elementary matrix  E  and a permutation matrix  P  such that  E A P = I , B , where  I R R r × R r B R R r × ( N R r ) , and  R r  is the rank of the matrix  A . Then  B  either:
1. 
Contains at least one non-negative row;
2. 
Contains at least one negative column vector, or there exists one nonpositive column vector, but the same row position where the zero lies will be negative in some other columns of  B .
Proof. 
We shall prove the result by induction. Without loss of generality, we can assume that the matrix  A  has full rank. Indeed, the proof is based on the following steps: (i) Iteration  N M = 1 , i.e.,  EAP = ( I , b ) , where  b R M . If  b  contains at least one positive element, then  A  has full cover. If  b  is a negative vector, then  A  has zero cover. If  b  is a nonpositive vector, then the cover order of  A  equals the number of zero terms. (ii) The result holds true for iteration  N M = K + 1  given that it holds true for iteration  N M = K . We prove the desired result in the following:
Suppose that for  N M = K , the above conclusion holds, i.e., if  R c ( A ) > 0 , then  A  can be transformed into  I , B . Let  b i j  be the  i j -th element in  B , for  i = 1 , 2 , , M j = 1 , , K . We have Equation (8) hold and  R c ( A ) = i = 1 s N i + s . In addition, let  B ¯  be the form in Equation (9) with  K = N M , and  B ¯  either contains at least one negative column or has one nonpositive column, but the position where the zero lies will be negative in some other column of  B ¯ . In the following, we will prove that if the above conclusion holds for  N M = K , then this conclusion will also hold when  N M = K + 1 .
When  N M = K + 1 , we can assume that  A = I , b 1 , , b K , b K + 1 , where  I R M × M  and  b i R M , for  i = 1 , , K + 1 . Without considering the  b K + 1 , we denote the remaining part in  A  as  A ¯ , which equals  ( I , b 1 , , b K ) . According to the assumption for  N M = K A ¯  can be transformed into échelon form and we let  R c ( A ¯ ) = i = 1 s N i + s . By considering the corresponding  b K + 1  with the échelon form of  A ¯  (apply the same row permutation to  b K + 1  as  A ¯  permutes in the échelon transformation and we still use  b K + 1  to denote it after the permutation), we can notice that if  b 1 , K + 1 > 0 R c ( A ) = i = 1 s N i + s + 1 . If  b 1 , K + 1 = 0 , we can perform the following process from the second row. Therefore, in the following, we will consider the case when  b 1 , K + 1 < 0 . According to Theorem 4 and Lemma 3, the following steps can be taken to make the first row nonpositive and move it to the last row without affecting the cover order of  A .
Step 1
Let  m = max b j 1 b 11 , , b j , K + 1 b 1 , K + 1 , where  b j 1 b 11 < 0 ,⋯,  b j , K + 1 b 1 , K + 1 < 0 . Multiply the first row of  A  with m, and add the product to i-th row, for  i = 2 , 3 , , M . We will have  A ( 1 ) .
Step 2
Use (-1) times the first row of  A ( 1 )  to obtain  A ( 2 ) .
Step 3
Let  t 2 = b 2 , K + 1 b 1 , K + 1 + m , t 3 = b 3 , K + 1 b 1 , K + 1 + m , , t M = b M , K + 1 b 1 , K + 1 + m  and let  a j T  be the j-th row of  A ( 2 ) . Then by adding  a 1 T t j  to the j-the row in  A ( 2 ) , where  j = 2 , 3 , , M . We will have  A ( 3 ) .
Step 4
Multiplying the first row of  A ( 3 )  with  1 b 1 , K + 1  and exchanging the position of the first column with the last column, we will obtain  A ( 4 ) .
Step 5
Permuting the rows and columns so that the first row in the right-hand side of  A ( 4 )  is moved to the last row, as well as securing the left-hand side identity matrix structure. After this, we will have  A ( 5 ) .
Step 6
Without considering the last column of  A ( 5 ) , rearranging the rows and columns of the first  ( M + K )  columns of it, we will obtain a new échelon form matrix  A ¯ ( 5 )  and  R c ( A ¯ ( 5 ) ) = i = 1 s ( 2 ) N i ( 2 ) + s ( 2 ) . By considering the corresponding  b ¯ K + 1  with the échelon form of  A ¯ ( 5 ) , we can notice that if  b ¯ 1 , K + 1 > 0 , then  R c ( A ) = i = 1 s ( 2 ) N i ( 2 ) + s ( 2 ) + 1 . If  b ¯ 1 , K + 1 < 0 , we can repeat the above steps.
Finally, after either t times transformation, there exists one  b 1 , K + 1 > 0 , such that the first row of the new matrix is non-negative and  R c ( A ) = i = 1 s ( t ) N i ( t ) + s ( t ) + 1 , or  R c ( A ) = 0  and there exists at least one column ( ( K + 1 ) -th column of  A ) which is negative.    □
An important indication given by Theorem 5 is that when the first scenario occurs, then the cover order of  A  can be determined by the steps as shown in the above échelon transformation; otherwise,  R c ( A ) = 0 .

4.3. Some Properties of the Échelon Form

We observe from the results of échelon transformation that the final échelon form of a matrix is not unique, and under different circumstances, different forms may be required. It is thus interesting to investigate the specific échelon form for special cases, especially for a low-rank matrix  A .
1.
Let  A R M × N  and  R r ( A ) = 2 . If  A  has full cover, then  A  can be transformed into:
A I 2 B 2 × ( N 2 ) 0 ( M 2 ) × 2 0 ( M 2 ) × ( N 2 )
where  B  is a non-negative matrix.
Proof. 
Without loss of generality, suppose that  M = R r ( A ) = 2 . If  A  has full cover, then by Theorem 5,  A  can be transformed into:
1 0 b 11 b 12 b 1 , ( N 2 ) 0 1 b 21 b 22 b 2 , ( N 2 )
where  b 11 , b 12 , , b 1 , ( N 2 ) > 0 . If  b 2 i < 0 , where  i { 1 , , N 2 } , let:
t = max i { 1 , , N 2 } b 2 i b 1 i , b 2 i < 0 = b 2 j b 1 j
Using t times the first row of  A  and adding the product to the second row of it, in the next step, multiply the first row of the above matrix with  1 b 1 j  and exchange the first column with the j-th column so that the identity matrix structure in the left-hand side part can be guaranteed. After these steps, we obtain a matrix  B  which has two non-negative row vectors.    □
2.
For a rank-2 matrix  A , we also have the following property:
Theorem 6. 
Let  A R M × N  and  R r ( A ) = 2 . Then  A  has zero cover if and only if it can be transformed into the form:
E A P = I 2 B + B 0 N 2 × 2 0 0
where all the elements in  B +  are non-negative, while the elements in  B  are all nonpositive. Specifically,  B  contains at least one column which is a negative vector, or two non-negative vectors with their negative terms lie in different rows.
Proof. 
Without loss of generality, we can assume that  A  has full rank.
Sufficiency:  Given a zero-cover matrix  A , then by Theorem 5, it can be transformed into  ( I , B ) , with  b i j  being the  i j -th element in  B , where  i { 1 , 2 } j { 1 , , N 2 } . Let  t = max i { 1 , , N 2 } { b 2 i b 1 i , b 1 i b 2 i < 0 } = b 2 j b 1 j . We multiply the first row of the above transformed matrix with t, and add the product to the second row. Then, we multiply the first row of the resulting matrix with  1 b 1 j  and exchange the first column with the j-th column to ensure that the left-hand side part remains an identity matrix. This results in a matrix of the form:
1 0 b 11 b 1 j 1 b 1 j b 1 , N 2 b 1 j 0 1 b 21 b 11 b 2 j b 1 j b 2 j b 1 j b 2 , N 2 b 1 , N 2 b 2 j b 1 j
If  b 1 i > 0  and  i { 1 , , N 2 } , then  b 2 i b 1 i b 2 j b 1 j = b 1 i b 2 i b 1 i b 2 j b 1 j > 0 . If  b 1 i < 0 , then for those  b 2 i < 0 , we have  b 2 i b 1 i b 2 j b 1 j < 0 , and if  b 2 i > 0 , the sign of  b 2 i b 1 i b 2 j b 1 j  is uncertain. After the above steps, and by performing some certain column permutations, the matrix  A  can be transformed into the form  I , B ( 1 ) , B ( 2 ) , B ( 3 ) , where  B ( 1 )  is a non-negative matrix,  B ( 3 )  is a nonpositive matrix, and the elements in the first row of  B ( 2 )  are all negative, while the elements in the second row of it are all positive. To simplify the discussion, we can write the above matrix as  ( I , B ( 1 ) ) , with  b i j ( 1 )  being the  i j -th element in  B ( 1 ) , where  i { 1 , 2 } j { 1 , , N 2 } . Then, we let  m = max i { 1 , , N 2 } { b 1 i ( 1 ) b 2 i ( 1 ) , b 2 i ( 1 ) b 1 i ( 1 ) < 0 } = b 1 s ( 1 ) b 2 s ( 1 ) . We multiply the second row of the above matrix by m and add the product to the first row. Then, multiplying the second row of the resulting matrix with  1 b 2 s ( 1 )  and exchanging the second column with the s-th column so that the identity matrix structure on the left-hand part can be ensured, we arrive at the following matrix:
1 0 b 11 ( 1 ) b 21 ( 1 ) b 1 s ( 1 ) b 2 s ( 1 ) b 1 s ( 1 ) b 2 s ( 1 ) b 1 , N 2 ( 1 ) b 2 , N 2 ( 1 ) b 1 s ( 1 ) b 2 s ( 1 ) 0 1 b 21 ( 1 ) b 2 s ( 1 ) 1 b 2 s ( 1 ) b 2 , N 1 ( 1 ) b 2 s ( 1 )
In the above matrix, we have if  b 2 i ( 1 ) > 0 , where  i { 1 , , N 2 } , then  b 1 i ( 1 ) b 2 i ( 1 ) b 1 s ( 1 ) b 2 s ( 1 ) > 0 , and if  b 2 i ( 1 ) < 0 , then  b 1 i ( 1 ) b 2 i ( 1 ) b 1 s ( 1 ) b 2 s ( 1 ) < 0 . As a result, if  A  has zero cover, then it can be transformed into the right-hand side form in Equation (11).
Necessity: Suppose that  A  can be transformed into the form in Equation (11). Then, consider the case when  B  contains two nonpositive vectors having their negative terms in different rows. Then  Ax = 0  can be written as:
I 2 × 2 B + B ¯ b 1 b 2 0 N 2 × 2 0 0 0 0 × u 1 , u 2 , x + , x , v 1 , v 2 T = 0
where  b 1 = b 1 , 0 T , b 2 = 0 , b 2 T , and  b 1 , b 2  are negative, and  B ¯  is the matrix formed by deleting  b 1  and  b 2  from  B . Then we will have:
u 1 u 2 = v 1 v 2 B + x + B ¯ x
Now, since  x  is the solution and must be positive, we can let the elements in  x +  and  x  take any positive value. If we let  v 1  and  v 2  be positive and large enough, we can still obtain positive  u 1  and  u 2 . In this case, all elements in  x  are positive and satisfy the equation  Ax = 0 . By Theorem 1,  A  has zero cover.
This completes the proof of the theorem.    □

5. Cover Order and Linear Programming

In this part, we present a systematic procedure using the concept of hyper-rectangle cover for solving LP problems.

5.1. Linear Programming (LP) Problem

The LP problem [16], in general, can be stated as:
min c T x subject to Ax = b x 0
where  A R M × N , with  M < N b R M , and  c , x R N . We can assume that  A  has full rank in general since redundant or inconsistent linear equations can always be detected and removed. The feasibility set of the above LP problem is:
F 1 = x R + N : Ax = b R + N
From the necessary and sufficient condition for the existence of non-negative solution for a nonhomogeneous system of linear equations developed in Theorem 2, we can directly obtain the necessary and sufficient condition that guarantees the nonempty feasibility set of the LP problem. This is stated in the following theorem:
Theorem 7. 
The feasibility set  F 1  of the LP problem is nonempty iff  R c ( A ˜ ) R c ( A ) , where  A ˜ = ( A , b ) .
Letting  z = c T x ; then, by adding the objective function into the constraints, the above LP problem can be restated as:
min z subject to A c T x = b z x 0
We denote  A c T  as  A c A b c T z  as  A ( z ) . By applying the échelon transformation to  A ( z )  without changing the position of the last row and the last column, we have:
A ( z ) I ( M + 1 ) × ( M + 1 ) B ( M + 1 ) × ( N M 1 ) f z + g
where  f  and  g  are  ( M + 1 ) × 1  column vectors. To simplify the analysis, in the following, we separate  A ( z )  into two parts and let:
A ˜ = I ( M + 1 ) × ( M + 1 ) B ( M + 1 ) × ( N M 1 ) and b ˜ = f z + g
We have the following observations:
Property 1. 
From Theorem 2, in order to have a nonempty feasibility set for this LP problem, adding  b ˜  to the right-hand side of  A ˜  should not increase the cover order of  A ˜ . In other words, the cover order of  A ( z )  should be less than or equal to the cover order of  A c .
Property 2. 
In a minimization problem, if the uncovered variable has a negative coefficient in the objective function and has negative or zero coefficients in all constraints in the échelon form, then the objective function is unbounded over the feasible region.

5.2. Three Possibilities of the Solution

Based on Property 1, we now analyze the possibilities of the solutions and the optimal value of the objective function of the LP problem under the three conditions: (1)  A ˜  has full cover; (2)  0 < R c ( A ˜ ) < N ; (3)  A ˜  has zero cover, resulting in the following theorem:
Theorem 8. 
For the LP problem given by Equation (14): If  A ˜  has full cover and the matrix  B  in  A ˜  is a non-negative matrix, then the LP problem has optimal solution if and only if  b ˜ i = f i z + g i 0 i = 1 , 2 , , R r . By solving these inequalities, we will have the range of z, which is:
max g i f i ( + ) , i { 1 , , R r } z min g i f i ( ) , i { 1 , , R r } ,
where  f i ( + )  and  f i ( )  are the positive and negative terms in the first  R r  elements of  b ˜ , respectively.
Proof. 
The proof of the above theorem follows directly from Property 1.    □
It should be noted that if the constraint of z in Theorem 8 is contradictory, i.e., if
min g i f i ( ) , i { 1 , , R r } < max g i f i ( + ) , i { 1 , , R r } ,
then the feasibility set of this linear program is empty: i.e.,  F 1 = . In other words, we are not able to find any feasible solution to this LP problem in this case. In addition, if there is no lower bound of z, i.e.,  max g i f i ( + ) , i { 1 , , R r }  in Equation (17) can be negative infinity, then the objective function in this minimization problem is unbounded.
By the same argument, obtaining the maximum value of z can also be achieved by solving the above inequalities. The maximum value will then be:
max z = min { g i f i ( ) , i { 1 , , R r } } .
If  A ˜  has full cover, but the matrix  B  is not a non-negative matrix, then let  I { 1 , , R r }  be the index set of the non-negative rows in  A ˜ . According to the assumption in échelon transformation, the first non-negative row vector in  A ˜  contains the largest number of positive terms and the number is  N 1 . Then the optimal value of the LP problem can be obtained by performing the following steps.
Cover Method (Minimization Form)
Step 1.
Solving  f i z + g i 0 , for  i I  and a candidate minimal value of z is:
z 0 = max g i f i ( ) , i I = g s f s ( )
Step 2.
If  z 0  satisfies:
max g k f k ( ) , k { 1 , , R r } I z 0 min g k f k ( + ) , k { 1 , , R r } I
then the process ends and the optimal value is obtained, which is
z m i n = z 0 = max g i f i ( ) , i I
Otherwise, there exist some  k { 1 , , R r } I  such that  f k z 0 + g k > 0 , i.e., we have  R c A ( z ) > R c ( A c ) , then the process continues.
Step 3.
Choose column  j k  to pivot in (i.e., introduce into the basis variable) by:
b 1 , j k b k , j k = min b 1 j b k j , b k j < 0 , 1 j N 1
Step 4.
Choose row  k ¯  to pivot in (i.e., drop from the basis variable) by:
f k ¯ z 0 + g k ¯ b k ¯ , j k = min f k z 0 + g k b k j k , b k j < 0 , f k z 0 + g k > 0
Step 5.
Replace the  k ¯ -th column with the  ( M + j k ) -th column and re-establish the échelon form.
Step 6.
If the matrix  B  is a non-negative matrix in the new échelon form, then the process ends and the optimal value is obtained, which is
z 0 = max g i n e w f i ( ) n e w , i { 1 , 2 , , R r }
Otherwise, the process continues.
Step 7.
Return to step 1.
The whole pivot process each time is performed by using  b i , j k b k ¯ , j k  times the  k ¯ -th row in  A ˜ , and adding the product into i-th row, for  i = 1 , 2 , , R r . Then we divide the  k ¯ -th row with  b k ¯ , j k , and the  ( M + j k ) -th column becomes  e k ¯ . Next, we exchange the position of the  ( M + j k ) -th column and the  k ¯ -th column. After this process,  f k ¯ z + g k ¯  is negative and the structure of the identity matrix ahead is reserved. The above computational procedures of the cover method in solving the LP problem can be summarized in the flow chart of Figure 3.
This simple step-by-step method provides an attractive alternative approach to the LP problem.
The following example provides a clear illustration of the cover method procedure. Here,  A ˜  has full cover but the matrix  B  in  A ˜  is not a non-negative matrix.
Example 3. 
min x 1 x 2 2 x 1 + x 2 + x 3 = 12 x 1 + 2 x 2 + x 4 = 9 x 1 , x 2 , x 3 , x 4 0
Letting  z = c T x  and adding the objective function into the constraints, we will have the following augmented matrix:
A ˜ ( z ) = 2 1 1 0 12 1 2 0 1 9 1 1 0 0 z
Applying the échelon transformation to  A ˜ ( z )  without changing the position of the last column, we have
A ˜ ( z ) 1 0 0 1 9 z 0 1 0 1 9 + 2 z 0 0 1 1 21 3 z
Since we exchange the position of the first two rows during this transformation, the corresponding positions of variables are also exchanged. According to Theorem 8, in order to have the feasible solutions for this LP problem, the following two conditions should be satisfied at the same time:  9 z 0  and  21 3 z 0 . By solving these two inequalities, we will have a candidate optimal value of z, which is  z 0 = max { 9 , 7 } = 7 . Since  z 0 min g 2 f 2 = 9 2 , then the optimal value of the objective function is  z * = 7 , and the corresponding optimal solution is  x * = ( 5 , 2 , 0 , 0 ) T .
Similarly, for the case when  0 < R c ( A ˜ ) < N , we can also apply the above procedures to obtain the optimal value of the objective function and the optimal solution towards the LP problem by changing the definition of the index set  I  and the range of k. For this case, we consider  i J , and  J { 1 , , s }  is the index set of the non-negative rows in the first s rows of  A ˜ , where s is obtained through the échelon transformation, and  k { 1 , 2 , , s } J .
For the zero-cover matrix, the status of the solution for the LP problem is given in the following theorem.
Theorem 9. 
For a full rank matrix  A ˜ , if it has zero cover, then the LP problem is feasible but unbounded.
Proof. 
Since adding any column to the right-hand side of a zero-cover matrix will still arrive at a matrix with zero-cover, the feasibility set  F 1  is always nonempty in this case. However, according to Theorem 5, a zero-cover matrix can be transformed into a structure which has at least one negative column or has one nonpositive column, but the same row position where the zero lies will be negative in some other column(s) of this structure. Thus, by Property 2, the objective function is unbounded over the feasible domain for the case when  A ˜  has zero cover.    □

5.3. Feasible Solutions for the LP Problem

With the échelon form and the specific structure of the zero-cover matrix, we are able to obtain a series of feasible solutions for any given LP problem. The detailed process is given in the following:
As we know,  A ˜ , b ˜  is an échelon form matrix, where  A ˜ R ( M + 1 ) × N  and  b ˜ R M + 1 . Then the échelon form can be divided into the following blocks:
A ˜ , b ˜ = I s 0 B ( 1 ) B ( 3 ) b ˜ ( 1 ) 0 I ( M + 1 s ) B ( 2 ) B ( 4 ) b ˜ ( 2 ) ,
where s is obtained through échelon transformation. Then by Theorem 1, in the system  A ˜ , b ˜ x = 0 x 0 , the covered variables  x i  are all zeros. As a result, we can ignore those covered column vectors in  A ˜ , b ˜ , which correspond to  I s B ( 1 ) 0 B ( 2 ) . In Equation (20), since  B ( 3 )  is a zero matrix and  b ˜ ( 1 )  is a zero vector, we only need to consider the remaining part of  A ˜ , b ˜  which is  I ( M + 1 s ) , B ( 4 ) , b ˜ ( 2 ) . Let us denote this part as  A ¯ , b ¯ = I , B ¯ , b ¯ = I ( M + 1 s ) , B ( 4 ) , b ˜ ( 2 ) . The cover order of this matrix is zero. Thus, in order to obtain the feasible solution for the LP problem, we only need to solve the following system of linear equations, where the non-negative vector  x ¯  is the uncovered part in  x :
I , B ¯ , b ¯ x ¯ = 0
For simplicity of discussion, we can assume that  I R m × m B ¯ R m × ( n m ) , and  b ¯ R m . From Theorem 5, we know that the zero-cover matrix can be transformed to the form which contains at least one negative column, or has one nonpositive column, but the row position where the zero lies will be negative in some other column of the matrix. Without loss of generality, we can assume that the negative column appears in the first column of  B ¯ , i.e.,  b ¯ 11 , b ¯ 21 , , b ¯ m 1 T  is a negative column vector. Then the following procedure enables us to obtain a series of feasible solutions to the LP problem.
Suppose that  x ¯ = x ¯ 1 , , x ¯ m , x ¯ m + 1 , x ¯ m + 2 , x ¯ n , x ¯ n + 1 T , where  x ¯ 1 , , x ¯ m  correspond to the column vectors in the  m × m  identity matrix,  x ¯ m + 1 , x ¯ m + 2 , , x ¯ n  correspond to the column vectors in  B ¯ , and  x ¯ n + 1  corresponds to  b ¯  in the multiplication  I , B ¯ , b ¯ x ¯ . Then, by Equation (21), the first m elements in  x ¯  can be expressed as a linear combination of  x ¯ m + 1 , , x ¯ n + 1 :
x ¯ 1 = b ¯ 11 x ¯ m + 1 b ¯ 12 x ¯ m + 2 b ¯ 1 ( n m ) x ¯ n b ¯ 1 x ¯ n + 1 x ¯ m = b ¯ m 1 x ¯ m + 1 b ¯ m 2 x ¯ m + 2 b ¯ m ( n m ) x ¯ n b ¯ m x ¯ n + 1
In order to obtain a linearly independent feasible solution set, we first let the vector  x ¯ m + 1 , , x ¯ n + 1 T  be a set of linearly independent vectors  L , 1 , 0 , , 0 T ( L , 0 , 1 , , 0 ) T , ⋯,  ( L , 0 , 0 , , 1 ) T  successively. In addition, in order to satisfy the non-negativity constraints on the variable  x ¯ i i = 1 , , n + 1 , we let:
L = max b ¯ i 2 b ¯ 11 , b ¯ i 3 b ¯ 21 , , , b ¯ i ( n m ) b ¯ m 1 , b ¯ i b ¯ i 1 , i = 1 , 2 , , m
We can then obtain a set of linear independent basic feasible solutions:
α 1 = b ¯ 11 L b ¯ 12 , , b ¯ m 1 L b ¯ m 2 , L , 1 , 0 , , 0 T α 2 = b ¯ 11 L b ¯ 13 , , b ¯ m 1 L b ¯ m 3 , L , 0 , 1 , , 0 T α n m 1 = b ¯ 11 L b ¯ 1 n , , b ¯ m 1 L b ¯ m n , L , 0 , 0 , , 1 , 0 T α n m = b ¯ 11 L b ¯ 1 , , b ¯ m 1 L b ¯ m , L , 0 , 0 , , 0 , 1 T
Any convex combination of those basic feasible solutions, i.e.,
x ¯ = k 1 α 1 + k 2 α 2 + + k n m α n m
where the real numbers  k i  satisfy  k i 0  and  k 1 + + k n m = 1 , is thus a solution of Equation (21). By padding the covered variables into  x ¯ , we obtain a series of feasible solutions to the LP problem.

5.4. The Simplex Method and the Cover Method

In 1947, Dantzig developed an algorithm to solve the LP problem efficiently, called the simplex method.
The LP problem is to find the extreme point of this polytope where the objective function is the smallest (or largest) in value if such an extreme point exists. By moving along the edge of the polytope, the simplex method identifies these extreme points with better objective values. The process continues until the optimum objective value is reached, or an unbounded edge is visited. For an LP problem having a nonempty feasible region, the algorithm always terminates because of the finite number of extreme points in the polytope. In practice, the simplex method has shown remarkable efficiency. However, in 1972, Klee and Minty gave an example, the Klee–Minty cube [44], showing that the worst-case complexity of the simplex method is exponential time.
While the simplex method regards the objective value z in the canonical tableau of the LP problem as a variable, the cover method treats it as a constant. Given a linear program, the cover method first rewrites an LP problem into the form of Equation (14), and then  A ( z )  is transformed into its échelon form. At this stage, if the matrix  B  in this échelon form is a non-negative matrix, then the optimum objective value can be determined directly according to Theorem 8. Thus, the computational complexity of this case is almost entirely determined by the complexity of échelon transformation. In the following, we will review the échelon transformation and analyze its computation complexity.
Consider a full rank matrix  A R M × N  with  M < N . The complexity of transforming  A  into an échelon form is  O ( M 2 N ) . In the structure arrangement process of the échelon transformation, the row having the greatest number of nonzero elements is moved to the first row, while the nonzero elements in this row have been moved to the left side of  B . Meanwhile, the corresponding column permutation such that the identity matrix structure could be preserved is performed. Thus, the selection of the row having the greatest number of nonzero elements is completed. The next step takes away the columns of  A  corresponding to these nonzero elements in the first row and performs an échelon transformation on the remaining part of  A . Such an iteration of échelon transformation, each time taking a lower complexity, continues until the desired form is achieved. The complexity of the structural arrangement process is  O ( M 2 ( N M ) ) . Thus, the computation complexity of échelon transformation in solving this type of LP problem by cover method is  O ( M 2 N ) .
It is observed, however, that if the matrix  B  is not a non-negative matrix, then the cover method for solving the LP problem will involve pivoting steps for which the complexity of the algorithm is no longer polynomial.

6. Cover Length

We first encountered the concept of cover length in Definition 1. In this section, we propose a method to determine the cover length of the covered variable  x i  associated with the i-th column vector  a i  in  A x . In addition, we find a strong relationship between the problem of cover length determination and the non-negative least square (NNLS) problem such that we can obtain an analytical result of the NNLS problem by simply determining the cover length of the corresponding variable. We also include a discussion of the various algorithms for solving the NNLS problem and the cover length method developed here.

6.1. Determination of Cover Length

In general, the cover length is obtained by solving the following optimization problem:
Problem 1. 
Let  A R M × N x = x 1 , x 2 , , x N T R + N  and  x N  be covered in  A x .
max x N subject to x T A T A x 1
where  x n 0  for  n = 1 , 2 , , N .
The maximum value of  x N  within the constraints is the cover length of the covered variable  x N . To solve the above optimization problem, let us form a Lagrangian function:  L ( x , λ ) = x N n = 1 N λ n x n + λ N + 1 ( x T A T A x 1 ) / 2 , where  λ n > 0  for  n = 1 , 2 , , N + 1 . Then, the necessary and sufficient condition for  x *  to be an optimal solution is that the following Karush–Kuhn–Tucker (KKT) condition must be satisfied:
L ( x , λ ) | x = x * , λ = λ * = e N λ * + λ N + 1 * A T A x * = 0 x n * λ n * = 0 for n = 1 , 2 , , N λ N + 1 * ( x * ) T A T A x * 1 = 0 ( x * ) T A T A x * 1 x * 0 λ N + 1 * 0 , λ * 0
where the non-negative vector  λ * R + N  is associated with the optimal vector  x *  such that  L ( x * , λ * )  is a stationary point of  L ( x , λ ) . On the other hand, we notice that
x T A T A x = p N N x N + p ¯ N T x ¯ N p N N 2 + x ¯ N T P ¯ N N p ¯ N p ¯ N T p N N x ¯ N
where  P = A T A  is an  N × N  positive semidefinite (PSD) matrix,  p N N  is the  N N -th element in  P P ¯ N N  is the  ( N 1 ) × ( N 1 )  sub-matrix of  P  by deleting the N-th row and N-th column from it,  p ¯ N  is the  ( N 1 ) × 1  vector generated by deleting the N-th entry from the N-th row of  P , and  x ¯ N  denotes the  ( N 1 ) × 1  vector obtained by deleting the N-th entry from  x . Therefore, we can represent the KKT condition alternatively as:
λ ¯ N * + λ N + 1 * x N * + p ¯ N T x ¯ N * p N N p ¯ N + P N N p ¯ N p ¯ N T p N N x ¯ N * = 0 1 λ N * + λ N + 1 * p N N x N * + p ¯ N T x ¯ N * = 0 x n * λ n * = 0 λ N + 1 * ( x * ) T A T A x * 1 = 0 ( x * ) T A T A x * 1 x * 0 λ N + 1 * 0 , λ * 0
Here,  λ ¯ N *  and  x ¯ N *  denote, respectively, the  ( N 1 ) × 1  vectors obtained by deleting the N-th entry from  λ *  and  x * . Since  x N * 0 , we have  λ N + 1 * 0 , and, thus,  x N * = λ N + 1 * . Using the KKT condition, the solution to Problem 1 is given in the following theorem:
Theorem 10. 
Let  A  be an  M × N  real matrix with its rank being R. Then,  x n  is covered in  A x  if and only if there exists an invertible principal sub-matrix  P i 1 i 2 i r  of order r in  A T A  that includes the  n n -th element  [ A T A ] n n , such that the following two conditions are satisfied simultaneously:
1. 
P i 1 i 2 i r | i j = n 1 e j 0  and  P i 1 i 2 i r | i j = n 1 e j j > 0 ;
2. 
det ( P i 1 i r | i j = n k ) 0  for  k { 1 , , N } { i 1 , , i r } , where  P i 1 i r | i j = n k  is the sub-matrix of  A T A  by replacing the old row:  p n , i 1 , , p n , i r  in  P i 1 i r | i j = n  with the new row  ( p k , i 1 , , p k , i r ) .
Then the cover length is given by  c n ( x n ) = P i 1 i 2 i r | i j = n 1 n n .
Proof. 
The KKT condition of Problem 1 can be simplified as:  P x = b , x 0 , b 0 , b N > 0  and  x i b i = 0  for  i = 1 , 2 , , N 1 , where  b R + N . Let  N ¯  be the set consisting of all the indices of  x i  which are all positive in the variable  x . Denoting the cardinality of a set as  | · | , we are able to find an  | N ¯ | × | N ¯ |  sub-matrix  P ¯  of  P , such that  P ¯ x ¯ = e | N | ¯  with all  x i  in  x ¯  being uncovered variables in  x  and  x ¯ | N | ¯ = x N , i.e., the last entry in  x ¯  is equivalent to the last one in  x . Then there exists a full column rank matrix  T R | N ¯ | × r , where  r | N ¯ | , containing  P ¯ | N | ¯  in  P ¯ . Without loss of generality, we can let  T = t 1 , , t r 1 , P ¯ | N ¯ |  and we will have  T x ˜ = e | N | ¯ , where  x ˜ R + r  and  x ˜ r > 0 . This can be proved in the following: Let  T  be the smallest set containing  P ¯ | N ¯ | , s.t.  e | N | ¯ cone T , where  cone T = cone t 1 , , t r 1 , P ¯ | N ¯ | = { θ 1 t 1 + + θ r 1 t r 1 + θ r P ¯ | N ¯ | | θ i 0 for i = 1 , , r } . Then  T  is linearly independent; otherwise, there are  μ j R  (not all 0), s.t.  j = 1 r 1 μ j t j + μ r P ¯ | N ¯ | = 0 . And there are  λ j 0 , s.t.  j = 1 r 1 λ j t j + λ r P ¯ | N ¯ | = e | N | ¯ , where  λ r > 0 . Then, we have  j = 1 r 1 ( α μ j + λ j ) t j + ( α μ r + λ r ) P ¯ | N ¯ | = e | N | ¯ . If  μ r 0 , then let  α = max 1 j r 1 { λ j μ j , μ j > 0 } = λ i μ i . Thus, for every  1 j r 1 λ j + α μ j 0  and  λ i + α μ i = 0 . Then, we can have a new  x ˜ R r 1  with the last element of it, which is  λ r + α μ r , being positive, while the others are  r 2  elements, which are expressed as  λ j + α μ j , being non-negative. When  μ r < 0 , we let  α = max 1 j r 1 { λ j μ j , μ j > 0 } = λ i μ i . Then, we have a new  x ˜ R r 1  with the last element of it being positive while the others are non-negative in the same manner as the case when  μ r 0 . As a result, we can always find a smaller set  T ˜  containing  P ¯ | N ¯ | , s.t.  e | N | ¯ cone T ˜ . Thus,  T  is linearly independent. According to the constraints of  x x ˜  should be equivalent to  x ¯  and  T = P ¯ . As a result,  P ¯  is invertible. Since we have  P ¯ x ¯ = e | N | ¯ , where  x ¯ R + r  and  x ¯ r > 0 , we will have  P i 1 i 2 i r | i j = n 1 e j 0  and the j-th element in it is positive. Until now, the first statement has been proven.
By using the new row  ( p k , i 1 , , p k , i r )  to replace the old row  ( p n , i 1 , , p n , i r )  in  P i 1 i 2 i r | i j = n , we will have  P i 1 i 2 i r | i j = n k x ¯ = b k e j . To simplify the expression, we denote  P i 1 i 2 i r | i j = n k  as  P ¯ k . If  P ¯ k  is invertible, then  x n = b k P ¯ ( r 1 ) × ( r 1 ) P ¯ k , where  P ¯ ( r 1 ) × ( r 1 )  is the  ( r 1 ) -th order leading principle sub-matrix of  P i 1 i 2 i r . Since  P i 1 i 2 i r  is a positive definite matrix, it follows that  P ¯ ( r 1 ) × ( r 1 ) > 0 . In addition,  b k 0 x n > 0 , then we have det ( P ¯ k ) > 0 . When  b k = 0 , det ( P ¯ k ) = 0 . As a result, det ( P ¯ k ) 0 , for   k = 1 , 2 , , N  but  k i 1 , i 2 , , i r .
When the above conditions are all satisfied, the cover length of  x n  can be obtained directly, which is  P i 1 i 2 i r | i j = n 1 n n .    □
From the above result, we can also conclude that if there is no principal sub-matrix that can satisfy all the conditions, then the corresponding variable is uncovered within the constraint.
The following example illustrates how the above method can be used to obtain the cover length of the covered variable:
Example 4. 
Determine the cover length of covered variable  x 4  given the following  4 × 4  matrix and its PSD matrix:
A = 3 2 5 2 3 5 0 4 1 3 1 3 2 2 1 4 , P = A T A = 23 2 18 1 2 42 15 23 18 15 27 11 1 23 11 45
We need to find out the principal sub-matrix that can satisfy all the conditions listed in Theorem 10. We first examine all the  2 × 2  principal sub-matrices of  P  containing a negative element in the right upper side corner, since only such a principle sub-matrix of order-2 could satisfy the condition that the last column of its inverse is a positive vector. Inspection of the above PSD matrix shows that there is only one such principal sub-matrix:  P 14 = 23 1 1 45 . We verify that  P 14  above is invertible and the last column of its inverse matrix is a positive column vector. Then, we replace the second row in  P 14  with the other rows, resulting in  P 14 | 4 2 = 23 1 2 23  and  P 14 | 4 3 = 23 1 18 11 . The determinants of both are verified to be non-negative. From the above discussion, we can see that the invertible  2 × 2  principal sub-matrix  P 14  satisfies all the conditions in Theorem 10 and we have:  P 14 1 = 45 1034 1 1034 1 1034 23 1034 . Thus, we can conclude that the cover length of  x 4  is  c 4 ( x 4 ) = 23 1034 .
Lemma 4. 
For any  A R M × N  and  x R + N ,
1. 
If all the entries of  A T A  are positive, then the cover length of  x n  is  c n ( x n ) = 1 A T A n n .
2. 
If  A T A  has full rank and all the entries in the n-th column of  A T A 1  are positive, then the cover length is  c n ( x n ) = A T A 1 n n .
Proof. 
To prove the first statement, given an  M × N  real matrix  A  and  x R + N , we can rewrite  x T A T A x  as:
x T A T A x = x ¯ T A ¯ T A ¯ x ¯ + x ¯ T A ¯ T a n x n + a n T A ¯ x ¯ x n + a n T a n x n 2
where  A ¯  is the  M × ( N 1 )  sub-matrix formed by deleting the n-th column of  A x ¯  denotes an  ( N 1 ) × 1  vector obtained by deleting n-th entry from  x  and  a n  is the n-th column of  A .
According to the assumption of Statement 1, i.e., all the entries of  A T A  are positive, then all terms in the above equations are non-negative. Thus, for any given positive real-valued number  τ > 0 x T A T A x τ 2  implies that  a n T a n x n 2 τ 2 , which gives  x n τ a n T a n . Therefore, according to the definition of cover length in Definition 1, the cover length of  x n  is given by  c n ( x n ) = 1 A T A n n .
The second statement can be obtained from Theorem 10 directly.    □

6.2. Cover Length Problem and NNLS Problem

The NNLS problem is a constrained least squares regression problem in which all the variables can only take non-negative values. Specifically, the NNLS problem can be stated as follows [45]:
Problem 2 (Non-negative Least Squares (NNLS)). 
Given  B R M × N  and  b R M , find a non-negative vector  u R + N  such that
min Bu b 2 2 subject to u 0
In the following, we show by introducing a new variable that the NNLS problem can be turned into a problem of determining the cover length of the corresponding variable. In so doing, a connection between cover length determination and the NNLS problem is established, providing us with a method to arrive at the closed-form optimal value of the objective function.
First, we let
τ 2 = Bu b 2 2
When  τ = 0 , Problem 2, is equivalent to the problem of finding solutions for the nonhomogeneous system of linear equations  Bu = b  with non-negative constraints on  u . Let us consider the case when  τ > 0 : by dividing  τ 2  on both sides of Equation (30), we will have
B u τ b 1 τ 2 2 = 1
Introducing a new variable  x = u τ , 1 τ T , the origin problem can be transformed into:
Problem 3. 
max x N + 1 subject to A x 2 2 = 1
where  x n 0  for  n = 1 , 2 , , N x N + 1 = 1 τ > 0  and  A = ( B , b ) .
We observe that Problem 3 is of the same form as Problem 1 and is consistent with Problem 2. Thus, the NNLS and the cover length determination problem are equivalent. By solving the cover length of the corresponding variable  x N + 1 , we obtain the equivalent closed-form optimal value of the objective function in the NNLS problem. If we are not able to find the cover length of this variable,  x N + 1  is unbounded within the constraint and the optimum value of the objective function in the NNLS problem is almost zero.
Example 5. 
The cover length determination problem in Example 4 is consistent with the NNLS problem:  min u R + 3 Bu b 2 2 , where  B = 3 2 5 3 5 0 1 3 1 2 2 1  and  b = ( 2 , 4 , 3 , 4 ) T . Let  τ 2 = Bu b 2 2  and  x = ( x 1 , x 2 , x 3 , x 4 ) T = ( u τ , 1 τ ) T . The cover length of  x 4  is  c 4 ( x 4 ) = 1 23 / 1034 = 1 τ ; thus, the optimal value of this NNLS problem is  τ 2 = Bu b 2 2 = 1 c 4 ( x 4 ) 2 = 23 / 1034 .
The above example demonstrates how to convert the cover length determination of a desired variable into the optimal value of the corresponding NNLS problem and verifies the equivalence of the two problems. For certain types of matrices, using this equivalence, we can even directly obtain the analytical optimal value of the NNLS problem. This is demonstrated by the example of the M-matrix in the following. Let us first define the Z- and the M-matrices [46]:
Definition 4 
(Z-matrix). An  N × N  real matrix in which the off-diagonal entries are less than or equal to zero, i.e., a matrix of the form  A = ( a i j )  with  a i j 0 i j , 1 i , j N , is a real Z-matrix.
Definition 5 
(M-matrix). Let  A  be an  N × N  real Z-matrix. Then  A  is also an M-matrix if it can be expressed in the form  A = s I T , where  T = ( t i j )  with  t i j 0 , for all  i j 1 i , j N , where s is at least as large as the maximum of the moduli of the eigenvalues of  T , and  I  is an identity matrix.
Theorem 11. 
Let  A R N × N  be a Z-matrix; then, the following statements are equivalent to  A  being a nonsingular M-matrix:
1. 
All the principal minors of  A  are positive. That is, the determinant of each sub-matrix of  A  obtained by deleting a set, possibly empty, of corresponding rows and columns of  A  is positive.
2. 
A  is inverse-positive. That is,  A 1  exists and  A 1  is a non-negative matrix.
Then, with the properties of M-matrix and cover length, we have the following result.
Theorem 12. 
Let matrix  B R N × ( N 1 )  and vector  b R N . Denote  A  as  ( B , b ) . Supposing that  A  is a nonsingular M-matrix, then the optimal value of the NNLS problem  min u R + N 1 Bu b 2 2  is exactly equal to  1 ( A T A ) 1 N N .
Proof. 
By assumption,  A  is a nonsingular M-matrix, therefore  A T A  is invertible and all the elements in  ( A T A ) 1  are positive. Reformulating the NNLS problem to the problem of determining the cover length of  x N  by applying Lemma 4, the cover length of the corresponding variable  x N  is given by  c N ( x N ) = A T A 1 N N . Thus, the optimal value  τ 2 = ( 1 c N ( x N ) ) 2 = 1 ( A T A ) 1 N N .    □

6.3. Comparison with the Active-Set Method

There are several normally used active-set methods for solving the NNLS problem. A typical example is the algorithm lsqnonneg in Matlab, which aims at creating an active set and using it to arrive at an approximate solution to the NNLS problem. lsqnonneg starts with an all-zero vector and computes the associated negative gradient vector  w . Then it finds the index of the position where the maximum value in  w  occurs and moves this index from the inactive set to the active set. By solving the corresponding least squares problem with the current active set, one non-negative solution candidate can be obtained. The active set and inactive set can be updated with the current candidate solution and continue the whole process until all the elements in  w  are nonpositive or the inactive set is empty. As Lawson and Hanson showed, this algorithm always converges and terminates in finite steps. However, there is no upper limit on the possible number of iterations that the algorithm might need to reach the point of the optimum solution, and it might be very slow in practice, owing largely to the computation of the pseudo-inverse. With regard to the computational complexity, since the exact running time required for the NNLS solver is unknown, the computational cost cannot be specified exactly. In many standard implementations of NNLS solvers (and particularly those based on active-set methods), the cost is typically  O ( M N 2 )  per iteration [47].
Compared with the active-set method, the cover length determination method is finite, and once we find one principal sub-matrix that can satisfy the conditions in Theorem 10, then the computation stops. Furthermore, we can find an upper limit on the possible number of steps that the algorithm needs and obtain a closed-form optimal value of the NNLS problem.
From the perspective of computation complexity, there is no clear advantage of the cover length method compared with the lsqnonneg since it involves the combination and permutation operations. However, while the accuracy of the lsqnonneg solution depends on a prescribed tolerance  ϵ , the cover length method yields the exact optimal value of the objective function.
We now present some numerical results illustrating the performance of the cover length method and lsqnonneg in solving the NNLS problem.
Table 1 shows the average running time (seconds) and average error of the lsqnonneg and the cover length method for the matrices and vectors randomly generated by Matlab’s rand function. The results shown here are averaged over 100 random samples with varying number of columns (from one to three) of  B  in the NNLS problem. The default termination tolerance on the solution of lsqnonneg is  10 × i j | a i j | × N × e p s , where  e p s = 2.22 × 10 16 , N is the row number of the matrix  B , and  a i j  is the element in  A = ( B , b ) . Table 1 also includes the computation complexity (number of maximum operations) of the cover length method in solving the NNLS problem. It is clear from the table that the advantage of the cover length method over lsqnonneg lies in the accuracy of the optimal value since cover length yields a closed-form one.

7. Conclusions

Linear systems of equations with non-negativity constraints on solutions is an area of study in linear algebra. Such problems arise frequently in many fields of science and engineering. In our consideration of such problems, we discovered the hyper-rectangle cover theory of a matrix, which is presented in this paper. The two main concepts in the hyper-rectangle cover theory, viz., the cover order and the cover length, were first defined, and many of their important properties were introduced. Based on this theory, several novel approaches to analyzing the above typical problems were proposed. The necessary and sufficient conditions under which a unique solution for a system of linear equations with non-negativity constraints exists were identified. We also showed how the specific échelon form of the matrix is constructed, and with this échelon form, the cover order of any given matrix can be determined.
With the help of cover theory, the emptiness of the feasibility set and the various possibilities of the solution for the LP problem were analyzed in detail. In addition, with the property of zero-cover matrix, a series of feasible solutions to the LP problem can be obtained.
Our study on the cover length led us to the development of a method to find the cover length of a covered variable. We also showed the equivalence between cover length determination and the NNLS problem so that the NNLS problem can be solved with the cover length method. This provides us with the analytical optimal value obtainable from the structure of the matrix rather than a numerical result having a finite accuracy. The development of the hyper-rectangle cover theory, thus, not only provides us with an efficient method to solve the system of linear equations with non-negativity constraints, it also suggests to us attractive alternative approaches to the LP and the NNLS problems.

Author Contributions

Conceptualization, X.C., K.M.W. and J.Z.; methodology, X.C., K.M.W., J.C. and J.Z.; software, X.C.; validation, X.C. and K.M.W.; formal analysis, X.C., K.M.W. and J.Z.; investigation, X.C., K.M.W. and J.Z.; resources, X.C. and J.Z.; data curation, X.C.; writing—original draft preparation, X.C.; writing—review and editing, K.M.W., J.C. and J.Z.; visualization, X.C. and K.M.W.; supervision, K.M.W.; project administration, X.C. and K.M.W.; funding acquisition, K.M.W. and J.Z.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Acknowledgments

This paper was written with profound gratitude to, and in fond memory of, Jiankang Zhang who, while pioneering much of the original research in hyper-rectangle cover theory, guided the first author (X.C.) with the utmost care in her research, and introduced, with great enthusiasm, some intriguing ideas in the subject to the second author (K.M.W.). His sudden departure from life left those who worked with him with a feeling of loss. He will always be remembered as a great teacher and a great colleague.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lay, D.C. Linear Algebra and Its Applications, 5th ed.; Pearson: New York, NY, USA, 2016. [Google Scholar]
  2. Bebikhov, Y.V.; Semenov, A.; Yakushev, I.; Kugusheva, N.; Pavlova, S.; Glazun, M. The application of mathematical simulation for solution of linear algebraic and ordinary differential equations in electrical engineering. In Proceedings of the IOP Conference Series: Materials Science and Engineering, Wuhan, China, 10–12 October 2019; IOP Publishing: Bristol, UK, 2019; Volume 643, p. 012067. [Google Scholar]
  3. Dianat, S.A.; Saber, E. Advanced Linear Algebra for Engineers with MATLAB; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar]
  4. Golomb, S.W.; Gong, G. Signal Design for Good Correlation: For Wireless Communication, Cryptography, and Radar; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar]
  5. Bardsley, J.M.; Knepper, S.; Nagy, J. Structured linear algebra problems in adaptive optics imaging. Adv. Comput. Math. 2011, 35, 103. [Google Scholar] [CrossRef]
  6. Datta, B.N. Linear and numerical linear algebra in control theory: Some research problems. Linear Algebra Its Appl. 1994, 197, 755–790. [Google Scholar] [CrossRef]
  7. Joshi, H.; Yavuz, M.; Townley, S.; Jha, B.K. Stability analysis of a non-singular fractional-order COVID-19 model with nonlinear incidence and treatment rate. Phys. Scr. 2023, 98, 045216. [Google Scholar] [CrossRef]
  8. Joshi, H.; Jha, B.K.; Yavuz, M. Modelling and analysis of fractional-order vaccination model for control of COVID-19 outbreak using real data. Math. Biosci. Eng. 2023, 20, 213–240. [Google Scholar] [CrossRef]
  9. Joshi, H.; Jha, B.K. 2D memory-based mathematical analysis for the combined impact of calcium influx and efflux on nerve cells. Comput. Math. Appl. 2023, 134, 33–44. [Google Scholar] [CrossRef]
  10. Anton, H.; Rorres, C. Elementary Linear Algebra: Applications Version; John Wiley & Sons: Hoboken, NJ, USA, 2013. [Google Scholar]
  11. Demmel, J.W. Matrix Computations (Gene H. Golub and Charles F. van Loan). SIAM Rev. 1986, 28, 252–255. [Google Scholar] [CrossRef]
  12. Horn, R.A.; Johnson, C.R. Matrix Analysis; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
  13. Roman, S.; Axler, S.; Gehring, F. Advanced Linear Algebra; Springer: Berlin/Heidelberg, Germany, 2005; Volume 3. [Google Scholar]
  14. Dines, L.L. On Positive Solutions of a System of Linear Equations. Ann. Math. 1926, 28, 386–392. [Google Scholar] [CrossRef]
  15. Schrijver, A. Theory of Linear and Integer Programming; John Wiley & Sons: Hoboken, NJ, USA, 1998. [Google Scholar]
  16. Dantzig, G.B.; Thapa, M.N. Linear Programming 1: Introduction; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  17. Karmarkar, N. A new polynomial-time algorithm for linear programming. In Proceedings of the Sixteenth Annual ACM Symposium on Theory of Computing, Washington, DC, USA, 30 April–2 May 1984; pp. 302–311. [Google Scholar]
  18. Khachiyan, L.G. Polynomial algorithms in linear programming. USSR Comput. Math. Math. Phys. 1980, 20, 53–72. [Google Scholar] [CrossRef]
  19. Potra, F.A.; Wright, S.J. Interior-point methods. J. Comput. Appl. Math. 2000, 124, 281–302. [Google Scholar] [CrossRef]
  20. Wright, M. The interior-point revolution in optimization: History, recent developments, and lasting consequences. Bull. Am. Math. Soc. 2005, 42, 39–56. [Google Scholar] [CrossRef]
  21. Dantzig, G. Linear Programming and Extensions; Princeton University Press: Princeton, NJ, USA, 2016. [Google Scholar]
  22. Dantzig, G.B.; Thapa, M.N. Linear Programming 2: Theory and Extensions; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  23. Murty, K.G. Linear Programming; Springer: Berlin/Heidelberg, Germany, 1983. [Google Scholar]
  24. Bro, R.; De Jong, S. A fast non-negativity-constrained least squares algorithm. J. Chemom. J. Chemom. Soc. 1997, 11, 393–401. [Google Scholar] [CrossRef]
  25. Gill, P.E.; Murray, W.; Wright, M.H. Practical Optimization; SIAM: Philadelphia, PA, USA, 2019. [Google Scholar]
  26. Van Benthem, M.H.; Keenan, M.R. Fast algorithm for the solution of large-scale non-negativity-constrained least squares problems. J. Chemom. J. Chemom. Soc. 2004, 18, 441–450. [Google Scholar] [CrossRef]
  27. Franc, V.; Hlaváč, V.; Navara, M. Sequential coordinate-wise algorithm for the non-negative least squares problem. In Proceedings of the International Conference on Computer Analysis of Images and Patterns, Versailles, France, 5–8 September 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 407–414. [Google Scholar]
  28. Kim, D.; Sra, S.; Dhillon, I.S. A New Projected Quasi-Newton Approach for the Nonnegative Least Squares Problem; Citeseer: Princeton, NJ, USA, 2006. [Google Scholar]
  29. Bellavia, S.; Macconi, M.; Morini, B. An interior point Newton-like method for non-negative least-squares problems with degenerate solution. Numer. Linear Algebra Appl. 2006, 13, 825–846. [Google Scholar] [CrossRef]
  30. Cantarella, J.; Piatek, M. Tsnnls: A solver for large sparse least squares problems with non-negative variables. arXiv 2004, arXiv:cs/0408029. [Google Scholar]
  31. Chen, D.; Plemmons, R.J. Nonnegativity constraints in numerical analysis. In The Birth of Numerical Analysis; World Scientific: Singapore, 2010; pp. 109–139. [Google Scholar]
  32. Portugal, L.F.; Judice, J.J.; Vicente, L.N. A comparison of block pivoting and interior-point algorithms for linear least squares problems with nonnegative variables. Math. Comput. 1994, 63, 625–643. [Google Scholar] [CrossRef]
  33. Lawson, C.; Hanson, R. Solving Least-Squares Problems; Prentice-Hall: Upper Saddle River, NJ, USA, 1974; Chapter 23. [Google Scholar]
  34. Zhang, Y.Y.; Yu, H.Y.; Zhang, J.K.; Wang, J.L. Reliable MIMO Optical Wireless Communications Through Super-Rectangular Cover. arXiv 2016, arXiv:1607.04206. [Google Scholar]
  35. Chu, X. Hyper-Rectangle Cover Theory and Its Applications. Ph.D. Thesis, McMaster University, Hamilton, ON, Canada, 2022. [Google Scholar]
  36. Grünbaum, B.; Klee, V.; Perles, M.A.; Shephard, G.C. Convex Polytopes; Springer: Berlin/Heidelberg, Germany, 1967; Volume 16. [Google Scholar]
  37. Roman, S. Positive Solutions to Linear Systems: Convexity and Separation. In Advanced Linear Algebra; Springer: New York, NY, USA, 2005; pp. 395–408. [Google Scholar] [CrossRef]
  38. Chu, M.; Plemmons, R. Nonnegative matrix factorization and applications. Bull. Int. Linear Algebra Soc. 2005, 34, 26. [Google Scholar]
  39. Petrou, M.M.; Petrou, C. Image Processing: The Fundamentals; John Wiley & Sons: Chichester, West Sussex, UK, 2010. [Google Scholar]
  40. Fu, X.; Huang, K.; Sidiropoulos, N.D.; Ma, W.K. Nonnegative Matrix Factorization for Signal and Data Analytics: Identifiability, Algorithms, and Applications. IEEE Signal Process. Mag. 2019, 36, 59–80. [Google Scholar] [CrossRef]
  41. Difonzo, F.V. A note on attractivity for the intersection of two discontinuity manifolds. Opusc. Math. 2020, 40, 685–702. [Google Scholar] [CrossRef]
  42. Dieci, L.; Difonzo, F. The moments sliding vector field on the intersection of two manifolds. J. Dyn. Differ. Equ. 2017, 29, 169–201. [Google Scholar] [CrossRef]
  43. Trefethen, L.N.; Bau, D., III. Numerical Linear Algebra; SIAM: Philadelphia, PA, USA, 1997; Volume 50. [Google Scholar]
  44. Klee, V.; Minty, G.J. How good is the simplex algorithm. Inequalities 1972, 3, 159–175. [Google Scholar]
  45. Lawson, C.L.; Hanson, R.J. Solving Least Squares Problems; SIAM: Philadelphia, PA, USA, 1995. [Google Scholar]
  46. Plemmons, R.J. M-matrix characterizations. I—nonsingular M-matrices. Linear Algebra Its Appl. 1977, 18, 175–188. [Google Scholar] [CrossRef]
  47. Björck, Å. Numerical Methods for Least Squares Problems; SIAM: Philadelphia, PA, USA, 1996. [Google Scholar]
Figure 1. Example of a  2 × 2  full-cover matrix.
Figure 1. Example of a  2 × 2  full-cover matrix.
Mathematics 11 02338 g001
Figure 2. Example of a zero-cover matrix.
Figure 2. Example of a zero-cover matrix.
Mathematics 11 02338 g002
Figure 3. Flow diagram of the cover method.
Figure 3. Flow diagram of the cover method.
Mathematics 11 02338 g003
Table 1. Comparison between Matlab’s lsqnonneg and cover length method in solving the NNLS problem.
Table 1. Comparison between Matlab’s lsqnonneg and cover length method in solving the NNLS problem.
Column Number of B123
ComplexityCover length method616589
Running timeCover length method   1.20 × 10 4   2.40 × 10 4 0.0024
lsqnonneg   2.10 × 10 4   3.17 × 10 4   4.18 × 10 4
Average errorCover length method   0.0000   0.0000 0.0000
lsqnonneg   1.1102 × 10 16   2.6645 × 10 15   2.8422 × 10 14
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chu, X.; Wong, K.M.; Chen, J.; Zhang, J. Systems of Linear Equations with Non-Negativity Constraints: Hyper-Rectangle Cover Theory and Its Applications. Mathematics 2023, 11, 2338. https://doi.org/10.3390/math11102338

AMA Style

Chu X, Wong KM, Chen J, Zhang J. Systems of Linear Equations with Non-Negativity Constraints: Hyper-Rectangle Cover Theory and Its Applications. Mathematics. 2023; 11(10):2338. https://doi.org/10.3390/math11102338

Chicago/Turabian Style

Chu, Xiaoxuan, Kon Max Wong, Jun Chen, and Jiankang Zhang. 2023. "Systems of Linear Equations with Non-Negativity Constraints: Hyper-Rectangle Cover Theory and Its Applications" Mathematics 11, no. 10: 2338. https://doi.org/10.3390/math11102338

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop