1. Introduction and Preliminaries
There are many research articles on the linear operators which preserve certain matrix properties over some matrix spaces (see [
1,
2,
3,
4,
5]). However, there are few articles for the characterizations of the linear operators that preserve arctic rank of matrices [
3,
4]. In this paper, we study some properties of arctic rank of Boolean matrices. We compare the arctic rank with Boolean rank and term rank of a given Boolean matrix. Furthermore, we obtain some characterizations of linear operators that preserve arctic rank on Boolean matrix space.
The Boolean algebra consists of the set  equipped with two binary operations, addition and multiplication. Two operations are defined as usual except that .
Every matrix whose entries are in  is called a Boolean matrix, and we write  for the set of all  Boolean matrices. The usual definitions for addition and multiplication of matrices over fields are applied to . If , we use  instead of . We denote the  identity matrix by , and the  matrix all of whose entries are 1 by . The zero matrices of any size are denoted by O.  is the transpose of .
A Boolean matrix in 
 is called a cell [
4] if it has exactly one nonzero entry. We denote the cell whose nonzero entry is in 
-th position by 
. The cell 
 is called diagonal or off-diagonal according as 
 or 
.
Every element in  is called a vector, and it is represented as lowercase and boldface letter. That is,  is a column vector, and  is the row vector for . Let  denote the number of nonzero entries in . In particular,  with  is the vector whose only nonzero entry is in i-th position for .
The Boolean rank [
1] of nonzero 
, 
, is the least integer 
k such that 
A is the sum of 
k Boolean matrices of rank one. That is, 
 is the least integer 
k for which there exist nonzero vectors 
 and 
 such that
      
Thus, if  is a Boolean rank-1 matrix, then  for some nonzero vectors  and . It is easy to verity that these vectors  and  are uniquely determined by A. Every nonzero Boolean matrix  can be written as a sum of Boolean rank-1 matrices. Here, the sum is called a decomposition of A. As an easy example,  is a decomposition of A, where  is the i-th column of A.
The arctic rank ([
6] of nonzero 
, 
, is the least value of
      
      over all decompositions, 
, of 
A. Evidently, 
 for every cell 
 since 
 is the unique decomposition of 
, and 
 since 
 is the unique decomposition of 
, where 
 and 
 are vectors all of whose entries are 1. Clearly, 
 is the unique decomposition of 
, and hence 
. In above definition, the value of 
l providing optimal decomposition of 
A can be greater than the Boolean rank of 
A, see Example 2.6 in [
6]. Note that the arctic rank of 
 may not be an integer, and may be much larger than 
. For example, consider the 
 Boolean matrix 
. As shown in Proposition 1 below, we have 
. Thus, the arctic rank of 
A is not integer and is larger than 4. In general, for any nonzero 
, the value of 
 is of the form 
, where 
 is a positive integer. It is well-known [
6] that the arctic rank of a Boolean matrix in 
 cannot be grater than 
.
In this research, we show that the arctic rank of a matrix is equal or greater than both the Boolean rank (Theorem 1) and the term rank (Theorem 2). This implies that the arctic rank of a matrix is an upper bound of these matrix functions. So, arctic rank can be used to estimate the Boolean rank or the term rank of a Boolean matrix. In addition, in the linear preserver problems, this research is the beginning of the arctic rank preserver problems over integers or real field that is more complex than the Boolean case. So, this paper can also help in revitalizing the research on linear preserver problems.
  2. Basic Results for Arctic Ranks of Boolean Matrices
In this section, we study the basic behavior of the arctic rank of Boolean matrices. In Boolean matrix space, we compare the arctic rank with the Boolean rank and the term rank.
Theorem 1. If  is nonzero, then .
 Proof.  Let  be an optimal decomposition of A such that . Since  and  are nonzero, we have  and  for all . If , then  by the definition of Boolean rank. It follows that .    □
 Two cells  and  in  are called collinear if either  or . This terminology can be extended to cells more than two cells.
Lemma 1. Let  be nonzero. Then the followings hold:
- (1)
- if  is a Boolean matrix which is obtained from A by deleting some zero rows or columns, then  
- (2)
- ; 
- (3)
- if  and  are permutation matrices, then ; 
- (4)
- if A is a sum of  cells which are collinear, then . 
 Proof.  - (1)
- It is an easy exercise to check. 
- (2)
- Note that  is a decomposition of A if and only if  is a decomposition of . This implies . 
- (3)
- First, suppose that  -  is a decomposition of  A- . Then
             - 
            is a decomposition of  - . Next, suppose that  -  is a decomposition of  - . Since  P-  and  Q-  are permutation matrices, we have  -  and  - . It follows that
             - 
            is a decomposition of  A- . From these two inclusions, we see that the set of decompositions of  -  is equal to that of  A- . This shows  - . 
- (4)
- Clearly . By (2) and (3), without loss of generality we assume that . Then for the vector  with ,  is the unique decomposition of A. Thus . 
        □
      
 Lemma 2. Let  be nonzero. Then .
 Proof.  If  and  are decompositions of A and B, respectively, then their sum is a decomposition of . This shows .    □
 For , let  denote the number of nonzero entries in A.
Corollary 1. If  is nonzero, then .
 Proof.  Let A be a sum of h cells  so that . Since  for all , by Lemma 2, we have .    □
 For Boolean matrices  and , their concatenation is denoted by  so that it is an  Boolean matrix. Then by Lemma 1, we have  and .
From now on, we will assume that 
 unless specified otherwise. Consider 
 for which 
 if 
 and 
 otherwise. Then for 
, we can write 
, where 
 is the zero vector in 
. Hence by Lemmas 1 and 2, we have
      
In fact,  by the below Proposition 1.
If 
 and 
 are Boolean matrices in 
, we say that 
A dominates [
5] 
B, written 
, if 
 implies 
 for all 
i and 
j. Equivalently, 
 if and only if 
.
Proposition 1. For , if , then .
 Proof.  By Lemma 1, we may assume that 
 is the unique zero entry of 
A. For any decomposition, 
, of 
A, consider the value 
. Let 
I be a subset of 
 such that 
 if the first entry of 
 is 1. Without loss of generality, we assume that 
 with 
. In this situation, we let 
, 
, 
 and 
. Then we have
        
Note that the first entry of 
 is 0 by definitions of 
I and 
. This implies that the 
th entry of 
 is 0. Suppose that the 
th entry of 
 is 1. Then the first entry of 
 is 1, and hence there exists an index 
 such that the first entry of 
 is 1. The first entry of 
 is 1 by definition of 
I, and thus the 
-th entry of 
 is 1. This contradicts 
. That is, the 
-th entry of 
 is 0, and hence the 
-th entry of 
 is 0. This implies 
 since 
 is the unique zero entry of 
A. Consequently, 
, and we conclude that the value 
 can be optimal when 
. Let 
 be an optimal decomposition of 
A so that 
. Put 
, 
, 
 and 
. If 
 or 
, then we can easily check that 
 is not equal to 
A. Thus we can say that 
 and 
 so that
        
This implies , ,  and . Then  for every  and . In this situation, the optimal is when  and  for every  and . That is, we have , ,  and . It follows that .    □
 If A and B are Boolean matrices of suitable size, then their direct sum is denoted by  so that . In this case, we can easily check that .
The term rank [
2] of nonzero 
, 
, is the minimum number of lines that needed to include all nonzero entries of 
A. Here, a line of 
A is a row or column of 
A. It follows from 
 that
      
      for all nonzero 
. It is clear that if 
 has term rank 
h, then there exist permutation matrices 
 and 
, such that 
.
Theorem 2. If  is nonzero, then .
 Proof.  If , without loss of generality, we assume that . Let  be a decomposition of A. Then for any , there exist  such that . It follows from  that , and hence .    □
 For , let  denote the set of Boolean matrices in  whose arctic rank is k. Let  be the set of members in  with fewest number of nonzero entries. That is, if , then  for all . Notice that any member in  must be a cell, and hence . Any member in  must be a sum of two cells that are collinear, and thus . If , then  is not equal to . For example, consider two members  and  in . Then , while .
If , using Theorem 2, we can obtain the specific structure of elements in  as following:
Proposition 2. Let . Then for , we have  if and only if .
 Proof.  Assume . Then there exist permutation matrices  and  such that . Thus by Lemma 1, . If , then  for all . Then, by Corollary 1,  and thus , a contradiction. Hence we must have .
Conversely, assume 
. Then 
 and so 
 by Theorem 2. Consider 
 with 
. Then 
 and thus 
. Hence we can write 
 for distinct cells 
. If 
, then there exist two cells in 
 that are collinear. Without loss of generality, we assume that 
 and 
 are collinear. Then, by Lemma 2,
        
        which is impossible. Hence we must have 
.    □
 Lemma 3. For  and , let . Then for , if  and A dominates  cells that are collinear, then .
 Proof.  Since 
, we can assume 
. Suppose 
 with 
. From 
 and 
, we can write 
 for distinct off-diagonal cells 
. Since 
A dominates 
 cells that are collinear, it follows that 
 are collinear and they must be collinear to a cell in 
. Without loss of generality, we can write 
 as 
 for all 
 and 
 so that 
. Let 
 be the vector whose all nonzero entries are located in the first and 
th positions for all 
. Then 
 and we see that 
 is an optimal decomposition of 
A. Therefore, we have
        
        and hence 
.    □
 Using Lemma 3 and Theorem 2, we can obtain the specific structure of elements in  with  as following:
Proposition 3. Let . Then for , we have  and  if and only if .
 Proof.  Assume that  and . Then A dominates two cells which are collinear. Thus by Lemma 3, . If , then  for all . Then, by Corollary 1,  and thus , a contradiction. Hence we must have .
Conversely, assume 
. Then 
 by Theorem 2. Since 
 is an integer, we have 
. Consider 
 with 
 and 
. Then 
 and thus 
. Hence we can write 
 for distinct cells 
. If 
, then there exist three cells in 
 that are collinear, or there exist four cells in 
 such that two cells are collinear and the others are collinear. For the first case, we can assume that 
 are collinear. Then by Lemmas 1 and 2,
        
        which is impossible. For the second case, we can assume that 
 are collinear and 
 are collinear. Again by Lemma 2, 
 which is impossible. Hence we must have 
.    □
 The sum of four cells in  is called a rectangle if the sum is of the form . Equivalently,  is a rectangle if and only if there exist permutation matrices  and , such that . This shows that  for every rectangle .
Proposition 4. For , let  with  and . Then  if and only if A dominates a rectangle.
 Proof.  Assume 
 and let 
 be an optimal decomposition of 
A. Then we have 
, equivalently, 
. Since 
 and 
 are nonzero, we have 
 and 
 for all 
. Hence the above equality is impossible if 
. If 
, then 
 for all 
, equivalently, 
 is a cell for all 
. This implies 
, a contradiction to the fact that 
. Hence 
. Since 
 and 
, it follows that 
 and so 
. Thus we must have 
 so that
        
For this equality, three cases arise:
        
- (1)
- ,  (or , ) for some i, and  for all ; 
- (2)
-  for some distinct , and  for all ; 
- (3)
-  for some i, and  for all . 
If (1) holds, then  is a sum of three cells, and  is a cell for all . These imply that  dominates at most  cells, a contradiction to the fact that . If (2) holds, then both  and  are a sum of two cells, and  is a cell for all . These imply that  dominates at most  cells, a contradiction to the fact that . Consequently, (3) must hold. In this case, there exist nonzero entries  of , and nonzero entries  of . These imply that A dominates the rectangle .
Conversely, assume that A dominates a rectangle. Since , we can assume . Thus, the rectangle must be of the form  for distinct . Without loss of generality, we assume that  and  so that . Thus we have  and so .    □
   3. Linear Preservers of Arctic Ranks 1 and 
with 
An operator T on  is a map from  into . For an operator T on , we say that it is linear if  and  for all A and B. In particular, every linear operator on  is completely determined by its behavior on the set of cells in .
Let T be an operator on . Then we say that:
- (1)
- T preserves arctic rank k if  for all ; 
- (2)
- T preserves arctic rank if  for all . 
If T is an operator on , then it is called a -operator if there exist permutation matrices  and , such that  for all X, or  and  for all X. It is obvious that every -operator on  is linear.
Theorem 3. If T is a -operator on , then T preserves arctic rank.
 Proof.  Since T is a -operator, if , then  is of the form either  or  for some permutation matrices  and . By Lemma 1,  and  since . These two results imply . Thus, T preserves arctic rank.    □
 Proposition 5. For , let  with  and . If A does not dominate a rectangle, then .
 Proof.  Since , we have  by Theorem 2. Suppose that A does not dominate a rectangle. Then by Proposition 4,  and thus .
If 
, then there is an optimal decomposition 
 of 
A such that 
, equivalently, 
. By a similar argument in the proof of Proposition 4, we see that 
. That is, we have
        
Then, there must be an index  such that  and  for all . These mean that  is a sum of two cells, and  is a cell for all . That is,  dominates at most  cells, a contradiction to the fact that . Therefore, .    □
 Example 1. Consider a linear operator T on  defined byfor every . Since T maps cells to cells, T preserves arctic rank 1. Note that all members of  are ,  and . All of them are mapped to . Thus T preserves arctic rank 2. For the member , we have . Therefore T does not preserve arctic rank  and hence T is not a -operator by Theorem 3.  Example 2. Consider a liner operator T on  defined byfor every . Obviously, T preserves arctic rank 1. Clearly,  and . Now, we can easily check that if A is a member of  different from , then  and . There are 18 such members, and each of them is mapped to a member in . Thus, T preserves arctic rank . For the member , we have . Therefore, T does not preserve arctic rank 2, and hence T is not a -operator by Theorem 3.  Henceforth, in order to study linear preservers of arctic ranks of Boolean matrices, we need the condition, . For a linear operator T on , it is clear that if there exists a member  such that , then T does not preserve arctic rank k.
Let . That is,  is the set of all cells in .
Lemma 4. For , let T be a linear operator on  which preserves arctic ranks 1 and  with . Then T is bijective on .
 Proof.  Since T preserves arctic rank 1, it shows that T maps cells to cells. Now we will show that T is bijective on , equivalently T is injective on  since  is finite.
Case 1:  for some . That is, T preserves arctic rank h. Suppose that T is not injective on . Then  for distinct cells  and . If these cells are not collinear, then there is a member  such that . Then, since , we have , a contradiction. Therefore  and  must be collinear. Without loss of generality we may assume that  (or ). Since  and T preserves arctic rank h, by Proposition 2, we can write  for all .  since .
Let  so that T preserves arctic rank 2. Since , it follows from  that  is either  or . From  and its image, we have  with . If we combine these two results, we must have . Then, for , we have , a contradiction to the fact that T preserves arctic rank 2. Hence, T is injective on  for the case of .
Now, let 
. Consider 
. Then by Lemma 1 and Proposition 1, 
. Hence 
 and thus 
. Since 
 for all 
 and 
, by Proposition 4, we have either 
 or there exist distinct 
 such that 
. However, if we consider the member
        
        by Lemma 3, it follows from 
 that 
 with 
. Combining these results, we see that 
 and hence 
. By a parallel argument, we have 
. If 
, then 
 by Lemma 3, while 
, a contradiction to the fact that 
T preserves arctic rank 3. If 
, then 
 and 
 are not collinear, and hence we can choose a member 
 with 
. Then 
 since 
. This contradicts that 
T preserves arctic rank 
h. Thus 
T is injective on 
.
Case 2:  for some . That is, T preserves arctic rank . If T is not injective on , then  for distinct cells E and F. Now, we can take  dominating , such that  and . Then  by Proposition 3, while  since . This contradicts that T preserves arctic rank . Thus T is injective on .    □
 Proposition 6. Let A be a Boolean matrix in , where .
- (1)
- If , then  if and only if there exist permutation matrices  and , such that  or  
- (2)
- If , then  if and only if there exist permutation matrices  and , such that  or , where E is a cell which is not dominated by  or . 
 Proof.  (1) Suppose . Let  be an optimal decomposition of A so that , equivalently, . Since  and  are nonzero vectors,  is impossible. If , then . This implies that the only one in  has the value 2, and the others have the value 1. Then  which is impossible. Hence we must have  so that . By an easy examination, we have ( and ) or ( and ). This is a result which we desire. The converse is obvious since .
(2) The proof is similar to (1).    □
 A Boolean matrix L in  is called a line matrix if  for some , or  for some ;  is an i-th row matrix and  is a j-th column matrix.
Lemma 5. For  and , let T be a linear operator on  which preserves arctic ranks 1 and h. Then the image of each line matrix is dominated by a line matrix.
 Proof.  Since T preserves arctic ranks 1 and h, by Lemma 4, T is bijective on . Suppose that the image of some line matrix is not dominated by a line matrix. Then by the bijection of T on , there exist two cells E and F which are not collinear such that  and  are collinear. Now, we can choose a member  with  such that A dominates E and F. Then we have  and  since  and  are collinear. Hence by Proposition 2, , a contradiction to the fact that T preserves arctic rank h. Therefore, the result follows.    □
 Lemma 6. For  and , let T be a linear operator on  which preserves arctic ranks 1 and . Then the image of each line matrix is dominated by a line matrix.
 Proof.  Since T preserves arctic ranks 1 and , by Lemma 4, T is bijective on . Suppose that the image of some line matrix is not dominated by a line matrix. Then there exist two cells E and F such that  and . Without loss of generality, we assume that  and  so that  and  are not collinear.
First, let 
 so that 
T preserves arctic rank 
. If we consider the member 
 and its image, by Proposition 6(1), we can write 
 (or 
). Since 
 and 
 are not collinear, without loss of generality, we assume that 
, and 
. This implies that there are two cells 
 and 
 in 
 such that 
. By an easy examination, we see that if 
X is a sum of 
 and two cells in 
, then 
. It follows that 
, while
        
        by Lemma 1, a contradiction to the fact that 
T preserves arctic rank 
. Hence, the result holds for the case of 
.
Next, if , then T preserves arctic rank . If we use Proposition 6(2), by a similar argument as the case of , we have the same result for the case of .
Now, let . In this case,  and T preserves arctic rank . Since  and T is bijective on , by Proposition 3,  with  and . Since , we can find  which is the sum of h cells in  such that B dominates , and . Let . We note that ,  and C dominates 4 cells that are collinear. Thus, by Lemma 3,  and thus . We have  and  since  and . Now, if , then  by Theorem 2, and if , then  or  by Propositions 4 and 5. These contradict that T preserves arctic rank . Hence the result follows.    □
 Now, we are ready to prove the main result:
Theorem 4. Let T be a linear operator on  with . Then the followings are equivalent:
- (1)
- T preserves arctic rank; 
- (2)
- T preserves arctic ranks 1 and  with ; 
- (3)
- T is a -operator. 
 Proof.  It is obvious that (1) implies (2). (3) implies (1) by Theorem 3. To show that (2) implies (3), suppose that T preserves arctic ranks 1 and  with . By Lemma 4, T is bijective on , and by Lemmas 5 and 6, the image of each line matrix is dominated by a line matrix. Thus we have either  for some i, or  for some j.
Case 1: 
. In this case, 
 since 
T is bijective on 
. Suppose that there exist a column matrix 
 such that 
 is dominated by a row matrix, say that 
. Since 
 and 
, we must have 
, and hence 
. This contradicts that 
T is bijective on 
. Thus, the image of each column matrix is dominated by a column matrix. Furthermore, by the bijection of 
T on 
, we conclude that the image of each column matrix is a column matrix. By a similar argument, the image of each row matrix is a row matrix. Thus, there exist permutations 
 and 
 of 
 and 
, respectively such that 
 and 
 for all 
i and 
j. Let 
P and 
Q be permutation matrices corresponding to 
 and 
, respectively. Then we have
        
        for all cells 
. Let 
 be any Boolean matrix in 
. Then by the action of 
T on the cells, we have 
. Therefore 
T is a 
-operator.
Case 2: . In this case, we must have  and  by the bijection of T on . By a parallel argument in Case 1, the image of each row matrix is a column matrix and the image of each column matrix is a row matrix. Furthermore, there exist permutation matrices P and Q in  such that  for all X. Hence T is a -operator.    □
   4. Linear Preservers of Arctic Ranks 1 and 
Notice that the possible values of arctic ranks of nonzero Boolean matrices are . In previous section, we have characterized linear operator on  that preserve arctic ranks 1 and  with  except for .
In this section, we investigate the characterization of linear operators on  that preserve arctic ranks 1 and .
We remaind that  is the set of a sum of two cells in  that are collinear. For example, all members of  in  are , ,  and .
Example 3. Define a linear operator T on  byfor every . An easy observation implies that T preserves arctic ranks 1 and , and the image of T is not a single line matrix.  Example 4. Consider a linear operator T on  defined byfor every . By a simple examination, we see that T preserves arctic ranks 1 and . Furthermore, the image of T is the first row matrix.  Let T be an operator on . Then we say that T is line-injective in a line matrix L if  and  for all distinct cells . It is obvious that if T is a -operator, then T is line-injective in any line matrix. We see that the operators, in the above two examples, are line-injective in any line matrix.
We can extend Example 4 as following: For a fixed value 
, define a linear operator 
T on 
 by
      
      for all cell 
, where 
 (mod 
n) and 
. We can check that 
T preserves arctic ranks 1 and 
. 
T is line-injective in any line matrix, and the image of 
T is the 
k-th row matrix.
Now, look again at the linear operator T on , in Example 3, which preserves arctic ranks 1 and . We see that T is not a -operator, and the image of T is not a single line matrix. For a linear operator T on , if  and , then the following holds:
Theorem 5. For  and , let T be a linear operator on  which preserves arctic ranks 1 and . Then T is line-injective in any line matrix. Furthermore, either T is a -operator or the image of T is a single line matrix.
 Proof.  Since T preserves arctic rank 1, it shows that T maps cells to cells. Let  be an arbitrary line matrix. If , then there exist two cells  such that . Then , while , a contradiction to the fact that T preserves arctic rank . Thus we have . If  for distinct cells , then , while , a contradiction. Hence  for all distinct cells . Consequently, T is line-injective in L. Since L is an arbitrary line matrix, T is line-injective in any line matrix.
That is, we have established that the image of each line matrix is dominated by a line matrix. Thus, if T is not a -operator, then there exist two-line matrices such that their images are dominated by the same line matrix. There are three possibilities:
Case 1: Suppose that the images of two row matrices are dominated by the same row matrix. Without loss of generality, we may assume that  and . In this case, we must have  since T is line-injective in . Now we show that , equivalently,  for all  and . If  or 2, the result is obvious. Thus we assume . Since  and T is line-injective in , we can write  and . If we consider the member  and its image, we can write  for some , or  for some . If , then we see that , while . This contradicts that T preserves arctic rank , and hence . Since  is an arbitrary cell, we conclude that the image of T is a single row matrix.
If the images of two-row matrices are dominated by the same column matrix, then we must have  and, by a similar argument as above, the image of T is a single column matrix.
Case 2: Suppose that the images of two-column matrices are dominated by the same line matrix. Then by a parallel proof of Case 1, we see that the image of T is a single line matrix.
Case 3: Suppose that the images of a row matrix and a column matrix are dominated by the same row matrix. Without loss of generality, we may assume that  and . In this case, we have . Now, we show that , and then by Case 1, the image of T is a single row matrix. Since  and T is line-injective in , we can write ,  and . If we consider the member  and its image, we have  for some  or  for some . If , then by considering the member  and its image, we see that  since . Then  shows . Thus, we can write  for some . Now, for the member , we have . This contradicts that T preserves arctic rank , and thus we must have . Consequently, we have  since .
If the images of a row matrix and a column matrix are dominated by the same column matrix, then we have  and, by a similar argument as above, the image of T is a single column matrix.    □
 The converse of Theorem 5 is not true. For example, we define a linear operator 
T on 
 by
      
      for every 
. Then 
T is line-injective in any line matrix, and the image of 
T is the first row matrix. However, 
T preserves neither arctic rank 1 nor arctic rank 
.