Abstract
Let be a matrix function over the field of complex numbers, where are a family of matrices with variable entries. The purpose of this paper is to propose and investigate the relationships between certain linear matrix functions that regularly appear in matrix theory and its applications. We shall derive a series of meaningful, necessary, and sufficient conditions for the collections of values of two given matrix functions to be equal through the cogent use of some highly selective formulas and facts regarding ranks, ranges, and generalized inverses of block matrix operations. As applications, we discuss some concrete topics concerning the algebraic connections between general solutions of a given linear matrix equation and its reduced equations.
Keywords:
block matrix; general solution; generalized inverse; matrix equation; matrix expression; range; rank MSC:
15A09; 15A10; 15A24
1. Introduction
Throughout this paper, we adopt the following notation: denotes the collection of all matrices over the field of complex numbers; and denote the transpose and the conjugate transpose of , respectively; denotes the rank of a matrix ; denotes the range of a matrix ; denotes the identity matrix of order m; denotes a partitioned matrix consisting of two submatrices A and B; the Moore–Penrose generalized inverse of a matrix , denoted by , is defined as the unique matrix that satisfies the following four Penrose equations:
In addition, we denote by and the two orthogonal projectors (Hermitian idempotent matrices) induced from A. For more detailed information regarding the generalized inverses of matrices, we refer the reader to [1,2,3,4]. The Kronecker product of any two matrices and is defined as . The vectorization operator of a matrix is defined as , where denotes the i-th column of A, . A well-known property of the vec operator of a triple matrix product is known as l; see, e.g., [5,6].
Consider a matrix function
where are matrices of appropriate sizes with variable entries from the field of complex numbers, and Z denotes the matrix value of the matrix function corresponding to the k variable matrices. Further, we denote the collection of all possible values of the function corresponding to the k variable matrices by
and call it the domain of the matrix function. Given a matrix function as such, algebraists would like to know their algebraic properties and performances and then employ them when solving problems related to matrix functions in theoretical and computational mathematics.
Given such a matrix function, there are some fundamental questions that we may necessarily ask:
- (I)
- When is the matrix value of (2) unique with respect to all the variable matrices ?
- (II)
- What is the solvability condition of the matrix equation , and what is the general solution of when it is solvable?
- (III)
- Given two matrix functions and of the same size with the domains and , respectively, what are the necessary and sufficient conditions for the four statementsto hold, respectively?
Theoretically speaking, concrete matrix functions can be arbitrarily constructed according to various ordinary algebraic operations of matrices, while algebraists can propose or encounter numerous specified matrix functions when dealing with theoretical and computational problems in matrix analysis and its applications. In comparison, linear matrix functions (LMFs) are a class of simple forms of all matrix functions, and they can be routinely defined according to the additions and multiplications of matrices. Let us just mention here a typical example of LMFs:
where , , and are given, and are variable matrices; . Correspondingly, the domain of the LMF is denoted by
The LMF in (5) includes many simple and well-known matrix expressions of this kind with variable entries as its special cases, such as , , and , as well as various partially specified matrices, such as , , , where the symbol * denotes a unspecified sub-matrix (cf. [7,8,9,10,11]).
Nowadays, we have some powerful algebraic tools and techniques for characterizing relationships between different matrix functions and matrix equations. Among them are the simple but surprisingly effective matrix rank method, as well as the matrix range method and the block matrix representation method. The purpose of this paper is to propose and study some fundamental problems related to the domains of certain specified cases of (5) through the nicely organized employment of some known formulas and facts for ranges and ranks of matrices. As applications, we also discuss the connections among general solutions of some linear matrix equations and their reduced linear matrix equations. This paper is organized as follows. In Section 2, the authors give some preliminary results and facts about generalized inverses, rank formulas, and matrix equations. In Section 3, we first present some known and new results and facts about the relationships between two matrix sets generated from (5), and then we discuss the relationships between general solutions of some well-known linear matrix equations that occur in linear algebra and matrix analysis and their applications. Section 4 discusses the relationships between the general solutions of some basic linear matrix equations and their reduced equations.
2. Notation and Some Preliminary Results
As we know in linear algebra and matrix theory, a matrix is null if and only if its rank is zero. As a direct consequence of this elementary fact, we easily came to the conclusion that two given matrices A and B of the same size are equal if and only if . In view of this fact, we may figure out that if certain nontrivial algebraic formulas for calculating the rank of the difference are obtained, we can reasonably utilize them to describe essential links between the two matrices and, especially, to characterize the matrix quality in a convenient manner. A solid underpinning of this proposed method is that we are really able to determine or compute the rank of a matrix through various elementary operations of matrices and to obtain analytical formulas for expressing the ranks of matrices in many cases. Recall, in addition, that block matrices and matrix equations are two types of basic conceptual objects in linear algebra. Correspondingly, the matrix rank method (for short, MRM), the block matrix representation method (for short, BMRM), and the matrix equation method (for short, MEM) are three basic and traditional analytic tools and techniques that are extensively employed in matrix theory and its applications because they give algebraists the capacity to construct and analyze various complicated matrix expressions and matrix equalities in a subtle and computationally tractable way. On the other hand, it has been realized since the 1960s that generalized inverses of matrices can be adopted to derive numerous exact and analytical expansion formulas for calculating the ranks of block matrices. In the following, we present a series of seminal equalities and facts about the ranks of matrices and matrix equations.
Lemma 1
([12]). Let and Then,
In particular, the following results hold.
- (a)
- (b)
- (c)
Lemma 2
([13]). Let and denote Then,
In particular, the following three statements are equivalent:
- (a)
- (b)
- (c)
Lemma 3
([14]). Let
be a given linear matrix equation, where and are known matrices, and is an unknown matrix. Then, the following four statements are equivalent:
- (a)
- (b)
- (c)
- (d)
Lemma 4
([14]). Let
be a two-sided linear matrix equation, where and are given, and is an unknown matrix.
Then, the following four statements are equivalent:
- (a)
- (b)
- Both and
- (c)
- Both and
- (d)
Lemma 5
([15]). The linear matrix equation
is solvable for the two unknown matrices and of appropriate sizes if and only if
hold, or equivalently,
holds. In particular, (15) holds for all matrices and if and only if
Lemma 6
([16]). The linear matrix equation
is solvable the two unknown matrices and of appropriate sizes if and only if the following four matrix rank equalities
hold, or equivalently, the following four matrix equalities
hold, where and
Lemma 7
([17,18]). Equation (18) holds for all matrices and of appropriate sizes if and only if any one of the following four block matrix equalities
holds.
Lemma 8
([19]). The linear matrix equation
is solvable for the four unknown matrices and of appropriate sizes if and only if the following four matrix rank equalities hold:
3. Main Results
We start by presenting two groups of fundamental results and facts regarding the relationships between two matrix sets generated from the two simplest cases in (5).
Lemma 9
([20]). Assume that two LMFs and their domains are given by
where and are known matrices, and and are variable matrices. Then, we have the following results.
- (a)
- i.e., there exist and such that if and only if
- (b)
- if and only if .
- (c)
- if and only if .
Lemma 10
([20]). Assume that two LMFs and their domains are given by
where and are known matrices, and are variable matrices; Then, we have the following results.
- (a)
- if and only if the following four conditions hold:
- (b)
- if and only if any one of the following three conditions holds:
- (i)
- and
- (ii)
- and
- (iii)
- and
- (c)
- if and only if any one of the following five conditions holds:
- (i)
- and
- (ii)
- and
- (iii)
- and
- (iv)
- and
- (v)
- and
As an extension of these known facts, we proceed to derive the following results and facts about relationships between the domains of two general matrix functions, which we shall use in the latter part of the article.
Theorem 1.
Assume that two LMFs and their domains are given by
where and are known matrices. Then, we have the following results.
- (a)
- if and only if the following four conditions hold:
- (b)
- if and only if any one of the following four conditions holds:
- (c)
- if and only if any one of the following four groups of conditions holds:
- (d)
- if and only if both (b) and (c) hold.
Proof.
The condition is obviously equivalent to for some variable matrices , , , and . We rewrite this equation as
In this case, applying Lemma 8 to (44) leads to (a).
By (15)–(17) and (44), the condition is equivalent to the fact that
holds for all variable matrices and . Further, by Lemma 7, the matrix equality in (45) holds for all variable matrices and if and only if the following four matrix equalities,
hold, where
hold by Lemma 1(c). Substituting these four matrix rank equalities into (46)–(49) leads to the equivalences of (36)–(39) and (46)–(49), respectively.
Applying (19)–(21) to (44), we see that the condition holds if and only if any one of the following four equations,
holds for all matrices and , where and . By Lemma 7, the matrix equality in (50) holds for all matrices and if and only if any one of the following two conditions holds:
which are further equivalent to
by (7), (9), and Lemma 1(a) and (c), as is required for (40); the matrix equality in (51) holds for all matrices and if and only if any one of the following three conditions holds:
which are further equivalent to
by (7), (9), and Lemma 1(a) and (c), thus establishing (41); the matrix equality in (52) holds for all matrices and if and only if any one of the following three conditions holds:
These three equalities are further equivalent to
by (7), (9), and Lemma 1(a) and (c), as is required for (42); the matrix equality in (53) holds for all matrices and if and only if any one of the following four conditions holds:
which are further equivalent to
by (8), (9), and Lemma 1(b) and (c), thus establishing (43). □
It has been well known since Penrose [14] that general solutions of linear matrix equations can be derived and represented by certain algebraic linear matrix expressions that are composed of the given matrices in the matrix equations and their generalized inverses. In view of this basic fact, we are able to employ the preceding formulas, results, and facts to describe and characterize various possible relationships between solutions of different linear matrix equations.
We remark that there exist many types of linear matrix equations for which we can represent their general solutions in certain explicit linear matrix functions, as given in (54). In this section, we present a selection of results and facts on the relationships between certain linear transformations of the solutions of some fundamental linear matrix equations.
Theorem 2.
Assume that the two linear matrix equations
are solvable for and respectively, where and are given, We also denote
where and are given; Then, we have the following results.
- (a)
- if and only if
- (b)
- if and only if
- (c)
- if and only if
Proof.
By Lemma 3, the general solutions of the two linear matrix equations in (54) can be expressed as
where and are arbitrary matrices. Then, the two sets in (55) can be represented as
Applying Lemma 9(a) to (57), we obtain that if and only if
where by (8),
Substitution of (59) and (60) into (58) yields
thus establishing (a).
Corollary 1.
Assume that and in (54) are solvable for and respectively, and denote
Then, we have the following results.
- (a)
- (b)
- if and only if i.e., and
- (c)
- if and only if i.e., and
Corollary 2.
Let and be given, and suppose that is solvable for In addition, we denote
where and are two given matrices. Then, the following results hold.
- (a)
- always holds.
- (b)
- if and only if
Corollary 3.
Let and be given, and suppose that is solvable for In addition, we denote
where Then, we have the following results.
- (a)
- always holds.
- (b)
- if and only if
4. Relationships between Solutions of Some Linear Matrix Equations and Their Reduced Equations
Let us partition the matrix equation in (11) as
where , , and are unknown matrices with and , . In this case, pre-multiplying (66) by yields the following reduced linear matrix equations:
where . Now, assume that (67) is solvable for X. Then, the equations in (67) are solvable for . Correspondingly, we denote
for the matrix sets composed of the partial solutions of (66) and (67), respectively, , and denote
In this section, we first discuss the relationships between and in (68), , as well as the two sets in (69).
Theorem 3.
Proof.
Theorem 4.
Assume that the matrix equation in (66) is solvable for and let and be as given in (67) and (69), respectively, for Then, we have the following results.
- (a)
- always holds.
- (b)
- The following four statements are equivalent:
- (i)
- (ii)
- (iii)
- (iv)
Proof.
By Lemma 3, the general solutions of (67) are given by
where are arbitrary, . Substitution of (12) and (71) into (69) yields
Applying Lemma 9(b) to (72) and (73), we see that if and only if
where by (8), we obtain
and
Both (75) and (76) mean that (74) is an identity, thus establishing (a). Substitution of (71) into (66) gives
It is obvious that holds if and only if the matrix equation in (77) holds for all , which, by Lemma 3, is equivalent to
where by (8), we obtain
Substituting it into (78), we see that (78) is equivalent to the rank equality . Combining this fact with (a) leads to the equivalence of (i) and (ii) in Result (b). The equivalences of (ii), (iii), and (iv) in Result (b) follow from Lemma 2. □
Equation (18) is well known in matrix theory and its applications, while the solvability condition and the general solution of this equation were precisely established by implementing certain calculations of the ranks, ranges, and generalized inverses of the given matrices in this matrix equation; see, e.g., [15,16,18,21,22] and the relevant literature quoted there.
It is easy to see that we can construct from (18) some small or transformed linear matrix equations. For instance, pre- and post-multiplying (18) by and , respectively, yields the following four reduced matrix equations:
respectively. Each of (79)–(82) is consistent as well if the matrix equation in (18) is consistent. Concerning the relationships among the solutions of (18) and (79)–(82), we have the following results.
Theorem 5.
Assume that the matrix equation in (18) is solvable for and and we denote by
the collections of all pairs of solutions of (18) and (79)–(82), respectively. Then, we have the following results.
- (a)
- always hold;
- (b)
- if and only if or
- (c)
- if and only if or or and
- (d)
- if and only if or or and
- (e)
- if and only if or
Proof.
Result (a) follows directly from (79)–(82). By Lemma 4, the general solutions of (79)–(82) are given by
respectively, where and are arbitrary matrices; . Substitution of (88)–(91) into (18) gives the following four matrix equations:
respectively. By Lemma 7, the equality in (92) holds for all and if and only if any one of the following four equalities holds:
In addition, it is easy to verify that the ranks of the left-hand sides of (96)–(99) are given by
Combining (96)–(99) with (101)–(104) leads to the equivalence in (b).
By Lemma 7, the equality in (93) holds for all and if and only if any one of the following four equalities holds:
In addition, it is easy to verify that the ranks of the left-hand sides of (105)–(108) are given by
Combining (105)–(108) with (109)–(112) leads to the equivalence in (c). Results (d) and (e) can be established with a similar approach. □
Theorem 6.
Assume that the matrix equation in (18) is solvable and and let
Then, we have the following results.
- (a)
- The matrix set equalities always hold;
- (b)
- always holds.
- (c)
- if and only if
Proof.
By the vec operation of matrix, (18) can be equivalently expressed as
which is a special case of (66). Pre-multiplying (119) with , , yields the following two reduced linear matrix equations:
Now, we denote
Then, we obtain from Theorem 3 that
always hold. On the other hand, it is easy to verify that
and
hold. Thus, the two equations in (120) denoted by the matrix vectorization operations are equivalent to
respectively. Hence, the two set equalities in (124) are equivalent to the set equalities in (a). Results (b) and (c) follow from applying Theorem 4 to (119). □
5. Conclusions
In the preceding sections, we described and studied some relationships between two different linear matrix functions and presented a series of clear explanations and solutions to some concrete problems in this subject area. The whole work covers some principal cases that are often encountered in the theory of matrix functions and their applications, while the results and facts obtained deeply reveal the inherent natures and properties of some basic linear matrix functions and their connections. Notice that the derivations of the main results and facts are based on various precise algebraic calculations of the ranks of certain block matrices related to the matrix equalities and matrix set inclusions, which substantially avoid certain complicated matrix operations occurring in matrix expressions and matrix equalities. Hence, they clearly demonstrate that the matrix rank method and the block matrix representation method are useful and effective for solving various matrix equality and matrix set inclusion problems. Actually, the two fundamental algebraic methods have been recognized as reliable and efficient tools and techniques in the descriptions and investigations of many matrix problems in theoretical and computational mathematics.
Finally, the two authors remark that the research results and facts in this article offer certain deep and valuable insights into various intrinsic links among different matrix functions that could not be seen previously, and therefore, we believe that this study will have a profound impact on the explorations of the connections among domains of general matrix functions under various specified assumptions.
Author Contributions
Methodology, Y.T.; Validation, R.Y.; Formal analysis, Y.T. and R.Y.; Investigation, Y.T. and R.Y.; Writing—original draft, Y.T.; writing—review and editing, Y.T. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Acknowledgments
The two authors would like to thank the four referees for their helpful comments and suggestions on an earlier version of this article.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Ben-Israel, A.; Greville, T.N.E. Generalized Inverses: Theory and Applications, 2nd ed.; Springer: New York, NY, USA, 2003. [Google Scholar]
- Bernstein, D.S. Scalar, Vector, and Matrix Mathematics: Theory, Facts, and Formulas–Revised and Expanded Edition; Princeton University Press: Princeton, NJ, USA; Oxford, UK, 2018. [Google Scholar]
- Campbell, S.L.; Meyer, C.D. Generalized Inverses of Linear Transformations; SIAM: Philadelphia, PA, USA, 2009. [Google Scholar]
- Rao, C.R.; Mitra, S.K. Generalized Inverse of Matrices and Its Applications; Wiley: New York, NY, USA, 1971. [Google Scholar]
- Horn, R.; Johnson, C.R. Topics in Matrix Analysis; Cambridge University Press: Cambridge, UK, 1991. [Google Scholar]
- Steeb, W.-H. Matrix Calculus and Kronecker Product with Applications and C++ Programs; World Scientific Publishing: Singapore, 1997. [Google Scholar]
- Bakonyi, M.; Woerdeman, H.J. Matrix Completions, Moments, and Sums of Hermitian Squares; Princeton University Press: Princeton, NJ, USA, 2011. [Google Scholar]
- Dancis, J. Choosing the inertias for completions of certain partially specified matrices. SIAM J. Matrix Anal. Appl. 1993, 14, 813–829. [Google Scholar] [CrossRef]
- Gohberg, I.; Kaashoek, M.A.; Schagen, F.V. Partially Specified Matrices and Operators: Classification, Completion, Applications; Birkhäuser: Basel, Switzerland, 1995. [Google Scholar]
- Jordán, C.; Torregrosa, J.R.; Urbano, A. On the Jordan form of completions of partial upper triangular matrices. Linear Algebra Appl. 1997, 254, 241–250. [Google Scholar] [CrossRef]
- Krupnik, M. Geometric multiplicities of completions of partial triangular matrices. Linear Algebra Appl. 1995, 220, 215–227. [Google Scholar] [CrossRef]
- Marsaglia, G.; Styan, G.P.H. Equalities and inequalities for ranks of matrices. Linear Multilinear Algebra 1974, 2, 269–292. [Google Scholar] [CrossRef]
- Tian, Y. Formulas for calculating the dimensions of the sums and the intersections of a family of linear subspaces with applications. Contrib. Algebra Geom. 2019, 60, 471–485. [Google Scholar] [CrossRef]
- Penrose, R. A generalized inverse for matrices. In Mathematical Proceedings of the Cambridge Philosophical Society; Cambridge University Press: Cambridge, UK, 1955; Volume 51, pp. 406–413. [Google Scholar]
- Baksalary, J.K.; Kala, R. The matrix equation AXB + CYD = E. Linear Algebra Appl. 1980, 30, 141–147. [Google Scholar] [CrossRef]
- Özgüler, A.B. The matrix equation AXB + CYD = E over a principal ideal domain. SIAM J. Matrix Anal. Appl. 1991, 12, 581–591. [Google Scholar] [CrossRef]
- Jiang, B.; Tian, Y. Necessary and sufficient conditions for nonlinear matrix identities to always hold. Aequ. Math. 2019, 93, 587–600. [Google Scholar] [CrossRef]
- Tian, Y. Upper and lower bounds for ranks of matrix expressions using generalized inverses. Linear Algebra Appl. 2002, 355, 187–214. [Google Scholar] [CrossRef]
- Tian, Y. Solvability of two linear matrix equations. Linear Multilinear Algebra 2000, 48, 123–147. [Google Scholar] [CrossRef]
- Tian, Y. Relations between matrix sets generated from linear matrix expressions and their applications. Comput. Math. Appl. 2011, 61, 1493–1501. [Google Scholar] [CrossRef]
- Tian, Y. Ranks and independence of solutions of the matrix equation AXB + CYD = M. Acta Math. Univ. Comen. 2006, 75, 75–84. [Google Scholar]
- Xu, G.; Wei, M.; Zheng, D. On solution of matrix equation AXB + CYD = F. Linear Algebra Appl. 1998, 279, 93–109. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).