1. Introduction
Throughout this article, let
denote the collection of all
matrices over the field of complex numbers;
denote the conjugate transpose of
;
denote the rank of
A, i.e., the maximum order of the invertible submatrix of
A;
and
denote the range and the null space of a matrix
, respectively;
denote the identity matrix of order
m;
denote a columnwise partitioned matrix consisting of the two submatrices
A and
B. For an
, the Moore–Penrose generalized inverse of
A is denoted as
and is defined to be the unique matrix
satisfying the four Penrose equations
The Moore–Penrose inverse of a matrix
A was specially studied and recognized because
,
,
, and
are four orthogonal projectors onto the ranges and kernels of
A and
, respectively, so that we can utilize it to denote and optimize a number of algebraic properties and performances of many matrix computations. Additionally, it can be used to represent other generalized inverses by means of certain algebraic operations of
A and
. In brief, we let
,
, and
denote the three orthogonal projectors induced from
A which can help in briefly denoting calculation processes related to generalized inverses of matrices. For more detailed information about generalized inverses of matrices, we refer to [
1,
2,
3,
4,
5].
In this paper, we consider the following linear matrix equation
where
and
are given matrices. As we know, the subject of matrix equations today is still a vital and active part of matrix theory, while (
1) is one of the famous and classical algebraic equations with single unknown matrix in mathematics and applications. An equation as such was first proposed and studied by Sir Penrose in his seminal paper [
6], and he obtained a group of exact and analytical results concerning the solvability condition and the general solution of the equation using the ranks, the ranges, and the generalized inverses of the given matrices in the equation. These results were recognized as the fundamental and classic materials in the research field of various linear matrix equations in the 1950s, and extensively prompted many deep-going investigations of linear matrix equations from theoretical and computational points of view in the past several decades; see [
6,
7,
8,
9,
10,
11,
12,
13,
14] for several earlier and recent papers regarding this subject.
Below, we describe our motivation in the study of (
1). We first partition the two coefficient matrices
A,
B, and the unknown matrix
X into the following forms
where
,
,
. Correspondingly, (
1) can be represented in the following partitioned form
Based on the partitioned representation in (
3), we are able to propose and approach the unknown submatrices in (
1) and their performances, separately, such as the nonsingularity and nullity of the submatrices, the independence of the submatrices in the general solution of (
1), etc. In fact, many explicit results and facts about the expressions and properties of the submatrices
in (
3) were established in the literature, in particular, a family of exact and analytical formulas for calculating the maximum and minimum ranks of
were given in [
11] for
.
In order to deeply reveal the algebraic properties of the four unknown matrices in (
3), we pre- and post-multiply both sides of (
3) with
and
respectively and note that
and
for
to yield the following four small linear matrix equations (transformed equations)
Apparently, each of the four matrix equations involves only one of the four unknown submatrices in (
2), and thus the four equations of this kind are often called reduced equations of (
3), constructed with matrix partition and transformation methods.
The purpose of this article is to provide a meaningful study on intrinsic connections between the general solutions of the full equation in (
3) and its four reduced equations in (
4)–(7). The rest of the article is organized as follows. In
Section 2, we introduce some known results regarding matrix rank formulas and matrix equations that we shall use in the main contents. In
Section 3, we first divide (
1) into a partitioned form, and present the general solutions of four reduced matrix equations associated with (
1). We then discuss the relationships between the general solutions of (
1) and the reduced equations using the matrix analysis methods and techniques mentioned above. As an application, we discuss the relationships among generalized inverses of a block matrix and its four submatrices.
2. Preliminary Results
As we know, block matrices, ranks of matrices, and matrix equations are basic concepts and objects in linear algebra, while the block matrix representation method (for short, BMRM), the matrix rank method (for short, MRM), and the matrix equation method (for short, MEM) are three fundamental and traditional analytic methods that are widely used in matrix theory and applications because they give us the ability to construct and analyze various complicated matrix expressions and matrix equalities in a subtle and computationally tractable way. On the other hand, it has been realized since the 1960s that generalized inverses of matrices can be employed to construct many precise and analytical expansion formulas for calculating the ranks of various partitioned matrices. These matrix rank formulas can be used to deal with a wide variety of theoretical and computational problems in matrix theory and applications; see the seminal paper [
15].
In this section, we summarize some relevant formulas and facts on ranks and ranges of matrices, as well as matrix equations and matrix functions. We first introduce a group of well-known useful formulas for calculating ranks of partitioned matrices and their consequences, which are easily understandable in the discipline of linear algebra and generalized inverses of matrices, and can be used as fundamental tools in dealing with various concrete problems in relation to ranks of matrices.
Lemma 1 ([
15]).
Let and Then,In particular, the following results hold.
- (a)
- (b)
In the following, we present a group of known results that are represented in terms of generalized inverses and ranks of matrices on the solvability condition and the general solution of (
1).
Lemma 2 ([
6]).
Let be as given in (1). Then, the following four statements are equivalent:- (i)
Equation (1) is solvable for - (ii)
and
- (iii)
and
- (iv)
In this case, the general solution of (1) can be written in the parametric formwhere are two arbitrary matrices. Lemma 3 ([
16,
17,
18]).
Let be a given linear matrix equation, where and are known matrices. Then, the following results hold.- (a)
Equation (11) is solvable for and if and only if the following four conditions holdwhere and - (b)
Equation (11) holds for all matrices and if and only if any one of the following four block matrix equalities holds - (c)
Under the assumptions that and (11) holds for all matrices and if and only if any one of the following three block matrix equalities holds:
Lemma 4 ([
19]).
Let and be two linear matrix expressions, where and are given, and and are four variable matrices. Additionally, let be the domains of two linear matrix functions, and Then, the following results hold.- (a)
if and only if
- (b)
if and only if any one of the three conditions(i)(ii) and(iii) holds.
- (c)
if and only if any one of the three conditions (i) (ii) (iii) and any one of the three conditions(iv) (v) (vi) hold.
In the following, we derive the solvability conditions and the general solutions of the four reduced equations in (
4)–(7).
Lemma 5. Let be as given in (1), and let its four reduced linear matrix equations be given as in (4)–(7). Then, the following results hold. - (a)
The following five statements are equivalent:
- (i)
Equation (4) is solvable for - (ii)
and
- (iii)
and
- (iv)
and
- (v)
In this case, the general solution of (4) can be written in the following parametric formwhere are two arbitrary matrices. - (b)
The following five statements are equivalent:
- (i)
Equation (5) is solvable for
- (ii)
and
- (iii)
and
- (iv)
and
- (v)
In this case, the general solution of (5) can be written in the following parametric formwhere are two arbitrary matrices. - (c)
The following five statements are equivalent:
- (i)
Equation (6) is solvable for
- (ii)
and
- (iii)
and
- (iv)
and
- (v)
In this case, the general solution of (6) can be written in the following parametric formwhere are two arbitrary matrices. - (d)
The following five statements are equivalent:
- (i)
Equation (7) is solvable for
- (ii)
and
- (iii)
and
- (iv)
and
- (v)
In this case, the general solution of (7) can be written in the following parametric formwhere are two arbitrary matrices.
Proof. Note that (
4)–(7) are special cases of the linear matrix equation in (
1) with different coefficient matrices and constant matrices. In this situation, applying the results in Lemma 2, we obtain the four groups of concrete facts and formulas in (a)–(d) on the solvability conditions and the general solutions of (
4)–(7). □
3. Main Results
It is easy to see that if (
1) is solvable for
X, then the four equations in (
4)–(7) are all solvable for
as well. In this situation, we denote the collections of all submatrices in the solutions of (
1) and (
4)–(7) as
and denote the collections of all solutions of (
1) and the block matrices composed of all solutions of (
4)–(7) as
In view of this notation, we obtain the following results and facts on the relationships between the solution sets in (
16)–(20), as well as (
21) and (22).
Theorem 1. Assume that the matrix equation in (1) is solvable for and let and be as given in (16)–(20), respectively. Then, we have the following results. - (a)
always hold,
- (b)
always holds.
- (c)
if and only if or or and
Proof. The matrix set inclusions
follow directly from the constructions of
,
We next show
holds for
By (
1), the general expression of the submatrix
in (
3) can be written as
where
and
and
and
are two arbitrary matrices. Thus,
is equivalent to
By Lemma 4(b), the matrix set inclusion in (
23) holds if and only if any one of the following three conditions
holds. By (
8)–(10) and elementary block matrix operations, we obtain the following rank equalities
Combining (
27)–(29), we see that (26) is an identity for the ranks of matrices. Hence, (
24) and (25) are of no use anymore in the description of (
23), and thus (
27) always holds. The set inclusions
,
, and
can be shown by a similar way. So that the four matrix set equalities in (a) hold. Result (b) is obvious from the construction of
.
Substituting (
12)–(
15) into (
3) gives the following matrix equation
where
Thus,
if and only if (
30) holds for all
and
,
. By Lemma 3(c), we see that (
30) holds for all matrices
and
if and only if any one of the following three conditions
holds. In this situation, it is easy to verify that
and by (
8)–(10) and elementary block matrix operations that
Substituting (
34)–(
36) into (
31)–(33) and then simplifying yield the results in (c). □
One of the fundamental research topics in the domain of generalized inverses is to characterize relationships between generalized inverses of two matrices. As one example in this regard, we let
, and partition
A and its generalized inverse as
for
. Additionally, denote
and
Referring to Theorem 1, we obtain the following result.
Corollary 1. Let with and let and be as given in (37)–(43). Then, we have the following results. - (a)
always hold,
- (b)
always holds.
- (c)
if and only if and
Proof. Setting in Lemma 5 and Theorem 1 and then simplifying lead to the results in (a), (b), and (c). □
4. Concluding Remarks
We described and studied in the preceding sections a number of theoretical problems regarding the relationships between the full matrix equation in (
1) and its four reduced equations in (
2)–(
5) through the well-organized employment of various well-known or established formulas and facts in relation to ranks, ranges, and generalized inverses of matrices. The obtained results provide some profound insights into the construction of the general solution of (
1), and therefore, they can be viewed as certain original theoretical contributions with direct or potential value in applications. We believe that this study enables us to use the reduced equations instead of the full equation under some assumptions, and in turn, hope that they can improve computational efficiency in many issues in relation to matrix equations.
In addition to (
4)–(7), we pre- and post-multiply the matrix equation in (
3) by
,
,
, and
, respectively, and note that
and
for
and
to yield a group of new reduced linear matrix equations as follows
In comparison, the constructions of these four reduced equations are different from the four equations in (
4)–(
7). Hence, the four equations are not necessarily equivalent to these in (
4)–(
7). In this situation, it would be of interest to describe the relationships between the general solution of the matrix equation in (
3) and the general solutions of the four reduced equations in (
44)–(
47) by means of the powerful and effective matrix rank methodology.
Moreover, we are able to decompose the matrix equation in (
1) into the following form
and then to construct a family of reduced linear matrix equations with the multiplication transformation method. In this light, it is necessary to explore the relationships between the general solution of (
48) and the general solutions of the reduced linear matrix equations by means of the effective methods and techniques used in this article, but there will be a mass of complicated matrix calculations in the study of the matrix equation.
Finally, we remark that, prompted by the comparison problem described in this article, there are many similar topics that can be proposed and examined concerning the connections of solutions of other kinds of matrix equations and their reduced equations.