Abstract
This paper provides a complete matrix analysis on equivalence problems of estimation and inference results under a true multivariate linear model and its misspecified form with an augmentation part through the cogent use of various algebraic formulas and facts in matrix analysis. The coverage of this study includes the matrix derivations of the best linear unbiased estimators under the true and misspecified models, and the establishment of necessary and sufficient conditions for the different estimators to be equivalent under the model assumptions.
Keywords:
augmentation; BLUE; BLUP; Kronecker product; multivariate linear model; parameter matrix; rank; reduced model MSC:
15A10; 62J05; 62H12
1. Introduction
Throughout this article, we use to stand for the collection of all matrices with real numbers; , , and to stand for the transpose, the rank, and the range (column space) of a matrix , respectively; and to denote the identity matrix of order m. For two symmetric matrices and of the same size, they are said to satisfy the inequality in the Löwner partial ordering if is positive semi-definite. The Kronecker product of any two matrices and is defined to be . The vectorization operator of a matrix is defined to be . A well-known property on the vec operator of a triple matrix product is . The Moore–Penrose inverse of , denoted by , is defined to be the unique solution of the four matrix equations , , , and . We also denote by , , and the three orthogonal projectors induced from , respectively, which will help in briefly denoting calculation processes related to generalized inverses of matrices. We also adopt the notation when is a block matrix. Further information about the orthogonal projectors , , and with their applications in the linear statistical models can be found e.g., in [1,2,3].
In this paper, we consider the multivariate general linear model
where is an observable random matrix (a longitudinal data set), is a known model matrix of arbitrary rank (, is a matrix of fixed but unknown parameters, and denote the expectation vector and the dispersion matrix of the random error matrix , and are two known positive semi-definite matrices of arbitrary ranks, and is an arbitrary positive scaling factor. As we know, the multivariate general linear model (for short, MGLM) as such in (1) is a relative direct extension of the most welcome type of univariate general linear models.
The assumption in (1) is typical in the estimation and statistical inference under a multivariate linear regression framework. In statistical practice, we may meet with the situation where a true regression model is misspecified in some other forms due to certain unforeseeable reasons, and, therefore, we face with the task of comparing estimation and inference results and establishing certain links between them for the purpose of reasonably explaining and utilizing the misspecified regression model. In this light, one of the situations in relation to model misspecification problems appears by adding or deleting regressors in the model. As such an example, if taking (1) as a true model and misspecifically adding a multiple new regressor part in (1), we obtain an over-parameterized (over-fitted) form of as
where is a known matrix of arbitrary rank, and is a matrix of fixed but unknown parameters. Given (1) and (2), we proposed and studied some research problems in [4] on the equivalence of inference results that are obtained from the two competing MGLMs.
As we know, a commonly-used technique of handling a partitioned model is to multiply a certain annihilating matrix and to transform the the model equation into a reduced model form. As a new exploration regarding the equivalence problem, we introduce the commonly-used technique into the study of (1) and (2). To do so, we pre-multiply to the both sides of the model equation and noting that to obtain a reduced model as follows:
It should be pointed out that estimation and inference results that we derive from the triple models in (1)–(3) are not necessarily identical. Thus, it is a primary requirement to describe the links between the models and to propose and describe possible equalities among estimation and inference results under three MGLMs.
Before approaching comparison problems of estimation and inference results under the triple models in (1)–(3), we mention a well-known and effective method that was widely used in the investigation of multivariate general linear models. Recall that the Kronecker products and vec operations of matrices are popular tools in dealing with matrix operations in relation to multivariate general linear models. Referring to these operations, we can alternatively represent the triple models in (1)–(3) in the following three standard linear statistical models:
As a common fact in statistical analysis, we know that the first step in the inference of (1) is to estimate/predict certain functions of the unknown parameter matrices and . Based on this consideration, it is of great interest to identify their estimators and predictors simultaneously. For this purpose, we construct a general parametric matrix that involves both and as follows:
where and are and matrices, respectively. In this situation, we easily obtain that
hold. Under the assumptions in the triple models in (1)–(3), the corresponding predictions of in (7) are not necessarily identical, and we even use the same optimality criterion to derive the predictors of under the triple competing models, and therefore, this fact leads us to propose and study a series of research problems regarding the comparison and equivalence issues about inference results obtained from the triple models. In order to obtain general results and facts under (1)–(8), we do not require probability distributions of the random variables in the MGLMs although they are necessary for further discussing identification and test problems.
The purpose of this paper is to consider some concrete problems on the comparisons of the best linear unbiased estimators derived from (1) and those derived from (2) and (3). Historically, there were some previous investigations on establishing possible equalities of estimations of unknown parameter matrices in two competing linear models; see e.g., [5,6], while equalities of estimations of unknown parameter vectors under linear models with new regressors (augmentation by nuisance parameters) were approached in [7,8,9,10,11,12,13,14,15,16,17]. Particularly, the present two authors studied in [4] the equivalences of estimation and inference results under (1), (2), (4), and (5). As an updated work on this subject, we introduce the two reduced models in (3) and (6), and carry out a new analysis of the equivalences of estimators under (1)–(6).
The remaining of this paper is constructed as follows: In Section 2, we introduce some matrix analysis tools that can be used to characterize equalities that involve algebraic operations of matrices and their generalized inverses. In Section 3, the authors present a standard procedure to describe the predictability and estimability of parametric matrices under the triple models in (1)–(3), and then show how to establish analytical expressions for calculating best linear unbiased predictors and best linear unbiased estimators of parametric matrices under the triple models in (1)–(3). In Section 4, the authors discuss a group of problems on the equivalences of the BLUEs under (1)–(3).
2. Some Preliminaries
In order to establish the proposed mathematical equalities for predictors/estimators udder the triple models in (1)–(3), we need to use a series of basic rank equalities in the following two lemmas:
Lemma 1
([18]). Let and Then,
In particular, the following results hold:
- (a)
- (b)
- (c)
- =
- (d)
A special consequence of (14) is given below, which we shall use to simplify some complicated matrix expressions that involve generalized inverses in the sequel.
Lemma 2.
Assume that five matrices and of appropriate sizes satisfy the conditions and Then,
Hence,
Matrix rank formulas and their consequences, as displayed in Lemmas 1 and 2, now are highly recognized as useful techniques to construct and characterize various simple or complicated algebraic equalities for matrices and their operations. We refer the reader to [19] and the references therein on the matrix rank method in the investigations of various linear statistical models.
Lemma 3
([20]). The linear matrix equation is consistent if and only if or equivalently, In this case, the general solution of the equation can be written in the following parametric form: where is an arbitrary matrix.
Finally, we present the following established result on constrained quadratic matrix-valued function minimization problem.
Lemma 4
([21,22]). Let
where , , are given, is positive semi-definite, and the matrix equation is consistent. Then, there always exists a solution of such that
holds for all solutions of . In this case, the matrix satisfying (18) is determined by the following consistent matrix equation:
while the general expression of and the corresponding are given by
where and is arbitrary.
3. The Precise Theory of Predictability, Estimability, and BLUP/BLUE
In this section, we present a standard procedure of establishing predictability, estimability, and BLUP theory under an MGLM for the purpose of solving the comparison problems proposed in Section 1. Most of the materials given below are routine illustrations of various known conceptions, definitions, and fundamental results and facts on MGLMs; see e.g., [4].
Definition 1.
Let be as given in (7). Then,
Definition 2.
Let be as given in (7). Then,
- (a)
- (b)
Recall that the unbiasedness of given predictors/estimators and the lowest covariance matrices formulated in (23) and (24) are intrinsic requirements in statistic analysis of parametric regression models, which can be regarded as some special cases of mathematical optimization problems on constrained quadratic matrix-valued functions in the Löwner partial ordering. Note from (1) and (7) that and can be rewritten as
Hence, the expectations of and can be expressed as
The dispersion matrix of can be expressed as
where .
Concerning the predictability of in (7), we have the following known result.
Lemma 5
- (a)
- Φ is predictable by in (1).
- (b)
- (c)
Theorem 1.
Assume Φ in (7) is predictable. Then,
The matrix equation in (29), called the BLUP equation associated with is consistent as well, i.e.,
holds under Lemma 5(c), while the general expressions of and the corresponding can be written as
where is arbitrary. In particular,
where is arbitrary. Furthermore, the following results hold.
- (a)
- and
- (b)
- is unique if and only if
- (c)
- is unique if and only if holds with probability
- (d)
- The expectation, the dispersion matrices of and as well as the covariance matrix between and are unique, and are given by
- (e)
- and satisfy
- (f)
- holds for any matrix
Proof.
We obtain from (28) that the constrained minimization problem in (23) is equivalent to
which is further reduced to
Since is a non-null nonnegative definite matrix, we apply Lemma 4 to (43) to yield the matrix equation
as required for (29). Equations (32) and (33) follow directly from (31). Result (a) is well known on the matrix ; see, e.g., [1,2].
Note that
by (10). Combining this fact with (31) leads to (b). Setting the term in (31) leads to (c).
From (1) and (31),
establishing (36). Combining (8) and (35) yields (37). Substituting (31) into (28) and simplifying, we obtain
Concerning the BLUEs of the mean matrix , and the BLUP of the error matrix in (1), we have the following results.
Corollary 1.
In this case, the matrix equation
is consistent, and the following results hold.
- (a)
- The general expression of can be written aswithwhere is arbitrary.
- (b)
- The general expression of can be written aswithwhere is arbitrary.
- (c)
- The general expression of can be written aswithwhere is arbitrary.
- (d)
- and satisfy
The BLUEs under the over-parameterized model (2) can be formulated from the standard results on the BLUEs under the true model as follows.
Theorem 2.
In this case, the matrix equation
is consistent, and a BLUE of under (2) is
where is arbitrary. In particular,
where is arbitrary.
Proof.
Note that can be rewritten as under (2). Hence, is estimable under (2) if and only if by Lemma 5, or equivalently,
In addition, note that and by (10). Hence, (65) is further equivalent to , as required for (a). Let in (65). Then, we obtain from (65) that
Hence, (66) is equivalent to as required for (b). Equations (58)–(64) follow from the standard results on BLUEs in Corollary 1. □
The BLUEs of unknown parameter matrices under the transformed model in (3) can be formulated from the standard results on the BLUEs under the true model as follows.
4. The Equivalence Analysis of BLUEs under True and Misspecified Models
Concerning equalities between two linear statistics and , the following three possible situations should be addressed.
Definition 3.
Let be a random matrix.
- (a)
- The equality is said to hold definitely iff .
- (b)
- The equality is said to hold with probability 1 iff both and hold.
- (c)
- and are said to have the same expectation and covariance matrix iff both and hold.
Assume that is estimable under (2). Then, it can be seen from Corollary 1(i), Theorem 2(a), and Theorem 3(a) that is estimable under (1) and (3) as well. In this case, the BLUE of can be written as (46), while the BLUEs of under the misspecified models can be written as (59) and (68), respectively. The triple BLUEs are not necessarily the same, and thus it is natural to consider relations among the three BLUEs by Definition 3.
Theorem 4.
Assume that is estimable under (2) for and let and be as given in (46), (59), and (68), respectively. In addition, let
Then, the following eight statements are equivalent:
- (a)
- holds definitely;
- (b)
- holds definitely;
- (c)
- holds with probability
- (d)
- holds with probability
- (e)
- (f)
- (g)
- (h)
Proof.
Combining (45) and (58), we obtain a new equation for :
This matrix equation has a solution for if and only if
Simplifying both sides by and elementary matrix operations, we obtain
by (11) and , and
Hence, (a) is equivalent to the rank equality in (h).
By Definition 3(a), hold definitely iff the coefficient matrix in (68) satisfies (45)
which is further reduced to
Note that
Then, (73) is simplified by (10), (11), (13), and (14) as
Setting (74) equal to zero, we obtain that (b) is equivalent to the rank equality in (h).
By Definition 3(b), we see that holds with probability 1 if and only if
holds with probability 1 if and only if
The first equalities in (75) and (76) hold naturally. Thus, (c), (d), and (e) are equivalent. Furthermore, the two equalities in (e) are equivalent to
Notice that . Equations (77) and (78) are equivalent to
Applying (14) to the difference on the left-hand side of (79), and simplifying by elementary matrix operations, (11) and (13), we obtain
by (13). Similarly, we can show that
by (13). Setting the right-hand sides of (81) and (82) equal to zero, we obtain the equivalence of (e) and (h).
It follows from (48) and (61) that
Hence,
where a rank formula for the matrix difference in (84) is
see [13]. Substituting (85) into (84) yields
Similarly, we can obtain
Setting the right-hand side equal to zero, we obtain the equivalences of (f), (g), and (h). □
Finally, we give a special case of Theorem 4 for as follows.
Corollary 2.
Assume that the mean vector is estimable under (2), i.e., holds. Then, the following statements are equivalent:
- (a)
- holds definitely (with probability
- (b)
- holds definitely (with probability
- (c)
- (d)
- (e)
- (f)
Proof.
Set in Theorem 4(h) and, simplifying by (13), we obtain
Hence, in Theorem 4(h) is reduced to the rank formula in (f). □
5. Conclusions
The comparison and equivalence analysis of statistical inference results under true and misspecified linear models can be proposed from the theoretical and applied point of view, as illustrated in this article, while there are many mathematical methods and techniques that are available to address the problems of such kind under more and less serious statistical assumptions. As a concrete topic in this regard, we reconsidered in the previous sections some equivalence analysis problems under a true multivariate linear model and its two misspecified forms. The key step of this study is to convert the equivalence analysis problems under the three models into certain algebraic matrix equalities or equations, and then to obtain the corresponding results and facts from the three true and misspecified models by means of some effective matrix analysis tools, including the matrix equation method and the matrix rank method. Because conclusions in the preceding sections are all presented through certain explicit expressions and equalities, we believe that the contributions in this article are easy to understand and can serve as a group of theoretical references in the statistical analysis of various subsequent problems regarding MGLMs. Because all the formulas and facts in the preceding theorems are represented in certain analytical expressions or formulas, they can easily be reduced to various specified conclusions when the model matrices and covariance matrix in (1) are given in certain prescribed formulations. For example, let
in (1), respectively, which are regularly assumed in various concrete MGLMs.
We believe that the resultful studies on the equivalences of BLUPs/BLUEs provide significant advances to algebraical methodology in the statistical analysis of MGLMs, which will bring enough enabling methodological improvements and advances in the field of multivariate analysis. Finally, we propose a further problem on comparison and equivalence analysis of statistical inference results under the following two competing constrained MGLMs:
where is a consistent matrix equation for the unknown parameter matrix .
Author Contributions
Conceptualization, B.J.; methodology, Y.T.; investigation, B.J. and Y.T.; writing original draft, B.J.; writing, review and editing, Y.T. All authors have read and agreed to the published version of the manuscript.
Funding
This work was funded by the Shandong Provincial Natural Science Foundation #ZR2019MA065.
Acknowledgments
The authors are grateful to three referees for their helpful reports to an earlier version of this article.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Markiewicz, A.; Puntanen, S. All about the ⊥ with its applications in the linear statistical models. Open Math. 2015, 13, 33–50. [Google Scholar] [CrossRef]
- Puntanen, S.; Styan, G.P.H.; Isotalo, J. Matrix Tricks for Linear Statistical Models; Our Personal Top Twenty; Springer: Berlin, Germany, 2011. [Google Scholar]
- Rao, C.R.; Mitra, S.K. Generalized Inverse of Matrices and Its Applications; Wiley: New York, NY, USA, 1971. [Google Scholar]
- Jiang, B.; Tian, Y. On equivalence of predictors/estimators under a multivariate general linear model with augmentation. J. Korean Stat. Soc. 2017, 46, 551–561. [Google Scholar] [CrossRef]
- Nel, D.G. Tests for equality of parameter matrices in two multivariate linear models. J. Multivar. Anal. 1997, 61, 29–37. [Google Scholar] [CrossRef]
- Gamage, J.; Ananda, M.M.A. An exact test for testing the equality of parameter matrices in two multivariate linear models. Linear Algebra Appl. 2006, 418, 882–885. [Google Scholar] [CrossRef]
- Isotalo, J.; Puntanen, S.; Styan, G.P.H. Effect of adding regressors on the equality of the OLSE and BLUE. Int. J. Stat. Sci. 2010, 6, 193–201. [Google Scholar]
- Jammalamadaka, S.R.; Sengupta, D. Changes in the general linear model: A unified approach. Linear Algebra Appl. 1999, 289, 225–242. [Google Scholar] [CrossRef]
- Jammalamadaka, S.R.; Sengupta, D. Inclusion and exclusion of data or parameters in the general linear model. Stat. Probab. Lett. 2007, 77, 1235–1247. [Google Scholar] [CrossRef]
- Gan, S.; Sun, Y.; Tian, Y. Equivalence of predictors under real and over-parameterized linear models. Commun. Stat. Theor. Meth. 2017, 46, 5368–5383. [Google Scholar] [CrossRef]
- Jun, S.J.; Pinkse, J. Adding regressors to obtain efficiency. Econom. Theory 2009, 25, 298–301. [Google Scholar] [CrossRef][Green Version]
- Li, W.; Tian, Y.; Yuan, R. Statistical analysis of a linear regression model with restrictions and superfluous variables. J. Ind. Manag. Optim. 2023, 19, 3107–3127. [Google Scholar] [CrossRef]
- Lu, C.; Gan, S.; Tian, Y. Some remarks on general linear model with new regressors. Stat. Prob. Lett. 2015, 97, 16–24. [Google Scholar] [CrossRef]
- Magnus, J.R.; Durbin, J. Estimation of regression coefficients of interest when other regression coefficients are of no interest. Econometrica 1999, 67, 639–643. [Google Scholar] [CrossRef]
- Baksalary, J.K. A study of the equivalence between a Gauss–Markoff model and its augmentation by nuisance parameters. Statistics 1984, 15, 3–35. [Google Scholar]
- Bhimasankaram, P.; Jammalamadaka, S.R. Updates of statistics in a general linear model: A statistical interpretation and applications. Commun. Stat. Simul. Comput. 1994, 23, 789–801. [Google Scholar] [CrossRef]
- Haslett, S.J.; Puntanen, S. Effect of adding regressors on the equality of the BLUEs under two linear models. J. Stat. Plann. Inference 2010, 140, 104–110. [Google Scholar] [CrossRef]
- Marsaglia, G.; Styan, G.P.H. Equalities and inequalities for ranks of matrices. Linear Multilinear Algebra 1974, 2, 269–292. [Google Scholar] [CrossRef]
- Tian, Y. Matrix rank and inertia formulas in the analysis of general linear models. Open Math. 2017, 15, 126–150. [Google Scholar] [CrossRef]
- Penrose, R. A generalized inverse for matrices. Proc. Camb. Phil. Soc. 1955, 51, 406–413. [Google Scholar] [CrossRef]
- Tian, Y. A new derivation of BLUPs under random-effects model. Metrika 2015, 78, 905–918. [Google Scholar] [CrossRef]
- Tian, Y. A matrix handling of predictions under a general linear random-effects model with new observations. Electron. J. Linear Algebra 2015, 29, 30–45. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).