Abstract
Symmetry is a crucial concept in various fields of mathematics, offering a systematic approach to understanding structural properties and simplifying complex problems. This study focuses on linear mixed models, emphasizing the role of symmetry in the design of experiments and the structure of variance–covariance matrices. Building on foundational works, this paper introduces the concept of nested models and employs Wishart matrices for variance component estimation, enhancing the efficiency of complex model analysis. The methodology is particularly applicable when variance–covariance matrices conform to a commutative Jordan algebra (CJA). The practical significance of this modeling approach is demonstrated through numerical applications using both simulated and real-world data.
1. Introduction
Symmetry plays a fundamental role in various branches of mathematics and their applications, providing a framework for understanding structural properties and simplifying complex problems. In the context of linear mixed models, symmetry can manifest in the design of experiments, the structure of variance–covariance matrices, and the underlying algebraic structures used in estimation procedures.
Linear mixed models have been extensively explored in the literature, with numerous seminal contributions enhancing the theoretical and practical understanding of them. For instance, Brown and Prescott [1] provide a thorough exploration of how linear mixed models can be effectively applied in medical research, offering invaluable insights into both theoretical foundations and practical applications. Pinheiro and Bates [2] extend the applicability of mixed-effects models by delving into their implementation in S and S-PLUS, highlighting the versatility and practicality of these models. Similarly, Rao and Kleffe [3] make a foundational contribution by advancing the estimation of variance components and their applications, laying a critical groundwork for subsequent developments in the field. Sahai and Ageel [4] enrich the discourse on linear mixed models by addressing the complexities of fixed, random, and mixed models in their analysis of variance, providing a comprehensive treatment of these methodologies. Additionally, Demidenko [5] presents a modern perspective by integrating theoretical insights with practical implementations in R, bridging the gap between abstract theory and applied statistics.
Recent advancements have also addressed computational challenges in estimating variance components efficiently. Tack and Müller [6] developed fast restricted maximum likelihood (REML) estimation techniques for linear mixed models with Kronecker product covariance structures, improving computational performance in large-scale applications. Additionally, Lee et al. [7] provide a unified approach to generalized linear mixed models, offering a comprehensive treatment that extends classical methods to accommodate non-normal responses, thus broadening the scope of applications.
Beyond these influential works on linear mixed models, Fonseca et al. [8], for example, investigate the properties and applications of binary operations in linear algebra, specifically examining the scope of orthogonal normal models. Furthermore, Mexia et al. [9] contribute to this domain by developing the COBS methodology, which incorporates segregation, matching, crossing, and nesting, streamlining complex problems through innovative techniques. More recently, Ferreira et al. [10] extend these advancements by exploring inference in nonorthogonal mixed models, addressing key challenges in model estimation and interpretation. Additionally, Ferreira et al. [11] further enhance statistical methodologies by examining inference in mixed models with a mixture of distributions and controlled heteroscedasticity. Complementing these developments, Bailey et al. [12] focus on experimental design, proposing designs for half-diallel experiments with commutative orthogonal block structures, which optimize efficiency in statistical analyses.
In our study, we build on these foundational works by introducing the concept of nested models and utilizing Wishart matrices for estimating variance components. By simplifying complex models as building blocks, our approach streamlines the estimation process and makes it more efficient. Our method is particularly effective when variance–covariance matrices fall under a commutative Jordan algebra, known as CJA, and allows for the structure of variance–covariance matrices to be captured and manipulated in a way that respects both symmetry and commutativity, facilitating the estimation of variance components.
In the next section, we analyze the structure of linear models with nested random effects models, which refer to models with random effects that are hierarchically structured within each level of the model. Moreover, we also analyze how to estimate the variance components associated with them. Then, we discuss the testing of hypotheses on the effects and interactions of fixed effect factors using F-tests and introduce a measure of relevance for the hypotheses, which becomes especially useful when there are several rejected null hypotheses. In the special case section, we focus on a specific case where the mean and the vector of residuals to the mean are independent for every block of observations, and analyze the implications of this assumption for the model. The Multiple Regression Designs section extends these results to multi-treatment regression models, with nested random effect factors, see Mexia [13]. Finally, we illustrate our methodology through a numerical application, followed by concluding remarks.
2. Variance Components
Let b represent the number of blocks in a mixed model that correspond to the level combinations of fixed effect factors. The observation vector for each block, , , is assumed to be normally distributed and independent, with mean vector
where r represents the number of observations within each block, is a column vector of ones with r elements, is the overall average response, and represents a vector of unknown fixed effects.
The variance–covariance structure of is given by
where w represents the number of variance components in the model, , is a known design matrix, and represents the jth variance component.
Each block l has the same factor structure, so the design matrix for the random effects in block l, denoted by , is related to the design matrices through
where represents the contribution of the jth random effect in block l. The linear mixed model for each block may then be written as
where is a vector of random effects, assumed to be normally distributed with mean zero and variance–covariance matrix , and represents the error term for block l. The elements of are typically assumed to be independent and normally distributed with mean zero and variance .
Symmetry is explicitly leveraged in this model through the use of orthogonal transformations. Let be a matrix whose row vectors constitute an orthogonal basis for the orthogonal complement, , of the range space of the column matrix . The transformed vectors
are normally distributed and independent with mean zero and variance–covariance matrix
where
This transformation simplifies the variance–covariance structure while preserving the symmetry in the model. The orthogonal decomposition facilitated by ensures that the variance components remain independent, reflecting the fundamental principles of symmetry.
Next, consider a symmetric matrix of dimension with elements . The half-vectorization of , denoted as , is a column vector of dimension obtained by extracting the upper triangular elements (including the diagonal) of
Thus, contains the main diagonal and upper triangle of .
Now, defining
and
we obtain
If are linearly independent, then is full-rank, and the variance components can be estimated using the Moore–Penrose inverse
where + denotes the Moore–Penrose inverse.
We have a set of b independent normally distributed random variables with null mean vectors and variance–covariance matrix , as given by Equation (2). For each of the random variables , we can calculate its variance–covariance matrix as
where E denotes the expectation operator. Using the linearity of expectation we can then calculate the expected variance–covariance matrix of all the random variables as
So the usual estimator of is then simply the sample variance–covariance matrix, given by
Using Equation (12), and substituting the estimator for , we obtain the estimator
for .
To understand the properties of the estimator , we analyze the distribution and covariance structure of the sample variance–covariance matrix and its role in the estimation process.
We can write
where follows a Wishart distribution. Given that is an unbiased estimator of , we have
So the covariance matrix of the estimator is
see Anderson [14].
Furthermore, since are independent and have null mean vectors, we have
Note that the final step follows from the fact that the expected value of the sample variance–covariance matrix is just the variance–covariance matrix itself. Now, substituting into the expectation of the estimator, we have
Recalling Equation (12), , we then have
Therefore, the covariance matrix of the estimator is
This result establishes that the sample variance–covariance matrix is unbiased and provides the foundation for deriving the properties of the estimator , including its covariance structure, as shown in subsequent equations.
The estimator derived in Equation (16) is both efficient and effective in estimating the variance components. Efficiency in this context refers to the estimator having minimal variance among the class of unbiased estimators, ensuring that the variance components are estimated with the greatest possible precision given the available data. Effectiveness pertains to how well the estimator captures the underlying variance structure while maintaining desirable statistical properties such as unbiasedness and consistency. By leveraging the transformation matrix and ensuring independence in variance decomposition, the method provides an optimal approach for estimating variance components in mixed models. The explicit use of symmetry further enhances the reliability and interpretability of the estimates, making the approach both statistically sound and computationally feasible. Moreover, we can consider as a least square estimator, measuring the quality of by its determination coefficient, , as well as in the fixed effects models, to obtain the partitions
where
However, in general, the and , , are not independent. To quantify the variability of the block means , we define as their variance. This variance can be expressed in terms of the variance components as follows
with
So, to estimate , we define the estimator
which depends on the estimated variance components . However, in general, this estimator is not independent of the vector of block means
because the variance components themselves are influenced by the variability within blocks. This lack of independence introduces a complication in interpreting the variance structure.
Special Case
An interesting and important special case arises when the conditionr
is satisfied. Under this condition, the matrices simplify to
which implies that the transformed variance–covariance matrix becomes
In this case, the block means are independent of the deviations , . Consequently, the estimator becomes independent of the vector of block means . This independence is a desirable property because it simplifies the interpretation of the variance components. Furthermore, under this condition, the estimator is unbiased, as its expectation equals the true variance
This result underscores the importance of the condition in Equation , as it ensures that accurately reflects the variability of the block means without being confounded by the within-block deviations.
3. Multiple Regression Designs
The application of regression models with fixed and random effects has a long history in the statistical literature. Early foundational work, such as that by Henderson [15], introduced best linear unbiased predictors (BLUPs) for mixed models, setting the stage for their use in various disciplines. Later, Searle, Casella, and McCulloch [16] provided comprehensive methodologies for variance component estimation and model inference. These developments have been applied to diverse fields, including genetics, where mixed models are used to estimate heritability [17], and industrial quality control, where random effects are incorporated to account for batch-to-batch variability [18].
A fundamental principle underlying multiple regression designs is the concept of symmetry, which plays a crucial role in variance component estimation. Symmetry in experimental designs allows for an equitable decomposition of effects and interactions, ensuring orthogonality and simplifying inference. As outlined by Scheffé [19], the decomposition of effects into orthogonal components facilitates hypothesis testing and interpretation in complex models. More recently, Mexia [13] extended these ideas to encompass generalized least squares estimators and their properties under specific experimental conditions.
In this section, we analyze multiple regression designs in the context of the base model defined earlier. The focus is on assessing the influence of factors and their interactions on estimable functions. We derive least squares estimators, establish their statistical properties, and construct hypothesis tests and confidence regions. These results are critical for understanding the effects of the experimental design and for validating the model’s assumptions.
Consider the case where the observation vector , for each of the c treatments, satisfies Equation (33). The mean vector of is , and the variance–covariance matrix is , with having k linearly independent column vectors. This type of model has been applied in many situations, see Mexia [13]. To simplify computation, we replace with , derived using the Gram–Schmidt orthonormalization technique, ensuring . This yields the least squares estimator
with mean vector and variance–covariance matrix . The estimators , are mutually independent and independent from the
which will be the product by of independent central chi squares with degrees of freedom, . Thus,
with .
To analyze the behavior of the individual components of the regression coefficients, we define the vectors and . These quantities summarize the behavior of the estimators across treatments and allow us to study their distributional properties. Specifically, we define
where represents the true regression coefficients for the i-th component across all c treatments, and represents their least squares estimators.
The estimators , , will be normally distributed, with mean vectors , , and variance-covariance matrix . Furthermore, the estimators are independent of one another and also independent of the residual sums of squares . As a result, they are also independent of the overall residual sum of squares, S.
To study the influence of factors and interactions in the base design, we introduce an orthogonal partition of the space . This partition allows us to isolate and analyze the contributions of different effects and interactions through appropriate test statistics. Specifically, we assume the orthogonal partition
where ⊞ denotes the orthogonal direct sum of subspaces associated with the effects and interactions of the factors. If the row vectors of form an orthonormal basis for , , we have
if and only if The symmetric structure of this partition ensures that estimates remain unbiased and efficiently computed.
To analyze specific effects, we define the quantities
where and Using these quantities, we define
and
The hypotheses to be tested are
where rejection of the null hypotheses indicates a significant effect or interaction.
To evaluate these hypotheses, we construct the test statistics
where the test statistics follow an -distribution with and g [or and g] degrees of freedom and non-centrality parameters [or ] for and
The uniformity in variance structure, a direct consequence of symmetry, enables the construction of uniformly most powerful (UMP) tests. Specifically, as noted by Lehmann-EL and Romano [20], the -tests for the hypotheses , , and , , are uniformly most powerful (UMP) within the class of tests whose power is determined by non-centrality parameters. Furthermore, their power increases proportionally with these parameters, as established in Mexia [21].
Let denote the p-th quantile for the central -distribution with and g degrees of freedom. Since
and the quantities are independent from S, we have
also independent of
The pivot variables
follow a central -distribution with degrees of freedom [or ] and
This result provides a framework for constructing confidence regions for the parameters Specifically, we have
where
To validate the model, we use the values introduced earlier and the Bartlett homoscedasticity test, see Bartlett and Kendall [22]. Assuming
with independent chi squares, we test the hypothesis
Rejection of indicates heteroscedasticity, suggesting that the model may need refinement.
Finally, we construct confidence regions for estimable functions. Let
and assume
The pivot variable
has a central -distribution with g and degrees of freedom. This leads to the confidence region
This confidence region quantifies the uncertainty in the estimable functions, providing a practical tool for inference.
4. Numerical Application
In this section, we first illustrate the theory using simulated data and then apply the proposed model to a real dataset.
All computations and simulations were conducted using the R software 4.3.0. The program estimates variance components within a balanced experimental design, defining both a full model and a two-block model. It constructs the necessary design matrices, simulates responses from a normal distribution, and applies the formulas presented in the previous sections to estimate the variance components.
4.1. Simulated Data
In agricultural experiments, it is common to study crop yields under different fertilizers while accounting for the variability introduced by different plots of land. In this example, let us consider such data, where plots are treated as fixed effects and fertilizers as random effects. The dataset consists of crop yields for four plots (fixed effect factor) under four different fertilizers (random effect factor), as shown in Table 1. To investigate potential differences in variability across subsets of the data, the dataset was divided into two blocks: Block 1, which includes data from Plot 1 and Plot 2, and Block 2, which includes data from Plot 3 and Plot 4.
Table 1.
Crop yields under different fertilizers.
The crop yield data in Table 1 were generated based on the following:
where is a column vector of ones with eight elements , representing the intercept term; contains the observed crop yields for block l; is a vector of unknown fixed effects associated with the plots; and is a vector of random effects associated with the fertilizers, assumed to be normally distributed with mean zero and variance–covariance matrix , where , with being the variance component for the random effects and being the identity matrix. Additionally, represents the error term for block l, with its elements being independent and identically distributed, following a normal distribution with mean zero and variance .
The design matrices for the fixed and random effects are, respectively,
For data generation, the variance components were set as and . The crop yield data for each block, , were simulated in R software 4.3.0 according to Equation (59). and were estimated as indicated in (16) and the orthogonal transformation matrix used was
This transformation ensures that the variance–covariance structure becomes clearer. For example, the variance of is represented as
where
Using the transformed data, the variance components and were estimated by solving
with
To analyze the entire model (i.e., using data from all four plots without splitting into blocks), the same linear mixed model framework was applied. The dataset was treated as a single block, where the observation vector for all plots was modeled as
with being a column vector of ones with 16 elements, encoding the fixed effects for all plots, and representing the random effects for the fertilizers,
The variance components and were estimated using the same method described before.
The estimation procedure was repeated 1000 times, yielding the mean variance component estimates in Table 2.
Table 2.
Variance component estimates for different blocks and the entire model.
As can be seen, the estimates for the variance components and differed slightly when calculated within the blocks as compared to the entire model. When considering the entire model, the estimates for the variance components were more averaged, representing a general overview of the variability across all plots. However, the block-wise analysis yielded slightly more refined estimates for each block, indicating the possibility of different variance structures within the blocks. So, while the differences in estimates were not large, the block approach can be beneficial for situations where there are substantial differences between groups (blocks), as it may lead to more accurate and tailored estimates of the variance components.
4.2. Real Data
In this subsection, we apply the proposed model to a real dataset related to housing affordability across different European countries. Specifically, we analyze the standardized house-price-to-income ratio for four countries (Croatia, Spain, Ireland, and Poland) over four years (2020–2023). This ratio measures how affordable housing is in each country, with higher values indicating higher house prices relative to income.
The dataset, sourced from Eurostat [23], 2025, is presented in Table 3.
Table 3.
Standardized house-price-to-income ratio—annual data.
Here, years are treated as a fixed effect, while countries are modeled as a random effect. The analysis follows the same methodology as in the simulated data case. That is, considering the entire dataset and the block approach, the dataset is divided into Block 1 (2020 and 2021) and Block 2 (2022 and 2023). The results are shown in Table 4.
Table 4.
Variance component estimates for different blocks and the wntire model.
The random effect variance , which accounts for country-level differences, varies significantly across blocks (15.4309 for Block 1 vs. 19.2157 for Block 2). The error variance also differs (higher in Block 1, lower in Block 2), suggesting changes in overall variability over time. The entire model provides an averaged estimate of variance components (, ), potentially smoothing over important structural differences across years.
The block-wise model provides a more refined view of the variance structure across different periods, which is valuable when significant shifts occur in the data. In contrast, the entire model offers a more general picture, which may be useful for broad trend analysis. In this case, since and differ across blocks, using a separated model can be advantageous if the goal is to capture year-specific variability. However, if the primary interest is in overall trends, the entire model remains a reasonable choice.
5. Limitations and Future Directions
While the models proposed in this paper offer significant advancements, several limitations should be acknowledged. First, the assumptions regarding the structure of the data, such as the CJA condition for variance–covariance matrices, may not always hold in practical applications. Additionally, the computational complexity of implementing these models could present challenges, particularly in large-scale datasets or when dealing with high-dimensional random effects. Future work could explore relaxing these assumptions.
6. Final Remarks
This paper offers a comprehensive guide to linear models with nested random effects. The statistical methods developed in this paper rely on symmetry principles. From the orthogonal transformations that simplify the variance structure to the independence and unbiasedness of estimators, symmetry plays a pivotal role in ensuring the robustness and elegance of the statistical framework. A significant innovative aspect is the extension of results to multi-treatment regression models with nested random effect factors, providing a more versatile modeling approach. The models developed are particularly practical when variance–covariance matrices fall under a commutative Jordan algebra (CJA), expanding their applicability. A numerical application is also provided to illustrate the practical implementation of these concepts and methods.
Funding
Nece and this work are supported by FCT—Fundação para a Ciência e a Tecnologia, I.P. by project references UIDB/00212/2020, UIDB/04630/2020 and UIDB/00297/2020.
Data Availability Statement
Eurostat Database. Available online: https://ec.europa.eu/eurostat/databrowser/view/tipsho60/default/table?lang=en&category=prc.prc_hpi (accessed on 1 February 2025).
Conflicts of Interest
The author declares no conflicts of interest.
References
- Brown, H.; Prescott, R. Applied Mixed Models in Medicine; Wiley: New York, NY, USA, 1999. [Google Scholar]
- Pinheiro, J.C.; Bates, D.M. Mixed-Effects Models in S and S-PLUS; Springer: New York, NY, USA, 2000. [Google Scholar]
- Rao, C.R.; Kleffe, J. Estimation of Variance Components and Applications; North-Holland: Amsterdam, The Netherlands, 1988. [Google Scholar] [CrossRef]
- Sahai, H.; Ageel, M.I. Analysis of Variance: Fixed, Random and Mixed Models; Birkhäuser: Cambridge, MA, USA, 2000. [Google Scholar]
- Demidenko, E. Mixed Models: Theory and Applications with R; Wiley: New York, NY, USA, 2013. [Google Scholar]
- Tack, L.; Müller, S. Fast REML estimation for linear mixed models with kronecker product covariance structures. J. Comput. Graph. Stat. 2021, 30, 1148–1160. [Google Scholar]
- Lee, Y.; Nelder, J.A.; Pawitan, Y. Generalized Linear Mixed Models: A Unified Approach; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar]
- Fonseca, M.; Mexia, J.T.; Zmyslony, R. Binary operation on Jordan algebras and orthogonal normal models. Linear Algebra Appl. 2006, 417, 75–86. [Google Scholar] [CrossRef]
- Mexia, J.T.; Vaquinhas, R.; Fonseca, M.; Zmyslony, R. COBS: Segregation, Matching, Crossing and Nesting. Latest Trends on Applied Mathematics, Simulation, Modeling. In Proceedings of the 4th International Conference on Applied Mathematics, Simulation, Modelling, Corfu Island, Greece, 22–25 July 2010. [Google Scholar]
- Ferreira, D.; Ferreira, S.S.; Nunes, C.; Mexia, J.T. Inference in nonorthogonal mixed models. Math. Methods Appl. Sci. 2022, 45, 3183–3196. [Google Scholar] [CrossRef]
- Ferreira, D.; Ferreira, S.S.; Antunes, P.; Oliveira, T.A.; Mexia, J.T. Inference in mixed models with a mixture of distributions and controlled heteroscedasticity. Commun. Stat.—Theory Methods 2024. [Google Scholar] [CrossRef]
- Bailey, R.A.; Cameron, P.J.; Ferreira, D.; Ferreira, S.S.; Nunes, C. Designs for Half-Diallel Experiments with Commutative Orthogonal Block Structure. J. Stat. Plan. Inference 2024, 231, 106139. [Google Scholar] [CrossRef]
- Mexia, J.T. Multi-Treatment Regression Designs; Universidade Nova de Lisboa: Lisbon, Portugal, 1987. [Google Scholar]
- Anderson, T.W. Introduction to Multivariate Statistical Analysis; John Wiley & Sons, Inc.: New York, NY, USA, 1958. [Google Scholar]
- Henderson, C.R. Estimation of genetic parameters. Ann. Math. Stat. 1950, 21, 309–310. [Google Scholar]
- Searle, S.R.; Casella, G.; McCulloch, C.E. Variance Components; John Wiley & Sons: New York, NY, USA, 1992. [Google Scholar]
- Lynch, M.; Walsh, B. Genetics and Analysis of Quantitative Traits; Sinauer Associates: Sunderland, MA, USA, 1998. [Google Scholar]
- Montgomery, D.C. Design and Analysis of Experiments, 8th ed.; John Wiley & Sons: New York, NY, USA, 2008. [Google Scholar]
- Scheffé, H. The Analysis of Variance; Wiley: New York, NY, USA, 1959. [Google Scholar]
- Lehmann, E.L.; Romano, J.P. Testing Statistical Hypotheses; Springer: New York, NY, USA, 2005. [Google Scholar]
- Mexia, J.T. Controlled Heteroscedasticity, Quocient Vector Spaces and F Tests for Hypothesis on Mean Vectors. Ph.D. Thesis, Universidade Nova de Lisboa, Lisbon, Portugal, 1987. [Google Scholar]
- Bartlett, M.S.; Kendall, D.G. The Statistical Analysis of Variance Heterogeneity and the Logarithmic Transformation. Suppl. J. R. Stat. Soc. 1946, 8, 128–138. [Google Scholar] [CrossRef]
- Eurostat Database. Available online: https://ec.europa.eu/eurostat/databrowser/view/tipsho60/default/table?lang=en&category=prc.prc_hpi (accessed on 1 February 2025).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).