1. Introduction
Symmetry plays a fundamental role in various branches of mathematics and their applications, providing a framework for understanding structural properties and simplifying complex problems. In the context of linear mixed models, symmetry can manifest in the design of experiments, the structure of variance–covariance matrices, and the underlying algebraic structures used in estimation procedures.
Linear mixed models have been extensively explored in the literature, with numerous seminal contributions enhancing the theoretical and practical understanding of them. For instance, Brown and Prescott [
1] provide a thorough exploration of how linear mixed models can be effectively applied in medical research, offering invaluable insights into both theoretical foundations and practical applications. Pinheiro and Bates [
2] extend the applicability of mixed-effects models by delving into their implementation in S and S-PLUS, highlighting the versatility and practicality of these models. Similarly, Rao and Kleffe [
3] make a foundational contribution by advancing the estimation of variance components and their applications, laying a critical groundwork for subsequent developments in the field. Sahai and Ageel [
4] enrich the discourse on linear mixed models by addressing the complexities of fixed, random, and mixed models in their analysis of variance, providing a comprehensive treatment of these methodologies. Additionally, Demidenko [
5] presents a modern perspective by integrating theoretical insights with practical implementations in R, bridging the gap between abstract theory and applied statistics.
Recent advancements have also addressed computational challenges in estimating variance components efficiently. Tack and Müller [
6] developed fast restricted maximum likelihood (REML) estimation techniques for linear mixed models with Kronecker product covariance structures, improving computational performance in large-scale applications. Additionally, Lee et al. [
7] provide a unified approach to generalized linear mixed models, offering a comprehensive treatment that extends classical methods to accommodate non-normal responses, thus broadening the scope of applications.
Beyond these influential works on linear mixed models, Fonseca et al. [
8], for example, investigate the properties and applications of binary operations in linear algebra, specifically examining the scope of orthogonal normal models. Furthermore, Mexia et al. [
9] contribute to this domain by developing the COBS methodology, which incorporates segregation, matching, crossing, and nesting, streamlining complex problems through innovative techniques. More recently, Ferreira et al. [
10] extend these advancements by exploring inference in nonorthogonal mixed models, addressing key challenges in model estimation and interpretation. Additionally, Ferreira et al. [
11] further enhance statistical methodologies by examining inference in mixed models with a mixture of distributions and controlled heteroscedasticity. Complementing these developments, Bailey et al. [
12] focus on experimental design, proposing designs for half-diallel experiments with commutative orthogonal block structures, which optimize efficiency in statistical analyses.
In our study, we build on these foundational works by introducing the concept of nested models and utilizing Wishart matrices for estimating variance components. By simplifying complex models as building blocks, our approach streamlines the estimation process and makes it more efficient. Our method is particularly effective when variance–covariance matrices fall under a commutative Jordan algebra, known as CJA, and allows for the structure of variance–covariance matrices to be captured and manipulated in a way that respects both symmetry and commutativity, facilitating the estimation of variance components.
In the next section, we analyze the structure of linear models with nested random effects models, which refer to models with random effects that are hierarchically structured within each level of the model. Moreover, we also analyze how to estimate the variance components associated with them. Then, we discuss the testing of hypotheses on the effects and interactions of fixed effect factors using F-tests and introduce a measure of relevance for the hypotheses, which becomes especially useful when there are several rejected null hypotheses. In the special case section, we focus on a specific case where the mean and the vector of residuals to the mean are independent for every block of observations, and analyze the implications of this assumption for the model. The Multiple Regression Designs section extends these results to multi-treatment regression models, with nested random effect factors, see Mexia [
13]. Finally, we illustrate our methodology through a numerical application, followed by concluding remarks.
2. Variance Components
Let
b represent the number of blocks in a mixed model that correspond to the level combinations of fixed effect factors. The observation vector for each block,
,
, is assumed to be normally distributed and independent, with mean vector
where
r represents the number of observations within each block,
is a column vector of ones with
r elements,
is the overall average response, and
represents a vector of unknown fixed effects.
The variance–covariance structure of
is given by
where
w represents the number of variance components in the model,
,
is a known design matrix, and
represents the
jth variance component.
Each block
l has the same factor structure, so the design matrix for the random effects in block
l, denoted by
, is related to the design matrices
through
where
represents the contribution of the
jth random effect in block
l. The linear mixed model for each block may then be written as
where
is a vector of random effects, assumed to be normally distributed with mean zero and variance–covariance matrix
, and
represents the error term for block
l. The elements of
are typically assumed to be independent and normally distributed with mean zero and variance
.
Symmetry is explicitly leveraged in this model through the use of orthogonal transformations. Let
be a
matrix whose row vectors constitute an orthogonal basis for the orthogonal complement,
, of the range space of the column matrix
. The transformed vectors
are normally distributed and independent with mean zero and variance–covariance matrix
where
This transformation simplifies the variance–covariance structure while preserving the symmetry in the model. The orthogonal decomposition facilitated by ensures that the variance components remain independent, reflecting the fundamental principles of symmetry.
Next, consider a symmetric matrix
of dimension
with elements
. The half-vectorization of
, denoted as
, is a column vector of dimension
obtained by extracting the upper triangular elements (including the diagonal) of
Thus, contains the main diagonal and upper triangle of .
Now, defining
and
we obtain
If
are linearly independent, then
is full-rank, and the variance components can be estimated using the Moore–Penrose inverse
where + denotes the Moore–Penrose inverse.
We have a set of
b independent normally distributed random variables
with null mean vectors and variance–covariance matrix
, as given by Equation (
2). For each of the random variables
, we can calculate its variance–covariance matrix as
where
E denotes the expectation operator. Using the linearity of expectation we can then calculate the expected variance–covariance matrix of all the random variables as
So the usual estimator of
is then simply the sample variance–covariance matrix, given by
Using Equation (
12), and substituting the estimator
for
, we obtain the estimator
for
.
To understand the properties of the estimator , we analyze the distribution and covariance structure of the sample variance–covariance matrix and its role in the estimation process.
We can write
where
follows a Wishart distribution. Given that
is an unbiased estimator of
, we have
So the covariance matrix of the estimator
is
see Anderson [
14].
Furthermore, since
are independent and have null mean vectors, we have
Note that the final step follows from the fact that the expected value of the sample variance–covariance matrix is just the variance–covariance matrix itself. Now, substituting
into the expectation of the estimator, we have
Recalling Equation (
12),
, we then have
Therefore, the covariance matrix of the estimator
is
This result establishes that the sample variance–covariance matrix is unbiased and provides the foundation for deriving the properties of the estimator , including its covariance structure, as shown in subsequent equations.
The estimator
derived in Equation (
16) is both efficient and effective in estimating the variance components. Efficiency in this context refers to the estimator having minimal variance among the class of unbiased estimators, ensuring that the variance components are estimated with the greatest possible precision given the available data. Effectiveness pertains to how well the estimator captures the underlying variance structure while maintaining desirable statistical properties such as unbiasedness and consistency. By leveraging the transformation matrix
and ensuring independence in variance decomposition, the method provides an optimal approach for estimating variance components in mixed models. The explicit use of symmetry further enhances the reliability and interpretability of the estimates, making the approach both statistically sound and computationally feasible. Moreover, we can consider
as a least square estimator, measuring the quality of
by its determination coefficient,
, as well as in the fixed effects models, to obtain the partitions
where
However, in general, the
and
,
, are not independent. To quantify the variability of the block means
, we define
as their variance. This variance can be expressed in terms of the variance components
as follows
with
So, to estimate
, we define the estimator
which depends on the estimated variance components
. However, in general, this estimator
is not independent of the vector of block means
because the variance components
themselves are influenced by the variability within blocks. This lack of independence introduces a complication in interpreting the variance structure.
Special Case
An interesting and important special case arises when the conditionr
is satisfied. Under this condition, the matrices
simplify to
which implies that the transformed variance–covariance matrix becomes
In this case, the block means
are independent of the deviations
,
. Consequently, the estimator
becomes independent of the vector of block means
. This independence is a desirable property because it simplifies the interpretation of the variance components. Furthermore, under this condition, the estimator
is unbiased, as its expectation equals the true variance
This result underscores the importance of the condition in Equation , as it ensures that accurately reflects the variability of the block means without being confounded by the within-block deviations.
3. Multiple Regression Designs
The application of regression models with fixed and random effects has a long history in the statistical literature. Early foundational work, such as that by Henderson [
15], introduced best linear unbiased predictors (BLUPs) for mixed models, setting the stage for their use in various disciplines. Later, Searle, Casella, and McCulloch [
16] provided comprehensive methodologies for variance component estimation and model inference. These developments have been applied to diverse fields, including genetics, where mixed models are used to estimate heritability [
17], and industrial quality control, where random effects are incorporated to account for batch-to-batch variability [
18].
A fundamental principle underlying multiple regression designs is the concept of symmetry, which plays a crucial role in variance component estimation. Symmetry in experimental designs allows for an equitable decomposition of effects and interactions, ensuring orthogonality and simplifying inference. As outlined by Scheffé [
19], the decomposition of effects into orthogonal components facilitates hypothesis testing and interpretation in complex models. More recently, Mexia [
13] extended these ideas to encompass generalized least squares estimators and their properties under specific experimental conditions.
In this section, we analyze multiple regression designs in the context of the base model defined earlier. The focus is on assessing the influence of factors and their interactions on estimable functions. We derive least squares estimators, establish their statistical properties, and construct hypothesis tests and confidence regions. These results are critical for understanding the effects of the experimental design and for validating the model’s assumptions.
Consider the case where the observation vector
, for each of the
c treatments, satisfies Equation (
33). The mean vector of
is
, and the variance–covariance matrix is
, with
having
k linearly independent column vectors. This type of model has been applied in many situations, see Mexia [
13]. To simplify computation, we replace
with
, derived using the Gram–Schmidt orthonormalization technique, ensuring
. This yields the least squares estimator
with mean vector
and variance–covariance matrix
. The estimators
,
are mutually independent and independent from the
which will be the product by
of independent central chi squares with
degrees of freedom,
. Thus,
with
.
To analyze the behavior of the individual components of the regression coefficients, we define the vectors
and
. These quantities summarize the behavior of the estimators across treatments and allow us to study their distributional properties. Specifically, we define
where
represents the true regression coefficients for the
i-th component across all
c treatments, and
represents their least squares estimators.
The estimators , , will be normally distributed, with mean vectors , , and variance-covariance matrix . Furthermore, the estimators are independent of one another and also independent of the residual sums of squares . As a result, they are also independent of the overall residual sum of squares, S.
To study the influence of factors and interactions in the base design, we introduce an orthogonal partition of the space
. This partition allows us to isolate and analyze the contributions of different effects and interactions through appropriate test statistics. Specifically, we assume the orthogonal partition
where ⊞ denotes the orthogonal direct sum of subspaces associated with the effects and interactions of the factors. If the
row vectors of
form an orthonormal basis for
,
, we have
if and only if
The symmetric structure of this partition ensures that estimates remain unbiased and efficiently computed.
To analyze specific effects, we define the quantities
where
and
Using these quantities, we define
and
The hypotheses to be tested are
where rejection of the null hypotheses indicates a significant effect or interaction.
To evaluate these hypotheses, we construct the test statistics
where the test statistics follow an
-distribution with
and
g [or
and
g] degrees of freedom and non-centrality parameters
[or
] for
and
The uniformity in variance structure, a direct consequence of symmetry, enables the construction of uniformly most powerful (UMP) tests. Specifically, as noted by Lehmann-EL and Romano [
20], the
-tests for the hypotheses
,
, and
,
, are uniformly most powerful (UMP) within the class of tests whose power is determined by non-centrality parameters. Furthermore, their power increases proportionally with these parameters, as established in Mexia [
21].
Let
denote the
p-th quantile for the central
-distribution with
and
g degrees of freedom. Since
and the quantities are independent from
S, we have
also independent of
The pivot variables
follow a central
-distribution with degrees of freedom
[or
] and
This result provides a framework for constructing confidence regions for the parameters
Specifically, we have
where
To validate the model, we use the
values introduced earlier and the Bartlett homoscedasticity test, see Bartlett and Kendall [
22]. Assuming
with independent chi squares, we test the hypothesis
Rejection of indicates heteroscedasticity, suggesting that the model may need refinement.
Finally, we construct confidence regions for estimable functions. Let
and assume
The pivot variable
has a central
-distribution with
g and
degrees of freedom. This leads to the
confidence region
This confidence region quantifies the uncertainty in the estimable functions, providing a practical tool for inference.
4. Numerical Application
In this section, we first illustrate the theory using simulated data and then apply the proposed model to a real dataset.
All computations and simulations were conducted using the R software 4.3.0. The program estimates variance components within a balanced experimental design, defining both a full model and a two-block model. It constructs the necessary design matrices, simulates responses from a normal distribution, and applies the formulas presented in the previous sections to estimate the variance components.
4.1. Simulated Data
In agricultural experiments, it is common to study crop yields under different fertilizers while accounting for the variability introduced by different plots of land. In this example, let us consider such data, where plots are treated as fixed effects and fertilizers as random effects. The dataset consists of crop yields for four plots (fixed effect factor) under four different fertilizers (random effect factor), as shown in
Table 1. To investigate potential differences in variability across subsets of the data, the dataset was divided into two blocks: Block 1, which includes data from Plot 1 and Plot 2, and Block 2, which includes data from Plot 3 and Plot 4.
The crop yield data in
Table 1 were generated based on the following:
where
is a column vector of ones with eight elements
, representing the intercept term;
contains the observed crop yields for block
l;
is a vector of unknown fixed effects associated with the plots; and
is a vector of random effects associated with the fertilizers, assumed to be normally distributed with mean zero and variance–covariance matrix
, where
, with
being the variance component for the random effects and
being the
identity matrix. Additionally,
represents the error term for block
l, with its elements being independent and identically distributed, following a normal distribution with mean zero and variance
.
The design matrices for the fixed and random effects are, respectively,
For data generation, the variance components were set as
and
. The crop yield data for each block,
, were simulated in R software 4.3.0 according to Equation (
59).
and
were estimated as indicated in (
16) and the orthogonal transformation matrix used was
This transformation ensures that the variance–covariance structure becomes clearer. For example, the variance of
is represented as
where
Using the transformed data, the variance components
and
were estimated by solving
with
To analyze the entire model (i.e., using data from all four plots without splitting into blocks), the same linear mixed model framework was applied. The dataset was treated as a single block, where the observation vector
for all plots was modeled as
with
being a column vector of ones with 16 elements,
encoding the fixed effects for all plots, and
representing the random effects for the fertilizers,
The variance components and were estimated using the same method described before.
The estimation procedure was repeated 1000 times, yielding the mean variance component estimates in
Table 2.
As can be seen, the estimates for the variance components and differed slightly when calculated within the blocks as compared to the entire model. When considering the entire model, the estimates for the variance components were more averaged, representing a general overview of the variability across all plots. However, the block-wise analysis yielded slightly more refined estimates for each block, indicating the possibility of different variance structures within the blocks. So, while the differences in estimates were not large, the block approach can be beneficial for situations where there are substantial differences between groups (blocks), as it may lead to more accurate and tailored estimates of the variance components.
4.2. Real Data
In this subsection, we apply the proposed model to a real dataset related to housing affordability across different European countries. Specifically, we analyze the standardized house-price-to-income ratio for four countries (Croatia, Spain, Ireland, and Poland) over four years (2020–2023). This ratio measures how affordable housing is in each country, with higher values indicating higher house prices relative to income.
The dataset, sourced from Eurostat [
23], 2025, is presented in
Table 3.
Here, years are treated as a fixed effect, while countries are modeled as a random effect. The analysis follows the same methodology as in the simulated data case. That is, considering the entire dataset and the block approach, the dataset is divided into Block 1 (2020 and 2021) and Block 2 (2022 and 2023). The results are shown in
Table 4.
The random effect variance , which accounts for country-level differences, varies significantly across blocks (15.4309 for Block 1 vs. 19.2157 for Block 2). The error variance also differs (higher in Block 1, lower in Block 2), suggesting changes in overall variability over time. The entire model provides an averaged estimate of variance components (, ), potentially smoothing over important structural differences across years.
The block-wise model provides a more refined view of the variance structure across different periods, which is valuable when significant shifts occur in the data. In contrast, the entire model offers a more general picture, which may be useful for broad trend analysis. In this case, since and differ across blocks, using a separated model can be advantageous if the goal is to capture year-specific variability. However, if the primary interest is in overall trends, the entire model remains a reasonable choice.