1. Introduction
The study of sphericity is a fundamental problem in multivariate analysis, with applications in various fields such as statistics, for example, in repeated measures ANOVA, where it is assumed that the variances of the differences between all combinations of related groups are equal; in engineering, for example, in signal processing, where often sphericity is assumed to ensure that the noise in the signals is uniformly distributed; and computer science, particularly in machine learning, where the concept of sphericity is applied in clustering algorithms like k-means. Moreover, sphericity is a stronger assumption than the homogeneity of variances. Suppose you have k samples with the same size. The data are said to have the property of sphericity if the pairwise differences in variance between the samples are all equal. If sphericity is violated, the calculations of variance can be distorted. Therefore, the need to measure sphericity plays a fundamental role in several statistical problems. Sphericity tests have been extensively studied in the literature, and several measures have been proposed to assess the degree of sphericity in a matrix. For example, Dawid [
1] considered an inference about the parameters of a multivariate linear model, in which the usual assumption of normality for the errors was replaced by a weaker assumption of spherical symmetry. Two Tukey multiple comparison procedures, as well as a Bonferroni and multivariate approach, were compared for their rates of Type I error and any-pairs power when multisample sphericity was not satisfied, and the design was unbalanced in Keselman and Keselman [
2]. Cornell et al. [
3] conducted a Monte Carlo simulation to investigate the relative power of eight tests for sphericity in randomized block designs. Chen et al. [
4] proposed nonparametric tests, assuming a specific parametric distribution for the data, for the sphericity and identity of high-dimensional covariance matrices. A test for the null of sphericity in a fixed effects panel data model is presented in Baltagi et al. [
5]. Onatski et al. [
6] studied the asymptotic power of tests of sphericity against perturbations in a single unknown direction as both the dimensionality of data and the number of observations go to infinity. The problem of testing the null hypothesis of sphericity of a high-dimensional covariance matrix against an alternative of unspecified multiple symmetry-breaking directions (multispiked alternatives) was considered in Onatski et al. [
7]. Zou et al. [
8] proposed tests for sphericity in cases where the data dimension is larger than the sample size.
In this paper, we propose a measure of sphericity, denoted . Additionally, we include a comparison between the proposed measure and other widely used measures of sphericity, namely, the Frobenius measure and the spectral measure.
While existing measures for sphericity are valuable, our proposed measure is tailored specifically for positive semi-definite matrices and offers insights into their geometric properties that traditional measures may not capture. Traditional measures are often designed for symmetric matrices, particularly covariance matrices in factor analysis, and may not adequately address the unique challenges of positive semi-definite matrices. In contrast, our measure directly considers the spread of eigenvalues, providing a range from 0 to 1 and offering a clear interpretation of the degree of sphericity. Additionally, it exhibits desirable properties such as invariance under orthogonal transformations and scaling, making it suitable for various statistical and machine learning tasks in high-dimensional settings.
The remainder of this paper is organized as follows: In
Section 2, we introduce the measure of sphericity
and establish its properties in finite-dimensional cases. In
Section 3, we extend our analysis to high-dimensional scenarios and examine the behavior of the proposed measure as the dimensionality increases. In
Section 4, we explore the stochastic case and investigate the convergence properties of the sphericity measure for sequences of random matrices. In
Section 5, we present a numerical comparison between the proposed approach and the Frobenius and the spectral measures. Finally, in
Section 6, we provide some final remarks.
2. Measure of Sphericity for Positive Semi-Definite Matrices
In this section, we introduce a measure of sphericity for positive semi-definite matrices. Let
denote the family of
positive semi-definite [definite] matrices. Given
, let
be the vector of its eigenvalues. Consequently, the trace and norm of
are given by
and
respectively.
Defining the Frobenius measure as
we now establish
Proposition 1. If then we have .
Proof. Let
with eigenvalues
. Using Lagrange multipliers to minimize
subject to
, we have
Solving
gives
, and since
, we obtain
. Thus,
Therefore,
Now, since
for all
j,
Thus,
and so we obtain
and the proof is completed. □
We can now use
to measure the sphericity of
.
When , indicating maximum anisotropy, equals 0. On the other hand, when , indicating perfect sphericity, equals 1. Therefore, ranges from 0 to 1, where higher values signify greater sphericity. This bounded interval, , ensures a clear and interpretable metric for assessing the sphericity of a given matrix .
Referring to Puntanen et al. [
9], if
is an orthogonal matrix, then,
will have the same eigenvalues as
. Consequently, we have
and thus, according to Equation (
1),
From this, it follows that
This invariance under orthogonal transformations implies that
is a geometric property of the matrix that does not depend on the coordinate system, making it a robust measure for assessing the sphericity of positive semi-definite matrices.
It is clear that, if
, we have
and hence,
This property indicates that
is a scale-invariant measure.
Representing the Kronecker matrix product by ⊗, when
,
, we have
so that
and that
Thus,
This multiplicative property under Kronecker products is useful in higher-dimensional sphericity measures.
Let us now establish
Proposition 2. Given , symmetric matrices, we have if, and only if, , with orthogonal.
Proof. If matrix exists, it diagonalizes both and , resulting in equal eigenvalues . Let be such that , where is a diagonal matrix with the eigenvalues of on its diagonal.
Then, we have . Since is orthogonal, its transpose is its inverse, so .
Since is orthogonal, denoted by , we have , where is orthogonal, completing the proof. □
Let us denote
if their eigenvalues, ordered in decreasing order, are equal. This establishes an equivalence relation, as confirmed by Proposition 2. Under this relation, if
, we observe that
as well as
This implies that
is solely determined by the eigenvalues of
, independent of its specific structure or arrangement. Thus, the metric
offers a robust assessment of sphericity, relying solely on the intrinsic properties of the matrix’s eigenvalues.
3. High-Dimensional Case
In this section, we extend our discussion to the high-dimensional case, where k becomes large. We explore whether there are notable differences between the finite k case and the scenario where k tends to infinity. Additionally, we investigate the properties exhibited when k is large.
In the finite k scenario, the measure effectively ranges between 0 and 1, allowing us to discern varying degrees of sphericity among matrices. However, as k approaches infinity, distinct behaviors emerge in the properties of .
Let us start by examining the expression for
, given in Equation (
2). The behavior of
is influenced by the behavior of
. As
k approaches infinity, we need to analyze how
behaves, which can be analyzed based on the relative magnitudes of the Frobenius norm and the trace of
.
For any matrix , the inequality holds, with equality only when is a scalar multiple of the identity matrix. As k tends to infinity, this inequality becomes increasingly significant for non-scalar multiples of the identity matrix. Consequently, tends towards 1 for matrices deviating from being scalar multiples of the identity, reflecting their departure from sphericity. Conversely, for matrices closely resembling scalar multiples of the identity (i.e., exhibiting high sphericity), approaches 0.
This behavior of underpins the changing nature of in the high-dimensional regime. As k grows large, becomes more sensitive to deviations from sphericity, reflecting the increasing diversity and complexity of matrices in higher dimensions. Additionally, the range of values may expand or contract as k tends to infinity, reflecting the evolving landscape of matrix sphericity in high-dimensional spaces.
Overall, the behavior of in the high-dimensional scenario offers valuable insights into the intrinsic properties of matrices and their representations in complex spaces. A further exploration and analysis of these patterns can provide a deeper understanding and appreciation of matrix geometry in high dimensions.
4. Stochastic Matrices
In this section, we discuss asymptotic normality for a sequence of
random matrices
, where
represents the random matrices in the sequence, and
represents the expectation of
. Specifically, we consider the scaling behavior as the dimension
d goes to infinity. So, we have
where
represents a scaling parameter, indicating convergence in distribution to a normal distribution with mean vector
and covariance matrix
. We define
If
can be diagonalized, then
where
is an orthogonal matrix, and
is a diagonal matrix containing the eigenvalues of
. Given this diagonalization, we have
We point out that (
4) holds whenever
whatever
k. Furthermore, if
can be expressed as a Kronecker product,
with
then,
where
Moreover, if
is given by
we have
For instance, consider the case with independent vectors
,
with
invertible. Define
and the centered vectors
Following Kollo and Von Rosen [
10], we obtain the Wishart matrix
with parameters
where
is the non-centrality (stochastic) parameter.
Now, as the trace
tends to infinity, we have
indicating that
is the normalized form of
when scaled by its trace. Consequently,
where
denotes the vectorization of
, transforming the matrix
into a column vector by stacking its columns on top of each other. Here,
and
with
having its
components null except for the
v-th, which is 1, and
For more details, see Dhrymes [
11].
Given that
is orthogonal,
Suppose that, for instance,
Then,
since
Now,
depends on the choice of
and takes values in the range
. Assuming that
we have
as well as
Now,
with
gives
Thus, we may pick for
any value in the range
.
Assuming
we find that the limit density for
is
.
Now, since
is orthogonal, we have
Thus, instead of using
, we can work with
This form is easier to handle because it represents a product of independent normal distributions for each component
.
We are particularly interested in the distribution of the ratio
where the numerator represents the sum of squared components, and the denominator is the square of the sum of a specific subset of components, indexed by set
C,
The cumulative distribution function
of
Z is given by
where the integration domain
is defined as
These equations are useful because they allow us to characterize the distribution of the statistic
Z, which can be used to test hypotheses or make inferences about the underlying matrices
and
. The form of
indicates how likely it is to observe a given value of
Z, providing a basis for statistical decision-making.
5. Comparison of Sphericity Measures
In this section, we compare three measures of sphericity for positive semi-definite matrices: the proposed measure, given in (
2), the Frobenius measure, given in (
1), and the spectral measure defined as follows:
where
is the largest eigenvalue of
.
The ranges and interpretations for the three measures are as follows:
The proposed measure ranges between 0 and 1. Perfect sphericity is indicated by a value of 1.
The Frobenius measure ranges between and 1, where k is the size of the matrix. Perfect sphericity is indicated by .
The range of the spectral measure depends on the largest eigenvalue and the size of the matrix. For example, as we can see on
Table 1, if we have a
matrix with both eigenvalues 1, the perfect sphericity is indicated by
.
To compare the three measures of sphericity, we considered several cases. The following table contains the size of the matrix, the considered eigenvalues, and the three measures of sphericity for each case.
For matrices with equal eigenvalues (e.g., 1, 1 for ; 1, 1, 1 for ), the proposed measure is consistently 1, indicating perfect sphericity. The Frobenius and spectral measures decrease with increasing matrix size due to their dependence on the number of eigenvalues.
For matrices with one eigenvalue significantly different from the others (e.g., 2, 1 for ; 2, 1, 1 for ), the proposed measure decreases slightly from 1 but remains relatively high. The Frobenius and spectral measures show a more significant decrease compared to the perfectly spherical cases, reflecting the perturbation.
For matrices with highly varied eigenvalues (e.g., 10, 1 for ; 10, 5, 1 for ), the proposed measure decreases more substantially, reflecting the lack of sphericity. The Frobenius and spectral measures show higher values, indicating greater sensitivity to anisotropy.
The great advantage of the proposed measure is that it is scale-invariant, as discussed earlier, maintaining consistency regardless of the magnitude of the eigenvalues. The Frobenius and spectral measures do not explicitly account for scale, which can affect their interpretability across different matrix sizes.
In terms of computational complexity, the proposed measure is comparable to the Frobenius measure and the spectral measure. All measures involve operations that scale with the size k of the matrix, primarily dominated by the computation of norms, traces, and, in the case of the spectral measure, eigenvalues. Therefore, while the proposed measure offers unique normalization and interpretation advantages, its computational demands are not significantly different from the other measures discussed. This parity ensures that the proposed measure remains feasible for practical implementation across various matrix sizes commonly encountered in statistical and machine learning applications.
6. Final Remarks
The investigation into the measure of sphericity for positive semi-definite matrices yields valuable insights into their geometric properties, especially in high-dimensional contexts. The introduced measure, , provides a robust means of quantifying the spread of eigenvalues, being invariant under orthogonal transformations and scaling and exhibiting predictable behavior under matrix operations such as the Kronecker product.
In finite-dimensional cases, effectively discriminates between varying degrees of sphericity, with values ranging from 0 to 1. Furthermore, in the realm of stochastic matrices, the asymptotic normality of random matrices allows for insightful analyses. The convergence properties of and the behavior of under such convergence offer valuable tools for understanding the statistical properties of high-dimensional data.
The proposed sphericity measure presents several advantages:
Normalization to Unit Interval: Values closer to 1 indicate high sphericity, while values closer to 0 indicate low sphericity.
Scaling with Dimension: The normalization factor makes the measure more comparable across matrices of different sizes.
Proportional Interpretation: This provides a potentially more intuitive understanding of sphericity.
Invariant Properties: remains unchanged under orthogonal transformations and scaling, ensuring consistency across different bases and magnitudes.
Predictable Behavior under Operations: This exhibits predictable behavior under matrix operations such as the Kronecker product, making it a versatile tool in various applications.
These advantages highlight how compares favorably to other sphericity measures, providing more comprehensive and robust insights into the matrix properties.
The analysis primarily utilized normal distributions due to their mathematical tractability and the availability of well-established results in random matrix theory. However, the proposed measure is not limited to normal distributions. Future research could explore its application to other probability distributions, such as the Wishart distribution or distributions with heavy tails, like the multivariate t-distribution. Extending the analysis to these distributions may offer new insights into the behavior of the sphericity measure and its robustness under different statistical models. By considering different distributions, we can broaden the applicability of the sphericity measure and enhance its relevance in various practical scenarios where the normality assumption may not hold. Such extensions would significantly contribute to the understanding and utilization of sphericity measures in multivariate analysis and high-dimensional statistics.