Next Article in Journal
On a Symmetric, Nonlinear Birth-Death Process with Bimodal Transition Probabilities
Previous Article in Journal
Tetraquark Spectroscopy: A Symmetry Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Testing Group Symmetry of a Multivariate Distribution

by
Lyudmila Sakhanenko
Department of Statistics and Probability, Michigan State University, East Lansing, MI 48824-1027, USA
Symmetry 2009, 1(2), 180-200; https://doi.org/10.3390/sym1020180
Submission received: 15 September 2009 / Accepted: 24 November 2009 / Published: 26 November 2009

Abstract

:
We propose and study a general class of tests for group symmetry of a multivariate distribution, which encompasses different types of symmetry, such as ellipsoidal and permutation symmetries among others. Our approach is based on supremum norms of special empirical processes combined with bootstrap. We show that these tests are consistent against any fixed alternative. This work generalizes the methodology of Koltchinskii and Sakhanenko [7], developed for ellipsoidal symmetry to the case of group symmetry. It also provides a unified approach to testing different types of symmetry of a multivariate distribution.
Classification:
MSC 62G10, 60F05, 60F99

1. Introduction

Let S be a compact group of linear transformations (operators) from R d to R d . A Borel probability measure P on R d is called S -symmetric if and only if there exist an affine nonsingular transformation A 0 from R d onto R d and a S -invariant Borel probability measure Π 0 such that P = Π 0 A 0 . In other words, if X is a random vector with the distribution P , then there exists an affine nonsingular transformation A 0 such that the random vector Z = A 0 X is S -invariant, which in its turn means that Z = d S Z for all transformations S from the group S . Clearly, an affine transformation A 0 is not unique, as one could take instead S A 0 for an arbitrary transformation S from the group S . But, on the other hand, one could always fix A 0 by normalizing it in any proper way.
Obviously, we can define Π 0 = Π 0 , P as the distribution of the random variable Z = A 0 X . The couple ( A 0 , Π 0 ) will be called parameters (specifiers) of the S -symmetric probability measure P . We denote by E ( R d ) the set of all S -symmetric distributions on R d .
We are interested in the following problem. Given an independent identically distributed (i.i.d.) sample ( X 1 , , X n ) from the distribution P , defined on the probability space ( Ω , Σ , P ) , construct and study tests for S -symmetry of the distribution P.
Our group approach will unify the theory for different tests for different types of symmetry. Below we give some examples of common in literature types of symmetry S with corresponding choices of A 0 .
Example 1.
Let S = { S : S ( x 1 , , x d ) = ( ± x 1 , , ± x d ) } . Then S -invariant probability measure Π 0 is a so called sign change symmetrical measure. Probability measure P is S -symmetric if there exists an affine transformation A 0 in the form A 0 x = x θ 0 with some θ 0 R d such that P = Π 0 A 0 . In this case, P is often called diagonally or reflectively symmetrical measure.
Let | x | denote the Euclidean norm of a vector x R d . If E | X | < + , one can define θ 0 = θ 0 ( P ) : = E X = R d x P ( d x ) . Obviously, the parameter θ 0 and the transformation A 0 are uniquely defined for any P such that the corresponding X is integrable (not necessarily S -symmetric). We can also define in such a generality Π 0 as the distribution of the random variable Z = A 0 X .
Example 2.
Let S be the group of all orthogonal transformations in R d . Then S -symmetrical probability measure P is often called ellipsoidally symmetric or elliptically symmetric or elliptically contoured measure and Π 0 is spherically symmetric one. In this case, if E | X | 2 < + , one can define θ 0 = θ 0 ( P ) as the mean E X = R d x P ( d x ) and V 0 = V 0 ( P ) as the square root of the covariance operator of X , so that A 0 x = A 0 ( P ) x = V 0 1 ( x θ 0 ) for any x from R d . Define Π 0 as the distribution of the random variable Z = A 0 X . As in the previous case, these parameters are defined for any P such that the corresponding X is square-integrable (not necessarily S -symmetric).
We refer to the papers [1,2,3,4,5,6,7,8,9] for results on ellipsoidal and spherical symmetry testing.
Example 3.
Let d = 2 and let S k be a group of all transformations translating a regular polygon with k vertices centered at 0 into itself. Clearly, S k is a subgroup of the group of all orthogonal transformations. Thus, an affine transformation A 0 can be fixed in the same way as in example 2.
Example 4.
Let S be a group of all reflections about hyperplanes { x R d : x k = x l , } k , l = 1 , , d . Then for each S S there exists a permutation ( i 1 , , i d ) such that S ( x 1 , , x d ) = ( x i 1 , , x i d ) for all x R d . In this case S -invariant probability measure Π 0 is called permutation symmetric measure.
As in examples 2 and 3 we can define A 0 x = A 0 ( P ) x = V 0 1 ( x θ 0 ) for any x from R d , where θ 0 = θ 0 ( P ) : = E X = R d x P ( d x ) and V 0 = V 0 ( P ) is the square root of the covariance operator of X , provided E | X | 2 < + (also see [10,11]).
Tests for symmetry of a multivariate distribution play an important role in statistics and in various fields of science. To name a few, in finance theory log-returns of assets are assumed to be ellipsoidally symmetric. In genetics it is assumed that gene expression values are diagonally symmetrically distributed. In image analysis components are assumed to be spherically symmetric. In linear programming it is assumed that the distribution of feasible solutions is permutation symmetric. In statistics sliced inverse regression method due to Li, see [12], works for ellipsoidally symmetric distributions. Also since tests for normality are extended to tests for ellipsoidal symmetry, any research field that employs multivariate analysis based on normality assumption can benefit from relaxing this assumption to ellipsoidal symmetry assumption. So clearly symmetry tests are needed in applications. See [13] for a detailed survey on the use of symmetry in various scientific fields.
The rest of the paper is organized in the following way. We give notations and construct test statistics with examples in Section 2. Main results and bootstrapped test statistics are given in Section 3, followed by a detailed example in Section 4. The proofs are in Section 5. The closing remarks are in Section 6, followed by technical details in Appendix.

2. Notations and Preliminaries

Let m denote the uniform distribution (the normalized Haar measure) on the group S . Given a bounded Borel function f on R d , we define
m f ( y ) : = S f ( S y ) m ( d S ) , y R d .
It is easy to check that for a S -symmetric P distribution with specifiers ( A 0 , Π 0 ) we have
R d f ( A 0 x ) P ( d x ) = R d m f ( y ) Π 0 ( d y )
for any bounded Borel function f. Indeed, R d f ( z ) Π 0 ( d z ) = R d f ( S y ) Π 0 ( d y ) for any S S . Since S is a compact group, one can integrate the equality over S with respect to the uniform measure. Thus, R d S f ( S y ) m ( d S ) Π 0 ( d y ) = R d m f ( y ) Π 0 ( d y ) implies (2).
As a result, if a class F characterizes the distribution, i.e.
R d f d Q 1 = R d f d Q 2 for all f F
implies that Q 1 = Q 2 , then P is S -symmetric if and only if (2) holds for all f F . In general, we call P a F -asymmetric distribution if and only if there exists a function f F such that (2) does not hold. This observation is the key idea behind the tests that we construct and study. Naturally, a class F should be rich enough and possess good properties for further analysis. Let us describe it. Let F be a semialgebraic subgraph class as introduced in [7]. Basically, for a function from such a class, its subgraph can be constructed from a union of intersections of a finite number of subgraphs of polynomials of a finite degree in R d . The same should be true for a product of two functions from such a class. For instance, one can use polynomials of bounded degree or trigonometric functions of bounded frequency as F . The precise definition can be found in Appendix. We will provide a few examples later on.
Let P n be the empirical distribution based on the sample ( X 1 , , X n ) . Assume in what follows that A n = A n ( P ) is a n 1 / 2 -consistent estimator of A 0 = A 0 ( P ) . And, furthermore, there exists such a function γ A : R d R that E γ A 2 ( X ) < + and as n
n 1 / 2 ( A n A 0 ) = n 1 / 2 R d γ A ( x ) d ( P n P ) + o P ( 1 ) .
For example, assuming E | X | 2 < + let us define θ n and V n as
θ n : = θ n ( X 1 , , X n ) : = X ¯ n : = n 1 i = 1 n X i
and
V n 2 : = V n 2 ( X 1 , , X n ) : = n 1 i = 1 n ( X i X ¯ ) ( X i X ¯ ) T ,
where all vectors are columns and superscript T denotes transposition. Then one can define A n x = A n ( X 1 , , X n ) x = V n 1 ( x θ n ) in examples 2-4 and A n x = x θ n in example 1 for any x R d . Under the condition E | X | 4 < + , V n is a n 1 / 2 -consistent estimator of V 0 . Weaker moment assumptions on P can be imposed if other statistics are considered for estimation of θ 0 , V 0 , such as a sample median and an M-estimator for the covariance matrix. See, e.g., [14].
The scaled residuals of the observations ( X 1 , , X n ) are defined as
Z j : = Z j , n : = A n X j , j = 1 , n .
Let Π n denote the empirical distribution based on the sample ( Z 1 , , Z n ) .
Our approach to the problem of testing for S -symmetry will be to use the sup-norms of the stochastic process
ξ n ( f ) : = n 1 / 2 R d f ( A n x ) P n ( d x ) R d m f ( y ) d Π n ( y ) = n 1 / 2 j = 1 n f ( Z j ) m f ( Z j ) , f F
as test statistics
T n ( F ) : = sup f F | ξ n ( f ) | .
Such functionals can be viewed as “measures of asymmetry” of the empirical distribution because of the relationship (2).
Note that a nonsingular affine transformation of the data ( X 1 , , X n ) results in an orthogonal transformation of the scaled residuals. If a class F is invariant with respect to all orthogonal transformations (i.e. for all f F and any orthogonal transformation O we have f O F ), then the test statistic defined as the sup-norm of the process ξ n is affine invariant. This is the case in the following examples.
Example 1.1.
Consider S from example 1. Let
H : = { x R d : x , u c } : u S d 1 , c R
be the class of all half-spaces in R d , where S d 1 denotes the unit sphere in R d . Consider the class
F : = { I H ( x ) : H H } .
For f ( · ) = I H ( · ) , we have m f ( y ) = 1 2 d ε i = ± 1 I H ( ε 1 y 1 , , ε d y d ) = : 1 2 d ε i = ± 1 I H ( ε y ) , y R d . The process ξ n becomes
ξ n ( H ) = n 1 / 2 2 d j = 1 n ε i = ± 1 ( I H ( Z j ) I H ( ε Z j ) ) , H H .
The test statistic is represented as
T n ( F ) : = sup H H | ξ n ( H ) | = n 1 / 2 2 d sup H H j = 1 n ε i = ± 1 ( I H ( Z j ) I H ( ε Z j ) ) .
Example 1.2.
Consider S from example 1. Let H be the class of all half-spaces in R d as in the previous example. Let
f ( x ) = ε i = ± 1 ( 1 ) | ε | I ε H ( x )
for ε H = { x : x , ε u c } with | ε | = ε 1 + + ε d . Then we have m f ( y ) = 0 . The process ξ n becomes
ξ n ( H ) = n 1 / 2 j = 1 n ε i = ± 1 ( 1 ) | ε | I ε H ( Z j ) , H H
and the test statistic looks like
T n ( F ) = n 1 / 2 sup H H j = 1 n ε i = ± 1 ( 1 ) | ε | I ε H ( Z j ) .
In one-dimensional case d = 1 this test statistic becomes
T n : = n 1 / 2 sup c R j = 1 n ( I ( Z j c ) I ( Z j c ) ) .
One gets the expression that resembles a well known test for symmetry based on the empirical distribution function P n ,
T ˜ n = n 1 / 2 sup x 0 | n ( P n ( x ) + P n ( x ) 1 ) | .
See, for instance the discussion in the paper [15].
Example 2.1.
Consider S from example 2. Let
C : = { { v S d 1 : v , l c } : l S d 1 , c R + }
be the class of “caps” on the unit sphere S d 1 . Consider the class
F : = { I C ( x | x | ) I { 0 < | x | t } : C C , t > 0 } .
For f ( · ) = I C ( · | · | ) I { 0 < | · | t } , we have m f ( ρ ) = m ( C ) I ( 0 , t ] ( ρ ) , ρ > 0 . The process ξ n becomes now
ξ n ( C , t ) = n 1 / 2 1 n ( I C ( Z j | Z j | ) m ( C ) ) I { 0 < | Z j | t } , C C , t > 0 .
The test statistic T n ( F ) : = sup t > 0 , C C | ξ n ( C , t ) | can also be represented as
T n ( F ) = n 1 / 2 max 1 j n sup C C k = 1 j ( I C ( Z [ k ] | Z [ k ] | ) m ( C ) ) ,
where Z [ j ] , j = 1 , , n , is the rearrangement of Z 1 , , Z n , such that | Z [ 1 ] | | Z [ n ] | . These tests were studied in [7] and [9].
Example 2.2.
Consider S from example 2. Let G l denote the linear space of spherical harmonics of degree less than or equal to l in R d , and let B l be the unit ball in G l L 2 ( S d 1 , d m ) . Denote
F : = { I { 0 < | x | t } ψ ( x | x | ) : ψ B l , t > 0 } .
Then for f F , we have f ( x ) = I { 0 < | x | t } ψ ( x | x | ) , x R d , and m f ( ρ ) = m ( ψ ) I ( 0 , t ] ( ρ ) , where m ( ψ ) is the average of ψ on S d 1 . In this case, the process ξ n becomes
ξ n ( t , ψ ) = n 1 / 2 1 n I { 0 < | Z j | t } ( ψ ( Z j | Z j | ) m ( ψ ) ) , t > 0 , ψ B l .
The statistic T n ( F ) : = sup t > 0 , ψ B l | ξ n ( t ; ψ ) | becomes
T n ( F ) = n 1 / 2 max 1 j n s = 1 dim ( G l ) k = 1 j ψ s ( Z [ k ] | Z [ k ] | ) δ s 1 ) 2 1 / 2 ,
where { ψ s , s = 1 , , dim ( G l ) } denotes an orthonormal basis of the space G l , ψ 1 1 , δ i j = 1 for i = j and 0 otherwise, and dim of a set denotes the number of elements of the set. These tests were studied in [7] and [9], where their superiority in level preservation and power performance over other tests both theoretically and in a simulation study, was shown. A similar approach was used to test for multivariate normality in [16]. The authors of [6] developed a different kind of tests for ellipsoidal symmetry based on spherical harmonics.
Example 2.3.
Consider S from example 2. Let H be the class of all half-spaces in R d as in the example 1.1. For f : = I H , where H : = { x R d : x , u c } , we have m f ( ρ ) : = γ ( c ρ ) , where γ ( c ) : = m { v : v , u c } . The process ξ n in this case is
ξ n ( u , c ) : = n 1 / 2 j = 1 n I { Z j , u c } γ ( c | Z j | ) , u S d 1 , c R
and the test statistic can be defined as
T n : = sup c R sup u S d 1 | ξ n ( u , c ) | .
This type of test statistics was systematically studied in papers [7,9,10].
Example 2.4.
Consider S from example 2. Let
F : = | · t | | · | | t | : t R d , t 0 .
For f ( · ) = | · t | | · | | t | , we have m f ( ρ ) = ρ β ( t ) 1 | t | , where β ( t ) : = S d 1 | v t | m ( d v ) . Thus, the process ξ n becomes
ξ n ( t ) : = n 1 / 2 | t | 1 i = 1 n | Z i t | | Z i | β ( t ) , t R d
and the test statistic can be chosen as
T n ( F ) : = sup t R d , t 0 | ξ n ( t ) | .
Example 2.5.
Consider S from example 2. Consider the class
F : = { e i t , x , x R d : | t | 1 } .
For f ( · ) = e i t , · we have m f ( ρ ) = c d J d / 2 1 ( ρ | t | ) ( ρ | t | ) d / 2 1 , ρ > 0 , where J l denotes the Bessel function of the l-th order, the constant c d depends only on d. The process ξ n becomes
ξ n ( t ) = n 1 / 2 j = 1 n exp { i t , Z j } c d J d / 2 1 ( | Z j | | t | ) ( | Z j | | t | ) d / 2 1 , | t | 1
and the test statistic can be chosen as
T n : = sup | t | 1 | ξ n ( t ) | .
Consider S from example 3. Due to similarity between examples 2 and 3, one can choose the same classes of functions for S k from example 3. We give just one of the examples as an illustration.
Example 3.1.
Consider the class F from example 2.1. Then for f ( · ) = I C ( · | · | ) I { 0 < | · | t } , we have m f ( y ) = 1 k l = 0 k 1 I C ( O l y / | y | ) I ( 0 , t ] ( | y | ) for all y R d , where O l is the rotation on angle 2 π l / k , l = 0 , , k 1 . In this case, the process ξ n is
ξ n ( C , t ) = n 1 / 2 j = 1 n ( I C ( Z j | Z j | ) 1 k l = 0 k 1 I C ( O l Z j / | Z j | ) ) I { 0 < | Z j | t } , C C , t > 0 .
The test statistic T n : = sup t > 0 , C C | ξ n ( C , t ) | can be also represented as
T n ( F ( = n 1 / 2 max 1 j n sup C C m = 1 j ( I C ( Z [ m ] | Z [ m ] | ) 1 k l = 0 k 1 I C ( O l Z [ m ] / | Z [ m ] | ) ) ,
where Z [ j ] , j = 1 , , n is the rearrangement of Z 1 , , Z n , such that | Z [ 1 ] | | Z [ n ] | .
Example 4.1.
Consider S from example 4. Let H be the class of all half-spaces in R d as in the example 1.1. Denote
F : = { I H ( y ) I ( y j 1 y j k ) : H H , ( j 1 , , j k ) is any combination from ( 1 , , d ) } .
Then for f F such that f ( y ) : = I H ( y ) I ( y j 1 y j k ) for y R d we have
m f ( y ) = m H , i 1 , . . . , i k ( y ) = 1 d ! ( i 1 , . . . , i d ) I H ( ( y i 1 , . . . , y i d ) ) I ( y i j 1 y i j k ) ,
where the summation is over all permutations ( i 1 , , i d ) of ( 1 , , d ) . In this case the process ξ n is
ξ n ( H , i 1 , , i k ) : = n 1 / 2 j = 1 n ( I H ( Z j ) I ( ( Z j ) j 1 . . . ( Z j ) j k ) m H , i 1 , . . . , i k ( Z j ) ) ,
H H , ( j 1 , , j k ) is any combination from ( 1 , , d )
and the test statistic can be defined as
T n : = sup H H sup 2 k d sup ( j 1 , . . . , j k ) | ξ n ( H , j 1 , . . . , j k ) | ,
where the last supremum is taken over all combinations ( j 1 , , j k ) out of ( 1 , , d ) . Well known and frequently used Friedman’s rank tests are based on the similar choice of a class F . For reference see the papers [17] and [11].
It is not hard to see that the function classes defined in examples 1.1, 1.2, 2.1–2.4, 3.1, 3.2, and 4.1 are semialgebraic subgraph. In addition, classes F characterize the distribution in the case of examples 1.1, 2.1, 2.3, 2.4, 3.1, 4.1 above.
We say the class of transformations S preserves the semialgebraic property if for any polynomial p on R d of degree less than or equal to r the set { ( x , t ) R d + 1 : p ( S x , t ) 0 } belongs to S A q , d + 1 , l for some q and l (see Appendix for the definition). Classes S , defined in examples 1–4, preserve the semialgebraic property.
Let
E ( f ; A ) : = R d f ( A x ) m f ( A x ) P ( d x ) .
It follows from (2) that, for a S -symmetric distribution P and for all f
E ( f ; A 0 ) = 0 .
Let W P denote the P-Brownian bridge, i.e. a centered Gaussian process indexed by functions in L 2 ( R d ; d P ) with the covariance
E W P ( f ) W P ( g ) = P ( f g ) P ( f ) P ( g ) .
We will frequently use integral notation for W P ( f ) :
W P ( f ) = R d f ( x ) W P ( d x ) .
As always, ( F ) denotes the space of all uniformly bounded functions on F with the sup-norm Y F : = sup f F | Y ( f ) | , Y ( F ) .
A sequence of stochastic processes ζ n : F R is said to converge weakly in ( F ) (in the sense of Hoffmann-Jørgensen) to the stochastic process ζ : F R if and only if there exists a Radon probability measure γ on ( F ) such that γ is the distribution of ζ and, for all bounded and · F -continuous functionals Φ : ( F ) R , we have E * Φ ( ζ n ) Φ ( x ) γ ( d x ) , where E * stands for the outer expectation, which is defined as E * Ψ = inf { E U : U Ψ , Ψ : Ω [ , ] measurable and E U exists } for a Ψ : Ω [ , ] . See for instance [18].
We assume in what follows that the class F satisfies standard measurability assumptions used in the theory of empirical processes (see [19] or [18]). We also need smoothness conditions (S) on P and F , which are given in Appendix.

3. Main Results

Theorem 1
Suppose that F is a semialgebraic subgraph class, the smoothness conditions (S) hold and E γ A 2 ( X ) < + . Define a Gaussian stochastic process
ξ P ( f ) : = W P ( f ( A 0 · ) m f ( A 0 · ) ) + E A ( f ; A 0 ) ( R d γ A ( x ) W P ( d x ) ) ,
whose distribution is a Radon measure in ( F ) . Then the sequence of stochastic processes
{ ξ n ( f ) n 1 / 2 E ( f ; A 0 ) : f F }
converges weakly in the space ( F ) to the process ξ P . In particular, if P is S -symmetric with specifiers ( A 0 , Π 0 ) , then the sequence ξ n converges weakly in the space ( F ) to the process ξ P .
Define the test statistics
T n : = ξ n F .
Given α > 0 , let
t α : = inf t : P { ξ P F t } α .
Let H 0 be the hypothesis that P E ( R d ) and let H a be the alternative that P E ( R d ) . Also, denote by H a ( F ) the alternative that P is F -asymmetric.
Theorem 1 and the well-known theorem of Cirel’son on continuity of the distribution of the sup-norm of Gaussian processes, see [20], imply the following.
Corollary 1
Suppose all conditions of Theorem 1 hold. Under the hypothesis H 0
P { T n t α } α
and under the alternative H a ( F )
P { T n t α } 1 as n .
In particular, if F characterizes the distribution, then under the alternative H a , i.e. for a fixed S -asymmetric distribution P,
P { T n t α } 1 as n .
In most cases, however, the limit distributions of such statistics as T n depend on the unknown parameters of the distribution P . Thus, to implement the test one has to evaluate the distribution of the test statistic using, for instance, a bootstrap method. We describe below a version of the conditional bootstrap for S -symmetry testing. It is a generalization of the bootstrap method proposed in [7].
Given P , let P s denote the S -symmetric distribution with specifiers ( A 0 , Π 0 ) . It will be called the S -symmetrization of P . Denote by P n s the S -symmetric distribution with specifiers ( A n , Π n ) . Let ( X 1 s , …, X n s ) be an i.i.d. sample from the distribution P n s , defined on a probability space ( Ω ^ , Σ ^ , P ^ ) . One can construct such a sample using the following procedure. Take an i.i.d. sample ( Y 1 , , Y n ) from Π n , which is a resampling from ( Z 1 , , Z n ) . Define
X j s : = A n 1 Y j , j = 1 , , n .
Then conditionally on ( X 1 , , X n ) , ( X 1 s , , X n s ) is an i.i.d. sample from the S -symmetric distribution P n s .
In particular, for S from example 1
X j s : = θ n + ε j Y j , j = 1 , , n ,
where ( ε 1 , , ε n ) is a Rademacher i.i.d. sample, that is ε j = 1 or 1 with probability 1/2, j = 1 , , n , independent of ( Y 1 , , Y n ) .
For S from example 2 one can take an i.i.d. sample ( U 1 , , U n ) uniformly distributed on S d 1 and an i.i.d. sample ( Y 1 , , Y n ) from Π ˜ n , the empirical distribution based on ( | Z 1 | , , | Z n | ) , independent of ( U 1 , , U n ) . In other words, ( Y 1 , , Y n ) is the resampling from the sample ( | Z 1 | , , | Z n | ) . Then
X j s : = θ n + V n U j Y j , j = 1 , , n .
For S from example 3 let ( ε 1 , , ε n ) be an i.i.d. sample uniformly distributed on { 0 , 1 , , k 1 } independent of ( Y 1 , , Y n ) , then
X j s : = θ n + V n O j Y j , j = 1 , , n ,
where O j is a rotation on the angle 2 π ε j k about 0.
Finally, for S from example 4 consider n independent permutations { i 1 ( j ) , , i d ( j ) } of ( 1 , , d ) , j = 1 , , n , independent of ( Y 1 , , Y n ) . Then
X j s : = θ n + R j Y j , j = 1 , , n ,
where R j is a reflection transformation such that R j ( x 1 , , x d ) = ( x i 1 ( j ) , , x i d ( j ) ) for x R d .
Let P ^ n denote the empirical measure based on the sample ( X 1 s , , X n s ) , and let
A ^ n : = A n ( X 1 s , , X n s ) .
Define the bootstrapped scaled residuals as
Z ^ j : = Z ^ j , n : = A ^ n X j s , j = 1 , , n .
Let Π ^ n denote the empirical distribution based on the sample ( Z ^ 1 , , Z ^ n ) .
The bootstrap version of ξ n is the process
ξ ^ n ( f ) : = n 1 / 2 ( R d f ( A ^ n x ) P ^ n ( d x ) R d m f ( y ) d Π ^ n ( y ) ) = n 1 / 2 j = 1 n f ( Z ^ j ) m f ( Z ^ j ) , f F .
Let B L 1 ( ( F ) ) denote the set of all functionals Φ : ( F ) R such that | Φ ( Y ) | 1 for all Y ( F ) and | Φ ( Y 1 ) Φ ( Y 2 ) | Y 1 Y 2 F for all Y 1 , Y 2 ( F ) . Given two stochastic processes ζ 1 , ζ 2 : Ω × Ω ^ × F R , we define the following bounded Lipschitz distance:
d B L ( ζ 1 , ζ 2 ) : = sup Φ B L 1 ( ( F ) ) | E ^ * Φ ( ζ 1 ) E ^ * Φ ( ζ 2 ) | ,
where E * denotes the outer expectation.
Now we are going to consider a bootstrap version of Theorem 1.
Theorem 2
Suppose that F is a semialgebraic subgraph class, the smoothness conditions (S) hold and E γ A 2 ( X ) < . Then the sequence of stochastic processes { ξ ^ n } converges weakly in the space ( F ) to a version ξ ^ P s of the process ξ P s (defined on the probability space ( Ω ^ , Σ ^ , P ^ ) ) in probability P . More precisely,
d B L ( ξ ^ n ; ξ ^ P s ) 0 as n in probability P .
In particular, if P is S -symmetric, ξ ^ n converges weakly to a version of the process ξ P .
Define test statistics
T ^ n : = ξ ^ n F .
Given α > 0 , let
t ^ n , α : = inf t : P ^ { T ^ n t } α .
In other words, t ^ n , α is a ( 1 α ) -quantile of the distribution of T ^ n conditional on the sample ( X 1 , , X n ) .
Then Theorems 1 and 2 imply the following.
Corollary 2
Suppose all the conditions of Theorems 1 and 2 hold. Under the hypothesis H 0
P { T n t ^ n , α } α
and under the alternative H a ( F )
P { T n t ^ n , α } 1 as n .
In particular, if F characterizes the distribution, the bootstrap test is consistent against any asymmetric alternative (subject to the smoothness conditions (S)): under the alternative H a
P { T n t ^ n , α } 1 as n .
Thus, our method provides tests that are consistent against any S -asymmetric alternative.

4. Detailed example

In this section we provide an example for which we verify all the assumptions and supply a step-by-step computational algorithm. Let d = 2 and consider the problem of testing whether P is elliptically contoured measure (Example 2). For a vector x R 2 let ( r , φ ) [ 0 , ) × [ 0 , 2 π ) be its polar coordinates. For a fixed integer l let
F : = lin ( { I { 0 < r t } cos ( k φ ) , I { 0 < | x | t } sin ( k φ ) : 1 k l , t > 0 } ) ,
where lin ( G ) denotes the linear span of G with all the functions bounded by 1, see Example 2.2. This class F satisfies the following assumptions.
1. It characterizes distribution only for l = . For a finite l it does not characterize the distribution, since one might find two different distributions Q 1 , 2 such that
0 t 0 2 π cos ( k φ ) d Q 1 ( r cos φ , r sin φ ) = 0 t 0 2 π cos ( k φ ) d Q 2 ( r cos φ , r sin φ )
and
0 t 0 2 π sin ( k φ ) d Q 1 ( r cos φ , r sin φ ) = 0 t 0 2 π sin ( k φ ) d Q 2 ( r cos φ , r sin φ )
for all t > 0 , 1 k l .
2. F is a semialgebraic subgraph class. Indeed, for ( x 1 , x 2 ) the sets
{ ( x 1 , x 2 , s ) : I ( 0 < x 1 2 + x 2 2 t 2 ) cos ( k arctan ( x 2 / x 1 ) ) s 0 or I ( 0 < x 1 2 + x 2 2 t 2 ) cos ( k arctan ( x 2 / x 1 ) ) s 0 }
can be represented as unions of finite number of intersections of polynomial sets of finite degree. For instance, for k = 1 we have have the following representation
{ 1 s 2 0 , t 2 x 1 2 x 2 2 0 , s 0 , x 1 2 s 2 ( x 1 2 + x 2 2 ) 0 } { 1 s 2 0 , t 2 x 1 2 x 2 2 0 , s 0 , s 2 ( x 1 2 + x 2 2 ) x 1 2 0 } .
The representations for any 1 k l can be obtained similarly using trigonometric identities. Obviously, similar arguments work for sines, linear combinations of sines and cosines and products of any two functions from F .
3. F is invariant with respect to all orthogonal transformations, which are rotations on the unit circle. Indeed, for any rotation on an angle τ a vector ( r , φ ) is transformed into the vector ( r , φ + τ ) . So for any t > 0 , 1 k l we have f ( r , φ + τ ) = I { 0 < r t } cos ( k ( φ + τ ) ) or f ( r , φ + τ ) = I { 0 < r t } sin ( k ( φ + τ ) ) , where both functions belong to the linear span of F and are bounded by 1. So any linear combination of such functions, which is bounded by 1, would also lie in F .
4. Condition (S2) holds. Indeed, F is uniformly bounded by 1. For any δ > 0 and any f F we have ω f ( · ; δ ) = C δ for some constant C > 0 , so that the measure of the set defined in (S2) is zero for δ < ε .
Also note that the group S of all orthonormal transformations of R 2 preserves semialgebraic property. Indeed, for any rotation on an angle τ the sets
{ ( x 1 , x 2 , s ) : I ( 0 < x 1 2 + x 2 2 t 2 ) cos ( k arctan ( x 2 / x 1 ) + τ ) s 0 or I ( 0 < x 1 2 + x 2 2 t 2 ) cos ( k arctan ( x 2 / x 1 ) + τ ) s 0 }
are semialgebraic. The same holds for sines and linear combinations of sines and cosines.
We also require the following two conditions on P: (S1) holds and E γ A 2 ( X ) < . The first condition is satisfied for absolutely continuous P with the uniformly bounded and continuously differentiable Lebesgue density with the corresponding derivative approaching zero at infinity faster than | x | 3 , x R 2 . For example, distributions with densities on a finite support and normal distributions satisfy (S1). The last condition is satisfied if E | X | 4 < .
Given a random sample ( X 1 , , X n ) from a distribution P let us describe a step-by-step testing algorithm.
1. Obtain θ n , V n . In our example θ n is the sample mean and V n is the square root of the sample covariance for the sample ( X 1 , , X n ) .
2. Calculate residuals Z j = V n 1 ( X j θ n ) , j = 1 , , n .
3. Find the test statistics, which can be simplified as follows
T n ( F ) = n 1 / 2 max 1 j n k = 1 l i = 1 j cos ( k φ i ) 2 + i = 1 j sin ( k φ i ) 2 1 / 2 ,
where φ i is the polar coordinate of Z i .
4. Choose a number of bootstrap repetitions, say M. On practice we often take a large number, for instance M = 10000 . Then the next four steps are repeated M times.
4.1. Generate a sample ( X 1 s , , X n s ) . In this example, first, generate a sample ( U 1 , , U n ) from a uniform distribution on the unit circle, independent of ( | Z 1 | , , | Z n | ) . Secondly, resample with replacement from ( | Z 1 | , , | Z n | ) to obtain ( R 1 , , R n ) . Thirdly, X j s = θ n + V n R j U j , j = 1 , , n .
4.2. Obtain θ ^ n , V ^ n . In our example θ ^ n is the sample mean and V ^ n is the square root of the sample covariance for the sample ( X 1 s , , X n s ) .
4.3. Calculate residuals Z ^ j = V ^ n 1 ( X j s θ ^ n ) , j = 1 , , n .
4.4. Find the bootstrapped test statistics, which can be simplified as follows
T ^ n ( F ) = n 1 / 2 max 1 j n k = 1 l i = 1 j cos ( k φ ^ i ) 2 + i = 1 j sin ( k φ ^ i ) 2 1 / 2 ,
where φ ^ i is the polar coordinate of Z ^ i .
5. Based on ( T ^ n ( 1 ) , , T ^ n ( M ) ) find the empirical ( 1 α ) -quantile of the distribution of T ^ n ( F ) , conditional on ( X 1 , , X n ) . Let us denote it as t ^ n , α , M .
6. If T n ( F ) t ^ n , α , M then reject H 0 : P is elliptically contoured distribution, at the significance level α .

5. Proofs

We use ideas and methods of the work [7]. Their technique was developed for ellipsoidal symmetry and is needed to be adjusted for group symmetry. Basically one should change V 1 ( x θ ) to A x throughout the proofs. However, there are technical difficulties associated with using transformations A instead of ( θ , V ) , they are hidden in the proofs of lemmas. We give a few details for completeness.
Let SP denote a subset of all nonsingular linear transformations in R d . Given a transformation A SP denote τ A f ( · ) : = f ( A · ) . For a function f on R d , let
f ˜ ( x ) : = f ( x ) m f ( x ) , x R d .
Given a class G of functions on R d , define G ˜ : = { g ˜ : g G } ,
G aff : = { τ A g : A SP } .
Now the process ξ n is represented as
ξ n ( f ) = n 1 / 2 R d τ A n f ˜ d P n , f F
and the process ξ ^ n as
ξ ^ n ( f ) = n 1 / 2 R d τ A ^ n f ˜ d P ^ n , f F .
Clearly,
E ( f ; A ) = R d τ A f ˜ d P .
Define
E s ( f ; A ) = R d τ A f ˜ d P s .
Given a function g on R d , we can write
R d g ( x ) P n s ( d x ) = R d S g ( A n 1 S y ) m ( d S ) Π n ( d y ) = R d M g ( A n ; x ) P n ( d x ) ,
where
M g ( A ; x ) : = S g ( A 1 S A x ) m ( d S ) .
A similar computation shows that
R d g ( x ) P s ( d x ) = R d M g ( A 0 ; x ) P ( d x ) = R d M g ( A 0 ; x ) P s ( d x ) .
Let
Γ ( g ; A ) : = R d M g ( A ; x ) P ( d x ) .
Given a class G of functions, define
G 2 : = { g h : g G , h G } ,
M ( G ) : = { M g ( A , · ) : g G , A SP } .
We reformulate the following versions of lemmas from [7] that describe smoothness properties of the functions introduced above and Donsker properties of the classes of functions given above. The convergence of transformations is with respect to the operator norm on the set of all linear transformations. Smoothness conditions are used in the proof of Lemma 1. Properties of Vapnik-Chervonenkis subgraph classes are used in the proof of Lemma 2. See [18] for details on Vapnik-Chervonenkis, Glivenko-Cantelli, and Donsker classes of functions.
Lemma 1
Suppose that P and F satisfy the smoothness conditions (S). Then the following statements hold:
(C1) If A A 0 then
sup f F R d | f ˜ A f ˜ A 0 | 2 P ( d x ) 0
and
sup f F R d | f ˜ A f ˜ A 0 | 2 P s ( d x ) 0 .
(C2) The function E ( f ; A ) is differentiable at the point A 0 for any f F , and the Taylor expansion of the first order
E ( f ; A ) = E ( f ; A 0 ) + E A ( f ; A 0 ) ( A A 0 ) + o ( | A A 0 | )
holds uniformly in f F .
(C3) Similarly, the function E s ( f ; A ) is differentiable at the point A 0 for any f F , and the Taylor expansion of the first order
E s ( f ; A ) = E s ( f ; A 0 ) + ( E s ) A ( f ; A 0 ) ( A A 0 ) + o ( | A A 0 | )
holds uniformly in f F .
(C4) The function Γ ( g ; A ) is continuous with respect to A at A 0 uniformly in g ( ( F ˜ ) aff ) 2 .
(C5) The function Γ ( g ; A ) is differentiable at the point A 0 for any g ( F ˜ ) aff , and, moreover, the Taylor expansion of the first order
Γ ( g ; A ) = Γ ( g ; A 0 ) + Γ A ( g ; A 0 ) ( A A 0 ) + o ( | A A 0 | )
holds uniformly in g ( F ˜ ) aff . Moreover, the matrix-valued function A Γ A ( τ A f ˜ ; A 0 ) is continuous at A 0 uniformly in f F .
(C6) if A A 0 , then for all δ > 0
sup R d | M τ B f ˜ ( A ; x ) M τ B f ˜ ( A 0 ; x ) | 2 P ( d x ) : B SP , | B A 0 | δ 0
and
sup f F R d | M τ A f ˜ ( A 0 ; x ) M τ A 0 f ˜ ( A 0 ; x ) | 2 P ( d x ) 0 .
Lemma 2
For a uniformly bounded semialgebraic subgraph class F , the classes ( F ˜ ) aff , M ( ( F ˜ ) aff ) and M ( ( F ˜ ) aff 2 ) are uniformly Donsker and uniformly Glivenko–Cantelli.
Proof of Theorem 1.
Define a process
η n ( f ; A ) : = n 1 / 2 ( P n P ) ( τ A f ˜ ) , f F , A SP .
(C1) and ( F ˜ ) aff being a P-Donsker class by Lemma 2 imply that we can use asymptotic equicontinuity to obtain
lim δ 0 lim sup n P * ( { sup | A A 0 | δ sup f F | η n ( f ; A ) η n ( f ; A 0 ) | ε } ) = 0
for all ε > 0 . Clearly,
ξ n ( f ) n 1 / 2 E ( f ; A 0 ) = η n ( f ; A n ) + n 1 / 2 E ( f ; A n ) E ( f ; A 0 ) = η n ( f ; A 0 ) + n 1 / 2 ( E ( f ; A n ) E ( f ; A 0 ) ) + ( η n ( f ; A n ) η n ( f ; A 0 ) ) .
If | A n A 0 | δ , we have
sup f F | ξ n ( f ) n 1 / 2 E ( f ; A 0 ) η n ( f ; A 0 ) n 1 / 2 ( E ( f ; A n ) E ( f ; A 0 ) ) | sup | A A 0 | δ sup f F | η n ( f ; A ) η n ( f ; A 0 ) | .
Note that
η n ( f ; A 0 ) = n 1 / 2 R d τ A 0 f ˜ d ( P n P ) .
Using (5) and n 1 / 2 -consistency of A n we obtain
ξ n ( f ) n 1 / 2 E ( f ; A 0 ) = n 1 / 2 R d τ A 0 f ˜ d ( P n P ) + n 1 / 2 ( E ( f ; A n ) E ( f ; A 0 ) ) + o p ( 1 )
as n uniformly in f F . It follows from (C1) and n 1 / 2 -consistency of A n that
n 1 / 2 ( E ( f ; A n ) E ( f ; A 0 ) ) = n 1 / 2 E A ( f , A 0 ) ( A n A 0 ) + o p ( 1 ) , n
uniformly in f F . Representations (6) and (7), the fact that ( F ˜ ) aff is a uniformly Donsker class from Lemma 2 and (C1) imply that the sequence ξ n ( f ) n 1 / 2 E ( f ; A 0 ) converges weakly in the space ( F ) to the Gaussian stochastic process ξ P . This implies the first statement of the theorem. If P is ellipsoidally symmetric then E ( f ; A 0 ) = 0 , which concludes the proof of Theorem 1.
Proof of Theorem 2.
Define a process
η ^ n ( f ; A ) : = n 1 / 2 ( P ^ n P n s ) ( τ A f ˜ ) , f F , A SP .
By Lemma 2, the class M ( ( ( F ˜ ) aff ) 2 ) is uniformly Glivenko–Cantelli. This together with (C4) and representations (3), (4) implies that
R d g h d P n s R d g h d P s as n a . s .
uniformly in g , h ( F ˜ ) aff . Similarly, since the class M ( ( F ˜ ) aff ) is uniformly Glivenko–Cantelli, by (C5) and representations (3), (4), we obtain
R d g d P n s R d g d P s as n a . s .
uniformly in g ( F ˜ ) aff .
Since ( F ˜ ) aff is a uniformly Donsker class, we can use Corollary 2.7 in [21] to prove that a.s. n 1 / 2 ( P ^ n P n s ) converges weakly in the space ( ( F ˜ ) aff ) to the same limit as n 1 / 2 ( P ˜ n P s ) , where P ˜ n is the empirical measure based on a sample from P s , i.e. to the P s -Brownian bridge W P s . Asymptotic equicontinuity and (C1) yield that for all ε > 0 P a.s.
lim δ 0 lim sup n P ^ * ( { sup | A A n | δ sup f F | η ^ n ( f ; A ) η ^ n ( f ; A n ) | ε } ) = 0 .
Define
E ^ n ( f ; A ) : = R d τ A f ˜ d P n s .
Since E ^ n ( f ; A n ) = 0 , we can write
ξ ^ n ( f ) = η ^ n ( f ; A ^ n ) + n 1 / 2 E ^ n ( f ; A ^ n ) = η ^ n ( f ; A n ) + n 1 / 2 ( E ^ n ( f ; A ^ n ) E ^ n ( f ; A n ) ) + ( η ^ n ( f ; A ^ n ) η ^ n ( f ; A n ) ) .
If | A ^ n A n | δ , we have
sup f F | ξ ^ n ( f ) η ^ n ( f ; A n ) n 1 / 2 ( E ^ n ( f ; A ^ n ) E ^ n ( f ; A n ) ) | sup | A A n | δ sup f F | η ^ n ( f ; A ) η ^ n ( f ; A n ) | .
Note that
η ^ n ( f ; A n ) = n 1 / 2 R d τ A n f ˜ d ( P ^ n P n s ) .
Using (8), (9) and standard asymptotic properties of the estimators A n we obtain
ξ ^ n ( f ) = n 1 / 2 R d τ A n f ˜ d ( P ^ n P n s ) + n 1 / 2 ( E ^ n ( f ; A ^ n ) E ^ n ( f ; A n ) ) + o p ( 1 )
as n uniformly in f F . Here and in what follows the remainder term o p ( 1 ) converges to 0 as n uniformly in f F in probability P × P ^ .
Applying the asymptotic equicontinuity condition to the process n 1 / 2 ( P ^ n P n s ) and using (C1), we obtain
n 1 / 2 R d τ A n f ˜ d ( P ^ n P n s ) = n 1 / 2 R d τ A 0 f ˜ d ( P ^ n P n s ) + o p ( 1 )
as n . Now we can write
n 1 / 2 ( E ^ n ( f ; A ^ n ) E ^ n ( f ; A n ) ) = n 1 / 2 ( E ^ n ( f ; A ^ n ) E s ( f ; A ^ n ) ) n 1 / 2 ( E ^ n ( f ; A n ) E s ( f ; A n ) ) + n 1 / 2 ( E s ( f ; A ^ n ) E s ( f ; A n ) ) .
Note that by (3) and (4)
n 1 / 2 ( E ^ n ( f ; A ) E s ( f ; A ) ) = n 1 / 2 ( P n P ) ( M τ A f ˜ ( A n ; · ) ) + n 1 / 2 ( Γ ( τ A f ˜ ; A n ) Γ ( τ A f ˜ ; A 0 ) ) = n 1 / 2 ( P n P ) ( M τ A f ˜ ( A 0 ; · ) ) + n 1 / 2 ( P n P ) ( M τ A f ˜ ( A n ; · ) ) n 1 / 2 ( P n P ) ( M τ A f ˜ ( A 0 ; · ) ) + n 1 / 2 ( Γ ( τ A f ˜ ; A n ) Γ ( τ A f ˜ ; A 0 ) ) .
Since, by Lemma 2, M ( ( F ˜ ) aff ) is uniformly Donsker class and since (C5) and (C6) hold, it is easy to prove the weak convergence of the processes
n 1 / 2 ( E ^ n ( f ; A ) E s ( f ; A ) ) , f F , A B ( A 0 )
in the space ( F × B ( A 0 ) ) , where B ( A 0 ) is a ball in SP with the center A 0 . Using the asymptotic equicontinuity and (C6), we obtain
n 1 / 2 ( E ^ n ( f ; A ^ n ) E s ( f ; A ^ n ) ) n 1 / 2 ( E ^ n ( f ; A n ) E s ( f ; A n ) ) = o p ( 1 ) as n .
It follows from (C3) and standard asymptotic properties of the estimators A n , A ^ n that
n 1 / 2 ( E s ( f ; A ^ n ) E s ( f ; A n ) ) = n 1 / 2 ( E s ) A ( f , A 0 ) ( A ^ n A n ) + o p ( 1 ) = ( E s ) A ( f , A 0 ) n 1 / 2 R d γ A ( x ) d ( P ^ n P n s ) + o p ( 1 ) , n ,
uniformly in f F . Relationships (10)–(14) along with, again, Corollary 2.7 in [21], imply the statement of the theorem.

6. Conclusion

We propose and study a general class of tests for group symmetry, which encompasses different types of symmetry, such as ellipsoidal and permutation symmetries. Our approach is based on supremum norms of special empirical processes combined with bootstrap.
There are several advantages to our methodology. First, the test statistics are indexed by classes of functions that are rich enough and still relatively simple to use. This provides some flexibility in choosing a suitable class of functions, thereby giving an appropriate test. Secondly, these tests are consistent against all possible asymmetric alternatives. Thirdly, they enjoy the property of affine invariance. Fourth, these are bootstrap tests, which could be considered as a drawback but it is a way to deal with complex nature of asymptotic null distribution of a non-bootstrap semiparametric test, and these tests have good theoretical properties. Fifth, this approach gathers separate ideas and methods developed for various types of symmetry under one umbrella. It provides a unified theory for studying statistical properties of seemingly different tests for different types of symmetry.

7. Appendix

Definition of a semialgebraic set. For any polynomial p on R m of degree less than or equal to r we will call { x R m : p ( x ) 0 } a polynomial set of degree less than or equal to r in R m . Let P r , m denote the class of all polynomial sets in R m of degree less than or equal to r. Then any set from the union { A ( B 1 , , B l ) : B 1 , , B l P r , m } is called a semialgebraic set of degree less than or equal to r and order less than or equal to l, A ( B 1 , , B l ) being the minimal set algebra generated by B 1 , , B l . Let S A r , m , l denote the class of all semialgebraic sets of degree less than or equal to r and order less than or equal to l in R m .
A class G of functions on R d is a semialgebraic subgraph class if and only if for some r , l for all functions g from G the set { ( x , t ) : g ( x ) t 0 or g ( x ) t 0 } belongs to S A r , d + 1 , l and for all functions g 1 , g 2 from G the set { ( x , y , t ) : g 1 ( x ) g 2 ( y ) t 0 or g 1 ( x ) g 2 ( y ) t 0 } belongs to S A r , 2 d + 1 , l .
Conditions on P and F . We also introduce the following smoothness conditions on the distribution P and the class F :
(S1) P is absolutely continuous with a uniformly bounded and continuously differentiable density p such that for some C A > d + 1
sup x R d ( 1 + | x | ) C A | p ( x ) | < + ,
where p denotes the derivative of the density p.
(S2) The class F is uniformly bounded and for all ε > 0 and R > 0
sup f F mes x R d : | x | R and ω f ( x ; δ ) ε 0 as δ 0 .
Here mes denotes Lebesgue measure in R d and
ω f ( x ; δ ) : = sup { | f ( x 1 ) f ( x 2 ) | : | x 1 x | δ , | x 2 x | δ } .
The classes F characterize the distribution. The classes F characterize the distribution in the case of the examples 1.1, 2.1, 2.3, 2.4, 3.1, 4.1 above. Indeed, this is a well-known property of the classes used in examples 1.1, 2.3 and 4.1. As to the Example 2.4, we refer, e.g., to the paper [22] for similar statements. To prove that this is the case in Example 2.1 (and in Example 3.1, similarly), consider the map R d { 0 } x ( | x | , x | x | ) R + × S d 1 . Since this map is a Borel isomorphism (even a homeomorphism), it suffices to show that for any two finite measures P , Q in R + × S d 1 the condition
P ( ( 0 , t ] × C ) = Q ( ( 0 , t ] × C ) for all t > 0 , C C
implies P = Q . We will prove that, in fact, for any two finite measures P , Q on R + × R d the condition
P ( ( 0 , t ] × H ) = Q ( ( 0 , t ] × H ) for all t > 0 , H H ,
where H is the class of all half-spaces in R d , implies that P = Q (the previous statement then follows, since one can consider two measures in R + × R d both supported in R + × S d 1 ). The condition
P ( ( 0 , t ] × H ) = Q ( ( 0 , t ] × H ) for all t > 0 , H H
is equivalent to the following one
R + R d I ( 0 , t ] ( u ) I ( , c ] ( l , x ) P ( d u , d x )
= R + R d I ( 0 , t ] ( u ) I ( , c ] ( l , x ) Q ( d u , d x )
for all l S d 1 , t > 0 , c R . Using a standard approximation of Borel functions by simple functions, we extend this to the equality
R + R d φ ( u ) ψ ( l , x ) P ( d u , d x ) = R + R d φ ( u ) ψ ( l , x ) Q ( d u , d x )
that holds for all bounded Borel functions φ , ψ . If we set φ ( u ) : = e i s u and ψ ( u ) : = e i u , we obtain that the characteristic functions of P and Q are equal, which implies that P = Q .

Acknowledgments

The author would like to acknowledge the support by NSF grant DMS-0806176 and to thank anonymous referees for a number of comments and suggestions that improved this paper.

References

  1. Beran, R. Testing for ellipsoidal symmetry of a multivariate density. Ann. Statist. 1979, 7, 150–162. [Google Scholar] [CrossRef]
  2. Romano, J. Bootstrap and randomization tests of some nonparametric hypothesis. Ann. Statist. 1989, 17, 141–159. [Google Scholar] [CrossRef]
  3. Baringhaus, L. Testing for spherical symmetry of a multivariate distribution. Ann. Statist. 1991, 19, 899–917. [Google Scholar] [CrossRef]
  4. Heathcote, C.R.; Rachev, S.T.; Cheng, B. Testing Multivariate Symmetry. J. Multivariate Analysis 1995, 54, 91–112. [Google Scholar] [CrossRef]
  5. Koltchinskii, V.; Li, L. Testing for spherical symmetry of a multivariate distribution. J. Multivariate Analysis 1998, 65, 228–244. [Google Scholar] [CrossRef]
  6. Manzotti, A.; Pérez, F.J.; Quiroz, A.J. A procedure for testing the null hypothesis of elliptical symmetry. J. Multivariate Analysis 2002, 81, 274–285. [Google Scholar] [CrossRef]
  7. Koltchinskii, V.; Sakhanenko, L. Testing for ellipsoidal symmetry of a multivariate distribution. In High Dimensional Probability II.; Giné, E., Mason, D., Wellner, J., Eds.; Progress in probability; Birkhäuser: Boston, MA, USA, 2000; pp. 493–510. [Google Scholar]
  8. Huffer, F.W.; Park, C. A test for elliptical symmetry. J. Multivariate Analysis 2007, 98, 256–281. [Google Scholar] [CrossRef]
  9. Sakhanenko, L. Testing for Ellipsoidal Symmetry: A comparison study. Comp. Stat. & Data Analysis 2008, 53, 565–581. [Google Scholar]
  10. Beran, R.; Millar, P. Multivariate symmetry models. In Festschrift for Lucien le Cam. Research papers in Probability and Statistics; Pollard, D., Torgersen, E., Yang, G., Eds.; Springer: New York, NY, USA, 1997; pp. 13–42. [Google Scholar]
  11. Wormleighton, R. Some tests of permutation symmetry. Ann. Math. Statist. 1959, 30, 1005–1017. [Google Scholar] [CrossRef]
  12. Li, K.-C. Sliced inverse regression for dimensional reduction (with discussion). J. Am. Statist. Assoc. 1991, 86, 316–342. [Google Scholar] [CrossRef]
  13. Petitjean, M. Chirality and Symmetry Measures: a Transdisciplinary Review. Entropy 2003, 5, 271–312. [Google Scholar] [CrossRef]
  14. Maronna, R. Robust M-estimators of multivariate location and scatter. Ann. Statist. 1976, 4, 51–67. [Google Scholar] [CrossRef]
  15. Hill, D.; Rao, P. Tests of symmetry based on Cramer-von Mises statistics. Biometrica 1977, 64, 489–494. [Google Scholar]
  16. Quiroz, A.J.; Dudley, R. Some New Tests for Multivariate Normality. Probab. Theory and Related Fields 1991, 87, 521–546. [Google Scholar] [CrossRef]
  17. Friedman, M. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. J. Am. Statist. Assoc. 1937, 32, 675–701. [Google Scholar] [CrossRef]
  18. van der Vaart, A.W.; Wellner, J. Weak Convergence and Empirical Processes with Applications to Statistics; Springer-Verlag: New York, NY, USA, 1996. [Google Scholar]
  19. Dudley, R.M. Uniform Central Limit Theorem; Cambridge University Press: Cambridge, UK, 1999. [Google Scholar]
  20. Cirel’son, B. The density of the distribution of the maximum of a Gaussian process. Theory of Probability and Appl. 1975, 20, 847–855. [Google Scholar] [CrossRef]
  21. Giné, E.; Zinn, J. Gaussian characterization of uniform Donsker classes. Ann. Probab. 1991, 19, 758–782. [Google Scholar] [CrossRef]
  22. Koldobskii, A.L. Inverse problem for potentials of measures in Banach spaces. In Probab. Theory and Math. Stat.; Grigelionis, B., Prohorov, Y., Sazonov, V., Eds.; VSP-Mokslas: Vilnius, Lithuania, 1990; Vol. 1, pp. 627–637. [Google Scholar]

Share and Cite

MDPI and ACS Style

Sakhanenko, L. Testing Group Symmetry of a Multivariate Distribution. Symmetry 2009, 1, 180-200. https://doi.org/10.3390/sym1020180

AMA Style

Sakhanenko L. Testing Group Symmetry of a Multivariate Distribution. Symmetry. 2009; 1(2):180-200. https://doi.org/10.3390/sym1020180

Chicago/Turabian Style

Sakhanenko, Lyudmila. 2009. "Testing Group Symmetry of a Multivariate Distribution" Symmetry 1, no. 2: 180-200. https://doi.org/10.3390/sym1020180

Article Metrics

Back to TopTop