Next Article in Journal
Trapezoidal Intuitionistic Fuzzy Power Heronian Aggregation Operator and Its Applications to Multiple-Attribute Group Decision-Making
Next Article in Special Issue
Enhanced Lot Acceptance Testing Based on Defect Counts and Posterior Odds Ratios
Previous Article in Journal
Variational Iteration Method for Solving Fractional Integro-Differential Equations with Conformable Differointegration
Previous Article in Special Issue
Statistical Analysis of Alpha Power Exponential Parameters Using Progressive First-Failure Censoring with Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Testing Multivariate Normality Based on t-Representative Points

1
Department of Statistics and Data Science, BNU-HKBU United International College, Zhuhai 519087, China
2
Guangdong Provincial Key Laboratory of Interdisciplinary Research and Application for Data Science, BNU-HKBU United International College, Zhuhai 519087, China
*
Author to whom correspondence should be addressed.
J.L. and P.H. are deeply indebted to our Ph.D. student J.Y. who finished all simulation work in this paper but unfortunately passed away on 1 January 2022 due to illness.
Axioms 2022, 11(11), 587; https://doi.org/10.3390/axioms11110587
Submission received: 4 September 2022 / Revised: 16 October 2022 / Accepted: 21 October 2022 / Published: 24 October 2022
(This article belongs to the Special Issue Computational Statistics & Data Analysis)

Abstract

:
Testing multivariate normality is an ever-lasting interest in the goodness-of-fit area since the classical Pearson’s chi-squared test. Among the numerous approaches in the construction of tests for multivariate normality, normal characterization is one of the common approaches, which can be divided into the necessary and sufficient characterization and necessary-only characterization. We construct a test for multivariate normality by combining the necessary-only characterization and the idea of statistical representative points in this paper. The main idea is to transform a high-dimensional sample into a one-dimensional one through the necessary normal characterization and then employ the representative-point-based Pearson’s chi-squared test. A limited Monte Carlo study shows a considerable power improvement of the representative-point-based chi-square test over the traditional one. An illustrative example is given to show the supplemental function of the new test when used together with existing ones in the literature.

1. Introduction

The multivariate normal distribution can be characterized many different ways. For example, (1) one of the most well-known characterizations is that a p-dimensional random vector x = ( X 1 , , X p ) ( p × 1 ) has a p-dimensional normal distribution if and only if all of its linear combinations a x ( a R p , the p-dimensional Euclidean space) has a univariate normal distribution; (2) Anderson [1] gave the multivariate normal characterizations by its mean vector and its covariance matrix, respectively; (3) Shao and Zhou [2] gave a characterization for a random vector without a multivariate normal distribution; to name a few. Most of the existing tests for MVN (multivariate normality) in the literature are more or less based on some kind of characterizations for MVN ([3,4]); some are based on necessary and sufficient characterizations for MVN; some others are only based on necessary characterizations for MVN ([5,6,7,8,9,10,11,12,13,14,15,16,17]). An MVN test based on a necessary characterization is called a necessary test in the literature. By reviewing a wide class of MVN tests, Ebner and Henze [18] discuss many of the most updated tests for MVN, which consists of mostly necessary ones. Necessary tests for MVN are usually easier to construct and their exact null distributions or asymptotic null distributions are easier to obtain with simple analytical expressions. However, a common drawback of all necessary tests for MVN is that if the null hypothesis of MVN is not rejected, there is no guarantee of MVN for the original data. Therefore, a characterization-based MVN test may have better power performance than many non-characterization-based MVN-tests. This is our motivation for developing a new characterization-based MVN test.
Yang, Fang, and Liang [19] gave a characterization for MVN by using the the sampling distributions of a series of specially constructed random vectors from an i.i.d. (independently identically distributed) normal sample from a zero-mean p-dimensional normal distribution N p ( 0 , Σ ) with an unknown covariance matrix Σ > 0 (positive definite). This characterization can be employed to construct Q-Q (quantile-quantile) plots for detecting non-multinormality ([20]). In this paper, we will develop further application of the MVN characterization in [19] through employing the idea of statistical representative points (RP for short, [21]), the properties of spherical distributions ([22]), and the classical Pearson–Fisher statistic ([23,24]). This paper is organized as follows. A brief review on the MVN characterization in [19] is given in Section 2. The development of the new test is given in Section 3. Section 4 is devoted to a Monte Carlo study on the empirical performance of the new test and its application. Some concluding remarks are given in the last section.

2. A Brief Review on the MVN Characterization

Let continuous random vectors ( d × 1 ) x 1 , , x n be independently identically distributed (i.i.d.) according to a probability density function (p.d.f.) f ( x ) , x R d (the d-dimensional Euclidean space). For simplicity, we write x 1 , , x n i.i.d. f ( x ) . It is assumed that the p.d.f. f ( x ) has a zero mean and finite second-order moments. Let:
S k = i = 1 k x i x i , y k = S k 1 2 x k , z k = y k 1 y k y k ,
for k = d + 1 , , n , where S k 1 2 = ( S k 1 2 ) 1 and S k 1 2 stands for the positive definite square root of S k .
Theorem 1
([19]). Let x 1 , , x n be i.i.d. N d ( 0 , Σ ) . Define the random vectors y k and z k by (1). Then,
1.
y d + 1 , , y n are mutually independent and y k ( k d + 1 ) has a symmetric multivariate Pearson Type II distribution with a p.d.f.,
f k ( y ) = Γ ( k 2 ) π d 2 Γ ( k d 2 ) 1 y y k d 2 2 , y R d , y y < 1 .
2.
z d + 1 , , z n are mutually independent and z k ( k d + 1 ) has a symmetric multivariate Pearson Type VII distribution with a p.d.f.,
h k ( z ) = Γ ( k 2 ) π d 2 Γ ( k d 2 ) 1 + z z k 2 , z R d .
3.
Let x 1 , , x n be i.i.d. in R d with a p.d.f. v ( x ) , which is continuous in x R d and v ( 0 ) > 0 . Define the random vectors z k by (1). If z k and z k 1 ( k d + 2 ) have p.d.f.’s h k ( z ) and h k 1 ( z ) defined by (3), respectively, then x i ( i = 1 , , n ) has a multivariate normal distribution.
From the above theorem, we can derive the following corollary.
Corollary 1.
Assume that x 1 , , x n are i.i.d. N d ( μ , Σ ) . Define the random vectors:
u i = x 1 + + x i i x i + 1 i ( i + 1 ) , i = 1 , , n 1 .
and,
S k = i = 1 k u i u i , y k = S k 1 2 u k , k = d + 1 , , n 1 .
Let y k = ( y k 1 , , y k d ) : d × 1 ( k = d + 1 , , n 1 ) and:
y ¯ k = 1 d i = 1 d y k i , t k = d y ¯ k v k , v k 2 = 1 d 1 i = 1 d ( y k i y ¯ k ) 2 ;
Then { t k : k = d + 1 , , n 1 } are i.i.d. t ( d 1 ) , the Student’s t-distribution with d.f. (degrees of freedom) d 1 .
Proof. 
By the definition of the u i in (4), it is easy to verify that { u i : i = 1 , , n 1 } are i.i.d. N d ( 0 , Σ ) . By the above theorem, { y k : k = d + 1 , , n 1 } are mutually independent and y k has a Pearson Type II distribution with a p.d.f. given by (2), which is a spherical distribution. For each fixed k, the random variable t k only depends on y k . We write it as:
t k = t k ( y k ) , k = d + 1 , , n 1 .
The independence of { y k : k = d + 1 , , n 1 } results in the independence of { t k ( y k ) : k = d + 1 , , n 1 } . Note that for each fixed k, t k ( y k ) is scale invariant. That is, t k ( a y k ) = t k ( y k ) holds for any constant a > 0 . According to Theorem 2.22 of [22] (p. 51), we have:
t k ( y k ) = d t k ( z 0 ) ,
where z 0 N d ( 0 , I d ) (the d-dimensional standard normal), and the sign “ = d ” means that both sides of the equality have the same distribution. By the definitions of the t-distribution it is obvious that:
t k ( y k ) t ( d 1 ) .
This completes the proof. □

3. The RP-Based Chi-Square Test

The above corollary provides a way to construct a necessary test for MVN of the original i.i.d. sample { x 1 , , x n } . Suppose that we want to test the hypothesis:
H 0 : { x 1 , , x n } is a sample from N d ( μ , Σ )
against the alternative hypothesis that { x 1 , , x n } is not a normal sample. This hypothesis can be transferred to testing:
H 0 : { t k : k = d + 1 , , n 1 } is a sample from t ( d 1 )
versus the alternative that H 0 in (8) is not true. It is obvious that a test for (8) is a necessary one for (7); that is, if hypothesis (8) is rejected, hypothesis (7) is also rejected. However, if hypothesis (8) is not rejected, there is no guarantee for the truth of hypothesis (7).
To test hypothesis (8), we can employ the Pearson–Fisher test (simply called PF-test) for assessing if the set of i.i.d. t-type variates { t k : k = d + 1 , , n 1 } in (6) is from t ( d 1 ) . The traditional PF-test is facing with numerous choices of cells for grouping an i.i.d. sample. The choice of equiprobable cells is recommended in [25], which means that each cell is assigned an equal probability. If the number of cells, m, is pre-assigned, the probability for each cell is 1 / m . Under the null hypothesis, the transformed sample { t k : k = d + 1 , , n 1 } in (6) is from t ( d 1 ) . Then the endpoints { a 1 = , a 2 , , a m 1 , a m = + } of all cells can be computed as follows.
a i a i + 1 f t ( x ; d 1 ) d x = 1 m , i = 1 , , m 1 ,
where f t ( x ; d 1 ) stands for the density function of the Student’s t-distribution t ( d 1 ) .
Because the representative points (RP) ([18]; or called principal points in [26]) of a probability distribution have the property of minimizing some kind of quadratic loss function, we propose to employ the RP of the Student’s t-distribution t ( d 1 ) as the endpoints of intervals for the Pearson–Fisher statistic. The t-RP are a set of points { R 1 , , R m } (for a selected number of points m) that minimize the quadratic loss function:
ϕ ( u 1 , , u m ) = + min 1 i m ( u i u ) 2 f t ( u ; d 1 ) d u ,
ϕ ( R 1 , , R m ) = min 1 i m { ϕ ( u 1 , , u m ) : < u 1 < < u m < + } .
Zhou and Wang [27] gave an algorithm for computing the t-RP { R 1 , , R m } . Define the following intervals:
I 1 = , R 1 + R 2 2 , I 2 = R 1 + R 2 2 , R 2 + R 3 2 , , I m 1 = R m 2 + R m 1 2 , R m 1 + R m 2 , I m = R m 1 + R m 2 , +
and the probabilities:
p i = I i f t ( x ; d 1 ) d x , i = 1 , , m .
According to [21], { p 1 , , p m } can be considered as a set of “representative probabilities" for the Student’s t density function f t ( · ; d 1 ) . The χ 2 -statistic for testing hypothesis (8) is computed by:
χ R 2 = i = 1 m ( n i n p i ) 2 n p i ,
where n i is the frequency of the transformed sample points { t k : k = d + 1 , , n 1 } given by (6) that are located in the interval I i in (10). It is known that χ R 2 χ 2 ( m 1 ) ( n ) in the distribution under some regular conditions. The p-value for testing (8) is computed by:
P ( χ R 2 , ν ) = K χ R 2 z ν 2 1 exp ( z 2 ) d z , with ν = m 1 , K = 2 ν 2 Γ ( ν 2 ) 1 .

4. A Monte Carlo Study and an Illustrative Example

In order to compare the χ R 2 -test (12) under the “representative probabilities” { p 1 , , p m } in (11) with the traditional chi-squared test, we choose the equiprobable cells for computing the traditional chi-square test. For a selected number of representative points m, define the interval endpoints:
a 1 satisfies P ( χ 2 ( m 1 ) < a 1 ) = 1 m ; a 2 satisfies P ( a 1 < χ 2 ( m 1 ) < a 2 ) = 1 m ; a m 1 satisfies P ( a m 2 < χ 2 ( m 1 ) < a m 1 ) = 1 m ; a m satisfies P ( χ 2 ( m 1 ) > a m ) = 1 m .
Denote the traditional chi-square test based on the interval endpoints (14) by χ T 2 :
χ T 2 = i = 1 m ( N i n / m ) 2 n / m ,
which is also an approximate χ 2 ( m 1 ) for a large sample size, where N i stands for the frequency of the observed sample points that are located in the intervals defined by the endpoints (14).

4.1. A Comparison between Empirical Type I Error Rates

Because the chi-square test based on the transformed sample points { t k : k = d + 1 , , n 1 } given by (6) is affine invariant under any nonsingular linear transformation of the original i.i.d. sample { x 1 , , x n } , we only need to generate samples from a d-dimensional standard normal N d ( 0 , I d ) ( I d stands for the d × d identity matrix). The simulation results under 2000 replications for each case are summarized in Table 1, Table 2 and Table 3 for significance levels α = 0.01, 0.05, and 0.10, respectively. It can be roughly concluded that both the traditional chi-square statistic χ T 2 and the RP chi-square statistic χ R 2 show reasonable control of type I error rates under different choices of the number of cells and relatively large sample sizes. Because the Pearson–Fisher chi-square test is an approximate test for goodness-of-fit under its asymptotic null distribution, we do not show its small-sample empirical performance.

4.2. A Simple Power Comparison

To show the benefit of employing the RP-idea for grouping cells for the Pearson–Fisher chi-square test, we carry out a simple Monte Carlo study by selecting the following alternative distributions, which consist of three types of distributions (symmetric about the origin; skewed distributions; distributions with normal marginals):
(1)
[symmetric] The multivariate Cauchy distribution ([22]) has a density function of the form:
f c ( x ) = C 1 1 + x 2 m d + 1 2 ,
where “ · ” stands for the Euclidean norm of a vector, C 1 is a normalizing constant depending on the dimension d.
(2)
[symmetric] The β -generalized normal distribution N d ( 0 , I d , 1 / 2 ) with β = 1 / 4 has a density function of the form by ([28]):
f ( x 1 , , x d ) = β d r d / β 2 d Γ d ( 1 / β ) · exp r i = 1 d | x i | β , ( x 1 , , x d ) R d ,
where r > 0 is a parameter. Let r = 1 / 2 in the simulation and denote it by β -normal.
(3)
[symmetric] Multivariate double Weibull distribution consisting of i.i.d. univariate double Weibull distributions ([29]), its density function is given by:
f d ( x ) = α 2 β d i = 1 d | x i | β α 1 exp i = 1 d | x i | β α , x = ( x 1 , , x d ) R d ,
where α and β are the shape parameter and scale parameter, respectively. Let α = 1 / 2 and β = 1 in the simulation.
(4)
[skewed] The shifted i.i.d. χ 2 ( 1 ) with i.i.d. marginals, each marginal has the same distribution as that of the random variable Y = X E ( X ) , where X χ 2 ( 1 ) , the univariate chi-square distribution with 1 degree of freedom and E ( X ) = 1 .
(5)
[skewed] The shifted i.i.d. exp ( 1 ) with i.i.d. marginals, each marginal has the same distribution as that of the random variable Y = X E ( X ) , where X exp ( 1 ) , the univariate exponential distribution.
(6)
[skewed] The shifted i.i.d. F-distribution with i.i.d. marginals F ( 4 , 3 ) with i.i.d. marginals, Y = X E ( X ) , where X F ( 4 , 3 ) , E ( X ) = E [ F ( 4 , 3 ) ] = 3 .
(7)
[A distribution with normal marginals] The distribution N ( 0 , 1 ) + χ 2 ( 2 ) consists of i.i.d. [ d / 2 ] normal N ( 0 , 1 ) marginals and d [ d / 2 ] i.i.d. χ 2 ( 2 ) 2 marginals, where [ d / 2 ] stands for the integer part of d / 2 .
(8)
[A distribution with normal marginals] The distribution N ( 0 , 1 ) + exp ( 1 ) consists of i.i.d. [ d / 2 ] normal N ( 0 , 1 ) marginals and d [ d / 2 ] i.i.d. exp ( 1 ) 1 marginals.
(9)
[A distribution with normal marginals] The distribution N ( 0 , 1 ) + F ( 4 , 3 ) consists of i.i.d. [ d / 2 ] normal N ( 0 , 1 ) marginals and d [ d / 2 ] i.i.d. F ( 4 , 3 ) 3 marginals.
For each of these alternative distributions, choose the sample size n = 50 , 70 , , 400 . Plot the power values versus the sample size n for both statistics χ R 2 in (12) and χ T 2 in (15).
A visual observation on the following Figure 1, Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9 immediately leads to the following two empirical conclusions:
(1)
The RP chi-square test χ R 2 is comparable to (or slightly better than) the traditional test χ T 2 for symmetric alternative distributions;
(2)
The RP chi-square test χ R 2 is able to improve the traditional test χ T 2 significantly for both skewed and normal+skewed alternative distributions.

4.3. An Illustrative Example

The data were collected from the measurements of various body circumferences of 252 men and were supplied by Dr. A. Garth Fisher. A complete list of the data can be found on the website: http://lib.stat.cmu.edu/datasets/ accessed on 23 October 2022 (bodyfat). In our analysis of the data, the variables are listed as follows:
X 1 : density determined from underwater weighing X 2 : percent body fat X 3 : age ( years ) X 4 : weight ( lbs ) X 5 : height ( inches ) X 6 : neck circumference ( cm ) X 7 : chest circumference ( cm ) X 8 : abdomen 2 circumference ( cm ) X 9 : hip circumference ( cm ) X 10 : thigh circumference ( cm ) X 11 : knee circumference ( cm ) X 12 : ankle circumference ( cm ) X 13 : biceps ( extended ) circumference ( cm ) X 14 : forearm circumference ( cm ) X 15 : wrist circumference ( cm )
The observations from the plotting method for detecting non-multinormality in [30] are summarized as follows.
(1)
The 5-dimensional random vector ( X 4 , , X 8 ) can be approximately considered as 5-dimensional normal;
(2)
The 5-dimensional random vector ( X 4 , X 5 , X 8 , X 9 , X 10 ) shows evidence of non-MVN;
(3)
The 5-dimensional random vector ( X 4 , X 5 , X 9 , X 10 , X 11 ) shows evidence of non-MVN;
(4)
The 5-dimensional random vector ( X 8 , , X 12 ) shows evidence of non-MVN;
(5)
The 10-dimensional random vector ( X 4 , , X 13 ) shows evidence of non-MVN;
(6)
The 10-dimensional random vector ( X 5 , , X 14 ) can be approximately considered as 10-dimensional normal;
(7)
The 10-dimensional random vector ( X 6 , , X 15 ) shows evidence of non-MVN;
(8)
The 10-dimensional random vector ( X 4 , X 6 , , X 11 , X 13 , X 14 , X 15 ) can be approximately considered as 10-dimensional normal.
The chi-square analysis by the two statistics χ R 2 and χ T 2 with three different choices of the number of cells ( m = 5, 10, and 15) is presented in the following Table 4. We obtain mostly consistent results with those from [30] under the significance level 0.05:
(1)
The 5-dimensional random vector ( X 4 , , X 8 ) can be approximately considered as 5-dimensional normal by χ R 2 for m = 10 and m = 15 , and by χ T 2 for m = 15 ;
(2)
The 5-dimensional random vector ( X 4 , X 5 , X 8 , X 9 , X 10 ) shows evidence of non-MVN by χ R 2 for m = 15 . χ T 2 fails to detect the non-MVN for all three choices of m;
(3)
The 5-dimensional random vector ( X 4 , X 5 , X 9 , X 10 , X 11 ) shows evidence of non-MVN by both χ R 2 and χ T 2 for all three choices of m;
(4)
The 5-dimensional random vector ( X 8 , , X 12 ) shows evidence of non-MVN by χ R 2 for m = 5 and m = 10 , and by χ T 2 for m = 10 and m = 15 ;
(5)
The 10-dimensional random vector ( X 4 , , X 13 ) shows evidence of non-MVN by χ R 2 for m = 15 . χ T 2 fails to detect the non-MVN for all three choices of m;
(6)
The 10-dimensional random vector ( X 5 , , X 14 ) shows evidence of non-MVN by χ R 2 for m = 10 and m = 15 . χ T 2 fails to detect the non-MVN for all three choices of m;
(7)
The 10-dimensional random vector ( X 6 , , X 15 ) can be approximately considered as 10-dimensional normal by both χ R 2 and χ T 2 for all three choices of m;
(8)
The 10-dimensional random vector ( X 4 , X 6 , , X 11 , X 13 , X 14 , X 15 ) can be approximately considered as 10-dimensional normal by both χ R 2 and χ T 2 .
The inconsistency from two different tests in the above conclusions (5) and (6) may come from the general drawback for the chi-square test in optimal cell selection. Some good properties were discussed for equiprobable cell selection compared to some random cell selection ([31]) under some conditions. RP-cell selection is based on the idea of minimizing some kind of expected quadratic loss when quantizing a continuous probability distribution ([32]). This kind of quantization gives some good properties in approximating a continuous probability distribution by a set of discrete points. Both methods for cell selection in the chi-square statistic construction are valid in capturing lack of fit between a set of observed frequencies and the set of expected frequencies. A captured lack of fit by any chi-square test always indicates some kind of discrepancy between the null hypothesis and the underlying distribution of sample data. Therefore, the sample data in the above observations (5) and (6) indicate evidence of non-MVN. It is also observed that different choices of the number of cells also result in different p-values. It is possible that a different number of cells could give completely different conclusions in applying the chi-square test. However, a lack of fit from any grouping of data by the chi-square test always indicates some kind of discrepancy between the null hypothesis and the underlying distribution of sample data. More discussion on selecting the number of cells can refer to some early studies on the application of Pearson’s chi-square test ([33,34,35]). A more complete illustration on applying various MVN tests to real data analysis by the R language can be found in [36].

5. Concluding Remarks

The RP-based chi-square test in this paper was developed for the purpose of demonstrating the application of statistical representative points (or principal points) in goodness-of-fit problems. It shows a competitive benefit compared to the same goodness-of-fit methods without employing the RP idea. It can be considered as a successful improvement from the point of view of the significant power increase in the Monte Carlo study. Because the RP-based chi-squared test is a necessary one for testing MVN, it cannot avoid the common weakness of all necessary tests for MVN in the literature. The real-data analysis in the illustrative example shows that the RP-based chi-square test can be a good supplemental test when used together with some existing tests in the literature as reviewed in the introduction section. While many statistics for testing MVN against some general alternative distributions reviewed by [18] perform very well, none of them are perfect. As summarized in [18], MVN test statistics can be classified into two major types: univariate and multivariate approaches. Univariate approaches are based on transformed data from multivariate observations. Therefore, univariate approaches are all necessary ones, implying that no rejection of the null hypothesis cannot conclude MVN. The RP-based chi-square test in this paper belongs to this family. One of the benefits of univariate approaches may be the simple asymptotic null distributions of test statistics with relative fast convergence. This can be found from the simulation of the type I error rates (see Table 1, Table 2 and Table 3 in this paper) based on the critical values of the asymptotic null distributions of the test statistics. However, loss of the original data information seems to be another common drawback of univariate approaches. As a result, univariate approaches usually lead to more power loss than do many multivariate approaches. The most representative multivariate approaches may be the Marida’s [7] multivariate skewness and kurtosis statistics. Many subsequent multivariate approaches were developed after [7], which are more or less related to the sample covariance matrix ([14,18]). One of the drawbacks of the sample-covariance matrix-related statistics is that convergence of the sample covariance matrix to the true population covariance matrix is very slow, and it becomes slower with the increase in data dimensions. As a result, almost all multivariate approaches require very large sample size to control type I error rates and their asymptotic null distributions do not help very much in real applications with finite sample sizes ([9]). The RP-based chi-square test for MVN in this paper can be considered as one of the miscellaneous results for testing MVN reviewed in [18]. It was developed through a necessary and sufficient characterization of MVN and data transformation. This unique characterization-based data transformation guarantees that non-normal multivariate data will not result in the same set of transformed data with the Student’s t-distribution. This is the motivation of proposing the RP-based chi-square test for MVN in this paper. The choice of the number of RPs for constructing the RP chi-square test is not unique. A large number of RPs may result in a zero frequency of transformed data points in some cells of the RP chi-square statistic, causing the numerical computation of the chi-square statistic broken under a small size. This is a common drawback of Pearson–Fisher’s chi-square statistics. After more MNV tests were developed since Pearson’s [23] initial chi-square test, and with the extensive Monte Carlo studies available in the literature, it is arguable that the Pearson chi-square test has become out-of-date. The RP-idea seems to inject new energy into the old Pearson chi-square test. A challenging application of the Pearson chi-square test is the situation that the null distribution contains unknown parameter(s), which has (have) to be estimated before implementing the chi-square test. As a result of the parameter estimation, the asymptotic null distribution of Pearson’s [23] classical chi-square test is no longer an exact chi-square distribution, but a linear combination of independent chi-squares [37]. Our future research direction is to find a way to employ the RP idea to improve other types of chi-square tests, as studied in [38]. Although it is not the purpose of this paper to develop a superior MVN test to any existing MVN test in the literature, which requires an extensive Monte Carlo study, the RP-idea to improve the oldest goodness-of-fit test in statistical history may shed some additional light to the nonparametric statistical inference.

Author Contributions

J.L. developed the theory for the RP chi-squared test and wrote the initial draft. P.H. supervised J.Y. to finish the simulation and the real data analysis and edited all figures. J.Y. finished all simulation and real data analysis. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by a UIC New Faculty Start-up Research Fund R72021106, and in part by the Guangdong Provincial Key Laboratory of Interdisciplinary Research and Application for Data Science, BNU-HKBU United International College (UIC), project code 2022B1212010006.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
i.i.d.Independent identically distributed
MVNMultivariate normality
p.d.f.Probability density function
RPRepresentative points

References

  1. Anderson, M.R. A characterization of the multivariate normal distribution. Ann. Math. Stat. 1971, 42, 824–827. [Google Scholar] [CrossRef]
  2. Shao, Y.; Zhou, M. A characterization of multivariate normality through univariate projections. J. Multivar. Anal. 2010, 101, 2637–2640. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Malkovich, J.F.; Afifi, A.A. On tests for multivariate normality. J. Am. Stat. Assoc. 1973, 68, 176–179. [Google Scholar] [CrossRef]
  4. Cox, D.R.; Small, N.J.H. Testing multivariate normality. Biometrika 1978, 65, 263–272. [Google Scholar] [CrossRef]
  5. Andrews, D.F.; Gnanadesikan, R.; Warner, J.L. Methods for assessing multivariate normality. In Proceedings of the Third International Symposium on Multivariate Analysis, Dayton, OH, USA, 19–24 June 1972; Volume 3, pp. 95–116. [Google Scholar]
  6. Gnanadesikan, R. Methods for Statistical Data Analysis of Multivariate Observations; Wiley: New York, NY, USA, 1977. [Google Scholar]
  7. Mardia, K.V. Measures of multivariate skewness and kurtosis with applications. Biometrika 1970, 57, 519–530. [Google Scholar] [CrossRef]
  8. Mardia, K.V. Tests of univariate and multivariate normality. In Handbook of Statistics; Krishnaiah, P.R., Ed.; North-Holland Publishing Company: Amsterdam, The Netherlands, 1980; Volume 1, pp. 279–320. [Google Scholar]
  9. Romeu, J.L.; Ozturk, A. A comparative study of goodness-of-fit tests for multivariate normality. J. Multivar. Anal. 1993, 46, 309–334. [Google Scholar] [CrossRef] [Green Version]
  10. Horswell, R.L.; Looney, S.W. A comparison of tests for multivariate normality that are based on measures of multivariate skewness and kurtosis. J. Stat. Comput. Simul. 1992, 42, 21–38. [Google Scholar] [CrossRef]
  11. Looney, S.W. How to use tests for univariate normality to assess multivariate normality. Am. Stat. 1995, 39, 75–79. [Google Scholar]
  12. Liang, J.; Li, R.; Fang, H.; Fang, K.T. Testing multinormality based on low-dimensional projection. J. Stat. Plann. Infer. 2000, 86, 129–141. [Google Scholar] [CrossRef]
  13. Srivastava, D.K.; Mudholkar, G.S. Goodness-of-fit tests for univariate and multivariate normal models. Handb. Stat. 2003, 22, 869–906. [Google Scholar]
  14. Mecklin, C.J.; Mundfrom, D.J. An appraisal and bibliography of tests for multivariate normality. Int. Stat. Rev. 2004, 72, 123–138. [Google Scholar] [CrossRef]
  15. Batsidis, A.; Martin, N.; Pardo, L.; Zografos, K. A Necessary power divergence type family tests of multivariate normality. Commun. Stat. Simul. Comput. 2013, 42, 2253–2271. [Google Scholar] [CrossRef]
  16. Al-Labadi, L.; Fazeli Asl, F.; Saberi, Z. A necessary Bayesian nonparametric test for assessing multivariate normality. Math. Methods Stat. 2021, 30, 64–81. [Google Scholar] [CrossRef]
  17. Doomik, J.A.; Hansen, H. An omnibus test for univariate and multivariate normality. Oxf. Bull. Econ. Stat. 2008, 70, 927–939. [Google Scholar]
  18. Ebner, B.; Henze, N. Tests for multivariate normality? a critical review with emphasis on weighted L2-statistics. Test 2020, 29, 845–892. [Google Scholar] [CrossRef]
  19. Yang, Z.H.; Fang, K.T.; Liang, J. A characterization of multivariate normal distribution and its application. Stat. Prob. Lett. 1996, 30, 347–352. [Google Scholar] [CrossRef]
  20. Liang, J.; Pan, W.; Yang, Z.H. Characterization-based Q-Q plots for testing multinormality. Stat. Prob. Lett. 2004, 70, 183–190. [Google Scholar] [CrossRef]
  21. Fang, K.T.; He, S.D. The Problem of Selecting a Given Number of Representative Points in a Normal Distribution and a Generalized Mill’s Ratio; Technical Report; Department of Statistics, Stanford University: Stanford, CA, USA, 1982. [Google Scholar]
  22. Fang, K.T.; Kotz, S.; Ng, K.W. Symmetric Multivariate and Related Distributions; Chapman and Hall: London, UK; New York, NY, USA, 1990. [Google Scholar]
  23. Pearson, K. On the criterion that a given system of deviations from the probable in the case of a correlated system of variables is such that it can be reasonably supposed to have arisen from random sampling. Philos. Mag. 1900, 50, 157–175. [Google Scholar] [CrossRef] [Green Version]
  24. Fisher, R.A. The condition under which χ2 measures the discrepancy between observation and hypothesis. J. R. Stat. Soc. 1924, 87, 442–450. [Google Scholar]
  25. Voinov, V.; Pya, N.; Alloyarova, R. A comparative study of some modified chi-squared tests. Commun. Stat. Simul. Comput. 2009, 38, 355–367. [Google Scholar] [CrossRef]
  26. Flury, B. Principal points. Biometrika 1990, 77, 33–41. [Google Scholar] [CrossRef]
  27. Zhou, M.; Wang, W. Representative points of the Student’s tn distribution and their applications in statistical simulation. Acta Math. Appl. Sin. 2016, 39, 620–640. (In Chinese) [Google Scholar]
  28. Goodman, I.R.; Kotz, S. Multivariate θ-generalized normal distribution. J. Multivar. Anal. 1973, 3, 204–219. [Google Scholar] [CrossRef] [Green Version]
  29. Perveen, Z.; Munir, M.; Ahmad, M. Double Weibull distribution: Properties and application. Pak. J. Sci. 2017, 69, 95–100. [Google Scholar]
  30. Liang, J.; Bentler, P.M. A t-distribution plot to detect non-multinormality. Comput. Stat. Data Anal. 1999, 30, 31–44. [Google Scholar] [CrossRef]
  31. Koehler, K.; Gann, F. Chi-squared goodness-of-fit tests: Cell Selection and Power. J. Commun. Stat. Simul. 1990, 19, 1265–1278. [Google Scholar] [CrossRef]
  32. Graf, S.; Luschgy, H. Foundations of Quantization for Probability Distributions; Springer: Berlin, Germany, 2000. [Google Scholar]
  33. Mann, H.; Wald, A. On the choice of the number of class intervals in the application of the chi-square test. Ann. Math. Stat. 1942, 13, 306–317. [Google Scholar] [CrossRef]
  34. Dahiya, R.C.; Gurland, J. How many classes in the Pearson chi-square test? J. Am. Stat. Assoc. 1973, 68, 707–712. [Google Scholar]
  35. Kallenberg, W.; Oosterhoff, J.; Schriever, B. The number of classes in chi-squared goodness-of-fit tests. J. Am. Stat. Assoc. 1985, 80, 959–968. [Google Scholar] [CrossRef]
  36. Korkmaz, S.; Goksuluk, D.; Zararsiz, G. MVN: An R package for assessing multivariate normality. R J. 2014, 6, 151–162. [Google Scholar] [CrossRef] [Green Version]
  37. Chernoff, H.; Lehmann, E.L. The use of maximum likelihood estimates in tests for goodness of fit. Ann. Math. Stat. 1954, 25, 579–589. [Google Scholar] [CrossRef]
  38. Voinov, V.; Nikulin, M.; Balakrishnan, N. Chi-Squared Goodness of Fit Tests with Applications; Academic Press: New York, NY, USA, 2013. [Google Scholar]
Figure 1. The Cauchy distribution. Red line for RP-chi-square, blue line for traditional chi-square.
Figure 1. The Cauchy distribution. Red line for RP-chi-square, blue line for traditional chi-square.
Axioms 11 00587 g001aAxioms 11 00587 g001b
Figure 2. The β -generalized normal distribution with β = 1 / 4 . Red line for RP-chi-square, blue line for traditional chi-square.
Figure 2. The β -generalized normal distribution with β = 1 / 4 . Red line for RP-chi-square, blue line for traditional chi-square.
Axioms 11 00587 g002
Figure 3. The double Weibulll distribution. Red line for RP-chi-square, blue line for traditional chi-square.
Figure 3. The double Weibulll distribution. Red line for RP-chi-square, blue line for traditional chi-square.
Axioms 11 00587 g003
Figure 4. The shifted chi-square distribution with degree of freedom 1. Red line for RP-chi-square, blue line for traditional chi-square.
Figure 4. The shifted chi-square distribution with degree of freedom 1. Red line for RP-chi-square, blue line for traditional chi-square.
Axioms 11 00587 g004aAxioms 11 00587 g004b
Figure 5. The shifted exponential distribution exp ( 1 ) . Red line for RP-chi-square, blue line for traditional chi-square.
Figure 5. The shifted exponential distribution exp ( 1 ) . Red line for RP-chi-square, blue line for traditional chi-square.
Axioms 11 00587 g005aAxioms 11 00587 g005b
Figure 6. The shifted F distribution F ( 4 , 3 ) . Red line for RP-chi-square, blue line for traditional chi-square.
Figure 6. The shifted F distribution F ( 4 , 3 ) . Red line for RP-chi-square, blue line for traditional chi-square.
Axioms 11 00587 g006aAxioms 11 00587 g006b
Figure 7. The distribution N ( 0 , 1 ) + χ 2 ( 2 ) . Red line for RP-chi-square, blue line for traditional chi-square.
Figure 7. The distribution N ( 0 , 1 ) + χ 2 ( 2 ) . Red line for RP-chi-square, blue line for traditional chi-square.
Axioms 11 00587 g007
Figure 8. The distribution N ( 0 , 1 ) + e x p ( 1 ) . Red line for RP-chi-square, blue line for traditional chi-square.
Figure 8. The distribution N ( 0 , 1 ) + e x p ( 1 ) . Red line for RP-chi-square, blue line for traditional chi-square.
Axioms 11 00587 g008
Figure 9. The distribution N ( 0 , 1 ) + F ( 4 , 3 ) . Red line for RP-chi-square, blue line for traditional chi-square.
Figure 9. The distribution N ( 0 , 1 ) + F ( 4 , 3 ) . Red line for RP-chi-square, blue line for traditional chi-square.
Axioms 11 00587 g009aAxioms 11 00587 g009b
Table 1. Empirical type I error rates ( α = 0.01 ).
Table 1. Empirical type I error rates ( α = 0.01 ).
nm χ 2 d = 5 d = 10 d = 15 d = 20
n = 50 m = 5 χ R 2 0.01550.01100.00750.0070
χ T 2 0.00750.00700.00650.0075
m = 10 χ R 2 0.03200.01450.01200.0125
χ T 2 0.00900.01050.00750.0085
m = 15 χ R 2 0.05250.02850.02100.0235
χ T 2 0.00800.01050.01150.0090
m = 20 χ R 2 0.08350.04650.02550.0340
χ T 2 0.01150.01100.00900.0130
n = 100 m = 5 χ R 2 0.01300.01000.01200.0105
χ T 2 0.00700.01250.01050.0150
m = 10 χ R 2 0.03500.01700.01300.0125
χ T 2 0.01150.01050.01100.0075
m = 15 χ R 2 0.08100.02300.01800.0185
χ T 2 0.01050.01250.01200.0090
m = 20 χ R 2 0.03700.02250.02450.0190
χ T 2 0.01100.01350.01100.0100
n = 200 m = 5 χ R 2 0.00750.00450.00600.0065
χ T 2 0.01050.00900.00750.0085
m = 10 χ R 2 0.02400.01500.01300.0130
χ T 2 0.01150.01350.01100.0105
m = 15 χ R 2 0.04300.01250.01550.0155
χ T 2 0.01200.00800.01250.0110
m = 20 χ R 2 0.04550.01450.01500.0170
χ T 2 0.00750.01200.00950.0120
n = 400 m = 5 χ R 2 0.01100.01550.01400.0085
χ T 2 0.01050.00800.01150.0090
m = 10 χ R 2 0.01850.01050.01400.0105
χ T 2 0.00700.00900.01500.0095
m = 15 χ R 2 0.02300.01400.01550.0145
χ T 2 0.01400.01050.01150.0125
m = 20 χ R 2 0.05700.01400.01200.0175
χ T 2 0.00850.01500.00750.0130
Table 2. Empirical type I error rates ( α = 0.05 ).
Table 2. Empirical type I error rates ( α = 0.05 ).
nm χ 2 d = 5 d = 10 d = 15 d = 20
n = 50 m = 5 χ R 2 0.04300.05500.04600.0450
χ T 2 0.04000.05550.03900.0460
m = 10 χ R 2 0.07300.05050.05650.0450
χ T 2 0.03500.04400.04800.0340
m = 15 χ R 2 0.06600.07500.07250.0675
χ T 2 0.04450.05750.04550.0545
m = 20 χ R 2 0.09300.11150.07900.0710
χ T 2 0.04900.05700.04950.0365
n = 100 m = 5 χ R 2 0.05550.04350.04650.0410
χ T 2 0.04700.04850.05400.0505
m = 10 χ R 2 0.07500.05300.05000.0560
χ T 2 0.05950.05100.05150.0480
m = 15 χ R 2 0.09700.06850.06100.0560
χ T 2 0.05300.05500.05400.0525
m = 20 χ R 2 0.07550.07350.06950.0665
χ T 2 0.05400.04650.05450.0420
n = 200 m = 5 χ R 2 0.04600.04850.04500.0465
χ T 2 0.05650.04950.04650.0505
m = 10 χ R 2 0.05800.05300.04000.0495
χ T 2 0.04900.05200.04250.0530
m = 15 χ R 2 0.11350.06250.05500.0565
χ T 2 0.05300.04800.04850.0470
m = 20 χ R 2 0.07150.06350.05600.0595
χ T 2 0.04850.04500.06000.0505
n = 400 m = 5 χ R 2 0.05500.04700.04950.0475
χ T 2 0.04850.05200.04500.0375
m = 10 χ R 2 0.05900.05250.05150.0475
χ T 2 0.05450.05650.05100.0460
m = 15 χ R 2 0.07400.04600.04750.0495
χ T 2 0.04650.05350.04600.0520
m = 20 χ R 2 0.08800.06700.05800.0505
χ T 2 0.05150.04700.04950.0475
Table 3. Empirical type I error rates ( α = 0.10 ).
Table 3. Empirical type I error rates ( α = 0.10 ).
nm χ 2 d = 5 d = 10 d = 15 d = 20
n = 50 m = 5 χ R 2 0.08350.08850.09150.0880
χ T 2 0.08550.10050.08450.0865
m = 10 χ R 2 0.13250.10450.09700.0985
χ T 2 0.08950.08850.09300.0960
m = 15 χ R 2 0.08700.10650.10550.1130
χ T 2 0.09400.08900.09850.1080
m = 20 χ R 2 0.11550.13750.13300.1290
χ T 2 0.08750.08150.07350.0835
n = 100 m = 5 χ R 2 0.09800.09650.09650.1010
χ T 2 0.11550.10850.10350.0885
m = 10 χ R 2 0.10650.10100.09500.0940
χ T 2 0.09850.09550.09050.0905
m = 15 χ R 2 0.11300.08950.11000.1070
χ T 2 0.09850.09300.10350.1010
m = 20 χ R 2 0.11350.10000.11400.1095
χ T 2 0.10150.09800.10050.0885
n = 200 m = 5 χ R 2 0.09650.09800.09650.1050
χ T 2 0.08950.08800.10550.0940
m = 10 χ R 2 0.10350.10400.09250.0965
χ T 2 0.09800.09700.10100.1040
m = 15 χ R 2 0.15750.10400.08650.1015
χ T 2 0.10400.09800.09400.0930
m = 20 χ R 2 0.10050.10500.09650.1035
χ T 2 0.09400.09650.11100.0990
n = 400 m = 5 χ R 2 0.08650.09950.09400.0930
χ T 2 0.10100.09450.09750.0955
m = 10 χ R 2 0.10200.10550.09750.1095
χ T 2 0.09800.10000.10800.0975
m = 15 χ R 2 0.10450.09500.09700.0930
χ T 2 0.09000.10250.10500.1005
m = 20 χ R 2 0.12100.09650.10850.1025
χ T 2 0.10350.09100.10750.1055
Table 4. p-values from the two chi-square tests (data: bodyfat).
Table 4. p-values from the two chi-square tests (data: bodyfat).
Subsets χ 2 -Test m = 5 m = 10 m = 15
( X 4 , , X 8 ) χ R 2 0.01410.09820.1707
χ T 2 0.02580.02050.0507
( X 4 , X 5 , X 8 , X 9 , X 10 ) χ R 2 0.14140.06410.0012
χ T 2 0.18220.25990.5342
( X 4 , X 5 , X 9 , X 10 , X 11 ) χ R 2 4.0736 × 10 6 9.0387 × 10 13 5.4589 × 10 11
χ T 2 0.07830.0067 6.7792 × 10 5
( X 8 , , X 12 ) χ R 2 9.8107 × 10 4 8.6962 × 10 4 0.0775
χ T 2 0.27340.00510.0022
( X 4 , , X 13 ) χ R 2 0.09800.1271 1.1012 × 10 4
χ T 2 0.71140.49250.2635
( X 5 , , X 14 ) χ R 2 0.36170.02580.0274
χ T 2 0.57820.31830.3534
( X 6 , , X 15 ) χ R 2 0.34350.94090.3285
χ T 2 0.21590.25420.5657
( X 4 , X 6 , , X 11 , X 13 , X 14 , X 15 ) χ R 2 0.20290.03620.3270
χ T 2 0.19980.11730.2191
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liang, J.; He, P.; Yang, J. Testing Multivariate Normality Based on t-Representative Points. Axioms 2022, 11, 587. https://doi.org/10.3390/axioms11110587

AMA Style

Liang J, He P, Yang J. Testing Multivariate Normality Based on t-Representative Points. Axioms. 2022; 11(11):587. https://doi.org/10.3390/axioms11110587

Chicago/Turabian Style

Liang, Jiajuan, Ping He, and Jun Yang. 2022. "Testing Multivariate Normality Based on t-Representative Points" Axioms 11, no. 11: 587. https://doi.org/10.3390/axioms11110587

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop