Next Article in Journal
Decoding Linear Codes over Chain Rings Given by Parity Check Matrices
Next Article in Special Issue
Compositional Data Modeling through Dirichlet Innovations
Previous Article in Journal
Analytical Solution for Wave Diffraction by a Concentric Three-Cylinder System near a Vertical Wall
Previous Article in Special Issue
Estimating the Gerber-Shiu Function in Lévy Insurance Risk Model by Fourier-Cosine Series Expansion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Dimensional Mahalanobis Distances of Complex Random Vectors

1
Department of Economics and Statistics, Linnaeus University, 35195 Växjö, Sweden
2
Department of Statistics, Örebro Univeristy, 70281 Örebro, Sweden
*
Author to whom correspondence should be addressed.
Mathematics 2021, 9(16), 1877; https://doi.org/10.3390/math9161877
Submission received: 27 May 2021 / Revised: 21 July 2021 / Accepted: 4 August 2021 / Published: 7 August 2021
(This article belongs to the Special Issue Mathematical and Computational Statistics and Their Applications)

Abstract

:
In this paper, we investigate the asymptotic distributions of two types of Mahalanobis distance (MD): leave-one-out MD and classical MD with both Gaussian- and non-Gaussian-distributed complex random vectors, when the sample size n and the dimension of variables p increase under a fixed ratio c = p / n . We investigate the distributional properties of complex MD when the random samples are independent, but not necessarily identically distributed. Some results regarding the F-matrix F = S 2 1 S 1 —the product of a sample covariance matrix S 1 (from the independent variable array ( b e ( Z i ) 1 × n ) with the inverse of another covariance matrix S 2 (from the independent variable array ( Z j i ) p × n )—are used to develop the asymptotic distributions of MDs. We generalize the F-matrix results so that the independence between the two components S 1 and S 2 of the F-matrix is not required.

1. Introduction

Mahalanobis distance (MD) is a fundamental statistic in multivariate analysis. It is used to measure the distance between two random vectors or the distance between a random vector and its center of distribution. MD has received wide attention since it was proposed by Mahalanobis [1] in the 1930’s. After decades of development, MD has been applied as a distance metric in various research areas, including propensity score analysis, as well as applications in matching [2], classification and discriminant analysis [3]. In the former case, MD is used to calculate the propensity scores as a measurement of the difference between two objects. In the latter case, the differences among the clusters are investigated based on MD. The applications of MD are also related to multivariate calibration [4], psychological analysis [5] and the construction of multivariate process control charts [6]; it is a standard method to assess the similarity between the observations.
As an essential scale for distinguishing objects from each other, the robustness of MD is important for guaranteeing the accuracy of an analysis. Thus, the properties of MD have been investigated in the last decade. Gath and Hayes [7] investigated the bounds for extreme MDs. Dai et al. [8] derived a number of marginal moments of the sample MD and showed that it behaves unexpectedly when p is relatively larger than n. Dai and Holgersson [9] examined the marginal behavior of MDs and determined their density functions. With the distribution derived by Dai and Holgersson [9], one can set the inferential boundaries for the MD itself and acquire robust criteria. Beside the developments above, some attention has been paid to the Fisher matrix (F-matrix) and Beta matrix which can be considered as the generalizations of MD.
We provide the definition of the Fisher matrix (see [10]) which is essential here. The Fisher matrices, or simply F -matrices, are an ensemble of matrices with two components each. Let Z j k , j , k = 1 , 2 , be either both real or both complex random variable arrays. For p 1 and n 1 , let
Z = Z j k : 1 j p , 1 k n = Z · 1 , , Z · n , z ( i ) = z i : 1 i n , i k = Z · i ,
be the arrays, with column vectors Z · k . The z ( i ) corresponds to the ith observation in Z . This observation is isolated from the rest of the observations, so that z ( i ) , i k , ( p × 1 ) and Z p × ( n 1 ) are two independent samples. We determine the following matrices:
S 1 = z ( i ) , i k z ( i ) , i k * = z ( i ) z ( i ) * , S 2 = 1 n k = 1 n Z · k Z · k * = 1 n Z Z * ,
where * stands for complex conjugate transpose. These two matrices are both of size p × p . Then the Fisher matrix (or F -matrix) is defined as
F : = S 1 S 2 1 .
Pillai [11], Pillai and Flury [12] and Bai et al. [13] have studied the spectral properties of the F-distribution. Johansson [14] investigated the central limit theorem (CLT) for the random Hermitian matrices, including the Gaussian unitary ensemble. Guionnet [15] established the CLT for the non-commutative functional of Gaussian large random matrices. Bai et al. [16] obtain the CLT for the linear spectral statistics of the Wigner matrix. Zheng [10] extends the moments of the F-matrix into non-Gaussian circumstances with the assumption that the two components S 1 and S 2 of an F-matrix F = S 2 1 S 1 are independent. Gallego et al. [17] extended the MD into the condition of a multidimensional Normal distribution and studied the properties of MD for this case.
However, to the best of our knowledge, the properties of MDs with non-Gaussian complex random variables have not been studied much in the literature. In this paper, we derive the CLT for an MD with complex random variables under a more general circumstance. The common restrictions on the random vector, such as independent and identically distributed (i.i.d.) for the random sample, are not required. The independence between the two components S 1 and S 2 of the F-matrix F = S 2 1 S 1 is relaxed. We investigate the distributional properties of complex MD without assuming normal distribution. The fourth moments of the MDs are allowed to be an arbitrary value.
This paper is organized as follows. In Section 2, we introduce the basic definitions concerning different types of complex random vectors, their covariance matrix, and the corresponding MDs. In Section 3, the first moments of different types of MDs and their distribution properties are given. The connection of leave-one-out MD and classical MD is derived, and their respective asymptotic distributions under general assumptions are investigated. We end up this paper by giving some concluding remarks and discussions in Section 4.

Some Examples of Mahalanobis Distance in Signal Processing

MD has been applied in different research areas. We give two examples that are related to the applications of MD in signal processing.
Example 1.
We illustrate our idea by using an example. In Figure 1, each point corresponds to the real and comlex part of an MD. The circle in red is an inferential boundary with the radius calculated based on MDs. The points that are outside the boundary are considered to be signals, while the points that are lying inside the circle are detected as noise.
This example has been used in some research; for example, in evaluating the capacity of multiple-input, multiple-output (MIMO) wireless communication systems (see [18] for more details). Denote the number of inputs (or transmitters) and the number of outputs (or receivers) of the MIMO wireless communication system by n t and n r respectively, and assume that the channel coefficients are correlated at both the transmitter and the receiver ends. Then the MIMO channel can be represented by an n r × n t complex random matrix H with corresponding covariance matrices Σ r and Σ t . For most of the cases, Σ r and Σ t should be replaced by the sample covariance matrices S r and S t to represent the channel correlations at the receiver and transmitter ends, respectively. The information processed by this random channel is a random quantity which can be measured by MD.
Example 2.
Zhao et al. [19] use MD in a fuzzy clustering algorithm in order to reduce the effect of noise on image segmentation. They compare the boundaries of clusters calculated by both the Euclidean distance and MD. The boundaries calculated by the Euclidean distance are straight lines which misclassify the observations that diverge far from their cluster center, while the boundaries based on the MDs are curves that fit better with the covariance of the cluster. This example implies that the MD is more accurate with regards to the measure of dissimilarity for image segmentation. It can be extended to the application of clustering when the channel signals are complex random variables.

2. Preliminaries

In this section, some important definitions of complex random variables and related concepts are introduced, including the covariance matrix of a complex random vector and the corresponding MDs.
We first define the covariance matrix of a general complex random vector on an n × p dimension complex random variable z where p is the number of variables and n is the sample size of the dataset, as given in [20]:
Definition 1.
Let z j = ( z 1 , , z p ) C p , j = 1 , , n be a complex random vector with known mean E z j = μ z , j where z j = x j + i y j , i = 1 and x j , y j R p . Let Γ p × p be the covariance matrix and C p × p be the relation matrix of z j respectively. Then Γ p × p and C p × p are defined respectively as follows:
Γ = E ( z j μ z , j ) ( z j μ z , j ) * ,
C = E z j μ z , j z j μ z , j ,
where is the transpose.
Definition 1 shows a general presentation of the complex random variables without imposing any distributional assumption on z j . If we set the components of z j , the random variables x j and y j , as normally distributed, then we are in a more familiar context in the complex space: circular symmetricity. The circular symmetricity of a complex normal random variable is an assumption that is used for the standardized form of complex Gaussian-distributed random variables. The definition of circular symmetricity is the following.
Definition 2.
A p-dimensional complex random variable z p × 1 = x p × 1 + i y p × 1 is a circularly symmetric complex normal one if the vector v e c [ x y ] is normally distributed, i.e.,
x p × 1 y p × 1 N Re μ z , p × 1 Im μ z , p × 1 , 1 2 Re Γ z , p × 1 Im Γ z , p × 1 Im Γ z , p × 1 Re Γ z , p × 1 ,
where μ z , p × 1 = E [ z ] and Γ z , p × p = E ( z μ z ) ( z μ z ) * , “ R e ” stands for the real part and “ I m ” stands for the imaginary part.
The circularly symmetric normally distributed complex random variable can be used in the circumstance of a standard normal distribution in real space to simplify the derivations on complex random variables. Based on this condition, we acquire a simplified probability density function (hereinafter p.d.f.) on a complex normal random vector which is presented in the following example.
Example 3.
The circularly symmetric complex random vector z = ( z 1 , , z p ) C p with mean vector μ z = 0 and relation matrix C = 0 has the p.d.f. as follows:
f ( z ) = 1 π p | z | e x p ( z * Γ z 1 z ) ,
where Γ z is the covariance matrix of z and “ | . | ” is the determinant.
Based on Example 3, one can see the possibility of transforming the expression of a complex random vector into a form consisting of vector multiplication between a complex constant vector and a real random vector. Let z j be a complex random variable; then, the transformation between a complex random variable and its real random variables can be presented as follows:
z j = J x j y j , w h e r e J = 1 i .
The transformation offers a different way of inspecting the complex random vector [21]. The complex random vector can be considered as the bivariate form of a real random vector pre-multiplied by a constant vector. This transformation offers another way to present the complex covariance matrix in the form of real random vectors. The idea is presented by the example as follows.
Define matrices Γ x x = E x Re μ x Re μ , Γ y y = E y Im μ y Im μ , Γ x y = E x Re μ y Im μ and Γ y x = E y Im μ x Re μ . The covariance matrix Γ of a p-dimensional complex random vector can be represented in the form of real random vectors x and y , as follows:
Γ z , 2 p × 2 p = Γ x x Γ x y Γ y x Γ y y .

3. MD on Complex Random Variables

In this section, we introduce two types of MDs given in Definitions 5 and 6. We start with the classical MD which is based on a known mean and a known covariance matrix. The definition is given as follows.
Definition 3.
The MD of the complex random vector z j C p , j = 1 , , n with known mean μ and known covariance matrix Γ z is defined as follows:
D Γ z , z j , μ = z j μ * Γ z 1 z j μ .
There are both real and complex parts for a complex random vector. Thus, we define the corresponding MDs of the two components in a complex random variable.
Definition 4.
The MDs on the real part and imaginary part x j , y j R p , j = 1 , , n of a complex random vector z j C p , j = 1 , , n with known mean μ and known covariance matrix Γ . . are defined as follows:
D Γ x x , x j , Re μ = x j Re μ * Γ x x 1 x j Re μ ,
D Γ y y , y j , Im μ = y j Im μ * Γ y y 1 y j Im μ .
Under the definitions above, we can derive the distribution of the MDs on a complex random vector with known mean and covariance. We present the distribution as follows.
Proposition 1.
Let Z be an n × p row-wise double array of i.i.d. complex normally distributed random variables with E z . j = 0 , j = 1 , , p and complex positive definite Hermitian covariance matrix Γ z . Then we have the distribution of the MD in (1) as D Γ z , z i , 0 χ 2 p 2 .
Proof. 
For the proof of this proposition, the reader can refer to reference [22]. □
The result of Proposition 1 is employed here to derive the moments of the MDs on the real and complex parts below.
Proposition 2.
The first moment of D Γ z , z j , 0 defined in (1) is given as follows:
E D Γ z , z j , 0 = 2 p .
Proof. 
The result follows Proposition 1. □
Theorem 1.
Set the random variables x i and y i to be normally distributed, and the covariance of D Γ x x , x i , 0 and D Γ y y , y i , 0 as given in (2) and (3). Then their distributions are given as follows:
C o v D Γ x x , x i , 0 , D Γ y y , y i , 0 = t r Γ x x 1 Γ y y 1 Φ p 2 ,
D Γ x x , x j , 0 D Γ y y , y j , 0 χ p 2 χ p 2 ,
whereis the tensor product, Φ = E x j x j y j y j and t r ( . ) stands for trace.
Proof. 
Proposition 1 shows D Γ y y , y j , 0 χ p 2 and D Γ x x , x j , 0 χ p 2 . In order to derive the covariance of these two MDs, we need to first derive their cross moments.
E x Γ x x 1 x y Γ y y 1 y = E t r x y Γ x x 1 Γ y y 1 x y = E t r Γ x x 1 Γ y y 1 x y x y = E t r Γ x x 1 Γ y y 1 x x y y = t r Γ x x 1 Γ y y 1 E x x y y .
Thus, the covariance is given as follows:
C o v D Γ x x , x j , 0 , D Γ y y , y j , 0 = E D Γ x x , x j , 0 D Γ y y , y j , 0 E D Γ x x , x j , 0 E D Γ y y , y j , 0 = E D Γ x x , x j , 0 D Γ y y , y j , 0 p 2 = t r Γ x x 1 Γ y y 1 Φ p 2 ,
which completes the proof. □
The results presented so far concern a complex random vector with known mean and known covariance matrix. In practice, the mean μ and covariance matrix Γ z are not available all the time. Thus, some alternative statistics such as sample mean z ¯ = n 1 j = 1 n z j and sample covariance S z = n 1 j = 1 n z j z ¯ z j z ¯ * are used as substitutions of the population mean and variance when building the MDs.
We introduce the definitions of MDs with sample mean and sample covariance matrix in the following.
Definition 5.
Let z j C p , j = 1 , , n , be a complex normally distributed random sample. The MD on the complex random vector z j with sample mean z ¯ = n 1 j = 1 n z j and sample covariance matrix S z = ( n 1 ) 1 j = 1 n z j z ¯ z j z ¯ * is defined as follows:
D j = D S z , z j , z ¯ = z j z ¯ * S z 1 z j z ¯ .
For the purpose of deriving further distributional properties, we give an alternative definition of an MD: the leave-one-out case. Leave-one-out here means that we remove one observation each time and use the rest of the observations for constructing the MD.
Definition 6.
Let the leave-one-out sample mean of a complex random vector be z ¯ ( i ) = n 1 1 j = 1 , j i n z j and the corresponding leave-one-out sample covariance matrix S i = n 2 1 j = 1 , j i n z j z ¯ ( i ) z j z ¯ ( i ) * . The leave-one-out MD on the complex random vector z i is defined as follows:
D ( i ) = D S i , z j , z ¯ i = z i z ¯ i * S i 1 z i z ¯ i .
The advantage of leaving one observation out of the dataset is the achieved independence between the removed observation and the sample covariance matrix of the other observations. The similarity in structure between the estimated and the leave-one-out MDs can be explored in theorems on their distributions. The distribution of the leave-one-out MD is derived as in the following theorem.
Theorem 2.
( n 1 ) 2 n 1 D j given in (4) follows a Beta distribution B ( p / 2 , ( n p 1 ) / 2 ) .
The proofs of this theorem and others are relegated to Appendix A, so that readers can perceive the results more smoothly. The distribution of a leave-one-out MD can be used to derive the distribution of an estimated MD. We show this in Theorems 3 and 4.
We present the main results of this paper in the following two theorems. The assumption of a normal distribution on the complex random variable is released. Instead, we introduce two more general assumptions. Assume:
(i)
The entries of complex random matrix Z n × p are independent complex random variables, but not necessarily identically distributed with mean 0 and variance 1. Let the 4th moment of the entries have arbitrary value β . The limiting ratio of their dimensions is p / n c ( 0 , 1 ) .
(ii)
For any η ,
lim n n 1 η 1 j k E | z j k n | 4 I ( | z j k n | η n ) = 0 ,
where I ( . ) is the indicator function. This assumption is a standard Linderberg-type condition that guarantees the convergence of the random variable without the assumption of identical distribution.
Theorem 3.
Under the assumptions (i) and (ii), set the 4th moment of the complex random vector E z 4 = β < . Then the asymptotic distribution of D ( i ) in (5) is:
lim n , p p υ 1 / 2 p 1 D ( i ) τ N 0 , 1 ,
where h = p + c p c ,
τ = β · p h 2 i 1 + i 2 + i 3 = 2 i 3 ! ( 1 ) i 3 h 2 1 c 1 i 1 2 1 i 1 ! 1 i 2 ! h 2 c 1 i 2 c 1 i 3 + i 1 + i 2 + i 3 = 0 2 + i 3 ! ( 1 ) i 3 1 i 1 ! 1 i 2 ! 2 · h 1 + i 1 i 2 h c 3 + i 3 + β 2 1 c × i 1 + + i 4 = 1 i 4 ! ( 1 ) i 4 h i 1 1 c 1 i 1 1 i 1 ! 1 i 2 ! h 2 c h 1 i 2 c c h 1 i 3 h c 1 + i 4 + i 1 + + i 4 = 1 i 4 ! ( 1 ) i 4 h i 1 1 c 1 i 1 1 i 1 ! 1 i 2 ! h 2 c h 1 i 2 c c h 1 i 3 h c 1 + i 4 + i 1 + + i 4 = 0 i 4 + 1 ! ( 1 ) i 4 1 i 1 ! 1 i 2 ! · h 1 + i 1 i 2 c h 1 i 3 + c h 1 i 3 h c 2 + i 4 i 1 + + i 5 = 2 i 5 ! 1 c 1 i 1 1 i 1 ! 1 i 2 ! ( 1 ) i 4 + 2 i 5 h 2 c 1 i 2 c c 1 i 3 c + c 1 i 4 c 1 i 5 i 1 + + i 5 = 0 i 5 + 2 ! ( 1 ) i 5 1 i 1 ! 1 i 2 ! · h 1 + i 1 i 2 c h 1 i 3 c h 1 i 4 h c 3 + i 5 ,
and
υ = 1 1 c 4 i 1 + i 2 + i 3 = 0 j 1 + j 2 = 2 + i 3 i 3 + 1 ! h 2 + i 1 i 2 + j 1 j 2 1 i 1 ! 1 i 2 ! 2 + i 3 ! 1 j 1 ! 1 j 2 ! + β p + c 1 c 2 h 2 i 1 + i 2 + i 3 = 1 i 3 ! ( 1 ) i 3 1 i 1 ! 1 i 2 ! h i 1 1 c 1 i 1 h 2 c h 1 i 2 h c 1 + i 3 + i 1 + i 2 + i 3 = 0 1 + i 3 ! ( 1 ) i 3 1 i 1 ! 1 i 2 ! h 1 + i 1 i 2 h c 2 + i 3 2 .
By using the results above, we derive the asymptotic distribution of the estimated MD in (4).
Theorem 4.
The asymptotic distribution of D j in (4) is given as follows:
p υ 1 / 2 p 1 n + τ 2 / ( n n 1 ) D j n 1 τ / n + τ 2 N 0 , 1 ,
where τ and υ are given in Theorem 3.
In the F-matrix F = S 2 1 S 1 = S 1 z j z ¯ z j z ¯ * , the two component S 1 and z j z ¯ z j z ¯ * are assumed to be independent. Theorem 3 is derived under this restriction, while the results in Theorem 4 extend the circumstance and release the assumption of independence in Theorem 3.

4. Summary and Conclusions

This paper defines different types of MDs on complex random vectors with either known or unknown mean and covariance matrix. The MDs’ first moments and the distributions of MD with known mean and covariance matrix are derived. Further, some asymptotic distributions of the estimated and leave-one-out MDs under complex non-normal distribution are investigated. We have relaxed several assumptions from our previous work [9]. The random variables in the MD are required to be independent but not necessarily identically distributed. The fourth moment of our random variables can be of arbitrary value. The independence between the two components S 2 and S 1 of the F-matrix F = S 2 1 S 1 can be ignored in our results.
In conclusion, the MDs on complex random vectors are useful tools when dealing with complex random vectors in many situations, for example, robust analysis with signal processing [23,24]. The asymptotic properties of MDs can be used in some inferential studies for finance theory [25]. Further studies could develop upon the real and imaginary parts of MDs over a complex random sample.

Author Contributions

Conceptualization, D.D.; Methodology, D.D. and Y.L.; Writing—original draft, D.D. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the referees and the Academic Editor for their constructive suggestions on how to improve the quality and the presentation of this paper. Yuli Liang’s research is supported by the strategic funding of Örebro University.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Proof of Theorem 2

Proof. 
Follow the setting in (4). Then Hardin and Rocke [26] shows that n 1 ( n 1 ) 2 D j with a scalar follows a Beta distribution B e t a ( p / 2 , ( n p 1 ) / 2 ) . □

Appendix A.2. Proof of Theorem 3

Proof. 
By using the Theorem 3.2 from [10], we can acquire the mean and covariance of the complex F matrix F = S i 1 z i z ¯ i z i z ¯ i * . Set β to be arbitrary. Then we receive
τ = lim r 1 1 4 π i | ξ | = 1 ( 1 + h ξ ) ( ξ + h ) 1 c 2 · ξ 1 ξ + r 1 + 1 ξ r 1 2 ξ + c / h d ξ + β p 1 c 2 2 π i · h 2 | ξ | = 1 ( 1 + h ξ ) ( ξ + h ) 1 c 2 · ξ 1 ξ + c / h 3 d ξ + β 1 c 4 π i | ξ | = 1 ( 1 + h ξ ) ( ξ + h ) 1 c 2 · ξ ξ 2 c / h 2 ξ + c / h 2 × 1 ξ c / h + 1 ξ + c / h 2 ξ + c / h d ξ = β p h 2 i 1 + i 2 + i 3 = 2 i 3 ! ( 1 ) i 3 h i 1 1 c 1 i 1 2 1 i 1 ! 1 i 2 ! h 2 c h 1 i 2 h c 1 + i 3 + i 1 + i 2 + i 3 = 0 2 + i 3 ! ( 1 ) i 3 1 i 1 ! 1 i 2 ! 2 · h 1 + i 1 i 2 h c 3 + i 3 + β 2 1 c × i 1 + + i 4 = 1 i 4 ! ( 1 ) i 4 h i 1 1 c 1 i 1 1 i 1 ! 1 i 2 ! h 2 c h 1 i 2 c c h 1 i 3 h c 1 + i 4 + i 1 + + i 4 = 1 i 4 ! ( 1 ) i 4 h i 1 1 c 1 i 1 1 i 1 ! 1 i 2 ! h 2 c h 1 i 2 c c h 1 i 3 h c 1 + i 4 + i 1 + + i 4 = 0 i 4 + 1 ! ( 1 ) i 4 1 i 1 ! 1 i 2 ! · h 1 + i 1 i 2 c h 1 i 3 + c h 1 i 3 h c 2 + i 4 i 1 + + i 5 = 2 ( 1 ) 2 i 4 i 5 ! 1 c 1 i 1 1 i 1 ! 1 i 2 ! h 2 c 1 i 2 c c 1 i 3 c + c 1 i 4 c 1 i 5 i 1 + + i 5 = 0 i 5 + 2 ! ( 1 ) i 5 1 i 1 ! 1 i 2 ! · h 1 + i 1 i 2 c h 1 i 3 c h 1 i 4 h c 3 + i 5 = β p h 2 i 1 + i 2 + i 3 = 2 i 3 ! ( 1 ) i 3 h 2 1 c 1 i 1 2 1 i 1 ! 1 i 2 ! h 2 c 1 i 2 c 1 i 3 + i 1 + i 2 + i 3 = 0 2 + i 3 ! ( 1 ) i 3 1 i 1 ! 1 i 2 ! 2 · h 1 + i 1 i 2 h c 3 + i 3 + β 2 1 c × i 1 + + i 4 = 1 i 4 ! ( 1 ) i 4 h i 1 1 c 1 i 1 1 i 1 ! 1 i 2 ! h 2 c h 1 i 2 c c h 1 i 3 h c 1 + i 4 + i 1 + + i 4 = 1 i 4 ! ( 1 ) i 4 h i 1 1 c 1 i 1 1 i 1 ! 1 i 2 ! h 2 c h 1 i 2 c c h 1 i 3 h c 1 + i 4 + i 1 + + i 4 = 0 i 4 + 1 ! ( 1 ) i 4 1 i 1 ! 1 i 2 ! · h 1 + i 1 i 2 c h 1 i 3 + c h 1 i 3 h c 2 + i 4 i 1 + + i 5 = 2 ( 1 ) 2 i 4 i 5 ! 1 c 1 i 1 1 i 1 ! 1 i 2 ! h 2 c 1 i 2 c c 1 i 3 c + c 1 i 4 c 1 i 5 i 1 + + i 5 = 0 i 5 + 2 ! ( 1 ) i 5 1 i 1 ! 1 i 2 ! · h 1 + i 1 i 2 c h 1 i 3 c h 1 i 4 h c 3 + i 5 .
For the variance, we have
υ = 1 4 π 2 lim r 1 ξ 1 = ξ 2 = 1 1 + h ξ 1 ξ 1 + h 1 + h ξ 2 ξ 2 + h 1 c 4 ξ 1 r ξ 2 2 ξ 1 ξ 2 d ξ 1 d ξ 2 β p + c 1 c 2 4 π 2 h 2 ξ 1 = 1 1 + h ξ 1 ξ 1 + h ξ 1 + c / h 2 ξ 1 d ξ 1 ξ 2 = 1 1 + h ξ 2 ξ 2 + h ξ 2 + c / h 2 ξ 2 d ξ 2 ,
where
1 4 π 2 lim ξ 1 = ξ 2 = 1 1 + h ξ 1 ξ 1 + h 1 + h ξ 2 ξ 2 + h 1 c 4 ξ 1 r ξ 2 2 ξ 1 ξ 2 d ξ 1 d ξ 2 = i 2 ξ 2 = 1 1 + h ξ 2 ξ 2 + h ξ 2 ξ 1 = 1 1 + h ξ 1 ξ 1 + h ξ 1 ξ 1 r ξ 2 2 d ξ 1 d ξ 2 = i 2 π 1 c 4 i 1 + i 2 + i 3 = 0 i 3 + 1 ! h 1 + i 1 i 2 1 i 1 ! 1 i 2 ! ξ 2 = 1 1 + h ξ 2 ξ 2 + h ξ 2 3 + i 3 d ξ 2 = 1 1 c 4 i 1 + i 2 + i 3 = 0 j 1 + j 2 = 2 + i 3 i 3 + 1 ! h 2 + i 1 i 2 + j 1 j 2 1 i 1 ! 1 i 2 ! 2 + i 3 ! 1 j 1 ! 1 j 2 ! ,
and
ξ 1 = 1 1 + h ξ 1 ξ 1 + h ξ 1 + c / h 2 ξ 1 d ξ 1 = 2 π i i 1 + i 2 + i 3 = 1 i 3 ! ( 1 ) i 3 1 i 1 ! 1 i 2 ! h i 1 1 c 1 i 1 h 2 c h 1 i 2 h c 1 + i 3 + i 1 + i 2 + i 3 = 0 1 + i 3 ! ( 1 ) i 3 1 i 1 ! 1 i 2 ! h 1 + i 1 i 2 h c 2 + i 3 .
So we obtain
υ = 1 c 4 i 1 + i 2 + i 3 = 0 j 1 + j 2 = 2 + i 3 i 3 + 1 ! h 2 + i 1 i 2 + j 1 j 2 1 i 1 ! 1 i 2 ! 2 + i 3 ! 1 j 1 ! 1 j 2 ! + β p + c 1 c 2 h 2 × i 1 + i 2 + i 3 = 1 i 3 ! ( 1 ) i 3 1 i 1 ! 1 i 2 ! h i 1 1 c 1 i 1 h 2 c h 1 i 2 h c 1 + i 3 + i 1 + i 2 + i 3 = 0 1 + i 3 ! ( 1 ) i 3 1 i 1 ! 1 i 2 ! h 1 + i 1 i 2 h c 2 + i 3 2 ,
which completes the proof. □

Appendix A.3. Proof of Theorem 4

Proof. 
Regarding asymptotic distribution discussed, one may ignore the rank-1 matrix ZZ * in the definition of the sample covariance matrix and define the sample covariance matrix to be S = n 1 j = 1 n z j z ¯ z j z ¯ * . This idea is given in [27]. Using the results from [9], set W = n S and W i = n 1 S i = j = 1 , j i n y j y j * where y j = z j z ¯ . Let D i = y i * S i 1 y i . Then W i + y i y * i = W i 1 + y * i W i 1 y i and W y i y * i = W 1 y * i W 1 y i so that W i + y i y * i W i + y i y * i W i W i = 1 + ( n 1 ) 1 D i and W y i y * i W Y i Y i W W = 1 n 1 D i . Hence 1 c p 1 D i 1 + c p 1 D i = 1 . Now D i is not independent of observation i, since X i is included in the sample mean vector X ¯ , and hence does not fully represent the leave-one-out estimator of D i . However, using the identity X i X ¯ = n 1 n 1 n n X i X ¯ i and substituting it into the above expressions, we find that 1 n 1 D i 1 + D i n 1 1 + D i n 2 = 1 . This can be used to derive properties of D i as a function of D i , and vice versa. Note that D i is used in the proof to make the derivation easier to follow. This subscript can be replaced by D j in the theorem.
Set g ( p 1 D i ) = p 1 n 1 D i / n + D i ; we acquire its first derivative that g ˙ = g ( D i ) / D i = p 1 n 1 n / n + D i 2 . The same applies to the function g ( τ ) , so that we can acquire g ˙ ( τ ) = p 1 n 1 n / n + τ 2 . With the results in Theorem 3 and the identity from above, we apply Cramer’s theorem [28] and derive g p 1 D i g τ N 0 , g ˙ 2 τ .
As a result, p υ 1 / 2 n + τ 2 / ( n n 1 ) p 1 D i p 1 n 1 τ / n + τ 2 N 0 , 1 as n , p , which completes the proof. □

References

  1. Mahalanobis, P. On tests and measures of group divergence. Int. J. Asiat. Soc. Bengal 1930, 26, 541–588. [Google Scholar]
  2. Stuart, E.A. Matching methods for causal inference: A review and a look forward. Stat. Sci. A Rev. J. Inst. Math. Stat. 2010, 25, 1. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. McLachlan, G. Discriminant Analysis and Statistical Pattern Recognition; John Wiley & Sons: New York, NY, USA, 2004. [Google Scholar]
  4. De Maesschalck, R.; Jouan-Rimbaud, D.; Massart, D.L. The Mahalanobis distance. Chemom. Intell. Lab. Syst. 2000, 50, 1–18. [Google Scholar] [CrossRef]
  5. Schinka, J.A.; Velicer, W.F. Research Methods in Psychology. In Handbook of Psychology; John Wiley & Sons: Hoboken, NJ, USA, 2003; Volume 2. [Google Scholar]
  6. Cabana, E.; Lillo, R.E. Robust multivariate control chart based on shrinkage for individual observations. J. Qual. Technol. 2021, 1–26. [Google Scholar] [CrossRef]
  7. Gath, E.G.; Hayes, K. Bounds for the largest Mahalanobis distance. Linear Algebra Appl. 2006, 419, 93–106. [Google Scholar] [CrossRef] [Green Version]
  8. Dai, D.; Holgersson, T.; Karlsson, P. Expected and unexpected values of individual Mahalanobis distances. Commun. Stat. Theory Methods 2017, 46, 8999–9006. [Google Scholar] [CrossRef]
  9. Dai, D.; Holgersson, T. High-dimensional CLTs for individual Mahalanobis distances. In Trends and Perspectives in Linear Statistical Inference; Springer: Berlin/Heidelberg, Germany, 2018; pp. 57–68. [Google Scholar]
  10. Zheng, S. Central Limit Theorems for Linear Spectral Statistics of Large Dimensional F-Matrices. Annales de l’IHP Probabilités et Statistiques 2012, 48, 444–476. [Google Scholar] [CrossRef]
  11. Pillai, K.S. Upper percentage points of the largest root of a matrix in multivariate analysis. Biometrika 1967, 54, 189–194. [Google Scholar] [CrossRef] [PubMed]
  12. Pillai, K.; Flury, B.N. Percentage points of the largest characteristic root of the multivariate beta matrix. Commun. Stat. Theory Methods 1984, 13, 2199–2237. [Google Scholar] [CrossRef]
  13. Bai, Z.D.; Yin, Y.Q.; Krishnaiah, P.R. On limiting spectral distribution of product of two random matrices when the underlying distribution is isotropic. J. Multivar. Anal. 1986, 19, 189–200. [Google Scholar] [CrossRef] [Green Version]
  14. Johansson, K. On fluctuations of eigenvalues of random Hermitian matrices. Duke Math. J. 1998, 91, 151–204. [Google Scholar] [CrossRef]
  15. Guionnet, A. Large deviations upper bounds and central limit theorems for non-commutative functionals of Gaussian large random matrices. In Annales de l’Institut Henri Poincare (B) Probability and Statistics; Elsevier: Amsterdam, The Netherlands, 2002; Volume 38, pp. 341–384. [Google Scholar]
  16. Bai, Z.; Yao, J. On the convergence of the spectral empirical process of Wigner matrices. Bernoulli 2005, 11, 1059–1092. [Google Scholar] [CrossRef]
  17. Gallego, G.; Cuevas, C.; Mohedano, R.; Garcia, N. On the Mahalanobis distance classification criterion for multidimensional normal distributions. IEEE Trans. Signal Process. 2013, 61, 4387–4396. [Google Scholar] [CrossRef] [Green Version]
  18. Ratnarajah, T.; Vaillancourt, R. Quadratic forms on complex random matrices and multiple-antenna systems. IEEE Trans. Inf. Theory 2005, 51, 2976–2984. [Google Scholar] [CrossRef]
  19. Zhao, X.; Li, Y.; Zhao, Q. Mahalanobis distance based on fuzzy clustering algorithm for image segmentation. Digit. Signal Process. 2015, 43, 8–16. [Google Scholar] [CrossRef]
  20. Goodman, N.R. Statistical analysis based on a certain multivariate complex Gaussian distribution (an introduction). Ann. Math. Stat. 1963, 34, 152–177. [Google Scholar] [CrossRef]
  21. Chave, A.D.; Thomson, D.J. Bounded influence magnetotelluric response function estimation. Geophys. J. Int. 2004, 157, 988–1006. [Google Scholar] [CrossRef]
  22. Giri, N. On the complex analogues of T2-and R2-tests. Ann. Math. Stat. 1965, 36, 664–670. [Google Scholar] [CrossRef]
  23. Zoubir, A.M.; Koivunen, V.; Ollila, E.; Muma, M. Robust Statistics for Signal Processing; Cambridge University Press: Cambridge, UK, 2018. [Google Scholar]
  24. Ma, Q.; Qin, S.; Jin, T. Complex Zhang neural networks for complex-variable dynamic quadratic programming. Neurocomputing 2019, 330, 56–69. [Google Scholar] [CrossRef]
  25. Bai, Z.; Silverstein, J.W. Spectral Analysis of Large Dimensional Random Matrices, 2nd ed.; Springer: New York, NY, USA, 2010. [Google Scholar]
  26. Hardin, J.; Rocke, D.M. The distribution of robust distances. J. Comput. Graph. Stat. 2005, 14, 928–946. [Google Scholar] [CrossRef] [Green Version]
  27. Yao, J.; Zheng, S.; Bai, Z. Sample Covariance Matrices and High-Dimensional Data Analysis; Cambridge University Press Cambridge: Cambridge, UK, 2015. [Google Scholar]
  28. Birke, M.; Dette, H. A note on testing the covariance matrix for large dimension. Stat. Probab. Lett. 2005, 74, 281–289. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Inferential on the signals based on MD.
Figure 1. Inferential on the signals based on MD.
Mathematics 09 01877 g001
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dai, D.; Liang, Y. High-Dimensional Mahalanobis Distances of Complex Random Vectors. Mathematics 2021, 9, 1877. https://doi.org/10.3390/math9161877

AMA Style

Dai D, Liang Y. High-Dimensional Mahalanobis Distances of Complex Random Vectors. Mathematics. 2021; 9(16):1877. https://doi.org/10.3390/math9161877

Chicago/Turabian Style

Dai, Deliang, and Yuli Liang. 2021. "High-Dimensional Mahalanobis Distances of Complex Random Vectors" Mathematics 9, no. 16: 1877. https://doi.org/10.3390/math9161877

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop