Abstract
This paper deals with the extension of principal component analysis, canonical correlation analysis, the Cramer–Rao inequality, and a few other statistical concepts in the real domain to the corresponding complex domain. Optimizations of Hermitian forms under a linear constraint, a bilinear form under Hermitian-form constraints, and similar maxima/minima problems in the complex domain are discussed. Some vector/matrix differential operators are developed to handle the above types of problems. These operators in the complex domain and the optimization problems in the complex domain are believed to be new and novel. These operators will also be useful in maximum likelihood estimation problems, which will be illustrated in the concluding remarks. Detailed steps are given in the derivations so that the methods are easily accessible to everyone.
Keywords:
principal component analysis; canonical correlation analysis; extensions to complex domain; linear forms; Hermitian forms; vector; matrix differential operators; Cramer–Rao inequality in the complex domain MSC:
62H25; 62H20; 62J10; 15B52; 15B57
1. Introduction
In most textbooks on statistics, scalar/vector/matrix-variate random variables in the complex domain and the corresponding statistical analysis in the complex domain, are not discussed. But, in a large number of physical situations, it is natural or more convenient to represent variables in the complex domain. Hence, statistical techniques in the complex domain are required for data analysis in these situations. In the physical science and engineering literature, there are a number of papers dealing with random variables in the complex domain. Data reduction techniques such as principal component analysis in the complex domain seem to be the main topic in these areas. Most of the papers in these applied areas, concentrate on developing algorithms for computing eigenvalues and eigenvectors, which are useful and relevant in principal component analysis, independent component analysis, factor analysis, and so on. Statistical analysis in the complex domain is widely used in the analysis of multi-look return signals in radar [1], in multi-task learning in artificial intelligence and machine learning [2], in problems such as signal processing [3], in principal component analysis and independent component analysis in analyzing meteorological data in the complex domain [4], in optimal allocation of resources, especially energy resources [5], in holography, microscopy and optical metrology [6], in delayed mixing in speech processing, in biomedical signal analysis, and in financial data modeling, etc. [7].
In the present paper, vector/matrix differential operators in the complex domain are defined. Then, these are applied in optimizing a Hermitian form under a linear constraint, a bilinear form under Hermitian-form constraints, etc. Then, as applications of these optimization problems, the real-domain techniques of principal component analysis and canonical correlation analysis are extended to the complex domain. Other statistical concepts such as the Cramer–Rao inequality, least square procedure, and related aspects are extended to the complex domain. Detailed derivations are given so that the methods will be accessible even to beginners.
The following notation is used in this paper: Scalar variables, whether mathematical or random, are denoted by lower-case letters such as . Vector/matrix variables are denoted by capital letters such as . Scalar constants are denoted by , etc., and vector, matrix constants by , etc. The wedge product of the differentials and is defined as , where x and y are two real scalar variables, so that . Let be an matrix with distinct real scalar variables as elements, then . The transpose of a matrix X is denoted by a prime as . For a matrix X, if (symmetric), then . Variables in the complex domain are written with a tilde, such as . If is a matrix in the complex domain, then are real, then is defined as . The determinant of a real matrix Y is written as or as and if is in the complex domain, then the absolute value of the determinant of is written as . Also, means the trace of the square matrix Y. A real matrix A being positive definite is written as , and indicates that the matrix , in the complex domain, is Hermitian positive definite. Other notation is explained when it occurs for the first time.
This paper is organized as follows: Section 2 starts with defining a vector differential operator in the complex domain. Then, with the help of this operator, some optimization problems such as optimizing a linear form subject to Hermitian-form constraint, a Hermitian form under Hermitian-form constraint, and a bilinear form with Hermitian-form constraints are discussed. Then, a matrix differential operator in the complex domain is defined. Differentiations of the trace of a product of matrices and determinant of a matrix, by using the matrix differential operator in the complex domain, are dealt with. Section 3 deals with the extension of principal component analysis to the complex domain. Section 4 delves into the extension of canonical correlation analysis to the complex domain. Section 5 examines the extension of the Cramer–Rao inequality, Cauchy–Schwarz inequality, and the least square estimation procedure to the complex domain. Detailed steps are given in each situation.
2. Optimization Involving Linear Forms, Traces, and Determinants
Here, we will consider linear forms and their differentiations first, and then, we will consider some situations of optimizing a Hermitian form with linear constraint and optimization of a linear form with Hermitian-form constraint. We consider some basic results to start with. Let
where are real vectors. Assume that A is a non-null arbitrary coefficient vector and is a known vector random variable in the complex domain.
Theorem 1.
For A and as defined above,
Proof.
This can be easily seen from the following: Then,
from the corresponding real variable case. That is,
□
Now, we consider a slightly more general result. Consider . Let and . Then, from note that and ; and , where means the complex conjugate of A.
Theorem 2.
For and A as defined above, let . Let be a matrix, free of the elements in A and . Then,
Proof.
Since and are free of the elements in A, we may take , real. Then, from Theorem 1 the result follows. □
Now, consider a vector A and a matrix and the Hermitian form in the complex domain where is free of the elements in A. Then, we have the following result:
Theorem 3.
For the Hermitian form as defined above, where are real,
Proof.
Opening up , we have the following:
Since , we have (real symmetric), (real skew symmetric). because due to being real skew symmetric. Similarly, . Then, from the results in the real case, we have the following:
Consider
But, and and real , and hence, . Also, because , and hence, . Therefore,
Then, from the real case,
from the corresponding real case. Then, from (2)–(5),
This establishes the result. □
2.1. Optimization of a Linear Form Subject to Hermitian-Form Constraint
Consider a linear form , where A and are non-null vectors, with A being arbitrary, and being a known vector variable with variance , where is the covariance matrix in . Consider the optimization of , subject to the constraint that the variance of is fixed, say unity, that is, . Then, we have the following result:
Theorem 4.
For the linear form and the constraint , as defined above,
Proof.
Let , where is a Lagrangian multiplier. Observe that and we assume that it is Hermitian positive definite. From previous results (1) and (2),
That is,
But from , which means that the maximum of our linear form is the largest and the minimum of our linear form is the smallest . But, from and , , where means the absolute value of and is the complex conjugate of .
being positive definite Hermitian form. This completes the proof. □
2.2. Optimization of a Hermitian Form Subject to Linear Constraint
Now, consider a problem of optimizing subject to fixed, where A and are vectors and is a Hermitian positive definite matrix, free of the elements of A. Then, we have the following result:
Theorem 5.
For the Hermitian form and linear form constraint , we have
Proof.
Let be a Lagrangian multiplier and consider . From (1) and (2), we have
From ,
because is Hermitian positive definite, where and denote the absolute values of and , respectively. Then, since A is arbitrary and since the linear restriction cannot eliminate the effect of A fully. Hence, we look for the minimum. From ,
Hence,
This completes the proof. □
2.3. Differentiation of Traces of Matrix Products and Matrix Differential Operator
Now, we consider a few matrix-variate cases, where also the problem is essentially optimization involving vector variable situations. Let be a matrix of arbitrary elements . Let be a matrix in the complex domain with distinct scalar complex variables as elements. Consider . Let and be the j-th columns of A and , respectively. Then, . Let are real vectors. Let . Consider the operator
and then, since we are taking partial derivatives. Then, we have the following result:
Theorem 6.
Let be as defined above. Then,
Proof.
From the previous result,
from Theorem 1. But, . Now, stack up these side by side as q columns for . Note that , and , and hence, . This completes the proof. □
For the next result to be established, we need a result on matrices, which will be stated here as a lemma.
Lemma 1.
For two real matrices A and B, where (symmetric) and (skew symmetric), .
Proof.
For two matrices and , it is well known that and . Then, for our matrices A and B, , we have . That is, . □
For the next problem, let be a Hermitian matrix of arbitrary elements . Let be a Hermitian matrix in the complex domain with distinct complex scalar variables as elements except for the Hermitian property. Let , where are real. Then, (symmetric), (skew symmetric). Let . Then, we have the following result:
Theorem 7.
Let , and be as defined above. Then,
where means a diagonal matrix consisting of only the diagonal elements from .
Proof.
Opening up , we have the following:
since by Lemma 1. Now, from the known result for symmetric matrices in the real case, we have
If the -th element, for , in is , then the -th element in is and the -th element in is . Hence, . Hence,
since the diagonal elements in and are already zeros. Combining and , we have
which corresponds to the result in the real case. □
2.4. Differentiation of a Determinant in the Complex Domain
Here, we will start with defining the derivative of a scalar quantity with respect to a matrix or we will define a matrix differential operator first. Let be an real matrix and let be an matrix in the complex domain. Then, we can always write , where are real matrices. Then, matrix differential operators and will be defined as the following:
Consider a nonsingular matrix in the complex domain and let be its determinant. Then, we have the following result:
Theorem 8.
For the nonsingular matrix in the complex domain,
and
Proof.
The cofactor expansion of a determinant holds in the real and complex domains and it is the following:
where is the cofactor of for all j and k. Let be the matrix of cofactors. For a general matrix, all ’s are distinct and the corresponding ’s are also distinct. Hence, in the real case, taking the partial derivative of the j-th line in we have for all j and k. Hence, in the real case
This is a known result. But, in the complex case, the situation is different. Before we tackle the complex-domain situation, we will develop some necessary tools. It is a known result that . But when (Hermitian), the eigenvalues are real, and hence, the determinant is real. Then, is Hermitian, thereby is also Hermitian. The following results will be helpful when we apply our matrix differential operators to scalar functions of matrices in the complex domain. For two scalar complex variables and the following can be easily verified.
For convenience, let us consider the term
where , with being real scalar quantities. Note that when is Hermitian. Hence, for example, when we differentiate with respect to it is equivalent to differentiating with respect to and vice versa. When is differentiated with respect to it is equivalent to differentiating with respect to , and so on. Then,
This is the derivative of the complex conjugate of the -th element in for all j and k when is a general matrix with distinct scalar complex variables as elements. Then,
When (Hermitian) we have . Then, the result in holds for the k-th term in the j-th line in . But, when is Hermitian, there is one more element contributing to and . This is the j-th term in the k-th line of . Thus, the sum of the contributions coming from these two terms is the derivative of with respect to . The sum of the contributions is the following, observing that :
Now,
for all . When , the diagonal elements in and are real, and hence, the term occurs only once, and therefore,
Since we are taking the partial derivative, the derivatives are the same as differentiation of . Hence,
and then,
Here, we have used two properties that for a matrix B, and if B is Hermitian. Then, we have the following theorem: □
Theorem 9.
When the nonsingular matrix in the complex domain is Hermitian, that is, , then,
In the following sections, we will consider some applications of the results obtained in Section 2.
3. Principal Component Analysis in the Complex Domain
This is an application of the mathematical problem of optimization of a Hermitian form under a Hermitian-form constraint.
In many physical situations, variables occur in pairs, such as time and phase, and hence, the most appropriate representation of such variables is through complex variables because a scalar complex variable can be taken as a pair of real variables. Let are real scalar variables, be a scalar complex variable. Then, a statistical density associated with is a real-valued scalar function of such that over the entire complex plane, something like a hill on the complex plane, so that the total volume under the surface is unity, that is, , where , the wedge product of the differentials and , respectively. Then, the center of gravity of the hill is at and the square of the measure of scatter in is , where means the expected value when is a scalar complex random variable and is its density.
If the scatter is small, then the variable is concentrated near the center of gravity . If is large, then is spread thin far and wide, and hence, it is more or less unrecognizable. If a large number of scalar variables are being considered as possible variables to be included into a model, then the ones with the larger scatter are the important variables to be included into the model. For convenience, one can consider linear functions of such variables because linear functions also contain individual variables. A linear function is of the form , where are constants with being scalar complex variables. For example, gives , where a * indicates the conjugate transpose and for a scalar quantity y, only, where c in the exponent indicates the complex conjugate. We may write the linear function as , where
where a prime indicates the transpose. The expected value of is and the variance–covariance matrix or the covariance matrix in is denoted by (Hermitian positive definite). Then, the variance of , denoted by , is given by . The most important linear function is that linear function having the maximum variance. Hence, we may compute . But, since we have a positive definite Hermitian form in and its maximum over A is at . Thus, unrestricted maximization does not make sense. Without loss of generality we may take because this can always be achieved for any non-null A. Then, the maximization amounts to maximizing within the unit sphere . There will be a maximum and a minimum in this case. We may incorporate the restriction by using a Lagrangian multiplier. Consider
where is a Lagrangian multiplier. Maximization will be achieved by using the following result, which will be stated as a lemma.
Lemma 2.
Let are vectors with distinct real scalar variables as elements. Let be a constant Hermitian matrix. Let the partial differential operator be defined as
Then,
Proof.
Let . When B is Hermitian, , that is, is real symmetric and is real skew symmetric. Let are real. Then, , where because due to . From the real case it is well known that . Similarly, . Now,
Similarly,
where we have used two properties. When , that is, it is real symmetric, we have because both are real, and hence, each is equal to its transpose, and therefore, the difference is zero. When we have , because both are real, and hence, each is equal to its transpose. Then, from the real operator operating on a real linear form the result follows. Similar results hold when differentiating with respect to the operator . Thus, we have the following:
Therefore,
Hence, the result. For , the identity matrix, the result on follows, that is, . Now, by using Lemma 1 we can differentiate in . That is,
where means the determinant of the square matrix . From , and pre-multiplying by and using the fact that , we have . Hence,
where is the largest eigenvalue of and is the smallest eigenvalue of . When is Hermitian, all its eigenvalues are real and when it is Hermitian positive definite, all its eigenvalues are real positive also. Hence, the procedure is the following: Take the largest eigenvalue of , say . Then, through (23), compute an eigenvector corresponding to , that is, solve for an A. Then, normalize this eigenvector through . That is, if an eigenvector corresponding to is , then compute . This is the normalized eigenvector corresponding to . Now, consider . This is the first principal component in the sense of the linear function having the maximum variance. Now, take the second largest eigenvalue . Go through the same procedure and construct the normalized eigenvector corresponding to . Then, is the second principal component. Continue the process and stop the process when the variance of , namely, , falls below a preassigned number. If there is no preassigned number, then will be the p principal components. Here, we have assumed that the eigenvalues of are distinct. When the eigenvalues are distinct, we can show that the eigenvectors corresponding to the distinct eigenvalues of a symmetric or Hermitian matrix are orthogonal to each other. Hence, our principal components will be orthogonal to each other in the sense that the joint dispersion in the pair is 0 for or the covariance between and is zero when . The covariance is defined as the following: Let be a vector in the complex domain and let be a vector in the complex domain. Then, the covariance of on is defined and denoted as , whenever this expected value exists, so that when , then , the covariance matrix in , and when , it is the variance of the scalar complex variable . □
When the covariance matrix is unknown, then we may construct sample principal components. Let our population be the vector , where are distinct scalar complex variables. Consider n independently and identically distributed (iid) such p-vectors. Then, we have a simple random sample of size n from . Then, the sample matrix is the matrix, denoted as the following:
Let the sample average be denoted by and the matrix of sample averages be denoted by the bold letter . Then, the sample sum of products matrix is given by
where . The motivation in using the sample sum of products matrix is that is an unbiased estimator of . Since we will be normalizing the eigenvectors, operate with itself. Compute the eigenvalues of . Take the largest eigenvalue of . Call it . Construct an eigenvector corresponding to and normalize it through , where is the normalized eigenvector corresponding to . Then, is the first sample principal component. When the columns of the sample matrix are not linearly related then we have (Hermitian positive definite) and all eigenvalues will be positive. We assume that the eigenvalues are distinct . This will be true almost surely. Now, take and construct and the second principal component and continue the process. We can show that the covariances between and will be zeros for all . This property follows from the fact that when the matrix is symmetric or Hermitian, the eigenvectors corresponding to distinct eigenvalues are orthogonal. When the population is p-variate complex Gaussian, then we can show that will be a complex Wishart distributed with degrees of freedom and parameter matrix . The distributions of the largest, smallest, and j-th largest eigenvalues and the corresponding eigenvectors of in the complex domain are given in [8].
4. Canonical Correlation Analysis in the Complex Domain
This is an application of the mathematical problem of optimization of a bilinear form in the complex domain, under two Hermitian-form constraints. The following application is regarding the prediction of one set of variables by using another set of variables.
Consider two sets of scalar complex variables
where p need not be equal to q. Consider the appended vector
where , and , and vice-versa. That is, is the covariance matrix in , is the covariance matrix in , is the covariance matrix in , and so on. Also, . Our aim is to predict the variables in the set by using the variables in the set and vice-versa, and obtain the “best” predictors; “best” in the sense of having the maximum joint dispersion. In order to represent each set and , we will take arbitrary linear functions of the variables in each set. Consider linear functions and , where
where and are listed above already. Since and are scalar constant quantities, and . Variances for linear functions are already seen in Section 2. Therefore, and . Here, and can be taken as measures of joint dispersion or joint variation between and and as the joint dispersion between and . As a criterion for the “best” predictor of by using and vice versa, we may take that the pair and have the maximum joint variation. The best predictor of by using as that pair having the maximum and the best predictor of by using as that pair having the maximum value for . Since covariances depend upon the units of measurements of the variables involved, we may take a scale-free covariance by taking
Further, as explained in Section 2, without loss of generality we may take and or confine the bilinear form (hyperboloid) within unit positive definite Hermitian forms (ellipsoids) in order to prevent them from going to . Hence, our procedure simplifies to optimizing subject to the conditions and computing that pair of A and B which will maximize . As before, we may use the Lagrangian multipliers and and consider the function
In order to optimize this we need one result on differentiation of a bilinear form, which will be stated as a lemma.
Lemma 3.
Let are real, vectors of distinct real scalar variables and , respectively, where . Let are real, vectors of distinct real scalar variables and , respectively, where , where are real. Let the partial differential operators be as defined in Section 2, namely,
and similar operators involving . Then,
Proof.
Opening up we have the following:
Then, from the known results in the real case, we have the following:
irrespective of whether A is real or in the complex domain. Then,
Similarly, This completes the proof. □
Now, differentiating in , we have the following:
Now, premultiply by and by to obtain the following, observing that :
or . Take B from and substitute in to obtain the following:
This shows that is an eigenvalue of , where the matrix is . From symmetry, it follows that is also an eigenvalue of , where the matrix is . Hence, all the nonzero ’s are common to both of these matrices. Hence, the procedure is the following: If , then compute the eigenvalues of , otherwise, compute the eigenvalues of the other matrix , both will give the same nonzero eigenvalues. Let be the largest and be the smallest nonzero eigenvalues. Then, we have the results
and
Then, the procedure is the following: If , then compute all the eigenvalues of . Let the largest eigenvalue be . Then, compute one eigenvector corresponding to . Use the equation . Let it be . Then, normalize it through . That is, compute . Then, compute . Then, use the same eigenvalue and compute one eigenvector from the equation . Let it be . Then, normalize it through , that is, compute . Now, compute . Then, is the first pair of canonical variables in the sense is the best predictor of and vice-versa. Now, take the second largest eigenvalue of . Then, compute one eigenvector corresponding to , that is, solve the equation Let be that eigenvector. Normalize it, that is, compute . Now, compute . Use the same and solve for B from the equation . Let be one solution. Then, normalize it, that is, compute . Now, compute . Then, is the second pair of canonical variables. Continue the process until falls below a preassigned limit. If there is no such preassigned limit, then compute all the pairs, that is, p if ; otherwise q. If , then start with the computation of the eigenvalues of and proceed parallel to the steps used in the case . Observe that the symmetric format of is . This form is also available from the same starting Equation . The symmetric format can always be written in the form for some matrix C, and hence, the symmetric form is either Hermitian positive definite or Hermitian positive semi-definite, and therefore, all the nonzero eigenvalues are positive. Let us assume that the eigenvalues are distinct, if . It is a known result that eigenvectors corresponding to distinct eigenvalues of Hermitian or symmetric matrices are orthogonal to each other. Hence, are non-correlated. Similarly, are non-correlated.
If are not available, then take a simple random sample of size n, , from . Then, compute the sample sum of products matrix:
as for the case of principal component analysis. Now, continue with ’s as with ’s. Then, we will obtain the pairs of sample canonical variables. Some distributional aspects of sample canonical variables and the sample canonical correlation matrix are discussed in [8]. Note that when is , then we have the square of the multiple correlation coefficient in , which is scalar. In this case, our starting set will have only one complex scalar variable and the set will have q variables. Here, the problem is to predict one variable in by using the variables in . The exact distributions of the canonical correlation matrix and the square of the absolute value of the multiple correlation coefficient are available in explicit forms in [8].
5. Covariance and Correlation in the Complex Domain
Consider scalar complex variables first. Let and be two scalar complex variables or scalar variables in the complex domain. Then, the mean values or expected values of and are, respectively, , where means the expected value of , and f and g are the densities of and , respectively. Let . Let are real scalar variables, and let are real scalar variables. Then, . Then, the variance, denoted by , and covariance, denoted by , are the following:
where, for example, , with * indicating the conjugate transpose and c indicating the conjugate only. Here, we have only scalar variables, and hence, the complex conjugate transpose is only complex conjugate. For convenience, we have used the notation . Then, . For a scalar complex variable . Let us examine the variances of sum and difference. Let
Then,
This can be simplified as the following:
From and we have
Let us examine the quantity, Note that
Hence,
Therefore, we may define the correlation coefficient in the complex domain, denoted by , as the following:
where denotes the real part in . Also, note that . This will motivate us to examine the dot product of two vectors in the complex domain and the Cauchy–Schwarz inequality in the complex domain.
5.1. The Cauchy–Schwarz Inequality in the Complex Domain
Let , and be two vectors in the complex domain. We will define the dot product between and , and it will be denoted and defined as the following: and
Thus, the Cauchy–Schwarz inequality holds for the complex domain also.
5.2. Minimum-Variance Unbiased Estimators in the Complex Domain
In the class of linear estimators for the parametric function , which linear function is the minimum-variance unbiased estimator for ? Let
and where is the covariance matrix in . Our aim here is to minimize subject to the constraint (given). Let be a Lagrangian multiplier and let . Then,
That is, . Hence, multiplied by the minimum value of gives the minimum of . From ,
Then,
is the minimum value of the variance of our linear estimator. Hence, the minimum-variance unbiased estimator is
Note also that
The first inequality follows from the inequality established in Section 5.1.
5.3. Cramer–Rao-Type Inequality in the Complex Domain
Let be a simple random sample from the population designated by the density , where are independently and identically distributed (iid). Then, the joint density . Let , denoted as , be a statistic with the expected value . That is, . Differentiating with respect to , we have
where we have assumed that the support of is free of and differentiation inside the integral is valid. But, from the total integral being one, we have
From and , we have because for any two scalar random variables u and v real, or and in the complex domain, and the corresponding results in the complex domain also hold. Therefore, since . Then,
Note that
This shows that the Cramer–Rao-type inequality holds in the complex domain also.
5.4. Least Square Estimation in Linear Models in the Complex Domain
Let us examine whether the least square procedure, in the class of linear models, holds in the complex domain also. Let be preassigned complex numbers or observations on k random variables in the complex domain. Let be a scalar complex variable. Then, a linear model of for predicting , can be of the following form:
But, in order to predict by using linear predictors we must know the conditional distribution of , given , and also the conditional expectation must be linear in . If the conditional distribution is not known, then we may use a distribution-free procedure. One such procedure is the estimation procedure by using the least square method. In this method, we set up a corresponding model of the following form for the j-th observation on , namely, , for , where is the sample size.
corresponding to the linear model in the real case, where is the random part or the sum total contributions coming from unknown factors, corresponding to . Then, if we sum up the observations and divide by n we obtain the sample averages , where, for example, . Then, from , we have
We have taken the error sum as zero without much loss of generality. Since is available from , we may rewrite as follows:
We may write all the equations in together as , where
Note that is an matrix, is an matrix, is a matrix, and is an matrix. Then, the sum of squares of the absolute values of the errors is the following:
In the least square procedure, we minimize this error sum of squares of the absolute values of the errors, and then, estimate the parameter vector . Note that and
where we have assumed that is a nonsingular matrix because ’s are preassigned numbers, and hence, can be taken as a full-rank matrix with rank . Then, the estimated , as per the least square estimate, again denoted by , is and the estimated model for is or the estimated is
where is available from and , . This shows that the least square procedure in the complex domain also runs parallel to that in the real domain. If is assumed to have an n-variate complex Gaussian distribution, then the inference problems also runs parallel to those in the real domain.
6. Concluding Remarks
In this paper, we have introduced vector/matrix differential operators in the complex domain. These differential operators in the complex domain are believed to be new. With the help of these operators, we have examined the optimization of a linear form with Hermitian-form constraint, optimization of a Hermitian form with linear form as well as Hermitian-form constraint, and optimization of a bilinear form with Hermitian-form constraints, where the linear forms and bilinear forms involve vectors and matrices in the complex domain. As applications of these optimization problems, we have extended principal component analysis and canonical correlation analysis to the complex domain. Also extended to the complex domain are the Cramer–Rao inequality, the Cauchy–Schwarz inequality, minimum-variance unbiased estimation, and least square analysis. If we use the general definition of a density as a real-valued scalar function such that in the domain of X and , where the argument X may be scalar or vector or matrix or a sequence of matrices in the real or complex domains [8], then the structures of the joint density, marginal density, conditional density, etc., will be parallel to those in the real domain. Then, we will be able to extend Bayesian analysis to the complex domain. One can also explore extending other multivariate statistical techniques such as factor analysis, classification problems, cluster analysis, analysis of variance, analysis of covariance, etc., to the complex domain. These are some of the open problems. Since the likelihood function L is a product of densities at the observed sample point, in the simple random sample case, this L will be a real-valued scalar function. Then, one can extend the maximum likelihood method of estimation to the complex domain. For example, in the p-variate complex Gaussian case, in the case of a simple random sample of size n, we have
where c is a constant, is the sample sum of squares and cross-products matrix, is the population mean value, and is the Hermitian positive definite covariance matrix. We have already established results on vector and matrix differential operators in the complex domain, operating on the trace and determinant involving a Hermitian positive definite matrix. Hence, all the terms in and are defined and such equations are already solved in our discussion of our operators operating on traces and determinants. Thus, we will see that yields the sample average as the estimate/estimator of , and , with the estimate on , yields the estimate/estimator of as . One can also examine the maximum likelihood estimation involving other scalar/vector/matrix-variate densities in the complex domain. The above are some of the open problems.
Funding
This research received no external funding.
Data Availability Statement
Data are contained within the article.
Conflicts of Interest
The author declares no conflicts of interest.
References
- Deng, X. Texture Analysis and Physical Interpretation of Polarimetric SAR Data. Ph.D Thesis, Universitat Politechnica de Catalunya, Barcelona, Spain, 2016. [Google Scholar]
- Du, L.; Liu, H.; Wang, P.; Feng, B.; Pan, M.; Bao, Z. Noise robust radar HRRP target recognition based on multitask factor analysis with small training data size. IEEE Trans. Signal Process. 2012, 60, 3546–3560. [Google Scholar]
- Hellings, C.; Gogler, P.; Utschick, W. Composite real principal component analysis of complex signals. In Proceedings of the 23rd European Signal Processing Conference (EUSIPCO), Nice, France, 31 August–4 September 2015; pp. 2216–2220. [Google Scholar]
- Horel, J.D. Complex principal component anslysis:theory and examples. J. Clim. Appl. Meteorol. 1984, 23, 21660–21673. [Google Scholar] [CrossRef]
- Katkovnik, V.; Ponomarenko, M.; Egiazarian, K. Complex-valued image denosing based on group-wise complex-domain sparsity. arXiv 2017, arXiv:1711.00362vl. [Google Scholar]
- Liu, J.; Xu, X.; Zhang, F.; Gao, Y.; Gao, W. Modeling of spatial distribution characteristics of high proportion renewable energy based on complex principal component analysis. In Proceedings of the 2020 IEEE Sustainable Power and Energy Conference (iSPEC), Chengdu, China, 23–25 November 2020; pp. 193–198. [Google Scholar]
- Morup, M.; Madsen, K.H.; Hansen, L.K. Shifted independent component analysis. In Proceedings of the 7th International Conference on Independent Component Analysis and Signal Separation: ICA2007, London, UK, 9–12 September 2007; pp. 89–96. [Google Scholar]
- Mathai, A.M.; Provost, S.B.; Haubold, H.J. Multivariate Statistical Analysis in the Real and Complex Domains; Springer Nature: Cham, Switzerland, 2022. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).