Next Article in Journal
Modeling Information Diffusion on Social Media: The Role of the Saturation Effect
Previous Article in Journal
Complex Dynamics and Intelligent Control: Advances, Challenges, and Applications in Mining and Industrial Processes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Locality-Preserving Multiprojection Discriminant Analysis

School of Computer Science and Engineering, Xi’an Technological University, Xi’an 710021, China
Mathematics 2025, 13(6), 962; https://doi.org/10.3390/math13060962
Submission received: 18 February 2025 / Revised: 10 March 2025 / Accepted: 13 March 2025 / Published: 14 March 2025

Abstract

:
Linear discriminant analysis (LDA), as an effective feature extraction method, has been widely applied in high-dimensional data analysis. However, its discriminative performance is still severely limited by the following factors. First, the restriction on the total number of features available from LDA has seriously limited its application to problems where the feature dimension is much larger than the number of classes. Second, LDA cannot deal with data containing multiple clusters (or subclasses) within a class because it cannot correctly depict the local structure of the data. To alleviate this issue, we propose a locality-preserving multiprojection discriminant analysis (LPMDA) model to extract more discriminative features preserving local structure. Specifically, LPMDA rephrases the objective function of LDA as a convex discriminant analysis framework from the perspective of metric learning, allowing for extracting more features than the number of classes. Furthermore, an auto-optimized graph technique is also integrated into the discriminant analysis framework to explore the local structure of the data. An efficient iterative optimization algorithm is presented to solve LPMDA. Extensive experiments on several benchmark datasets confirm the effectiveness of the proposed method.

1. Introduction

Dimensionality reduction (DR) [1,2,3] is an important technique for solving the curse of the dimensionality problem by transforming the original high-dimensional features into low-dimensional features that retain the most important information, thus reducing learning complexity and memory consumption. Depending on whether label information is available, DR methods can be divided into unsupervised and supervised methods. The former learns the low-dimensional features by exploiting the topological structures of unlabeled instances. In contrast, supervised methods can fully explore potential associations between categories and features by exploiting the additional labels, resulting in more discriminative low-dimensional features [4]. Consequently, supervised DR methods are receiving increasing attention.
Linear discriminant analysis (LDA) [5], as a classical supervised DR method, has been widely applied in various fields such as face recognition [6] and seismic object detection [7,8]. LDA tries to find the optimal projection directions, which maximize the between-class variance and minimize the within-class variance in the projected space. By fine-tuning the design, some enhanced variants of LDA have been proposed. For example, Yang et al. [9] combined the adjacency graph model with LDA to improve the robustness to intraclass distributional differences and thus improve the discriminative ability for class-overlapping data. Daniela et al. [10] introduced regularization terms ( l 1 and fused lasso penalties) into the LDA to further enhance its generalization ability to complex distributed data. Zhang et al. [11] proposed a Fisher discrimination multiple kernel dictionary learning framework to handle nonlinear features. Ju et al. [12] proposed a robust probabilistic linear discriminant analysis model by embedding a Kronecker decomposable component in the model for tensor data. Wang et al. [13] proposed a novel ratio sum formula for linear discriminant analysis, which maximizes the sum of the ratio of between-class differences to within-class differences in each dimension after projecting. Li et al. [14] presented a discriminative feature extraction model (MMC) based on the maximum margin criterion, which effectively tackles the issue of small sample size. Recently, benefiting from the powerful representation ability of deep neural networks (DNNs), deep discriminant analysis methods [15,16,17] have received increasing attention from researchers. However, they all require a large training dataset and training time, even with high-performance equipment (e.g., GPUs or TPUs). Although LDA and its variants are effective in many applications, there are two serious limitations. One is that LDA can only extract features one less than the number of classes because of the rank deficiency in the between-class scattering matrix [6,18,19,20,21]. This limitation may not be an obvious obstacle when the number of classes is large, but it becomes a fatal bottleneck when the number of classes is small, such as the odd–even recognition problem of handwritten digits in the subsequent experiments. Another limitation is that LDA can only preserve the global structure of data while ignoring the local structure of the data [22,23,24,25]. Regrettably, the distribution of real-world data is diverse, even for data in the same class. As a result, the class arithmetic mean point of each cluster might overlap, making LDA incapable of handling the multimodal data, which degrade its performance in practice.
To deal with the first problem, several recursive discriminant analysis methods and alternative metrics to LDA have been proposed. Okada et al. [26] presented a discriminant analysis with orthonormal coordinate axes (OFDAs) to extract an unlimited number of features. Xiang et al. [6] proposed a recursive Fisher linear discriminant algorithm (RFLD) to extract a greater number of features than the number of classes. Ohta et al. [27] proposed an incremental learning algorithm for RFLD, which can effectively extract the features online by recursively calculating the discriminant vectors. Li et al. [18] constructed a robust and sparse linear discriminant analysis (RSLDA) and obtained the multiple discriminant projections via a recursive procedure. In [28], the authors redefine the between-class scatter matrix using the Chernoff distance, which allows more features to be extracted. Zhu et al. [20] divided the data of each class into a set of subclasses to improve the rank deficiency problem and determined the optimal feature dimension using the leave-one-out-test criterion. However, the aforesaid recursive procedure or subclass partitioning requires additional prior knowledge and computational overhead. Zadeh et al. [29] characterized the similarity between data from the perspective of metric learning and proposed the geometric mean metric learning (GMML) model. GMML minimizes a strictly convex optimization problem that allows for efficient closed-form solutions but cannot be directly applied to feature extraction and does not consider the local structure of the data.
To solve the second problem, several methods based on preserving local manifold structures have been proposed, such as the local Fisher and the pairwise criteria [30,31]. Local Fisher discriminant analysis (LFDA) [22] uses label information to compute within- and between-class similarity and combines the ideas of FDA and LPP to learn projection matrices. Zhang et al. [32] further improved the LFDA by using the density region estimated from each class and neighborhood information. Zhu et al. [33] proposed neighborhood linear discriminant analysis (NLDA), which constructs the scatter matrices on a neighborhood of inverse nearest neighbors to represent the internal structure of instances. In [31], a pairwise formulation of LDA based on the nearest neighbor maximum projection (NMMP) is proposed. Zhou et al. [34] presented the manifold partition discriminant analysis (MPDA) model to learn a manifold subspace and generate the low-dimensional features. However, the similarity matrices used in the above method are derived from the original data and are sensitive to noise, resulting in suboptimal results.
Despite the excellent performance of the previous works in dimensionality reduction, most of them are still based on maximizing the trace ratio or difference between the between-class and within-class scatter matrices, making it difficult to effectively solve the above problems simultaneously. To this end, we propose a novel locality-preserving multiprojection discriminant analysis (LPMDA) model to extract a flexible number of discriminative features that fully explore the local structure of data. Specifically, LPMDA builds a novel convex discriminant analysis framework from a metric learning perspective that can extract the optimal number of discriminant features, breaking the limitation of the number of classes. In addition, an auto-optimized graph technique is introduced to the convex discriminant analysis framework to preserve the local structure of instances in the low-dimensional subspace. Finally, an alternating optimization strategy is used to solve the joint optimization model. The contributions of LPMDA are highlighted as follows:
  • A novel convex discriminant analysis framework is established from the perspective of metric learning, which can learn a flexible number of discriminative projections to extract more discriminative features.
  • An auto-optimized graph mechanism is cleverly integrated into the discriminant analysis framework, which automatically exploits the neighborships of each instance and further enhances the discriminative ability of the extracted features.
  • An efficient iterative strategy is designed to solve the resultant optimization problem. Extensive experiments were conducted on the benchmark datasets to demonstrate the superiority of the proposed method.
The rest of this article is as follows. Section 2 briefly reviews related work, including the LDA and GMML and also introduces some notations used in this article. Section 3 introduces the details of our LPMDA. Section 4 presents the experimental results and analysis. The conclusion is given in Section 5.

2. Related Works

For convenience, some notations used in this article are provided. Matrices and vectors are written in bold uppercase and bold lowercase, respectively. X = [ x 1 , x 2 , , x n ] n × d denotes a data matrix composed of n instances with d-dimensional features. Assuming that X can be classified into c classes, we record X i = [ x 1 i , x 2 i , , x n i i ] n i × d as the data submatrix with n i instances belonging to the ith class. Additionally, 1 denotes a vector with all elements that are 1. For matrix M , tr ( M ) and M 1 denote M ’s trace and inverse, respectively.

2.1. Linear Discriminant Analysis

LDA aims to find a low-dimensional representation of the data so that points within the same class are clustered together and points between different classes are pushed as far apart as possible. Therefore, it is easy to implement by following the trace ratio formulation:
max W T r ( W S b W ) T r ( W S w W ) ,
where the optimal solution W d × r is made of the eigenvectors corresponding to the first r largest generalized eigenvalues of S w 1 S b , if  S w is full rank. S b and S w denote the between-class and within-class scatter matrix, respectively, and they are defined as
S b = i = 1 c n i ( x ¯ i x ¯ ) ( x ¯ i x ¯ ) , S w = i = 1 c j = 1 n i ( x j i x ¯ i ) ( x j i x ¯ i )
Here, x ¯ i denotes the mean instance of the ith class, and  x ¯ is the mean instance of all data. To better figure out our method, we also reformulate the between-class and within-class scatter matrices in a pairwise manner with the following lemma.
Lemma 1. 
S b and S w defined by (2) can be equivalently redescribed as
S b = 1 2 n i = 1 c k = 1 c n i n k ( x ¯ i x ¯ k ) ( x ¯ i x ¯ k ) , S w = 1 2 i = 1 c j = 1 n i h = 1 n i 1 n i ( x j i x h i ) ( x j i x h i )
Proof. 
Please refer to Appendix A for the detailed proof of Lemma 1.    □
From (1) and (3), it can be observed that LDA faces the following two problems: (1) LDA can extract up to c 1 features, since the maximum rank of S b is c 1 ; when the feature dimension is much larger than the number of classes, the discriminative performance of LDA deteriorates severely. (2) LDA only considers the global structure while overlooking the local structure of the instances; when evaluating the within-class scatter matrix (3) from a graph perspective, the graph constructed in each class is fully connected with the weights all equal to 1 n i and thus cannot explore the local structure of the data.

2.2. Geometric Mean Metric Learning

GMML [29] seeks a metric matrix such that it yields smaller distances for similar instances and larger distances for dissimilar instances, which can be formulated as follows:
min M 0 T r ( M P ) + T r ( M 1 Q ) + μ D s l d ( M , M 0 )
where M is a symmetric positive definite (SPD) matrix to be learned, μ is a regularization parameter, M 0 is a prior SPD matrix about M , D s l d ( M , M 0 ) is the symmetrized LogDet divergence, and P and Q are defined as follows:
P = Σ ( x i , x j ) S ( x i x j ) ( x i x j ) , Q = Σ ( x i , x j ) D ( x i x j ) ( x i x j )
Here, S denotes the set of instances of the same class, and D denotes the set of instances from different classes. GMML tries to learn a Mahalanobis distance, i.e., find an SPD matrix M that yields small distances for instances in the same class. Different from most existing metric learning methods, which treat the instances of different classes asymmetrically, GMML measures the distances between the instances of different classes using M 1 . This operation turns (4) into a strictly (geodesically) convex optimization problem and allows for a closed-form solution. For details about the strictly geodesically convex problem, this can be found in [29]. Given the excellent performance of GMML, we introduce it as a similarity measurement tool in our LPMDA.

3. Proposed Method

In this section, we propose a novel discriminant analysis framework for dimensionality reduction, which can extract more features than the number of classes, while fully preserving the local structure between instances. The proposed LPMDA mainly consists of two basic modules, i.e., multiprojection learning with adaptive metric learning and local structure exploration with auto-optimized graphs. Figure 1 illustrates the overall framework of our proposed method. The technical details of each component are detailed as follows.

3.1. Multiprojection Discriminant Analysis

As mentioned above, it is difficult to extract enough discriminative features by extension directly from LDA. To address this problem, we make a simple but crucial transformation to the optimization problem of LDA. According to the properties of the trace operation, the optimization objective (1) can be reformulated as
max W T r ( W W S b ) T r ( W W S w ) ,
To generate multiple discriminant projections for each instance, we first introduce an SPD matrix M d × d to replace W W in (6) and obtain
max M 0 T r ( M S b ) T r ( M S w ) .
It should be noted that optimization problem (7) is not convex with respect to M , and solving it usually involves iterative techniques, leading to considerable computational complexity. To maximize the variance of the intraclass class while maximizing those of the interclass, we embed a metric matrix M in the numerator term and its inverse matrix M 1 in the denominator term of (7), respectively, and sum them. Therefore, it becomes
min M 0 T r ( M S w ) + T r ( M 1 S b ) + μ D s l d ( M , M 0 ) .
where μ is a regularization parameter, M 0 is the prior knowledge of the SPD matrix, and  D s l d ( M , M 0 ) is the symmetrized LogDet divergence [29], i.e.,
D s l d ( M , M 0 ) = t r ( M M 0 1 ) + t r ( M 1 M 0 ) 2 d
Replacing the tracking ratio operation in (7) with the asymmetric operation in GMML [29] effectively avoids the rank deficiency in between-class scattering and thus yields more discriminative projections. Model (8) aims to find a metric space measured by M such that instances of each class are close to the class centroid while the centroids of different classes are far apart in this metric space. Our goal is to obtain the multiple discriminant projections for each instance, which is nothing but the Cholesky decomposition of M . Therefore, the goal of model (8) fully aligns with the core motivation of LDA but can extract any number of discriminative features. This is a breakthrough improvement over the original LDA. It is worth noting that the metric matrix M in our model is essentially different from that in GMML. This is because S w and S b in (8) are only the within-class scatter matrix and the between-class scatter matrix, whereas P and Q in (4) are scaled second sample moments of the differences between similar instances and the differences between dissimilar instances. Therefore, model (8) provides a more refined description of data distribution and class information.

3.2. Local Structure Exploitation

Referring to Equation (3), classical LDA methods calculate the within-class scatter matrix S w using a fully connected equal-weight graph, which cannot explore the local geometric structure of the instances. To this end, we embed an auto-optimized graph into the within-class scatter to adaptively exploit the local geometric structure of the instances as follows:
S ˜ w = i = 1 c j = 1 n i h = 1 n i Φ j , h i n i ( x j i x h i ) ( x j i x h i ) , s . t . Φ j , : i 1 = k , Φ j , h i { 0 , 1 } ,
where Φ j , h i is a kNN graph to be learned, which can be viewed as the weight between x j i and x h i , and Φ j , : i denotes the weights of x j i . In short, the kNN graph Φ can be automatically optimized, and the optimal neighborships between instances can be found as the algorithm converges, thus exploring the submanifold structure of the instances.

3.3. Objective Function

By incorporating both the multiprojection discriminant analysis and local structure exploitation, we remodel LDA and achieve the objective of the proposed LPMDA as
min M 0 T r ( M S ˜ w ) + T r ( M 1 S ˜ b ) + μ D s l d ( M , M 0 ) , s . t . Φ j , : i 1 = k , Φ j , h i { 0 , 1 }
The proposed model (11) is more appealing than the LDA on the basis of the following two considerations. First, the total number of features available in LPMDA is independent of the number of classes, thus overcoming the major drawback that the total number of features available from LDA is limited to the number of classes minus one. Second, LPMDA embeds an auto-optimized k-nearest neighbor graph in the framework, which allows the model to fully exploit not only the differences between instances from different classes but also the local structural information between instances from the same class, further enhancing the discriminativeness of the extracted features. Model (11) combines metric learning and adaptive graph learning to construct a unified discriminant analysis framework, which can also be refined by introducing more sophisticated tricks such as nonlinear metric learning and hypergraph learning.

3.4. Optimization for LPMDA

In this section, we present an iterative optimization algorithm to solve problem (11). First, the distance matrix in the original space is calculated by Euclidean distance, and the similarity matrix Φ is initialized by the rule that if instance x h i is the kNN point of x j i , then the element Φ j , h i is equal to 1; otherwise, it is 0. We will then iteratively optimize M and Φ as follows.
When fixing Φ , M is updated by solving the following problem:
min M 0 T r ( M S ˜ w ) + T r ( M 1 S ˜ b ) + μ D s l d ( M , M 0 ) .
The optimal solution of (12) is the midpoint of the geodesic joining S ˜ w 1 to S ˜ b [35]; that is,
M = ( S ˜ w + μ M 0 1 ) 1 α ( S ˜ b + μ M 0 ) ,
where α ( · ) is defined as A α B = A 1 / 2 ( A 1 / 2 B A 1 / 2 ) α A 1 / 2 , and α ( 0 , 1 ) is a weighting parameter.
When fixing M , Φ is updated by solving the following problem:
min Φ i = 1 c 1 n i j = 1 n i h = 1 n i Φ j , h i T r ( M ( x j i x h i ) ( x j i x h i ) ) , s . t . Φ j , : i 1 = k , Φ j , h i { 0 , 1 }
which is equivalent to solving the following subproblem:
min Φ j , : i h = 1 n i Φ j , h i x j i x h i M 2 , s . t . Φ j , : i 1 = k , Φ j , h i { 0 , 1 } ,
where x j i x h i M 2 = ( x j i x h i ) M ( x j i x h i ) denotes the square Mahalanobis distance between x j i and x h i . It is obvious that there are only k elements equal to 1 in the vector Φ j , : i , and the others are all 0, because of the constraints on Φ j , : i . Assuming that the indices of the non-zero elements in the vector Φ j , : i are { τ 1 , τ 2 , , τ m , , τ k } ( 1 τ m n i , τ m j ) , then problem (15) can be reduced as
min τ m m = 1 k x j i x τ m i M 2 .
Obviously, the optimal solution τ m * of problem (16) is the indices of the instances of the kNNs of x j i under the Mahalanobis distance based on M . Then, the optimal solution to (15) is
Φ j , h i = 1 if h = τ m * 0 otherwise ,
which is consistent with the motivation to find neighbors in the optimal subspace to eliminate the effects of noise.
Following the iterative optimization of M and Φ , our method is able to automatically assign k optimal neighbors to each instance. Based on the optimal M , we can obtain the projection matrix W with flexible dimensionality and the capability to preserve the local structure of the instances, which is exactly the Cholesky decomposition of M , i.e., satisfying M = W W . Algorithm 1 summarizes the procedure of LPMDA in which the convergence condition used in our experiments is W k + 1 W k F W k F 10 4 .
Algorithm 1 Algorithm for solving LPMDA
Input: Training instance matrix X , label matrix Y , regularization parameter μ , weighting parameter α ( 0 , 1 ) , neighbors number k, reduced dimension r.
Initialization:  Φ n × n
While not converged do
1. Update M by solving problem (12)
2. Calculate Φ according to Equation (15).
end while
W : = C h o l ( M ) , where C h o l ( · ) denotes the Cholesky decomposition operation.
Output: W d × r .

3.5. Complexity and Convergence Analysis

Obviously, the cost of LPMDA consists of two parts. The first part is to find the SPD matrix M , which requires the Cholesky decomposition of S ˜ b and S ˜ w , with a complexity of approximately O ( d 3 ) . The second part is to solve Φ , which only involves some basic operations such as vector subtraction and multiplication, which can be ignored compared to the first part. Thus, the total computational complexity of LPMDA is O ( t d 3 ) , where t is the number of iterations. For a more detailed comparison, Table 1 reports the time complexity of the comparison methods in this paper. It is obvious that our model will eventually converge to a local optimum, since the optimal solution can be obtained at each optimization step.

3.6. Connections to Previous Discriminators

To provide a deeper insight into our approach, this section details the differences between our LPMDA and several closely related discriminators. Indeed, most existing discriminators can be unified into the following framework:
min W L ( S b , S w , W ) s . t . W Ω
where L ( · ) denotes the optimization criterion; S b and S w are the generalized between- and within-class scatter matrices of instances, respectively, which may be defined differently in different methods; W is the projection matrix; and Ω denotes the constraint for W .
Table 2 summarizes some representative methods and our LPMDA in the form of framework (18). These methods typically improve feature extraction from the following two core points: (1) Designing different optimization criteria or iterative strategies to extract more features. For example, MMC [14] avoids the small sample size problem by employing the maximum margin criterion; RFLD [6] and RSLDA [18] extract more discriminative features using an iterative strategy. (2) Embedding graph similarity or neighborhood information in the scatter matrix to fully exploit the local structure of instances. For example, NMMP [31], NLDA [33], and MPDA [34] introduce graph similarity into the scatter matrix to fully exploit the local structure of instances. Our LPMDA starts from a new perspective of metric learning, incorporating adaptive graph learning to extract more discriminative features, taking into account local information, without requiring iterative learning.

4. Experiments

In this section, LPMDA is compared with several representative related methods, including classical LDA [21], MMC [14], RFLD [6], and RSLDA [18], which can extract more features than the number of classes, and NMMP [31], NLDA [33], and MPDA [34], which take into account the local structure of the sample. All parameters of the comparative methods were set according to the suggestions of the relevant literature. In our LPMDA, the regularization parameter μ is set to μ = 10 4 to avoid matrix singularity, α ( 0 , 1 ) is the weighting parameter and is selected from the set { 10 6 , 10 5 , , 10 2 } , and the number of neighbors per class k is set to 5. The prior matrix M 0 was set to the identity matrix throughout the experiment. All the parameters in these models are selected from their own candidate set by the five-fold cross-validation technique on the corresponding training datasets. First, we evaluate the ability of LPMDA to explore the local structural information of instances in multimodal classes. Specifically, two handwritten digit recognition datasets are used to identify whether a digit is odd or even. Then, we further evaluate LPMDA on six public benchmark datasets, for which we do not know whether they contain a multimodal class or not.
All experiments are run 10 times with random data splits of training and test samples, and then the mean classification results with standard deviation are reported on different datasets. All experiments are implemented on MATLAB R2017a with Win7 system, Inter Core i7-8550 CPU, and 8 GB RAM.

4.1. Experiments on Handwritten Digit Recognition

In this subsection, we will use two handwritten digit datasets, including USPS and MNIST, to evaluate the proposed method. The details of these datasets are given in Table 3, where ♯ indicates the corresponding number of instances. For each dataset, we aim to identify whether the digit is odd or even. Therefore, handwritten digit recognition is transformed into a binary classification problem, where each class consists of five subclasses, and some digital images are shown in Figure 2.
In these experiments, each gray image is resized to 16 × 16 with 256-D features. For fair comparison, the reduced dimensions of LDA are set to c 1 ( c = 2 ), and the optimal dimensions of other methods are searched from [10:10:200]. All the experiments are repeated 20 times. Finally, the average classification accuracies with standard deviations are reported in Table 4. From Table 4, we can see that LDA performs poorly on these two datasets, mainly because each class is multimodal and contains five subclasses, while LDA does not consider the local structure between instances. In addition, the number of features generated by LDA is much smaller than the feature dimension, which is another major reason for its performance degradation. Two locality-preserved methods including NMMP and NLDA outperform LDA and MMC, confirming the need to explore the local structure to discriminate the multimodal data. LPMDA achieves the best performance among all the comparison methods. Specifically, for the USPS data, the accuracy of LPMDA improves by 4.87% over DLSR; about 3.56% over MMC; about more 2% over NMMP, RFLD and RSLDA; about more 1% over NLDA; and about 0.88% over the best competitor, MPDA. For MNIST data, LPMDA increased by 6.57% compared to LDA, 5.23% compared to MMC, and even more compared to NMMP; RFLD increased by more than 4%, which is more than 2% higher than RSLDA and NLDA, and about 1.22% higher than the best competitor, MPDA. This is because LPMDA can extract enough discriminative features that preserve the local structure of the instances.
To further validate the ability of our model LPMDA to capture local structures between instances, we conducted a 2-D visualization experiment on a subset of the USPS, which merges the ‘3’ and ‘9’ into an odd class (‘’ and ‘’) and denotes the ‘6’ as the even class (‘’). We compare the embedding results of LPMDA with those of LDA, NMMP, MPDA, NLDA, and evaluate the between-class separability (i.e., ‘’ and ‘’ are well separated from ‘’) and the within-class multimodality preservation ability (i.e., ‘’ and ‘’ are well grouped). Figure 3 shows the instances embedded in the two-dimensional space found by each method. The horizontal axis is the first feature found by each method, while the vertical axis is the second feature. LDA allows us to extract only one meaningful feature in binary class classification problems (see Section 2.1), so here, we choose the second feature at random. We also visualize the t-SNE of the original features [36] as a reference. From Figure 3, we can see that LDA tends to mix instances from different classes, which would be caused by the within-class multimodality. NMMP and MPDA separate ‘’ and ‘’ from ‘’ well, but the within-class multimodality is lost, i.e., ‘’ and ‘’ are mixed. NLDA separates odd classes from even classes well and preserves the within-class multimodality to some extent. LPMDA nicely separates instances in different classes from each other, while clearly preserving the within-class multimodality.

4.2. Experiments on Benchmark Datasets

In this subsection, we use four datasets, i.e., Connect 4, DNA, SVMGuide2, and Segment, which are available from the UCI database website, as well as one object data Coil20 and one face image dataset PIE [37], to evaluate the proposed discriminator. The PIE face dataset contains 68 people with 41368 face images in total. We select a subset of this dataset, where each person has 170 images that are collected under five different poses (C05, C07, C09, C27, and C29). The details of these datasets are listed in Table 5. The number of classes ranges from 3 to 68. The number of features ranges from 19 to 1024. The reduced dimensions of LDA are set to c 1 . The optimal dimensions of all the other compared methods are searched from [2:2:100] on Connect 4 and DNA, [2:1:20] on SVMGuide2, [2:2:18] on Segment, [10:5:200] on Coil20, and [10:10:200] on PIE. We randomly selected 60 % and 40 % of the samples from all the data for training and testing, and the experimental results are shown in Figure 4. We can see that LPMDA outperforms the other discriminators on all the benchmark datasets. These methods, including MMC, NMMP, RFLD, RSLDA, MPDA, NLDA, and LPMDA, eliminate the limitation on the number of available features and preserve more discriminative information, resulting in a higher average accuracy than LDA, especially for small-class datasets. MPDA, NLDA and LPMDA fully consider the local structural information of the instances and perform better than the other methods. Our LPMDA achieved the best performance mainly due to its combination of auto-optimized graph mechanisms and metric learning, making it capable of extracting an infinite number of discriminative features containing local structural information.

4.3. Statistical Significance

The Friedman test [38,39] was used as the statistical test to systematically analyze the relative performance between the proposed method and baselines. In the Friedman test, the null hypothesis states that all of the algorithms are equivalent, with equal average ranks. If the null hypothesis is rejected, a post hoc Nemenyi test is used for the identification of algorithms that are significantly different. In the experiments, two algorithms are considered to have significantly different performances if the average ranks of the two algorithms differ by at least one critical difference (CD). Figure 5 shows the CD diagrams for the eight comparison methods on the eight benchmark datasets, where the average rank of each comparison method is marked along the axis. The axis is turned so that the lowest ranks (best performance) are on the right. Algorithms whose average ranks are within one CD of LPMDA are connected by a red line, and those that are not are considered to have a significantly different performance to LPMDA. As shown in Figure 5, our proposed method, LPMDA, ranks first and significantly outperforms NMMP, RFLD, RSLDA, MMC, and LDA. LPMDA formulates the discriminative feature extraction as a strict (geodesic) convex optimization problem with an auto-optimized kNN graph constraint, which can extract the features that are not limited in number by the number of classes and explores the local structural information between instances. This makes LPMDA achieve a competitive performance against the compared methods.

4.4. Convergence and Computational Performance

In this section, we provide the convergence curves of the proposed method over all the used datasets. Considering that the main focus of our discriminant analysis model is to learn the discriminative projection matrix W , we directly use the relative error W k + 1 W k F W k F 10 4 as the convergence condition for Algorithm 1, where W k denotes the value of W at the k-th iteration. Figure 6 shows the convergence curves of the relative error versus the number of iterations on different datasets. From Figure 6, we can see that our iterative optimization algorithm converges in less than ten iterations for most of the datasets, which indicates the efficiency of our iterative optimization algorithm.
To further investigate the computational performance of our method, Table 6 reports the training time of LPMDA and other related methods on the eight benchmark datasets. All experiments were set up as in the previous sections, and the results in Table 6 are the average training time over 20 repeated experiments. LDA has the shortest training time while MPDA has the longest training time. This is because MPDA requires computationally intensive matrix factorization operations. Our LPMDA runs slightly more slowly than LDA and MMC but achieves a marked improvement in discriminative performance. RFLD and RSLDA require an iterative solving of projection vectors, and NMMP and NLDA involve a complex computation of similarity graphs, and thus, both run slower than LPMDA.

4.5. Ablation Study

In LPMDA, the objective function contains two key schemes: geometric mean metric learning and auto-optimized graph embedding. To study the role each scheme plays in the dimension reduction problem, we conducted ablation experiments. The two degenerate versions of LPMDA are as follows:
  • LPMDA1: LPMDA1 only replaces the trace ratio operation in LDA with the GMML similarity measure to obtain optimization problem (8), which can yield more features than the number of classes.
  • LPMDA2: LPMDA2 incorporates auto-optimized graph embedding in LDA, resulting in the following optimization problem:
    max W , Φ T r ( W S b W ) T r ( W S ˜ w W ) , s . t . Φ j , : i 1 = k , Φ j , h i { 0 , 1 }
where S b is defined as (2), and S ˜ w = i = 1 c j = 1 n i h = 1 n i Φ j , h i n i ( x j i x h i ) ( x j i x h i ) .
Table 7 reports the ablation results of the proposed LPMDA on the eight benchmark datasets. From the table, it can be seen that LPMDA achieved the best performance on all datasets. For data with small class sizes, e.g., USPS MNIST, Connect 4, and DNA SVMGuide2, LPMDA 1 outperforms LPMDA 2. This is because LPMDA2 can extract much fewer features than LPMDA1, resulting in insufficient discriminative information. For Segment and Coil20, the number of features extracted by LPMDA2 is not significantly different from LPMDA1, but LPMDA2 embeds the additional local information, giving slightly better performance than LPMDA1. Therefore, we can conclude that both geometric mean metric learning and auto-optimized graph embedding can improve the performance of LPMDA.

4.6. Parameter Sensitiveness

In order to obtain a more detailed view, we carried out experiments to see the effect of k on the performance of the proposed LPMDA. The accuracy varies with the number of neighbors per class (k), as shown in Figure 7. It can be seen that on these datasets, the performance of the LPMDA first improves and then stabilizes as k increases from 1 to 10. Furthermore, there is no significant difference in the performance of LPMDA on most datasets as k increases from 5 to 10. To balance the performance and efficiency, k = 5 was set in the experiments.

5. Conclusions and Future Work

In this paper, we present a novel locality-preserving multiprojection discriminant analysis (LPMDA) framework that can directly extract enough discriminative features that preserve local structural information. Unlike the existing LDA-based methods, our proposed LPMDA maximizes the variance of the intraclass class while maximizing those of the interclass under the geometric mean metric learning framework, which can extract more features than the number of classes. Furthermore, an auto-optimized graph technique is introduced into the learning framework to fully exploit the local structure of the data from each class. More importantly, LPMDA has only one parameter to tune, making it very convenient to implement in real-world applications. Regarding the experiment, the odd–even digit recognition experiments on USPS and MNIST verified the effectiveness of LPMDA for data containing multiple clusters (or subclasses) within a class. The results and corresponding quantitative analysis in Table 4 indicate that LPMDA is significantly superior to other baselines, further validating its effectiveness in handling intraclass multimodal and small-class data. The experimental results on the six benchmark datasets in Figure 4 and the statistical significance results in Figure 5 further demonstrate the effectiveness of LPMDA in feature extraction for real data. The results in Table 6 indicate that the training time of LPMDA is slightly longer than that of LDA and MMC. However, all the computational costs associated with LPMDA only occur in the learning of the projection matrix. Once the features are extracted and ready for subsequent tasks, there will be no extra cost. In addition, the superior performance of LPMDA is sufficient to convince us that it is worth putting up with the computation overhead.
LPMDA provides a unified framework for feature extraction, and some more sophisticated metric learning or graph learning techniques can be easily plugged in to further extend it for some complex industrial big data. For example, introducing nonlinear metric learning or hypergraph learning techniques further enhances the model’s ability to handle nonlinear or class-overlapping data. The main computational complexity of LPMDA is focused on matrix operations, and the use of the instance selection technique to reduce its computational complexity is another major line of research.

Funding

This research was funded by the Education Department of Shaanxi Province, grant number 23JK0475.

Data Availability Statement

The USPS and MNIST datasets are openly available at http://www.cad.zju.edu.cn/home/dengcai/Data/MLData.html, accessed on 11 November 2024; Coil20 and PIE datasets are available upon request from the authors; and the other datasets are available from the UCI database website at https://archive.ics.uci.edu/datasets, accessed on 22 November 2024.

Acknowledgments

The authors would like to thank the editor, the associate editor, and the anonymous reviewers for their constructive comments and suggestions.

Conflicts of Interest

The authors declare that there are no conflicts of interest in this manuscript.

Appendix A

Proof. 
According to the formulation of S w , we have
S w = i = 1 c j = 1 n i ( x j i x ¯ i ) ( x j i x ¯ i ) = i = 1 c j = 1 n i ( x j i 1 n i h = 1 n i x h i ) ( x j i 1 n i h = 1 n i x h i ) = i = 1 c j = 1 n i x j i x j i i = 1 c 1 n i j = 1 n i x j i h = 1 n i x h i = i = 1 c 1 n i j = 1 n i h = 1 n i x j i x j i i = 1 c 1 n i j = 1 n i x j i h = 1 n i x h i = 1 2 i = 1 c j = 1 n i h = 1 n i 1 n i ( x j i x h i ) ( x j i x h i )
Similarly, according to the definition of S b , we have
S b = i = 1 c n i ( x ¯ i x ¯ ) ( x ¯ i x ¯ ) = i = 1 c n i x ¯ i x ¯ i 1 n i = 1 c n i x ¯ i k = 1 c n k x ¯ k = 1 2 n i = 1 c k = 1 c n i n k ( x ¯ i x ¯ k ) ( x ¯ i x ¯ k ) .
Thus, the proof is complete. □

References

  1. Wang, J.; Wang, L.; Nie, F.; Li, X. Fast Unsupervised Projection for Large-Scale Data. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 3634–3644. [Google Scholar] [CrossRef] [PubMed]
  2. Dorrity, M.W.; Saunders, L.M.; Queitsch, C.; Fields, S.; Trapnell, C. Dimensionality reduction by UMAP to visualize physical and genetic interactions. Nat. Commun. 2020, 11, 1537. [Google Scholar] [CrossRef] [PubMed]
  3. Yang, A.; Yu, W.; Zi, Y.; Chow, T. An Enhanced Trace Ratio Linear Discriminant Analysis for Fault Diagnosis: An Illustrated Example Using HDD Data. IEEE Trans. Instrum. Meas. 2019, 68, 4629–4639. [Google Scholar] [CrossRef]
  4. Ma, J.; Xu, F.; Rong, X. Discriminative multi-label feature selection with adaptive graph diffusion. Pattern Recognit. 2024, 148, 110154. [Google Scholar] [CrossRef]
  5. Yin, W.; Ma, Z.; Liu, Q. Discriminative subspace learning via optimization on Riemannian manifold. Pattern Recognit. 2023, 139, 109450. [Google Scholar] [CrossRef]
  6. Xiang, C.; Fan, X.A.; Lee, T.H. Face Recognition Using Recursive Fisher Linear Discriminant. IEEE Trans. Image Process. 2006, 15, 2097–2105. [Google Scholar] [CrossRef]
  7. Hemmatpour, S.; Hashemi, H. Using PCA and RDA feature reduction techniques for ranking seismic attributes. J. Earth Space Phys. 2011, 37, 217–227. [Google Scholar]
  8. Fakhari, M.G.; Hashemi, H. Fisher Discriminant Analysis (FDA), a supervised feature reduction method in seismic object detection. Geopersia 2019, 9, 141–149. [Google Scholar]
  9. Yan, S.; Xu, D.; Zhang, B.; Zhang, H.; Yang, Q.; Lin, S. Graph Embedding and Extensions: A General Framework for Dimensionality Reduction. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 40–51. [Google Scholar] [CrossRef]
  10. Witten, D.M.; Tibshirani, R. Penalized Classification Using Fisher’s Linear Discriminant. J. R. Stat. Soc. Ser.-Stat. Methodol. 2011, 73, 753–772. [Google Scholar] [CrossRef]
  11. Zhang, X.; Yang, T.; Long, H.; Shi, H.; Wang, J.; Yang, L. Fisher discrimination multiple kernel dictionary learning for robust identification of nonlinear features in machinery health monitoring. Inf. Sci. 2024, 677, 120862. [Google Scholar] [CrossRef]
  12. Ju, F.; Sun, Y.; Gao, J.; Hu, Y.; Yin, B. Kronecker-decomposable robust probabilistic tensor discriminant analysis. Inf. Sci. 2021, 561, 196–210. [Google Scholar] [CrossRef]
  13. Nie, F.; Wang, J.; Wang, H.; Li, X. Ratio Sum Versus Sum Ratio for Linear Discriminant Analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1936, 44, 10171–10185. [Google Scholar]
  14. Li, H.; Jiang, T.; Zhang, K. Efficient and robust feature extraction by maximum margin criterion. IEEE Trans. Neural Netw. Learn. Syst. 2006, 17, 157–165. [Google Scholar] [CrossRef]
  15. Hayes, T.L.; Kanan, C. Lifelong Machine Learning with Deep Streaming Linear Discriminant Analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 887–896. [Google Scholar]
  16. Bartan, B.; Pilanci, M. Neural Fisher Discriminant Analysis: Optimal Neural Network Embeddings in Polynomial Time. In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; pp. 1647–1663. [Google Scholar]
  17. Uzun, B.; Cevikalp, H.; Saribas, H. Deep Discriminative Feature Models (ddfms) for Set Based Face Recognition and Distance Metric Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 5594–5608. [Google Scholar] [CrossRef] [PubMed]
  18. Li, C.; Shao, Y.; Yin, W.; Liu, M. Robust and Sparse Linear Discriminant Analysis via an Alternating Direction Method of Multipliers. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 915–926. [Google Scholar] [CrossRef]
  19. Loog, M.; Ginneken, B.V.; Duin, R. Dimensionality reduction of image features using the canonical contextual correlation projection. Pattern Recognit. 2005, 38, 2409–2418. [Google Scholar] [CrossRef]
  20. Zhu, M.; Martinez, A.M. Subclass Discriminant Analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1274–1286. [Google Scholar]
  21. Webb, A.R. Introduction to Statistical Pattern Recognition. In Statistical Pattern Recognition; John Wiley & Sons: Hoboken, NJ, USA, 2003. [Google Scholar]
  22. Sugiyama, M. Dimensionality reduction of multimodal labeled data by local fisher discriminant analysis. J. Mach. Learn. Res. 2007, 8, 1027–1061. [Google Scholar]
  23. Li, Z.; Nie, F.; Chang, X.; Yi, Y. Beyond trace ratio: Weighted harmonic mean of trace ratios for multiclass discriminant analysis. IEEE Trans. Knowl. Data Eng. 2017, 29, 2100–2110. [Google Scholar] [CrossRef]
  24. Wa, H.; Wang, H.; Guo, G.; Wei, X. Separability-Oriented Subclass Discriminant Analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 409–422. [Google Scholar]
  25. Luo, T.; Hou, C.; Nie, F.; Yi, D. Dimension reduction for non-gaussian data by adaptive discriminative analysis. IEEE Trans. Cybern. 2019, 49, 933–946. [Google Scholar] [CrossRef]
  26. Okada, T.; Tomita, S. An optimal orthonormal system for discriminant analysis. Pattern Recognit. 1985, 18, 139–144. [Google Scholar] [CrossRef]
  27. Ohta, R.; Ozawa, S. An Incremental Learning Algorithm of Recursive Fisher Linear Discriminant. In Proceedings of the International Joint Conference on Neural Networks, Atlanta, GA, USA, 14–19 June 2009; pp. 2310–2315. [Google Scholar]
  28. Loog, M.; Duin, R. Linear Dimensionality Reduction via a Heteroscedastic Extension of LDA: The Chernoff Criterion. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 732–739. [Google Scholar]
  29. Zadeh, P.; Hosseini, R.; Sra, S. Geometric mean metric learning. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016; pp. 2464–2471. [Google Scholar]
  30. Fan, Z.; Xu, Y.; Zhang, D. Local Linear Discriminant Analysis Framework Using Sample Neighbors. IEEE Trans. Neural Netw. Learn. Syst. 2011, 22, 1119–1132. [Google Scholar] [CrossRef]
  31. Nie, F.; Xiang, S.; Zhang, C. Neighborhood MinMax Projections. In Proceedings of the International Joint Conference on Artificial Intelligence, Hyderabad, India, 6–12 January 2007; pp. 993–998. [Google Scholar]
  32. Zhao, Z.; Chow, T. Robust linearly optimized discriminant analysis. Neurocomputing 2012, 79, 140–157. [Google Scholar]
  33. Zhu, F.; Gao, J.; Yang, J.; Ye, N. Neighborhood linear discriminant analysis. Pattern Recognit. 2022, 123, 108422. [Google Scholar] [CrossRef]
  34. Zhou, Y.; Sun, S. Manifold Partition Discriminant Analysis. IEEE Trans. Cybern. 2017, 47, 830–840. [Google Scholar] [CrossRef]
  35. Bhatia, R. Positive Definite Matrices. In Princeton; Princeton University Press: Princeton, NJ, USA, 2009; Volume 24. [Google Scholar]
  36. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  37. Sim, T.; Baker, S.; Bsat, M. The cmu pose, illumination, and expression database. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1615–1618. [Google Scholar]
  38. Demiar, J.; Schuurmans, D. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  39. Ma, J.; Zhou, S. Discriminative least squares regression for multiclass classification based on within-class scatter minimization. Appl. Intell. 2022, 52, 622–635. [Google Scholar] [CrossRef]
Figure 1. The overall framework of the proposed LPMDA. LPMDA aims to extract the low-dimensional discriminative features that preserve the local structure information between instances.
Figure 1. The overall framework of the proposed LPMDA. LPMDA aims to extract the low-dimensional discriminative features that preserve the local structure information between instances.
Mathematics 13 00962 g001
Figure 2. The digit samples in USPS and MNIST.
Figure 2. The digit samples in USPS and MNIST.
Mathematics 13 00962 g002
Figure 3. Visualization results of (a) T-SNE results of the original features; embedded features generated by (b) LDA, (c) NMMP, (d) MPDA, (e) NLDA, and (f) LPMDA. Note that all features are normalized to [−1,1].
Figure 3. Visualization results of (a) T-SNE results of the original features; embedded features generated by (b) LDA, (c) NMMP, (d) MPDA, (e) NLDA, and (f) LPMDA. Note that all features are normalized to [−1,1].
Mathematics 13 00962 g003aMathematics 13 00962 g003b
Figure 4. Average accuracy with standard deviations (%) of seven baselines and our LPMDA on the six benchmark datasets. The red numbers below the x-axis denote the optimal reduced dimensions of different methods on the corresponding data.
Figure 4. Average accuracy with standard deviations (%) of seven baselines and our LPMDA on the six benchmark datasets. The red numbers below the x-axis denote the optimal reduced dimensions of different methods on the corresponding data.
Mathematics 13 00962 g004
Figure 5. CD diagram of different methods with significance level α = 0.05 .
Figure 5. CD diagram of different methods with significance level α = 0.05 .
Mathematics 13 00962 g005
Figure 6. Convergence curve of USPS, MNIST, Connect 4, DNA, SVMGuide2, Segment, Coil20, and PIE generated by LPMDA.
Figure 6. Convergence curve of USPS, MNIST, Connect 4, DNA, SVMGuide2, Segment, Coil20, and PIE generated by LPMDA.
Mathematics 13 00962 g006
Figure 7. Accuracy varies with the number of neighbors per class (k) on the datasets.
Figure 7. Accuracy varies with the number of neighbors per class (k) on the datasets.
Mathematics 13 00962 g007
Table 1. Time complexity of the comparison methods, where t denotes the number of iterations, and N p is the number of submanifold partitions in MPDA.
Table 1. Time complexity of the comparison methods, where t denotes the number of iterations, and N p is the number of submanifold partitions in MPDA.
MethodComputational Complexity
LDA [21] O ( n d 2 + d 3 )
MMC [14] O ( n d 2 + t d 3 )
NMMP [31] O ( n 2 d + t n r d )
RFLD [6] O ( t ( n d 2 + d 3 ) )
RSLDA [18] O ( n 3 + t n c 2 )
MPDA [34] O ( ( n + d ) 3 + p = 1 P ( d 2 N p + N p 2 d ) )
NLDA [33] O ( n 2 + d 3 )
LPMDA O ( t d 3 )
Table 2. Descriptions of representative advanced discriminators and our LPMDA based on framework (18). ‘♯ Features’ denotes the maximum number of available features, x ¯ i and x ¯ are the mean vectors of X i and X , and N N k ( x ) denotes the k-nearest neighbors of x .
Table 2. Descriptions of representative advanced discriminators and our LPMDA based on framework (18). ‘♯ Features’ denotes the maximum number of available features, x ¯ i and x ¯ are the mean vectors of X i and X , and N N k ( x ) denotes the k-nearest neighbors of x .
Methods L ( X , Y , W , V ) Definition of S b and S w ♯ FeaturesStructure
Preservation
LDA [21] W = arg max W T r ( W S b W ) T r ( W S w W )   S b = i = 1 c n i ( x ¯ i x ¯ ) ( x ¯ i x ¯ )
S w = i = 1 c j = 1 n i ( x j i x ¯ i ) ( x j i x ¯ i )  
c 1 global
MMC [14] W = arg max W W = I T r ( W ( S b S w ) W )   S b = i = 1 c n i ( x ¯ i x ¯ ) ( x ¯ i x ¯ )
S w = i = 1 c j = 1 n i ( x j i x ¯ i ) ( x j i x ¯ i )  
c global
RFLD [6] w ( k + 1 ) = arg max w w ( k ) = 1 w S b ( k ) w w S w ( k ) w   x i ( k + 1 ) = x i ( k ) w ( k ) x i ( k ) w ( k )
S b ( k + 1 ) = i = 1 c n i ( x ¯ i ( k + 1 ) x ¯ ( k + 1 ) ) ( x ¯ i ( k + 1 ) x ¯ ( k + 1 ) )
S w ( k + 1 ) = i = 1 c j = 1 n i ( x j i ( k + 1 ) x ¯ i ( k + 1 ) ) ( x j i ( k + 1 ) x ¯ i ( k + 1 ) )  
c global
RSLDA [18] w ( k + 1 ) = arg min w i = 1 c n i | w x i ( k + 1 ) x ¯ ( k + 1 ) |
+ λ w S w ( k + 1 ) w + δ w 1
  x i ( k + 1 ) = x i ( k ) j = 1 k w ( j ) x i ( k ) w ( j )
S w ( k + 1 ) = i = 1 c j = 1 n i ( x j i ( k + 1 ) x ¯ i ( k + 1 ) ) ( x j i ( k + 1 ) x ¯ i ( k + 1 ) )  
c global
NMMP [31] W = arg max W W = I T r ( W S b W ) T r ( W S w W )   N b ( C j ) = { x q | x q N N k ( x j ) , C q C j }
N w ( C j ) = { x q | x q N N k ( x j ) , C q = C j }
S b = i , j : x i N b ( C j ) & x j N b ( C i ) ( x i x j ) ( x i x j )
S w = i , j : x i N w ( C j ) & x j N w ( C i ) ( x i x j ) ( x i x j )  
c local
NLDA [33] W = arg max W T r ( ( W S w W ) 1 ( W S b W ) )   R N N k ( x i , X ) = { x q | x q X x i , x i N N k ( x q ) }
m i = 1 | R N N k ( x i , X y i ) | x j R N N k ( x i , X y i ) x j
S b = i = 1 , | R N N k ( x i , X y i ) | t n j = 1 , | R N N k ( x j , X y j ) | t , y i y j n ( m i m j ) ( m i m j )
S w = i = 1 , | R N N k ( x i , X y i ) | t n x j R N N k ( x i , X y i ) ( x j m i ) ( x j m i )  
c local
LPMDA M = arg min M 0 , Φ T r ( M S ˜ w ) + T r ( M 1 S ˜ b )
+ μ D s l d ( M , M 0 )
  S ˜ b = 1 2 n i = 1 c k = 1 c n i n k ( x ¯ i x ¯ k ) ( x ¯ i x ¯ k )
S ˜ w = i = 1 c j = 1 n i h = 1 n i Φ j , h i n i ( x j i x h i ) ( x j i x h i )
Φ j , : i 1 = k , Φ j , h i { 0 , 1 }  
c local
Table 3. Brief description of the handwritten digit datasets.
Table 3. Brief description of the handwritten digit datasets.
ClassUSPSMNIST
♯Training♯Test♯Training♯Test
06441775923980
1119435967421135
2100526459581032
373119861311010
46581665842982
56522005421892
65561605918958
766417062651028
86451475851974
954216659491009
Table 4. Experimental results (%) (best average accuracy ± standard deviations (%); optimal reduced dimensions) on handwritten digit datasets; the best results are highlighted in bold.
Table 4. Experimental results (%) (best average accuracy ± standard deviations (%); optimal reduced dimensions) on handwritten digit datasets; the best results are highlighted in bold.
DataLDAMMCNMMPRFLDRSLDAMPDANLDALPMDA
USPS94.02 ± 1.2195.33 ± 1.3396.73 ± 1.5196.55 ± 1.3197.21 ± 1.5198.01 ± 1.6297.21 ± 1.5598.89 ± 0.81
(1)(140)(160)(160)(130)(110)(160)(150)
MNIST83.18 ± 1.6684.52 ± 1.7385.67 ± 1.5585.72 ± 1.4587.32 ± 1.6288.53 ± 1.7187.51 ± 1.6089.75 ± 0.85
(1)(120)(130)(140)(110)(100)(130)(140)
Table 5. Brief description of the benchmark datasets used in experiments.
Table 5. Brief description of the benchmark datasets used in experiments.
Dataset# Classes# Instances# Features
Connect 4344,473126
DNA33186180
SVMGuide2339120
Segment7231019
Coil20201440256
PIE6811,5541024
Table 6. Running time (seconds) of eight methods on eight benchmark datasets.
Table 6. Running time (seconds) of eight methods on eight benchmark datasets.
DatasetLDAMMCNMMPRFLDRSLDAMPDANLDALPMDA
USPS0.2130.3871.6750.7211.3692.9122.1630.392
MNIST0.3760.4331.8321.0261.5763.6852.8250.453
Connect41.3681.8306.1224.6515.21612.3776.9331.836
DNA1.2311.6825.3313.6854.92111.9376.3871.721
SVMGuide20.0860.1220.4680.3290.4530.9750.6370.151
Segment0.1310.1730.6210.4670.5261.2650.8610.210
Coil200.1680.2610.8210.5870.7331.7331.2330.311
PIE11.61216.3389.37560.3172.891179.32192.33117.833
Table 7. Results (%) (best average accuracy ± standard deviations%; optimal reduced dimensions) of the ablation study on benchmark datasets; the best results are highlighted in bold.
Table 7. Results (%) (best average accuracy ± standard deviations%; optimal reduced dimensions) of the ablation study on benchmark datasets; the best results are highlighted in bold.
DatasetUSPSMNISTConnect 4DNASVMGuide2SegmentCoil20PIE
LPMDA196.30 ± 0.9287.03 ± 0.8771.21 ± 1.1282.11 ± 1.6779.61 ± 1.2288.63 ± 0.7675.81 ± 1.2692.36 ± 1.02
(150)(140)(8)(7)(6)(6)(40)(135)
LPMDA295.51 ± 2.3385.61 ± 1.0169.01 ± 1.2180.38 ± 2.0778.51 ± 0.9189.26 ± 0.7876.85 ± 0.9891.48 ± 0.85
(1)(1)(2)(2)(2)(6)(19)(67)
LPMDA98.89 ± 0.8189.75 ± 0.8574.21 ± 1.0284.36 ± 1.7181.37 ± 0.8791.33 ± 0.6977.23 ± 1.8394.23 ± 0.72
(150)(140)(6)(4)(4)(6)(35)(120)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, J. Locality-Preserving Multiprojection Discriminant Analysis. Mathematics 2025, 13, 962. https://doi.org/10.3390/math13060962

AMA Style

Ma J. Locality-Preserving Multiprojection Discriminant Analysis. Mathematics. 2025; 13(6):962. https://doi.org/10.3390/math13060962

Chicago/Turabian Style

Ma, Jiajun. 2025. "Locality-Preserving Multiprojection Discriminant Analysis" Mathematics 13, no. 6: 962. https://doi.org/10.3390/math13060962

APA Style

Ma, J. (2025). Locality-Preserving Multiprojection Discriminant Analysis. Mathematics, 13(6), 962. https://doi.org/10.3390/math13060962

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop