Next Article in Journal
Evaluation of the First Metacarpal Bone Head and Distal Radius Bone Architecture Using Fractal Analysis of Adolescent Hand–Wrist Radiographs
Previous Article in Journal
Morphodynamic Features of Contrast-Enhanced Mammography and Their Correlation with Breast Cancer Histopathology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distributed Sparse Manifold-Constrained Optimization Algorithm in Linear Discriminant Analysis

1
State Key Laboratory of Integrated Chips and Systems, School of Microelectronics, Fudan University, Shanghai 200433, China
2
Shanghai Key Laboratory of Automobile Intelligent Network Interaction Chip and System, School of Microelectronics, Shanghai University, Shanghai 200444, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2025, 11(3), 81; https://doi.org/10.3390/jimaging11030081
Submission received: 20 January 2025 / Revised: 4 March 2025 / Accepted: 11 March 2025 / Published: 13 March 2025

Abstract

:
In the field of video image processing, high definition is one of the main directions for future development. Faced with the curse of dimensionality caused by the increasingly large amount of ultra-high-definition video data, effective dimensionality reduction techniques have become increasingly important. Linear discriminant analysis (LDA) is a supervised learning dimensionality reduction technique that has been widely used in data preprocessing for dimensionality reduction and video image processing tasks. However, traditional LDA methods are not suitable for the dimensionality reduction and processing of small high-dimensional samples. In order to improve the accuracy and robustness of linear discriminant analysis, this paper proposes a new distributed sparse manifold constraint (DSC) optimization LDA method, called DSCLDA, which introduces L 2 , 0 -norm regularization for local sparse feature representation and manifold regularization for global feature constraints. By iterating the hard threshold operator and transforming the original problem into an approximate non-convex sparse optimization problem, the manifold proximal gradient (ManPG) method is used as a distributed iterative solution. Each step of the algorithm has an explicit solution. Simulation experiments have verified the correctness and effectiveness of this method. Compared with several advanced sparse linear discriminant analysis methods, this method effectively improves the average classification accuracy by at least 0.90 % .

1. Introduction

An ultra-high-definition video image processing system relies on the ability to detect information from multiple targets and over long distances [1,2]. However, the current chip computing power cannot support complex computational imaging processing methods, resulting in the inability to meet the real-time requirements of the system, and the processing performance has not yet reached the demand for high-dimensional information detection and perception. Therefore, it is necessary to quickly and adaptively process high-dimensional video images. Nowadays, complex image data are inherently redundant and non-Gaussian, leading to the unstable performance of traditional methods such as principal component analysis (PCA), linear discriminant analysis (LDA) [3,4], Fisher discriminant analysis (FDA) [5], orthogonal linear discriminant analysis (OLDA) [6], and uncorrelated linear discriminant analysis (ULDA) [7], which has affected the actual video processing performance. Therefore, it is urgent to explore how to utilize high-dimensional spatial data, establish new sparse discrimination models, and design effective optimization schemes to improve existing detection and classification strategies.
From the perspective of data analysis, the key to processing and analyzing high-dimensional data lies in dimensionality reduction and feature extraction, with a focus on sparsity [8,9]. As an emerging optimization branch, sparse constraints have attracted much attention due to their ability to break through traditional Shannon sampling and achieve efficient transmission. Nowadays, sparsity constraints have been widely used in pattern recognition image processing, and its applicability in many other fields has been recognized [10,11]. Sparse constraints refer to the majority of elements being zero. For high-dimensional data, it is necessary to consider sparsity, such as through sparse linear discriminant analysis (SLDA) [12], sparse uncorrelated linear discriminant analysis (SULDA) [13], robust sparse linear discriminant analysis (RSLDA) [14], intra-class and inter-class kernel constraints (IIKCs) [15], hypergraph Laplacian-based semi-supervised discriminant analysis (HSDAFS) [16], adaptive and fuzzy locality discriminant analysis (AFLDA) [17], etc. Compared with traditional LDA methods, sparse discriminant analysis greatly improves the identification ability of the system. However, sparse discriminant analysis methods usually replace the L 1 -norm with the L 0 -norm to obtain convex optimization problems. Obviously, the L 0 -norm can select the most representative feature variables and optimize faster than L 1 -norm-constrained optimization. Examples include a sparse signal recovery framework based on segmented threshold L 0 gradient approximation [18], image non-negative matrix factorization with alternating smooth L 0 -norm constraints [19], and sparse feature selection based on fast embedding spectral analysis [20]. The above methods also prove that the L 0 -norm does indeed have better feature vector selection and faster optimization algorithm capabilities.
From the perspective of feature extraction, sparse analysis-based methods have significant data analysis capabilities but can not reveal potential causal relationships between variables during the analysis process [21,22,23]. To address this problem, manifold learning can be introduced to learn local features of potential information in high-dimensional space. To characterize such data, a practicable solution is to map the linear inseparable features in high-dimensional space to a low-dimensional nonlinear feature space, such as through robust sparse manifold discriminative analysis (RSMDA) [24], which captures both global and local geometric information through manifold learning. Zhang et al. [25] proposed a coupled discriminative manifold alignment (CDMA) method, which focuses on aligning the manifold structures of a low resolution (LR) and high resolution (HR) in a common feature subspace. In order to use manifold learning methods, many optimization schemes have been proposed, such as the projection algorithm [26], precise penalty algorithm [27], augmented Lagrangian algorithm [28], iterative hard threshold [29], Newton hard threshold pursuit [30], etc. In addition, in the field of video image processing, the problem of high-dimensional pixel videos essentially requires a minimization optimization with Stiefel manifolds, including the partial least squares [31], principal component analysis [32], and canonical correlation analysis [33]. Manifold-constrained optimization can even be seen frequently in reinforcement learning [34] and federated learning [35].
The optimization methods mentioned above mostly focus on a single constraint. Currently, there is limited research on problems that consider both manifold constraints and sparse constraints. On the one hand, this is because both constraints are non-convex, non-smooth, and even NP-hard, making joint algorithm design difficult. On the other hand, joint constraints require two constraints to share the same variable, making theoretical analysis more difficult. To solve these difficulties, this paper proposes a new distributed sparse manifold-constrained optimization algorithm and explores effective numerical solutions. The proposed joint constraints are introduced to LDA, and the novel method is called distributed sparse manifold-constrained linear discriminant analysis (DSCLDA). The proposed method first divides the process monitoring data into multiple data nodes and performs distributed parallel operations simultaneously. Afterward, a L 2 , 0 -norm sparse constraint is constructed to regulate local features and preserve the local structure of variables. In addition, by using manifold constraints on global variables, the proposed method can capture the causal correlation and reduce the data structure loss during the projection. By using the manifold proximal gradient (ManPG) to combine local and global variables, sparse constraints and manifold constraints are incorporated into the calculation process during optimization, and explicit solutions for each variable are obtained. The contributions of the proposed method can be presented as follows:
  • This paper proposes a novel distributed sparse manifold-constrained linear discriminant analysis (DSCLDA) method, which introduces sparse and manifold constraints to maintain the local and global structure.
  • We designed an effective solution scheme that combines local and global variables using the manifold proximal gradient (ManPG) to obtain explicit solutions for each subproblem.
  • We conducted a series of experiments on several public datasets to verify the effectiveness of the proposed method and discuss the convergence and feature distribution.
The rest of this paper is organized as follows. Section 2 introduces the notations and related works. Section 3 details the problem of the proposed method and the corresponding optimization algorithm. Section 4 evaluates and discusses the performance of the proposed method. Section 5 concludes this paper.

2. Notations and Preliminaries

2.1. Notations

For convenience, we will define some symbols required for this section. For the matrix X R n × p , X i is represented as the ith row, and X i j is represented as an element of the ith row and jth column. O n × p is written as an all-zero matrix in n × p , I n × p represents the identity matrix with the dimensions n × p , and I p represents the identity matrix with the dimensions p × p . X represents the transpose of X, and vec ( X ) represents the vectorization of X. For set T , T ¯ is used to represent the complement of T. In addition, for matrices X , Y R n × p , the inner product is defined as X , Y = tr ( X Y ) = i = 1 n i = 1 p X i j Y i j , where tr ( · ) represents the trace of the matrix.

2.2. Preliminaries

LDA, as a supervised learning method, can use prior experience of categories in the dimensionality reduction process, while unsupervised learning cannot use prior experience of categories. Compared with other methods, a feature of LDA is to learn discriminative projections by maximizing the inter-class distance while minimizing the intra-class distance, thereby achieving a more effective dimensionality reduction ability. Define the inter-class distance matrix S b and intra-class distance matrix S w for the training samples, and these two matrices can be defined as
S b = 1 n i = 1 c n i ( x ¯ i x ¯ ) ( x ¯ i x ¯ ) ,
S w = 1 n i = 1 c j = 1 n i ( x ¯ i j x ¯ i ) ( x ¯ i j x ¯ i ) .
LDA attempts to find a suitable projection direction that minimizes intra-class dispersion and maximizes inter-class dispersion after projection. This search process can be expressed as follows:
X = arg max X X = I Tr X S b X Tr X S w X .
To avoid the distortion of S w , problem (3) can also be extended in the following form:
min X Tr X ( S w μ S b ) X s . t . X X = I .
However, LDA still has some shortcomings. For example, LDA can only reduce the dimensionality of data with a category of k to k 1 at most. Therefore, LDA cannot be used when reducing the dimensionality below k 1 . In addition, if the original sample size is too small, the dimensionality reduction results of LDA are prone to overfitting. Therefore, a common modification is to add sparse constraints to LDA, commonly known as SLDA. In common SLDA methods, the L 1 -norm is applied in LDA to induce sparsity, which can remove redundant features from the data and improve the performance of video image processing. The formulation of SLDA, which introduces the L 2 , 1 -norm, is expressed as follows:
min X Tr X ( S w μ S b ) X + λ X 2 , 1 s . t . X X = I .
To effectively eliminate noise and outliers in SLDA and improve robustness in discriminant analysis, reference [14] proposed RSLDA, which is expressed in the form of
min P , X , E Tr X ( S w μ S b ) X + λ 1 X 2 , 1 + λ 2 E 1 s . t . R = P X R + E , P P = I ,
where · 1 is the L 1 -norm. By selecting different parameters of λ 1 and λ 2 , RSLDA can select important features and effectively eliminate noise and outliers, thereby achieving excellent performance in the field of image classification.
Another method to improve the performance of SLDA is to incorporate manifold constraints into the optimization problem, such as in the RSMDA method from reference [24], which is represented as
min P , X , E Tr X ( S w μ S b ) X + Tr Q X ( S w μ S b ) Q X + λ 1 X 2 , 1 + λ 2 E 1 s . t . Q = P X Q + E , P P = I .
Inspired by the above methods, this paper proposes an LDA variant that utilizes joint sparsity and manifold constraints. The specific optimization problem will be described in detail in Section 3.

3. Methodology

3.1. Optimization Problem

In this paper, for the random matrix X, the proposed distributed sparse manifold constraints can be expressed as the following problem:
min X i = 1 l f i ( X ) + λ g ( X ) , s . t . X 2 , 0 s , X T X = I p ,
where l represents the total number of distributed representations of X. Distributed sparse manifold constraints can fully utilize the spatial information of the current extended variables, further improving the interpretability of variables from the process monitoring data. Therefore, combined with regular LDA, distributed sparse manifold linear discriminant analysis (DSCLDA) is proposed, which can fully utilize the local and global information of process monitoring observations and take into account both causal and structural relationships between variables.
In this model, f i R n × p R ( i = 1 , 2 , , l ) is the given Lipschitz locally continuous function, and g R n × p is the given global function. X 2 , 0 s is introduced as the sparse constraint, and X T X = I p is used as the manifold constraint. Substituting problem (4) into the distributed sparse constraint yields
min X i = 1 l f i ( X ) + λ g ( X ) , s . t . X = A + E , X 2 , 0 s , X T X = I p .

3.2. Optimization Algorithm

To obtain an effective algorithm, the distributed variable X i and the global variable Y are introduced to transform problem (8) into
min X i i = 1 l f i ( X i ) + λ g ( Y ) , s . t . X i = A i + E i , X i 2 , 0 s , Y T Y = I p ,
where X i represents the variables of the ith distribution. In problem (10), the sparse constraint only includes the local variable X i , and the manifold constraint only includes the global variable Y. Therefore, further consideration can be given to the optimization problem of the following penalty function:
min X i , Y i = 1 l f i ( X i ) + λ g ( Y ) + i = 1 l μ i X i Y F 2 , s . t . X i = A i + E i , X i 2 , 0 s , Y T Y = I p ,
in which μ i is the penalty parameter corresponding to each branch.

3.2.1. Updating X i

Problem (30) is an NP-hard problem, and there is no explicit solution. Inspired by the Newton hard threshold tracking method, the proposed optimization algorithm extends it to matrices. Assuming the objective function is h i ( X i ) , then the gradient of this function is represented as
h i ( X i ) = f i ( X i ) + 2 μ i ( X i Y ) .
The Hessian expression of problem (12) can be written as
2 h i ( X i ) = f i ( X i ) + 2 μ i I n p .
If X i P ʃ ( X i α i h i ( X i ) ) is satisfied (where α i > 0 is the step size parameter), then X i can be considered as the stable point for problem (29). Let T S ( X i , α i ) represent the set of indicators for the first s rows of X i α i h i ( X i ) under L 2 -norm constraints; then, for any T i T s ( X i , α i ) , this satisfies a nonlinear relationship, which is written as
H i ( X i , T i ) = ( h i ( X i ) ) T i ( X i ) T ¯ i = 0 ,
in which ( h i ( X i ) ) T i represents the submatrix in h i ( X i ) , and T i is the corresponding indicator set. ( X i ) T ¯ i R n a × p indicates the submatrix in X i with T ¯ i as the indicator set. The gradient of H i ( X i , T ) in X i can be expressed as
H i ( X i , T ) = ( 2 h i ( X i ) ) T i T i ( 2 h i ( X i ) ) T i T ¯ i O ( n s ) p × s p I ( n s ) p ,
where ( 2 h i ( X i ) ) T i T i R s p × s p represents the Hessian submatrix with the indicator set T i T i . Define
X i ( α ) = ( X i ) T i + α ( D ) T i ( O ) T i ,
where D represents the descending direction. The minimum X i can be obtained using a sparse proximal gradient (SpaPG). The descending direction D is obtained from
H i ( X i , T ) vec ( D ) = vec ( H i ( X i , T ) ) ,
Then, the k + 1 th X i , X i k + 1 , should be represented as
X i k + 1 = X i k ( α i k ) ,
in which α i k = ρ τ , while τ is the smallest positive integer that satisfies the following equation, written as
h i ( X i ( ρ τ ) ) h i ( X i ) + σ ρ τ h i ( X i ) , D k

3.2.2. Updating Y

Set M = { T | Y T Y = I p } ; then, the tangent space of manifold M at Y is expressed as T Y M = { Z | Z T Y + Y T Z = 0 } . Assuming the objective function is ϕ ( Y ) , it has the following approximation function:
ϕ ( Y k ) + ϕ ( Y k ) , Y Y k + 1 2 t Y Y k F 2 ,
where 1 L t > 0 is a parameter. To obtain the descending direction D, define
min D R R × P ϕ ( Y k ) , D + 1 2 t D F 2 s . t . D T Y k M ,
Based on the definition of T Y k M , set D T Y k + Y k T D = 0 , and Equation (21) can be represented as
min D R R × P ϕ ( Y k ) , D + 1 2 t D F 2 s . t . D T Y k + Y k T D = 0 ,
Based on Equation (22), the Lagrange function can be obtained, which is written as
L ( D , Λ ) = ϕ ( Y k ) , D + 1 2 t D F 2 Λ , D T Y k + Y k T D ,
in which Λ R n × p is the Lagrange multiplier. Then, the corresponding Karush–Kuhn–Tucker (KKT) system for the Lagrangian function above is represented as
0 D L ( D , Λ ) , 0 = D T Y k + Y k T D
By synthesizing Equation (24), the optimization problem for { D , Λ , Y } can be obtained, which is written as
D ( Λ ) T Y k + Y k T D ( Λ ) = 0 .
Equation (25) can be solved using the manifold proximal gradient (ManPG) algorithm, and the k + 1 th Y can be represented as
Y k + 1 = R Y k ( γ k D k ) ,
in which the mapping R Y : T Y M M represents Retraction. R Y maps the vectors in the tangent space to the manifold, allowing the problem to maintain orthogonality during the optimization. In Equation (26), γ k = γ η q , and q is the smallest positive integer that satisfies the following equation, expressed as
ϕ ( Y k + 1 ) ϕ ( Y k ) γ η q 2 t D k F 2 .

3.3. Convergence Analysis

According to the updates of X i and Y, the optimization algorithm of Equation (11) can be expressed as Algorithm 1. In addition, according to the literature [36], if ( X i * , Y * ) satisfies
0 P S ( f i ( X i * ) + μ i ( X i * Y * ) ) , 0 P M ( λ g ( Y * ) μ i ( X * Y * ) ) ,
then ( X i * , Y * ) can be considered as the stable point of Equation (11). The experimental verification of the convergence analysis can be found in Section 4.5.

3.4. Complexity Analysis

To verify how distributed sparse constraints can enhance the performance of existing methods, this section compares the complexity and computational cost of the proposed method with that of the baseline LDA method. For Equation (9), given the original data with a dimensionality of d and n samples, the computational complexity of the objective function is O ( n d 2 ) . The sparse constraint X 2 , 0 s , which checks the number of non-zero elements, has a complexity of O ( n d ) . The manifold constraint X T X = I p implies that X is an orthogonal matrix, which, enforcing orthogonality through methods such as QR decomposition, has a complexity of O ( n d 2 ) . Therefore, the overall complexity of the proposed method is O ( n d 2 ) . In contrast, the computational complexity of traditional LDA-based methods is primarily determined by calculating the within-class scatter matrix S w , the between-class scatter matrix S b , and solving the generalized eigenvalue problem. The complexity of computing S w is O ( n d 2 ) , while that of computing S b is O ( d 2 ) , due to the calculation based on class means and the global mean. The complexity of solving the eigenvalues and eigenvectors of S w 1 S b is O ( d 3 ) , which is the most time-consuming part of LDA. The overall complexity of LDA is O ( n d 2 + d 3 ) . The proposed distributed sparse constraint method demonstrates superior computational efficiency over traditional LDA methods by reducing the overall complexity from O ( n d 2 + d 3 ) to O ( n d 2 ) through the enforcement of sparsity and orthogonality constraints, thereby eliminating the most time-consuming generalized eigenvalue problem in LDA.
Algorithm 1 Optimization algorithm for (11)
Input: Data X, parameters s,l, λ , μ i > 0 .
Initialize: Data Y 0 , parameter k = 0 .
Output: Data Y.
While not converged do
1:
According to Algorithm 2, update X i k + 1 by
min X i f i ( X i ) + μ i X i Y F 2 , s . t .   X i = A i + E i , X i 2 , 0 s .
2:
According to Algorithm 3, update Y k + 1 by
min Y   λ g ( Y ) + i = 1 l μ i X i Y F 2 , s . t .   Y T Y = I p .
3:
If the process meets the shutdown criteria X i 2 , 0 s , stop; Otherwise, let k = k + 1 and return to Step 1.
End while
Algorithm 2 Optimization algorithm for (12)
Input: Data X, parameters μ , α > 0 , ρ ( 0 , 1 ) , σ ( 0 , 1 / 2 ) .
Initialize:  X i 0 , T i 0 T S ( X i 0 , α ) , when k = 0 .
Output: X i k
While not converged do
1:
Obtain the nonlinear relationship H i ( X i k , T k ) and the gradient H i ( X i k , T i k ) , according to (14), (15);
2:
Obtain the descent direction D k , according to (17);
3:
Update the local variable X i k + 1 , according to (9);
4:
If the process meets the shutdown criteria, stop; Otherwise, let k = k + 1 , update T i k T S ( X i k , α k ) , and return to Step 1.
End while
Algorithm 3 Optimization algorithm for (20)
Input: Parameters γ , t > 0 , η ( 0 , 1 ) .
Initialize:  Y 0 , Λ 0 , k = 0 .
Output:  X k , Y k .
While not converged do
1:
Obtain the descent direction D, according to (25);
2:
Update the global variable Y k + 1 , according to (26),
3:
If the process meets the shutdown criteria, stop; Otherwise, let k = k + 1 and return to Step 1.
End while

4. Simulation Studies

In the experiments, DSCLDA was compared with traditional LDA and six LDA variants, including AFLDA [17], ERSLDA [37], RSLDA+IIKC [15], RSMDA [24], RSLDA [14], SULDA [13], and SLDA [12]. The optimization problem and constraint of each method are shown in Table 1. The datasets used in the experiments in this paper are shown in Table 2, and examples of each dataset are shown in Figure 1. In this experiment, a self-built vehicle dataset, called the CAR_image dataset, was introduced.

4.1. Experiment Settings

Due to the fact that the datasets were divided into D parts as data nodes for distributed computing during the experiment, D was added before all method names to indicate distributed performance, such as DERSLDA and DRSLDA. In the simulation verification, each method was executed 10 times, with different random samples selected from the same dataset for each run; then, the average classification accuracy was calculated. To improve computational efficiency, all datasets were preconverted into grayscale images. In addition, to improve computational efficiency and achieve better classification accuracy, this experiment used PCA to perform dimensionality reduction on all image datasets, retaining 95% of the original data information. Furthermore, due to the large and inconsistent image sizes in the Car_image dataset, the unified resolution of the images in this dataset was 64 × 128 .
In selecting experimental parameters, the selection of the parameters λ and μ was carried out through a ten-fold cross-validation method based on the content and size of each dataset. The range of the parameters λ and μ is denoted as 10 5 , 10 4 , 10 3 , 10 2 , 10 1 , 1 , 10 1 , 10 2 , 10 3 , 10 4 , and 10 5 . Prior to numerical validation, a strategy of fixing the value of λ while varying μ was employed to ascertain the corresponding accuracy for each configuration, serving as a basis for evaluation. The experimental results on the COIL20 image dataset are shown in Figure 2. Based on the experimental results, it can be determined that the selection of the parameters λ and μ should be within the range of 10 5 , 10 4 , 10 3 , and 10 2 to achieve better image processing performance. Specifically, for the COIL20 image dataset, the most suitable parameter combination was identified as 10 3 and 10 5 , and a similar method for parameter selection was applied to other datasets under investigation. In addition, the shutdown criterion in this experiment was set so that 100 iterations would be reached or the overall objective function value would be less than 10 3 .

4.2. Experiment Based on Sample Size

This experiment used the k nearest neighbors (KNN) classifier to analyze the classification accuracy of the dimensionality reduction results of various methods. The knn classifier is a supervised machine learning algorithm that assigns a new data point to the class most common among the k nearest neighbors in the feature space, based on a distance metric such as the Euclidean distance. In this experiment, four different sample sizes were randomly selected for each dataset as the training set, and the remaining samples were used as the testing set. The classification experiment results under different sample sizes are shown in Table 3, where the highest-performing results are highlighted in bold. The simple image datasets used in the experiment, including the Mnist dataset, Hand Gesture Recognition dataset, and COIL20 dataset, have simple content, a monotonous background, and obvious features. Therefore, each method could achieve better classification performance on the above three datasets. The image features of the NEU surface defect dataset, Car_image dataset, and Caltech-101 dataset are relatively complex or have a high proportion of the background, resulting in relatively low classification accuracy for each model on these datasets. However, the experimental results show that the DSCLDA model still had improvements in the classification performance compared to other methods on these datasets.
Compared with other methods, DSCLDA improved by at least 0.51 % on the Mnist dataset; improved by at least 0.44 % on the Hand Gesture Recognition dataset; improved by at least 0.85 % on the COIL20 image dataset; improved by at least 0.86 % on the NEU surface defect dataset; improved by at least 2.16 % on the Car_image dataset; and improved by at least 0.55 % on the Caltech-101 image dataset. The classification performance of DSCLDA was further improved on two difficult datasets, namely the NEU surface defect dataset and the Car_image dataset. The results can be explained by the fact that the DSCLDA model, which simultaneously extracts features from both global and local structures, can obtain more representative feature data when processing complex images or images with unclear features, thereby achieving better classification performance. The experimental results also demonstrate that DSCLDA divides process monitoring data into multiple data nodes and performs distributed parallel operations, which not only improves computational efficiency but also provides better adaptation to the processing needs of large-scale data.
Compared to other methods, the average classification performance of the proposed DSCLDA method improved by at least 0.90 % , which proves that the proposed method achieves satisfactory classification performance by introducing joint sparse and manifold constraints. In addition, compared with DRSLDA, DRSMDA, DRSLDA+IIKC, and DERSLDA, the proposed DSCLDA still had a significant improvement, indicating that the proposed method can demonstrate advantages when compared with some of the latest SLDA variants.

4.3. Experiment Based on the Number of Dimensions

In this experiment, ( 50 , 100 ) , ( 4 , 6 ) , ( 4 , 6 ) , ( 50 , 100 ) , ( 10 , 20 ) , and ( 10 , 20 ) samples were selected as training sets for each type on the six public image datasets, and the remaining samples were used as testing sets with dimensions ranging from 5 to 200. The classification experiment results are shown in Figure 3, Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8. The experimental results indicate that the proposed DSCLDA method achieved relatively better classification performance on the six publicly available datasets mentioned above. From the experimental results, it can be seen that the classification performance of DLDA and DSLDA is very sensitive to the choice of dimensionality. As the dimensionality increases, the classification performance of these two methods may even even decrease. However, the proposed DSCLDA method can still maintain classification accuracy in the presence of dimensional changes. The experimental results demonstrate the flexibility of the proposed method in dimension selection. For the NEU surface defect and Caltech-101 image datasets, the classification performance of DSCLDA did not show significant improvement compared to that of other methods because the features of these two datasets are relatively complex and not clear enough, resulting in similar classification results for the above methods. However, on several other publicly available datasets, the proposed DSCLDA method still performed relatively better in terms of its classification accuracy. DSCLDA normalizes local features by constructing L 2 , 0 -norm sparse constraints, preserves the local structure of variables, and utilizes manifold constraints to capture causal correlations between global variables, reducing data structure loss during the projection process.

4.4. Experiments with Deep Learning Methods

Deep learning methods, such as Transformer-based feature extraction models, provide new perspectives and powerful tools for feature extraction and dimensionality reduction, which can provide valuable benchmarks. These deep learning methods typically have better feature-learning capabilities and stronger robustness and can achieve excellent performance on large-scale datasets. Therefore, this paper also compares DSCLDA with deep learning-based dimensionality reduction techniques, such as R3D-CNN [43], I3D [44], and Transformer [45], to demonstrate its broader applicability in different scenarios. Through experiments on the Hand Gesture Recognition (HGR) and CIFAR-100 [46] datasets, we validated the advantages of DSCLDA in feature extraction and dimensionality reduction, as well as its competitiveness with deep learning methods. Table 4 demonstrates that the gesture recognition dataset may have certain limitations, such as its number of samples, diversity, and representativeness. On smaller datasets, simpler or more traditional models, such as DSCLDA, may perform better because of their lower complexity. On the other hand, models supported by deep learning methods may be more suitable for handling large and complex datasets, capable of capturing more subtle patterns and relationships.

4.5. Convergence Analysis

This section describes how we conducted experimental verification of the convergence analysis proposed in Section 3.3. In the proposed DSCLDA method, the most computationally expensive step is the calculation of the projection matrix X, while the most computationally intensive task is the process of solving the inverse matrix, which significantly affects the computational efficiency of DSCLDA. In this experiment, the computational efficiency of DSCLDA was reflected in the speed of function value reduction and the convergence speed of the classification accuracy. In order to visually demonstrate the convergence of the proposed DSCLDA method, Figure 9 shows the curves of the objective function value and classification accuracy of the functions. As the number of iterations increased, the objective function value of the proposed DSCLDA method rapidly decreased and reached its minimum value, and the classification accuracy also reached its maximum value and converged within 30 iterations. The experimental results validate the fast convergence of DSCLDA.

4.6. t-SNE Comparison

In addition, to further validate the principle and effectiveness of the proposed method, the t-SNE method was utilized to visualize the data distribution before and after projection. The experiment used the top five types of data from the Mnist dataset and randomly selected 100 samples for each type as the training set and the remaining samples as the testing set. The corresponding classification accuracies of each method were 85.85% (DRSLDA), 90.10% (DERSLDA), and 90.20% (DSCLDA), respectively. The experimental results are shown in Figure 10. It can be seen that when not projected, the inter-class and intra-class distributions of the Mnist dataset were not significant. When using the DRSLDA method for projection, it can reduce the distance between types and increase the distance between different types, but DRSLDA cannot fully classify all data, and this classification method is not satisfactory in terms of distribution. With the introduction of sparse constraints, the inter-class distance between different types of data becomes larger, while the intra-class spacing becomes smaller. In the t-SNE distribution of the proposed DERSLDA method, the intra-class spacing is relatively small, but the inter-class distance is not large enough, so there is still a possibility of data confusion during the classification process. In the t-SNE distribution of the proposed DSCLDA method, the distance between different types is the largest, and the distance between types is the smallest, such as between type 1 and type 2, which is more conducive to determining the data category during the classification process. The experimental results show that the proposed method has relatively better classification performance.

5. Conclusions

In this paper, we constructed a novel distributed sparse manifold constraint and a novel LDA variant, called DSCLDA. The proposed method trains discriminative projections by introducing manifold constraints and L 2 , 0 -norm sparse constraints, which can obtain the most discriminative features for process monitoring. In addition, in this paper, we designed and developed a novel manifold proximal gradient algorithm to handle the proposed optimization model, while distributed parallel computing could significantly improve computational efficiency. The advantages of DSCs and DSCLDA have been demonstrated through numerical experiments on several public datasets. Compared with other existing LDA methods, the proposed DSCLDA method improves the image classification accuracy by at least 0.90 % and also has significant advantages in convergence and feature distribution.
However, the proposed method currently exhibits limitations in terms of its image processing efficiency and feature classification accuracy, necessitating integration with deep learning techniques for improvement. In the future, we will attempt to combine the proposed method with deep learning methods to improve the efficiency of image processing and the accuracy of feature classification. Furthermore, deployment on hardware platforms may be constrained by computational complexity and insufficient flexibility, highlighting the need for further optimization to enhance the processing efficiency and applicability. In addition, this method will also be considered for deployment on hardware to improve the efficiency of the method’s processing and the flexibility of the method’s usage.

Author Contributions

Methodology, M.F. and J.L.; software, M.F.; writing—original draft, M.F. and J.L.; writing—review and editing, Y.Z. and X.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62204044 and in part by the State Key Laboratory of Integrated Chips and Systems under Grant SKLICS-K202302.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All datasets used are available online with open access.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yu, W.; Zhu, Q.; Zheng, N.; Huang, J.; Zhou, M.; Zhao, F. Learning non-uniform-sampling for ultra-high-definition image enhancement. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 1412–1421. [Google Scholar]
  2. Yu, X.; Dai, P.; Li, W.; Ma, L.; Shen, J.; Li, J.; Qi, X. Towards efficient and scale-robust ultra-high-definition image demoiréing. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 646–662. [Google Scholar]
  3. McLachlan, G.J. Discriminant Analysis and Statistical Pattern Recognition; John Wiley & Sons: Hoboken, NJ, USA, 2005. [Google Scholar]
  4. Ullah, S.; Ahmad, Z.; Kim, J.M. Fault Diagnosis of a Multistage Centrifugal Pump Using Explanatory Ratio Linear Discriminant Analysis. Sensors 2024, 24, 1830. [Google Scholar] [CrossRef] [PubMed]
  5. Mai, Q.; Zou, H. A note on the connection and equivalence of three sparse linear discriminant analysis methods. Technometrics 2013, 55, 243–246. [Google Scholar] [CrossRef]
  6. Ye, J.; Xiong, T. Null space versus orthogonal linear discriminant analysis. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 1073–1080. [Google Scholar]
  7. Ye, J.; Janardan, R.; Li, Q.; Park, H. Feature reduction via generalized uncorrelated linear discriminant analysis. IEEE Trans. Knowl. Data Eng. 2006, 18, 1312–1322. [Google Scholar]
  8. Shi, Y.; Huang, W.; Ye, H.; Ruan, C.; Xing, N.; Geng, Y.; Dong, Y.; Peng, D. Partial least square discriminant analysis based on normalized two-stage vegetation indices for mapping damage from rice diseases using PlanetScope datasets. Sensors 2018, 18, 1901. [Google Scholar] [CrossRef]
  9. Bach, F. High-dimensional analysis of double descent for linear regression with random projections. SIAM J. Math. Data Sci. 2024, 6, 26–50. [Google Scholar] [CrossRef]
  10. Xu, H.L.; Chen, G.Y.; Cheng, S.Q.; Gan, M.; Chen, J. Variable projection algorithms with sparse constraint for separable nonlinear models. Control Theory Technol. 2024, 22, 135–146. [Google Scholar] [CrossRef]
  11. Zhang, L.; Wei, Y.; Liu, J.; Wu, J.; An, D. A hyperspectral band selection method based on sparse band attention network for maize seed variety identification. Expert Syst. Appl. 2024, 238, 122273. [Google Scholar] [CrossRef]
  12. Clemmensen, L.; Hastie, T.; Witten, D.; Ersbøll, B. Sparse discriminant analysis. Technometrics 2011, 53, 406–413. [Google Scholar] [CrossRef]
  13. Zhang, X.; Chu, D.; Tan, R.C. Sparse uncorrelated linear discriminant analysis for undersampled problems. IEEE Trans. Neural Netw. Learn. Syst. 2015, 27, 1469–1485. [Google Scholar] [CrossRef]
  14. Wen, J.; Fang, X.; Cui, J.; Fei, L.; Yan, K.; Chen, Y.; Xu, Y. Robust sparse linear discriminant analysis. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 390–403. [Google Scholar] [CrossRef]
  15. Li, S.; Zhang, H.; Ma, R.; Zhou, J.; Wen, J.; Zhang, B. Linear discriminant analysis with generalized kernel constraint for robust image classification. Pattern Recognit. 2023, 136, 109196. [Google Scholar] [CrossRef]
  16. Sheikhpour, R.; Berahmand, K.; Mohammadi, M.; Khosravi, H. Sparse feature selection using hypergraph Laplacian-based semi-supervised discriminant analysis. Pattern Recognit. 2025, 157, 110882. [Google Scholar] [CrossRef]
  17. Wang, J.; Yin, H.; Nie, F.; Li, X. Adaptive and fuzzy locality discriminant analysis for dimensionality reduction. Pattern Recognit. 2024, 151, 110382. [Google Scholar] [CrossRef]
  18. Vivekanand, V.; Mishra, D. Framework for Segmented threshold L0 gradient approximation based network for sparse signal recovery. Neural Netw. 2023, 162, 425–442. [Google Scholar]
  19. Chen, K.; Che, H.; Li, X.; Leung, M.F. Graph non-negative matrix factorization with alternative smoothed L0 regularizations. Neural Comput. Appl. 2023, 35, 9995–10009. [Google Scholar] [CrossRef]
  20. Wang, J.; Wang, H.; Nie, F.; Li, X. Sparse feature selection via fast embedding spectral analysis. Pattern Recognit. 2023, 139, 109472. [Google Scholar] [CrossRef]
  21. Chen, D.W.; Miao, R.; Yang, W.Q.; Liang, Y.; Chen, H.H.; Huang, L.; Deng, C.J.; Han, N. A feature extraction method based on differential entropy and linear discriminant analysis for emotion recognition. Sensors 2019, 19, 1631. [Google Scholar] [CrossRef]
  22. Zheng, W.; Lu, S.; Yang, Y.; Yin, Z.; Yin, L. Lightweight transformer image feature extraction network. PeerJ Comput. Sci. 2024, 10, e1755. [Google Scholar] [CrossRef]
  23. Zhou, J.; Zhang, Q.; Zeng, S.; Zhang, B.; Fang, L. Latent linear discriminant analysis for feature extraction via isometric structural learning. Pattern Recognit. 2024, 149, 110218. [Google Scholar] [CrossRef]
  24. Wang, J.; Liu, Z.; Zhang, K.; Wu, Q.; Zhang, M. Robust sparse manifold discriminant analysis. Multimed. Tools Appl. 2022, 81, 20781–20796. [Google Scholar] [CrossRef]
  25. Zhang, K.; Zheng, D.; Li, J.; Gao, X.; Lu, J. Coupled discriminative manifold alignment for low-resolution face recognition. Pattern Recognit. 2024, 147, 110049. [Google Scholar] [CrossRef]
  26. Chen, S.; Ma, S.; Man-Cho So, A.; Zhang, T. Proximal gradient method for nonsmooth optimization over the Stiefel manifold. SIAM J. Optim. 2020, 30, 210–239. [Google Scholar] [CrossRef]
  27. Xiao, N.; Liu, X.; Yuan, Y.x. Exact Penalty Function for L2,1 Norm Minimization over the Stiefel Manifold. SIAM J. Optim. 2021, 31, 3097–3126. [Google Scholar] [CrossRef]
  28. Wang, L.; Liu, X. Decentralized optimization over the Stiefel manifold by an approximate augmented Lagrangian function. IEEE Trans. Signal Process. 2022, 70, 3029–3041. [Google Scholar] [CrossRef]
  29. Beck, A.; Eldar, Y.C. Sparsity constrained nonlinear optimization: Optimality conditions and algorithms. SIAM J. Optim. 2013, 23, 1480–1509. [Google Scholar] [CrossRef]
  30. Zhou, S.; Xiu, N.; Qi, H.D. Global and quadratic convergence of Newton hard-thresholding pursuit. J. Mach. Learn. Res. 2021, 22, 1–45. [Google Scholar]
  31. Li, G.; Qin, S.J.; Zhou, D. Geometric properties of partial least squares for process monitoring. Automatica 2010, 46, 204–210. [Google Scholar] [CrossRef]
  32. Liu, Y.; Zeng, J.; Xie, L.; Luo, S.; Su, H. Structured joint sparse principal component analysis for fault detection and isolation. IEEE Trans. Ind. Inform. 2018, 15, 2721–2731. [Google Scholar] [CrossRef]
  33. Chen, Z.; Ding, S.X.; Peng, T.; Yang, C.; Gui, W. Fault detection for non-Gaussian processes using generalized canonical correlation analysis and randomized algorithms. IEEE Trans. Ind. Electron. 2017, 65, 1559–1567. [Google Scholar] [CrossRef]
  34. Li, H.; Liu, D.; Wang, D. Manifold regularized reinforcement learning. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 932–943. [Google Scholar] [CrossRef]
  35. Li, J.; Ma, S. Federated learning on Riemannian manifolds. arXiv 2022, arXiv:2206.05668. [Google Scholar]
  36. Rockafellar, R.T.; Wets, R.J.B. Variational Analysis; Springer Science & Business Media: Berlin, Germany, 2009; Volume 317. [Google Scholar]
  37. Liu, J.; Feng, M.; Xiu, X.; Liu, W.; Zeng, X. Efficient and Robust Sparse Linear Discriminant Analysis for Data Classification. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 9, 617–629. [Google Scholar] [CrossRef]
  38. Deng, L. The mnist database of handwritten digit images for machine learning research [best of the web]. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
  39. Mantecón, T.; del Blanco, C.R.; Jaureguizar, F.; García, N. Hand gesture recognition using infrared imagery provided by leap motion controller. In Proceedings of the Advanced Concepts for Intelligent Vision Systems: 17th International Conference, ACIVS 2016, Lecce, Italy, 24–27 October 2016; Proceedings 17. Springer: Berlin/Heidelberg, Germany, 2016; pp. 47–57. [Google Scholar]
  40. Nene, S.A.; Nayar, S.K.; Murase, H. Columbia Object Image Library (Coil-20); Department of Computer Science, Columbia University: New York, NY, USA, 1996. [Google Scholar]
  41. Bao, Y.; Song, K.; Liu, J.; Wang, Y.; Yan, Y.; Yu, H.; Li, X. Triplet-graph reasoning network for few-shot metal generic surface defect segmentation. IEEE Trans. Instrum. Meas. 2021, 70, 1–11. [Google Scholar] [CrossRef]
  42. Kinnunen, T.; Kamarainen, J.K.; Lensu, L.; Lankinen, J.; Käviäinen, H. Making visual object categorization more challenging: Randomized caltech-101 data set. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 476–479. [Google Scholar]
  43. Molchanov, P.; Yang, X.; Gupta, S.; Kim, K.; Tyree, S.; Kautz, J. Online detection and classification of dynamic hand gestures with recurrent 3d convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4207–4215. [Google Scholar]
  44. Carreira, J.; Zisserman, A. Quo vadis, action recognition? A new model and the kinetics dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6299–6308. [Google Scholar]
  45. D’Eusanio, A.; Simoni, A.; Pini, S.; Borghi, G.; Vezzani, R.; Cucchiara, R. A transformer-based network for dynamic hand gesture recognition. In Proceedings of the 2020 International Conference on 3D Vision (3DV), Fukuoka, Japan, 25–28 November 2020; pp. 623–632. [Google Scholar]
  46. Zhang, Z.; Sabuncu, M. Generalized cross entropy loss for training deep neural networks with noisy labels. Adv. Neural Inf. Process. Syst. 2018, 31, 1–11. [Google Scholar]
Figure 1. Some image examples from the datasets used in the experiment. (a) Mnist, (b) Hand Gesture Recognition, (c) COIL20, (d) NEU surface defects, (e) Car_image, (f) Caltech-101.
Figure 1. Some image examples from the datasets used in the experiment. (a) Mnist, (b) Hand Gesture Recognition, (c) COIL20, (d) NEU surface defects, (e) Car_image, (f) Caltech-101.
Jimaging 11 00081 g001
Figure 2. Parameter cross-validation on the COIL20 image dataset. The parameters λ and μ are derived from Equation (11) In this figure, green indicates high value and blue indicates low value.
Figure 2. Parameter cross-validation on the COIL20 image dataset. The parameters λ and μ are derived from Equation (11) In this figure, green indicates high value and blue indicates low value.
Jimaging 11 00081 g002
Figure 3. Classification accuracy on the Mnist dataset. (a) Number of samples: 50; (b) number of samples: 100.
Figure 3. Classification accuracy on the Mnist dataset. (a) Number of samples: 50; (b) number of samples: 100.
Jimaging 11 00081 g003
Figure 4. Classification accuracy on the Hand Gesture Recognition dataset. (a) Number of samples: 4; (b) number of samples: 6.
Figure 4. Classification accuracy on the Hand Gesture Recognition dataset. (a) Number of samples: 4; (b) number of samples: 6.
Jimaging 11 00081 g004
Figure 5. Classification accuracy on COIL20 image dataset. (a) Number of samples: 4; (b) number of samples: 6.
Figure 5. Classification accuracy on COIL20 image dataset. (a) Number of samples: 4; (b) number of samples: 6.
Jimaging 11 00081 g005
Figure 6. Classification accuracy on the NEU surface defect dataset. (a) Number of samples: 50; (b) number of samples: 100.
Figure 6. Classification accuracy on the NEU surface defect dataset. (a) Number of samples: 50; (b) number of samples: 100.
Jimaging 11 00081 g006
Figure 7. Classification accuracy on the Car_image dataset. (a) Number of samples: 10; (b) number of samples: 20.
Figure 7. Classification accuracy on the Car_image dataset. (a) Number of samples: 10; (b) number of samples: 20.
Jimaging 11 00081 g007
Figure 8. Classification accuracy on the Caltech-101 image dataset. (a) Number of samples: 10; (b) number of samples: 20.
Figure 8. Classification accuracy on the Caltech-101 image dataset. (a) Number of samples: 10; (b) number of samples: 20.
Jimaging 11 00081 g008
Figure 9. The relationship between the objective function value, classification accuracy, and the number of iterations. (a) Mnist, (b) Hand Gesture Recognition, (c) COIL20, (d) NEU surface defects, (e) Car_image, (f) Caltech-101.
Figure 9. The relationship between the objective function value, classification accuracy, and the number of iterations. (a) Mnist, (b) Hand Gesture Recognition, (c) COIL20, (d) NEU surface defects, (e) Car_image, (f) Caltech-101.
Jimaging 11 00081 g009
Figure 10. The data distribution displayed using the t-SNE method. The images correspond to (a) the local data of the original Mnist dataset; (b) the distribution of the corresponding data after DRSLDA projection; (c) the distribution of the corresponding data after DERSLDA projection; (d) and the corresponding data distribution projected through DSCLDA.
Figure 10. The data distribution displayed using the t-SNE method. The images correspond to (a) the local data of the original Mnist dataset; (b) the distribution of the corresponding data after DRSLDA projection; (c) the distribution of the corresponding data after DERSLDA projection; (d) and the corresponding data distribution projected through DSCLDA.
Jimaging 11 00081 g010
Table 1. Information on all comparison methods used in this experiment. The bold method is the proposed method.
Table 1. Information on all comparison methods used in this experiment. The bold method is the proposed method.
MethodOptimization ProblemConstraint
LDA min X tr X ( S w μ S b ) X s . t . X X = I
SLDA min X Tr X ( S w μ S b ) X + λ X 2 , 1 s . t . X X = I
SULDA min G s . t . U 1 T G = t 1 P 1 Z , Z T Z = I
RSLDA min P , X , E Tr X ( S w μ S b ) X s . t . R = P X R + E , P P = I
+ λ 1 X 2 , 1 + λ 2 E 1
RSMDA min P , X , E Tr X ( S w μ S b ) X s . t . R = P X R + E , P P = I
+ Tr X R ( L w L b ) R X
+ λ 1 X 2 , 1 + λ 2 E 1
RSLDA+IIKC min P , X , E Tr X ( S w μ S b ) X s . t . R = P X R + E , P P = I
+ λ 1 X 2 , 1 + λ 2 E 1
+ α Tr X ( S w μ S b ) X
ERSLDA min P , X , E , N Tr X ( S w μ S b ) X s . t . R = P X R + E + N , P P = I
+ λ 1 X 2 , p p + λ 2 E p p + η N F 2
DSCLDA min X i = 1 d tr ( X ( S i w τ S i b ) X ) s . t . X 2 , 0 s , X X = I p
Table 2. Information related to the dataset used in this experiment.
Table 2. Information related to the dataset used in this experiment.
DatasetImage TypesImagesColor TypeOriginal Resolution
Mnist [38]1060,000Gray 28 × 28
Hand Gesture Recognition [39]1020,000Gray 240 × 640
Coil20 [40]201440Gray 128 × 128
NEU surface defects [41]61200Gray 32 × 32
Car_image10200RGB 800 × 600 to 5000 × 3000
Caltech-101 [42]1019146RGB and grayAbout 300 × 200
Table 3. The classification accuracy obtained on six datasets. The bold value represents the highest value of the column.
Table 3. The classification accuracy obtained on six datasets. The bold value represents the highest value of the column.
MethodsMnistHand Gesture RecognitionCOIL20
1050100200456736912
DLDA75.2085.5984.0283.7875.9580.3887.8189.3565.2974.9781.5880.40
DSLDA80.4285.6484.8584.4681.5083.6990.6992.2370.8478.2884.4684.02
DSULDA87.7887.3486.3793.3884.7688.6290.1191.6574.1083.2183.8886.37
DRSLDA84.0385.8588.4096.5887.4089.7388.9890.5276.7484.3282.7586.75
DRSMDA83.6286.5690.7797.5187.6285.4990.8592.3976.9680.0884.6287.30
DRSLDA+IIKC73.2783.8585.9296.9490.3090.8692.8994.4379.6485.4586.6686.94
DERSLDA86.7790.1092.0697.6288.1290.2689.8191.3577.4684.8583.5888.73
DAFLDA85.9189.1190.8496.3487.9388.2888.7690.6776.2482.1783.3685.23
DSCLDA86.9290.2092.9497.8290.3791.3993.4895.0279.7185.9887.2590.95
MethodsNEU Surface DefectsCar_IMAGECaltech-101
2550751001015202510152025
DLDA41.2738.8743.2648.0820.6425.0034.5044.7751.5458.1662.8265.21
DSLDA43.0344.9348.1550.5337.0339.9342.1542.0855.5667.6070.0274.32
DSULDA42.1848.7352.4554.8237.1842.7347.4548.8267.8977.0980.6986.11
DRSLDA42.7346.2050.8956.5036.7340.2044.8950.5069.2283.6083.7087.04
DRSMDA52.1853.3357.8560.8347.1847.3351.8554.8371.8683.3384.5186.25
DRSLDA+IIKC52.3057.8061.7064.9246.3051.8055.7059.9274.4587.5290.3291.02
DERSLDA47.5254.5355.8562.9242.5249.5350.8556.9273.2085.1085.2088.25
DAFLDA45.6450.9152.6856.1240.5846.3748.2553.7670.4583.6884.7185.99
DSCLDA54.5557.8062.2265.5850.3253.5757.9961.3574.7988.2890.9891.47
Table 4. The accuracy of DSCLDA, R3D-CNN, I3D, and Transformer on the HGR and CIFAR-100 datasets. The bold value represents the highest value of the column.
Table 4. The accuracy of DSCLDA, R3D-CNN, I3D, and Transformer on the HGR and CIFAR-100 datasets. The bold value represents the highest value of the column.
HGRCIFAR-100
MethodAcc. (%)MethodAcc. (%)
DSCLDA90.37DSCLDA63.45
R3D-CNN83.80R3D-CNN90.62
I3D85.70I3D94.82
Transformer87.60Transformer95.03
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Chen, X.; Feng, M.; Liu, J. Distributed Sparse Manifold-Constrained Optimization Algorithm in Linear Discriminant Analysis. J. Imaging 2025, 11, 81. https://doi.org/10.3390/jimaging11030081

AMA Style

Zhang Y, Chen X, Feng M, Liu J. Distributed Sparse Manifold-Constrained Optimization Algorithm in Linear Discriminant Analysis. Journal of Imaging. 2025; 11(3):81. https://doi.org/10.3390/jimaging11030081

Chicago/Turabian Style

Zhang, Yuhao, Xiaoxiang Chen, Manlong Feng, and Jingjing Liu. 2025. "Distributed Sparse Manifold-Constrained Optimization Algorithm in Linear Discriminant Analysis" Journal of Imaging 11, no. 3: 81. https://doi.org/10.3390/jimaging11030081

APA Style

Zhang, Y., Chen, X., Feng, M., & Liu, J. (2025). Distributed Sparse Manifold-Constrained Optimization Algorithm in Linear Discriminant Analysis. Journal of Imaging, 11(3), 81. https://doi.org/10.3390/jimaging11030081

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop