Next Article in Journal
Transient Dynamic Response Analysis of Two-Dimensional Saturated Soil with Singular Boundary Method
Previous Article in Journal
Distributionally Robust Optimization Model for a Minimum Cost Consensus with Asymmetric Adjustment Costs Based on the Wasserstein Metric
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exponential Graph Regularized Non-Negative Low-Rank Factorization for Robust Latent Representation

1
School of Computer Science (School of Intelligent Auditing), Nanjing Audit University, Nanjing 211815, China
2
Jiangsu Modern Intelligent Audit Integrated Application Technology Engineering Research Center, Nanjing Audit University, Nanjing 211815, China
3
School of Electronic Information, Qingdao University, Qingdao 266071, China
4
Key Laboratory of Intelligent Information Processing, Nanjing Xiaozhuang University, Nanjing 211171, China
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(22), 4314; https://doi.org/10.3390/math10224314
Submission received: 21 October 2022 / Revised: 11 November 2022 / Accepted: 15 November 2022 / Published: 17 November 2022

Abstract

:
Non-negative matrix factorization (NMF) is a fundamental theory that has received much attention and is widely used in image engineering, pattern recognition and other fields. However, the classical NMF has limitations such as only focusing on local information, sensitivity to noise and small sample size (SSS) problems. Therefore, how to develop the NMF to improve the performance and robustness of the algorithm is a worthy challenge. Based on the bottlenecks above, we propose an exponential graph regularization non-negative low-rank factorization algorithm (EGNLRF) combining sparseness, low rank and matrix exponential. Firstly, based on the assumption that the data is corroded, we decompose a given raw data item with a data error fitting noise matrix, applying a low-rank constraint to the denoising data. Then, we perform a non-negative factorization on the resulting low-rank matrix, from which we derive the low-dimensional representation of the original matrix. Finally, we use the low-dimensional representation for graph embedding to maintain the geometry between samples. The graph embedding terms are matrix exponentiated to cope with SSS problems and nearest neighbor sensitivity. The above three steps will be incorporated into a joint framework to validate and optimize each other; therefore, we can learn latent data representations that are undisturbed by noise and preserve the local structure of known samples. We conducted simulation experiments on different datasets and verified the effectiveness of the algorithm by comparing the proposed with the lasting ones related to NMF, low rank and graph embedding.

1. Introduction

With the continuous development of science and technology, various electronic devices are constantly bringing forth new ideas, the volume of data obtained is growing explosively and the “dimensionality curse” is becoming more and more significant [1,2]. Therefore, how to efficiently obtain effective feature representations from massive data has always been the focus of research. In addition, the practical application scenario is usually a typical small sample size (SSS) scenario [3,4,5]: abundant unlabeled data and a small amount of labeled data, which undoubtedly brings huge challenges to the multivariate data conversion algorithm. Generally speaking, a good data transformation method will have the basic characteristics of making the data structure clearer and reducing the data dimension. The more classic data dimensionality reduction algorithms include principal component analysis (PCA) [6], linear discriminant analysis (LDA) [7], locality preserving projections (LPP) [8], independent component analysis (ICA) [9], projection pursuit (PP) [10] and non-negative matrix factorization (NMF) [11], etc., which lead to differences in statistical properties due to different constraints imposed.
NMF stands out due to its natural non-negativity and its part-based representation properties. On the one hand, in many real-world data, negative values are meaningless; on the other hand, the theory that the perception of the whole is made up of the perception of the parts that make up the whole captures, in a sense, the essence of intelligent data description [12]. Due to the enhancement of semantic interpretability under non-negativity and the resulting sparsity that restrains the adverse effects of external changes to a certain extent, NMF has become a basic tool for multivariate data analysis and has been successfully applied to research in image engineering [13], pattern recognition [14], data mining [15], spectral data analysis [16], complex networks analysis [17] and other fields. However, the disadvantages of NMF are also significant. Non-negative constraints may help some models learn part-based representations, but this does not mean that only non-negative constraints are sufficient; NMF uses Euclidean distance as a measure of loss, resulting in sensitivity to noise or corrosion in the original data; the processing objects of NMF are essentially vectors, which makes encountering SSS problems inevitable [5].
Many NMF-based extension efforts have been developed to address these challenges. Generally, adding a regularization term based on the NMF framework, called constrained NMF, can bring about the effect of improving performance. Li et al. [18] presented a local learning NMF (LNMF) which, based on the NMF framework, defines an objective function to impose localization constraints for learning spatial localization. Immediately after, Hoyer et al. [19] proposed to combine the concept of sparseness to improve NMF and made relevant theoretical proofs and algorithm implementations. Wang et al. [20] imposed Fisher constraints on the NMF algorithm and proposed a fisher NMF (FNMF) to encode the discriminative information of classification problems. Since the above only considers the Euclidean space structure and related literature find that data generally have potential popular structures in high-dimensional data, NMF has a new direction for improvement. Cai et al. [21] proposed a graph regularization NMF (GNMF) that encodes geometric information by constructing an affinity graph. However, the way GNMF constructs graphs only considers the relationship between pairs of points, ignoring the higher-order relationship between multiple points. Zeng et al. [22] developed an improved GNMF, which introduced the concept of a hypergraph, called hypergraph regularization NMF (HNMF).
Soon, it was discovered that the original data contained many disturbances or corrosions, for which the NMF algorithm exhibited extremely unstable performance [23]. Hence, a robust NMF is on the agenda. Kong et al. [24] proposed an NMF model using the L 2 , 1 norm loss function to handle noise and outliers with its sparsity. Since the aforementioned algorithms [23,24] are based on local learning and global information is superior to local information in robustness, Lu et al. [25] consider an NMF extension that integrates global information and partial information. Different from sparse representation-based algorithms [26], low-rank representation (LRR) [27] aims to find the lowest rank representation of a vector set, which can capture global information well. There are many LRR-based extensions in the field of feature extraction and data transformation [28,29]. Inspired by LRR, they proposed a new low-rank nonnegative factorization (LRNF) method and used the L1 norm to sparsely constrain the assumed noise matrix. Further, to consider the label information of the samples, Lu et al. [30] introduced structural incoherence into LRNF and proposed a new algorithm called SILR-NMF. Li et al. [31] provided a unified framework of graph regularization non-negative low-rank matrix factorization (GNLMF) to improve the algorithm’s robustness to noise and popular structures. However, in its essence, it is only simple addition of low rank and GNMF and the two parts are not closely related. Recently, to learn a graph that can better represent the similar relationship between samples, Lu et al. [32] proposed a low-rank adaptive neighborhood graph embedding. He et al. [33] proposed low-rank NMF (LNMFS) based on the Stiefel manifold, using orthogonal constraints and graph smoothness to improve algorithm robustness.
Most of the algorithms mentioned above add various constraints to NMF, such as sparsity, low rank, graph embedding, linear discrimination, etc., while little is mentioned about how to improve NMF to improve its robustness to few-shot scenarios. These improve the algorithm’s performance, but NMF is carried out in the case of converting samples to column vectors, which inevitably faces the problem of SSS. In addition, we cannot ignore the graph structure of the sample and the global information, both of which can obtain distinct information from the original data and the combination of the two can achieve a win–win situation. The nearest neighbor-sensitive problem cannot be ignored when adding graph-related constraints to NMF to preserve geometric structure information. Since the above problems are more significant in dimensionality reduction algorithms such as LDA and LPP, and these algorithms have also been maturely developed  [34,35,36], the introduction of matrix exponential seems to bring unexpected benefits [37,38].
Inspired by the above, this paper combines low rank, graph embedding, matrix exponential and sparsity to propose a new robust NMF, called exponential graph regularization non-negative low-rank factorization (EGNLRF). Figure 1 performs the framework of the EGNLRF. Firstly, we incorporate low-rank into the NMF framework and use the L 2 , 1 norm to sparsely constrain the noise terms in the low-rank restoration. Secondly, we introduce graph embeddings using a regularization term to exploit local relations between samples as well as label information. To further improve the robustness of the algorithm in SSS scenarios, we perform matrix exponential processing on the graph embedding terms. Finally, we develop an alternating iterative algorithm to optimize the latent representation of denoised low-rank data.
The main contributions of this paper are as follows.
  • The matrix exponent operation is carried out in the graph embedding items to expand the SSS scenarios’ application and increase the robustness of the algorithm. The attribute characteristic of the matrix index enables it to solve the matrix singularity problem.
  • The nearest neighbor sensitivity caused by the introduction of graph structure can be alleviated because of the power series definition of the matrix exponent itself. Furthermore, the matrix index also provides more information for the algorithm, such as the potential hyper-order geometric structure among samples. The matrix index can spread the distance between samples and expand the margin between labels for classification.
  • An exponential graph regularization non-negative low-rank decomposition algorithm is proposed, which incorporates graph regularization and low rank into the NMF framework for joint optimization. The algorithm integrates the global and local information of samples; not only can it learn latent representations that are not disturbed by noise but it can also maintain the local structure of known samples.
  • The optimization process of the EGNLRF algorithm is derived in detail and its convergence is proved. Compared with other methods, the superiority of EGNLRF was verified by a simulation comparison experiment using the nearest neighbor classifier recognition rate as an evaluation index.

2. Related Work

NMF is one of the classical matrix factorization methods, which focuses on finding non-negative representations of factor matrices that approximate the original matrix.
Given an original data matrix X = [ x 1 , x 2 , , x n ] R + m × n , composed of n sample data column vectors. NMF aims to find two non-negative matrix factors U R m × k and V R n × k that can approximate the original matrix X, which can be described as
X U V T
For the convenience of calculation, the square of the F-norm of the difference between the two matrices is generally used as the loss function to quantify the performance.
F l o s s = X U V T F 2
In practical applications, NMF usually has k m and k n and the theory limits it to only an additive combination of basis vectors, so NMF essentially finds a partial-based latent low-dimensional representation of the original matrix. At present, this method has shown excellent performance in image engineering [13], pattern recognition [14] and other fields and has been widely used based on its simple and effective advantages.
Related research [39,40] shows that NMF algorithms can learn part-based representations, similar to how the brain perceives the world. In practical applications, to meet different needs, related scholars have studied the variants of NMF. This section will briefly introduce some representative NMF variant algorithms.
The geometric structure information of the samples is considered first. To reveal the hidden semantics, the GNMF algorithm that respects the local geometry between samples is proposed
m i n U , V 0 X U V T F 2 + λ t r ( V T L V ) .
where t r ( V T L V ) is the graph regularization term and L is the Laplacian matrix. Similarly, Zeng et al. [22] also proposed an improved algorithm for NMF from a geometrical point of view. GNMF uses affinity graphs to encode geometric information, but simple graphs can only represent relationships between pairs of points, ignoring higher-order information between multiple points. Therefore, they introduce hypergraphs and propose HNMF
m i n U , V 0 X U V T F 2 + λ t r ( V T L h y p e r V ) .
where L h y p e r is the hyper-graph Laplacian matrix.
Furthermore, since the above algorithms are sensitive to noise and have low robustness, some robust NMF has gradually entered people’s field of vision. The global representation of data is more robust than the part-based representation, so Lu et al. [25] proposed an LRNF algorithm that can learn both part- and global-based.
m i n A , E , U , V A U V T F 2 + α A * + β E 1 s . t . X = A + E , U , V 0 .
where E is the noise matrix and A is the clean data matrix after denoising. Similar work was conducted by Li et al. [31], who proposed an NMF algorithm based on low-rank recovery. To further improve the robustness of the algorithm to popular geometric structures, the GNLMF algorithm is proposed.
m i n A , E , U , V X A E F 2 + α t r V T L V + β t r U U T s . t . A = U V T , A , U , V 0 r a n k ( A ) r , c a r d ( E ) s
where t r V T L V is the graph embedding regularization term and t r U U T is the Tikhonov regularization term of U. GNLMF only incorporates the graph embedding into the objective function, while the low-rank restoration is only a constraint of the objective function. In essence, GNLMF is a simple combination of low-rank recovery and GNMF, without joint optimization of the objective function.
It is undeniable that the above methods all improve NMF from different angles. However, they either do not consider geometric information or do not take into account the noise sensitivity. GNLMF takes all of these into account, but only in simple combinations and is incapable of dealing with the problem of SSS. In addition, a series of problems brought about by the introduction of graph embeddings [38,41], such as nearest neighbor sensitivity and matrix singularity, have not been taken seriously. In response to these issues, we propose an exponential graph regularization non-negative low-rank factorization algorithm combining sparseness, low rank and matrix exponential.

3. Exponential Graph Regularized Non-Negative Low-Rank Factorization

In this section, we first introduce the composition process of the objective function of EGNLRF. Then, the optimization algorithm of the function and the solution process are given. Finally, we analyze the time complexity and convergence of the algorithm.

3.1. Formulations and Problem Setting

In the real world, the original image information is often more or less corroded by noise, illumination or occlusion, etc. Hence, a related robust NMF algorithm is proposed. To improve the robustness of the algorithm, GNMF, LRNF and GNLMF combine different algorithm methods to explore the improvement and performance optimization of NMF from different angles. GNMF combines graph embedding and NMF to consider the local relationship between training data; LRNF introduces noise terms and low rank into NMF, which mainly focuses on global information. GNLMF synthesizes global information and local relations, but it simply combines low-rank and graph regularization NMF without joint optimization on the objective function.
Given an original data matrix X = [ x 1 , x 2 , , x n ] R + m × n , NMF decomposes it into two non-negative factor matrices. Related research has proved that the extended regularization term in NMF can improve the robustness of NMF. Therefore, inspired by the above algorithms, we combine sparse, low-rank and graph embeddings to obtain NMF variants as shown below
m i n A , E , U , V A U V T F 2 + α r a n k ( A ) + β c a r d ( E ) + λ t r ( V T L V ) s . t . X = A + E , U , V 0 .
It is assumed that the noise is sparse, which means that the sparseness of the noise term is a key factor affecting the performance of the algorithm. Therefore, the noise matrix E is sparsely constrained using the L 2 , 1 norm. In addition, since Equation (7) belongs to the NP problem, this paper uses a convex optimization problem to replace it
m i n A , E , U , V A U V T F 2 + α A * + β E 2 , 1 + λ t r ( V T L V ) s . t . X = A + E , U , V 0 .
The above three algorithms all have their problems when finding low-rank latent representations. The LRNF algorithm does not consider the information from the original training samples. When experiments are performed on the same dataset, the latent representation obtained with LRNF does not change under the division of different training sets and test sets. This can be improved by introducing graph embeddings. GNMF and GNLMF algorithms introduce graph embedding to extract local information of samples, but the SSS problem, nearest neighbor sensitivity and matrix singularity brought by the introduction of graph embedding have not been solved. In addition, graph embedding only considers the relationship between pairs of points, ignoring the higher-order information between multiple points. Therefore, we introduce exponential matrices into graph embeddings to address these challenges.

3.2. Matrix Exponential

As a theory widely used in ordinary differential equations, Markov chain analysis and other fields, the exponential matrix also occupies an important seat in pattern recognition’s image matrix correlation algorithm. This section will briefly introduce the definition and properties of exponential matrices that are closely related to this paper. Given a square matrix H of order n, ts exponential form is expanded as follows:
e x p ( H ) = I + H + H 2 2 ! + + H m m ! +
where I is the identity matrix of order n. According to Equation (9), we can obtain some properties of the matrix exponential function.
  • Firstly, if matrix H has eigenvectors v 1 , v 2 , , v n corresponding to the eigenvalues λ 1 , λ 2 , , λ n , then e x p ( H ) has the same eigenvectors v 1 , v 2 , , v n corresponding to the eigenvalues e λ 1 , e λ 2 , , e λ n . This means that e x p ( V T L V ) can preserve and enhance the pivotal information of the graph, which is the key to matrix V T L V being replaced by e x p ( V T L V ) .
  • Secondly, e x p ( H ) is a finite full-rank matrix. The matrix singularity problem will be solved easily after introducing the matrix exponential function.
  • Finally, if D is a nonsingular matrix, we have
    e x p ( D 1 H D ) = D 1 e x p ( H ) D .
    Since symmetric matrices are diagonalizable, the Laplacian matrix exponent optimizes the operation by this property.
Furthermore, the power series definition of matrix exponents strongly suggests that e x p ( V T L V ) can improve small-shot problems and nearest-neighbor-sensitive problems and obtain hidden higher-order graph structure information. Hence, after introducing the matrix index, the objective function can be obtained as follows:
m i n A , E , U , V A U V T F 2 + α A * + β E 2 , 1 + λ t r ( e x p ( V T L V ) ) s . t . X = A + E , U 0 , V 0 .
The model synthesizes the global and local information of the samples, along with low-rank constraints, noise sparsity and exponential Laplacian matrices. The clean data matrix A, noise matrix E and NMF factor matrices U, V are obtained with Equation (10) and we use U T A or V T as the underlying representation for all instances as appropriate. The resulting latent representation will be used as the EGNLRF classification result for standard classification.

3.3. Optimization and Solution

There are four variables in Equation (10), which is not a joint convex optimization problem and cannot be solved directly. This paper considers using the alternating direction method of multipliers (ADMM) [42] or the augmented Lagrange multipliers (ALM) [43] to solve the multivariable optimization non-convex matter in Equation (10). To solve the nuclear norm problem, we first convert Equation (10) into an equivalent problem as follows:
m i n A , E , U , V A U V T F 2 + α J * + β E 2 , 1 + λ t r ( e x p ( V T L V ) ) s . t . X = A + E , A = J , U , V 0 .
which can be disposed of with the ALM function as shown below.
L ( A , J , E , U , V , Y 1 , Y 2 , μ ) = A U V T F 2 + α J * + β E 2 , 1 + λ t r ( e x p ( V T L V ) ) + μ 2 X A E + Y 1 μ F 2 + μ 2 A J + Y 2 μ F 2 1 2 μ ( Y 1 F 2 + Y 2 F 2 ) .
It can be observed that the ALM function is convex on any one variable when the rest of the variables are fixed in Equation (12). Therefore, the optimization problem can be solved by alternating iterative updates. Next, each variable’s solution details and updated formulation are given separately.

3.3.1. Update A, with Remaining Variables Fixed

When the remaining variables are fixed to compute A, the objective function is
m i n A A U V T F 2 + μ 2 A X + E Y 1 μ F 2 + μ 2 A J + Y 2 μ F 2 .
Using the Lagrangian method to find the local optimal solution, we can obtain
A = 2 U V T + μ X E + J + Y 1 μ Y 2 μ 2 + 2 μ

3.3.2. Update J, with Remaining Variables Fixed

It can be observed that when the remaining variables are fixed, the problem is transformed into a convex relaxation of the rank minimization problem and Equation (15) can be obtained.
m i n J α μ J * + 1 2 J A + Y 2 μ F 2 .
Aiming at this kind of problem, Cai et al. [44] proposed a singular value thresholding algorithm and proved its effectiveness and convergence. Therefore, J can be iterated through the following steps.
First, we give Q 1 = A + Y 2 μ and then perform a singular value decomposition operation on Q 1 .
Q 1 = U s v d Σ V s v d * , Σ = d i a g ( { σ i } 1 i r ) .
Then, for τ > 0 , the soft threshold operator D τ has the following definition
D τ ( Σ ) : = d i a g σ i τ + .
Thus, the iterative update rule for J is given by Equation (18).
J = U s v d D α μ ( Σ ) V s v d * ,

3.3.3. Update E, with Remaining Variables Fixed

After fixing the rest of the variables except E, we obtain the following function
m i n E β μ E 2 , 1 + 1 2 X A E + Y 1 μ F 2 .
For convex problems such as Equation (19), relevant scholars [27] have proved the existence of closed solutions. If E is the optimal solution of Equation (19), then each column of E is calculated as
E ( : , i ) = q i λ q i q i , if β μ < q i , 0 , otherwise .
where Q 2 = [ q 1 , q 2 , , q i , ] = X A + Y 1 μ .

3.3.4. Update U, with Remaining Variables Fixed

Fixing the remaining variables to solve U, we can obtain the following function
m i n U A U V T F 2 ,
which is similar to the objective function of NMF. Therefore, we can directly obtain the update rule of U:
u i k u i k ( X V ) i k ( U V T V ) i k .

3.3.5. Update V, with Remaining Variables Fixed

When the remaining variables are fixed to solve V, the objective function is
m i n V A U V T F 2 + λ t r ( e x p ( V T L V ) ) .
This formula is different from Equation (21), but it is still a convex function, which can be solved by a simple calculation.
Let ϕ be the Lagrange multiplier; the function can be obtained as
L ( V ) = t r A A T 2 t r t r A V U T + t r U V T V U T + λ t r e x p V T L V + t r ϕ V T .
Taking the partial derivative concerning L ( V ) and using the KKT condition, it can be obtained that
( V U T U ) j k v j k + λ L V e x p V T L V j k v j k = ( X T V ) j k v j k .
This results in the following update rule:
v j k v j k X T V + W V e x p V T L V j k V U T U + λ D V e x p V T L V j k .
For the uncertainty of the solution of the NMF objective function, the constraint that the Euclidean length of each column vector in the matrix U (or V) is 1 is generally added in practice. We add that restriction to U, which can be achieved in the following ways:
u i k u i k i u i k 2 , v j k v j k i u i k 2
Algorithm 1 summarizes the iterative process of the EGNLRF algorithm. Among them, U and V are initialized by random matrixes.
Algorithm 1 EGNLRF
Require: 
Data matrix X, parameter α , β , λ in objective function and the maximum number of iterations m a x t ;
Initialize: 
U, V, E = 0 , J = 0 , Y 1 = Y 2 = 0 , μ > 0 , ρ > 0 and t = 1 ;
1:
repeat
2:
    Update A by Equation (14);
3:
    Update J by Equation (18);
4:
    Update E by Equation (20);
5:
    Update U by Equation (22);
6:
    Update V by Equation (26);
7:
    Update Lagrange multipliers as follows:
             Y 1 Y 1 + μ ( X A E ) ;
             Y 2 Y 2 + μ ( A J ) ;
8:
     μ m i n ( ρ μ , m a x μ ) ;
9:
    t t + 1 ;
10:
until  t > m a x t or Convergence
11:
Normalize each column of U, keeping U V T unchanged
12:
Return the optional solution ( A , J , E , U , V )

3.4. Convergence and Computational Complexity

Problem (11) is a non-convex function on all variables, which is a typical NP-hard problem, so we consider iterative optimization to approximate it. Hence, the corresponding convergence requirements are relaxed appropriately, which requires that the value of the objective function does not increase during the iterations. When the input data is non-negative, it can be known that the objective function in Equation (11) is non-negative and has a lower bound. In the five steps of the iterative algorithm, the ADMM is used to decrease the value of the objective function. Steps 3 and 4 have been proven to have closed solutions [27,44], which are decreasing and convergent. The convergence proof of step 2 is similar to the study by Lu et al. [25], while the convergence of steps 5 and 6 is proved by Cai et al. [21]. To more intuitively reflect the convergence of the proposed algorithm, we verified it on different datasets, as shown in Figure 2. We can observe that the polyline of FERET in Figure 2 has an upward perturbation at the beginning of the iteration. Our zero-initialization of matrices J and E, which are values that J and E do not end up with, creates a situation where the objective function value is abnormally small in the first few iterations.
In addition, the time complexity of the algorithm is also worth paying attention to. In Algorithm 1, the time complexity of step 2 is O ( m n 2 ) . The time cost of step 3 is O ( n 3 ) , which has an SVD operation. The computational complexity of step 4 is O ( m n ) and the time complexity of steps 5 and 6 is O ( n 2 k + k 3 + m n k ) which is only a simple matrix multiplication operation in these steps. In general, there are some explicit m > > k , n > > k in the data transformation. Therefore, the overall time complexity of the algorithm is O ( t ( m n 2 + n 3 ) ) , where t is the number of iterations.

3.5. Discussion of Other Optimizations

The ADMM optimization method described above is only the most common method for solving non-convex problems at present, which is applied in many fields. The alternating nonnegative least squares (ANLS) algorithm is also an effective way to solve the non-convex problem of NMF. Many scholars have made innovations and developments in the ANLS framework, such as the projected gradient method, active-set-like method, hierarchical alternating least squares, etc. However, the key difference from ADMM is that there is no constraint between the variables that are solved alternately in ANLS. For the application of the objective function with additional constraints in the ANLS framework, the easiest way is to convert the constraints into corresponding regularization terms and add them to the objective function. The ADMM optimization method described above is only the most common method for solving non-convex problems at present, which is applied in many fields. The alternating nonnegative least squares (ANLS) framework [45,46,47] is also an effective way to solve the non-convex problem of NMF. Many scholars have made innovations and developments in the ANLS framework, such as the projected gradient method [45], active-set-like method [46], hierarchical alternating least squares [47], etc. However, the key difference from ADMM is that there is no constraint between the variables that are solved alternately in ANLS. For the application of the objective function with additional constraints in the ANLS framework, the easiest way is to convert the constraints into corresponding regularization terms and add them to the objective function.

4. Experiments and Evaluations

In this section, we use real-world datasets to verify the performance of the proposed algorithm. Furthermore, experiments include tests of different degrees of corrosion such as noise, occlusion, etc.

4.1. Real World Data Sets

We evaluate EGNLRF on four public datasets: object recognition (COIL20 [48]), palmprint recognition (PolyU palmprint database [49]) and face recognition (AR [50] and FERET [51]). Some of the original images for these four datasets are shown in Figure 3. It can be observed that AR is a dataset with natural occlusion. To verify the robustness of the algorithm, we add different erosions to the dataset for testing.
  • COIL Data Set [48]: The COIL20 database contains 20 objects, each with 72 gray images taken from different viewpoints.
  • AR Data Set [50]: The AR database contains more than 4000 facial images of 126 people, including frontal views with different facial expressions, lighting conditions and occlusions. The pictures are collected in two sessions two weeks apart and 120 individuals participated in both sessions.
  • FERET Data Set [51]: The FERET database contains 200 people, each with seven photos with different lighting conditions.
  • PolyU palmprint Data Set [49]: The PolyU palmprint database contains six palmprint photos of 100 people in varying shades of light.

4.2. Baseline Algorithms

The following are the selected benchmark algorithms for comparison with EGNLRF.
  • NMF [11]: NMF is a classical matrix factorization algorithm that achieves nonlinear dimensionality reduction through non-negative constraints.
  • ELPP [38]: ELPP is an improved algorithm of LPP to avoid matrix singularity.
  • GNMF [21]: GNMF takes into account the local structure between samples by combining affinity graph encoding and NMF.
  • LRNF [25]: LRNF enables the learning of partial and global latent representations by incorporating low-rank into the NMF framework.
  • LNMFS [33]: LNMFS algorithm based on the Stiefel manifold is proposed, which considers orthogonal constraints and image smoothness.
  • LRAGE [32]: LRAGE is an unsupervised feature extraction method using adaptive probabilistic neighborhood graph embedding and low-rank constraints.
Table 1 presents the objective functions and algorithm time complexity of these baseline algorithms. It is clearly observed that NMF has the lowest time complexity and the simplest objective function. The remaining algorithms either add various constraints or introduce special operations, resulting in more or less an increase in algorithm complexity. Among them, the time requirements of GNMF are always relatively small, while the time requirements of other algorithms will produce different sorting according to different m and n. In the SSS scenario, which is in the case of m > n or m > > n , ELPP, LRNF and LRAGE will have a high time cost. Since V is used for graph embedding and exponentiating, EGNLRF keeps the time cost at a moderate level.

4.3. Experimental Setup

Before reporting detailed experimental results, how to tune the hyperparameters of the algorithm must be explained. In the above comparison algorithms, GNMF, LNMFS and LRAGE have one parameter while LRNF has two parameters. We select the values of these parameters from { 10 3 , 10 2 , 10 1 , 1 , 10 , 10 2 , 10 3 } by referring to the article proposing these algorithms. In EGNLRF, this paper selects the three parameters α , β and λ from { 10 4 , 10 3 , 10 2 , 10 1 , 1 , 10 , 10 2 , 10 3 , 10 4 } through cross-validation. In the experiment, the nearest neighbor classification is used to calculate the recognition rate of the samples in the test set. Figure 4 shows the recognition rate on the COIL20 data set as a function of one parameter while the rest of the parameters are fixed. When conducting experiments in different databases, p samples will be randomly selected as training samples according to the size of the database and the rest constitute the test set. Among them, p = 2 , 3 , 4 in PolyU palmprint database and p = 9 , 12 , 15 in AR database. In FERET and COIL20 database, p = 2 , 3 , 4 , 5 and p = 20 , 30 , 40 , 50 , respectively. In specific experiments, the initialization of several algorithms will randomly generate U and V matrices. Hence, we uniformly initialize U and V matrices and input them into the algorithm as a fixed parameter to control variables.

4.4. Experimental Results

In this subsection, the EGNLRF algorithm will be tested on FERET, AR, COIL20 and PolyU palmprint, which cover diversified fields, such as face recognition, object recognition and palmprint recognition.
First, we conduct experiments on the FERET database. Face recognition does not require very high accuracy in the image, but it is usually corrupted by a variety of noise. The FERET database is a typical database used to verify whether the algorithm is robust to different lighting and camera angles. Furthermore, we added Gaussian noise of different densities to the images of the FERET database for experiments. Figure 5 shows part of the image of FERET with noise and the experimental results are shown in Figure 6. It can be observed that the image recognition rates of LNMFS, LRAGE and EGNLRF algorithms are at a superior level under different dimensionality reductions. As for the three figures with the number of training samples on the horizontal coordinate, the EGNLRF algorithm always has an advantage over the comparison algorithm, which is more obvious when the number of training samples is low or the image contains noise. Another common corrosion of images is occlusion.
In addition, the setting of the nearest neighbor parameter is a key point worth noting, so we give the algorithms with different nearest neighbor parameters for comparison. The nearest neighbor was evaluated from 2 to 10 on the FERET database and a boxplot is drawn using the result, as shown in Figure 7.
Considering more realistic scenarios, we choose to conduct experiments on face database AR containing natural occlusions, such as scarves and sunglasses, rather than adding occlusions randomly. Figure 5 shows part of the image of AR with noise. It can be observed in Table 2 and Table 3 that the performance of the proposed algorithm is superior in the face database with natural occlusion, which demonstrates that the EGNLRF algorithm is robust to occlusion corrosion. And we can obviously obtain that algorithms with low-rank constraints will outperform other algorithms without it in the performance of the AR database. The low-rank recovery framework can better deal with occlusion corrosion because the low-rank constraint can restore the occluded part to a certain extent.
Furthermore, to verify the algorithm’s effectiveness in data conversion, we also conducted comparative experiments on the COIL20 object and PolyU palmprint databases. Inevitably, we add noise corrosion or occlusion corrosion to the original database to verify the robustness of the algorithm, concerning which Figure 8 and Figure 9 show some images. The results of all the experiments on COIL20 and PolyU palmprint are shown in Figure 10 and Figure 11. We can find that EGNLRF is always in the dominant position in the original database, whether it changes with the dimension or with the number of training samples. It is undeniable that the algorithms LRNF, GNMF, LNMFS and LRAGE all show good anti-corrosion ability in COIL20 in (b) and (c) of Figure 10, while the performance of NMF and ELPP algorithms fluctuates sharply with the addition of noise or occlusion. However, the anti-jamming ability of the proposed algorithm is better than that of the contrast algorithms, which is theoretically due to the synthesis of a variety of regular term constraints. It is also reflected in (e) and (f) of Figure 10 and the gap in recognition rate will increase with the increase of interference. In the experiment of the palmprint database, because the palmprint itself has high similarity, it is more sensitive to noise and the requirements for data conversion are the most stringent in these databases. Hence, the corrosion parameter setting of the robustness experiment becomes smaller accordingly. It is easy to observe that the performance of most comparison algorithms changes dramatically with the addition of corrosion in Figure 11. LNMFS can avoid the significant degradation of its performance, while the proposed algorithm is better. By comparing all the subgraphs in Figure 11, it can be found that if there is only a piece of single geometric structure in a low-rank recovery structure or non-negative matrix decomposition structure in the algorithm, it can only deal with less noise corrosion and occlusion corrosion. Only by merging these three into a joint framework can we deal with more serious corrosion, such as (e) and (f) in Figure 11.

4.5. Observations and Discussions

Generally speaking, by observing and comparing Figure 6, Figure 7, Figure 10 and Figure 11 and Table 2 and Table 3 above, we can draw the following summaries.
In most cases, LNMFS and LRAGE perform better than earlier algorithms because they employ more considerations or improvements concerning data conversion than before. Most of these algorithms share the same theoretical basis, such as NMF, graph embedding and low rank. The addition of these regularization terms will improve the performance of the algorithm.
Joint optimization with multiple constraints has higher classification accuracy than single constraints such as NMF and ELPP. The part-based information of the image and the geometric structure information of the image are both important parts of the image. This implies that the combination of multiple constraints that can obtain different distinct image information can obtain a highly robust and efficient data conversion algorithm.
LNMFS incorporates a variety of constraints into the objective function, but it is still not as superior as the proposed algorithm. This is because the algorithm that proposes the indexed graph regularization can alleviate the SSS problems and improve the nearest neighbor-sensitive problem. In addition, the exponential graph regularization term can also improve the hidden high-order geometric structure information for the algorithm, which is determined by the definition of its power series.

5. Conclusions

The robustness of data transformation or feature representation is not strong and the matrix singularity problem in SSS scenarios is always a research bottleneck. To solve these problems, this paper considers a low-rank recovery graph embedded NMF framework and introduces matrix exponential to alleviate the SSS problems and nearest neighbor sensitivity and obtain more high-order geometric structure information. Simulation experiments show that the proposed algorithm has absolute advantages over the comparison algorithm in most cases of the four databases, which is especially obvious when the images are subjected to various corrosions. For future work, we expect to design a more concise federation framework to realize automatic and efficient data transformation, which can simplify self-service parameter tuning.

Author Contributions

Conceptualization, G.Y., L.Z. and M.W.; Methodology, G.Y., L.Z. and M.W.; Software, G.Y., L.Z. and M.W.; Validation, L.Z.; Data curation, L.Z.; Writing–original draft, G.Y., L.Z. and M.W.; Writing–review–editing, G.Y., L.Z. and M.W.; Supervision, G.Y. and M.W.; Project administration, G.Y., L.Z. and M.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Postgraduate Research & Practice Innovation Program of Jiang Su Province KYCX22_2219, the National Natural Science Foundation of China under Grant 62172229, 61876213, the Natural Science Foundation of Jiangsu Province BK20211295, BK20201397, the Jiangsu Key Laboratory of Image and Video Understanding for Social Safety of Nanjing University of Science and Technology under Grants J2021-4, funded by the Qing Lan Project of Jiangsu University and the Future Network Scientific Research Fund Project SRFP- 2021-YB-25.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses or interpretation of data; in the writing of the manuscript or in the decision to publish the results.

References

  1. Zebari, R.; Abdulazeez, A.; Zeebaree, D.; Zebari, D.; Saeed, J. A comprehensive review of dimensionality reduction techniques for feature selection and feature extraction. J. Appl. Sci. Technol. Trends 2020, 1, 56–70. [Google Scholar] [CrossRef]
  2. Luo, F.; Zhang, L.; Du, B.; Zhang, L. Dimensionality Reduction with Enhanced Hybrid-Graph Discriminant Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5336–5353. [Google Scholar] [CrossRef]
  3. Jia, S.; Jiang, S.; Lin, Z.; Li, N.; Xu, M.; Yu, S. A survey: Deep learning for hyperspectral image classification with few labeled samples. Neurocomputing 2021, 448, 179–204. [Google Scholar] [CrossRef]
  4. Konietschke, F.; Schwab, K.; Pauly, M. Small sample sizes: A big data problem in high-dimensional data analysis. Stat. Methods Med. Res. 2021, 30, 687–701. [Google Scholar] [CrossRef] [PubMed]
  5. Sharma, A.; Paliwal, K.K. Linear discriminant analysis for the small sample size problem: An overview. Int. J. Mach. Learn. Cybern. 2015, 6, 443–454. [Google Scholar] [CrossRef] [Green Version]
  6. Turk, M.; Pentland, A. Eigenfaces for Recognition. J. Cogn. Neurosci. 1991, 3, 71–86. [Google Scholar] [CrossRef]
  7. Martinez, A.; Kak, A. PCA versus LDA. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 228–233. [Google Scholar] [CrossRef] [Green Version]
  8. He, X.; Niyogi, P. Locality Preserving Projections. In Proceedings of the Advances in Neural Information Processing Systems NIPS, Vancouver, BC, Canada, 8–13 December 2003; Volume 16. [Google Scholar]
  9. Hyvärinen, A.; Oja, E. Independent component analysis: Algorithms and applications. Neural Netw. 2000, 13, 411–430. [Google Scholar] [CrossRef] [Green Version]
  10. Huber, P.J. Projection Pursuit. Ann. Stat. 1985, 13, 435–475. [Google Scholar] [CrossRef]
  11. Lee, D.; Seung, H.S. Algorithms for non-negative matrix factorization. In Proceedings of the Advances in Neural Information Processing Systems NIPS, Denver, CO, USA, 27 November–2 December 2000; Volume 13. [Google Scholar]
  12. Wang, Y.X.; Zhang, Y.J. Nonnegative matrix factorization: A comprehensive review. IEEE Trans. Knowl. Data Eng. 2012, 25, 1336–1353. [Google Scholar] [CrossRef]
  13. Wang, S.; Deng, C.; Lin, W.; Huang, G.B.; Zhao, B. NMF-based image quality assessment using extreme learning machine. IEEE Trans. Cybern. 2016, 47, 232–243. [Google Scholar] [CrossRef]
  14. You, M.; Wang, H.; Liu, Z.; Chen, C.; Liu, J.; Xu, X.H.; Qiu, Z.M. Novel feature extraction method for cough detection using NMF. IET Signal Process. 2017, 11, 515–520. [Google Scholar] [CrossRef]
  15. Berry, M.W.; Browne, M.; Langville, A.N.; Pauca, V.P.; Plemmons, R.J. Algorithms and applications for approximate nonnegative matrix factorization. Comput. Stat. Data Anal. 2007, 52, 155–173. [Google Scholar] [CrossRef] [Green Version]
  16. Feng, X.R.; Li, H.C.; Wang, R.; Du, Q.; Jia, X.; Plaza, A.J. Hyperspectral Unmixing Based on Nonnegative Matrix Factorization: A Comprehensive Review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 4414–4436. [Google Scholar] [CrossRef]
  17. He, C.; Fei, X.; Cheng, Q.; Li, H.; Hu, Z.; Tang, Y. A survey of community detection in complex networks using nonnegative matrix factorization. IEEE Trans. Comput. Soc. Syst. 2021, 9, 440–457. [Google Scholar] [CrossRef]
  18. Li, S.Z.; Hou, X.W.; Zhang, H.J.; Cheng, Q.S. Learning spatially localized, parts-based representation. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition CVPR, Kauai, HI, USA, 8–14 December 2001. [Google Scholar]
  19. Hoyer, P.O. Non-negative matrix factorization with sparseness constraints. J. Mach. Learn. Res. 2004, 5, 1457–1469. [Google Scholar]
  20. Wang, Y.; Jia, Y. Fisher non-negative matrix factorization for learning local features. In Proceedings of the Asian Conference on Computer Vision, Jeju, Korea, 27–30 January 2004; pp. 27–30. [Google Scholar]
  21. Cai, D.; He, X.; Han, J.; Huang, T.S. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1548–1560. [Google Scholar]
  22. Zeng, K.; Yu, J.; Li, C.; You, J.; Jin, T. Image clustering by hyper-graph regularized non-negative matrix factorization. Neurocomputing 2014, 138, 209–217. [Google Scholar] [CrossRef]
  23. Zhang, L.; Chen, Z.; Zheng, M.; He, X. Robust non-negative matrix factorization. Front. Electr. Electron. Eng. China 2011, 6, 192–200. [Google Scholar] [CrossRef]
  24. Kong, D.; Ding, C.; Huang, H. Robust nonnegative matrix factorization using l21-norm. In Proceedings of the 20th ACM International Conference on Information and Knowledge Management, Glasgow, UK, 24–28 October 2011; pp. 673–682. [Google Scholar]
  25. Lu, Y.; Lai, Z.; Li, X.; Zhang, D.; Wong, W.K.; Yuan, C. Learning parts-based and global representation for image classification. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 3345–3360. [Google Scholar] [CrossRef]
  26. Maihami, V.; Yaghmaee, F. A review on the application of structured sparse representation at image annotation. Artif. Intell. Rev. 2017, 48, 331–348. [Google Scholar] [CrossRef]
  27. Liu, G.; Lin, Z.; Yu, Y. Robust subspace segmentation by low-rank representation. In Proceedings of the ICML, Haifa, Israel, 21–24 June 2010. [Google Scholar]
  28. Wan, M.; Chen, X.; Zhan, T.; Yang, G.; Tan, H.; Zheng, H. Low-rank 2D local discriminant graph embedding for robust image feature extraction. Pattern Recognit. 2023, 133, 109034. [Google Scholar] [CrossRef]
  29. Wan, M.; Yao, Y.; Zhan, T.; Yang, G. Supervised low-rank embedded regression (SLRER) for robust subspace learning. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1917–1927. [Google Scholar] [CrossRef]
  30. Lu, Y.; Yuan, C.; Zhu, W.; Li, X. Structurally incoherent low-rank nonnegative matrix factorization for image classification. IEEE Trans. Image Process. 2018, 27, 5248–5260. [Google Scholar] [CrossRef] [PubMed]
  31. Li, X.; Cui, G.; Dong, Y. Graph regularized non-negative low-rank matrix factorization for image clustering. IEEE Trans. Cybern. 2016, 47, 3840–3853. [Google Scholar] [CrossRef]
  32. Lu, J.; Wang, H.; Zhou, J.; Chen, Y.; Lai, Z.; Hu, Q. Low-rank adaptive graph embedding for unsupervised feature extraction. Pattern Recognit. 2021, 113, 107758. [Google Scholar] [CrossRef]
  33. He, P.; Xu, X.; Ding, J.; Fan, B. Low-rank nonnegative matrix factorization on Stiefel manifold. Inf. Sci. 2020, 514, 131–148. [Google Scholar] [CrossRef]
  34. Wan, M.; Chen, X.; Zhan, T.; Xu, C.; Yang, G.; Zhou, H. Sparse fuzzy two-dimensional discriminant local preserving projection (SF2DDLPP) for robust image feature extraction. Inf. Sci. 2021, 563, 1–15. [Google Scholar] [CrossRef]
  35. Chen, X.; Wan, M.; Zheng, H.; Xu, C.; Sun, C.; Fan, Z. A New Bilinear Supervised Neighborhood Discrete Discriminant Hashing. Mathematics 2022, 10, 2110. [Google Scholar] [CrossRef]
  36. Wan, M.; Chen, X.; Zhao, C.; Zhan, T.; Yang, G. A new weakly supervised discrete discriminant hashing for robust data representation. Inf. Sci. 2022, 611, 335–348. [Google Scholar] [CrossRef]
  37. Zhang, T.; Fang, B.; Tang, Y.Y.; Shang, Z.; Xu, B. Generalized discriminant analysis: A matrix exponential approach. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2009, 40, 186–197. [Google Scholar] [CrossRef]
  38. Wang, S.J.; Chen, H.L.; Peng, X.J.; Zhou, C.G. Exponential locality preserving projections for small sample size problem. Neurocomputing 2011, 74, 3654–3662. [Google Scholar] [CrossRef]
  39. Lee, D.D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788–791. [Google Scholar] [CrossRef] [PubMed]
  40. DiCarlo, J.J.; Zoccolan, D.; Rust, N.C. How does the brain solve visual object recognition? Neuron 2012, 73, 415–434. [Google Scholar] [CrossRef] [Green Version]
  41. Wan, M.; Lai, Z.; Yang, G.; Yang, Z.; Zhang, F.; Zheng, H. Local graph embedding based on maximum margin criterion via fuzzy set. Fuzzy Sets Syst. 2017, 318, 120–131. [Google Scholar] [CrossRef] [Green Version]
  42. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. Found. Trends® Mach. Learn. 2011, 3, 1–122. [Google Scholar]
  43. Kuybeda, O.; Frank, G.A.; Bartesaghi, A.; Borgnia, M.; Subramaniam, S.; Sapiro, G. A collaborative framework for 3D alignment and classification of heterogeneous subvolumes in cryo-electron tomography. J. Struct. Biol. 2013, 181, 116–127. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Cai, J.F.; Candès, E.J.; Shen, Z. A singular value thresholding algorithm for matrix completion. SIAM J. Optim. 2010, 20, 1956–1982. [Google Scholar] [CrossRef]
  45. Zdunek, R.; Phan, A.H.; Cichocki, A. Image classification with nonnegative matrix factorization based on spectral projected gradient. In Artificial Neural Networks; Springer: Cham, Switzerland, 2015; pp. 31–50. [Google Scholar]
  46. Kim, J.; Park, H. Fast nonnegative matrix factorization: An active-set-like method and comparisons. SIAM J. Sci. Comput. 2011, 33, 3261–3281. [Google Scholar] [CrossRef] [Green Version]
  47. Teboulle, M.; Vaisbourd, Y. Novel proximal gradient methods for nonnegative matrix factorization with sparsity constraints. SIAM J. Imaging Sci. 2020, 13, 381–421. [Google Scholar] [CrossRef]
  48. Nene, S.A. Columbia Object Image Library(COIL-20); Tech. Rep. CUCS-006-96; Columbia University: New York, NY, USA, 1996. [Google Scholar]
  49. The PolyU Palmprint Database. Available online: http://www.comp.polyu.edu.hk/biometrics/ (accessed on 20 October 2022).
  50. Martinez, A.; Benavente, R. The AR Face Database; Tech. Rep. 24; AUTO University: Barcelona, Spain, 1998. [Google Scholar]
  51. Phillips, P.J.; Wechsler, H.; Huang, J.; Rauss, P.J. The FERET database and evaluation procedure for face-recognition algorithms. Image Vis. Comput. 1998, 16, 295–306. [Google Scholar] [CrossRef]
Figure 1. Framework of the EGNLRF. For simplicity, only two color-coded classes are chosen for illustration. The framework mainly consists of three important parts: low-rank, NMF and graph embedding. These three parts jointly optimize to learn an effective latent representation.
Figure 1. Framework of the EGNLRF. For simplicity, only two color-coded classes are chosen for illustration. The framework mainly consists of three important parts: low-rank, NMF and graph embedding. These three parts jointly optimize to learn an effective latent representation.
Mathematics 10 04314 g001
Figure 2. Convergence of EGNLMF iterations on public datasets FERET, AR, COIL20 and PolyU palmprint.
Figure 2. Convergence of EGNLMF iterations on public datasets FERET, AR, COIL20 and PolyU palmprint.
Mathematics 10 04314 g002
Figure 3. Some original images from the (a) FERET, (b) AR, (c) COIL20, (d) PolyU palmprint databases.
Figure 3. Some original images from the (a) FERET, (b) AR, (c) COIL20, (d) PolyU palmprint databases.
Mathematics 10 04314 g003
Figure 4. Performance concerning the COIL20 public database (a) α with β and λ fixed, (b) β with α and λ fixed, (c) λ with α and β fixed.
Figure 4. Performance concerning the COIL20 public database (a) α with β and λ fixed, (b) β with α and λ fixed, (c) λ with α and β fixed.
Mathematics 10 04314 g004
Figure 5. Some images from the (a) FERET (noisy density = 0.1), (b) FERET (noisy density = 0.5), (c) AR (noisy density = 0.1), (d) AR (noisy density = 0.5).
Figure 5. Some images from the (a) FERET (noisy density = 0.1), (b) FERET (noisy density = 0.5), (c) AR (noisy density = 0.1), (d) AR (noisy density = 0.5).
Mathematics 10 04314 g005
Figure 6. Performance of seven algorithms on FERET public database with different dimensions or numbers of training samples. (a,b) raw database, (c) with noisy density = 0.1, (d) with noisy density = 0.5.
Figure 6. Performance of seven algorithms on FERET public database with different dimensions or numbers of training samples. (a,b) raw database, (c) with noisy density = 0.1, (d) with noisy density = 0.5.
Mathematics 10 04314 g006
Figure 7. Performance of five algorithms on FERET public database with different numbers of training samples. We experimented with the nearest neighbor parameter from 2 to 10 and plotted a boxplot to show the sensitivity of various algorithms to the nearest neighbor.
Figure 7. Performance of five algorithms on FERET public database with different numbers of training samples. We experimented with the nearest neighbor parameter from 2 to 10 and plotted a boxplot to show the sensitivity of various algorithms to the nearest neighbor.
Mathematics 10 04314 g007
Figure 8. Some images from the (a) COIL20 (noisy density = 0.1), (b) COIL20 (noisy density = 0.5), (c) COIL20 (occlusion size = 10 × 10 ), (d) COIL20 (occlusion size = 30 × 30 ).
Figure 8. Some images from the (a) COIL20 (noisy density = 0.1), (b) COIL20 (noisy density = 0.5), (c) COIL20 (occlusion size = 10 × 10 ), (d) COIL20 (occlusion size = 30 × 30 ).
Mathematics 10 04314 g008
Figure 9. Some images from the (a) PolyU palmprint (noisy density = 0.1), (b) PolyU palmprint (noisy density = 0.2), (c) PolyU palmprint (occlusion size = 10 × 10 ), (d) PolyU palmprint (occlusion size = 15 × 15 ).
Figure 9. Some images from the (a) PolyU palmprint (noisy density = 0.1), (b) PolyU palmprint (noisy density = 0.2), (c) PolyU palmprint (occlusion size = 10 × 10 ), (d) PolyU palmprint (occlusion size = 15 × 15 ).
Mathematics 10 04314 g009
Figure 10. Performance of seven algorithms on COIL20 public database with different dimensions or numbers of training samples. (a,d) raw database, (b) with noisy density = 0.1, (e) with noisy density = 0.5, (c) with occlusion size = 10 × 10 , (f) with occlusion size = 30 × 30 .
Figure 10. Performance of seven algorithms on COIL20 public database with different dimensions or numbers of training samples. (a,d) raw database, (b) with noisy density = 0.1, (e) with noisy density = 0.5, (c) with occlusion size = 10 × 10 , (f) with occlusion size = 30 × 30 .
Mathematics 10 04314 g010
Figure 11. Performance of seven algorithms on PolyU palmprint public database with different dimensions or numbers of training samples. (a,d) raw database, (b) with noisy density = 0.1, (e) with noisy density = 0.5, (c) with occlusion size = 10 × 10 , (f) with occlusion size = 15 × 15 .
Figure 11. Performance of seven algorithms on PolyU palmprint public database with different dimensions or numbers of training samples. (a,d) raw database, (b) with noisy density = 0.1, (e) with noisy density = 0.5, (c) with occlusion size = 10 × 10 , (f) with occlusion size = 15 × 15 .
Mathematics 10 04314 g011
Table 1. Objective functions and algorithm time complexity of seven algorithms.
Table 1. Objective functions and algorithm time complexity of seven algorithms.
AlgorithmObjective FunctionTime Complexity *
NMF X U V T F 2 O ( t m n k )
ELPP P T e x p X T L X P ** O ( m 2 n + m n 2 + n 3 )
LRNF A U V T F 2 + α A * + β E 1 O ( t ( m 3 + m n 2 ) )
GNMF X U V T F 2 + λ t r ( V T L V ) O ( t m n k + m n 2 )
LNMFS X U V T F 2 + α V F 2 + β t r ( V T L V ) O ( t ( m n k + n 2 k ) )
LRAGE i , j X i T X j T P R 2 2 W i j + α W F 2 + β P R 2 , 1 *** O ( t ( m 2 n + m n 2 + m 3 ) )
EGNLRF A U V T F 2 + α A * + β E 2 , 1 + λ t r e x p V T L V O ( t ( m n 2 + n 3 ) )
* m >> k, n >> k. ** P is a mapping transformation matrix of size m × k. *** R is a regression matrix of size k × m and W is the graph affinity matrix of size n × n.
Table 2. Recognition accuracy (%) of different algorithms on the AR dataset and with different dimensions.
Table 2. Recognition accuracy (%) of different algorithms on the AR dataset and with different dimensions.
Dimension102030405060708090100110120
NMF36.0239.2645.3746.3046.8550.0953.2456.1158.9857.0457.3158.98
ELPP15.8329.4737.7345.9149.9251.5953.3354.3256.3657.8857.7359.62
LRNF36.3051.0256.9463.5263.7067.1368.6171.4874.5471.3973.9872.59
GNMF59.1768.0669.4471.4872.2272.5973.8972.8773.8072.0474.8172.41
LNMFS61.6767.2270.2871.3070.7471.1173.7072.9673.0672.5974.5473.89
LRAGE44.2658.8063.4366.5768.8971.6771.5772.6975.0075.7475.9375.74
EGNLRF65.0069.0771.4874.4475.3777.0477.4178.7080.0080.1981.0282.04
Table 3. Recognition accuracy (%) of different algorithms on the AR dataset and with different density noisy.
Table 3. Recognition accuracy (%) of different algorithms on the AR dataset and with different density noisy.
AlgrithmsRaw DatabaseGaussion Noisy (Density = 0.1)Gaussion Noisy (Density = 0.5)
912159121591215
NMF47.7052.7463.7140.2946.7955.1533.1538.6445.64
ELPP46.2357.0264.5534.4139.8845.8342.1647.6254.17
LRNF58.8760.6075.9059.1761.3174.0948.4357.4464.62
GNMF62.6069.1774.8556.9667.2072.4835.0554.5267.65
LNMFS53.2472.3879.0946.6766.5576.2935.9859.8269.47
LRAGE61.6768.1075.6157.8965.8971.9757.9462.6269.70
EGNLRF69.6176.8587.8867.3574.9486.3654.7168.1576.21
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, G.; Zhang, L.; Wan, M. Exponential Graph Regularized Non-Negative Low-Rank Factorization for Robust Latent Representation. Mathematics 2022, 10, 4314. https://doi.org/10.3390/math10224314

AMA Style

Yang G, Zhang L, Wan M. Exponential Graph Regularized Non-Negative Low-Rank Factorization for Robust Latent Representation. Mathematics. 2022; 10(22):4314. https://doi.org/10.3390/math10224314

Chicago/Turabian Style

Yang, Guowei, Lin Zhang, and Minghua Wan. 2022. "Exponential Graph Regularized Non-Negative Low-Rank Factorization for Robust Latent Representation" Mathematics 10, no. 22: 4314. https://doi.org/10.3390/math10224314

APA Style

Yang, G., Zhang, L., & Wan, M. (2022). Exponential Graph Regularized Non-Negative Low-Rank Factorization for Robust Latent Representation. Mathematics, 10(22), 4314. https://doi.org/10.3390/math10224314

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop