Next Article in Journal
Mathematical Modelling of Diffusion Flows in Two-Phase Stratified Bodies with Randomly Disposed Layers of Stochastically Set Thickness
Previous Article in Journal
Recursive Convex Model for Optimal Power Flow Solution in Monopolar DC Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Low-Rank Structurally Incoherent Algorithm for Robust Image Feature Extraction

1
School of Computer Science and Engineering, Southeast University, Nanjing 211189, China
2
School of Engineering Audit, Nanjing Audit University, Nanjing 211815, China
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(19), 3648; https://doi.org/10.3390/math10193648
Submission received: 20 August 2022 / Revised: 22 September 2022 / Accepted: 30 September 2022 / Published: 5 October 2022

Abstract

:
In order to solve the problem in which structurally incoherent low-rank non-negative matrix decomposition (SILR-NMF) algorithms only consider the non-negativity of the data and do not consider the manifold distribution of high-dimensional space data, a new structurally incoherent low rank two-dimensional local discriminant graph embedding (SILR-2DLDGE) is proposed in this paper. The algorithm consists of the following three parts. Firstly, it is vital to keep the intrinsic relationship between data points. By the token, we introduced the graph embedding (GE) framework to preserve locality information. Secondly, the algorithm alleviates the impact of noise and corruption uses the L1 norm as a constraint by low-rank learning. Finally, the algorithm improves the discriminant ability by encrypting the structurally incoherent parts of the data. In the meantime, we capture the theoretical basis of the algorithm and analyze the computational cost and convergence. The experimental results and discussions on several image databases show that the proposed algorithm is more effective than the SILR-NMF algorithm.

1. Introduction

In the past few decades, “curse of dimensionality” [1] problems have been a hot topic for many researchers. A popular dimensionality reduction method, 2D Local Preserving Projections (2DLPP) [2,3,4], has been widely utilized in pattern recognition [5,6,7] and computer vision [8,9,10]. However, on the basis of the 2DLPP unsupervised algorithm, some improved 2DLPP supervised algorithms have also been proposed [11,12,13].
Unfortunately, these above algorithms are similar in that they all use the L2 norm, which will hinder the algorithms’ performance when dealing with noise or abnormal data. Therefore, in order to solve the problem in which the L2 norm is affected by outliers, many researchers use L1-norm-based algorithms as dimension reduction algorithms of distance criteria, which are widely considered to be effective methods [14,15,16,17,18]. For example, L1-PCA [14] and PCA-L1 [15], based on the PCA algorithm, solve the noise and outlier sensitivity in the data through an optimization problem. Finally, the Rotation Invariant L1-norm PCA (R1-PCA) [16] algorithm proposed on the basis of L1-PCA has some PCA properties, which has some PCA properties. In order to solve the general L1 norm local preservation problem, Ref. [17] proposed the LPP algorithm (LPP-L1) based on the PCA-L1 algorithm to more effectively preserve the spatial topological structure. At the same time, in order to solve the outliers and corrosion problems in LPP-L1, a 2DLPP algorithm based on LPP-L1 algorithm (2DLPP-L1) was proposed [18].
In recent years, compared with algorithms based on L1 norm, many feature extraction algorithms using low rank representation (LRR) are conducive to extracting clean information, which has proved to be crucially important when there is noise in the data. The robustness of these algorithms has attracted considerable attention from researchers [19,20,21,22,23,24]. For example, in order to better maintain the lowest rank representation and global structure of data, the LRR in reference [19,20] introduced the single subspace clustering problem into multiple subspace clustering. The subspace structure was recovered from the data damaged by noise or occlusion. On the basis of LRR, a robust PCA (RPCA) was proposed by introducing kernel norm [21,22]. On the basis of LRR, the manifold structure was introduced as the regularization term, and Laplace regularization LRR [23] was proposed and applied to clustering data. Based on the combination of the NNLRS-graph and LRR, the non-negative concept was introduced to propose a non-negative low-rank sparse graph (NNLRS) [24] to maintain the global structure in the data.
A new matrix factorization method with structural incoherence (SILR-NMF) [25] has been introduced into face recognition to improve the recognition performance in the presence of corrupted data. In Ref. [26], a new structurally incoherent method was proposed to improve the additional discrimination ability in face recognition by introducing the structural incoherence. Motivated by [25] and [26], we propose a new algorithm dubbed structural incoherent low rank two-dimensional local discriminant mapping embedding (SILR-2DLDGE), which is superior to the SILR-NMF algorithm who has important limitations. Since the SILR-NMF algorithm can not make full use of the neighborhood relationship between data points. SILR-2DLDGE is implemented in three steps. Firstly, the intra-class weighted matrix graph and the inter-class weighted matrix graph are constructed to maintain the discriminant information of local neighborhood. Secondly, low-rank learning is used to eliminate noise and damage in the data. Finally, the structural incoherencies [25,26] are combined to optimize the discriminant information of different classes.
The four main contributions of this paper are as follows:
  • We present a novel algorithm-based 2DLPP, which can simultaneously perform structurally incoherent, optimal graph Laplacian, and low-rank functions in a unified strategy. The algorithm has stronger discriminant ability than SILR-NMF, which fully reveals the structure information of the neighborhood to improve the discrimination ability.
  • We propose to introduce intraclass and interclass graphs into the structurally incoherent model to make the data points in the same class more compact and different class data points as far away as possible.
  • We used the low-rank feature to ensure that the given data are treated as two parts composed of the low-rank matrix and sparse matrix, representing useful features and nasty noise, respectively, so as to improve the performance of the algorithm using kernel norms as a measure of the regularization term.
  • We designed a practical and simple algorithm to cater to the optimization process, and verified the algorithm on five datasets. Actually, our proposed method can achieve better performance.
The rest of study was planned as follows: Section 2 mainly introduces the early algorithms such as LRR, SILR-NMF, and LRMD-SI. Section 3 introduces the details of proposed algorithm, and analyzes the computational complexity and convergence of the SILR-2DLDGE algorithm. In Section 4, we compare the proposed algorithm with other state-of-the-art algorithms, and verify the promising performance of the SILR-2DLDGE algorithm on five databases. The summary and future work can be seen in the last section of the paper.

2. Related Works

2.1. Notation

To facilitate the understanding of this article, we will introduce the related work on some early algorithms, i.e., LRR, LRMD-SI, and SILR-NMF algorithms.
Firstly, we consider N original space samples X i R m × n into feature space Y i R n × d . Then, the linear transformation shown in Equation (1) can be obtained:
Y i = P T X i , i ( 1 , 2 , , N )
where P R d × n is a transformation matrix, and n > d .
Secondly, we define the matrix Z = [ z i j ]     R m × n , and then Z i or Z . j is used to represent the i th or j th row of Z , respectively.
Finally, we define the matrix Z 1 = i , j | z i j | , Z F = Z i . 2 2 , and Z 2 , 1 = j Z . j 2 , respectively. Additionally, Z * represents the kernel norm which is used to compute the sum of the singular matrix.

2.2. LRR

The LRR [19] algorithm is different from sparse representation [27]. LRR adopts a joint approach to transfer the recovery of damaged data to multiple subspaces, which is an effective subspace segmentation algorithm. At the same time, the global structure of data is obtained by searching the lowest rank representation of all data.
Each data vector in X can be learned by the linear combination of the T :
X = T Z ,
where Z = [ Z 1 , Z 2 , Z n ] is the coefficient matrix being, representation X :
min Z Z * ,   s . t .   X = T Z ,
Here, * denotes the nuclear norm.
The noise E can be separated from corrupted data as follows [19]:
min Z Z * + λ E   2 , 1           s . t .   Z = X - T E ,
where λ is a parameter.

2.3. SILR-NMF

In view of the fact that face is often affected by noise, SILR-NMF [25] is a better choice for feature representation and image classification. We can obtain Equation (5) as the final form of its objective function:
    min Z , E , U 0 , V 0         i , j X U V F 2 + i j   α Z i * + β E i 1 + γ i j Z j T Z i F 2       s . t .   Z = X E  
where α , β , and γ are positive parameters.

2.4. LRMD-SI

In view of the fact that face is often affected by noise, Wei et al. [26] introduced a structurally incoherent constraint based on low-rank matrix decomposition (LRMD-SI), which is considered to have good performance in image representation and classification. We can obtain Equation (6) as the final form of its objective function:
min Z , E         i = 1 c { Z i * + λ E i 1 } + η i j Z j T Z i F 2         s . t .   X i = E i + Z i ,      
where matrix . F is the F-norm of the matrix, η is the parameter, and i ( 1 , 2 , , c . ) .

3. Structurally Incoherent Low-Rank 2D Local Discriminant Graph Embedding

This section discusses the objective function of the SILR-2DLDGE algorithm, and describes its optimization process, computational complexity, and convergence.

3.1. The Objective Function of SILR-2DLDGE

In low-rank matrix recovery, the training dataset X = [ X ˜ 1 , X ˜ 2 , , X ˜ C ] , which is divided into the low-rank matrix Z = [ Z ˜ 1 , Z ˜ 2 , , Z ˜ C ] and the noise matrix E = [ E ˜ 1 , E ˜ 2 , , E ˜ C ] , where X ˜ i = [ X ˜ 1 , X ˜ 2 , , X ˜ N ] ( i = 1 , 2 , , C ) ,   C is the number of class of training sample   X , and N is the total number of data in the i class of training sample   X . Then, we have the following:
min Z ˜ i , E ˜ i         i = 1 C { Z ˜ i * + λ E ˜ i 1 }         s . t .   X ˜ i = Z ˜ i + E ˜ i
where λ is a parameter.
The following equation can be obtained by encoding the 2DLDGE algorithm into Equation (10) and introducing an orthogonal constraint on   Y :
    min Z ˜ i , E ˜ i , P ˜ i         i , j Y i Y j 2 2 S i j w γ Y i Y j 2 2 S i j b + i = 1 C { α Z ˜ i * + β E ˜ i 1 } s . t .   Z ˜ i = B ˜ i ,     Y ˜ i = P ˜ i T Z ˜ i     a n d     X ˜ i = Z ˜ i + E ˜ i    
where the balance parameters α > 0 and β > 0 .
We define the scatter matrix   S i j w (intra-class) and the scatter matrix   S i j b (inter-class) in Equation (8):
S i j w = { 1 , i f   X i N K c + ( X j ) o r X j N K c + ( X i )   0 , o t h e r w i s e .
S i j b = { 1 , X i N K b + ( X j ) o r X j N K b + ( X i ) 0 , o t h e r w i s e .
where N K c + ( X i ) or N K c + ( X j ) in the same class means the index set of the Kc nearest neighbors of the sample X i or X j , respectively. Additionally, N K b + ( X i ) or N K b + ( X j ) in the different class express the index set of the Kb nearest neighbors of the sample X i or X j , respectively.
We added structural incoherence to Equation (8), which is expressed as follows to improve the discrimination ability and further distinguish the different classes of the Equation (7) algorithm:
    min Z ˜ i , E ˜ i , P ˜ i   i , j Y i Y j 2 2 S i j w γ Y i Y j 2 2 S i j b +   i = 1 C { α Z ˜ i * + β E ˜ i 1 } + η i j Z ˜ j T Z ˜ i F 2 s . t .   Z ˜ i = B ˜ i ,   Y ˜ i = P ˜ i T Z ˜ i     a n d     X ˜ i = Z ˜ i + E ˜ i    
In Equation (11), the algorithm improves the incoherence of algorithm structure by adding the F-norm between different pairs of   Z ˜ i and   Z ˜ j . The first term is to maintain the neighborhood information in the low dimensional subspace consistent with the clean data structure. The second term is used to learn a low rank matrix to ensure that its noise interference is reduced as much as possible. The last term has the discrimination ability and further distinguishes the different classes in the data.

3.2. The Optimization of SILR-2DLDGE

Due to the product term   Z ˜ i T   Z ˜ j , the minimization Equation (11) is non-convex. We iteratively replaced low-rank matrices   Z ˜ i to solve the following problems:
    min Z ˜ i , E ˜ i , P ˜ i   i , j Y i Y j 2 2 S i j w γ Y i Y j 2 2 S i j b + α Z ˜ i * +   β E ˜ i 1 + η i j Z ˜ j T Z ˜ i F 2 s . t .   Z ˜ i = B ˜ i ,   Y ˜ i = P ˜ i T Z ˜ i     a n d     X ˜ i = Z ˜ i + E ˜ i    
By introducing the auxiliary variable   B ˜ i to solve Equation (12), the following equation can be obtained:
    min Z ˜ i , E ˜ i , P ˜ i   i , j Y i Y j 2 2 S i j w γ Y i Y j 2 2 S i j b + α Z ˜ i * + β E ˜ i 1 + η i j Z ˜ j T B ˜ i F 2 s . t .   Z ˜ i = B ˜ i ,   Y ˜ i = P ˜ i T Z ˜ i     a n d     X ˜ i = Z ˜ i + E ˜ i    
Augmented Lagrange multipliers or ADMM [28] and other techniques can be used to solve Equation (13). We solved Equation (13) by performing singular value decomposition (SVD) at each iteration step, but it may have high computational complexity. We used the nuclear norm property [29] to reduce the complexity of Equation (13) as follows [30]:
Z ˜ i * = min R . H 1 2 ( H F 2 + R F 2 ) ,     s . t .   Z ˜ i = R H  
where Z ˜ i R a × b N is an arbitrary matrix. If r   min ( a , b N ) of A ˜ i , the factorization   Z ˜ i = R a × r H r × b N is obtained by the minimum solution of the above equation.
We can write the equivalent of Equation (13) using the conclusion of Equation (14) as follows:
    min Z ˜ i , E ˜ i , P ˜   i , j Y i Y j 2 2 S i j w γ Y i Y j 2 2 S i j b + α 2 ( R F 2 + H F 2 ) + β E ˜ i 1 + η i j Z ˜ j T B ˜ i F 2 s . t .   Z ˜ i = B ˜ i ,   Y ˜ i = P ˜ i T Z ˜ i     a n d     X ˜ i = Z ˜ i + E ˜ i    
The augmented Lagrange function of Equation (15) is detailed as follows:
L ( Z ˜ i , B ˜ i , μ , E ˜ i , P ˜ i , R , H , M 1 , M 2 , M 3 )       =   i , j Y i Y j 2 2 S i j w γ Y i Y j 2 2 S i j b + α 2 ( R F 2 + H F 2 ) + β E ˜ i 1 + η i j Z ˜ j T B ˜ i F 2     + t r ( M 1 T ( X ˜ i Z ˜ i E ˜ i ) )           + t r ( M 2 T ( Z ˜ i B ˜ i ) ) + t r ( M 3 T ( Z ˜ i R H ) )     + μ 2 ( X ˜ i Z ˜ i E ˜ i F 2 + Z ˜ i B ˜ i F 2 + Z ˜ i R H F 2 )       = i , j Y i Y j 2 2 S i j w γ Y i Y j 2 2 S i j b + α 2 ( R F 2 + H F 2 ) + β E ˜ i 1 + η i j Z ˜ j T B ˜ i F 2 + μ 2 X ˜ i Z ˜ i E ˜ i + M 1 μ F 2           + μ 2 Z ˜ i B ˜ i + M 2 μ F 2   + μ 2 Z ˜ i R H + M 3 μ F 2 1 2 μ ( M 1 F 2 + M 2 F 2 + M 3 F 2 )       = 2 P ˜ i T Z ˜ i ( L w I ) Z ˜ i T P ˜ i 2 P ˜ i T Z ˜ i ( L b I ) Z ˜ i T P ˜ i + α 2 ( R F 2 + H F 2 ) + β E ˜ i 1 + η i j Z ˜ j T B ˜ i F 2           + μ 2 X ˜ i Z ˜ i E ˜ i + M 1 μ F 2         + μ 2 Z ˜ i B ˜ i + M 2 μ F 2   + μ 2 Z ˜ i R H + M 3 μ F 2 1 2 μ ( M 1 F 2 + M 2 F 2 + M 3 F 2 )
where the penalty parameter is   μ > 0 , and the Lagrange multipliers are   M 1 ,   M 2 , and   M 3 .
Next, we separated the solutions for each variable.
(1)
Fix Z ˜ i , E ˜ i , R , H , and P , and update B ˜ i : since Z ˜ i , E ˜ i , R , H , and P are fixed, we can obtain the following equation from Equation (16):
μ 2 Z ˜ i B ˜ i + M 2 μ F 2 + η i j Z ˜ j T B ˜ i F 2
Letting the partial derivative of variable B be zero in Equation (17):
μ ( B ˜ i Z ˜ i M 2 μ ) + 2 η i j Z ˜ j Z ˜ j T B ˜ i = 0
Then, we have the following:
B ˜ i = ( 2 η j i Z ˜ j Z ˜ j T + μ I ) 1 ( μ Z ˜ i + M 2 )
(2)
Fix B ˜ i , E ˜ i , R , H , and P , and update Z ˜ i : by fixing other variables, we can obtain the following equation from Equation (16):
2 P ˜ i T Z ˜ i ( L w I ) Z ˜ i T P ˜ i 2 P ˜ i T Z ˜ i ( L b I ) Z ˜ i T P ˜ i + X ˜ i Z ˜ i E ˜ i + M 1 μ F 2 + Z ˜ i B ˜ i + M 2 μ F 2 + Z ˜ i R H + M 3 μ F 2
In Equation (20), letting the partial derivative of   Z ˜ i be zero, we have:
  4 P ˜ i P ˜ i T Z ˜ i ( ( L w L b ) I ) + 3 Z ˜ i ( Q ˜ 1 + Q ˜ 2 + Q ˜ 3 ) = 0
where Q ˜ 3 = R H M 3 μ , Q ˜ 2 = B ˜ i M 2 μ and Q ˜ 1 = X ˜ i E ˜ i + M 1 μ .
(3)
Fix Z ˜ i , B ˜ i , R , H , and P , and update E ˜ i : to fix other variables, the optimal E ˜ i can be obtained in Equation (16); thus, we have:
β E ˜ i 1 + μ 2 X ˜ i Z ˜ i E ˜ i + M 1 μ F 2
To simplify the solution of the above problem, we introduce a shrinkage operator according to reference [31]. In addition, S ε [ X ] = max ( | X | ε , 0 ) . sgn ( X ) represents the soft threshold to achieve the above purposes [32]. Finally, it is converted into Equation (23):
E ˜ i = S β μ ( X ˜ i Z ˜ i + M 1 μ )
(4)
Fix Z ˜ i , B ˜ i , E ˜ i , H , and P , and update R : we obtain the following Equation (24) by fixing all variables except   R in Equation (16):
1 2 ( α R F 2 + μ A ˜ i R H + M 3 μ F 2 )
In Equation (24), letting the partial derivative of variable R be zero, we obtain:
μ R H H T + α R ( M 3 + μ Z ˜ i ) H T = 0 R = ( M 3 + μ Z ˜ i ) H ( μ R H H T + α R ) 1
(5)
Fix Z ˜ i , B ˜ i , E ˜ i , R , and P , and update H : we fix all other variables in Equation (16) except H ; thus, the following equation can be obtained:
1 2 ( α H F 2 + μ Z ˜ i R H + M 3 μ F 2 )
The following equation is obtained to use the partial derivative of Equation (26):
μ R T R H + α H ( μ Z ˜ i + M 3 ) H T = 0 H = ( α I + μ R T R ) 1 R T ( μ Z ˜ i + M 3 )
(6)
Fix Z ˜ i , B ˜ i , E ˜ i , R , and H , and update P : we simplify Equation (16) and obtain the solution of variable P :
Ψ   = arg   min P ˜ i   2 P ˜ i T Z ˜ ( L w I ) Z ˜ T P ˜ i 2 P ˜ i T Z ˜ ( L b I ) Z ˜ T P ˜ i
Equation   ( 28 )   will   be   converted   as   follows : 2 P ˜ i T Z ˜ ( ( L w L b ) I ) Z ˜ T P ˜ i =   2 P ˜ i T ( X ˜ E ˜ ) ( ( L w L b ) I ) ( X ˜ E ˜ ) T P ˜ i
In order to solve the equation better, we add a constraint adjustment; the equation is as follows:
  P ˜ i T ( D w D b ) ( X ˜ E ˜ ) T P ˜ i =       1
Equations (29) and (30) can be transformed into the following formula:
arg min p T ( X E ) ( D w D b ) ( X E ) T p = 1     P ˜ T ( X E ) ( ( L w L b ) I ) ( X E ) T P ˜
The solution of Equation (31) is the same as solving the generalized eigenvalue equation as follows:
( X E ) ( ( L w L b ) I ) ( X E ) T p = λ ( X E ) ( D w D b ) ( X E ) T p
Algorithm 1 gives the concrete steps of SILR-2DLDGE.
Algorithm 1 SILR-2DLDGE
Input: Parameters α , β , γ and η ; and Training set X in Equation (14).
Initialization: E ˜ i = 0, B ˜ i = 0, ρ > 0, M 1 = 0, M 2 = 0, M 3 = 0.
Repeat
Step 1. Fixing Z ˜ i , E ˜ i , R , H and P , update B ˜ i with Equation (17);
Step 2. Fixing B ˜ i , E ˜ i , R , H and P , update Z ˜ i with Equation (20);
Step 3. Fixing Z ˜ i , B ˜ i , R , H and P , update E ˜ i with Equation (22);
Step 4. Fixing Z ˜ i , B ˜ i , E ˜ i , H and P , update R with Equation (24);
Step 5. Fixing Z ˜ i , B ˜ i , E ˜ i , R and P , update H with Equation (26);
Step 6. Fixing Z ˜ i , B ˜ i , E ˜ i , R and H , update P with Equation (28);
Step 7. Update parameter M 1 , M 2 , and M 3 as follows:
Update M 1 through M 1 = μ ( X ˜ i A ˜ i E ˜ i ) + M 1 ;
Update M 2 through M 2 = μ ( A ˜ i B ˜ i ) + M 2 ;
Update M 3 through M 3 = μ ( A ˜ i R H ) + M 3 ;
Step 8. Update μ through μ = min ( ρ μ , max μ ) ;
Step 9. Update t through t = t + 1;
Until Equation (11) is converged.
Step 10. Obtain the ( B ˜ i , A ˜ i , E ˜ i , R , H , P ) optimal solution;
Output: Obtain projection matrix P
The flow chart of the SILR-2DLDGE method is shown in Figure 1.

3.3. The Convergence Analysis

Firstly, we analyzed and proved the weak convergence of the proposed SILR-2DLDGE algorithm. In some cases, any limit point of the iterative sequence generated by the SILR-2DLDGE algorithm is a stationary point satisfying the Karush–Kuhn–Tucker (KKT) condition [33].
Assuming that the SILR-2DLDGE algorithm reaches a stationary point, any convergence point must satisfy the KKT condition, which is a necessary condition for a local optimal solution. Therefore, we can derive the KKT condition of Equation (11) as follows (note that Lagrange multiplier is not involved in the process of solving   P ; thus, we do not prove its KKT condition):
Z ˜ i R H = 0 ,         Z ˜ i B ˜ i = 0 , X ˜ i Z ˜ i E ˜ i = 0 L Z ˜ i =   4 p p T Z ˜ i ( ( L w L b ) I ) + 3 Z ˜ i ( Q ˜ 1 + Q ˜ 2 + Q ˜ 3 ) = 0 L B ˜ i =   μ ( B ˜ i Z ˜ i M 2 μ ) + 2 η j i Z ˜ j Z ˜ j T B ˜ i = 0 L R = α R ( μ Z ˜ i + M 3 ) H T + μ R H H T = 0 ( X E ) ( ( L w L b ) I ) ( X E ) T p = λ ( X E ) ( D w D b ) ( X E ) T p L H = α H ( μ Z ˜ i + M 3 ) H T + μ R T R H = 0 M 1 β E ˜ i E ˜ i 1
It can be deduced from the last relationship in Equation (33) that:
M 1 μ + X ˜ i Z ˜ i β E ˜ i   X ˜ i Z ˜ i 1 μ + X ˜ i Z ˜ i Λ β μ ( X ˜ i A ˜ i )
where the application element to X ˜ i Z ˜ i is the scalar function Λ β μ ( t ) = t + β μ | t | . From [34], the following relationships will be learned:
E ˜ i = Λ β μ 1 ( M 1 μ + X ˜ i Z ˜ i ) = S ( M 1 μ + X ˜ i Z ˜ i , β μ )
where S ( x , ε ) = max ( | x | ε , 0 ) . s i g n ( x ) . Thus, the following equation can be obtained with the KKT condition:
Z ˜ i R H = 0 ,         Z ˜ i B ˜ i = 0 , X ˜ i Z ˜ i E ˜ i = 0 L Z ˜ i =   4 p p T Z ˜ i ( ( L w L b ) I ) + 3 Z ˜ i ( Q ˜ 1 + Q ˜ 2 + Q ˜ 3 ) = 0 L B ˜ i =   μ ( B ˜ i Z ˜ i M 2 μ ) + 2 η j i Z ˜ j Z ˜ j T B ˜ i = 0 L R = α R + μ R H H T ( μ Z ˜ i + M 3 ) H T = 0 L H = α H + μ R T R H ( μ Z ˜ i + M 3 ) H T = 0 E ˜ i = S ( X ˜ i Z ˜ i + M 1 μ , β μ )
On the basis of the above conditions, the point convergence satisfying the KKT condition is proved.
Theorem 1 [25].
Suppose that  { θ j } j = 1 is bounded to be generated by the SILR-2DLDGE algorithm and let θ ( Z ˜ i , B ˜ i , H , R , E ˜ i , P ˜ i , M 1 , M 2 , M 3 ) . If lim j { θ j + 1 θ j } = 0 satisfies KKT conditions and converges, then it can be obtained that any point converges to the KKT point in { θ j } j = 1 .
Proof of Theorem 1.
M i + is the next point of M i in the sequence { M i j } j = 1 . Firstly, M 1 , M 2 , and M 3 can be obtained, as shown in the following equation:
M 1 + μ ( X ˜ i Z ˜ i E ˜ i ) + M 1 M 2 + μ ( Z ˜ i B ˜ i ) + M 2 M 3 + μ ( Z ˜ i R H ) + M 3
If ( X ˜ i Z ˜ i E ˜ i ) 0 , ( Z ˜ i B ˜ i ) 0 and ( Z ˜ i R H ) 0 , then we can obtain that ( M 1 + M 1 ) 0 , ( M 2 + M 2 ) 0 and ( M 3 + M 3 ) 0 . Thus, the sequence variables of { M 1 j } j = 1 , { M 2 j } j = 1 and { M 3 j } j = 1 converge to the stationary point, and we obtain the first three conditions in Equation (37).
Next, according to the SILR-2DLDGE algorithm and the fourth KKT condition, we obtain Equation (38):
( 2 η j i Z ˜ j Z ˜ ij T + μ I ) ( B ˜ i + B ˜ i ) =   ( μ Z ˜ i + M 2 ) ( 2 η j i Z ˜ j Z ˜ ij T + μ I ) B ˜ i
When ( B ˜ i + B ˜ i ) 0 , we can derive   M 2 2 η j i Z ˜ j Z ˜ j T i B ˜ i 0 from the second condition.
The following equation is obtained by the third KKT condition:
2 μ ( Z ˜ i + Z ˜ i ) = μ ( Q 1 + Q 2 + Q 3 ) 4 p p T Z ˜ i ( ( L w L b ) I ) 2 μ Z ˜ i
Thus, 4 p p T Z ˜ i ( ( L w L b ) I ) + 2 μ Z ˜ i μ ( Q 1 + Q 2 + Q 3 ) 0 when ( Z ˜ i + Z ˜ i ) 0 .
Similarly, the following equation can be obtained:
( R + R ) ( μ H H T + α I ) = μ ( Z ˜ i + M 2 μ ) H T R ( μ H H T + α I )
When ( R + R ) 0 , α R + μ R H H T μ ( Z ˜ i + M 2 μ ) H T 0 can be obtained. Similarly:
( α I + μ R T R ) ( H + H ) = μ R T ( Z ˜ i + M 2 μ ) ( α I + μ R R T ) H
When ( H + H ) 0 , we have α H + μ R T R H μ R T ( Z ˜ i + M 2 μ ) 0 .
From the last condition in Equation (35), the following equation is as follows:
E ˜ i + E ˜ i = S ( X ˜ i Z ˜ i + M 1 μ , β μ ) E ˜ i
We can obtain the last condition when E ˜ i + E ˜ i 0 .
Both sides of Equations (38)–(42) indicate tending to 0 when lim j { θ j + 1 θ j } = 0 . Finally, the sequence variable of { θ j } j = 1 can be obtained asymptotically to satisfy the KKT condition in Equation (11).
QED □

3.4. Computational Complexity

The main computation steps of SILR-2DLDGE algorithm are shown in steps 2 and 6. In the SILR-2DLDGE algorithm, the complexity of step 2 is O(t(2a2l)) to solve the Sylvester equation problem, and the complexity of step 6 is O(t(2a3)). Therefore, we can obtain the total complexity of the SILR-2DLDGE algorithm, which is O(t(2(a3 + a2l))), where t is the number of iterations.

4. Experiment Results

In this section, we introduce several representative experiments to verify the robustness of the proposed SILR-2DLDGE algorithm. We compared the results of the proposed algorithm with 2DLPP [2,3,4], LPP-L1 [17], 2DLP-L1 [18], LRR [19,20], SILR-NMF [25], LRMD-SI [26], and LPP [35,36] algorithms on FERET [37], ORL [38], coil100 [39], Yale [40], AR [41], and PolyU [42] databases, respectively.

4.1. Selection of Samples and Parameters

In each subsequent experiment, we will randomly select T = 5, 6, 10, 5, 6, 3 samples from each class in FERET, ORL, COIL 100, Yale, AR, and PolyU databases for training. In each running experiment, NN classifier was used for classification, and the experiment was repeated 10 times.
In all the following experiments, we used an iterative algorithm to derive solutions of the L1-norm algorithms and set the maximum value of iterations of the relevant iterative algorithm to 500. The LPP, 2DLPP, LPP-L1, and 2DLPP-L1 algorithms based on local graph embedding are obtained using k-nearest neighbors, where k = T − 1 (T is the training sets of each databases) is well collected in the observation space [43]. We choose the parameters of LRR, LRMD-SI, and SILR-NMF models in the references. From the objective function of the SILR-2DLDGE algorithm in Equation (13), the parameter range of α, β, and η were selected from [0.001, 0.01, 0.1, 1, 10, 100, 1000], and the γ parameter was selected from [0.1, 0.2, ……, 0.9, 1] to evaluate the effect, the values of parameters in our algorithm are given in Table 1.

4.2. Experiments on Occlusion Databases

To verify the robustness of the algorithm, we will randomly add 10 × 10 blocks to different positions of images to carry out continuous occlusion experiments in FERET, ORL, and COIL100 databases.
  • FERET database
The FERET database is mainly used to study changes in pose, illumination, and facial expression. There are 200 classes in the FERET database, and each class has seven images with a resolution of 40 × 40 pixels, resulting in a total of 1400 gray-scale images.
  • ORL database
The ORL database is mainly used to study changes with different expressions, postures, and illuminations. There are 40 classes in the ORL database., and each class has 10 images with a resolution of 56 × 46 pixels, resulting in a total of 400 gray-scale images.
  • COIL 100 database
The COIL 100 object database is mainly used to study changes with different illuminations. There are 100 subjects, and each subject has 72 images with a resolution of 32 × 32 pixels, resulting in a total of 7200 gray-scale images. Figure 2 shows some occlusion images on the three different databases.
To verify the effectiveness of continuous occlusion in FERET, ORL, and COIL100 databases, respectively, we randomly added 10 × 10 blocks to different positions of images. Figure 3, Figure 4 and Figure 5 show the average recognition rates (%) of different dimension changes on the three different databases.
We ran a set of experimental results and compared them with seven methods; the results are presented in this section. According to the settings in Section 4.1, when each class in FERET, ORL, and COIL100 databases randomly selects T sample points to form a training sample set, the optimal average recognition rate (%) of the seven algorithms is shown in Table 2.

4.3. Experiments on Noise Databases

Experiments were carried out on different levels of random pixel corruptions to verify the robustness of the algorithm. “Salt & pepper” noise with a density of 0.1 was added to the Yale, AR, and PolyU databases for corrosion experiments.
  • Description of the Yale database
The Yale face database is mainly used to study changes in facial expressions and lighting conditions. There are 15 classes in the Yale database, and each class has 11 images with a resolution of 50 × 40 pixels, resulting in a total of 165 gray-scale images.
  • Description of the AR database
The AR database is mainly used to study the changes in lighting conditions, facial expressions, and occlusion. There are 70 men and 56 women with a total of 126 people in the AR database, including 4000 color images with a resolution of 50 × 40 pixels.
  • Description of the PolyU palmprint database
The PolyU database is mainly used to study image changes in two periods. There are 100 different palms in the PolyU database, and each palm has six samples, resulting in a total of 600 gray images with a resolution of 64 × 64 pixels. Figure 6 shows the original images and some corrosion images on the three different databases.
For continuous corrosion experiments, we randomly added a density of 0.1 of “salt & pepper” noise to the Yale, AR, and PolyU databases. Then, we ran a set of experiments and compared them with seven methods to evaluate the proposed SILR-2DLDGE algorithm. Figure 7, Figure 8 and Figure 9 show the average recognition rates (%) of different dimension changes on the three different databases.
As in the last experiment, we can obtain the best average recognition rates (%) of the eight algorithms according to the settings in Section 4.1, as shown in Table 3.

4.4. Convergence Study

In the last experiment, we will further prove the convergence of the SILR-2DLDGE algorithm on the FERET and AR databases, respectively. The objective function values decreased monotonically; Figure 10 shows that the proposed algorithms are convergent due to the increase in the number of iterations.
In addition, to further verify the effectiveness of our proposed algorithm, we took the first three training samples from each class on the PolyU palmprint database; Table 4 shows the CPU time of each method.

4.5. Observations and Discussion

The four contributions of the SILR-2DLDGE algorithm are as follows:
(1)
In Table 2 and Table 3, the maximum average recognition rate of the SILR-2DLDGE algorithm on different databases is the highest, which fully shows that our algorithm is optimal and robust. The key reason is that SILR-2DLDGE not only learns a base matrix with low-rank property and local discriminant ability, but also has the advantages of SILR-NMF, which can weaken the disturbance of noise.
(2)
The SILR-2DLDGE algorithm has the advantages of low-rank learning, sparse learning, and incoherent structure learning, similarly to the SILR-NMF algorithm, as well as the advantages of graph embedding, which shows that more discrimination information can be obtained by sparse learning combined with the L1 norm and L * norm. We can see from the curve changes in Figure 3, Figure 4, Figure 5, Figure 7, Figure 8 and Figure 9 that the average recognition rates of this algorithm in three noisy databases and three occlusion databases are higher than that of other algorithms, which fully show that our algorithm has stronger robustness than other algorithms.
(3)
The SILR-2DLDGE algorithm takes less time than the others (Table 4) and learns the sparse transformation matrix to encode the geometric structure of the data and effectively improve the classification accuracy, which can obtain clean data in cases of noise disturbance. At the same time, structural incoherence can ensure that it is easier to separate data points of different classes.
(4)
The maximum average recognition rate of the SILR-2DLDGE algorithm varies among different databases. For example, the SILR-2DLDGE algorithm has the highest average maximum recognition rate without occlusion on the ORL database, whereas it has the lowest average maximum recognition rate with an occlusion of 10 × 10 on the COIL20 database. At the same time, the SILR-2DLDGE algorithm has the highest average maximum recognition rate without noise on the YALE database, whereas it has the lowest average maximum recognition rate with added “salt and pepper” noise with a density of 0.1 on the AR database. The above reasons are due to the small number of ORL and YALE samples, the large AR and PolyU palmprint databases, the presence of glasses and other occlusions in AR, and the presence of illumination and other effects in the PolyU palmprint.

5. Conclusions

This study mainly combined the 2DLPP algorithm with low-rank representation learning and proposed a structurally incoherent low-rank two-dimensional local discriminant graph embedding (SILR-2DLDGE) algorithm based on subspace learning, graph embedding, low-rank sparsity, and structural incoherence. Identification information, local geometry information, low-rank representation information, and structural incoherence existed simultaneously in the proposed algorithm. In addition, the ultimate goal of the proposed algorithm is to make the data points as independent as possible from different classes. In particular, it used low-rank learning L1 norms as constraints to reduce the influence of noise and corruption. The noise and occlusion experiments on six public databases further proved that the proposed SILR-2DLDGE algorithm is more robust than other algorithms. The algorithm is sensitive to parameters. In the future, we will study various parameters and further improve the robustness of the algorithm.

Author Contributions

Conceptualization, H.G. and A.S.; methodology, H.G. and A.S.; software, H.G. and A.S.; validation, H.G. and A.S.; formal analysis, H.G.; investigation, H.G.; resources, H.G.; data curation, H.G.; writing—original draft preparation, H.G. and A.S.; writing—review and editing, H.G. and A.S.; visualization, H.G.; supervision, A.S.; project administration, H.G. and A.S.; funding acquisition, H.G. and A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (No. 71972102), the Research and Planning fund for Humanities and Social Sciences of the Ministry of Education (No. 19YJAZH100), and the Key University Science Research Project of Jiangsu Province (No. 20KJA520002).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses or interpretation of data; in the writing of the manuscript or in the decision to publish the results.

References

  1. Qian, J.; Zhu, S.; Wong, W.K.; Zhang, H.; Lai, Z.; Yang, J. Dual robust regression for pattern classifification. Inf. Sci. 2021, 546, 1014–1029. [Google Scholar] [CrossRef]
  2. Niu, B.; Yang, Q.; Shiu, S.C.K.; Pal, S.K. Two-dimensional Laplacianfaces algorithm for face recognition. Pattern Recognit. 2008, 41, 3237–3243. [Google Scholar] [CrossRef]
  3. Hu, D.; Feng, G.; Zhou, Z. Two-dimensional locality preserving projections (2DLPP) with its application to palmprint recognition. Pattern Recognit. 2007, 40, 339–342. [Google Scholar] [CrossRef]
  4. Wan, M.; Chen, X.; Zhan, T.; Xu, C.; Yang, G.; Zhou, H. Sparse Fuzzy Two-Dimensional Discriminant Local Preserving Projection (SF2DDLPP) for Robust Image Feature Extraction. Inf. Sci. 2021, 563, 1–15. [Google Scholar] [CrossRef]
  5. Zheng, H.; Wang, R.; Ji, W.; Zong, M.; Wong, W.K.; Lai, Z.; Lv, H. Discriminative deep multi-task learning for facial expression recognition. Inf. Sci. 2020, 533, 60–71. [Google Scholar] [CrossRef]
  6. Shen, F.; Zhou, X.; Yu, J.; Yang, Y.; Liu, L.; Shen, H.T. Scalable Zero-Shot Learning via Binary Visual-Semantic Embeddings. IEEE Trans. Image Process. 2019, 28, 3662–3674. [Google Scholar] [CrossRef] [PubMed]
  7. Zhao, C.; Wang, X.; Zuo, W.; Shen, F.; Shao, L.; Miao, D. Similarity learning with joint transfer constraints for person re-identification. Pattern Recognit. 2020, 97, 107014. [Google Scholar] [CrossRef]
  8. Zhou, J.; Lai, Z.; Miao, D.; Gao, C.; Yue, X. Multigranulation rough-fuzzy clustering based on shadowed sets. Inf. Sci. 2020, 507, 553–573. [Google Scholar] [CrossRef]
  9. Zhang, Z.; Zhang, Y.; Liu, G.; Tang, J.; Yan, S.; Wang, M. Joint Label Prediction Based Semi-Supervised Adaptive Concept Factorization for Robust Data Representation. IEEE Trans. Knowl. Data Eng. 2019, 32, 952–970. [Google Scholar] [CrossRef] [Green Version]
  10. Liu, Z.; Lai, Z.; Ou, W.; Zhang, K.; Zheng, R. Structured optimal graph based sparse feature extraction for semi-supervised learning. Signal Process 2020, 170, 107456. [Google Scholar] [CrossRef]
  11. Zhi, R.; Ruan, Q. Facial expression recognition based on two-dimensional discriminant locality preserving projections. Neurocomputing 2008, 71, 1730–1734. [Google Scholar] [CrossRef]
  12. Wan, M.; Lai, Z.; Shao, J.; Jin, Z. Two-dimensional local graph embedding discriminant analysis (2DLGEDA) with its application to face and Palm Biometrics. Neurocomputing 2009, 73, 197–203. [Google Scholar] [CrossRef]
  13. Jove, E.; Casteleiro-Roca, J.-L.; Quintián, H.; Méndez-Pérez, J.-A.; Calvo-Rolle, J.-L. A new method for anomaly detection based on non-convex boundaries with random two-dimensional projections. Inf. Fusion 2021, 65, 50–57. [Google Scholar] [CrossRef]
  14. Meng, D.; Zhao, Q.; Xu, Z. Improve robustness of sparse PCA by L1-norm maximization. Pattern Recognit. 2012, 45, 487–497. [Google Scholar] [CrossRef]
  15. Kwak, N. Principal Component Analysis Based on L1-Norm Maximization. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1672–1680. [Google Scholar] [CrossRef]
  16. Ding, C.; Zhou, D.; He, X.; Zha, H. R1-PCA: Rotational invariant L1-norm principal component analysis for robust subspace factorization. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 281–288. [Google Scholar]
  17. Pang, Y.; Yuan, Y. Outlier-resisting graph embedding. Neurocomputing 2010, 73, 968–974. [Google Scholar] [CrossRef]
  18. Zhao, H.X.; Xing, H.J.; Wang, X.Z.; Chen, J.F. L1-norm-based 2DLPP. In Proceedings of the 2011 Chinese Control and Decision Conference (CCDC), Mianyang, China, 23–25 May 2011; pp. 1259–1264. [Google Scholar]
  19. Liu, G.; Lin, Z.; Yu, Y. Robust subspace segmentation by low-rank representation. In Proceedings of the 27th International Conference on International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 663–670. [Google Scholar]
  20. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; Ma, Y. Robust Recovery of Subspace Structures by Low-Rank Representation. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 171–184. [Google Scholar] [CrossRef] [Green Version]
  21. Wright, J.; Ganesh, A.; Rao, S.; Peng, Y.; Ma, Y. Robust principal component analysis: Exact recovery of corrupted low-rank matrices via convex optimization. In Proceedings of the 22nd International Conference on Neural Information Processing Systems, Whistler, BC, Canada, 7–10 December 2009; pp. 1–9. [Google Scholar]
  22. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis. J. ACM 2009, 58, 1–17. [Google Scholar] [CrossRef]
  23. Liu, J.; Chen, Y.; Zhang, J.; Xu, Z. Enhancing Low-Rank Subspace Clustering by Manifold Regularization. IEEE Trans. Image Process. 2014, 23, 4022–4030. [Google Scholar] [CrossRef]
  24. Zhuang, L.; Gao, H.; Lin, Z.; Ma, Y.; Zhang, X.; Yu, N. Non-negative low-rank and sparse graph for semi-supervised learning. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2328–2335. [Google Scholar]
  25. Lu, Y.; Yuan, C.; Zhu, W.; Li, X. Structurally Incoherent Low-Rank Nonnegative Matrix Factorization for Image Classification. IEEE Trans. Image Process. 2018, 27, 5248–5260. [Google Scholar] [CrossRef]
  26. Wei, C.-P.; Chen, C.-F.; Wang, Y.-C.F. Robust face recognition with structurally incoherent low-rank matrix decomposition. IEEE Trans. Image Process. 2014, 23, 3294–3307. [Google Scholar] [CrossRef] [Green Version]
  27. Wright, J.; Yang, A.Y.; Ganesh, A.; Sastry, S.S.; Ma, Y. Robust face recognition via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 210–227. [Google Scholar] [CrossRef] [Green Version]
  28. Lin, Z.; Chen, M.; Ma, Y. The augmented Lagrange multiplier algorithm for exact recovery of corrupted low-rank matrices. arXiv 2009, arXiv:1009.5055. [Google Scholar]
  29. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed optimization and statistical learning via the alternating direction algorithm of multipliers. Found. Trends Mach. Learn. 2011, 3, 1–122. [Google Scholar] [CrossRef]
  30. Mazumder, R.; Hastie, T.; Tibshirani, R. Spectral Regularization Algorithms for Learning Large Incomplete Matrices. J. Mach. Learn. Res. 2010, 11, 2287–2322. [Google Scholar]
  31. Drineas, P.; Frieze, A.; Kannan, R.; Vempala, S.; Vinay, V. Clustering Large Graphs via the Singular Value Decomposition. Mach. Learn. 2004, 56, 9–33. [Google Scholar] [CrossRef] [Green Version]
  32. Hosseini-Asl, E.; Zurada, J.M.; Nasraoui, O. Deep Learning of Part-Based Representation of Data Using Sparse Autoencoders With Nonnegativity Constraints. IEEE Trans. Neural Netw. Learn. Syst. 2015, 27, 2486–2498. [Google Scholar] [CrossRef] [Green Version]
  33. Boyd, S.P.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  34. Kim, E.; Lee, M.; Oh, S. Elastic-net regularization of singular values for robust subspace learning. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 915–923. [Google Scholar]
  35. He, X.; Niyogi, P. Locality Preserving Projections. In Proceedings of the Advances in Neural Information Processing System 16, Vancouver, BC, Canada, 8–13 December 2003. [Google Scholar]
  36. He, X.; Yan, S.; Hu, Y.; Niyogi, P.; Zhang, H.J. Face recognition using laplacianfaces. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 328–340. [Google Scholar]
  37. Phillips, P.J.; Moon, H.; Rizvi, S.A.; Rauss, P.J. The FERET evaluation algorithmology for face-recognition algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1090–1104. [Google Scholar] [CrossRef]
  38. Samaria, F.S.; Harter, A.C. Parameterisation of a stochastic model for human face identifification. In Proceedings of the 1994 IEEE Workshop on Applications of Computer Vision, Sarasota, FL, USA, 5–7 December 1994; pp. 138–142. [Google Scholar]
  39. Nene, S.A.; Nayar, S.K.; Murase, H. Columbia Object Image Library (COIL-100); Columbia University: New York, NY, USA, 1996. [Google Scholar]
  40. Yang, J.; Zhang, D.; Frangi, A.F.; Yang, J.Y. Two-dimensional PCA: A new approach to appearance-based face representation and recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 131–137. [Google Scholar] [CrossRef] [Green Version]
  41. Martinez, A.M.; Benavente, R. The AR Face Database; Centrede Visió per Computador, Univ. Autonoma de Barcelona: Bellaterra, Barcelona, 1998. [Google Scholar]
  42. Liu, Z.; Ou, W.; Lu, W.; Wang, L. Discriminative feature extraction based on sparse and low-rank representation. Neurocomputing 2019, 362, 129–138. [Google Scholar] [CrossRef]
  43. Yang, J.; Zhang, D.; Yang, J.; Niu, B. Globally maximizing, locally minimizing: Unsupervised discriminant projection with applications to face and palm biometrics. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 650–664. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The flow chart of the SILR-2DLDGE method.
Figure 1. The flow chart of the SILR-2DLDGE method.
Mathematics 10 03648 g001
Figure 2. Some occlusion images from the three different databases. (a) FERET, (b) ORL, and (c) COIL 100.
Figure 2. Some occlusion images from the three different databases. (a) FERET, (b) ORL, and (c) COIL 100.
Mathematics 10 03648 g002
Figure 3. The average recognition rates (%) of the eight algorithms in the FERET database vary with the dimension.
Figure 3. The average recognition rates (%) of the eight algorithms in the FERET database vary with the dimension.
Mathematics 10 03648 g003
Figure 4. The average recognition rates (%) of the eight algorithms in the ORL database varies with the dimension.
Figure 4. The average recognition rates (%) of the eight algorithms in the ORL database varies with the dimension.
Mathematics 10 03648 g004
Figure 5. The average recognition rates (%) of the eight algorithms in COIL20 database vary with the dimension.
Figure 5. The average recognition rates (%) of the eight algorithms in COIL20 database vary with the dimension.
Mathematics 10 03648 g005
Figure 6. Some the original images and corrosion images on the three different databases. (a) Yale, (b) AR, and (c) PolyU.
Figure 6. Some the original images and corrosion images on the three different databases. (a) Yale, (b) AR, and (c) PolyU.
Mathematics 10 03648 g006
Figure 7. The average recognition rates (%) of the eight algorithms in the Yale database vary with the dimension.
Figure 7. The average recognition rates (%) of the eight algorithms in the Yale database vary with the dimension.
Mathematics 10 03648 g007
Figure 8. The average recognition rates (%) of the eight algorithms in the AR database vary with the dimension.
Figure 8. The average recognition rates (%) of the eight algorithms in the AR database vary with the dimension.
Mathematics 10 03648 g008
Figure 9. The average recognition rates (%) of the eight algorithms in the PolyU palmprint database vary with the dimension.
Figure 9. The average recognition rates (%) of the eight algorithms in the PolyU palmprint database vary with the dimension.
Mathematics 10 03648 g009
Figure 10. Convergence of the proposed algorithm. (a) FERET databases; (b) AR databases.
Figure 10. Convergence of the proposed algorithm. (a) FERET databases; (b) AR databases.
Mathematics 10 03648 g010
Table 1. Parameter values of the SILR-2DLDGE algorithm on different datasets.
Table 1. Parameter values of the SILR-2DLDGE algorithm on different datasets.
DatasetParameters of SILR-2DLDGE
αβγη
FERET0.0010.10.30.01
ORL0.010.10.20.001
COIL 1000.1100.5100
Yale0.0010.010.30.01
AR0.0010.10.80.01
PolyU0.010.10.610
Table 2. The best average recognition rates (%) of the eight algorithms in FERET, ORL, and COIL100 databases and the corresponding dimensions (D).
Table 2. The best average recognition rates (%) of the eight algorithms in FERET, ORL, and COIL100 databases and the corresponding dimensions (D).
AlgorithmDatabase
Without OcclusionOcclusion Size = 10 × 10
FERETORLCOIL20FERETORLCOIL20
LPP87.1689.5086.6984.3587.3685.90
25(D)18(D)20(D)26(D)20(D)24(D)
2DLPP89.2893.1188.3586.7690.5586.86
40 × 16(D)50 × 14(D)32 × 20(D)40 × 18(D)50 × 16(D)32 × 22(D)
LPP-L184.2586.3383.4582.1885.8781.65
28(D)22(D)26(D)30(D)24(D)28(D)
2DLPP-L185.4288.3684.3583.8287.5882.36
40 × 12(D)50 × 14(D)32 × 18(D)40 × 16(D)50 ×16(D)32 × 20(D)
LRR82.6586.8581.2080.6385.2579.95
32(D)16(D)28(D)36(D)18(D)32(D)
LRMD-SI88.7091.3586.8086.7889.8084.95
16(D)18(D)16(D)18(D)18(D)20(D)
SILR-NMF92.5293.9590.6590.2391.6888.28
418(D)26(D)20(D)22(D)30(D)20(D)
SILR-2DLDGE94.7596.6292.5491.6194.8690.58
40 × 18(D)50 × 16(D)32 × 20(D)40 × 16(D)50 × 20(D)32 × 18(D)
D: dimension.
Table 3. The best average recognition rates (%) of different algorithms and corresponding dimensions (D) on the Yale, AR, and PolyU databases and the corresponding dimensions.
Table 3. The best average recognition rates (%) of different algorithms and corresponding dimensions (D) on the Yale, AR, and PolyU databases and the corresponding dimensions.
AlgorithmDatabase
Without NoiseNoise Density = 0.1
YaleARPolyUYaleARPolyU
LPP90.3188.9389.0889.6585.4387.75
32(D)40(D)36(D)36(D)40(D)40(D)
2DLPP91.1890.3591.8890.5687.3988.86
56 × 26(D)50 × 24(D)64 × 30(D)56 × 30(D)50 × 28(D)64 × 36(D)
LPP-L191.8789.3690.5689.0886.4588.82
36(D)45(D)38(D)32(D)45(D)36(D)
2DLPP-L192.9890.1691.0890.6688.4589.56
56 × 20(D)50 × 18(D)64 × 24(D)56 × 26(D)50 × 22(D)64 × 28(D)
LRR87.6582.9683.2585.3680.4082.82
36(D)18(D)16(D)40(D)20(D)20(D)
LRMD-SI91.8086.4588.2389.0883.8285.66
20(D)20(D)18(D)22(D)28(D)26(D)
SILR-NMF92.5190.6991.4890.6587.0289.54
12(D)16(D)10(D)14(D)14(D)12(D)
SILR-2DLDGE95.2593.6094.7893.6590.2892.15
56 × 12(D)50 × 14(D)64 × 10(D)56 × 16(D)50 × 12(D)64 × 16(D)
D: dimension.
Table 4. The average CPU time consumed of the eight algorithms in the PolyU palmprint database.
Table 4. The average CPU time consumed of the eight algorithms in the PolyU palmprint database.
PolyU
ALGORITHM
LPP2DLPP LPP -   L 1 2 DLPP -   L 1 LRRLRMD-SISILR-NMFSILR-2DLDGE
Noise density = 0.1(S)0.9810.8181.2360.8520.7650.7830.7060.652
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ge, H.; Song, A. A New Low-Rank Structurally Incoherent Algorithm for Robust Image Feature Extraction. Mathematics 2022, 10, 3648. https://doi.org/10.3390/math10193648

AMA Style

Ge H, Song A. A New Low-Rank Structurally Incoherent Algorithm for Robust Image Feature Extraction. Mathematics. 2022; 10(19):3648. https://doi.org/10.3390/math10193648

Chicago/Turabian Style

Ge, Hongmei, and Aibo Song. 2022. "A New Low-Rank Structurally Incoherent Algorithm for Robust Image Feature Extraction" Mathematics 10, no. 19: 3648. https://doi.org/10.3390/math10193648

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop