Next Article in Journal
A Stochastic Analysis of the Effect of Trading Parameters on the Stability of the Financial Markets Using a Bayesian Approach
Previous Article in Journal
A New Spike Membership Function for the Recognition and Processing of Spatiotemporal Spike Patterns: Syllable-Based Speech Recognition Application
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual Space Latent Representation Learning for Image Representation

1
School of Mathematics and Information Science, North Minzu University, Yinchuan 750030, China
2
School of Mathematics and Computer Application, Shangluo University, Shangluo 726000, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(11), 2526; https://doi.org/10.3390/math11112526
Submission received: 3 May 2023 / Revised: 23 May 2023 / Accepted: 28 May 2023 / Published: 31 May 2023
(This article belongs to the Section E1: Mathematics and Computer Science)

Abstract

:
Semi-supervised non-negative matrix factorization (NMF) has achieved successful results due to the significant ability of image recognition by a small quantity of labeled information. However, there still exist problems to be solved such as the interconnection information not being fully explored and the inevitable mixed noise in the data, which deteriorates the performance of these methods. To circumvent this problem, we propose a novel semi-supervised method named DLRGNMF. Firstly, dual latent space is characterized by the affinity matrix to explicitly reflect the interrelationship between data instances and feature variables, which can exploit the global interconnection information in dual space and reduce the adverse impacts caused by noise and redundant information. Secondly, we embed the manifold regularization mechanism in the dual graph to steadily retain the local manifold structure of dual space. Moreover, the sparsity and the biorthogonal condition are integrated to constrain matrix factorization, which can greatly improve the algorithm’s accuracy and robustness. Lastly, an effective alternating iterative updating method is proposed, and the model is optimized. Empirical evaluation on nine benchmark datasets demonstrates that DLRGNMF is more effective than competitive methods.

1. Introduction

Dimensionality reduction for image representation is a fundamental task in machine learning. Dimensionality reduction can not only shorten computation time and reduce storage space, but also find out latent and discriminative features by removing the noise and irrelevant features. Over the past decade, many dimensionality reduction methods have been presented such as locally linear embedding (LLE) [1], non-negative matrix factorization (NMF) [2,3], feature selection (FS) [4], and principal component analysis (PCA) [5]. In particular, due to its excellent low-dimensional learning ability, NMF has attracted a lot of attention and been successfully applied to image analysis [6], text classification [7], face recognition [8], and biometrics [9].
NMF is expressed as low-dimensional representations of high-dimensional data [10]. Since there is a constraint that these low-dimensional matrices are negative in NMF, the part-of-whole interpretation can be guaranteed and is consistent with human perception [11]. However, there are some limitations to NMF. Firstly, NMF ignores local data structure [12] since the data with a non-Gaussian distribution do not satisfy the constraint condition that the samples are only restricted to the number of classes in NMF. Secondly, recent studies show that sparse representation can improve the robustness and recognition of the algorithm to some extent, whereas NMF is not able to generate sparse representation. Furthermore, NMF is an unsupervised method that fails to investigate partial label information. However, in the real world, a small amount of label information is easily available. Lastly, in NMF there is an assumption that data instances are distributed and independent, which results in ignoring the interconnection of data instances in the real world. Nonetheless, whether data instances come from homologous sources or heterogeneous sources, there will be some interconnection between them [13], as well as between feature variables. For data instances, each data instance exhibits a certain degree of correlation that stronger correlation among samples may occur in the same class and weaker correlation among samples may exist in different classes. For example, as shown in Figure 1, Figure 1a–c represent three kinds of faces belonging to three different classes in the ORL dataset [14]. At first glance, there exists a common object, i.e., eyeglasses, in Figure 1a,b, whereas there is no common object in Figure 1b,c. Despite these groups of images belonging to different classes, the eyeglasses feature exhibits a moderate correlation to some extent within these two categories in Figure 1a,b. However, the correlation in the eyeglasses feature between Figure 1b and Figure 1c is low. By leveraging NMF to learn these interconnected relationships as prior knowledge and performing matrix factorization, the resulting low-dimensional representation will produce satisfactory outcomes within the constraints of this information. Furthermore, due to the diversity of samples, as in Figure 1a, the eyes are unable to act as a dominant feature for classification because the glasses play an important role to some extent. Hence, the interconnection of the eyes and the glasses can successfully distinguish the faces in Figure 1a,c.
To solve the above problems in NMF, various NMF variants have been proposed. To establish local data structure, i.e., the first limitation of NMF, Huang et al. [15] and Cai et al. [16] presented new NMF variants with a neighborhood graph to enable a compact representation of similar data points in the data space consistent with coefficient vectors. Different from the above two methods, Liu et al. [17] introduced the local coordinate constraint, which can enable the base vectors close to the samples. To sufficiently exploit the local manifold structure in data space and feature space, Shang et al. [18] constructed a dual graph model by encoding geometric information. Since the sparsity of NMF is not enough, Hoyer [19] introduced the sparsity of the coefficient matrix into NMF to explore parts-based representations and control the degree of sparseness explicitly. On the basis of the method in [19], Meng et al. [20] obtained the sparse basis matrix by imposing the ℓ2,1-norm while considering the local manifold structure in the dual space. To utilize partial label information, some researchers [21,22,23] constructed a stronger constraint and embed it into NMF to perform semi-supervised NMF. Different from previous semi-supervised NMF algorithms that shared the same coordinates for data points with the same labels, an additional constraint in the new representation was imposed to distribute the samples of the same class on the same axis, which could improve the robustness of clustering [24]. Integrating the merits of dual graph regularization, sparse regression, and semi-supervised, Meng et al. [25] imposed biorthogonal constraints to perform a semi-supervised non-negative matrix factorization. Thus, a unique basis vector corresponded to each image, which efficiently improved the discrimination ability of clusters and the exclusion between different classes. In fact, the interconnection information in NMF has not been adequately investigated. Recently, latent representation has been applied in the data space [26,27] or feature space [28] due to its good performance in exploiting the interconnection information. However, how to employ a fraction of label information to associate a small number of data instances and feature variables in latent representation space remains a challenging problem.
In summary, there are still some problems to be solved in NMF. These methods related to NMF are listed in terms of their characteristics in Table 1 to enhance the comparative of overcoming the limitations of NMF.
Inspired by latent representation, we propose a novel method named semi-supervised NMF via dual space latent representation learning and dual graph regularization (DLRGNMF) to alleviate the above issues. The main contributions of our work are as follows:
(1)
The dual latent representation mechanism is embedded into the semi-supervised NMF framework by the affinity matrix to explicitly exploit global interconnection information in dual space, which can reduce the adverse impacts caused by noise.
(2)
To steadily describe local data structures, the dual graph is introduced into latent representation learning to further fully investigate the coherent information structure in dual space.
(3)
To achieve a sparse representation of matrix factorization, the ℓ2,1-norm is incorporated into the basis matrix U in the proposed framework, which can simplify the measurement process and improve the clustering performance.
In conclusion, our proposed method can overcome the limitations of NMF to some extent. In real applicated fields our proposed method can be applied to image analysis, text classification, attribute community detection, face recognition, and recommender system.
This paper is constructed as follows: we review the related studies in Section 2; Section 3 illustrates the DLRGNMF algorithm, gives the iterative update rules, and provides a detailed analysis of convergence analysis; Section 4 evaluates DLRGNMF in terms of clustering analysis and ablation experiments; lastly, we summarize the conclusion in Section 5.

2. Related Work

In this section, we briefly review the algorithms related to our work.

2.1. CNMF

Studies show that, in semi-supervised algorithms, a small fraction of label information can improve the accuracy of learning [37,38]. As an extension of NMF, CNMF [21] incorporates the label information to improve the discriminating power, thus attaching label information to NMF as hard constraints. The objective function of CNMF is formulated as follows:
O C N M F = X U Z T A T F 2 , s . t .   U 0 ,   Z 0 ,
where Z is a label constraint matrix and A is a label auxiliary matrix. This strategy by constraint of Z can require the samples with the same label to be consistent to coordinate in the mapping space. Liu et al. [21] gave proof of the update iteration rule for CNMF and its convergence. CNMF integrates the merits of NMF and a semi-supervised mechanism to improve its discriminating power.

2.2. DNMF

Motivated by manifold learning theory, Shang et al. proposed DNMF [18], which constructs neighborhood graphs according to the observation data on the structure of data and feature manifold. The objective function of DNMF is formulated as follows:
O D N M F = X U V T F 2 + λ T r V T L Z V + μ T r U T L U U , s . t .   U 0 ,   V 0 ,
where λ and μ are graph regularization parameters. DNMF integrates the virtues of dual graph regularization to perform matrix factorization to further enhance its learning ability.

2.3. SODNMF

Inspired by CNMF and DNMF, Meng et al. [25] designed SODNMF. The geometric manifold information can be effectively modeled by adopting dual graph regularization constraints to NMF. Furthermore, the sparse constraint is also embedded to guarantee the sparsity. The objective function of SODNMF is expressed as
O S O D N M F = X P R A T C T F 2 + α T r P T L P P + T r A T C T L S C A + θ P 2 , 1 2 1 2 , s . t .   P T P = I , A T C T C A = I ,
where α is the regularization parameter, and θ is the sparse constraint parameter. The two terms associated with α are dual graph regularization terms, and the term is controlled by θ is a sparse constraint.

2.4. LRLMR

Traditional unsupervised feature selection methods ignore the interrelationship between data instances. For this reason, Tang et al. presented unsupervised feature selection via latent representation learning and manifold regularization (LRLMR) [26], integrating latent representation learning into the local manifold information to perform feature selection, which can explore the global interrelationship of data space by constructing an affinity matrix. LRLMR formulates the objective function as follows:
O L R L M R = X W V F 2 + α W 2 , 1 + β A V V T F 2 + γ T r W T X T L X W , s . t .   V 0 ,
where α is the sparse constraint parameter, β is the latent representation learning parameter, and γ is the manifold regularization parameter. LRLMR learns the latent representation matrix V into a pseudo-label matrix to yield the clustering index and further reduce noise.

2.5. Latent Representation

Latent representations of data have yielded promising results and attracted considerable attention in machine learning tasks. In the real world, the purpose of latent representations is to establish the link information through DMvNM [31], DSSNMF [32], ADGCFFS [29], symmetric NMF [39], etc. Generally, latent representation constructs an affinity matrix P R n × n to describe the interconnection information between samples by decomposing P through a symmetric NMF, which is represented as follows:
R = P V V T F 2 ,
where V R n × k is the low-dimensional representation by mapping to the new representation space, while k and n denote latent factors and the number of samples, respectively. The affinity matrix P indicates the global interconnection information between samples. However, interconnection information exists not only in data instances but also between features [28]. Figure 2 illustrates Pearson correlation coefficients between data instances and feature variables of the Soybean dataset to demonstrate the interconnection relationship. It can be clearly observed that, in Figure 2a,b, the absolute values of Pearson correlation coefficients are higher than 0.3 in most cases, especially in Figure 2a, which implies that there is strongly correlated interconnection information between data instances and correlated interconnection information feature variables; that is, interconnection information exists in dual space.

3. Proposed Method

Inspired by the theory in the literature [28,39], we propose a novel model dual space based on latent representation learning named DLRGNMF, whose pipeline is visualized in Figure 3.

3.1. Dual Space Latent Representation Learning and Dual Graph Regularization

Conventional non-negative matrix factorization algorithms always assume that the data are independently and uniformly distributed by default. However, this kind of distribution is not ideal in actual application. Since noise is generated by various factors, the data instances derived from homology or heterogeneity are often interdependent [13,40]. Therefore, it is significant to exploit the intrinsic data structure and feature structure through link information. Some dimensionality reduction algorithms [26,27,41] implemented this idea by virtue of latent representation learning and achieved excellent results. Inspired by these algorithms, we embed dual space latent representation learning into the semi-supervised learning of NMF by mapping the original data to learn the dual latent space of matrix factorization. Thus, the interconnection of information is investigated in mining dual space simultaneously to reduce the influence of the noise, and to improve the robustness of matrix factorization.
To perform matrix factorization in the learned dual latent space, we construct the objective function of the dual space latent representation learning as follows:
R 1 = M C A A T C T F 2 + N U U T F 2 ,
where A R n + c l × c is label auxiliary matrix, U R m × c is the basis matrix, C R n × n + c l is the label constraint matrix [21], n is the number of samples, l is the number of labeled data points, and c is the number of categories. The first term is to the latent representation of the data space by constructing an affinity matrix M R n × n in the data space, which can reflect inherent information between instances. The unlabeled dataset is defined by X = x 1 , x 2 , , x n R m × n , where m is the feature dimension of the sample. The affinity matrix M is defined as follows:
M i j = e x p x i x j 2 2 σ 2 ,     i , j = 1 , 2 , , n .
The second term in Equation (6) is constructed by an affinity matrix N R m × m to learn the feature latent space and investigate inherent information between features. The affinity matrix N is defined as follows:
N i j = e x p y i y j 2 2 σ 2 ,     i , j = 1 , 2 , , m ,
where x i , y i , and σ denote the i -th sample, the i -th feature of the data matrix X , and a Gaussian bandwidth parameter, respectively, with 0 < M i j 1 ,   0 < N i j 1 . M i j and N i j represent the interrelation between the i -th and j -th data instances and the interrelation between the i -th and j -th feature vectors, respectively. Therefore, M i j and N i j describe the global interconnection information in the dual space and learn this information through the situation of the symmetric NMF; therefore, the corresponding low-dimensional representation can be accomplished by matrix factorization under the constraints of this information.
With the development of graph theory and spectral theory, graph regularization has been successfully applied to NMF. In Equation (6), we introduce latent representation learning which maps the raw data into the dual latent space, thus constructing a Laplacian regularization term for dual graph [18], which is represented as follows:
R 2 = T r U T L N U + T r A T C T L M C A ,
where L M R n × n is the Laplacian matrix of the data space and L M = D M S M , D M satisfies D M i i = j S M i j , and S M is the similarity matrix constructed by the k -neighborhood graph M k x i as follows:
S M i j = M i j ,     i f x j   M k x i   0 ,                                         o t h e r w i s e             i , j = 1 , 2 , 3 , , n .
Similarly, the Laplacian matrix of the feature space is denoted as L N R m × m and L N = D N S N , D N satisfies D N i i = j S N i j , and S N is the similarity matrix, constructed as follows:
S N i j = N i j ,     i f y j   N k y i   0 ,                                         o t h e r w i s e             i , j = 1 , 2 , 3 , , m ,
where N k y i is a k -neighborhood graph.

3.2. Objective Function

First, we construct the objective function for the semi-supervised NMF as follows [25]:
U * , R * , C * , A * = a r g   m i n   X U R A T C T F 2 , s . t .   U T U = I ,   A T C T C A = I .
We incorporate biorthogonal constraints in the semi-supervised NMF, which possesses the robust ability to classify basic and coefficient vectors simultaneously to improve cluster performance. Since biorthogonal constraints are strong constraints, diagonal scaling matrix R can be utilized to avoid the problem of generating unreliable solutions. To simplify the calculation, the ℓ2,1-norm is introduced to the basis matrix U , and the novel objective function is expressed as follows:
U * , R * , C * , A * = a r g   m i n   X U R A T C T F 2 + θ U 2 , 1 . s . t .   U T U = I ,   A T C T C A = I .
where θ is the sparse constraint parameter, and θ > 0 controls the sparsity of the model.
DLRGNMF integrates dual space latent representation learning and dual graph regularization into a semi-supervised NMF framework. Combining Equations (6), (9), and (13), the final objective function of DLRGNMF is transformed into the following:
U * , R * , C * , A * = a r g   m i n   X U R A T C T F 2 + α T r U T L N U + T r A T C T L M C A + γ 2 M C A A T C T F 2 + N U U T F 2 + θ U 2 , 1 ,
s . t .   U T U = I ,   A T C T C A = I ,
where α and γ are the dual graph regularization parameter and the dual space parameter, respectively, and α , γ , θ > 0 , which can balance the weight of the above terms.

3.3. Optimization

In this subsection, the optimization principle of DLRGNMF is illustrated. Since the objective function in Equation (14) is simultaneously nonconvex with U , R , C , and A , it is challenging to solve. To solve the objective function in Equation (14), the objective function with a single variable is convex and can be optimized by the alternate iterative method. Thus, the entire optimization problem can be decomposed into four subproblems. By optimizing this objective function of DLRGNMF, the parameter β is expressed as the biorthogonal parameter to restrict the biorthogonal term and β > 0 . Therefore, the Lagrange Equation (15) is constructed as follows:
L U , R , C , A = X U R A T C T F 2 + α T r U T L N U + α T r A T C T L M C A + β T r U T U I + β T r A T C T C A I + γ 2 M C A A T C T F 2 + γ 2 N U U T F 2 + θ T r U T Q U ,
where Q R m × m is a diagonal matrix, and we can calculate the i -th diagonal element q i i of Q as follows:
q i i = 1 2 U i 2 .
To avoid overflow, we introduce a small constant ε , and Equation (16) can be represented as
q i i = 1 2 max ( U i 2 , ε ) .
In order to optimize U , R , C , and A , the partial derivatives of the Lagrange function in Equation (15) are formulated, yielding
L U = 2 X C A R T + 2 U R A T C T C A R T + 2 α L N U + 2 β U + 2 θ Q U 2 γ N U + 2 γ U U T U ,
L R = 2 U T X C A + 2 U T U R A T C T C A ,
L A = 2 C T X T U R + 2 C T C A R T U T U R + 2 α C T L M C A + 2 β C T C A 2 γ C T M C A + 2 γ C T C A A T C T C A ,
L C = 2 X T U R A T + 2 C A R T U T U R A T + 2 α L M C A A T + 2 β C A A T 2 γ M C A A T + 2 γ C A A T C T C A A T .
Then, according to the KKT condition [42], the iterative update equations for U , R , C , and A are written as follows:
u i j u i j X C A R T + α W N U + γ N U i j U R A T C T C A R T + α D N U + β U + θ Q U + γ U U T U i j ,
r i j r i j U T X C A i j U T U R A T C T C A i j ,
a i j a i j C T X T U R + α C T W M C A + γ C T M C A i j C T C A R T U T U R + α C T D M C A + β C T C A + γ C T C A A T C T C A i j ,
c i j c i j X T U R A T + α W M C A A T + γ M C A A T i j C A R T U T U R A T + α D M C A A T + β C A A T + γ C A A T C T C A A T i j .
With the above analysis, the procedure of DLRGNMF is summarized in Table 2.

3.4. Convergence Analysis

In this section, we prove the convergence of DLRGNMF by demonstrating that under the update rules in Equations (22)–(25), the objective function in Equation (14) is monotonically decreasing.
First, we need to introduce a theorem [43], which theoretically guarantees the convergence of DLRGNMF.
Definition 1. 
If there is a function  G x ,   x  such that  F x  satisfies
G x , x F x , G x , x = F x ,
F  is a nonincreasing function when the updating formula is written as
x t + 1 = arg m i n x   G x , x t ,
where  G x ,   x  is an auxiliary function of  F x .
Proof. 
F x t + 1 G x t + 1 , x t G x t , x t = F x t .
If the objective function is proved to be monotonic, it is retained to contain the U term and transformed into
F U = X U R A T C T F 2 + α T r U T L N U + β T r U T U I + γ 2 N U U T F 2 + θ T r U T Q U .
We can get the partial derivatives of F U in the first order and second order with respect to
F i j = 2 X C A R T + 2 U R A T C T C A R T + 2 α L N U + 2 β U + 2 θ Q U 2 γ N U + 2 γ U U T U ,
F i j = 2 R A T C T C A R T j j + 2 α L N i i + 2 β I i i + 2 θ Q i i 2 γ N i i + 2 γ U U T i i .
Lemma 1. 
G U i j , U i j t = F i j U i j t + F i j U i j t U i j U i j t + U R A T C T C A R T + α D N U + β U + θ Q U + γ U U T U i j U i j t U i j U i j t 2 ,
where  G U i j ,   U i j t  is the auxiliary function of  F i j .
Proof. 
The Taylor series expansion of F i j U i j is
F i j U i j = F i j U i j t + F i j U i j t U i j U i j t + R A T C T C A R T j j + α L N i i + β I i i + θ Q i i γ N i i + γ U U T i i U i j U i j t 2 .
G U i j , U i j t F i j U i j is equivalent to
U R A T C T C A R T + α D N U + β U + θ Q U + γ U U T U i j U i j t R A T C T C A R T j j + α L N i i + β I i i + θ Q i i γ N i i + γ U U T i i .
Since,
U R A T C T C A R T i j = k U i k t R A T C T C A R T k j R A T C T C A R T i i U i j t α D N U i j = α k D U i k U k j t α L N i i U i j t β U + θ Q U i j = k β I + θ Q i k U k j t α β I + θ Q i i U i j t γ U U T U i j = γ k U U T i k U k j t γ U U T i i U i j t ,
we get G U i j , U i j t F i j U i j . □
Next, we prove that according to the iterative update rule in Equation (22) F i j is monotonically decreasing.
Substituting Equation (32) G U i j , U i j t into Lemma 1 with x t + 1 = arg m i n x   G x , x t ,
U i j t + 1 = U i j t U i j t F i j U i j t 2 U R A T C T C A R T + α D N U + β U + θ Q U + γ U U T U i j = U i j t X C A R T + α W N U + γ N U i j U R A T C T C A R T + α D N U + β U + θ Q U + γ U U T U i j .
It can be seen from the updating rules of U that F i j monotonically decreases under updating Equation (22). The proof of the updating rules of R , A , and C is similar to that of U ; thus, we can obtain updating Equations (23)–(25). Therefore, we can conclude that DLRGNMF is convergent. □

4. Experiments

In this subsection, the effectiveness of DLRGNMF is demonstrated by comparing it with nine state-of-the-art algorithms on nine public datasets. Through the low-dimensional representations, we use k-means to obtain clustering results. We implement experiments with MATLAB R2018b on a Windows machine with a 3.10 GHz i5-11300H and 16 GB main memory.

4.1. Results on the Synthetic Dataset

To demonstrate the clustering effectiveness and noise robustness of DLRGNMF, we perform clustering experiments on this synthetic dataset and calculate clustering accuracy (ACC) [35] for the compared algorithms. The synthetic dataset includes three categories, with each category consisting of 300 data instances and seven feature dimensions, which involves the former two dimensions generated by Gaussian distribution, as shown in Figure 4a, and five noise dimensions randomly generated (0–5). We compare DLRGNMF with CNMF [21], GNMF [16], DNMF [18], DSNMF [20], NMFAN [30], EWRNMF [33], and SODNMF [25], as described in detail in Section 4.2.
Under the same experimental environment, DLRGNMF was conducted and compared with other algorithms firstly in terms of the dimension reduction on the synthetic dataset, and the clustering results are illustrated in Figure 4b–i. From Figure 4, it is demonstrated that DLRGNMF, in Figure 4i, achieves the best clustering result because the samples of each category are accurately assigned to the corresponding clusters, while the clustering results of other comparison methods are incorrect. The reason may be that, for the comparison algorithms with the graph model, the graph model may not be reliable when the data noise is high. In other words, the neighborhood graph will result in inaccurate clustering results in the period of constructing the neighborhood graph.
Moreover, the first two dimensions in the synthetic dataset are generated by Gaussian distribution, which implies that there are tighter interconnections between the samples and features of the data than other dimensions. DLRGNMF investigates the global interconnection information of the data through latent representation learning in dual space to reduce noise interference and further preserve the local manifold structure of samples. Consequently, the clustering result of DLRGNMF is significantly improved, which tends to alleviate noisy dimensions and the discriminative of features, as well as promote clustering precision.

4.2. Results on Public Datasets Benchmarks

In this section, the experimental performance is performed in terms of clustering accuracy (ACC) and normalized mutual information (NMI) on nine public datasets [32].

4.2.1. Datasets

The datasets include COIL20, JAFFE50, Lung_dis, ORL, Soybean, UMIST, warpPIE10P, Yale32, and Yale64 [36,44,45,46], and their details are shown in Table 3.

4.2.2. Compared Algorithms

  • PCA [47]: Principal component analysis generates a low-rank subspace with the major information of the data by the projection of the direction of maximum variance.
  • K-Multiple Means (KMM) [48]: It is an improved K-means by introducing the local centroids into K-means.
  • LRLMR [26]: Latent representation learning is embedded in feature selection to extract global interconnection information between data instances to guide feature selection.
  • CNMF [21]: Semi-supervision is embedded into NMF as a robust constraint to improve the algorithm performance by a small amount of label data.
  • DNMF [18]: The local manifold information of dual space is preserved by constructing a dual graph regularization.
  • DSNMF [20]: The local manifold structures of dual space are preserved and retain the sparsity simultaneously.
  • NMFAN [30]: The local manifold information is exploited by constructing an adaptive graph regularization term to obtain the optimal neighborhood graph.
  • EWRNMF [33]: An adaptive weight NMF that guides the matrix factorization by assigning adaptive weights to each data instance.
  • SODNMF [25]: A semi-supervised NMF with dual graph regularization terms, sparse constraints and biorthogonal constraints.

4.2.3. Experimental Settings

DLRGNMF and other comparison algorithms are clustered by K-means, and the average and standard deviation of the clustering results over 20 iterations are calculated as the final result. For parameter setting, the neighborhood size k was set to 5, the maximum number of iterations was set to 100, and the ratio of training samples p e r was set to 0.1. For other parameters of the comparison algorithm, the ranges were set to be consistent with the corresponding literature. For DLRGNMF, we tuned the balance parameters α in the range of {100, 101, …, 106}, γ in the range of {10−3, 10−2, …, 102, 103}, β in the range of {10−8, 10−5, 10−3, 10−1, 100}, θ in the range of {100, 103, 108, 1018, 1028}, and the kernel function σ to 1.

4.2.4. Performance

The comparative clustering results are presented in Table 4 and Table 5 with the best clustering results in bold, and the computation time on real-world datasets is listed in Table 6. To improve the comprehensibility of Table 4, Table 5 and Table 6, we conducted a visual analysis of the results, as demonstrated in Figure 5, Figure 6 and Figure 7. From these tables and figures, DLRGNMF outperforms the compared algorithms in ACC and NMI on all datasets, which fully demonstrates that it can effectively extract features and has excellent low-dimensional learning ability. The detailed conclusions can be drawn as follows.
(1)
GNMF, DNMF, and DSNMF are three NMF variants with graph regularization terms. They outperformed LRLMR, KMM, and PCA on most datasets except for Lung_dis and ORL. The reason is that the graph models are constructed to retain the local manifold structure, which can achieve promising clustering results.
(2)
CNMF is superior to GNMF on the ORL, Yale32, and Yale64 datasets, which reflects the merits of semi-supervised NMF in improving clustering accuracy with less label information.
(3)
LRLMR achieves better performance than PCA, KMM, and CNMF on most test datasets since LRLMR exploits interconnection information by latent representation learning, which can further enhance the discriminative of the model.
(4)
NMFAN and EWRNMF are two adaptive methods that construct an adaptive graph and adaptive weights, respectively. Unsatisfactory results are achieved on most datasets because they fail to fully explore interconnection information and label information.
(5)
Compared with other methods, DLRGNMF achieves the best performance, especially on COIL20; the increases in ACC values are 19.09%, 21.25%, 16.65%, 19.98%, 6.91%, 2.61%, 1.83%, 19.37%, 17.90%, and 1.72% with respect to PCA, KMM, LRLMR, CNMF, GNMF, DNMF, DSNMF, NMFAN, EWRNMF, and SODNMF. DLRGNMF embeds dual space latent representation learning to guide matrix decomposition and reduce noise. Therefore, it learns more information about interconnection and achieves the best clustering results.
(6)
In terms of running time, DLRGNMF is a little slower than some compared methods on some datasets because latent representation learning inevitably entails more operations, whereas it outperforms SODNMF on Jaffe50, UMIST, and warpPIE10P datasets due to rapid convergence.

4.2.5. Intuitive Presentation

In Figure 8, the low-dimensional representation of DLRGNMF on nine benchmark datasets is processed with t-SNE to visualize the clustering results in the 2D plane. We can explicitly observe that the features extracted by DLRGNMF for COIL20, JAFFE50, Soybean, and warpPIE10P result in the intraclass and interclass distances of the samples differing significantly, which clearly implies that clustering boundaries and higher ACC values are obtained. However, the number of features in ORL, Yale32, and Yale64 is 10 times that of the samples in each class; thus, the low-dimensional representations may have fuzzy boundaries between some classes, which leads to inferior clustering results compared with other datasets. Overall, the low-dimensional representations extracted by DLRGNMF can clearly indicate the spatial differences between different classes of samples, enhancing its discrimination performance between various classes of samples.

4.2.6. Ablation Study

To investigate the influence of each item on the clustering performance of DLRGNMF, ablation experiments were conducted on nine datasets, and the comparison results are presented in Figure 9. DLRGNMF without dual graph regularization term is denoted as DLRG-1, the model without sparse constraints is denoted as DLRG-2, and the model without latent representation learning is denoted as DLRG-3.
As shown in Figure 9, DLRGNMF achieves the best performance. In terms of clustering performance, DLRG-2 is nearly similar to DLRGNMF, while DLRG-1 achieves the worst performance. It is demonstrated that the dual graph regularization term plays the most important role in the model because the dual graph regularization term can preserve both the local manifold structure and the global inherent information, which are also complementary by latent representation learning. By comparison, without latent representation learning, the performance of DLRG-3 leads to a more declined degree than DLRG-2. Consequently, the graph regularization term has the most significant contribution to the clustering performance. Meanwhile, latent representation learning is also contributive and yields a degree of improvement.
To validate the sparsity, we calculated the sparseness of the learned basis vectors using DLRGNMF and DLRG-2 on nine datasets according to Equation (27) from [49], and the experimental results are shown in Figure 10. Generally, a higher value of sparseness indicates better sparsity. It is evident that the sparseness of DLRGNMF is higher than that of DLRG-2 on all test datasets, which implies that DLRGNMF can yield better sparse representation.

4.2.7. Convergence Study

Figure 11 shows the convergence curves of DLRGNMF on nine datasets. The curves converge efficiently and stably on all datasets, with fewer than five iterations on most datasets, which further demonstrates that DLRGNMF is convergent, verifying Section 3.4.

4.2.8. Parameter Sensitivity Experiment

From ablation experiments, it can be concluded that a significant contribution to the clustering performance is provided by the dual graph regularization term and latent representation learning term in dual space. Hence, we conducted parameter sensitivity experiments on these two parameters, α and γ . The search ranges of parameter α and γ were respectively set as {100, 101, …, 106} and {10−3, 10−2, …, 102, 103}, while the other parameters were fixed at β = 10−3   and   θ = 103. Figure 12 and Figure 13 illustrate the varying ACC and NMI under different constituted values of α and γ . These figures imply that ACC is positively correlated with the trend of NMI, and that the clustering performance increases with parameter α in the COIL20, JAFFE50, ORL, UMIST, warpPIE10P, Yale32, and Yale64 datasets. Hence, the parameter α should not be set too small; otherwise, it tends to achieve bad clustering results. The suitable ranges of parameters α and γ are respectively [103, 106] and [101, 103].

4.2.9. Noise Test

To further verify the robustness of DLRGNMF, a noise test was conducted. Three block sizes of 8 × 8, 12 × 12, and 16 × 16 were chosen from images in the Yale32 and JAFFE50 datasets as noise. These noises were randomly synthesized in the original images as shown in Figure 14b–d and Figure 14f–h, and the comparison results on the noisy Yale32 and JAFFE50 datasets are shown in Table 7 and Table 8, where the best results are shown in bold. It can be observed DLRGNMF can achieve better clustering performance than the compared methods under different noise influences. Consequently, DLRGNMF has a stronger robustness of features by latent representation learning.

5. Conclusions

In this paper, a novel DLRGNMF algorithm was proposed. Due to the interrelation of data instances in the real world, dual space latent representation learning is embedded into semi-supervised NMF, which can fully exploit the global interconnection information of dual space by constructing dual latent space. In the mapped dual latent space, the local manifold information of dual space in the raw data is retained by dual graph regularization. Through latent representation learning and manifold regularization, DLRGNMF can reduce reductant information and further improve the low-dimensional learning ability of matrix factorization. Extensive experiments on different datasets verified its superiority and noise reduction capability.
The limitation of DLRGNMF is that there are some parameters to be tuned. In future work, it is desirable to incorporate a regular term without parameters to constrain NMF, which will significantly reduce the time cost of the model. Meanwhile, we would like to explore an efficient optimization method that can optimize all variables simultaneously.

Author Contributions

Y.H., software, data curation, and writing—original draft preparation; Z.M., conceptualization, methodology, writing—review and editing, and validation; H.L., visualization and investigation; J.W., supervision and writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Ningxia (Nos. 2020AAC03215 and 2022AAC03268), the National Natural Science Foundation of China (Nos. 61462002), and Basic Scientific Research in Central Universities of North Minzu University (Nos. 2021KJCX09 and FWNX21).

Data Availability Statement

The data and code that support the findings of this study are available from the corresponding author (Z.M) upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Roweis, S.; Saul, L. Nonlinear dimensionality reduction by locally linear embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef]
  2. Lee, D.; Seung, H. Algorithms for non-negative matrix factorization. Adv. Neural Inf. Process. Syst. 2000, 13, 556–562. [Google Scholar]
  3. Lee, D.; Seung, H. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788–791. [Google Scholar]
  4. Xu, Z.; King, I.; Lyu, M.; Jin, R. Discriminative semi-supervised feature selection via manifold regularization. IEEE Trans. Neural Netw. 2010, 21, 1033–1047. [Google Scholar]
  5. Lipovetsky, S. PCA and SVD with nonnegative loadings. Pattern Recognit. 2009, 42, 68–76. [Google Scholar] [CrossRef]
  6. Sandler, R.; Lindenbaum, M. Nonnegative matrix factorization with earth mover’s distance metric for Image analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 1590–1602. [Google Scholar]
  7. Tu, D.; Chen, L.; Lv, M.; Shi, H.; Chen, G. Hierarchical online NMF for detecting and tracking topic hierarchies in a text stream. Pattern Recognit. 2018, 76, 203–214. [Google Scholar] [CrossRef]
  8. Chen, W.; Zhao, Y.; Pan, B.; Chen, B. Supervised kernel nonnegative matrix factorization for face recognition. Neurocomputing 2016, 205, 165–181. [Google Scholar] [CrossRef]
  9. Huang, Y.; Yang, G.; Wang, K.; Liu, H.; Yin, Y. Robust multi-feature collective non-negative matrix factorization for ECG biometrics. Pattern Recognit. 2022, 123, 108376. [Google Scholar] [CrossRef]
  10. Zheng, Z.; Yang, J.; Zhu, Y. Initialization enhancer for non-negative matrix factorization. Eng. Appl. Artif. Intell. 2007, 20, 101–110. [Google Scholar] [CrossRef]
  11. Zdunek, R.; Sadowski, T. Segmented convex-hull algorithms for near-separable NMF and NTF. Neurocomputing 2019, 331, 150–164. [Google Scholar] [CrossRef]
  12. Chen, M.; Gong, M.; Li, X. Feature Weighted Non-Negative Matrix Factorization. IEEE Trans. Cybern. 2021, 53, 1093–1105. [Google Scholar]
  13. Jacob, Y.; Denoyer, L.; Gallinari, P. Learning latent representations of nodes for classifying in heterogeneous social networks. In Proceedings of the 7th ACM International Conference on Web Search and Data Mining, New York, NY, USA, 24 February 2014; pp. 373–382. [Google Scholar]
  14. Samaria, F.; Harter, A. Parameterisation of a stochastic model for human face identification. In Proceedings of the 2nd IEEE Workshop on Applications of Computer Vision, Sarasota, FL, USA, 1 January 1994; pp. 138–142. [Google Scholar]
  15. Huang, J.; Nie, F.; Huang, H.; Ding, C. Robust manifold nonnegative matrix factorization. ACM Trans. Knowl. Disc. Data 2014, 8, 1–21. [Google Scholar] [CrossRef]
  16. Cai, D.; He, X.; Han, J.; Huang, T.S. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 1548–1560. [Google Scholar]
  17. Liu, H.; Yang, Z.; Yang, J.; Wu, Z.; Li, X. Local coordinate concept factorization for image representation. IEEE Trans. Neural Netw. Learn. Syst. 2014, 25, 1071–1082. [Google Scholar]
  18. Shang, F.; Jiao, L.; Wang, F. Graph dual regularization nonnegative matrix factorization for co-clustering. Pattern Recognit. 2012, 45, 2237–2250. [Google Scholar] [CrossRef]
  19. Hoyer, P. Non-negative matrix factorization with sparseness constraints. J. Mach. Learn. Res. 2004, 5, 1457–1469. [Google Scholar]
  20. Meng, Y.; Shang, R.; Jiao, L.; Zhang, W.; Yuan, Y.; Yang, S. Feature selection based dual-graph sparse non-negative matrix factorization for local discriminative clustering. Neurocomputing 2018, 290, 87–99. [Google Scholar] [CrossRef]
  21. Liu, H.; Wu, Z.; Cai, D.; Huang, T. Constrained nonnegative matrix factorization for image representation. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1299–1311. [Google Scholar] [CrossRef]
  22. Peng, S.; Ser, W.; Chen, B.; Lin, Z. Robust semi-supervised nonnegative matrix factorization for image clustering. Pattern Recognit. 2021, 111, 107683. [Google Scholar] [CrossRef]
  23. Sun, J.; Wang, Z.; Sun, F.; Li, H. Sparse dual graph-regularized NMF for image co-clustering. Neurocomputing 2018, 316, 156–165. [Google Scholar] [CrossRef]
  24. Babaee, M.; Tsoukalas, S.; Babaee, M.; Rigoll, G.; Datcu, M. Discriminative non-negative matrix factorization for dimensionality reduction. Neurocomputing 2016, 173, 212–223. [Google Scholar] [CrossRef]
  25. Meng, Y.; Shang, R.; Jiao, L.; Zhang, W. Dual-graph regularized non-negative matrix factorization with sparse and orthogonal constraints. Eng. Appl. Artif. Intell. 2018, 69, 24–35. [Google Scholar] [CrossRef]
  26. Tang, C.; Bian, M.; Liu, X.; Li, M.; Zhou, H.; Wang, P.; Yin, H. Unsupervised feature selection via latent representation learning and manifold regularization. Neural Netw. 2019, 117, 163–178. [Google Scholar] [CrossRef] [PubMed]
  27. Ding, D.; Yang, X.; Xia, F.; Ma, T.; Liu, H.; Tang, C. Unsupervised feature selection via adaptive hypergraph regularized latent representation learning. Neurocomputing 2020, 378, 79–97. [Google Scholar] [CrossRef]
  28. Shang, R.; Wang, L.; Shang, F.; Jiao, L.; Li, Y. Dual space latent representation learning for unsupervised feature selection. Pattern Recognit. 2021, 114, 107873. [Google Scholar] [CrossRef]
  29. Ye, J.; Jin, Z. Feature selection for adaptive dual-graph regularized concept factorization for data representation. Neural Process. Lett. 2017, 45, 667–668. [Google Scholar] [CrossRef]
  30. Huang, S.; Xu, Z.; Wang, F. Nonnegative matrix factorization with adaptive neighbors. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14 May 2017; pp. 486–493. [Google Scholar]
  31. Luo, P.; Peng, J.; Guan, Z.; Fan, J. Dual-regularized multi-view non-negative matrix factorization. Neurocomputing 2018, 294, 1–11. [Google Scholar] [CrossRef]
  32. Xing, Z.; Wen, M.; Peng, J.; Feng, J. Discriminative semi-supervised non-negative matrix factorization for data clustering. Eng. Appl. Artif. Intell. 2021, 103, 104289. [Google Scholar] [CrossRef]
  33. Shen, T.; Li, J.; Tong, C.; He, Q.; Li, C.; Yao, Y.; Teng, Y. Adaptive weighted nonnegative matrix factorization for robust feature representation. arXiv 2022, arXiv:2206.03020. [Google Scholar] [CrossRef]
  34. Li, H.; Gao, Y.; Liu, J.; Zhang, J.; Li, C. Semi-supervised graph regularized nonnegative matrix factorization with local coordinate for image representation. Signal Process. Image Commun. 2022, 102, 116589. [Google Scholar] [CrossRef]
  35. Chavoshinejad, J.; Seyedi, S.; Tab, F.; Salahian, N. Self-supervised semi-supervised nonnegative matrix factorization for data clustering. Pattern Recognit. 2023, 137, 109282. [Google Scholar] [CrossRef]
  36. Li, S.; Li, W.; Lu, H.; Li, Y. Semi-supervised non-negative matrix tri-factorization with adaptive neighbors and block-diagonal learning. Eng. Appl. Artif. Intell. 2023, 121, 106043. [Google Scholar] [CrossRef]
  37. Lu, M.; Zhao, X.; Zhang, L.; Li, F. Semi-supervised concept factorization for document clustering. Inf. Sci. 2016, 331, 86–98. [Google Scholar] [CrossRef]
  38. Feng, X.; Jiao, Y.; Lv, C.; Zhou, D. Label consistent semi-supervised non-negative matrix factorization for maintenance activities identification. Eng. Appl. Artif. Intell. 2016, 52, 161–167. [Google Scholar] [CrossRef]
  39. Kuang, D.; Ding, C.; Park, H. Symmetric nonnegative matrix factorization for graph clustering. In Proceedings of the 2012 SIAM International Conference on Data Mining, Anaheim, CA, USA, 26–28 April 2012; pp. 106–117. [Google Scholar]
  40. Tang, L.; Liu, H. Relational learning via latent social dimensions. In Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Paris, France, 28 June–1 July 2009; pp. 817–826. [Google Scholar]
  41. Cui, J.; Zhu, Q.; Wang, D.; Li, Z. Learning robust latent representation for discriminative regression. Pattern Recognit. Lett. 2018, 117, 193–200. [Google Scholar] [CrossRef]
  42. Li, H.; Zhang, J.; Liu, J. Class-driven concept factorization for image representation. Neurocomputing 2016, 190, 197–208. [Google Scholar] [CrossRef]
  43. Shang, R.; Zhang, Z.; Jiao, L.; Liu, C.; Li, Y. Self-representation based dual-graph regularized feature selection clustering. Neurocomputing 2016, 171, 1242–1253. [Google Scholar] [CrossRef]
  44. Shang, R.; Wang, W.; Stolkin, R.; Jiao, L. Subspace learning-based graph regularized feature selection. Knowl. Based Syst. 2016, 112, 152–165. [Google Scholar] [CrossRef]
  45. Nie, F.; Wu, D.; Wang, R.; Li, X. Self-weighted clustering with adaptive neighbors. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 3428–3441. [Google Scholar] [CrossRef]
  46. Shang, R.; Zhang, Z.; Jiao, L.; Wang, W.; Yang, S. Global discriminative-based nonnegative spectral clustering. Pattern Recognit. 2016, 55, 172–182. [Google Scholar] [CrossRef]
  47. Gewers, F.; Ferreira, G.; Arruda, H.; Silva, F.; Comin, C.; Amancio, D.; Costa, L. Principal component analysis: A natural approach to data exploration. ACM Comput. Surv. 2021, 54, 1–34. [Google Scholar] [CrossRef]
  48. Nie, F.; Wang, C.; Li, X. K-multiple-means: A multiple-means clustering method with specified K clusters. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2019, Anchorage, AK, USA, 4–8 August 2019; pp. 959–967. [Google Scholar]
  49. Sun, J.; Gai, X.; Sun, F.; Hong, R. Dual graph-regularized constrained nonnegative matrix factorization for image clustering. KSII Trans. Internet Inf. Syst. 2017, 11, 2607–2627. [Google Scholar] [CrossRef]
Figure 1. Samples from the ORL dataset. (ac) are different classes in the ORL dataset.
Figure 1. Samples from the ORL dataset. (ac) are different classes in the ORL dataset.
Mathematics 11 02526 g001
Figure 2. Pearson correlation coefficient 3D diagram of the Soybean dataset: (a) Pearson correlation coefficient between samples; (b) Pearson correlation coefficient between features.
Figure 2. Pearson correlation coefficient 3D diagram of the Soybean dataset: (a) Pearson correlation coefficient between samples; (b) Pearson correlation coefficient between features.
Mathematics 11 02526 g002
Figure 3. The framework of the proposed DLRGNMF method.
Figure 3. The framework of the proposed DLRGNMF method.
Mathematics 11 02526 g003
Figure 4. Clustering performance of nine NMF variants methods on a Gaussian distributed synthetic dataset.
Figure 4. Clustering performance of nine NMF variants methods on a Gaussian distributed synthetic dataset.
Mathematics 11 02526 g004aMathematics 11 02526 g004b
Figure 5. Visualization of ACC (mean ± SD %) for different algorithms on real-world datasets.
Figure 5. Visualization of ACC (mean ± SD %) for different algorithms on real-world datasets.
Mathematics 11 02526 g005
Figure 6. Visualization of NMI (mean ± SD %) for different algorithms on real-world datasets.
Figure 6. Visualization of NMI (mean ± SD %) for different algorithms on real-world datasets.
Mathematics 11 02526 g006
Figure 7. Visualization of computation time (s) for different algorithms on real-world datasets.
Figure 7. Visualization of computation time (s) for different algorithms on real-world datasets.
Mathematics 11 02526 g007aMathematics 11 02526 g007b
Figure 8. The 2D presentation of clustering results by DLRGNMF on nine benchmark datasets. Different colors indicate different classes.
Figure 8. The 2D presentation of clustering results by DLRGNMF on nine benchmark datasets. Different colors indicate different classes.
Mathematics 11 02526 g008aMathematics 11 02526 g008b
Figure 9. Ablation experiments of DLRGNMF on nine datasets.
Figure 9. Ablation experiments of DLRGNMF on nine datasets.
Mathematics 11 02526 g009aMathematics 11 02526 g009b
Figure 10. A sparseness measure of DLRGNMF and DLRG-2 on nine datasets.
Figure 10. A sparseness measure of DLRGNMF and DLRG-2 on nine datasets.
Mathematics 11 02526 g010
Figure 11. Convergence curves of DLRGNMF on nine datasets.
Figure 11. Convergence curves of DLRGNMF on nine datasets.
Mathematics 11 02526 g011
Figure 12. The ACC results of DLRGNMF on the nine datasets under different parameters.
Figure 12. The ACC results of DLRGNMF on the nine datasets under different parameters.
Mathematics 11 02526 g012
Figure 13. The NMI results of DLRGNMF on the nine datasets under different parameters.
Figure 13. The NMI results of DLRGNMF on the nine datasets under different parameters.
Mathematics 11 02526 g013aMathematics 11 02526 g013b
Figure 14. Samples from Yale32 and JAFFE50 datasets with noise of different sizes.
Figure 14. Samples from Yale32 and JAFFE50 datasets with noise of different sizes.
Mathematics 11 02526 g014
Table 1. Summary of the state-of-the-art NMF variants.
Table 1. Summary of the state-of-the-art NMF variants.
MethodsSemi-SupervisedSparse RepresentationLocal Data StructureInterconnection Information
NMF with sparseness constraints (2004) [19]×××
Constrained NMF (CNMF) (2012) [21]×××
Graph regularized NMF (GNMF) (2011) [16]×××
Graph dual regularization NMF (DNMF) (2012) [18]×××
Robust manifold NMF (RMNMF) (2014) [15]×××
Local coordinate concept factorization (LCF) (2014) [17]××
Discriminative NMF (2016) [24]×××
Adaptive dual-graph regularized CF with FS (ADGCFFS) (2017) [29]×××
NMF with adaptive neighbors (NMFAN) (2017) [30]×××
Dual-graph sparse NMF (DSNMF) (2018) [20]××
Dual-regularized multi-view NMF (DMvNM) (2018) [31]×××
Dual-graph regularized NMF with sparse and orthogonal constraints (SODNMF) (2018) [25]×
Correntropy-based semi-supervised NMF (CSNMF) (2021) [22]×××
Discriminative semi-supervised NMF (DSSNMF) (2021) [32]×××
Feature weighted NMF (FNMF) (2021) [12]×××
Entropy weighted robust NMF (EWRNMF) (2022) [33]××××
Semi-supervised graph regularized NMF with local coordinate (SGLNMF) (2022) [34]×
Self-supervised semi-supervised NMF (S4NMF) (2023) [35]×××
Semi-supervised NMF with adaptive neighbors and block-diagonal (ABNMTF) (2023) [36]××
DLRGNMF (this study)
Table 2. DLRGNMF algorithm steps.
Table 2. DLRGNMF algorithm steps.
Algorithm 1 The optimization process of DLRGNMF
1: Input: Data matrix X R m × n , class number of samples c, neighborhood size k, balance parameters α, β, θ and γ, the maximum number of iterations N I t e r , and the ratio of training samples p e r .
2: Initialization: Normalize the data matrix X, generate U, R, and A, and pick up per% as the label information from the original data to construct matrix C, iteration time t = 0;
3: Construct the dual space latent representation learning;
4: Construct the dual-graph regularized model;
5: while not converged do
6:    Update U using Equation (22);
7:    Update R using Equation (23);
8:    Update A using Equation (24);
9:    Update C using Equation (25);
10:     Update Q using Equation (17);
11:     Update t by: t = t + 1, tNIter;
12: end while
13: Output: The label constraint matrix C, the basis matrix U, the diagonal scaling matrix R, and the label auxiliary matrix A.
Table 3. Details of nine datasets.
Table 3. Details of nine datasets.
DatasetsSamplesFeaturesClassesType
COIL201440102420Object image
JAFFE50213102410Face image
Lung_dis733257Biological
ORL400102440Face image
Soybean47354Digital image
UMIST575102420Face image
PIE10P210242010Face image
Yale32165102415Face image
Yale64165409615Face image
Table 4. ACC (mean ± SD %) of different algorithms on real-world datasets.
Table 4. ACC (mean ± SD %) of different algorithms on real-world datasets.
MethodsCOIL20JAFFE50Lung_disORLSoybeanUMISTPIE10PYale32Yale64
PCA64.54 ± 3.0089.60 ± 1.2582.53 ± 5.8555.15 ± 2.3873.09 ± 1.0440.03 ± 1.4429.19 ± 1.9742.18 ± 2.9852.94 ± 3.11
KMM62.82 ± 3.0985.92 ± 5.2978.90 ± 5.3452.11 ± 2.4776.28 ± 14.8341.32 ± 1.9831.90 ± 1.9539.03 ± 3.0447.70 ± 2.61
LRLMR66.49 ± 2.8984.08 ± 4.5282.19 ± 4.5854.43 ± 3.0489.15 ± 4.6855.48 ± 2.9248.74 ± 3.1241.09 ± 2.8550.94 ± 3.33
CNMF63.83 ± 2.5381.13 ± 6.7551.99 ± 5.6362.57 ± 2.9280.96 ± 7.2240.18 ± 1.7841.52 ± 2.5843.21 ± 3.0054.36 ± 4.11
GNMF74.26 ± 4.3186.20 ± 4.7468.22 ± 2.9055.13 ± 2.3388.72 ± 2.5953.71 ± 4.0978.14 ± 5.4340.21 ± 1.7551.52 ± 1.96
DNMF77.69 ± 3.3191.03 ± 5.5768.49 ± 6.0559.44 ± 1.7389.26 ± 0.4857.26 ± 4.2977.88 ± 3.4040.70 ± 2.5953.52 ± 3.15
DSNMF78.31 ± 2.5090.85 ± 5.0671.03 ± 0.9258.76 ± 2.2192.13 ± 1.9656.36 ± 3.1178.36 ± 3.7042.58 ± 2.5152.24 ± 2.04
NMFAN64.32 ± 3.8283.85 ± 4.3049.86 ± 4.5854.69 ± 2.4682.34 ± 8.2641.36 ± 2.1343.17 ± 3.3841.39 ± 3.2951.79 ± 3.54
EWRNMF65.49 ± 2.7283.22 ± 4.9549.25 ± 7.6353.09 ± 2.6580.32 ± 7.4641.37 ± 2.6140.69 ± 2.6042.27 ± 2.2752.42 ± 3.22
SODNMF78.40 ± 3.6192.51 ± 4.5873.56 ± 5.0364.04 ± 2.1894.15 ± 5.1156.73 ± 5.5479.40 ± 3.7843.24 ± 1.7255.36 ± 2.74
DLRGNMF79.77 ± 2.8993.33 ± 4.3382.67 ± 4.0665.04 ± 2.6694.79 ± 5.3758.23 ± 4.8580.05 ± 4.1745.12 ± 2.6956.33 ± 2.23
Table 5. NMI (mean ± SD %) of different algorithms on real-world datasets.
Table 5. NMI (mean ± SD %) of different algorithms on real-world datasets.
MethodsCOIL20JAFFE50Lung_disORLSoybeanUMISTPIE10PYale32Yale64
PCA74.96 ± 1.4586.85 ± 1.2675.16 ± 3.7273.42 ± 1.4071.22 ± 0.1759.78 ± 1.0831.17 ± 2.4747.26 ± 2.6156.71 ± 2.69
KMM74.04 ± 1.6085.40 ± 3.6072.04 ± 4.5971.62 ± 1.3868.74 ± 17.4060.63 ± 1.3836.14 ± 3.8944.51 ± 2.8951.88 ± 2.48
LRLMR75.29 ± 1.1583.19 ± 2.5874.86 ± 4.4773.25 ± 1.6985.38 ± 6.3067.77 ± 1.7860.45 ± 3.1045.03 ± 2.2854.71 ± 2.01
CNMF73.13 ± 1.5380.57 ± 3.6847.15 ± 4.5972.64 ± 1.8377.01 ± 7.2858.40 ± 1.5249.74 ± 3.2349.40 ± 2.3858.60 ± 3.41
GNMF85.04 ± 2.4183.31 ± 3.5766.23 ± 1.7473.38 ± 1.3780.74 ± 2.4968.81 ± 2.2782.90 ± 3.1945.52 ± 1.5754.27 ± 1.53
DNMF88.77 ± 1.3391.14 ± 2.9063.52 ± 7.3474.19 ± 1.2381.30 ± 0.8674.86 ± 2.0983.14 ± 1.5844.88 ± 2.4755.28 ± 1.79
DSNMF88.80 ± 1.3090.77 ± 1.9867.65 ± 1.2574.26 ± 1.2084.95 ± 3.0774.36 ± 1.9582.59 ± 2.7046.43 ± 1.8454.39 ± 1.87
NMFAN73.31 ± 2.0482.08 ± 2.8339.29 ± 4.1973.61 ± 1.3977.48 ± 7.5759.65 ± 1.6650.32 ± 2.6245.46 ± 2.6055.37 ± 3.10
EWRNMF74.53 ± 1.9580.85 ± 3.6943.86 ± 7.0971.89 ± 1.4375.38 ± 5.4360.52 ± 1.6246.67 ± 2.6947.15 ± 1.5754.77 ± 2.87
SODNMF88.79 ± 1.3491.78 ± 2.2469.72 ± 5.1177.65 ± 1.4589.35 ± 8.4575.46 ± 2.5283.42 ± 2.0848.95 ± 1.5757.52 ± 2.14
DLRGNMF89.34 ± 0.9892.13 ± 2.0975.22 ± 2.2878.37 ± 1.7090.94 ± 8.4675.67 ± 3.1083.89 ± 2.2650.28 ± 2.2557.97 ± 1.97
Table 6. Computation time (s) of different algorithms on real-world datasets.
Table 6. Computation time (s) of different algorithms on real-world datasets.
MethodsCOIL20Jaffe50Lung_disORLSoybeanUMISTPIE10PYale32Yale64
PCA0.20640.01640.02180.06710.00440.07440.03200.01650.0233
KMM4.68381.46970.97101.89790.01082.76228.89672.118026.5506
LRLMR7.37280.73370.01195.81500.00091.214615.09922.7130129.4183
CNMF42.41592.55300.17134.97340.01188.253910.07212.176425.4597
GNMF0.83450.08550.04440.42060.01560.32920.13600.09630.3016
DNMF5.68500.67850.08451.21010.00711.67072.37830.63556.3090
DSNMF5.79170.73140.08531.19190.00351.71082.56780.03280.2136
NMFAN15.48420.56600.45486.89500.03032.81731.92550.56620.9083
EWRNMF1.71530.22760.02620.54180.00450.64290.42730.19700.6678
SODNMF45.77550.93430.07020.25720.00848.558910.23580.09500.9597
DLRGNMF52.32440.41560.26530.34810.02935.91248.24030.16732.4980
Table 7. ACC and NMI (mean ± SD %) on the noised Yale32 datasets.
Table 7. ACC and NMI (mean ± SD %) on the noised Yale32 datasets.
MethodsAccuracy (%)Normalized Mutual Information (%)
8 × 8 Noise12 × 12 Noise16 × 16 Noise8 × 8 Noise12 × 12 Noise16 × 16 Noise
PCA41.24 ± 2.0340.91 ± 2.7139.76 ± 2.7746.98 ± 2.2947.02 ± 2.1546.14 ± 1.98
KMM35.24 ± 2.9432.18 ± 3.3831.52 ± 2.7240.63 ± 3.8737.70 ± 3.8036.24 ± 3.10
LRLMR40.21 ± 3.5038.15 ± 3.1935.76 ± 1.7744.28 ± 3.0342.06 ± 3.0238.31 ± 2.37
CNMF42.36 ± 3.9742.24 ± 3.5039.97 ± 3.7148.46 ± 2.8047.62 ± 2.5546.31 ± 3.01
GNMF40.09 ± 3.6239.55 ± 2.2939.09 ± 2.6545.19 ± 2.9244.45 ± 2.3644.89 ± 2.76
DNMF38.36 ± 2.6537.42 ± 1.6936.45 ± 1.9944.26 ± 2.5144.10 ± 1.3842.36 ± 1.89
DSNMF40.27 ± 2.1639.42 ± 2.2337.18 ± 1.7145.97 ± 1.9444.17 ± 2.1042.61 ± 2.00
NMFAN40.03 ± 2.7839.21 ± 2.9038.18 ± 3.6645.22 ± 2.1544.84 ± 2.8244.15 ± 3.01
EWRNMF39.79 ± 2.2838.76 ± 3.1438.39 ± 2.4044.96 ± 1.8344.06 ± 2.6143.28 ± 2.03
SODNMF42.94 ± 2.3940.58 ± 2.5138.48 ± 2.1249.04 ± 1.8647.11 ± 2.5445.91 ± 1.65
DLRGNMF44.58 ± 2.6442.39 ± 1.8840.24 ± 2.3050.43 ± 2.5147.88 ± 1.5546.37 ± 1.68
Table 8. ACC and NMI (mean ± SD %) on the noised JAFFE50 datasets.
Table 8. ACC and NMI (mean ± SD %) on the noised JAFFE50 datasets.
MethodsAccuracy (%)Normalized Mutual Information (%)
8 × 8 Noise12 × 12 Noise16 × 16 Noise8 × 8 Noise12 × 12 Noise16 × 16 Noise
PCA88.24 ± 5.3886.67 ± 2.2183.29 ± 4.3688.40 ± 3.3685.25 ± 2.8781.83 ± 2.06
KMM86.01 ± 5.0283.83 ± 3.3177.82 ± 5.5287.16 ± 2.6086.25 ± 2.7281.64 ± 3.87
LRLMR83.09 ± 3.4982.90 ± 3.9577.93 ± 3.3482.23 ± 2.8581.29 ± 2.7977.17 ± 2.86
CNMF83.10 ± 6.6081.08 ± 5.6478.97 ± 5.1881.50 ± 4.9880.07 ± 3.0877.80 ± 3.64
GNMF83.22 ± 3.9381.71 ± 3.0978.83 ± 6.4184.40 ± 2.4283.94 ± 2.2178.13 ± 3.82
DNMF90.96 ± 5.1290.54 ± 3.7684.25 ± 5.1091.31 ± 2.3190.32 ± 2.0985.37 ± 2.80
DSNMF90.12 ± 4.7088.52 ± 4.7084.77 ± 4.3590.68 ± 2.2689.07 ± 2.9785.39 ± 2.79
NMFAN83.31 ± 6.3183.00 ± 4.9279.69 ± 3.3381.99 ± 3.4680.60 ± 3.8277.84 ± 3.33
EWRNMF82.32 ± 5.0282.51 ± 5.1980.80 ± 5.4980.88 ± 3.6780.08 ± 3.6579.27 ± 3.61
SODNMF91.74 ± 3.2890.14 ± 3.7688.69 ± 4.9091.17 ± 1.9889.57 ± 2.1187.84 ± 3.32
DLRGNMF93.54 ± 3.1493.57 ± 3.2190.42 ± 4.0792.25 ± 1.8492.40 ± 1.8089.29 ± 2.12
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, Y.; Ma, Z.; Li, H.; Wang, J. Dual Space Latent Representation Learning for Image Representation. Mathematics 2023, 11, 2526. https://doi.org/10.3390/math11112526

AMA Style

Huang Y, Ma Z, Li H, Wang J. Dual Space Latent Representation Learning for Image Representation. Mathematics. 2023; 11(11):2526. https://doi.org/10.3390/math11112526

Chicago/Turabian Style

Huang, Yulei, Ziping Ma, Huirong Li, and Jingyu Wang. 2023. "Dual Space Latent Representation Learning for Image Representation" Mathematics 11, no. 11: 2526. https://doi.org/10.3390/math11112526

APA Style

Huang, Y., Ma, Z., Li, H., & Wang, J. (2023). Dual Space Latent Representation Learning for Image Representation. Mathematics, 11(11), 2526. https://doi.org/10.3390/math11112526

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop