Next Article in Journal
Navigating Cross-Border E-Commerce: Prioritizing Logistics Partners with Hybrid MCGDM
Previous Article in Journal
Brain Oscillations and Autonomic Synthonization via Comodulation in Collaborative Negotiation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Autoencoder-like Sparse Non-Negative Matrix Factorization with Structure Relationship Preservation

1
School of Statistics and Data Science, Lanzhou University of Finance and Economics, Lanzhou 730020, China
2
Key Laboratory of Digital Economy and Social Computing Science of Gansu, Lanzhou 730020, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(8), 875; https://doi.org/10.3390/e27080875
Submission received: 21 July 2025 / Revised: 12 August 2025 / Accepted: 17 August 2025 / Published: 19 August 2025
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

Clustering algorithms based on non-negative matrix factorization (NMF) have garnered significant attention in data mining due to their strong interpretability and computational simplicity. However, traditional NMF often struggles to effectively capture and preserve topological structure information between data during low-dimensional representation. Therefore, this paper proposes an autoencoder-like sparse non-negative matrix factorization with structure relationship preservation (ASNMF-SRP). Firstly, drawing on the principle of autoencoders, a “decoder-encoder” co-optimization matrix factorization framework is constructed to enhance the factorization stability and representation capability of the coefficient matrix. Then, a preference-adjusted random walk strategy is introduced to capture higher-order neighborhood relationships between samples, encoding multi-order topological structure information of the data through an optimal graph regularization term. Simultaneously, to mitigate the impact of noise and outliers, the l 2 , 1 -norm is used to constrain the feature correlation between low-dimensional representations and the original data, preserving feature relationships between data, and a sparse constraint is imposed on the coefficient matrix via the inner product. Finally, clustering experiments conducted on 8 public datasets demonstrate that ASNMF-SRP consistently exhibits favorable clustering performance.

1. Introduction

Clustering aims to reveal intrinsic relationships between data through similarity measurement, and it is widely applied in various fields such as marketing [1], gene expression [2], and pattern recognition [3]. However, with the continuous increase in data scale, traditional clustering algorithms face challenges posed by the “curse of dimensionality”, including increased computational complexity, feature redundancy, and amplified noise. Consequently, researchers have proposed a collaborative optimization paradigm of “dimensionality reduction followed by clustering”. Matrix factorization, as a crucial dimensionality reduction method, factorizes the original data matrix into low-rank submatrices to achieve the goal of “parts representing the whole”. Common matrix factorization methods include Principal Component Analysis (PCA) [4], Singular Value Decomposition (SVD) [5], and non-negative matrix factorization (NMF) [6]. Due to advantages such as non-negativity constraints, computational simplicity, and fast implementation, NMF is frequently used in clustering analysis.
In recent years, to further enhance the performance of NMF, researchers have proposed many improved NMF algorithms. For example, Hoyer et al. [7] proposed sparse non-negative matrix factorization (SNMF), utilizing l 1 -norm to improve the sparsity of NMF. Kong et al. [8] proposed robust non-negative matrix factorization (RNMF), replacing the Frobenius norm with the l 2 , 1 -norm to enhance the algorithm’s robustness. Ding et al. [9] proposed orthogonal non-negative matrix factorization (ONMF), reducing interference from redundant information by imposing orthogonality constraints on the factorized submatrices. These NMF algorithms primarily focus on the reconstruction error after matrix factorization but overlook the spatial structural relationships within the original data. Studies have shown that introducing graph regularization techniques into NMF can not only optimize reconstruction error but also further capture local geometric structures between data [10]. Cai et al. [11] proposed the graph regularized non-negative matrix factorization algorithm (GNMF), encoding topological information between data into the objective function to effectively improve clustering performance. Wu et al. [12] proposed robust manifold non-negative matrix factorization (MNMEL21), which considers manifold learning of data and employs the l2,1-norm to mitigate interference from noise and outliers. Li et al. [13] proposed graph regularized non-negative low-rank matrix factorization (GNLMF), enhancing clustering accuracy on image data by capturing low-rank structures. Liu et al. [14] proposed graph regularized discriminative non-negative matrix factorization (GDNMF), which incorporates local geometric structure information between samples and data label information to effectively improve clustering performance on image data. To enhance the quality of graph learning, Huang et al. [15] proposed non-negative matrix factorization with adaptive graph learning (NMFAN), dynamically adjusting graph structure learning to effectively improve algorithm performance. Ren et al. [16] proposed semi-supervised symmetric non-negative matrix factorization with graph quality improvement (S3NMFGC), improving clustering performance by dynamically generating and adaptively updating graph learning results. Mohammadi et al. [17] proposed a semi-supervised multi-view clustering method based on adaptive symmetric NMF (SSA-SNMF), which improves the performance of the algorithm by introducing a multi-constraint optimization strategy.
In the process of graph learning, the quality of the learned graph structure directly affects the accuracy and robustness of clustering results. Currently, most graph learning methods are based on measuring Euclidean distances between samples. As the complexity of data internal structures increases, some samples may exhibit distant geometric distances, yet their local neighborhood topologies might show highly similar characteristics. Therefore, with the development of Graph Convolutional Neural Networks (GCNNs) [18], higher-order topological relationships between samples have gained widespread attention among researchers. Wang et al. [19] proposed a robust high-order graph learning algorithm for multi-view clustering (RHGL), which improves clustering accuracy by learning high-order graphs. Zhan et al. [20] proposed a multi-view clustering method with optimal high-order graph embedding (Co-MSE), which enhances the quality of graph learning by modeling high-order correlations. Additionally, Wang et al. [21] indicated that autoencoder-based methods can not only reduce data dimensionality but also further learn latent information between data points. Thus, from the perspective of “decoder-encoder” collaborative optimization, although the aforementioned improved NMF-based algorithms enhance clustering performance through different regularization terms, their objective functions only include the “decoder” component. They lack an explicit “encoder” structure with constraints on the coefficient matrix, thus preventing the coefficient matrix from fully uncovering latent information in the raw data.
Research indicates that introducing sparsity constraints into NMF can effectively enhance the clustering performance of algorithms [7,22]. Meng et al. [23] proposed sparse and orthogonally constrained dual-graph regularized non-negative matrix factorization (SODNMF), which incorporates dual-graph regularization, sparsity constraints, and orthogonality constraints into the objective function, significantly improving the algorithm’s performance. Peng et al. [24] proposed logarithmic sparse non-negative matrix factorization (LS-NMF), designing a l log -(pseudo) norm to achieve sparse constraints. Xiong et al. [25] proposed dual-graph regularized sparse robust and adaptive non-negative matrix factorization (DRGSNMF), adopting the L 2 , p -norm to construct a sparse regularization term. Although the implementations of these algorithms differ, they all further enhance algorithm performance through sparse constraints.
Features often exhibit complex correlations or complementarity, such as in consumer behavior data. Therefore, in NMF, optimizing only the coefficient matrix may fail to effectively preserve structural relationships between features. To address this, Shang et al. [26] proposed dual-graph regularized non-negative matrix factorization (DNMF), which enhances the algorithm’s generalization capability by incorporating learning of the feature graph. Gu et al. [27] proposed a dual-graph regularized co-optimization clustering algorithm (DRCC), simultaneously considering the learning of both sample and feature graphs. Furthermore, to ensure the effective preservation of feature relationships from the original data matrix during dimensionality reduction, Hedjam et al. [28] proposed feature relationship-preserving non-negative matrix factorization (FR-NMF), and Salahian et al. [29] proposed a deep autoencoder non-negative matrix factorization with contrastive regularization and feature relationship preservation (DANMF-CRFR).
Therefore, this paper proposes an autoencoder-like sparse non-negative matrix factorization with structure relationship preservation (ASNMF-SRP). This method incorporates autoencoder principles into NMF, extending the constraints on the coefficient matrix from implicit to explicit. Through these dual constraints, it not only enhances the decomposition stability of the coefficient matrix but also further uncovers latent information in the raw data. Simultaneously, it preserves topological information in the original data space using higher-order graph regularization terms for both samples and features and incorporates an optimization term for feature relationship preservation. Finally, sparse optimization is applied to the coefficient matrix. Thus, ASNMF-SRP integrates autoencoder-like NMF, higher-order graph regularization, feature relationship preservation, and sparse constraints into a unified optimization framework.
The main contributions of the ASNMF-SRP algorithm are as follows:
(1) Optimization of the coefficient matrix. Firstly, this paper introduces an autoencoder-like NMF, adding an explicit “encoder” structure to constrain the coefficient matrix. Then, in the local structure learning process, a higher-order graph regularization method is proposed, enabling the algorithm to progressively extend from first-order to higher-order graph learning. Finally, an inner-product-based sparse constraint is introduced, preserving high-order collaborative relationships between data at the element level.
(2) Optimization of the basis matrix. This paper proposes a robust feature relationship preservation method, thereby establishing a topology-preserving relationship between the low-dimensional embedding space and the original high-dimensional data feature structure.
(3) This paper demonstrates the feasibility of ASNMF-SRP through convergence analysis, time complexity analysis, and other aspects. Simultaneously, in the experimental section, comparative analysis of clustering results, visualization analysis of low-dimensional representation data, parameter sensitivity analysis, and other experiments are conducted. Ultimately, the experimental results show that ASNMF-SRP exhibits excellent clustering performance.
The remaining work in this paper is organized as follows: Section 2 introduces some foundational work related to ASNMF-SRP. Section 3 primarily describes the derivation of the ASNMF-SRP algorithm, including the convergence and time complexity of submatrix update iterations. Section 4 presents a series of experiments conducted on eight public datasets. Section 5 provides a summary of the ASNMF-SRP algorithm.

2. Related Work

2.1. Non-Negative Matrix Factorization (NMF)

The objective of NMF is to factorize a given non-negative data matrix X = x 1 , x 2 , , x n + m × n into the product of two non-negative and low-rank matrices: a basis matrix U + m × r and a coefficient matrix V + r × n , i.e., X U V , where r min m , n . During the factorization process, NMF optimizes the approximate representation by minimizing the reconstruction error. Therefore, it employs the Frobenius norm to define the objective function as follows:
min X U V F 2 , s . t . U 0 , V 0
Here, F denotes the Frobenius norm of a matrix, which measures the reconstruction error after matrix factorization. For the optimization problem in Equation (1), Lee et al. [30] proposed an iterative algorithm based on multiplicative update rules, with the update formulas given by the following:
U i j U i j X V T i j U V V T i j , V i j V i j U T X i j U T U V i j

2.2. Robust Non-Negative Matrix Factorization (RNMF)

NMF measures reconstruction error based on Euclidean distance, making it highly susceptible to noise and outliers. To further enhance the robustness of the matrix factorization process, Kong et al. [8] proposed robust non-negative matrix factorization (RNMF), which employs the l 2 , 1 -norm to assess the quality of data reconstruction. The objective function of RNMF is as follows:
min X U V 2 , 1 , s . t . U 0 , V 0
The update formulas for the basis matrix U and coefficient matrix V in Equation (3) are as follows:
U i j U i j X G 1 V T i j U V G 1 V T i j , V i j V i j U T X G 1 i j U T U V G 1 i j
Here, G1 is a diagonal matrix with diagonal elements given by the following:
G 1 j j = 1 / X U V i 2

2.3. Graph Regularized Non-Negative Matrix Factorization (GNMF)

The standard NMF achieves the goal of “representing the whole by parts” by minimizing the reconstruction error, but this approach overlooks the latent structural information in the data. Therefore, Cai et al. [11] proposed graph regularized non-negative matrix factorization (GNMF). GNMF improves upon NMF by encoding the local geometric structure among samples into the objective function for optimization, thereby enhancing the representation quality of samples in the low-dimensional space. The objective function of GNMF is as follows:
min X U V F 2 + λ Tr V L V T , s . t . U 0 , V 0
Here, λ is a non-negative regularization parameter that adjusts the strength of graph learning. L is the Laplacian matrix, which satisfies L = D W . In this context, W is the similarity matrix reflecting the relationships between samples, and the degree matrix D = diag ( D 11 , D 22 , , D n n ) is a diagonal matrix formed by the row sums of W , i.e., D i i = j W i j . The update rules for GNMF are as follows:
U i j U i j X V T i j U V V T i j , V i j V i j U T X + λ W V i j U T U V + λ D V i j

3. Methodology

3.1. Autoencoder-like Non-Negative Matrix Factorization

In clustering analysis, the coefficient matrix V of NMF serves as the clustering indicator matrix. Therefore, the coefficient matrix V needs to effectively encode the structural features of the original data to ensure the accuracy of subsequent clustering. Inspired by the non-negative symmetric encoder-decoder method proposed by Sun et al. [31] in community detection and the principle of autoencoders, this paper imposes an explicit constraint on the coefficient matrix V , i.e., V f X = U T X . This approach not only improves the factorization stability of the coefficient matrix but also further uncovers the latent representations of the original data. The specific implementation process is as follows:
Linear Decoder:
g v = U v X g V = U V X U V F 2
Linear Encoder:
f x = U T x V f X = U T X V U T X F 2
In the “decoder-encoder” architecture, the coefficient matrix V acts as the encoder matrix. From Equations (8) and (9), it can be seen that the “decoder” part corresponds to standard NMF, aiming to minimize the reconstruction error between data, i.e., maximizing the approximation X U V . The “encoder” part, on the other hand, transforms the data matrix X into a distributed representation using the basis matrix U + m × r , thereby imposing an explicit constraint on the coefficient matrix, i.e., V U T X .
By combining Equations (8) and (9), an autoencoder-like non-negative matrix factorization framework is obtained, with the following objective function:
min | | X U V | | F 2 d e c o d e r + | | V U T X | | F 2 e n c o d e r , s . t . U 0 , V 0
Here, the decoder and encoder parts share the same basis matrix U , thereby forming a symmetry constraint. Figure 1 is a schematic diagram of Equation (10).

3.2. High-Order Graph Regularization

To further improve the quality of the data representation in the low-dimensional space, this paper introduces graph regularization techniques based on Equation (10) to learn the geometric structural information of the original high-dimensional space. Therefore, this paper captures the neighborhood relationships between samples by constructing a nearest neighbor graph. The first-order similarity matrix W 1 is defined as follows:
W 1 i j = 1 , x i N k x j or x j N k x i 0 , o t h e r w i s e
Here, N k x i denotes the set of k -nearest data points to sample x i . x i and x j are the i -th and j -th column samples of the original data matrix X = x 1 , x 2 , , x n + m × n , respectively.
The coefficient matrix V = v 1 , v 2 , , v n + r × n is the low-dimensional representation to be optimized. According to manifold learning and spectral graph theory, the smoothness of this low-dimensional representation is given by the following:
Z = 1 2 i , j = 1 n v i v j 2 W 1 i j = i = 1 n v i v i T D 1 i i i , j = 1 n v i v j T W 1 i j = Tr V D 1 V T Tr V W 1 V T = Tr V L 1 V T
Here, D 1 is a diagonal matrix with D 1 i i = j W i j . L 1 is defined as the first-order Laplacian matrix, which satisfies the equation L 1 = D 1 W 1 .
In real-world scenarios, samples often exhibit complex structural relationships. For instance, in online shopping, consumers from different regions may have no direct connection. However, by analyzing behavioral data, it is possible to identify neighborhood relationships where consumers share highly similar purchasing patterns. Such neighborhood relationships provide a scientific basis for businesses to segment target customer groups and implement precision marketing strategies. As shown in Figure 2, samples x i and x l are both connected to samples 1, 2, and 3, indicating that x i and x l share the same neighborhood structure. However, x i and x l are far apart and not within each other’s neighborhood sets. From Equation (11), it follows that W i l = 0 . Therefore, the first-order similarity matrix can only capture the neighborhood relationships between samples but fails to consider the similarity of their neighborhood structures. To further capture higher-order similarities between different neighborhood systems, this paper introduces a preference-adjusted random walk strategy. The random walk strategy [32,33] constructs sequences of sample points and encodes multi-step reachability into a higher-order similarity matrix, thereby reflecting higher-order structural information among samples. During the random walk process, the transition probability matrix P i j represents the probability of transitioning from sample x i to sample x j , defined as follows:
P i j = a i j k i
Here, a i j indicates the connection state between samples x i and x j : if x i and x j are connected, then a i j = 1 ; otherwise, a i j = 0 . k i represents the degree of sample x i . Since a i j takes values of 0 or 1, k i here denotes the number of samples connected to x i .
The transition probability matrix in Equation (13) is based on an equal-probability assumption, which often fails to accurately reflect the topological heterogeneity among samples when processing high-dimensional complex data. Therefore, this paper incorporates the Preferential Attachment (PA) model into Equation (13). By assigning connection preference weights between sample points, the transition probability matrix is further optimized to more realistically reflect higher-order neighborhood relationships among samples. Thus, the new expression is as follows:
P i j = P A s i m S c o r e M a x P A s i m S c o r e × a i j k i
Therefore, the d -th order similarity between samples x i and x j is as follows:
S i j d = P i k × P k l × × P z j
where d = 1 , 2 , , n . In practical applications, the value of d should not be too large, as an excessively large d may introduce noise interference, thereby affecting the quality of graph learning. The average of the d -th order adjacency matrices is as follows:
W 2 = S 1 + S 2 + + S d / d
In Equation (16), W 2 is referred to as the second-order similarity matrix, which aims to capture higher-order neighborhood relationships between samples. The second-order Laplacian matrix is defined as L 2 = D 2 W 2 , where D 2 is a diagonal matrix with entries D 2 i i = j W i j .
By considering both the first-order and second-order Laplacian matrices, the optimal Laplacian matrix is now defined as follows:
L = ω 1 L 1 + ω 2 L 2
where w 1 and w 2 are balancing parameters. Consequently, the optimal graph regularization term is given by the following:
min V 0 Tr V L V T

3.3. Feature Relationship Preservation

As data structures increase in complexity, feature information also plays a crucial role in clustering analysis. For example, when enterprises identify target customer groups, consumer characteristics, such as purchase frequency ( x 1 * ), interests ( x 2 * ), spending power ( x 3 * ), shopping preferences ( x 4 * ), and brand loyalty ( x 5 * ), are key dimensions. Therefore, when high-dimensional data are transformed into low-dimensional representations, preserving the intrinsic relationships between features becomes essential for effectively uncovering the latent structure of customer groups. This is illustrated in Figure 3.
In Figure 3, x i * represents the i-th row vector of matrix X , corresponding to the i-th feature information in the original high-dimensional data, where i = 1 , 2 , 3 , 4 , 5 . u i * represents the i -th row vector of the basis matrix U . As shown in Figure 3, the feature relationship between x i * and u i * remains consistent, which provides a theoretical basis for enterprises to identify target customer groups. Therefore, incorporating the feature relationships between data into the objective function for optimization can further preserve the feature relationships among the data.
X X T i j = l = 1 n x i l x j l represents the inner product between the i -th and j -th features of X . U U T i j = l = 1 n u i l u j l represents the inner product between the i -th and j -th rows of the basis matrix U . To maintain consistency between feature relationships in the low-dimensional space and original data, we construct the feature preservation constraint term:
min U 0 X X T β   U U T F 2
where β is a non-negative balance parameter.
Equation (19) defines the reconstruction error based on the Frobenius norm. Therefore, when seeking to minimize the objective function value, Equation (19) is susceptible to interference from noise and outliers. To address this, this paper employs the l 2 , 1 -norm to further optimize Equation (19), thereby enhancing its robustness. The specific expression is as follows:
min U 0 X X T β U U T 2 , 1

3.4. Sparsity of Coefficient Matrix

In the clustering process, the coefficient matrix V plays the role of a cluster indicator matrix. Therefore, the state of the coefficient matrix V affects the accuracy of clustering results. Imposing sparse constraints on the coefficient matrix V can make the representation of each sample in the low-dimensional space significantly correlated with only a few basis vectors, thereby weakening the interference of irrelevant information on clustering results. Common sparsity methods include: l 0 -norm, l 1 / 2 -norm, l 1 -norm, l 2 , 1 -norm, l log -(pseudo) norm, etc. A detailed introduction is shown in Table 1, where v i j represents the element in the i -th row and j -th column of the coefficient matrix V + r × n .
The five common sparsity methods in Table 1 primarily impose element-wise regularization on the coefficient matrix V but overlook the correlations within the internal structure of V . This element-level separate optimization struggles to capture potential high-order collaborative relationships in the sample space. Therefore, this paper proposes a sparse constraint based on inner product penalty.
For the coefficient matrix V = v 1 , v 2 , , v n + r × n , we have
V T V = v 1 , v 1 v 1 , v 2 v 1 , v n v 2 , v 1 v 2 , v 2 v 2 , v n v n , v 1 v n , v 2 v n , v n
In Equation (21), the diagonal elements of V T V reflect autocorrelation, while the off-diagonal elements reflect correlations between different vectors. Here, to avoid interference from autocorrelation, we remove the diagonal elements based on Equation (21) and only retain the off-diagonal elements for sparsification. Thus, we can further obtain the following:
i = 1 n j = 1 n v i , v j i = 1 n v i , v i = i = 1 n j = 1 j i n v i , v j = Tr V T V 1 n Tr V T V
where 1 n denotes an n × n matrix with all-one element.
Therefore, the sparse constraint term for the coefficient matrix V is as follows:
min V 0 Tr V T V 1 n Tr V T V

3.5. Objective Function

From Equations (10), (18), (20) and (23), the objective function of ASNMF-SRP is derived as follows:
min U , V 0 X U V F 2 + V U T X F 2 + α Tr V L V T + X X T β U U T 2 , 1   + λ Tr V T V 1 n Tr V T V
where α , β , and λ are all non-negative regularization parameters. Figure 4 is a schematic diagram of the ASNMF-SRP algorithm.
In the objective function Equation (24), each term plays an important role. The first term ( X U V F 2 ) aims to focus on the reconstruction error of matrix factorization. The second term ( V U T X F 2 ) represents an explicit constraint on the coefficient matrix, thereby improving the factorization stability of the coefficient matrix. The third term ( Tr V L V T ) encodes higher-order topological information between samples by constructing the optimal Laplacian matrix. The fourth term ( X X T β U U T 2 , 1 ) ensures that the feature relationship structure of the original high-dimensional data can be preserved in the low-dimensional space after dimensionality reduction. The fifth term ( Tr V T V 1 n Tr V T V ) imposes a structured sparse constraint on the coefficient matrix by introducing a sparse regularization term in the form of inner product.

3.6. Optimization Algorithm

ASNMF-SRP employs multiplicative update rules to iteratively optimize the basis matrix U and coefficient matrix V . The augmented Lagrangian function of the objective function Equation (24) is as follows:
P U , V = X U V F 2 + V U T X F 2 + α Tr V L V T + X X T β U U T 2 , 1 + λ Tr V T V 1 n Tr V T V Tr Λ 1 U T Tr Λ 2 V T
where Λ 1 + m × r and Λ 2 + r × n denote the Lagrange multipliers for U and V , respectively.
Expanding Equation (25) and omitting terms independent of U and V , we obtain the following:
L U , V = Tr V T U T U V 2 V T U T X + Tr V T V 2 V T U T X + X T U U T X + α Tr V L V T + Tr β 2 U U T G 2 U U T 2 β U U T G 2 X X T + λ Tr V T V 1 n Tr V T V Tr Λ 1 U T Tr Λ 2 V T
where G 2 is a diagonal matrix whose diagonal elements are as follows:
G 2 j j = 1 / X X T β U U T i 2
Taking partial derivatives of L U , V with respect to U and V , respectively, gives the following:
L U = 2 U V V T 4 X V T + 2 X X T U + 2 β 2 U U T G 2 U + 2 β 2 G 2 U U T U 2 β X X T G 2 U 2 β G 2 X X T U Λ 1
L V = 2 U T U V 4 U T X + 2 V + 2 α V L + 2 λ V 1 n 2 λ V Λ 2
To ensure the non-negativity of the coefficient matrix V , the optimal Laplacian matrix L in Equation (29) is factorized as L = L + L , where L + = ( | L | + L ) / 2 and L = ( | L | L ) / 2 .
Equation (29) can be rewritten as follows:
L V = 2 U T U V 4 U T X + 2 V + 2 α V L + 2 α V L + 2 λ V 1 n 2 λ V Λ 2
According to the KKT conditions Λ 1 i j U i j = 0 and Λ 2 i j V i j = 0 , we derive the following:
( 2 U V V T 4 X V T + 2 X X T U + 2 β 2 U U T G 2 U + 2 β 2 G 2 U U T U 2 β X X T G 2 U 2 β G 2 X X T U Λ 1 ) i j U i j = 0
2 U T U V 4 U T X + 2 V + 2 α V L + 2 α V L + 2 λ V 1 n 2 λ V Λ 2 i j V i j = 0
Therefore, we obtain the following:
( 2 U V V T 4 X V T + 2 X X T U + 2 β 2 U U T G 2 U + 2 β 2 G 2 U U T U 2 β X X T G 2 U 2 β G 2 X X T U Λ 1 ) i j U i j 2 = 0
2 U T U V 4 U T X + 2 V + 2 α V L + 2 α V L + 2 λ V 1 n 2 λ V Λ 2 i j V i j 2 = 0
From Equations (33) and (34), the update formulas for U and V are as follows:
U i j t + 1 U i j t 2 X V T + β X X T G 2 U + β G 2 X X T U i j U V V T + X X T U + β 2 U U T G 2 U + β 2 G 2 U U T U i j
V i j t + 1 V i j t 2 U T X + α V L + λ V i j U T U V + V + α V L + + λ V 1 n i j

3.7. Convergence Analysis

In this section, the convergence of U and V under the update rules in Equations (35) and (36) in the ASNMF-SRP algorithm is proven using the auxiliary function method.
Firstly, we prove that the coefficient matrix V is non-increasing under the update rule in Equation (36). According to the objective function Equation (24), we can derive a functional F i j V i j regarding the coefficient matrix V as follows:
F i j ( V i j ) = Tr ( V T U T U V 4 V T U T X + V T V + α V L + V T α V L V T + λ V T V 1 n λ V T V )
Proposition 1.
The constructed function  G V i j , V i j t  is an auxiliary function of  F i j V i j .
G V i j , V i j t = i j U T U V t i j V i j 2 V i j t 4 i j U T X i j V i j t 1 + log V i j V i j t + i j V i j 2 + α i j V t L + i j V i j 2 V i j t α i j l L j l V i j t V i l t 1 + log V j l V i l V j l t V i l t + λ i j V t 1 n i j V i j 2 V i j t λ i j V i j t V i j t 1 + log V i j V i j V i j t V i j t
Proof. 
When V i j = V i j t , we have G V i j , V i j = F i j V i j . The following proves the case G V i j , V i j t F i j V i j .
When A and B are symmetric matrices, we have the following:
i = 1 n j = 1 m A S t B i j S i j 2 S i j t Tr S T A S B , A + n × n , B + m × m , S t + n × m , S + n × m
From Equation (39), we obtain the following:
Tr V T U T U V i j U T U V t i j V i j 2 V i j t
Tr V T V i j V i j 2
α Tr V L + V T α i j V t L + i j V i j 2 V i j t
λ Tr V T V 1 n λ i j V t 1 n i j V i j 2 V i j t
From x > 1 + log x x > 0 , we obtain the following:
4 Tr V T U T X   = 4 Tr X T U V 4 i j U T X i j V i j t 1 + log V i j V i j t
α Tr V L V T α i j l L j l V i j t V i l t 1 + log V j l V i l V j l t V i l t
λ Tr V T V λ i j V i j t V i j t 1 + log V i j V i j V i j t V i j t
From Equations (40)–(46), it follows that G V i j , V i j t F i j V i j . Therefore, G V i j , V i j t is an auxiliary function of F i j V i j , and Proposition 1 is proven.
Let G V , V t V i j = 0 , we obtain the following:
V i j t + 1 V i j t 2 U T X + α V L + λ V i j U T U V + V + α V L + + λ V 1 n i j
Thus, the coefficient matrix V in the ASNMF-SRP algorithm is non-increasing under the update rule in Equation (36). Similarly, it can be proven that the basis matrix U is non-increasing under the update rule in Equation (35). Therefore, U and V converge under the update rules in Equations (35) and (36). The specific implementation of the ASNMF-SRP algorithm is described in Algorithm 1.
Algorithm 1 ASNMF-SRP
Input: Initial matrix X = x 1 , x 2 , , x n + m × n , number of classes r , neighborhood parameter k , regularization parameters α , β and λ , balance parameters w 1 and w 2 parameter d , threshold ε , maximum iterations t .
Output: Basis matrix U and coefficient matrix V .
1. Initialization: t = 0 , Randomly generate basis matrix U m × r and coefficient matrix V r × n ;
2. Obtain optimal Laplacian matrix L according to Equations (11)–(17);
3. For t = 1 , 2 , 3 , , maxIter
4.  U i j t + 1 U i j t 2 X V T + β X X T G 2 U + β G 2 X X T U i j U V V T + X X T U + β 2 U U T G 2 U + β 2 G 2 U U T U i j ;
5.  V i j t + 1 V i j t 2 U T X + α V L + λ V i j U T U V + V + α V L + + λ V 1 n i j ;
6. if U t U t 1 ε and V t V t 1 ε
Break and return ( U , V );
7. End if
8. End for

3.8. Time Complexity Analysis

When inputting a data matrix X m × n , where m is the feature dimension of samples and n is the number of samples, assume the number of classes for this dataset is r , and r min m , n . After t iterations, the time complexity of ASNMF-SRP is as follows:
(1) The autoencoder-like NMF part: complexities for updating U and V are O t m 2 r + t m r 2 and O t m 2 r + t m r 2 ;
(2) The higher-order graph regularization part: complexity is O m n 2 + m e d , where e denotes the average degree of nodes;
(3) The feature preservation part: complexity is O t m 2 r + t m 2 + m ;
(4) The sparse constraint term: complexity is O t n 2 r .
Therefore, the overall time complexity of ASNMF-SRP is O t m 2 r + t m 2 + m + t n 2 r + m n 2 + m e d .

4. Experiments

In this section, clustering experiments will be conducted on 8 public datasets to evaluate the clustering performance of the ASNMF-SRP algorithm. The experiments are implemented in Python 3.11 with the computer environment being Intel(R) Core (TM) i5-1135G7 @ 2.40 GHz 2.42 GHz, 16 GB RAM, and Windows 11 64-bit operating system.

4.1. Dataset

To verify the effectiveness of the algorithm, we selected 8 public datasets for comparative experiments. MSRA25 [37] dataset contains 1799 face images from 12 individuals. Semeion, Krvs, PenDigits, and Vehicle datasets are from the UCI Machine Learning Repository (https://archive.ics.uci.edu (accessed on 23 March 2025)). Hitech dataset is from the CLUTO toolkit (https://conservancy.umn.edu/items/4fbef165-f964-41ed-a239-86a8f931ffbe (accessed on 26 March 2025)). COIL20 (http://www.cs.columbia.edu/CAVE/software/softlib/coil-20.php (accessed on 15 December 2024)) and COIL100 (http://www.kaggle.com/jessicali9530/coil100/downloads/coil100.zip/2 (accessed on 15 December 2024)) datasets record images of 20 and 100 objects, respectively. The basic information of these 8 datasets is shown in Table 2.

4.2. Clustering Performance Evaluation Metrics

To compare the clustering performance of various algorithms, this paper selects four commonly used clustering performance evaluation metrics. Among them, larger values for these four clustering evaluation metrics indicate better clustering performance of the corresponding algorithm.

4.2.1. Clustering Accuracy (ACC)

ACC = i = 1 n δ m a p s i , r i n
where n is the total number of samples in the input data, r i is the true class label of the data, s i represents the result after clustering by the algorithm, and m a p denotes the mapping function. The expression δ m a p s i , r i in Equation (48) is as follows:
δ m a p s i , r i = 1 , m a p s i = r i 0 , m a p s i r i
From Equations (48) and (49), ACC is an evaluation metric used to measure the consistency between clustering results and true labels, with a value range of 0 , 1 .

4.2.2. Adjusted Rand Index (ARI)

ARI = i j n i j 2 i x i 2 j y i 2 / n 2 0.5 i x i 2 + j y i 2 i x i 2 j y i 2 / n 2
where x i represents the total number of samples in the i -th cluster after clustering, y i represents the total number of samples in the true j -th class, and n i j denotes the total number of samples common to both the cluster x i and the true class y j . From Equation (50), the value range of ARI is 1 , 1 .

4.2.3. Normalized Mutual Information (NMI)

NMI N , N * = MI N , N * max H N , H N *
where N * represents the data labels obtained through clustering, N represents the actual data labels, and H is defined as the entropy function.
The mutual information MI expression is as follows:
MI N , N * = n i N , n j * N * p n i , n j * log p n i , n j * p n i p n j *
where p n i and p n j * represent the marginal probabilities of the true class and clustering results, respectively, and p n i , n j * is the joint probability of the true class n i and cluster n j * . From Equations (51) and (52), the value range of NMI is 0 , 1 .

4.2.4. Clustering Purity (PUR)

PUR = i = 1 r max j n i j n
where n i represents the total number of samples, r is the number of clusters, and n i j represents the number of samples in cluster i that belong to the true class j . From Equation (53), the value range of PUR is 0 , 1 .

4.3. Comparison Algorithms and Parameter Settings

The detailed descriptions of the 9 algorithms participating in the comparative experiments are as follows:
(1) NMF [6] factorizes the original non-negative matrix into the product of two non-negative low-rank submatrices to achieve the goal of representing the whole by parts.
(2) ONMF [9] imposes orthogonality constraints on the factorized submatrices based on NMF.
(3) Hx-NMF [38] combines logarithm with NMF to improve the robustness of the algorithm.
(4) EMMF [39] is an algorithm based on entropy minimization matrix factorization.
(5) GNMF [11] encodes the local geometric structure between data into the objective function to improve the quality of low-dimensional representations.
(6) RMMMF [40] uses the l 2 , 1 -norm to enhance the algorithm’s robustness.
(7) DRCC [27] jointly optimizes the local geometric structures of samples and features by incorporating them into the objective function.
(8) FR-NMF [28] improves the clustering performance of the algorithm by introducing a feature preservation term.
(9) LS-NMF [24] imposes sparse constraints on the basis matrix and coefficient matrix using the l l o g -(pseudo) norm.
In the experimental process, we perform a grid search over the algorithm’s regularization parameter in the range 1 0 4 , 1 0 3 , 1 0 2 , 1 0 1 , 1 0 0 , 1 0 1 , 1 0 2 , 1 0 3 .
During the experiments, the parameter values for the higher-order graph regularization parameter α , feature relationship preservation regularization parameter β , and sparse constraint regularization parameter λ in the ASNMF-SRP algorithm are shown in Table 3.

4.4. Results and Analysis

Due to fluctuations in the clustering results of various algorithms, to obtain representative clustering outcomes, each algorithm was executed 20 times on the 8 datasets during the experiments. The mean and standard deviation were calculated as the final comparative clustering results. Table 4, Table 5, Table 6 and Table 7 present the specific values of ACC, ARI, NMI, and PUR for these 10 algorithms on the 8 datasets, respectively. In the tables, bold numbers indicate the optimal clustering results on that dataset, and the I-P (Improvement Percentage) column shows the percentage by which the ASNMF-SRP algorithm improves over the best result among the other 9 algorithms.
By comparing the data in Table 4, Table 5, Table 6 and Table 7, it can be seen that ASNMF-SRP exhibits excellent clustering performance. The specific data analysis is as follows:
(1) On image datasets (MSRA25, Semeion, COIL20, and COIL100), the values of all four clustering evaluation metrics for ASNMF-SRP are higher than those of the other nine algorithms. This indicates that ASNMF-SRP holds certain advantages in handling clustering problems related to images.
(2) On non-image datasets (Krvs, Hitech, PenDigits, and Vehicle), ASNMF-SRP fails to achieve optimal values in ACC on the Hitech dataset and in ARI and NMI on the Vehicle dataset. However, in most cases, ASNMF-SRP demonstrates optimal performance.
(3) FR-NMF incorporates an additional feature relationship preservation term compared to NMF. According to the data in Table 4, Table 5, Table 6 and Table 7, FR-NMF outperforms NMF in clustering performance on some datasets. ASNMF-SRP not only includes feature relationship preservation but also incorporates components such as autoencoder-like NMF, higher-order graph regularization, and sparse constraints. The data in Table 4, Table 5, Table 6 and Table 7 show that ASNMF-SRP achieves higher values across all four clustering evaluation metrics on these eight datasets compared to NMF and FR-NMF. This demonstrates that through the synergistic effects of various regularization terms, ASNMF-SRP not only enhances clustering performance but also improves generalization capability.
This paper employs the t-SNE method to project the low-dimensional representation matrices obtained by the ten algorithms on the MSRA25, COIL20, and Hitech datasets into a two-dimensional space for visualization.
From Figure 5, Figure 6 and Figure 7, it can be observed that on the MSRA25, COIL20, and Hitech datasets, the low-rank representation matrix obtained after dimensionality reduction using the ASNMF-SRP algorithm exhibits stronger discriminative power. Therefore, the ASNMF-SRP algorithm demonstrates excellent clustering performance.

4.5. Analysis of the Impact of Autoencoder-like NMF on Clustering Performance

Unlike traditional NMF, the objective function of ASNMF-SRP includes a “decoder- encoder” module. To investigate the influence of this autoencoder-like NMF architecture on clustering performance, we removed the “encoder” part in Equation (24), causing ASNMF-SRP to degenerate into ASNMF-SRP-1. The objective function of ASNMF-SRP-1 is as follows:
min U , V 0 X U V F 2 + α Tr V L V T + X X T β U U T 2 , 1 + λ Tr V T V 1 n Tr V T V
From Table 8, it can be seen that the clustering performance of ASNMF-SRP is significantly better than that of ASNMF-SRP-1. On the clustering evaluation metrics ACC, ARI, NMI, and PUR, it shows average improvements of 12.03%, 17.02%, 10.81%, and 10.56%, respectively. This indicates that introducing an NMF component with a structure similar to an autoencoder has a positive effect on ASNMF-SRP, effectively enhancing its clustering performance.

4.6. Analysis of the Impact of Higher-Order Graph Regularization on Clustering Performance

To further examine the influence of higher-order neighborhood relationships among samples on clustering performance, we set w 2 = 0 in Equation (17), thereby reducing ASNMF-SRP to ASNMF-SRP-2. At this point, the objective function of ASNMF-SRP-2 is as follows:
min U , V 0 X U V F 2 + V U T X F 2 + α Tr V L 1 V T + X X T β U U T 2 , 1 + λ Tr V T V 1 n Tr V T V
From Table 9, it can be observed that ASNMF-SRP exhibits superior clustering performance over ASNMF-SRP-2 on most datasets, indicating its stronger generalization ability. This also suggests that higher-order neighborhood relationships among samples can influence the clustering performance of the algorithm. However, on a few datasets (e.g., Krvs), ASNMF-SRP does not perform well. Therefore, future work should further explore how to design an optimal Laplacian matrix to effectively encode multi-order topological structure information, thereby improving the algorithm’s universality.

4.7. Robustness Analysis of ASNMF-SRP

To evaluate the robustness of ASNMF-SRP, we added salt-and-pepper noise with intensities of 10%, 20%, 30%, and 40% to the MSRA25, COIL20, and COIL100 datasets. To further compare the contributions of Equations (19) and (20) to the robustness of ASNMF-SRP, we conducted comparative experiments between ASNMF-SRP and ASNMF-SRP-3. The objective function of ASNMF-SRP-3 is as follows:
min U , V 0 X U V F 2 + V U T X F 2 + α Tr V L V T + X X T β U U T F 2 + λ Tr V T V 1 n Tr V T V
From Figure 8, it can be seen that as the noise intensity increases, the clustering evaluation metric values of both ASNMF-SRP and ASNMF-SRP-3 show a declining trend. However, the clustering evaluation metric values of ASNMF-SRP are consistently higher than those of ASNMF-SRP-3. Therefore, adopting the l 2 , 1 -norm to measure reconstruction errors in the “feature relationship preservation” component can enhance the robustness of ASNMF-SRP.

4.8. Parameter Sensitivity Analysis

ASNMF-SRP has three hyperparameters: the higher-order graph regularization coefficient α , the balance parameter for feature relationship preservation β , and the regularization parameter for sparse constraint λ . In the parameter sensitivity analysis experiment, the value ranges for α and β were set to { 10 0 , 10 1 , 10 2 , 10 3 } , and the range for λ was { 10 4 , 10 3 } . The parameter sensitivity concerning α , β , and λ is shown in Figure 9.
From Figure 9, we can obtain the following:
(1) On the MSRA25, COIL20, COIL100, and Hitech datasets, ASNMF-SRP demonstrates favorable clustering performance when the higher-order regularization parameter α = 1000 ; on the Semeion and Vehicle datasets, ASNMF-SRP exhibits good clustering performance when parameter α takes smaller values.
(2) With parameters α and β determined, different values of the sparse constraint regularization parameter λ significantly affect the clustering performance of the ASNMF-SRP algorithm. For example, on the COIL100 dataset, when λ = 0.0001 , the clustering results of ASNMF-SRP overall outperform those when λ = 0.001 .
(3) α and λ jointly constrain the coefficient matrix V , where α preserves the manifold structure of samples while λ balances the strength of sparsity, both optimizing the coefficient matrix through synergistic effects. Simultaneously, α and β collaboratively optimize the local geometric structure of data in both sample space and feature space. In parameter settings, the values of α and β are generally greater than that of λ to enhance manifold structure preservation capability, while the value of λ is kept relatively small to ensure appropriate sparse constraint intensity.

4.9. Empirical Convergence

This section will experimentally analyze the convergence of the ASNMF-SRP algorithm on 8 datasets to further verify the accuracy of the theoretical analysis in Section 3.7. In the empirical convergence experiment, the values of hyperparameters α , β , and λ in ASNMF-SRP remain consistent with those in Table 3.
From Figure 10, it can be observed that the ASNMF-SRP algorithm achieves convergence on all eight datasets. Specifically, on the MSRA25, Semeion, COIL20, COIL100, Krvs, Hitech, and PenDigits datasets, the ASNMF-SRP algorithm exhibits a relatively fast convergence rate during the initial iterations, with its corresponding objective function value rapidly decreasing, ultimately achieving convergence within 20 iterations. On the Vehicle dataset, the ASNMF-SRP algorithm shows a rapid decline in the objective function value during the initial iterations, but subsequently, the rate of decrease slows down, ultimately achieving convergence around 40 iterations.

5. Conclusions

This paper proposes an autoencoder-like sparse non-negative matrix factorization with structure relationship preservation (ASNMF-SRP). By integrating the principle of autoencoders, the algorithm enhances the optimization of the coefficient matrix through a linear “decoder-encoder” approach, which not only improves the factorization stability of the coefficient matrix but also further uncovers the latent representations of the original data. In the structural learning of the sample space, ASNMF-SRP effectively captures higher-order topological information between samples by constructing the optimal Laplacian matrix. In feature relationship learning, ASNMF-SRP employs the l 2 , 1 -norm to define the reconstruction error of feature correlations between the basis matrix and the original data matrix, ensuring consistency in feature relationships between the low-dimensional space and the original high-dimensional space. Furthermore, a sparse constraint based on inner product representation is imposed on the coefficient matrix, further enhancing the clustering performance of ASNMF-SRP. Finally, comparative experiments between ASNMF-SRP and nine other advanced clustering algorithms demonstrate, as evidenced by the data in Table 4, Table 5, Table 6 and Table 7, that ASNMF-SRP achieves favorable clustering performance. In future research work, we will conduct in-depth exploration from the following three aspects: (1) To further optimize the construction process of high-order graph regularization, enhancing the model’s capability to represent complex structures; (2) to extend ASNMF-SRP into a deep NMF framework, enabling better exploration of latent features in data; and (3) to design a multi-view extension model for ASNMF-SRP, thereby broadening the applicability of the algorithm.

Author Contributions

Conceptualization, L.Z. and H.G.; methodology, L.Z.; software, L.Z.; validation, L.Z. and H.G.; formal analysis, L.Z. and H.G.; investigation, L.Z. and H.G.; resources, L.Z. and H.G.; data curation, L.Z.; writing—original draft preparation, L.Z.; writing—review and editing, L.Z. and H.G.; visualization, L.Z.; supervision, H.G.; project administration, L.Z. and H.G.; funding acquisition, H.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Social Science Fund of China (NO. 21BTJ042), the Gansu Provincial Natural Science Foundation (NO. 23JRRA1186), and the Gansu Provincial Universities’ Young Doctor Support Program (NO. 2025QB-058).

Data Availability Statement

Publicly available datasets were analyzed in this study.

Acknowledgments

The authors sincerely appreciate the editors and reviewers for their valuable comments and professional suggestions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kim, S.; Shin, W.; Kim, H.W. Predicting online customer purchase: The integration of customer characteristics and browsing patterns. Decis. Support Syst. 2024, 177, 114105. [Google Scholar] [CrossRef]
  2. Zeng, Y.; Chen, J.; Pan, Z.; Yu, W.; Yang, Y. Integrating single-cell multi-omics data through self-supervised clustering. Appl. Soft Comput. 2025, 169, 112541. [Google Scholar] [CrossRef]
  3. Mardani, K.; Maghooli, K.; Farokhi, F. Segmentation of coronary arteries from X-ray angiographic images using density based spatial clustering of applications with noise (DBSCAN). Biomed. Signal Process. Control 2025, 101, 107175. [Google Scholar] [CrossRef]
  4. Wold, S.; Esbensen, K.; Geladi, P. Principal component analysis. Chemom. Intell. Lab. Syst. 1987, 2, 37–52. [Google Scholar] [CrossRef]
  5. Zhang, Q.; Wang, Y.; Levine, M.D.; Yuan, X.; Wang, L. Multisensor video fusion based on higher order singular value decomposition. Inf. Fusion 2015, 24, 54–71. [Google Scholar] [CrossRef]
  6. Lee, D.D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788–791. [Google Scholar] [CrossRef]
  7. Hoyer, P.O. Non-negative matrix factorization with sparseness constraints. J. Mach. Learn. Res. 2004, 5, 1457–1469. [Google Scholar] [CrossRef]
  8. Kong, D.; Ding, C.; Huang, H. Robust Nonnegative Matrix Factorization Using L21-norm. In Proceedings of the 20th ACM International Conference on Information and Knowledge Management, Glasgow, UK, 24–28 October 2011; pp. 673–682. [Google Scholar] [CrossRef]
  9. Ding, C.; Li, T.; Peng, W. Orthogonal nonnegative matrix t-factorizations for clustering. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Philadelphia, PA, USA, 20–23 August 2006; pp. 126–135. [Google Scholar] [CrossRef]
  10. Belkin, M.; Niyogi, P.; Sindhwani, V. Manifold Regularization: A Geometric Framework for Learning from Labeled and Unlabeled Examples. J. Mach. Learn. Res. 2006, 7, 2399–2434. [Google Scholar]
  11. Cai, D.; He, X.; Han, J.; Huang, T.S. Graph Regularized Nonnegative Matrix Factorization for Data Representation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1548–1560. [Google Scholar] [CrossRef]
  12. Wu, B.; Wang, E.; Zhu, Z.; Chen, W.; Xiao, P. Manifold NMF with L2,1 norm for clustering. Neurocomputing 2018, 273, 78–88. [Google Scholar] [CrossRef]
  13. Li, X.; Cui, G.; Dong, Y. Graph Regularized Non-Negative Low-Rank Matrix Factorization for Image Clustering. IEEE Trans. Cybern. 2017, 47, 3840–3853. [Google Scholar] [CrossRef]
  14. Liu, Z.; Zhu, F.; Xiong, H.; Chen, X.; Pelusi, D.; Vasilakos, A.V. Graph regularized discriminative nonnegative matrix factorization. Eng. Appl. Artif. Intell. 2025, 139, 109629. [Google Scholar] [CrossRef]
  15. Huang, S.; Xu, Z.; Zhao, K.; Ren, Y. Regularized nonnegative matrix factorization with adaptive local structure learning. Neurocomputing 2020, 382, 196–209. [Google Scholar] [CrossRef]
  16. Ren, X.; Yang, Y. Semi-supervised symmetric non-negative matrix factorization with graph quality improvement and constraints. Appl. Intell. 2025, 55, 397. [Google Scholar] [CrossRef]
  17. Mohammadi, M.; Berahmand, K.; Azizi, S.; Sheikhpour, R.; Khosravi, H. Semi-Supervised Adaptive Symmetric Nonnegative Matrix Factorization for Multi-View Clustering. IEEE Trans. Netw. Sci. Eng. 2025; early access. [Google Scholar] [CrossRef]
  18. Defferrard, M.; Bresson, X.; Vandergheynst, P. Convolutional neural networks on graphs with fast localized spectral filtering. Adv. Neural Inf. Process. Syst. 2016, 29, 3844–3852. [Google Scholar] [CrossRef]
  19. Wang, D.; Ren, F.; Zhuang, Y.; Liang, C. Robust high-order graph learning for incomplete multi-view clustering. Expert Syst. Appl. 2025, 280, 127580. [Google Scholar] [CrossRef]
  20. Zhan, S.; Jiang, H.; Shen, D. Co-regularized optimal high-order graph embedding for multi-view clustering. Pattern Recogn. 2025, 157, 110892. [Google Scholar] [CrossRef]
  21. Wang, Y.; Yao, H.; Zhao, S. Auto-encoder based dimensionality reduction. Neurocomputing 2016, 184, 232–242. [Google Scholar] [CrossRef]
  22. Chen, Y.; Qu, G.; Zhao, J. Orthogonal graph regularized non-negative matrix factorization under sparse constraints for clustering. Expert Syst. Appl. 2024, 249, 123797. [Google Scholar] [CrossRef]
  23. Meng, Y.; Shang, R.; Jiao, L.; Zhang, W.; Yang, S. Dual-graph regularized non-negative matrix factorization with sparse and orthogonal constraints. Eng. Appl. Artif. Intell. 2018, 69, 24–35. [Google Scholar] [CrossRef]
  24. Peng, C.; Zhang, Y.; Chen, Y.; Kang, Z.; Chen, C.; Cheng, Q. Log-based sparse nonnegative matrix factorization for data representation. Knowl.-Based Syst. 2022, 251, 109127. [Google Scholar] [CrossRef]
  25. Xiong, W.; Ma, Y.; Zhang, C.; Liu, S. Dual graph-regularized sparse robust adaptive non-negative matrix factorization. Expert Syst. Appl. 2025, 281, 127594. [Google Scholar] [CrossRef]
  26. Shang, F.; Jiao, L.C.; Wang, F. Graph dual regularization non-negative matrix factorization for co-clustering. Pattern Recognit. 2012, 45, 2237–2250. [Google Scholar] [CrossRef]
  27. Gu, Q.; Zhou, J. Co-clustering on manifolds. In Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Paris, France, 28 June–1 July 2009; pp. 359–368. [Google Scholar] [CrossRef]
  28. Hedjam, R.; Abdesselam, A.; Melgani, F. NMF with feature relationship preservation penalty term for clustering problems. Pattern Recognit. 2021, 112, 107814. [Google Scholar] [CrossRef]
  29. Salahian, N.; Tab, F.A.; Seyedi, S.A.; Chavoshinejad, J. Deep Autoencoder-like NMF with Contrastive Regularization and Feature Relationship Preservation. Expert Syst. Appl. 2023, 214, 119051. [Google Scholar] [CrossRef]
  30. Lee, D.D.; Seung, H.S. Algorithms for non-negative matrix factorization. In Proceedings of the 14th International Conference on Neural Information Processing Systems, Denver, CO, USA, 1 January 2000; pp. 535–541. [Google Scholar]
  31. Sun, B.J.; Shen, H.; Gao, J.; Ouyang, W.; Cheng, X. A non-negative symmetric encoder-decoder approach for community detection. In Proceedings of the 2017 ACM Conference on Information and Knowledge Management, New York, NY, USA, 6–10 November 2017; pp. 597–606. [Google Scholar] [CrossRef]
  32. Li, T.; Zhang, R.; Yao, Y.; Liu, Y.; Ma, J.; Tang, J. Graph regularized autoencoding-inspired non-negative matrix factorization for link prediction in complex networks using clustering information and biased random walk. J. Supercomput. 2024, 80, 14433–14469. [Google Scholar] [CrossRef]
  33. Zhang, H.; Kou, G.; Peng, Y.; Zhang, B. Role-aware random walk for network embedding. Inf. Sci. 2024, 652, 119765. [Google Scholar] [CrossRef]
  34. Hosein, M.; Massoud, B.Z.; Christian, J. A Fast Approach for Overcomplete Sparse Decomposition Based on Smoothed l~(0) Norm. IEEE Trans. Signal Process. 2009, 57, 289–301. [Google Scholar] [CrossRef]
  35. Lu, X.; Wu, H.; Yuan, Y.; Yan, P.; Li, X. Manifold Regularized Sparse NMF for Hyperspectral Unmixing. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2815–2826. [Google Scholar] [CrossRef]
  36. Meng, Y.; Shang, R.; Jiao, L.; Zhang, W.; Yuan, Y.; Yang, S. Feature selection based dual-graph sparse non-negative matrix factorization for local discriminative clustering. Neurocomputing 2018, 290, 87–99. [Google Scholar] [CrossRef]
  37. Wang, J.; Wang, L.; Nie, F.; Li, X. A novel formulation of trace ratio linear discriminant analysis. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 5568–5578. [Google Scholar] [CrossRef]
  38. Wang, Q.; He, X.; Jiang, X.; Li, X. Robust Bi-stochastic Graph Regularized Matrix Factorization for Data Clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 390–403. [Google Scholar] [CrossRef] [PubMed]
  39. Chen, M.; Li, X. Entropy Minimizing Matrix Factorization. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 9209–9222. [Google Scholar] [CrossRef] [PubMed]
  40. Huang, J.; Nie, F.; Huang, H.; Ding, C.H.Q. Robust Manifold Nonnegative Matrix Factorization. ACM Trans. Knowl. Discov. Data 2013, 8, 11. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of autoencoder-like non-negative matrix factorization.
Figure 1. Schematic diagram of autoencoder-like non-negative matrix factorization.
Entropy 27 00875 g001
Figure 2. Example of associations between samples.
Figure 2. Example of associations between samples.
Entropy 27 00875 g002
Figure 3. Example of feature relationship preservation in high-dimensional and low-dimensional spaces.
Figure 3. Example of feature relationship preservation in high-dimensional and low-dimensional spaces.
Entropy 27 00875 g003
Figure 4. Schematic diagram of the ASNMF-SRP algorithm.
Figure 4. Schematic diagram of the ASNMF-SRP algorithm.
Entropy 27 00875 g004
Figure 5. Visual comparison of low-dimensional representation matrices of various algorithms on the MSRA25 dataset.
Figure 5. Visual comparison of low-dimensional representation matrices of various algorithms on the MSRA25 dataset.
Entropy 27 00875 g005
Figure 6. Visual comparison of low-dimensional representation matrices of various algorithms on the COIL20 dataset.
Figure 6. Visual comparison of low-dimensional representation matrices of various algorithms on the COIL20 dataset.
Entropy 27 00875 g006
Figure 7. Visual comparison of low-dimensional representation matrices of various algorithms on the Hitech dataset.
Figure 7. Visual comparison of low-dimensional representation matrices of various algorithms on the Hitech dataset.
Entropy 27 00875 g007
Figure 8. Comparison of clustering results between ASNMF-SRP and ASNMF-SRP-3 under different noise intensities.
Figure 8. Comparison of clustering results between ASNMF-SRP and ASNMF-SRP-3 under different noise intensities.
Entropy 27 00875 g008
Figure 9. Parameter sensitivity analysis of ASNMF-SRP on various datasets.
Figure 9. Parameter sensitivity analysis of ASNMF-SRP on various datasets.
Entropy 27 00875 g009aEntropy 27 00875 g009b
Figure 10. Convergence curves of ASNMF-SRP on various datasets.
Figure 10. Convergence curves of ASNMF-SRP on various datasets.
Entropy 27 00875 g010aEntropy 27 00875 g010b
Table 1. Introduction to five common sparsity methods.
Table 1. Introduction to five common sparsity methods.
No.MethodIntroduction to Sparsity Method
1 l 0 -norm [34] V 0 denotes the number of non-zero elements in V .
2 l 1 / 2 -norm [35] V 1 / 2 = i = 1 r j = 1 n v i j 1 / 2
3 l 1 -norm [22] V 1 = i = 1 r j = 1 n v i j
4 l 2 , 1 -norm [36] V 2 , 1 = j = 1 n i = 1 r v i j 2
5 l log -(pseudo) norm [24] V log = i = 1 r j = 1 n log 1 + v i j
Table 2. Introduction of the 8 datasets.
Table 2. Introduction of the 8 datasets.
NO.DatasetSamples ( n )Features ( m )Classes ( r )Data TypeImage Size
1MSRA25179925612Face dataset16 × 16
2Semeion159325610Digit images16 × 16
3COIL201440102420Object images32 × 32
4COIL10072001024100Object images32 × 32
5Krvs3196362Network detection
6Hitech230122166Technology news
7PenDigits34981610Handwritten digits
8Vehicle846184Vehicle contours
Table 3. Parameter values of ASNMF-SRP on different datasets.
Table 3. Parameter values of ASNMF-SRP on different datasets.
No.DatasetHigher-Order Graph Regularization α Feature Relationship Preservation β Sparse
Constraint λ
1MSRA25100010000.001
2Semeion11000.0001
3COIL20100010000.001
4COIL100100010000.001
5Krvs100100.0001
6Hitech1000100.0001
7PenDigits1001000.001
8Vehicle110.001
Table 4. ACC of 10 algorithms on 8 datasets (mean ± standard deviation).
Table 4. ACC of 10 algorithms on 8 datasets (mean ± standard deviation).
AlgorithmNMFONMFHx-NMFEMMFGNMFRMNMFDRCCFR-NMFLS-NMFASNMF-SRPI-P
Dataset
MSRA250.508420.493750.520260.505920.539380.555500.289740.516590.540190.579044.24%
±0.023±0.026±0.029±0.018±0.032±0.034±0.029±0.022±0.022±0.045--
Semeion0.525080.569590.508070.519650.592090.279160.616010.535400.602510.6806310.49%
±0.042±0.032±0.042±0.040±0.039±0.057±0.037±0.037±0.036±0.050--
COIL200.664060.685310.658120.651420.768440.243850.790170.650280.773610.841746.53%
±0.029±0.028±0.030±0.019±0.013±0.094±0.034±0.026±0.014±0.012--
COIL1000.470260.491590.468770.478820.487380.412350.460990.470440.483060.6417330.54%
±0.014±0.011±0.012±0.018±0.014±0.012±0.014±0.015±0.011±0.010--
Krvs0.519090.537420.522230.521370.530820.518100.555940.525520.533870.568132.19%
±0.003±0.012±0.002±0.003±0.012±0.007±0.004±0.023±0.015±0.003--
Hitech0.233850.233750.234030.235440.231050.262730.240870.226030.232680.25367−3.45%
±0.002±0.005±0.008±0.004±0.004±0.001±0.004±0.007±0.002±0.002--
PenDigits0.662160.707290.680920.666300.679730.651830.731190.667910.685330.8044210.02%
±0.036±0.048±0.040±0.038±0.053±0.037±0.043±0.034±0.052±0.032--
Vehicle0.387940.437770.401420.390960.443970.359160.411940.435700.443680.452361.89%
±0.019±0.002±0.026±0.021±0.008±0.003±0.017±0.023±0.009±0.002--
Table 5. ARI of 10 algorithms on 8 datasets (mean ± standard deviation).
Table 5. ARI of 10 algorithms on 8 datasets (mean ± standard deviation).
AlgorithmNMFONMFHx-NMFEMMFGNMFRMNMFDRCCFR-NMFLS-NMFASNMF-SRPI-P
Dataset
MSRA250.345750.329330.357340.334820.406050.400580.120130.355620.405960.446629.99%
±0.026±0.027±0.027±0.020±0.037±0.034±0.023±0.019±0.034±0.052--
Semeion0.311980.350510.306550.311340.442800.118550.416510.319430.459210.488096.29%
±0.033±0.024±0.031±0.030±0.032±0.047±0.026±0.029±0.030±0.034--
COIL200.579890.625820.576270.570690.741600.164500.733720.566820.742340.802448.10%
±0.026±0.023±0.034±0.026±0.018±0.093±0.036±0.024±0.016±0.007--
COIL1000.395840.441570.395490.406970.423710.303290.391670.397090.420670.5357321.32%
±0.016±0.014±0.017±0.019±0.010±0.015±0.016±0.012±0.011±0.016--
Krvs0.001070.005790.001580.001470.00393−0.000380.012010.004100.005050.0181451.04%
±0.000±0.004±0.000±0.000±0.003±0.001±0.002±0.006±0.005±0.002--
Hitech−0.000950.000920.000210.00034−0.000410.000160.002520.00066−0.000490.00689173.41%
±0.001±0.001±0.001±0.001±0.001±0.001±0.002±0.001±0.001±0.000--
PenDigits0.523610.554280.523520.523290.541370.502380.574740.528890.558090.6848719.16%
±0.029±0.040±0.037±0.022±0.045±0.049±0.037±0.024±0.039±0.035--
Vehicle0.081690.124620.092070.083220.131030.063640.096910.113210.132210.12014−9.13%
±0.013±0.003±0.018±0.018±0.009±0.004±0.014±0.012±0.007±0.001--
Table 6. NMI of 10 algorithms on 8 datasets (mean ± standard deviation).
Table 6. NMI of 10 algorithms on 8 datasets (mean ± standard deviation).
AlgorithmNMFONMFHx-NMFEMMFGNMFRMNMFDRCCFR-NMFLS-NMFASNMF-SRPI-P
Dataset
MSRA250.569350.562960.577730.554670.651110.602950.237450.577150.646130.715129.83%
±0.021±0.028±0.023±0.021±0.031±0.029±0.033±0.017±0.031±0.026--
Semeion0.441620.488470.443120.449380.607900.201710.540140.449380.614890.632822.92%
±0.025±0.020±0.026±0.023±0.020±0.074±0.018±0.022±0.019±0.021--
COIL200.761120.795910.760670.754230.885380.313750.891310.755460.885000.915292.69%
±0.015±0.010±0.018±0.017±0.012±0.131±0.011±0.015±0.012±0.006--
COIL1000.752580.768350.754000.756460.772260.700610.746410.731170.769480.838358.56%
±0.005±0.005±0.006±0.006±0.004±0.009±0.006±0.005±0.004±0.003
Krvs0.000600.003970.000910.000940.002650.002030.008180.005920.003520.0125052.81%
±0.000±0.003±0.000±0.000±0.002±0.003±0.001±0.006±0.003±0.001--
Hitech0.007990.009890.008540.010490.008650.005580.012030.007860.007990.0193560.85%
±0.001±0.002±0.002±0.001±0.002±0.001±0.002±0.002±0.002±0.000--
PenDigits0.682510.693320.665760.676840.697510.642940.696150.684800.709530.8012012.92%
±0.022±0.019±0.026±0.018±0.027±0.043±0.016±0.021±0.018±0.021--
Vehicle0.116780.190000.136480.127140.190620.084080.148580.164040.191810.18544−3.32%
±0.014±0.006±0.019±0.018±0.015±0.005±0.019±0.018±0.013±0.000--
Table 7. PUR of 10 algorithms on 8 datasets (mean ± standard deviation).
Table 7. PUR of 10 algorithms on 8 datasets (mean ± standard deviation).
AlgorithmNMFONMFHx-NMFEMMFGNMFRMNMFDRCCFR-NMFLS-NMFASNMF-SRPI-P
Dataset
MSRA250.529850.521760.537550.522870.565870.579180.306980.538490.566510.614796.15%
±0.021±0.023±0.025±0.020±0.026±0.027±0.030±0.017±0.025±0.035--
Semeion0.537630.588040.534310.538070.637260.287380.636250.548620.643690.697398.34%
±0.036±0.025±0.030±0.031±0.024±0.061±0.027±0.028±0.026±0.031--
COIL200.691350.713400.686280.679970.807150.249030.828020.677530.808920.864554.41%
±0.024±0.023±0.021±0.021±0.017±0.095±0.017±0.023±0.018±0.011--
COIL1000.526630.547050.528160.533790.546540.479510.518490.516370.542470.6959927.23%
±0.013±0.010±0.011±0.013±0.012±0.012±0.012±0.012±0.009±0.006--
Krvs0.522450.537420.522890.522610.531370.522370.555940.533600.533870.568132.19%
±0.000±0.012±0.001±0.000±0.011±0.000±0.004±0.016±0.015±0.003--
Hitech0.266930.270340.267580.270170.268770.263800.270990.267880.266060.285255.26%
±0.003±0.003±0.003±0.003±0.003±0.001±0.004±0.005±0.003±0.001--
PenDigits0.692620.723400.694470.691180.706750.672260.738720.696260.712560.810959.78%
±0.026±0.031±0.031±0.026±0.035±0.035±0.029±0.024±0.035±0.023--
Vehicle0.392850.437770.405730.396810.443970.371450.418320.438530.443680.452361.89%
±0.015±0.002±0.020±0.018±0.008±0.005±0.020±0.022±0.009±0.002--
Table 8. Clustering results of ASNMF-SRP and ASNMF-SRP-1 (mean ± standard deviation).
Table 8. Clustering results of ASNMF-SRP and ASNMF-SRP-1 (mean ± standard deviation).
DatasetACCARINMIPUR
ASNMF-SRP-1ASNMF-SRPASNMF-SRP-1ASNMF-SRPASNMF-SRP-1ASNMF-SRPASNMF-SRP-1ASNMF-SRP
MSRA250.496800.579040.333560.446620.556120.715120.516560.61479
±0.021±0.045±0.016±0.052±0.019±0.026±0.019±0.035
Semeion0.647240.680630.481980.488090.642920.632820.689080.69739
±0.010±0.050±0.007±0.034±0.010±0.021±0.011±0.031
COIL200.805520.841740.763130.802440.886510.915290.835870.86455
±0.015±0.012±0.017±0.007±0.008±0.006±0.013±0.011
COIL1000.480070.641730.403110.535730.732550.838350.525560.69599
±0.009±0.010±0.011±0.016±0.005±0.003±0.007±0.006
Krvs0.528390.568130.003930.018140.003960.012500.533390.56813
±0.017±0.003±0.004±0.002±0.004±0.001±0.012±0.003
Hitech0.233010.253670.000140.006890.008930.019350.268470.28525
±0.005±0.002±0.001±0.000±0.001±0.000±0.003±0.001
PenDigits0.665330.804420.531300.684870.686840.801200.694950.81095
±0.040±0.032±0.030±0.035±0.020±0.021±0.026±0.023
Vehicle0.447460.452360.134450.120140.200320.185440.448880.45236
±0.004±0.002±0.003±0.001±0.008±0.000±0.006±0.002
Table 9. Clustering results of ASNMF-SRP and ASNMF-SRP-2 (mean ± standard deviation).
Table 9. Clustering results of ASNMF-SRP and ASNMF-SRP-2 (mean ± standard deviation).
DatasetACCARINMIPUR
ASNMF-SRP-2ASNMF-SRPASNMF-SRP-2ASNMF-SRPASNMF-SRP-2ASNMF-SRPASNMF-SRP-2ASNMF-SRP
MSRA250.575180.579040.444370.446620.720120.715120.613540.61479
±0.032±0.045±0.038±0.052±0.019±0.026±0.025±0.035
Semeion0.678030.680630.483800.488090.628830.632820.695670.69739
±0.044±0.050±0.029±0.034±0.018±0.021±0.025±0.031
COIL200.838650.841740.801550.802440.912850.915290.861940.86455
±0.013±0.012±0.008±0.007±0.004±0.006±0.013±0.011
COIL1000.638180.641730.543460.535730.831030.838350.690850.69599
±0.007±0.010±0.014±0.016±0.003±0.003±0.004±0.006
Krvs0.568840.568130.018500.018140.012740.012500.568840.56813
±0.000±0.003±0.000±0.002±0.000±0.001±0.000±0.003
Hitech0.247440.253670.004240.006890.018930.019350.279120.28525
±0.001±0.002±0.001±0.000±0.000±0.000±0.001±0.001
PenDigits0.799360.804420.675170.684870.794760.801200.804990.81095
±0.029±0.032±0.028±0.035±0.017±0.021±0.022±0.023
Vehicle0.452190.452360.120180.120140.185400.185440.452190.45236
±0.001±0.002±0.001±0.001±0.001±0.000±0.001±0.002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhong, L.; Gao, H. Autoencoder-like Sparse Non-Negative Matrix Factorization with Structure Relationship Preservation. Entropy 2025, 27, 875. https://doi.org/10.3390/e27080875

AMA Style

Zhong L, Gao H. Autoencoder-like Sparse Non-Negative Matrix Factorization with Structure Relationship Preservation. Entropy. 2025; 27(8):875. https://doi.org/10.3390/e27080875

Chicago/Turabian Style

Zhong, Ling, and Haiyan Gao. 2025. "Autoencoder-like Sparse Non-Negative Matrix Factorization with Structure Relationship Preservation" Entropy 27, no. 8: 875. https://doi.org/10.3390/e27080875

APA Style

Zhong, L., & Gao, H. (2025). Autoencoder-like Sparse Non-Negative Matrix Factorization with Structure Relationship Preservation. Entropy, 27(8), 875. https://doi.org/10.3390/e27080875

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop