Next Article in Journal
Refined One Relaxation Time-Fractional Theory for the Thermoelastic Response of Circular Cylinders with Variable Thermal Conductivity
Previous Article in Journal
Some Identities Involving Fubini Polynomials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrating Dual Graph Constraints into Sparse Non-Negative Tucker Decomposition for Enhanced Co-Clustering

1
School of Mathematical Sciences, Guizhou Normal University, Guiyang 550025, China
2
School of Mathematics and Physics, Anshun University, Anshun 561000, China
3
School of Mathematical Sciences, Xiamen University, Xiamen 361005, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2025, 13(21), 3494; https://doi.org/10.3390/math13213494
Submission received: 29 September 2025 / Revised: 29 October 2025 / Accepted: 30 October 2025 / Published: 1 November 2025

Abstract

Collaborative clustering is an ensemble technique that enhances clustering performance by simultaneously and synergistically processing multiple data dimensions or tasks. This is an active research area in artificial intelligence, machine learning, and data mining. A common approach to co-clustering is based on non-negative matrix factorization (NMF). While widely used, NMF-based co-clustering is limited by its bilinear nature and fails to capture the multilinear structure of data. With the objective of enhancing the effectiveness of non-negative Tucker decomposition (NTD) in image clustering tasks, in this paper, we propose a dual-graph constrained sparse non-negative Tucker decomposition NTD (GDSNTD) model for co-clustering. It integrates graph regularization, the Frobenius norm, and an l 1 norm constraint to simultaneously optimize the objective function. The GDSNTD mode, featuring graph regularization on both factor matrices, more effectively discovers meaningful latent structures in high-order data. The addition of the l 1 regularization constraint on the factor matrices may help identify the most critical original features, and the use of the Frobenius norm may produce a more highly stable and accurate solution to the optimization problem. Then, the convergence of the proposed method is proven, and the detailed derivation is provided. Finally, experimental results on public datasets demonstrate that the proposed model outperforms state-of-the-art methods in image clustering, achieving superior scores in accuracy and Normalized Mutual Information.

1. Introduction

Non-negative decomposition is a general term for a type of matrix or tensor decomposition method that requires all components in the decomposition result to be non-negative. The core idea is to decompose a non-negative data matrix or tensor into the product of several non-negative factor matrices or core tensors in order to extract interpretable latent features or components from non-negative data. The most representative non-negative decompositions are non-negative matrix factorization (NMF) and its extension in the high-dimensional domain non-negative Tucker decomposition (NTD). NMF is a common dimensionality reduction method that can extract the main features from the original data and has high interpretability [1]. When data requires more than two indices to be uniquely identified, it should be expressed in the form of a tensor. Thus, a tensor is naturally viewed as a generalization of vectors (first-order tensors) and matrices (second-order tensors) to higher orders. Non-negative tensor decomposition is a multi-dimensional data analysis method based on matrix decomposition, which can transform high-dimensional data into low-dimensional representations and retain the main features of the original data [2]. Shcherbakova  et al. investigated the advantages of non-negative Tucker decomposition [3]. Like matrix decomposition, the purpose of tensor decomposition is to extract the information or main components hidden in the original data [4,5]. The Tucker decomposition is a tensor factorization method designed for handling tensor data [6]. NTD is a concrete application of Tucker decomposition in non-negative matrix decomposition. Its core idea is to decompose a high-order non-negative tensor into the product of a core tensor and a series of factor matrices [7]. NTD not only retains the interpretability benefits of NMF but also effectively preserves the inherent multi-way structure of the original tensor data. Consequently, non-negative Tucker decomposition (NTD) is extensively utilized in a wide range of disciplines, including image analysis and text mining, to uncover the latent structures inherent in multi-way datasets.
To improve model performance, numerous regularized extensions of NMF and NTD have been developed in the recent literature. These models have demonstrated excellent performance in image clustering or data mining. Cai et al. introduced the graph-regularized non-negative matrix factorization (GNMF) algorithm, which incorporates a geometrically-based affinity graph into the NMF framework to preserve the local manifold structure of the data [8]. Sun et al. proposed graph-regularized and sparse non-negative matrix factorization with hard constraints (GSNMFC), jointly incorporating a graph regularizer and hard prior label information as well as a sparseness constraint as additional conditions to uncover the intrinsic geometrical and discriminative structures of the data space [9]. They also proposed sparse dual graph-regularized non-negative matrix factorization (SDGNMF), jointly incorporating the dual graph-regularized and sparseness constraints as additional conditions to uncover the intrinsic geometrical, discriminative structures of the data space [10]. Shang et al. proposed a novel algorithm, called graph dual regularization non-negative matrix factorization (DNMF), which simultaneously considers the geometric structures of both the data manifold and the feature manifold [11]. Long et al. proposed a novel constrained non-negative matrix factorization algorithm, called the graph-regularized discriminative non-negative matrix factorization (GDNMF), to incorporate into the NMF model both the intrinsic geometrical structure and discriminative information [12]. Saberi-Movahed et al. present a systematic analysis of NMF in dimensionality reduction, with a focus on both feature extraction and feature selection approaches [13]. Jing et al. proposed a novel semi-supervised NMF method that incorporates label regularization, basis regularization, and graph regularization [14]. Li et al. developed a manifold regularization non-negative Tucker decomposition (MR-NTD) model. To preserve the geometric information within tensor data, their method employs a manifold regularization term on the core tensor derived from the Tucker decomposition [15]. Yin and Ma incorporated this geometrically based Locally Linear Embedding (LLE) into the original NTD, thus proposing NTD-LLE for the clustering of image databases [16]. To enhance the representation learning of tensor data, Qiu et al. proposed a novel graph-regularized non-negative Tucker decomposition (GNTD) framework [17]. This method is designed to jointly extract low-dimensional parts-based representations and preserve the underlying manifold structure within the high-dimensional tensor data.
The above research demonstrates that the non-negative decomposition model significantly improves image clustering performance. However, advancements in scientific technologies have led to increasingly complex phenomena, rendering previous models inadequate for handling the resulting complexities. Building on this foundation, a common strategy to enhance image clustering accuracy has been the development of collaborative clustering (co-clustering) frameworks. This is typically achieved by incorporating specific regularization terms that capture the relationships between different data views or clusters. Co-clustering is an ensemble learning method that performs simultaneous clustering along multiple dimensions or data views [18]. When using the non-negative matrix model, the goal of co-clustering is to simultaneously identify the clusters of both the rows and columns of the two-dimensional data matrix [19]. Del Buono and Pio present a process which aims at enhancing the performance of three-factor NMF as a co-clustering method, by identifying a clearer correlation structure represented by the block matrix [20]. Deng et al. proposed graph-regularized sparse NMF (GSNMF) and graph-regularized sparse non-negative matrix tri-factorization (GSNMTF) models. By incorporating an L 1 norm constraint on the low-dimensional matrix, they aimed to scale the data eigenvalues and enforce sparsity. This co-clustering approach has been shown to enhance the performance of standard non-negative matrix factorization models [21]. Chachlakis, Dhanaraj, Prater-Bennette, and Markopoulos presented Dynamic L 1 -Tucker: an algorithm for dynamic and outlier-resistant Tucker analysis of tensor data [22]. The  L 1 norm is extensively used in convex optimization problems [23]. Ahmed et al. study a tensor-structured linear regression model over the space of sparse, low Tucker-rank tensors [24]. As deep learning models become larger and more complex, sparsity is emerging as a critical consideration for enhancing efficiency and scalability, making it a central theme in the development of new image processing and data analysis methods.
As illustrated by the above studies, building upon NMF, scholars have developed numerous models to address the evolving needs of various scientific fields. The performance of the NMF model can be significantly improved through the combined constraints of graph regularization and sparsity, which directly exploit the internal structure and inherent characteristics of the data. NTD is the extension of NMF to the high-dimensional domain. However, there are relatively few co-clustering methods for constructing high-performance NTD models that leverage the internal structure and inherent characteristics of the tensor data itself. While GNTD captures graph structure and sparse NMTF promotes sparsity, neither is designed to simultaneously learn from multiple graphs while enforcing directional sparsity patterns across different data modes. To address this gap, and inspired by advancements in co-clustering NMF, this paper proposes a GDSNTD model based on GNTD for enhanced co-clustering. The new model combines graph regularization, the Frobenius norm, and the  l 1 norm to simultaneously optimize the objective function. In NTD, graph regularization serves to preserve the multi-linear structure of the original data. The review of the prior literature revealed that imposing multiple graph constraints on NMF models enhances their clustering performance. Motivated by this finding, we consequently introduce dual graph constraints into the NTD framework, applying them directly to the factor matrices of a tensor. This approach allows the model to capture the intrinsic data geometry more clearly, thereby improving clustering accuracy. Furthermore, we provide updated iterative optimization rules and prove the convergence of the model. Experiments on public datasets demonstrate that the proposed method outperforms several leading state-of-the-art methods.
The main contributions of this study are as follows:
  • We introduce dual graph constraints into the NTD framework, applying them directly to the factor matrices of a tensor. To the best of our knowledge, no existing NTD framework integrates dual graph constraints with sparse regularization simultaneously. While graph-regularized and sparse factorization techniques exist, our model GDSNTD is the first to integrate them in a unified, constrained co-clustering optimization framework for NTD. We propose a new co-clustering version of the NTD model, equipped with three regularization terms: graph regularization, the Frobenius norm, and the  l 1 norm. The graph regularization term captures the internal geometric structure of high-dimensional data more accurately. The  l 1 norm term helps to scale original features in the factor matrices. The Frobenius norm improves the generalization ability of the model. Therefore, the co-clustering GDSNTD integrates strengths from graph-regularized and sparse factorization techniques so that the model yields a more accurate solution to the optimization problem.
  • In the novel, unified optimization objective, we leverage the L-Lipschitz condition to derive the update rules for the proposed co-clustering GDSNTD method. Subsequently, we establish the convergence of the proposed algorithm.
  • Experiments on public datasets demonstrate the effectiveness and superiority of the proposed method.
The remainder of the paper is organized as follows. In Section 2, we review the related models. In Section 3, the GDSNTD method is proposed, and its detailed inference process and the proof of convergence of the algorithm are illustrated. Section 4 presents the performance of the proposed model via experiments on various datasets. Finally, we present our conclusions in Section 5 and outline future work in Section 6.

2. Preliminaries

In this section, we review the related models.
(A)
NTD
  Given a non-negative data tensor X R + I 1 × I 2 × I 3 , non-negative tucker decomposition (NTD) aims at decomposing the non-negative tensor X into a non-negative core tensor G R + I 1 × I 2 × I 3 multiplied by N non-negative factor matrices U R + I 1 × J 1 , V R + I 2 × J 2 , and  W R + I 3 × J 3 along each mode [7]. NTD minimizes the sum of squared residues between the data tensor X and the multi-linear product of core tensor G and factor matrices U , V , and W , which is expressed as
F N T D = min G 0 , U , V , W 0 X G × 1 U × 2 V × 3 W F 2 .
  NTD provides an effective embedding and representation for image tensor data X . However, NTD does not consider the geometrical structure of image tensor data  X .
(B)
GNTD
  Qiu et al. [17] proposed the GNTD model, which incorporates the graph regularization term into the original NTD method. The GNTD is mathematically formulated as
F G N T D = min G 0 , U , V , W 0 X G × 1 U × 2 V × 3 W F 2 + λ T r ( W L W ) ,
where λ 0 is the regularization parameter for balancing the importance of the graph regularization term and the reconstruction error term, tr(·) denotes the trace of the matrix, and L is a Laplacian matrix which characterizes the data manifold. L is defined as L = H A , and  ( H ) i i = j ( A ) i j . A is the adjacency matrix of a p-nearest neighbor graph regarding the data points ( Y 1 , , Y 2 , ) constructed using the 0 1 scheme.
( A ) i j = 1 , if Y i N p ( Y j ) and Y j N p ( Y i ) , 0 , otherwise ,
where N p ( Y j ) represents the set of p-nearest neighbor points of ( Y j ) .
  The GNTD algorithm ingeniously combines graph constraints with non-negative constraints. This approach effectively preserves the local geometric structure of the data, enabling the learned features to better capture its intrinsic structure. As a result, GNTD significantly enhances the performance of traditional tensor decomposition models and has become a powerful tool for handling complex high-dimensional data, receiving considerable attention in machine learning and data mining.
(C)
GSNMTF
  The graph-regularized sparse non-negative matrix tri-factorization (GSNMTF) model incorporates graph regularization and the l 1 norm into the standard NMF objective function [21]. This integration not only preserves the geometric structure of the data and feature spaces but also promotes sparsity in the resulting factor matrices. The objective function of GSNMTF is defined as
F G S N M T F = min U 0 , S 0 , V 0 X U S V F 2 + λ U T r ( U L U U ) + λ V T r ( V L V V ) +
γ U U 1 1 + γ V V 1 1 ,
where tr(·) is the trace of the matrix, and γ U and γ V are the parameters for controlling the sparseness of the matrices U and V , respectively. · F and · 1 are the Frobenius norm and l 1 norm, respectively.

3. Dual Graph Constraints into Sparse Non-Negative Tucker Decomposition (GDSNTD) Model

This section introduces the proposed GDSNTD model. We begin with the problem setup. Given a third-order tensor X R + I 1 × I 2 × I 3 , where I 1 , I 2 , and I 3 are the numbers of rows, columns, and tubes of tensor, respectively. The Tucker decomposition of the tensor is given by
X G × 1 U × 2 V × 3 W .
It decomposes a tensor into a core tensor multiplied by a matrix along each mode. Here the tensor G R + J 1 × J 2 × J 3 is called the core tensor and its entries show the level of interaction between the different components. U R + I 1 × J 1 , V R + I 2 × J 2 , and  W R + I 3 × J 3 are the factor matrices (which are usually orthogonal) and can be thought of as the principal components in each mode, and  { J 1 , J 2 , J 3 } m i n { I 1 , I 2 , I 3 } .

3.1. Data and Feature Graphs

Many research studies have demonstrated that geometric structure not only affects the data space but also the feature space and that high-dimensional data are usually located in a low-dimensional sub-manifold of the ambient space, and this underlying geometrical information of data could be obtained by modeling a neighbor graph. We impose a graph regularization constraint on the factor matrix V . We construct a p-nearest neighbors data space, whose vertices are { X : , : , 1 , , X : , : , N } . Then, we encode their geometrical information by connecting each tensor subject with its p-nearest neighbors, thereby constructing a p-nearest neighbor graph [8] with a binary ( 0 1 ) weighting scheme. The weight matrix is defined as
[ C V ] i j k = 1 if X j , : , k N p ( X i , : , k ) and X i , : , k N p ( X j , : , k ) , 0 otherwise ,
where N p ( X j , : , k ) represents the set of p samples closest to X i , : , k in the graph. The graph Laplacian of the data graph is defined as L V = D V C V , where D V is the diagonal degree matrix whose elements are given by [ D V ] i i = j [ C V ] i j .
Similarly, we impose a graph regularization constraint on the factor matrix W . We can also construct a p-nearest neighbors graph of the feature space and define the weight matrix as follows:
[ C W ] i j k = 1 if X k , j , : N p ( X i , j , : ) and X i , j , : N p ( X k , j , : ) , 0 otherwise ,
where N p ( X k , j , : ) represents the set of p samples closest to X i , j , : in the graph. The graph Laplacian of the data graph is defined as L W = D W C W .

3.2. Objective Function of GDSNTD

In this part, we present our objective function of GDSNTD, which is designed to approximate the original high-dimensional tensor by yielding cleaner and sparser low-dimensional matrices. This model is able to present the geometric structure of tensors more clearly than GNTD and makes up for the deficiency of GSNMTF in handling high-dimensional data. The proposed model has advanced the development of the NTD framework by improving its convergence properties and has facilitated its application in complex tasks such as image clustering analysis. It integrates graph regularization, the Frobenius norm, and a sparsity constraint, and its objective function is defined as
F G D S N T D = min G 0 , U , V , W 0 X G × 1 U × 2 V × 3 W F 2 + α V V F 2 + α W W F 2 + λ V T r ( V L V V ) + λ W T r ( W L W W ) + γ V V 1 1 + γ W W 1 1 ,
where tr(·) is the trace of the matrix. α V is the parameter of V , α W is the parameter of W . λ V and λ W are the graph regularization parameters of V and W , respectively. They govern the trade-off between preserving the intrinsic graph structure of the data and optimizing other objective terms. γ V and γ W are coefficients of the l 1 norm. They serve as a trade-off between the sparsity of the factor matrices and other objective terms. It is worth noting that the norms in the objective function are entrywise norms [25]. · F and · 1 represent the Frobenius norm and l 1 norm, respectively.
Equivalently, Equation (2) is to be rewritten in matrix form as
F G D S N T D = min G 0 , U , V , W 0 X ( 1 ) U G ( 1 ) ( W V ) F 2 + α V V F 2 + α W W F 2 + λ V T r ( V L V V ) + λ W T r ( W L W W ) + γ V V 1 1 + γ W W 1 1 ,
where the tensor X is expanded according to mode-1. When the tensor X is expanded according to mode-2 or mode-3, it is to be rewritten in the following forms as Equations (4) and Equation (5), respectively:  
F G D S N T D = min G 0 , U , V , W 0 X ( 2 ) V G ( 2 ) ( W U ) F 2 + α V V F 2 + α W W F 2 + λ V T r ( V L V V ) + λ W T r ( W L W W ) + γ V V 1 1 + γ W W 1 1 ,
F G D S N T D = min G 0 , U , V , W 0 X ( 3 ) W G ( 3 ) ( V U ) F 2 + α V V F 2 + α W W F 2 + λ V T r ( V L V V ) + λ W T r ( W L W W ) + γ V V 1 1 + γ W W 1 1 .

3.3. Inference of GDSNTD

In this section, the detailed inference is illustrated, and then, a corresponding algorithm is designed.

3.3.1. Fix U R + I 1 × J 1 , V R + I 2 × J 2 , W R + I 3 × J 3 , Solve G

Letting λ G G be the Lagrange multiplier [26] for constraint G 0 , and our objective function is written in matrix form as
L ^ ( G ) = v e c ( X ) ( W V U ) v e c ( G ) F 2 + v e c ( G ) v e c ( λ G ) .
To obtain the optimal solution, the gradient descent algorithm [27] is used to solve Equation (6). The gradient is as follows:
L ^ v e c ( G ) = 2 ( W V U ) ( W V U ) v e c ( G ) 2 ( W V U ) v e c ( X ) + v e c ( λ G ) .
Let G ( t ) represent the corresponding value during the tth round of updates. Therefore, using the Karush–Kuhn–Tucker (KKT) condition [27], L v e c ( G ) = 0 and v e c ( G ) i v e c ( λ G ) i = 0 , we obtain the following updating rule:
( v e c ( G ) ) i ( v e c ( G ) ) i ( ( W V U ) v e c ( X ) ) i ( ( W V U ) ( W V U ) v e c ( G ) ) i .

3.3.2. Fix G , V R + I 2 × J 2  and W R + I 3 × J 3 , Solve U R + I 1 × J 1

At this stage, similar to the update of G , letting λ U U be the Lagrange multiplier for constraint U 0 , our optimization goal is to be written in matrix form as follows:
O ^ ( U ) = X ( 1 ) U G ( 1 ) ( W V ) F 2 + T r ( U λ U ) .
The function (10) is to be rewritten as
O ^ ( U ) = T r ( X ( 1 ) X ( 1 ) ) 2 T r ( X ( 1 ) ( W V ) G ( 1 ) U ) + T r ( U G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) U ) + r = 1 N T r ( U λ U ) .
Now, the partial derivative of O ^ ( U ) to U is
O ^ ( U ) ( U ) = 2 X ( 1 ) ( W V ) G ( 1 ) + 2 U G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) + λ U .
According to the (Karush–Kuhn–Tucker) (KKT) condition [27], O ^ ( U ) ( U ) = 0 , so
λ U = 2 X ( 1 ) ( W V ) G ( 1 ) 2 U G ( r ) ( ( W V ) ( W V ) ) G ( r ) .
And also because λ U U = 0 , ⊙ denotes the Hadamard product, and we can obtain the following:
( 2 X ( 1 ) ( W V ) G ( 1 ) 2 U G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) ) U = 0 .
Therefore, the update rule of U is
U i j U i j ( X ( 1 ) ( W V ) G ( 1 ) ) i j ( U G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) ) i j .

3.3.3. Fix G , U R + I 1 × J 1 and W R + I 3 × J 3 , Solve V R + I 2 × J 2

At this stage, our goal is to minimize the function P ( V ) . The optimization goal in (4) is to be defined as
P ( V ) = P ^ ( V ) + γ V V 1 1 ,
where
P ^ ( V ) = X ( 2 ) V G ( 2 ) ( W U ) F 2 + α V V F 2 + λ V T r ( V L V V ) + T r ( λ V ) .
Since P ^ ( V ) is a convex function with respect to V and it satisfies the L-Lipschitz condition [28] as shown in Equation (14), there exists a constant L 1 0 such that
P ^ ( V ) P ^ ( V ) L 1 V V , ( V , V ) ,
where P ^ ( V ) is to be by the approximated second-order Taylor expansion near V as
P ˜ ( V ) P ^ ( V ) + L 1 2 ( V V ) 2 + P ^ ( V ) ( V ) ( V V ) = L 1 2 ( V ( V 1 L 1 P ^ ( V ) V ) ) 2 + Φ ( V ) , Φ ( V ) = P ^ ( V ) 1 2 L 1 ( P ^ ( V ) V ) 2 .
Let V = V ( t ) , the minimum value of Equation (17) is then obtained as V ( t + 1 ) and
V ( t + 1 ) = V ( t ) 1 L 1 P ^ ( V ( t ) ) ( V ( t ) ) .
Since Φ ( V ( t ) ) is independent of the optimization variable V , it can be treated as a constant. Consequently, the objective function for updating V simplifies to
V ( t + 1 ) = arg min V [ L 1 2 V ( V ( t ) 1 L 1 P ^ ( V ( t ) ) V ( t ) ) 2 2 + γ V V 1 1 ] .
Let S ( t ) = V ( t ) 1 L 1 P ^ ( V t ) V ( t ) , then, for solving Equation (19), we first solve S ( t ) and then
V ( t + 1 ) = arg min V [ L 1 2 V S ( t ) 2 2 + γ V V 1 1 ] .
Now, the partial derivative of P ^ ( V ( t ) ) to V ( t ) is  
P ^ ( V ( t ) ) ( V ( t ) ) = 2 ( X ( 2 ) ( t ) V ( t ) G ( 2 ) ( t ) ( W ( t ) U ( t ) ) ) ( G ( 2 ) ( t ) ( W ( t ) U ( t ) ) ) + 2 α V V ( t ) + 2 λ V L V ( t ) V ( t ) + λ V ( t ) = 2 X ( 2 ) ( t ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + 2 V ( t ) G ( 2 ) ( t ) ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + 2 α V V ( t ) + 2 λ V ( D V ( t ) C V ( t ) ) V ( t ) + λ V ( t ) .
where D V ( t ) and C V ( t ) should be solved at the same time. According to the (Karush–Kuhn–Tucker) (KKT) condition [27], λ V V ( t ) = 0 , and  P ^ ( V ( t ) ) V ( t ) = 0 , we can obtain the following:
( 2 X ( 2 ) ( t ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) 2 λ V C V ( t ) V ( t ) + 2 V ( t ) G ( 2 ) ( t ) ( W ( t ) U ( t ) ) T ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + 2 α V V ( t ) + 2 λ V D V ( t ) V ( t ) ) V ( t ) = 0 .
Therefore, the update rule of S ( t ) is
S i j ( t ) = V i j ( t ) [ X ( 2 ) ( t ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + λ V C V ( t ) V ( t ) ] i j [ V ( t ) G ( 2 ) ( t ) ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + α V V ( t ) + λ V D V ( t ) V ( t ) ] i j .
To solve the objective function in Equation (19), we minimize the S i j ( t ) term using the gradient descent and apply a soft threshold [28] to minimize the V 1 1 regularizer. Consequently, the update rule for V is derived as follows:
V i j ( t + 1 ) = S i j ( t ) γ V L 1 , S i j ( t ) > γ V L 1 , 0 , S i j ( t ) γ V L 1 .

3.3.4. Fix G , U R + I 1 × J 1 and V R + I 2 × J 2 , Solve W R + I 3 × J 3

Similarly, to update the factor matrix W , we write the objective function Q ( W ) as follows:
Q ( W ) = Q ^ ( W ) + γ W W 1 1 ,
where
Q ^ ( W ) = X ( 3 ) W G ( 3 ) ( V U ) F 2 + α W W F 2 + λ W T r ( W L W W ) + T r ( λ W ) .
Since Q ^ ( W ) in Equation (25) is a convex function regarding W , and satisfies the L-Lipschitz condition [28], there exists a constant L 2 0 such that
Q ^ ( W ) Q ^ ( W ) L 2 W W , ( W , W ) ,
where Q ^ ( W ) is the derivative of W . Q ^ ( W ) is to be by the approximated second-order Taylor expansion near W as
Q ˜ ( W ) Q ^ ( W ) + L 2 2 ( W W ) 2 + Q ^ ( W ) ( W ) ( W W ) = L 2 2 ( W ( W 1 L 2 Q ^ ( W ) W ) ) 2 + Θ ( W ) , Θ ( W ) = Q ^ ( W ) 1 2 L 2 ( Q ^ ( W ) W ) 2 .
Let W = W ( t ) , then, the minimum value of Equation (28) is obtained as W ( t + 1 ) and
W ( t + 1 ) = W ( t ) 1 L 2 Q ^ ( W t ) ( W t ) .
Similarly, since Ψ ( W ( t ) ) is independent of W t , the objective function updated with W can be written as
W ( t + 1 ) = arg min W [ L 2 2 W ( W ( t ) 1 L 2 Q ^ ( W ( t ) ) W ( t ) ) 2 2 + γ W W 1 1 ] .
Let R ( t ) = W ( t ) 1 L 2 Q ^ ( W t ) W ( t ) , then, for solving Equation (30), we first solve R ( t ) and then
W ( t + 1 ) = arg min W [ L 2 2 W R ( t ) 2 2 + γ W W 1 1 ] .
Now, the partial derivative of Q ^ ( W ( t ) ) to W ( t ) is
Q ^ ( W ( t ) ) ( W ( t ) ) = 2 ( X ( 3 ) ( t ) W ( t ) G ( 3 ) ( t ) ( V ( t ) U ( t ) ) ) ( G ( 3 ) ( t ) ( V ( t ) U ( t ) ) ) + 2 α W W ( t ) + 2 λ W L W ( t ) W ( t ) + λ W ( t ) = 2 X ( 3 ) ( t ) ( V ( t ) U ( t ) ) G ( 3 ) ( t ) + 2 W ( t ) G ( 3 ) ( t ) ( V ( t ) U ( t ) ) ( V ( t ) U ( t ) ) G ( 3 ) ( t ) + 2 α 3 W ( t ) + 2 λ W ( D W ( t ) C W ( t ) ) W ( t ) + λ W ( t ) ,
where D W ( t ) and C W ( t ) should be solved at the same time. According to the (Karush–Kuhn–Tucker) (KKT) condition [27], λ W W ( t ) = 0 , and  Q ^ ( W ( t ) ) W ( t ) = 0 , we can obtain the following:
( 2 X ( 3 ) ( t ) ( V ( t ) U ( t ) ) G ( 3 ) ( t ) 2 λ W C W ( t ) W ( t ) + 2 W ( t ) G ( 3 ) ( t ) ( V ( t ) U ( t ) ) ( V ( t ) U ( t ) ) G ( 3 ) ( t ) + 2 α W A ( 3 ) ( t ) + 2 λ W D W ( t ) W ( t ) ) W ( t ) = 0 .
Therefore, the update rule of R ( t ) is
R i j ( t ) = W i j ( t ) [ X ( 3 ) ( t ) ( V ( t ) U ( t ) ) G ( 3 ) ( t ) + λ W C W ( t ) W ( t ) ] i j [ W ( t ) G ( 3 ) ( t ) ( V ( t ) U ( t ) ) ( V ( t ) U ( t ) ) G ( 3 ) ( t ) + α W W ( t ) + λ W D W ( t ) W ( t ) ] i j .
Similarly, we minimize R i j ( t ) and W 1 1 using gradient descent and soft thresholding, respectively. This leads to the following update rule for W :  
W i j ( t + 1 ) = R i j ( t ) γ W L 2 , R i j ( t ) > γ W L 2 , 0 , R i j ( t ) γ W L 2 .

3.4. Algorithm Design

According to the above reasoning, a clustering algorithm based on GDSNTD is summarized, with its detailed steps outlined in Algorithm 1.
Algorithm 1: GDSNTD for Clustering
  • Input:
  •       Data tensor X R + I 1 × I 2 × I 3 . The algorithm parameters J 1 , J 2 , J 3 , p, and regularization
  •      parameters α V , α W , λ V , λ W , γ V , γ W . The stopping criterion and the maximum number
  •       of iterations maxiter.
  • Output: Core tensor G and factors U , V , W ;
  • Ensure:
  •      Clustering results;
  •      repeat
  •      for k = 1 , 2 , , maxiter do
  • 1:     Update G , U , V , W , and using (8), (13), (24), (35), respectively.
  •       if r e < ϵ
  •            Return G , U , V , W
  •       end if
  •      end for
  •      until converges
  • 2:     Run k-means on W to get clustering result.
  • 3:     return Clustering results
We employ a stopping criterion based on the relative change in the objective function value. The criterion is defined as
r e = f ( k 1 ) f ( k ) f ( k ) ,
where f ( k 1 ) and f ( k ) are the objective function values at the ( k 1 ) -th and k-th iterations, respectively. The iterations are terminated when r e < ϵ , with ϵ being a predefined convergence tolerance [29].

3.5. Convergence Analysis

Since the update rule for G is identical to that in [17], this subsection employs the auxiliary function approach exclusively to prove the convergence of the updating rules given in Equations (8), (13), (24) and (35). We now present the following theorem:
Theorem 1.
For G 0 , U 0 , V 0 , and W 0 , the objective function in (2) is non-increasing under the updating rules in (8), (13), (24), and (35) if and only if U 0 , V 0 , W 0 , and G are at a stationary point.
To prove Theorem 1, we first give a definition and several lemmas.
Definition 1
([1]). H ( a , a ) is an auxiliary function for J ( a ) if the conditions H ( a , a ) J ( a ) and H ( a , a ) = J ( a ) are satisfied.
Lemma 1
([1]). If H ( a , a ) is an auxiliary function for J ( a ) , then J ( a ) is non-increasing under the updating rule
a t + 1 = a r g m i n a H ( a , a t ) .
Proof. 
J ( a t + 1 ) H ( a t + 1 , a t ) H ( a t , a t ) J ( a t ) . □
The equality J ( a t + 1 ) = J ( a t ) holds only if a t is a local minimum of H ( a , a t ) . By iterating the update rule (36), a t converges to the local minimum of H ( a , a t ) .
Next, we provide a detailed proof of the convergence for the above updating rule for U using an auxiliary function. Recall the function from Equation (3):
J ^ ( U ) = X ( 1 ) U G ( 1 ) ( W V ) F 2 .
We can obtain
J ^ i j ( U i j ( t ) ) = J ^ ( U i j ( t ) ) ( U i j ( t ) ) = [ 2 X ( 1 ) ( t ) ( W V ) ( t ) G ( 1 ) ( t ) + 2 U ( t ) G ( 1 ) ( t ) ( ( W V ) ( t ) ( W V ) ( t ) ) G ( r ) ( t ) ] i j .
J ^ i j ( U i j ( t ) ) = [ 2 G ( 1 ) ( t ) ( ( W V ) ( t ) ( W V ) ( t ) ) G ( 1 ) ( t ) ] j j .
Essentially, it is sufficient to prove that each J ^ i j ( U i j ) is non-increasing under the update rule.
Lemma 2.
The function
H ^ ( U i j , U i j t ) = J ^ i j ( U i j t ) + J ^ i j ( U i j t ) ( U i j U i j t ) + [ U t G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) ] j j U i j t ( U i j U i j t ) 2
is an auxiliary function for J ^ i j ( U i j ) , where the matrix U t = ( U i j t ) .
Proof. 
Obviously, H ˜ ( U i j t , U i j t ) = J ^ i j ( U i j t ) . According to Definition 1, we only need to show that H ˜ ( U i j , U i j t ) J ^ i j ( U i j ) . We first obtain the Taylor series expansion of J ^ i j ( U i j ) at U i j t to be
J ˜ i j ( U i j ) = J i j ( U i j t ) + J ^ i j ( U i j t ) ( U i j U i j t ) + 1 2 J i j ( U i j t ) ( U i j U i j t ) 2 = J i j ( U i j t ) + J i j ( U i j t ) ( U i j U i j t ) + [ G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) ] j j ( U i j U i j t ) 2 .
By Equation (40), H ( U i j , U i j t ) J i j ( U i j ) is equivalent to
[ U t G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) ] j j U i j t [ G ( r ) ( ( W V ) ( W V ) ) G ( 1 ) ] j j .
We can obtain
[ U t G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) ] j j = U i l t [ G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) ] l j U i j t [ G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) ] j j ,
Therefore, the inequality H ˜ ( U i j , U i j t ) J ^ i j ( U i j ) holds. □
Analogously, the convergence under the update rule in Equation (24) can be proven. Next, we provide a detailed proof of convergence for the V update rule using an auxiliary function. Consider the following function:
J ( V ) = J ^ + γ V V 1 1 ,
where
J ^ ( V ) = X ( 2 ) V G ( 2 ) ( W U ) F 2 + α V U F 2 + λ V T r ( V L V V ) ,
We derive
J ^ i j ( V i j ( t ) ) = J ^ i j ( V i j ( t ) ) ( V i j ( t ) ) = [ 2 X ( 2 ) ( t ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + 2 V ( t ) G ( 2 ) ( t ) ( ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) ) G ( 2 ) ( t ) + 2 α V V ( t ) + 2 λ V D V ( t ) V ( t ) 2 λ V C V ( t ) V ( t ) ] i j ,
J ^ i j ( V i j ( t ) ) = [ 2 G ( 2 ) ( t ) ( W ( t ) U ( t ) ) T ( W ( t ) U ( t ) ) G ( 2 ) ( t ) T + 2 α V + 2 λ V D V ( t ) 2 λ V C V ( t ) ] j j .
Essentially, the updating rule is element wise, so it is sufficient to prove each J ^ i j ( V i j ) is non-increasing under the update rule.
Lemma 3.
The function
H ^ ( V i j , V i j t ) = J ^ i j ( V i j t ) + J ^ i j ( V i j t ) ( V i j V i j t ) + M 1 + M 2 V i j t ( V i j V i j t ) 2
is an auxiliary function for J ^ i j ( V i j ) , where the matrix V t = ( V i j t ) , and
M 1 = V ( t ) G ( 2 ) ( t ) ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) ,
M 2 = α V V ( t ) + λ V D V ( t ) V ( t ) .
Proof. 
Obviously, H ˜ ( V i j t , V i j t ) = J ^ i j ( V i j t ) . According to Definition 1, we only need to show that H ˜ ( V i j , V i j t ) J ^ i j ( V i j ) . We first obtain the Taylor series expansion of J ^ i j ( V i j ) at V i j t to be
J ˜ i j ( V i j ) = J i j ( V i j t ) + J ^ i j ( V i j t ) ( V i j V i j t ) + 1 2 J i j ( V i j t ) ( V i j V i j t ) 2 = J i j ( V i j t ) + J i j ( V i j t ) ( V i j V i j t ) + [ G ( 2 ) ( t ) ( ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) ) G ( 2 ) ( t ) + α V + λ V D V ( t ) λ V C V ( t ) ] j j ( V i j V i j t ) 2 .
By Equation (48) and Equation (49), H ( V i j , V i j t ) J i j ( V i j ) is equivalent to
[ M 1 + M 2 ] i j V i j t [ G ( 2 ) ( t ) ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + α V + λ V D V ( t ) λ V C V ( t ) ] j j .
We can obtain
[ V ( t ) G ( 2 ) ( t ) ( ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) ) G ( 2 ) ( t ) ] i j = V i l t [ G ( 2 ) ( t ) ( ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) ) G ( 2 ) ( t ) ] l j V i j t [ G ( 2 ) ( t ) ( ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) ) G ( 2 ) ( t ) ] j j ,
and
[ λ V D V ( t ) V ( t ) ] i j = V i l t [ λ V D V ( t ) ] l j V i j t [ λ V D V ( t ) ] j j V i j t [ λ V ( D V ( t ) V V ( t ) ) ] j j ,
Therefore, the inequality G ( V i j , V i j t ) F i j ( V i j ) holds. □
Thus, Equation (50) holds, and H ( V i j , V i j t ) J i j ( V i j ) . According to Equation (23),
S i j ( t ) = V i j ( t ) [ X ( 2 ) ( t ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + λ V C V ( t ) V ( t ) ] i j [ V ( t ) G ( 2 ) ( t ) ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + α V V ( t ) + λ V D V ( t ) V ( t ) ] i j = V i j ( t ) V i j ( t ) J ^ i j ( V i j ( t ) ) [ V ( t ) G ( 2 ) ( t ) ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + α V V ( t ) + λ V D V ( t ) V ( t ) ] i j .
The soft thresholding operation ensures convergence as follows: if S i j ( t ) > γ V L , the update subtracts a constant value; otherwise, if S i j ( t ) γ V L , the value is set to zero. Both cases are consistent with the convergence proof. Because Equation (51) is an auxiliary function for J ^ i j , and according to the soft threshold, J i j is non-increasing under the updating rule stated in Equation (24). Analogously, we can prove the convergence under the updating rule in Equation (35).

4. Experiments

In this section, the experimental setup and analysis of the results are discussed in detail. The descriptions of these datasets are given as follows. First, we introduce the image datasets Coil20 (http://www.cad.zju.edu.cn/home/dengcai/Data/MLData.html, (accessed on 15 May 2024)), Georgia (https://www.cnblogs.com/kuangqiu/p/7776829.html, (accessed on 10 March 2025)), Iris (http://archive.ics.uci.edu/dataset/53/iris, (accessed 20 November 2024)) [30]. The datasets are summarized in Table 1.
The details of the public datasets used in the experiments are presented in Figure 1 and Figure 2.
In order to evaluate the effectiveness of our proposed GDSNTD scheme, we compared it with six classical or state-of-the-art clustering and co-clustering methods. All the simulations were performed on a computer with a 2.30-GHz Intel Core i7-11800H CPU and 32 GB memory of 64-bit MATLAB 2016a in Windows 10. Without special specifications, the maximum number of iterations is set to 1000.
  • Non-negative matrix factorization (NMF) [1]: NMF aims to decompose a matrix into two low-dimensional matrices and is now often used as a data processing method in machine learning.
    F N M F = min U 0 , V 0 X U V .
  • Non-negative Tucker decomposition (NTD) [7]: The NTD algorithm is considered as a generalization of NMF.
    F N T D = min G 0 , U , V , W 0 X G × 1 U × 2 V × 3 W F 2 .
  • Graph-regularized NTD (GNTD) [17]:
    F G N T D = min G 0 , U , V , W 0 X G × 1 U × 2 V × 3 W F 2 + λ T r ( W L W ) .
  • Graph dual regularized NMF (GDNMF) [12]: GDNMF simultaneously considers the geometric structures of both the data manifold and the feature manifold.
    F G D N M F = min U 0 , V 0 X U V F 2 + λ 1 T r ( V L V V ) + λ 2 T r ( U L U U ) .
  • Graph dual regularized non-negative matrix tri-factorization (GDNMTF) [12]: DNMTF is an extension of our DNMF algorithm and simultaneously incorporates two graph regularizers of both data manifold and feature manifold into its objective function.
    F G D N M T F = min U 0 , S 0 , V 0 X U S V F 2 + λ 1 T r ( V L V V ) + λ 2 T r ( U L U U ) .
  • Graph-regularized sparse non-negative matrix trifactorization (GSNMTF) [21]: The GSNMTF model introduces graph regularization and an l 1 norm constraint into the objective function.
    F G S N M T F = min U 0 , S 0 , V 0 X U S V F 2 + λ U T r ( U L U U ) + λ V T r ( V L V V ) +
    γ U U 1 1 + γ V V 1 1 .

4.1. Evaluation Measures

In this section, two widely used metrics, accuracy (AC) and Normalized Mutual Information (NMI), are used to evaluate the clustering performance. Accuracy aims to find a one-to-one relationship between classes and clusters. It is usually used to calculate the proportion of correct samples to the total number of samples. It is a relatively intuitive evaluation index. Accuracy is a clustering evaluation metric that measures the proportion of correctly assigned samples against the total, after a one-to-one mapping between clusters and true classes is established. It is an intuitive and widely used measure [31]. NMI is commonly used in clustering to measure the similarity between two clusterings.
AC is defined as
A C = 1 N i = 1 N δ ( g t i , m a p ( c i ) ) ,
where N is the total number of samples, g t i is the ground-truth label of sample i, c i is the cluster label assigned by the algorithm, δ ( · , · ) is the Dirac delta function, and m a p ( · ) is the optimal mapping function [32].
NMI is defined as
N M I ( B , T ) = 2 · ( I ( B ; T ) ) H ( B ) + H ( T ) ,
where I ( B ; T ) is the mutual information between B and T, and H ( B ) and H ( T ) are their entropies. A higher NMI indicates a better alignment between the clustering result and the true labels [33].

4.2. Experimental Setup and Clustering Results Analysis

In this part, the experimental setup is described, and the experimental results are discussed in detail. To ensure the fairness between models, our proposed algorithm and all comparison algorithms use the same random initialization matrix. Subsequently, each experiment is conducted ten times independently on original data, then K-means clustering is performed five times independently on this low-dimensional reduced data. The average and variance of the results are recorded. The standard deviation is set to 0 if it is less than 1 × 10 5 .
Table 2 and Table 3 summarize the accuracy and NMI results for each algorithm across all datasets. Table 4 lists the accuracy and standard deviation for each algorithm on the Georgia dataset, while Table 5 presents the corresponding NMI results. Similarly, results for the COIL20 dataset are detailed in Table 6 (accuracy) and Table 7 (NMI). From the results, we derive the following main conclusions:
1.
From Table 2, we can observe that the value of AC of GDSNTD is better than that of other methods on most datasets. The value of NMI reflects the proportion of correct decisions, which also proves that the proposed model performs better. The performance improvement is evident in the Georgia dataset, where GDSNTD improves the accuracy by 48.92% over the NMF co-clustering algorithm. The accuracy of GDSNTD is also improved by 26.1%, 5.15%, 7.02%, 3.71%, and 2.82% compared with other co-clustering algorithms NTD, GNTD, GDNMF, GDNMTF, and GSNMTF, respectively. In the Coil20 dataset, the accuracy of GDSNTD is also improved by 30.50%, 38.58%, 2.29%, 0.81%, 1.14%, and 0.57%, compared with other co-clustering algorithms NTD, GNTD, GDNMF, GDNMTF, and GSNMTF, respectively. In the Iris dataset, the accuracy of GDSNTD is improved by 17.22%, 32.77%, 2.84%, 2.10%, 0.17%, and 0.12% compared with other co-clustering algorithms NMF, NTD, GNTD, GDNMF, GDNMTF, and GSNMTF, respectively.
2.
From Table 3, we can observe that our proposed method of GDSNTD can also achieve higher accuracy. The NMI of GDSNTD is as high as 50.16% on the Iris dataset, which is 34.95%, 7.46%, 4.27%, 3.49%, and 3.42% better compared with NMF, GNTD, GDNMF, GDNMTF, and GSNMTF, respectively. In the Georgia dataset, the accuracy of GDSNTD is also improved by 26.67%, 15.01%, 3.12%, 4.79%, 4.54%, and 4.26% compared with other co-clustering algorithms NMF, NTD, GNTD, GDNMF, GDNMTF, and GSNMTF, respectively. In the Coil20 dataset, the accuracy of GDSNTD is improved by 21.02%, 23.85%, 2.75%, 0.3%, 0.2%, and 0.32%, compared with other co-clustering algorithms NMF, NTD, GNTD, GDNMF, GDNMTF, and GSNMTF, respectively.
3.
Table 4 shows that the average AC of GDSNTD is higher than that of the other methods on most datasets. In the Georgia dataset, the accuracy of the co-clustering algorithm GDSNTD is improved by an average of 27.4%, 22.6%, 2.2%, 8.7%, 5.11%, and 3.79% compared with NMF, NTD, GNTD, GDNMF, GDNMTF, and GSNMTF, respectively. From Table 6, it can be observed that on the Coil20 dataset, the accuracy of GDSNTD is, on average, 23.9%, 30.9%, 1.49%, 4.0%, 5.79%, and 3.18% higher than that of NMF, NTD, GNTD, GDNMF, GDNMTF, and GSNMTF, respectively.
4.
From Table 5, it can be observed that on the Georgia dataset, the accuracy of GDSNTD is, on average, 23.7%, 19.4%, 1.82%, 7.3%, 5.67%, and 4.63% higher than that of NMF, NTD, GNTD, GDNMF, GDNMTF, and GSNMTF, respectively. From Table 7, it can be observed that on the Coil20 dataset, the accuracy of GDSNTD is, on average, 20.8%, 31.5%, 1.42%, 3.2%, 6.14%, and 2.76% higher than that of NMF, NTD, GNTD, GDNMF, GDNMTF, and GSNMTF, respectively.
Figure 3 shows the clustering performance on the Georgia, COIL20, and Iris datasets. It can be observed that the proposed GDSNTD method surpasses all other compared methods.

4.3. Interpretation of GDSNTD’s Superior Performance

Across all datasets, our proposed method consistently outperforms all competing baselines of NMF, NTD, GNTD, GDNMF, GDNMTF, and GSNMTF in both standard clustering and co-clustering tasks. This empirical evidence strongly suggests that the GDSNTD model with dual graph constraints successfully leads to a more structured and discriminative latent space, thereby improving clustering accuracy.
First, applying manifold constraints to the factor matrices is equivalent to preserving the structure in the “essential features” of the data, resulting in greater accuracy. Second, l 1 regularization counterbalances the Frobenius norm, leading to a more stable optimization process and yielding factors within a more reasonable numerical range. Furthermore, l 1 regularization promotes the learning of “parts” that correspond to local structures within the data, thereby promoting a decomposition with enhanced coherence.

4.4. Parameter Selection

Every parameter is searched over a range of values from 0 to 10 5 . The choice of parameters has a pronounced effect on experimental performance; therefore, they were selected systematically through a grid search. Figure 4 shows the process of determining the optimal parameters using a grid search on the Iris dataset. From Figure 4, we can see that the accuracy is only 0.7772 when α V = α W = 10 0 and γ V = γ W = 10 0 . When the coefficient of the sparse regularization term is γ V = γ W = 10 2 , the accuracy improves to 0.7980 . From Figure 5, it is to be observed that the NMI is 0.8615 when α V = α W = 10 3 and γ V = γ W = 10 1 , and when the coefficient of the sparse regularization term is γ V = γ W = 10 2 , the accuracy is only 0.8776 .

4.5. Convergence Study

As described in Section 3, the convergence of the proposed algorithms has been theoretically proved. In this subsection, we experimentally analyze the convergence of the proposed algorithms by examining the relationship between the number of iterations and the value of the objective function. This relationship is visualized in Figure 6, Figure 7 and Figure 8. The convergence behavior of GSNTD on the Georgia dataset is illustrated in Figure 6. The convergence behavior of GSNTD on the Coil20 dataset is illustrated in Figure 7. The convergence behavior of GSNTD on the Iris dataset is illustrated in Figure 8. The observed monotonic decrease in the objective function value demonstrates that the algorithm converges effectively under the multiplicative update rules. This result provides empirical support for the convergence proof given in Theorem 1. We can find that GDSNTD is usually able to reach convergence within 1000 iterations.

4.6. Complexity Analysis

In this subsection, we analyze the computational complexity of GDSNTD. Note that X is a third-order I 1 × I 2 × I 3 -dimensional tensor, and G is a third-order J 1 × J 2 × J 3 -dimensional core tensor. Factor matrices are U R + I 1 × J 1 , V R + I 2 × J 2 , and W R + I 3 × J 3 .
Consider the updating rule in (8), the operation ( W V U ) v e c ( X ) corresponds to the tensor mode product X × 1 U × 2 V × 3 W . Its computational complexity is O ( I 1 I 2 I 3 J 1 + I 2 I 3 J 1 J 2 + I 3 J 1 J 2 J 3 ) . ( W V U ) ( W V U ) v e c ( G ) corresponds to v e c ( G × 1 U U × 2 V V × 3 W W ) . The computational complexity is J 1 J 2 J 3 ( J 1 + J 2 + J 3 ) . The total computational complexity of computing the update of G is bounded by O ( I 1 I 2 I 3 J 1 + I 2 I 3 J 1 J 2 + I 3 J 1 J 2 J 3 + J 1 J 2 J 3 ( J 1 + J 2 + J 3 ) ) .
Consider the updating rule in (13), X ( 1 ) ( W V ) G ( 1 ) corresponds to ( X × 2 V × 3 W ) ( 1 ) G ( 1 ) , which needs O ( I 1 I 2 I 3 J 2 + I 1 I 3 J 2 J 3 + I 1 J 1 J 2 J 3 ) operations. U G ( 1 ) ( ( W V ) ( W V ) ) G ( 1 ) corresponds to ( G × 1 U × 2 V V × 3 W W ) ( 1 ) G ( 1 ) , which needs O ( J 1 2 J 2 J 3 + J 1 J 2 2 J 3 + J 1 J 2 J 3 2 + I 1 J 1 2 ) operations. The total computational complexity of computing the update of U is bounded by O ( I 1 I 2 I 3 J 2 + I 1 I 3 J 2 J 3 + I 1 J 1 J 2 J 3 + J 1 2 J 2 J 3 + J 1 J 2 2 J 3 + J 1 J 2 J 3 2 + I 1 J 1 2 ) .
Consider the updating rule in (23), X ( 2 ) ( t ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + λ V C V ( t ) V ( t ) corresponds to ( X × 1 U × 3 W ) ( 2 ) G ( 2 ) + λ V C V V , and it takes O ( I 1 I 2 I 3 J 1 + I 2 I 3 J 1 J 3 + I 2 J 1 J 2 J 3 + I 2 2 J 2 ) operations. V ( t ) G ( 2 ) ( t ) ( W ( t ) U ( t ) ) ( W ( t ) U ( t ) ) G ( 2 ) ( t ) + α V V ( t ) + λ V D V ( t ) V ( t ) corresponds to ( G × 1 U U × 2 V × 3 W W ) ( 2 ) G ( 2 ) + α V V + λ V D V V , and it takes O ( J 1 2 J 2 J 3 + J 1 J 2 2 J 3 + J 1 J 2 J 3 2 + I 2 J 2 2 + I 2 2 J 2 + I 2 J 2 ) operations. The total computational complexity of computing the update of V is bounded by O ( I 1 I 2 I 3 J 1 + I 2 I 3 J 1 J 3 + I 2 J 1 J 2 J 3 + 2 I 2 2 J 2 + J 1 2 J 2 J 3 + J 1 J 2 2 J 3 + J 1 J 2 J 3 2 + I 2 J 2 2 + I 2 J 2 ) .
Similarly, consider the updating rule in (34), the total computational complexity of computing the update of W is bounded by O ( I 1 I 2 I 3 J 1 + I 2 I 3 J 1 J 2 + I 3 J 1 J 2 J 3 + 2 I 3 2 J 3 + J 1 2 J 2 J 3 + J 1 J 2 2 J 3 + J 1 J 2 J 3 2 + I 3 J 3 2 + I 3 J 3 ) .
Therefore, the total computational complexity of the proposed method is O ( I 1 I 2 I 3 ( 3 J 1 + J 2 ) + I 2 I 3 ( 2 J 2 + J 3 ) + I 1 I 3 J 2 J 3 + J 1 J 2 J 3 ( 2 I 2 + I 1 + I 2 + 4 ( J 1 + J 2 + J 3 ) ) + 2 I 2 2 J 2 + 2 I 3 2 J 3 + I 1 J 1 2 + I 2 J 2 2 + I 3 J 3 2 + I 2 J 2 + I 3 J 3 ) . Compared with GNTD, the total computational complexity of the GDSNTD is increased by 2 I 2 2 J 2 + 2 I 3 2 J 3 + I 2 J 2 + I 3 J 3 .
Figure 9 shows the relative performance in terms of clustering ACC and time consumption among NTD, GNTD, GDNMTF, GSNMTF, and GDSNTD on the Georgia dataset. Figure 10 shows the relative performance in terms of clustering NMI and time consumption among NTD, GNTD, GDNMTF, GSNMTF, and GDSNTD on the Coil20 dataset. In the figure, CPU(s) is the running time in seconds.

5. Conclusions

In this paper, we propose a co-clustering method via a dual graph-regularized sparse non-negative Tucker decomposition (GDSNTD) framework. It incorporates the Frobenius norm, graph regularization, and an l 1 norm constraint. The graph regularization term is used to maintain the intrinsic geometric structure of the data. This approach allows the model to represent the internal structure of high-dimensional data more accurately. The model not only adds the l 1 norm to achieve the adjustment of data eigenvalues but also embeds the Frobenius norm to improve the generalization ability of the model. We then present a detailed derivation of the solution, design the accompanying algorithm for GDSNTD, and provide a convergence proof. The experimental results show that selecting appropriate parameters can improve the model’s classification accuracy. This indicates that the selected rules have practical significance. Experiments on public datasets demonstrate that our proposed model achieves superior image clustering performance. The proposed method outperforms the state-of-the-art method GSNMTF by an average of 3.79% in clustering accuracy and 4.63% in Normalized Mutual Information on the Georgia dataset.

6. Future Work

Although the proposed GSNTTD method has shown good clustering performance, it operates on the assumption that the data resides in a well-formed latent vector space [34]. In recent years, deep learning-based clustering has experienced rapid advancement. The current popular deep learning-based methods are dominated by several key approaches, such as Graph Autoencoders (GAEs), graph convolutional network-based clustering (GCN-based clustering), and Deep Tensor Factorization.
GAEs aim to learn low-dimensional latent representations (embeddings) of graph-structured data in an unsupervised manner. It learns to compress input data into a lower-dimensional latent representation (encoding) and then reconstruct the original input from this representation (decoding). The critical distinction is that the input data is a graph, characterized by its topology (structure) and node attributes (features). Kipf et al. in [35] formally introduced the GAE and its variational counterpart, the variational graph autoencoder (VGAE). The field of GAE has evolved rapidly from foundational models to highly sophisticated and specialized architectures. Recent advancements have focused on developing more sophisticated architectures to address specific challenges. Zhou K et al. proposed a new causal representation method based on a graph autoencoder embedded autoencoder (GeAE). The GeAE employs a causal structure learning module to account for non-linear causal relationships present in the data [36].
Kipf and Welling in [35] introduced the variational graph autoencoder (VGAE), a framework for unsupervised learning on graph-structured data based on the variational autoencoder (VAE). In this work, they demonstrated this VGAE model using a GCN encoder and a simple inner product decoder. In [37], Kipf and Welling first introduced the efficient, first-order approximation-based graph convolutional layer. This formulation has laid the groundwork for numerous subsequent GCN studies. The core idea of a GCN encoder is to learn low-dimensional vector representations (i.e., node embeddings) for nodes by propagating and transforming information across the graph structure. The node’s embedding is determined not only by its own features but also jointly by the features of its neighbor nodes and the local graph structure. The core principle of GCN-based clustering is to unify node representation learning and cluster assignment into an end-to-end, jointly optimized framework and perform these tasks simultaneously.
The principle of Deep Tensor Factorization can be understood as using a deep neural network to perform the factorization and reconstruction process. Wu et al. proposed a Neural Tensor Factorization model, which incorporates a multi-layer perceptron structure to learn the non-linearities between different latent factors [38]. In [39], Jiang et al. proposed a generic architecture of deep transfer tensor factorization (DTTF), where the side information is embedded to provide effective compensation for the tensor sparsity.
In [40], Ballard et al. categorized feedforward neural networks, graph convolutional neural networks, and autoencoders into non-generative deep learning-based multi-omics integration methods. They found that deep learning-based approaches build off of previous statistical methods to integrate multi-omics data by enabling the modeling of complex and non-linear interactions between data types. Therefore, deep learning-based clustering has emerged as a rapidly advancing field. It integrates conventional cluster analysis with deep learning, leveraging its powerful feature representation and non-linear mapping capabilities to demonstrate superior performance in unsupervised clustering tasks. In the future, we intend to incorporate recent deep learning-based methods into our subsequent research to explore more up-to-date clustering methodologies.

Author Contributions

Conceptualization, J.H. and L.L.; methodology, L.L.; writing—original draft preparation J.H. and L.L. All authors have read and agreed to the published version of the manuscript.

Funding

The authors acknowledge the support from the National Natural Science Foundation of China under Grants (Nos. 12361044, 12161020, 12061025), Basic Research Project of Science and Technology Plan of Guizhou of China under Grants Qian Ke he foundation ZK[2023] General 022.

Data Availability Statement

The original contributions presented in this study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Lee, D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788–791. [Google Scholar] [CrossRef] [PubMed]
  2. Friedlander, M.P.; Hatz, K. Computing non-negative tensor factorizations. Optim. Methods Softw. 2008, 23, 631–647. [Google Scholar] [CrossRef]
  3. Shcherbakova, E.M.; Matveev, S.A.; Smirnov, A.P.; Tyrtyshnikov, E.E. Study of performance of low-rank nonnegative tensor factorization methods. Russ. J. Numer. Anal. Math. Model. 2023, 38, 231–239. [Google Scholar] [CrossRef]
  4. De Lathauwer, L. A survey of tensor methods. In Proceedings of the 2009 IEEE International Symposium on Circuits and Systems, Taipei, Taiwan, 24–27 May 2009; pp. 2773–2776. [Google Scholar] [CrossRef]
  5. Zhou, G.; Cichocki, A.; Zhao, Q.; Xie, S. Nonnegative matrix and tensor factorizations: An algorithmic perspective. IEEE Signal Process. Mag. 2014, 31, 54–65. [Google Scholar] [CrossRef]
  6. Kolda, T.G.; Bader, B.W. Tensor decompositions and applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  7. Kim, Y.D.; Choi, S. Nonnegative Tucker decomposition. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; IEEE: New York, NY, USA, 2007; pp. 1–8. [Google Scholar] [CrossRef]
  8. Cai, D.; He, X.; Han, J.; Huang, T.S. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1548–1560. [Google Scholar] [CrossRef]
  9. Sun, F.; Xu, M.; Hu, X.; Jiang, X. Graph regularized and sparse nonnegative matrix factorization with hard constraints for data representation. Neurocomputing 2016, 173, 233–244. [Google Scholar] [CrossRef]
  10. Sun, J.; Wang, Z.H.; Sun, F.; Li, H.J. Sparse dual graph-regularized NMF for image co-clustering. Neurocomputing 2018, 316, 156–165. [Google Scholar] [CrossRef]
  11. Shang, F.; Jiao, L.C.; Wang, J. Graph dual regularization non-negative matrix factorization for co-clustering. Pattern Recognit. 2012, 45, 2237–2250. [Google Scholar] [CrossRef]
  12. Long, X.; Lu, H.; Peng, Y.; Li, W. Graph regularized discriminative non-negative matrix factorization for face recognition. Multimed. Tools Appl. 2014, 72, 2679–2699. [Google Scholar] [CrossRef]
  13. Saberi-Movahed, F.; Berahm, K.; Sheikhpour, R.; Li, Y.; Pan, S. Nonnegative matrix factorization in dimensionality reduction: A survey. arXiv 2024, arXiv:2405.03615. [Google Scholar] [CrossRef]
  14. Jing, W.J.; Lu, L.; Ou, W. Semi-supervised non-negative matrix factorization with structure preserving for image clustering. Neural Netw. 2025, 187, 107340. [Google Scholar] [CrossRef]
  15. Li, X.; Ng, M.K.; Cong, G.; Ye, Y.; Wu, Q. MR-NTD: Manifold Regularization Nonnegative Tucker Decomposition for Tensor Data Dimension Reduction and Representation. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 1787–1800. [Google Scholar] [CrossRef]
  16. Yin, W.; Ma, Z. LE & LLE Regularized Nonnegative Tucker Decomposition for clustering of high dimensional datasets. Neurocomputing 2019, 364, 77–94. [Google Scholar] [CrossRef]
  17. Qiu, Y.; Zhou, G.; Zhang, Y.; Xie, S. Graph regularized nonnegative Tucker decomposition for tensor data representation. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; IEEE: New York, NY, USA, 2019; pp. 8613–8617. [Google Scholar] [CrossRef]
  18. Dhillon, I.S.; Mallela, S.; Modha, D.S. Information-theoretic co-clustering. In Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 24–27 August 2003; pp. 89–98. [Google Scholar] [CrossRef]
  19. Whang, J.J.; Dhillo, I.S. Non-exhaustive, Overlapping co-clustering. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 2367–2370. [Google Scholar] [CrossRef]
  20. Del Buono, N.; Pio, G. Non-negative matrix tri-factorization for co-clustering: An analysis of the block matrix. Inf. Sci. 2015, 301, 13–26. [Google Scholar] [CrossRef]
  21. Deng, P.; Li, T.R.; Wang, H.; Wang, D.; Horng, S.J.; Liu, R. Graph Regularized Sparse Non-Negative Matrix Factorization for Clustering. IEEE Trans. Comput. Soc. Syst. 2023, 10, 910–921. [Google Scholar] [CrossRef]
  22. Chachlakis, D.G.; Dhanaraj, M.; Prater-Bennette, A.; Markopoulos, P.P. Dynamic L1-norm Tucker tensor decomposition. IEEE J. Sel. Top. Signal Process. 2021, 15, 587–602. [Google Scholar] [CrossRef]
  23. Peng, X.; Lu, C.Y.; Yi, Z.; Tang, H.J. Connections Between Nuclear-Norm af Frobenius-Norm-Based Representations. IEEE Trans. Neural Netw. Learn. Syst. 2016, 29, 218–224. [Google Scholar] [CrossRef]
  24. Ahmed, T.; Raja, H.; Bajwa, W.U. Tensor regression using low-rank and sparse Tucker decompositions. SIAM J. Math. Data Sci. 2020, 2, 944–966. [Google Scholar] [CrossRef]
  25. Chachlakis, D.G.; Prater-Bennette, A.; Markopoulos, P.P. L1-norm Tucker tensor decomposition. IEEE Access 2019, 7, 178454–178465. [Google Scholar] [CrossRef]
  26. Rockafellar, R.T. Lagrange multipliers and optimality. SIAM Rev. 1993, 35, 183–238. [Google Scholar] [CrossRef]
  27. Stephen, B.; Lieven, V. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar] [CrossRef]
  28. Patrick, L.C.; Valérie, R.W. Signal recovery by proximal forward-backward splitting. Multiscale Model. Simul. 2005, 4, 1168–1200. [Google Scholar] [CrossRef]
  29. Kim, H.; Park, H. Nonnegative matrix factorization based on alternating nonnegativity constrained least squares and active set method. SIAM J. Matrix Anal. Appl. 2008, 30, 713–730. [Google Scholar] [CrossRef]
  30. Asuncion, A.; Newman, D. Iris; UCI Machine Learning Repository: Irvine, CA, USA, 1936. [Google Scholar] [CrossRef]
  31. Cai, D.; He, X.; Han, J. Document clustering using locality preserving indexing. IEEE Trans. Knowl. Data Eng. 2005, 17, 1624–1637. [Google Scholar] [CrossRef]
  32. Yang, C.; Yi, Z. Document clustering using locality preserving indexing and support vector machines. Soft Comput. 2008, 17, 677–683. [Google Scholar] [CrossRef]
  33. Estévez, P.A.; Tesmer, M.; Perez, C.A.; Zurada, J.M. Normalized mutual information feature selection. IEEE Trans. Neural Netw. 2009, 20, 189–201. [Google Scholar] [CrossRef]
  34. Dong, Y.; Deng, Y.; Dong, Y.; Wang, J. A survey of clustering based on deep learning. Comput. Appl. 2022, 42, 1021–1028. [Google Scholar] [CrossRef]
  35. Kipf, T.N.; Welling, M. Variational Graph Auto-Encoders. Neural Information Processing Systems (NeurIPS) Workshop on Bayesian Deep Learning. arXiv 2016, arXiv:1611.07308. [Google Scholar] [CrossRef]
  36. Zhou, K.; Jiang, M.; Gabrys, B.; Xu, Y. Learning causal representations based on a GAE embedded autoencoder. IEEE Trans. Knowl. Data Eng. 2025, 37, 3472–3484. [Google Scholar] [CrossRef]
  37. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907v4. [Google Scholar] [CrossRef]
  38. Wu, X.; Shi, B.; Dong, Y.; Huang, C.; Chawla, N.V. Neural tensor factorization. arXiv 2018, arXiv:1802.04416. [Google Scholar] [CrossRef]
  39. Jiang, P.; Xin, K.; Li, C. Deep Transfer Tensor Factorization for Multi-View Learning. IEEE Int. Conf. Data Min. Work. 2022, 459–466. [Google Scholar] [CrossRef]
  40. Ballard, J.L.; Wang, Z.; Li, W.; Shen, L.; Long, Q. Deep learning-based approaches for multi-omics data integration and analysis. Biodata Min. 2024, 17, 38. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The details of public datasets of Coil20 (a) and Georgia (b).
Figure 1. The details of public datasets of Coil20 (a) and Georgia (b).
Mathematics 13 03494 g001
Figure 2. Different kinds of Iris datasets.
Figure 2. Different kinds of Iris datasets.
Mathematics 13 03494 g002
Figure 3. The clustering performance on each dataset.
Figure 3. The clustering performance on each dataset.
Mathematics 13 03494 g003
Figure 4. Parameter tuning on dataset Coil20 regarding accuracy.
Figure 4. Parameter tuning on dataset Coil20 regarding accuracy.
Mathematics 13 03494 g004
Figure 5. Parameter tuning on dataset Coil20 regarding NMI.
Figure 5. Parameter tuning on dataset Coil20 regarding NMI.
Mathematics 13 03494 g005
Figure 6. Convergence curves of GDSNTD on Georgia dataset.
Figure 6. Convergence curves of GDSNTD on Georgia dataset.
Mathematics 13 03494 g006
Figure 7. Convergence curves of GDSNTD on Coil20 dataset.
Figure 7. Convergence curves of GDSNTD on Coil20 dataset.
Mathematics 13 03494 g007
Figure 8. Convergence curves of GDSNTD on Iris dataset.
Figure 8. Convergence curves of GDSNTD on Iris dataset.
Mathematics 13 03494 g008
Figure 9. Comparison of clustering performance and running time among NTD, GNTD, GDNMTF, GSNMTF, and GDSNTD on the Coil20 dataset.
Figure 9. Comparison of clustering performance and running time among NTD, GNTD, GDNMTF, GSNMTF, and GDSNTD on the Coil20 dataset.
Mathematics 13 03494 g009
Figure 10. Comparison of clustering performance and running time among NTD, GNTD, GDNMTF, GSNMTF, and GDSNTD on the Georgia dataset.
Figure 10. Comparison of clustering performance and running time among NTD, GNTD, GDNMTF, GSNMTF, and GDSNTD on the Georgia dataset.
Mathematics 13 03494 g010
Table 1. Summary of the datasets.
Table 1. Summary of the datasets.
IdxDatasetsSamplesFeaturesClasses
D01Georgia750102450
D02Coil201440102420
D03Iris15043
Table 2. AC (%) of each algorithm on each dataset.
Table 2. AC (%) of each algorithm on each dataset.
DataClassesNMFNTDGNTDGDNMFGDNMTFGSNMTFGDSNTD
D015037.18 ± 2.1743.91 ± 2.1052.66 ± 2.9851.74 ± 2.3953.39 ± 1.8653.85 ± 1.7955.37 ± 2.39
D022061.19 ± 4.0357.62 ± 4.1578.06 ± 3.6379.21 ± 3.6278.95 ± 2.5979.40 ± 3.0579.85 ± 2.79
D03382.13 ± 9.7572.51 ± 8.8193.61 ± 8.9094.29 ± 1.1096.11 ± 6.3896.15 ± 3.0596.27 ± 9.01
Table 3. NMI (%) of each algorithm on each dataset.
Table 3. NMI (%) of each algorithm on each dataset.
DataClassesNMFNTDGNTDGDNMFGDNMTFGSNMTFGDSNTD
D015057.63 ± 1.3563.47 ± 1.3770.79 ± 1.9569.66 ± 1.1769.83 ± 1.0970.02 ± 0.9373.00 ± 1.15
D012072.99 ± 1.8071.32 ± 2.2285.97 ± 1.7088.07 ± 1.4688.15 ± 1.0688.05 ± 1.1488.33 ± 1.07
D03367.53 ± 5.7860.69 ± 3.1084.80 ± 6.1287.40 ± 7.7988.56 ± 4.5988.37 ± 5.6191.13 ± 7.14
Table 4. AC (%) of each algorithm on Georgia dataset.
Table 4. AC (%) of each algorithm on Georgia dataset.
ClassesNMFNTDGNTDGDNMFGDNMTFGSNMTFGDSNTD
380.44 ± 11.3067.56 ± 11.8487.24 ± 14.3879.20 ± 12.6181.29 ± 11.8679.51 ± 12.5291.82 ± 9.18
767.05 ± 8.7658.51 ± 10.0366.36 ± 11.2364.63 ± 7.9767.75 ± 9.1269.22 ± 8.0067.03 ± 12.54
1155.5 ± 7.0161.73 ± 7.8770.29 ± 6.9760.57 ± 6.7061.26 ± 6.0166.41 ± 6.5870.64 ± 8.10
1554.35 ± 7.0253.5 ± 3.5263.31 ± 5.6559.37 ± 6.3659.50 ± 6.6158.45 ± 5.0464.99 ± 5.49
1949.36 ± 3.8252.32 ± 6.8660.45 ± 5.1359.69 ± 5.3058.81 ± 5.2660.64 ± 4.4061.95 ± 4.77
2346.94 ± 4.5149.01 ± 4.2460.99 ± 4.7055.91 ± 5.8854.11 ± 3.7955.43 ± 4.1559.16 ± 5.65
2744.78 ± 3.7649.64 ± 2.5258.21 ± 4.3853.29 ± 4.0257.12 ± 3.3459.71 ± 3.0160.25 ± 4.51
3142.69 ± 3.8546.65 ± 3.1557.18 ± 4.1155.31 ± 3.9356.64 ± 3.1056.68 ± 3.0058.16 ± 3.82
3541.64 ± 2.7047.14 ± 3.6055.05 ± 3.7251.58 ± 4.7156.45 ± 3.7156.94 ± 3.7057.31 ± 3.07
3941.6 ± 2.6844.74 ± 2.6053.66 ± 3.2250.90 ± 2.9954.81 ± 2.9854.57 ± 3.2955.59 ± 3.07
4339.88 ± 2.2543.62 ± 3.1753.35 ± 3.0251.50 ± 2.7054.13 ± 2.4054.19 ± 2.4254.20 ± 2.66
4738.97 ± 2.4243.81 ± 2.4853.56 ± 3.2250.51 ± 2.3852.89 ± 1.8953.17 ± 2.1854.11 ± 2.30
4937.53 ± 1.8043.46 ± 2.2252.60 ± 2.2650.10 ± 2.8553.39 ± 2.3453.28 ± 2.4352.88 ± 2.85
5037.18 ± 2.1743.91 ± 2.1052.66 ± 2.9851.74 ± 2.3953.39 ± 1.8653.85 ± 1.7955.37 ± 2.39
Avg.48.4250.3260.3556.7458.6859.4361.68
Table 5. NMI (%) of each algorithm on Georgia dataset.
Table 5. NMI (%) of each algorithm on Georgia dataset.
ClassesNMFNTDGNTDGDNMFGDNMTFGSNMTFGDSNTD
357.07 ± 20.2749.33 ± 16.3178.08 ± 21.4265.85 ± 18.0765.13 ± 15.6165.00 ± 12.4982.78 ± 15.77
760.4 ± 8.5454.68 ± 10.7365.48 ± 11.1763.46 ± 7.1666.87 ± 10.5167.55 ± 10.1867.1 ± 12.48
1160.44 ± 8.0465.62 ± 6.6875.79 ± 5.9165.82 ± 4.8866.83 ± 4.8570.14 ± 5.2475.93 ± 6.89
1560.79 ± 6.4759.30 ± 2.8471.45 ± 4.3968.13 ± 5.0567.71 ± 5.2367.87 ± 4.2673.05 ± 4.64
1959.70 ± 2.3162.60 ± 5.6869.98 ± 4.1370.93 ± 4.4968.83 ± 3.8171.29 ± 2.9771.64 ± 3.66
2358.57 ± 3.3660.9 ± 3.8772.16 ± 2.8668.54 ± 4.3666.93 ± 3.2567.61 ± 2.8571.65 ± 4.47
2758.57 ± 2.8562.99 ± 2.0171.60 ± 3.4766.23 ± 2.9370.04 ± 1.9571.88 ± 1.9473.61 ± 3.56
3158.04 ± 3.5661.92 ± 2.2571.02 ± 2.4269.94 ± 2.4370.50 ± 2.0670.55 ± 1.9372.31 ± 2.90
3558.31 ± 2.3162.89 ± 2.9570.63 ± 2.3567.54 ± 3.0170.86 ± 2.1270.93 ± 2.4572.14 ± 2.05
3958.90 ± 1.9361.8 ± 1.6470.65 ± 2.0567.50 ± 2.1870.81 ± 1.8970.73 ± 1.9471.69 ± 2.06
4358.67 ± 1.5361.74 ± 2.2670.67 ± 1.5768.62 ± 1.3670.15 ± 1.7570.31 ± 1.4771.27 ± 1.53
4758.40 ± 1.7962.81 ± 1.5771.43 ± 2.1968.45 ± 1.1169.45 ± 1.1769.52 ± 1.3271.40 ± 1.52
4957.60 ± 1.3362.78 ± 1.2370.49 ± 1.0368.21 ± 1.5569.80 ± 1.2869.85 ± 1.3170.78 ± 1.53
5057.63 ± 1.3563.47 ± 1.3770.79 ± 1.9569.66 ± 1.1769.83 ± 1.0970.02 ± 0.9373.00 ± 1.15
Avg.58.7960.9271.4467.7868.8469.5272.74
Table 6. AC (%) of each algorithm on Coil20 dataset.
Table 6. AC (%) of each algorithm on Coil20 dataset.
ClassesNMFNTDGNTDGDNMFGDNMTFGSNMTFGDSNTD
297.92 ± 4.2176.93 ± 14.221 ± 01 ± 086.49 ± 17.6299.86 ± 0.281 ± 0
479.26 ± 16.7379.07 ± 15.1593.65 ± 11.0985.87 ± 11.4274.72 ± 15.6177.62 ± 16.9395.40 ± 8.99
667.24 ± 12.0065.14 ± 12.0386.75 ± 9.7678.11 ± 13.4381.63 ± 11.3782.63 ± 11.3483.73 ± 15.22
872.17 ± 7.7165.94 ± 6.1485.94 ± 7.7686.87 ± 9.7088.00 ± 8.8288.90 ± 9.1386.32 ± 6.78
1067.52 ± 5.0969.11 ± 4.7887.62 ± 8.3286.64 ± 7.7685.75 ± 8.3085.93 ± 7.9088.74 ± 9.01
1266.43 ± 7.9264.14 ± 5.0783.05 ± 7.7978.69 ± 8.2278.40 ± 4.7179.42 ± 6.7386.39 ± 7.22
1464.58 ± 5.1563.58 ± 5.9879.30 ± 5.5281.04 ± 8.1682.28 ± 3.4782.44 ± 4.9581.37 ± 5.39
1663.86 ± 6.1460.67 ± 4.4882.45 ± 4.3778.67 ± 4.8980.86 ± 6.4580.96 ± 7.2884.10 ± 3.12
1861.71 ± 3.9361.09 ± 4.2879.12 ± 5.1978.58 ± 4.3179.59 ± 4.1581.90 ± 4.8781.02 ± 4.78
1962.06 ± 3.7959.67 ± 3.7876.58 ± 4.0076.32 ± 4.6777.81 ± 3.4478.09 ± 3.6579.37 ± 2.90
2061.19 ± 4.0357.62 ± 4.1578.06 ± 3.6379.21 ± 3.6278.95 ± 2.5979.40 ± 3.0579.85 ± 2.79
Avg.69.4565.7284.7782.7384.1783.3886.03
Table 7. NMI (%) of each algorithm on Coil20 dataset.
Table 7. NMI (%) of each algorithm on Coil20 dataset.
ClassesNMFNTDGNTDGDNMFGDNMTFGSNMTFGDSNTD
291.99 ± 16.1940.04 ± 30.5293.35 ± 8.581 ± 066.48 ± 42.5398.94 ± 2.141 ± 0
475.51 ± 18.5575.94 ± 14.8493.35 ± 9.4781.75 ± 11.1976.83 ± 16.3974.81 ± 20.3194.47 ± 8.55
665.27 ± 1.3464.32 ± 11.7986.17 ± 9.0878.73 ± 12.4283.13 ± 8.8783.88 ± 9.8885.39 ± 11.34
876.67 ± 6.9070.71 ± 5.5888.91 ± 4.8189.55 ± 7.0891.67 ± 6.0492.18 ± 5.7089.58 ± 4.03
1074.23 ± 3.0574.83 ± 4.0291.71 ± 5.2291.48 ± 4.6090.51 ± 5.3689.91 ± 5.7592.22 ± 5.61
1273.35 ± 6.1171.96 ± 2.9888.55 ± 5.9986.72 ± 4.5785.57 ± 3.7086.17 ± 4.5491.18 ± 5.31
1474.63 ± 3.1172.13 ± 5.2486.35 ± 3.1888.58 ± 4.9490.34 ± 2.4189.84 ± 3.0687.51 ± 3.13
1673.87 ± 3.7071.73 ± 2.7289.20 ± 2.3187.10 ± 2.5988.78 ± 4.0688.38 ± 4.4890.99 ± 1.88
1872.77 ± 2.6172.49 ± 2.8486.78 ± 3.1186.97 ± 2.1389.78 ± 2.5189.99 ± 2.8789.10 ± 2.45
1973.43 ± 2.3272.17 ± 2.5885.55 ± 1.9486.44 ± 2.0387.62 ± 1.3987.6 ± 1.8988.02 ± 1.90
2072.99 ± 1.8071.32 ± 2.2285.97 ± 1.7088.07 ± 1.4688.15 ± 1.0688.05 ± 1.1488.33 ± 1.07
Avg.74.9768.8889.3287.7689.6088.1690.59
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, J.; Lu, L. Integrating Dual Graph Constraints into Sparse Non-Negative Tucker Decomposition for Enhanced Co-Clustering. Mathematics 2025, 13, 3494. https://doi.org/10.3390/math13213494

AMA Style

Han J, Lu L. Integrating Dual Graph Constraints into Sparse Non-Negative Tucker Decomposition for Enhanced Co-Clustering. Mathematics. 2025; 13(21):3494. https://doi.org/10.3390/math13213494

Chicago/Turabian Style

Han, Jing, and Linzhang Lu. 2025. "Integrating Dual Graph Constraints into Sparse Non-Negative Tucker Decomposition for Enhanced Co-Clustering" Mathematics 13, no. 21: 3494. https://doi.org/10.3390/math13213494

APA Style

Han, J., & Lu, L. (2025). Integrating Dual Graph Constraints into Sparse Non-Negative Tucker Decomposition for Enhanced Co-Clustering. Mathematics, 13(21), 3494. https://doi.org/10.3390/math13213494

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop