You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

24 November 2025

Tensorized Consensus Graph Learning for Incomplete Multi-View Clustering with Confidence Integration

,
,
and
School of Computer and Artificial Intelligence, Changzhou University, Changzhou 213164, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Computing and Artificial Intelligence

Abstract

Graph-based multi-view clustering has gained significant attention in recent years due to its superior ability to reveal clustering structures. However, existing methods often incur high computational costs when capturing local information and overlook the higher-order correlations between multiple views. To address these issues, we propose Tensorized Consensus Graph Learning for Incomplete Multi-View Clustering with Confidence Integration (TCGL). This approach constructs adjacency and local heat kernel graphs by filtering missing samples to better capture local structures while leveraging a t-SVD-based weighted tensor nuclear norm sparsification method to reduce noise. Additionally, we introduce a matrix energy-based adjacency graph normalization strategy that utilizes common nearest neighbors to generate probability matrices, enhancing noise resistance and improving structural exploration. Experimental results demonstrate that TCGL effectively handles incomplete data and significantly outperforms state-of-the-art approaches across multiple datasets.

1. Introduction

Multi-view clustering (MVC) is a technique that enhances clustering performance by integrating data from multiple sources or feature representations, allowing for a more comprehensive understanding of complex objects []. Real-world entities, such as images, users or biological samples, can usually be described from multiple complementary perspectives []. For instance, an image may be represented using a variety of descriptors including color histograms, textural patterns, and shape descriptors [], while a user’s profile might encompass multi-faceted information [], such as social connections [], behavior logs and text content []. The objective of MVC is to leverage the complementarity and consistency between these views and thereby more accurately capture the intrinsic data structure.
In practical applications, multi-view data frequently encounters a range of challenges [], such as missing instances caused by sensor failures or data corruption, noise interference, and the inherent difficulty of aligning heterogeneous data sources []. Consequently, Incomplete Multi-view Clustering (IMVC) algorithms have been devised to address these issues []. Current IMVC methods can be generally divided into two principal categories []. The first approach relies on graph-based techniques [], where a similarity graph is constructed for each view, and clustering is subsequently achieved by leveraging the spectral properties of the graph Laplacian matrix []. For example, Liu et al. [] presented an efficient incomplete multi-view clustering method (EE-IMVC) that iteratively refines the base clustering matrix and the consensus clustering matrix to enhance clustering outcomes. Additionally, Wen et al. [] proposed a consensus graph learning framework grounded in high-confidence local structures (HCLS-CGL), which enhances noise robustness by learning cross-view consensus graphs and incorporating confidence graph constraints. The second approach employs non-negative matrix factorization (NMF) method for multi-view clustering task [], which uncovers latent data structures by decomposing the data matrix into two or more non-negative factor matrices. Within this paradigm, Khan et al. [] put forward the Manifold Regularized Multi-view Fuzzy Clustering Method (MRFCM). This approach maintains the innate geometric configuration of the data through manifold regularization and dimension reduction, while simultaneously mitigating overfitting and enhancing sparsity through a combined use of the Frobenius norm and L1-norm. Furthermore, Luong et al. [] introduced a deep non-negative matrix factorization-based multi-view clustering approach (ODD-NMF), which preserves the structural relationships within the data through multi-layer manifold learning, integrating diversity and orthogonality constraints to optimize clustering performance.
However, existing methods still have several limitations []. Many methods are limited to dual-view data or necessitate the completeness of at least one view. Additionally, learning graph structures directly from raw data can easily introduce a large amount of noise []. For instance, Gaussian-kernel-based graphs lack universal applicability, and the self-expressive property of original data is especially sensitive to anomalous points []. Furthermore, most methods focus on common representations or pairwise view relationships, inadequately capturing the higher-order interdependencies within multi-view data and consequently resulting in the loss of vital underlying semantic information []. Compounding these issues, the graph learning process in most methods is typically disconnected from the clustering objective [], as prior knowledge about the desired cluster structure is rarely incorporated. This disjunction ultimately results in severely suboptimal performance.
To tackle these challenges, we propose a novel method termed Tensorized Consensus Graph Learning (TCGL) for incomplete multi-view clustering. As illustrated in Figure 1, TCGL constructs adjacency graphs and local heat kernel graphs by filtering missing samples and retaining valid data, which aids in representing the data’s local structural characteristics. To this end, the local heat kernel matrices are assembled into a tensor. We then employ a weighted tensor nuclear norm minimization, grounded in t-SVD, to denoise the constructed tensor and preserve critical similarity structures that are often obscured in graphs derived directly from raw data. Furthermore, the adjacency graph of each view is normalized by its matrix energy to enhance multi-view consistency and clustering performance. Subsequently, the sparsified local heat kernel graphs and the probability matrices are fused into a consensus graph, which more accurately captures the underlying data structures across all views. The major contributions of this paper are summarized as follows:
Figure 1. Illustration of the proposed TCGL model. Given an incomplete multi-view dataset X , TCGL first processes the incomplete data to construct the adjacency graph H ( v ) and the local heat kernel graph Z ( v ) . The heat kernel matrices are then assembled into a tensor and refined by a t-SVD-based sparsification module to enhance high-similarity information and reduce noise. Concurrently, the adjacency graph H ( v ) is normalized to generate a probability matrix. Finally, these components are fused to learn a consensus graph for multi-view clustering.
  • We propose the Tensorized Consensus Graph Learning (TCGL) with a confidence integration framework for incomplete multi-view clustering, which leverages missing sample filtration to construct adjacency and local heat kernel graphs, thereby enhancing the preservation of local structural information.
  • A probability matrix generation strategy based on common nearest neighbors is employed, coupled with a t-SVD-based weighted tensor nuclear norm sparsification technique. This integration effectively mitigates noise, retains high-value similarity information, and strengthens structural representation learning.
  • TCGL framework is optimized with an alternating minimization algorithm. Extensive evaluations on several benchmark datasets validate that our approach surpasses existing state-of-the-art IMVC techniques.
This paper is structured as follows. Section 2 covers the preliminaries. Section 3 elaborates on the proposed TCGL method, followed by the description of the optimization algorithm in Section 4. Section 5 evaluates the experimental results, and Section 6 provides the concluding remarks and future work.

2. Preliminary

This section presents the notation and key preliminaries that form the foundation for our proposed method, is cataloged in Table 1. The input is an incomplete multi-view dataset X = { X ( 1 ) , X ( 2 ) , , X ( v ) } , where X ( v ) R d v × n denotes the feature matrix of the v-th view ( v = 1 , 2 , , V ). Here, d v represents the feature dimensionality of the v-th view, n is the total number of samples across all views, and each column X ( v ) : , i R d v ( i = 1 , 2 , , n ) corresponds to the feature representation of the i-th sample in the v-th view.

2.1. Notation

2.1.1. Basic Symbol Conventions

  • Calligraphic letters (e.g., A , X ) represent tensors (typically 3D tensors for multi-view graph modeling in this work);
  • Bold uppercase letters (e.g., A , X ( v ) ) denote matrices;
  • Bold lowercase letters (e.g., a , x i ( v ) ) signify vectors;
  • Plain lowercase letters (e.g., n , d v , V ) represent scalars (e.g., sample count, feature dimensionality, view number).
For a matrix A R m × n :
  • A : , i denotes the i-th column vector of A ;
  • A i , : denotes the i-th row vector of A ;
  • A i , j denotes the element at the i-th row and j-th column of A ;
  • | A | denotes the element-wise absolute value operation (i.e., | A | i , j = | A i , j | );
  • A 0 indicates that all elements of A are non-negative;
  • rank ( A ) denotes the rank of A ;
  • 1 m × n and 1 n are the compact notation for an all-ones matrix and an n-dimensional all-ones vector, respectively (with dimensions omitted when evident).
Table 1. List of matrix, tensor and set symbols.
Table 1. List of matrix, tensor and set symbols.
SymbolDimensionDescription
X setIncomplete multi-view dataset { X ( 1 ) , , X ( V ) }
X ( v ) R d v × n Feature matrix of view v
O R n × V Missing-indicator matrix
M ( v ) R n × n Diagonal mask for available samples in view v
Z ˜ ( v ) R n v × n v Local similarity between available samples in view v
Z ( v ) R n × n Full similarity graph of view v (zero-padded)
H ˜ ( v ) R n v × n v Co-neighbor probability in view v
H ( v ) R n × n Full co-neighbor confidence graph of view v
S R n × n Consensus similarity graph for clustering
L S R n × n Normalized Laplacian of consensus graph S
Z R n × n × V Tensor stacking all Z ( v ) graphs
J R n × n × V Auxiliary tensor for low-rank regularization
Q R n × n × V Lagrange-multiplier tensor in ALM
α R V View-weight vector [ α 1 , , α V ] T
ω R min ( n , V ) Weight vector for tensor nuclear norm shrinkage
F R n × c Spectral embedding matrix

2.1.2. Graph Laplacian Matrix

For a similarity matrix B R n × n (constructed to describe sample similarity in a single view), its symmetric normalized graph Laplacian is defined as:
L B = D B | B | + | B | T 2 ,
where D B R n × n is the degree matrix corresponding to B , and its i-th diagonal element is calculated as:
( D B ) i , i = j = 1 n | B i , j | + | B j , i | 2 .

2.1.3. Tensor-Related Operations

For a 3D tensor A R m × n × k (used to stack k view-specific similarity matrices in this work):
  • A ( t ) R m × n denotes the t-th frontal slice of A (i.e., the matrix obtained by fixing the third dimension at index t, t = 1 , 2 , , k );
  • Fast Fourier Transform (FFT) along the third dimension: A ¯ = fft ( A , [ ] , 3 ) , where A ¯ ( t ) is the FFT result of the t-th frontal slice;
  • Inverse FFT (IFFT) recovery: A = ifft ( A ¯ , [ ] , 3 ) , which reconstructs the original tensor from its frequency-domain representation A ¯ ;

2.1.4. Block Vectorization

Converts a 3D tensor into a 2D matrix by concatenating its frontal slices vertically, defined as:
bvec ( A ) = A ( 1 ) A ( 2 ) A ( k ) R m k × n .
The inverse operation (block folding) recovers the tensor from its block-vectorized form: bvfold ( bvec ( A ) ) = A .

2.1.5. Block Circulant Matrix

Constructed from the frontal slices of A to enable tensor singular value decomposition (t-SVD) via matrix operations, defined as:
bcirc ( A ) = A ( 1 ) A ( k ) A ( k 1 ) A ( 2 ) A ( 2 ) A ( 1 ) A ( k ) A ( 3 ) A ( k ) A ( k 1 ) A ( k 2 ) A ( 1 ) R m k × n k .
This structure ensures that the t-SVD of A is equivalent to the conventional SVD of bcirc ( A ) in the frequency domain, which simplifies the optimization of tensor nuclear norm regularization.
Definition 1
(Transpose of a Tensor []). For a tensor A R n 1 × n 2 × n 3 , its transpose A T R n 2 × n 1 × n 3 is acquired by initially transposing each frontal slice ( A ( k ) ) and subsequently reversing the sequence of the transposed slices from 2 to n 3 .
Definition 2
(Identity Tensor). The identity tensor I is an n 1 × n 1 × n 3 tensor, characterized by an identity matrix of size n 1 × n 1 as its first frontal slice, with all other frontal slices being null.
Definition 3
(Orthogonal Tensor). A tensor A is deemed orthogonal if it satisfies the conditions A T · A = A · A T = I .
Definition 4
(t-Product). For tensors A R n 1 × n 2 × n 3 and B R n 2 × n 4 × n 3 , their t-product A · B results in a tensor E of dimensions n 1 × n 4 × n 3 , defined as follows:
E = A B = bvfold ( bcirc ( A ) bvec ( B ) ) .
The t-product between A and B can be swiftly computed by multiplying corresponding frontal slices in the Fourier domain:
E ¯ ( k ) = A ¯ ( k ) B ¯ ( k ) , k = 1 , , n 3 ,
and then obtaining the t-product E = ifft ( E ; [ 2 , 3 ] ) .
Definition 5
(t-SVD []). A tensor A R n 1 × n 2 × n 3 can be factorized via its t-SVD as:
A = U S V T ,
where U R n 1 × n 1 × n 3 and V R n 2 × n 2 × n 3 are orthogonal tensors, and S R n 1 × n 2 × n 3 is an f-diagonal tensor with diagonal matrices as its frontal slices.
Following [], an efficient computation of the t-SVD leverages the matrix SVD within the Fourier domain.
A ¯ ( k ) = U ¯ ( k ) S ¯ ( k ) V ¯ ( k ) T , k = 1 , , n 3 .
Definition 6
(Tensor Multi-Rank). For a tensor A R n 1 × n 2 × n 3 , its multi-rank is defined as a vector r R n 3 × 1 , where the i-th element corresponds to the rank of the i-th frontal slice of A .
Definition 7
(t-SVD Based Tensor Nuclear Norm []). Given a tensor A R n 1 × n 2 × n 3 , its t-SVD based tensor nuclear norm (t-TNN) is expressed as:
A = k = 1 n 3 A ( k ) = k = 1 n 3 i = 1 min ( n 1 , n 2 ) σ i ( A ( k ) ) ,
where σ i ( A ( k ) ) denotes the i-th largest singular value of A ( k ) .
Additionally, the weighted t-SVD based tensor nuclear norm is defined as:
A ω , = k = 1 n 3 i = 1 min ( n 1 , n 2 ) w i σ i ( A ( k ) ) ,
where w i is the weight associated with σ i ( A ( k ) ) , allowing the incorporation of prior knowledge regarding the singular values of the matrices.
Remark 1.
In this work, the weight vector ω = [ w 1 , . . . , w i ] is initialized uniformly as w i = 1 / min ( n 1 , n 2 ) for i = 1 , , min ( n 1 , n 2 ) , ensuring equal contribution from all singular value components at the start of optimization. Furthermore, ω remains fixed throughout the learning process. This design provides a stable regularization framework. The model’s adaptability to the importance of different views is primarily handled through the optimization of the view-weighting vector α and the low-rank constraints on the tensor slices.

3. The Proposed Method

We present a graph-based consensus learning framework for incomplete multi-view clustering. TCGL comprises three key stages: addressing missing data, constructing KNN binary and confidence graphs, and consolidating the confidence graphs into a tensor for low-rank processing. The weighted confidence graph facilitates the generation of a consensus graph, which is subsequently clustered using K-means.
The existing approaches for dealing with incomplete data, such as imputation, distance-based adjustment [], weighted processing, latent model inference, and dimensionality reduction, are all subject to limitations. These limitations include introducing biases, escalating computational costs, losing information, and depending on subjective rules. To overcome these challenges, TCGL strategically excludes the dimensions containing missing values for each sample through the constructed index matrix. By constructing a location index matrix O { 0 , 1 } n × l , where O i , j = 1 indicates the j-th view of the i-th sample is not missing, and O i , j = 0 indicates it is missing. M i , j ( v ) is constructed by the prior location indices O for filtering complete data and aligning data, as follows:
M i , j ( v ) = 1 , i f t h e j - t h s a m l p e h a s a n a v a i l a b l e i n s t a n c e x i ( v ) i n t h e v - t h v i e w 0 , o t h e r w i s e .

Consensus Graph Learning Method

The standard procedure for graph-based spectral clustering [,] includes three key steps: affinity graph construction, eigen-decomposition of the graph Laplacian matrix, and final cluster assignment via k-means. The affinity graph is an n × n matrix, where n represents the number of samples in the data, and the elements of the matrix indicate the relationships between the samples, such as their similarity or distance. Among these steps, the construction of the affinity graph is particularly important as it significantly impacts the final clustering results.
In recent years, several methods for constructing high-quality affinity graphs have been proposed [,,] aiming to better capture the intrinsic relationships between the data points. One such method is the constrained Laplacian rank (CLR) approach []. CLR aims to learn a new similarity matrix that reflects the cluster structure by ensuring that the matrix has exactly c blocks, each corresponding to one of the c clusters. The formulation of this problem is as follows:
min S S G F 2 , s . t . 0 S 1 , S 1 = 1 , rank ( L S ) = n c , diag ( S ) = 0 ,
where S R n × n denotes the consensus high-quality graph to be learned. The constraints ensure that the elements of S lie within the range [ 0 , 1 ] ; each row sums to 1; and the rank of its Laplacian matrix L S equals n c , where n is the number of samples and c is the number of clusters. In addition, G R n × n represents the pre-constructed affinity graph of the v-th view, such as the k-nearest neighbor (k-NN) graph.
The rationale for imposing the rank constraint rank ( L S ) = n c lies in the theorem proposed by [], which states that the number of connected components in the graph S is equal to the number of zero eigenvalues of its Laplacian matrix. This constraint ensures exactly c connected components in the affinity graph, which is consistent with the intrinsic cluster structure of the data.
In graph-learning-based clustering methods, it has been shown that learning a high-quality graph capable of capturing the intrinsic relationships between data points is crucial for improving clustering performance. Therefore, under the assumption that different views should lead to a consistent clustering outcome, many approaches aim to learn a unified consensus graph S R n × n by effectively integrating information from all available views for multi-view clustering []. A commonly employed formulation is as follows:
min ϕ ( S , α ) v = 1 V α v S G ( v ) F 2 + λ 1 S F 2 , s . t . 0 S 1 , S 1 = 1 , rank ( L S ) = n c , 0 α v 1 , v = 1 V α v = 1 , diag ( S ) = 0 ,
where α v is defined as follows:
α v = 1 2 S G ( v ) F 2 .
In multi-view clustering, each view might contain varying amounts of information. To integrate this information effectively while acknowledging the differing contributions of each view to the final clustering result, assigning weights to each view is meaningful. This is why α v is introduced in the formula, where r typically represents the power of the weight, further adjusting the influence of each view. By tuning the values of α v , we can highlight views that significantly affect the clustering results while downplaying those with lesser impact, thereby enhancing the overall clustering performance.
Furthermore, the rank constraint rank ( L S ) = n c in Equation (10) can be transformed into the minimum optimization problem of min F T F = I Tr F T L S F . This transformation allows us to obtain the following equivalent optimization problem:
min ϕ ( S , α ) v = 1 V α v S G ( v ) F 2 + λ 1 S F 2 + λ 2 Tr F T L S F , s . t . 0 S 1 , S 1 = 1 , 0 α v 1 , v = 1 V α v = 1 , diag ( S ) = 0 , F T F = I ,
where ϕ ( S , α ) denotes the boundary constraints of variables S and α , and l is the number of views of the data. After obtaining the consensus graph S by optimizing the above objective, spectral clustering or connected component search methods can be applied to S to achieve the clustering result of the given data. The parameter r is tunable to control the distribution of the coefficient vector α . Additionally, λ 1 and λ 2 are penalty parameters introduced to regulate the optimization process.
Challenge and Motivation 1: Although the consensus graph learning model provides an effective multi-view clustering framework, it presents significant challenges under the Generalized Incomplete Multi-View Clustering (GIMC) setting. In GIMC, the core challenge stems from the heterogeneous and unaligned graphs built from available instances. While some techniques aim to recover these missing parts [], they are often computationally expensive. Moreover, perfect recovery is impractical, and imperfect reconstructions can lead to severe performance degradation.
To overcome these limitations, this paper employs a method to learn a consensus graph from the available similarity information of the non-missing instances.rather than relying on potentially imperfect recovered graphs. Specifically, we utilize the previously defined M matrix to indicate the availability of each sample in each view, where M i , j = 1 indicates that the j-th view of the i-th sample is not missing, and M i , j = 0 indicates that it is missing. Based on this, we formulate the following optimization problem to learn the consensus graph S R n × n using certain graph information from the available views:
min S , α v = 1 V α v S M ( v ) Z ˜ ( v ) M ( v ) T F 2 + λ 1 S F 2 + λ 2 Tr F T L S F , s . t . 0 α v 1 , v = 1 V α v = 1 , S T 1 = 1 , 0 S 1 , diag ( S ) = 0 , F T F = I ,
where Z ˜ ( v ) R n v × n v , Z ˜ i , j ( v ) represents the similarity between the i-th and j-th available instances in the v-th view. In our approach, Z ˜ i , j ( v ) is constructed by the recognizable k-nearest neighbor algorithm as follows:
Z ˜ i , j ( v ) = e x i ( v ) x j ( v ) 2 2 2 , if x i ( v ) ψ ( x j ( v ) ) or x j ( v ) ψ ( x i ( v ) ) 0 , otherwise .
Challenge and Motivation 2: For incomplete multi-view data, due to the uncontrollable nature of missing views and noise, the pre-constructed graph Z ˜ ( v ) may fail to accurately capture the true nearest neighbor relationships between samples. This limitation motivates us to integrate adaptive nearest neighbor graph learning into the consensus graph framework. However, existing solutions often suffer from high computational complexity and optimization difficulties.
Inspired by [], we propose an innovative strategy based on spatial proximity assumptions. The core hypothesis of our method is that intrinsically adjacent samples should share overlapping neighbor sets, and the degree of overlap is positively correlated with class consistency. Specifically, pairwise co-neighbor counts are computed by matrix multiplication H i , j = A i , : · A : , j , generating a binary adjacency matrix H (with the diagonal initialized to zero) from observable data instances. Subsequently, we apply max-normalization H ˜ = H / max ( H ) to transform the co-neighbor counts into probabilistic affinities, constructing a co-neighbor likelihood matrix H ˜ . This matrix encodes the probability of sample pairs sharing mutual neighbors and serves as structural prior knowledge. We then incorporate this topological information into the following optimization model:
min S , α v = 1 V α v ( S M ( v ) Z ˜ ( v ) M ( v ) T ) M ( v ) H ˜ ( v ) M ( v ) T F 2 + λ 1 S F 2 + λ 2 Tr F T L s F , s . t . 0 α v 1 , v = 1 V α v = 1 , S T 1 = 1 , 0 S 1 , diag ( S ) = 0 , F T F = I ,
where ⊙ denotes element-wise multiplication. This formulation effectively incorporates confidence-based structural information, providing a more robust framework for consensus graph learning in incomplete multi-view clustering scenarios. To derive an equivalent formulation of Equation (18), we first define the transformed variables Z ( v ) and H ( v ) . Specifically, let Z ( v ) = M ( v ) T Z ˜ ( v ) M ( v ) and H ( v ) = M ( v ) T H ˜ ( v ) M ( v ) . Substituting these into the original problem yields the following equivalent optimization problem:
min S , α v = 1 V α v ( S Z ( v ) ) H ( v ) F 2 + λ 1 S F 2 + λ 2 Tr F T L s F , s . t . 0 α v 1 , v = 1 V α v = 1 , S T 1 = 1 , 0 S 1 , diag ( S ) = 0 , F T F = I .
Challenge and Motivation 3: In multi-view clustering tasks, exploring the local connectivity of data has been proven to be an effective strategy. Neighbors of x i ( v ) R d v × 1 are typically described as the k-nearest data points in the dataset to x i ( v ) .
We previously introduced the specific solution formula for Z. Based on this, we can redefine the basic formula for computing the nearest neighbor confidence matrix Z ( v ) as follows:
min Z ( v ) v = 1 V i , j = 1 n κ Z ˜ ( v ) F 2 , s . t . 0 z i j ( v ) 1 , ( z i ( v ) ) T 1 = 1 .
The parameter κ serves as a regularization parameter to prevent the trivial solution Z ( v ) = I . Despite their impressive performance, most existing methods focus on learning common representations or pairwise correlations between views [,]. However, this approach often fails to capture the full range of deeper, higher-order correlations in multi-view data, leading to the loss of crucial semantic information. Furthermore, these methods typically require a separate post-processing step to obtain the final clustering results and cannot account for the interrelationships between multiple affinity matrices from different views in a unified framework. As a result, this often leads to suboptimal clustering performance. The constraint ( z i ( v ) ) T 1 = 1 is imposed to normalize sample-wise affinities, eliminate scale ambiguity, and maintain comparability across views. This normalization is a standard technique in graph learning and does not restrict the relative similarity structure.
To address the aforementioned issues, we leverage a low-rank tensor-based method for multi-view proximity learning. This method simultaneously optimizes the affinity matrices for each view to effectively capture higher-order correlations in multi-view data, thereby enhancing the final clustering performance. As illustrated in Figure 1, data samples X ( 1 ) , , X ( V ) from multiple feature subsets or data sources are first provided as input, and corresponding affinity matrices Z ( 1 ) , , Z ( V ) are generated through a multi-view proximity learning strategy. To further explore the higher-order correlations between different data points across views, we employ a tensor construction mechanism, where multiple affinity matrices are jointly constructed into a tensor Z R n × n × V . The mathematical formulation of the method is given as follows:
min Z ( v ) , Z v = 1 V κ Z ( v ) F 2 + β Ψ ( Z ) , s . t . 0 z i j ( v ) 1 , ( z i ( v ) ) T 1 = 1 , Z = Φ ( Z ( 1 ) , , Z ( V ) ) ,
where β is a regularization parameter, Ψ represents the specific constraint applied to the constructed tensor Z , and Φ refers to the process of constructing the tensor Z by merging multiple affinity matrices Z ( v ) into a third-order tensor. Inspired by the work in [], we employ a weighted t-SVD-based tensor nuclear norm to capture the higher-order correlations embedded within multi-view affinity matrices. The mathematical formulation for this approach is given as follows:
min Z ( v ) , Z v = 1 V κ Z ( v ) F 2 + β Z ω , , s . t . 0 z ij ( v ) 1 , ( z i ( v ) ) T 1 = 1 , Z = Φ ( Z ( 1 ) , , Z ( V ) ) .
To achieve the separability of Z , the variable-splitting technique is implemented, and an auxiliary tensor variable J R n × n × V is introduced as a substitute for Z . Consequently, the model presented in Equation (19) can be restructured into the subsequent optimization problem.
min Z ( v ) , J v = 1 V κ Z ( v ) F 2 + β J ω , , s . t . 0 z i j ( v ) 1 , ( z i ( v ) ) T 1 = 1 , J = Z .
This work integrates two complementary methodologies. The objective of consensus graph learning is to derive a unified graph that reflects cross-view relationships and the inherent clustering structure. Conversely, low-rank tensorization employs tensors to model higher-order correlations and complex interdependencies between views. Although demonstrably effective individually, these two strategies operate in isolation, limiting the exploitation of their complementary strengths. To overcome this limitation, we propose a unified framework integrating Consensus Graph Learning with Low-rank Tensorization. By jointly optimizing graph construction and tensor decomposition, our approach simultaneously captures local connectivity patterns and global higher-order correlations, providing a more robust and comprehensive solution for multi-view clustering. The details of the framework are as follows:
min α , S , Z ( v ) , F , J v = 1 V κ Z ( v ) F 2 + β J ω , Tensor Low - Rank Factorization + v = 1 V α v ( S Z ( v ) ) H ( v ) F 2 + λ 1 S F 2 + λ 2 Tr ( F T L s F ) Consensus Graph Learning , s . t . 0 α v 1 , v = 1 V α v = 1 , S T 1 = 1 , 0 S 1 , 0 z i j ( v ) 1 , ( z i ( v ) ) T 1 = 1 , J = Z , F T F = I .
Remark 2.
The proposed objective function incorporates a tensor rank regularization term to capture multi-dimensional interactions across heterogeneous data views. Specifically, the transformed representation Z ( v ) R n × n undergoes a spatial reorganization through tensor rotation. As demonstrated in Figure 1, the rotational operation yields Z ( r ) whose i-th frontal slice Z i ( r ) R n × V encapsulates cross-view associations between n instances. An optimized graph structure should maintain inter-sample relational consistency throughout these heterogeneous feature spaces.
To address the inherent structural divergence between different view-specific clusters, we impose a multi-rank minimization constraint on the third-order tensor Z . This formulation induces spatial low-rank characteristics in each rotated slice Z i ( r ) , enabling effective extraction of Supplementary Information from disparate viewpoints.
The adoption of third-order tensor representation fundamentally extends conventional matrix-based approaches by modeling triadic relationships rather than pairwise interactions. While matrix factorization techniques can only recover dyadic correlations, the proposed tensor rank minimization framework inherently preserves multi-way dependency patterns. This characteristic makes the tensor regularization term particularly suitable for revealing latent synergistic interactions and multi-dimensional correlations in multi-view learning scenarios.
Remark 3.
The rotation of the tensor along the third mode is introduced to make the model compatible with the t-SVD framework proposed by Kilmer and Martin []. By applying this rotation, each frontal slice of the transformed tensor corresponds to a mode-3 fiber in the original representation, allowing the tensor to be expressed as a sequence of circulant-structured matrices. This is essential because the t-SVD and the associated tensor nuclear norm are defined under this rotated representation, where convolution-like interactions across views become linear operations in the Fourier domain.
Intuitively, the rotation does not change the underlying information but reorganizes the tensor so that correlations across views are captured in a way that aligns with the algebra of the t-product. This enables the model to better exploit multi-view dependencies using a mathematically well-established framework, rather than relying on an arbitrary permutation of tensor dimensions.
Remark 4.
The proposed tensor formulation offers a different perspective compared with conventional weighted matrix approaches. While weighted combinations of matrices ( v = 1 V w v Z ( v ) ) are effective for modeling pairwise relationships, they mainly characterize dyadic correlations between views. In contrast, representing multi-view affinities as a third-order tensor Z R n × n × V provides a unified structure in which all views are jointly considered. The low-rank tensor constraint fosters a shared low-dimensional representation, effectively integrating multi-view information at once. This approach can capture specific higher-order interactions—for instance, cluster patterns shaped by three or more views—which pairwise combinations may not adequately represent.

4. Optimization

The Augmented Lagrangian Multiplier (ALM) in conjunction with the Alternative Direction Minimizing (ADM) optimization algorithm is utilized for addressing the optimization problem at hand, which involves updating one specific variable while keeping the others constant. Furthermore, the corresponding augmented Lagrangian function of the above model is
min α , S , Z ( v ) , F , J v = 1 V κ Z ( v ) F 2 + β J ω , + ρ 2 J ( Z + Q ρ ) F 2 , + v = 1 V α v ( S Z ( v ) ) H ( v ) F 2 + λ 1 S F 2 + λ 2 Tr F T L s F . s . t . 0 α v 1 , v = 1 V α v = 1 , S T 1 = 1 , 0 S 1 , 0 z i j ( v ) 1 , ( z i ( v ) ) T 1 = 1 , J = Z , F T F = I ,
where Q R n × n × V is the Lagrange multiplier, and ρ > 0 is the penalty factor. By means of the alternative minimization strategy, the problem in Equation (25) can be divided into the following five sub-problems, each of which alternately optimizes each variable while fixing the others.
S subproblem : By fixing the other variables except S , we can update S by the following problem:
min S v = 1 V α v S Z ( v ) H ( v ) F 2 + λ 1 S F 2 + λ 2 Tr F T L S F , s . t . S T 1 = 1 , 0 S 1 , diag ( S ) = 0 .
In Equation (26), we can simply infer that T r F T L S F = 1 2 i , j = 1 n F i , : F j , : 2 2 S i , j when S is a positive matrix. Defining E i , j = F i , : F j , : 2 2 , then we can transform Equation (26) as follows:
min S i , j = 1 n S i , j v = 1 V α v H i , j ( v ) 2 Z i , j λ 2 4 E i , j v = 1 V α v H i , j ( v ) 2 + λ 1 2 .
It is obvious that Equation (27) is equivalent to the following n independent sub-optimization problems with respect to each column of matrix S:
min S : , j i = 1 n S i , j v = 1 V α v H i , j ( v ) 2 Z i , j λ 2 4 E i , j v = 1 V α v H i , j ( v ) 2 + λ 1 2 , s . t . i = 1 n S i , j = 1 , 0 S i , j 1 , S j , j = 0 .
According to the Lagrangian algorithm, we can obtain the following closed form solution for Equation (28):
S i , j = ( T i , j + η j ) + , i j 0 , i = j ,
where T i , j = v = 1 V α v H i , j ( v ) 2 Z i , j λ 2 4 E i , j v = 1 V α v H i , j ( v ) 2 + λ 1 , function ( a ) + sets negative a to zero and preserves the non-negative a. According to the boundary constraint i = 1 n S i , j = 1 , 0 S i , j 1 , S j , j = 0 , we can obtain η j = 1 i = 1 , i j n T i , j n 1 .
Z ( v ) subproblem : By fixing the other variables except Z, we can update Z by the following problem:
min Z ( v ) v = 1 l α v S Z ( v ) H ( v ) F 2 + κ Z ( v ) F 2 + ρ 2 J ( Z + Q ρ ) F 2 ,
where Q R n × n × V is the Lagrange multiplier, and ρ > 0 is the penalty factor.
min Z ( v ) v = 1 l α v S Z ( v ) H ( v ) F 2 + κ Z ( v ) F 2 + ρ 2 J ( v ) Z ( v ) + Q ( v ) ρ F 2 .
Denote B ( v ) = J ( v ) Q ( v ) ρ , and for each view v, the problem in Equation (31) can be rewritten as:
min 0 z i j ( v ) 1 , z i ( v ) T 1 = 1 v = 1 l i , j = 1 n α v ( s i , j z i , j ( v ) ) h i , j ( v ) 2 + κ z i , j ( v ) 2 + ρ 2 b i , j ( v ) z i , j ( v ) 2 .
Accordingly, the problem in Equation (32) can be reformulated separately for each i:
min 0 z i ( v ) 1 , z i ( v ) T 1 = 1 v = 1 l α v s i 2 2 s i z i ( v ) + z i ( v ) 2 h i ( v ) 2 + κ z i ( v ) 2 + ρ 2 b i ( v ) 2 2 b i ( v ) z i ( v ) + z i ( v ) 2 .
After certain combination, the above problem is equivalent to solving the following model:
min 0 z i ( v ) 1 , z i ( v ) T 1 = 1 z i ( v ) 2 α v s i h i + ρ b i ( v ) 2 κ + 2 α v h i + ρ 2 2 .
Specifically, the problem in Equation (34) is an Euclidean projection problem on the simplex space [], whose Lagrangian function can be written as
L z i ( v ) , η , φ = 1 2 z i ( v ) 2 α v s i h i + ρ b i ( v ) 2 κ + 2 α v h i + ρ 2 2 η z i ( v ) T 1 1 φ T z i ( v ) ,
where η is a scalar and φ is a Lagrangian coefficient vector. According to the Karush-Kuhn-Tucker condition, the optimal solution z i ( v ) can be verified to be
z i ( v ) = 2 α v s i h i + ρ b i ( v ) 2 κ + 2 α v h i + ρ + η 1 + .
We employ the efficient sorting-based algorithm by Duchi et al. [] for the simplex projection.
J subproblem : By fixing the other variables, J can be updated by solving the following problem:
min J β ρ J ω , + 1 2 J A + Q ρ F 2 .
The solution to the aforementioned minimization problem is established by the following two theorems.
Theorem 1.
Given Y = U Y D Y V Y T being SVD of Y R m × n , τ > 0 , l = min ( m , n ) , we can define the standard weighted nuclear norm minimization problem as:
arg min X 1 2 X Y F 2 + τ X ω , .
Consequently, we arrive at the optimal solution for the model defined in Equation (39):
X = Γ τ ω [ Y ] = U Y P τ ω ( Y ) V Y T ,
where we have
P τ ω ( Y ) = diag ( ξ 1 , ξ 2 , , ξ l ) with ξ i = sign ( σ i ( Y ) ) max ( σ i ( Y ) τ ω i , 0 ) .
Theorem  2.
Given A R n 1 × n 2 × n 3 , l = min ( n 1 , n 2 ) , we have A = U S V T . Then the model
arg min X 1 2 X A F 2 + τ X ω , .
It can be defined as:
X = Γ τ ω ( A ) = U ifft ( P τ ω ( A ¯ ) ) V T ,
where A ¯ = fft ( A , [ ] , 3 ) , and P τ ω ( A ¯ ) is a tensor, in which P τ ω ( A ¯ ( i ) ) denotes the i-th frontal slice of P τ ω ( A ¯ ( i ) ) .
Proof. 
Specifically speaking, the optimization problem in Equation (39) can be reformulated in the Fourier domain as follows:
X ¯ = arg min X ¯ 1 2 X ¯ A ¯ F 2 + i = 1 n 3 j = 1 l τ ω j σ j ( X ¯ ( i ) ) ,
where X ¯ ( i ) denotes the i-th frontal slice of X ¯ , σ j ( X ¯ ( i ) ) is the j-th largest singular value of X ¯ ( i ) , and ω j stands for the weighting shrinkage coefficient of σ j ( X ¯ ( i ) ) .    □
The problem in Equation (43) can be transformed by employing the Frobenius norm scheme:
arg min X ¯ i = 1 n 3 1 2 X ¯ ( i ) A ¯ ( i ) F 2 + j = 1 l τ ω j σ j ( X ¯ ( i ) ) .
Notice that each variable X ¯ ( i ) in Equation (41) is independent, which means that we can divide this problem into n 3 independent subproblems. Thus, for the i-th problem, i = 1 , 2 , , n 3 , we have
X ¯ ( i ) = arg min X ¯ ( i ) 1 2 X ¯ ( i ) A ¯ ( i ) F 2 + j = 1 l τ ω j σ j ( X ¯ ( i ) ) .
According to Theorem 1, we can obtain the solution of the above equation, i.e.,
X ¯ ( i ) = Γ τ ω [ A ¯ ( i ) ] = U ¯ ( i ) P τ ω ( A ¯ ( i ) ) V ¯ ( i ) T ,
where X ¯ ( i ) denotes the i-th frontal slice of X ¯ , similarly for U ¯ ( i ) and V ¯ ( i ) . Hence, the solution in Equation (46) can be further written as:
X = Γ τ ω [ A ] = U ifft ( P τ ω ( A ¯ ) ) V T ,
where U = ifft ( U ¯ , [ ] , 3 ) and V = ifft ( V ¯ , [ ] , 3 ) .
The solution to the model in Equation (47) follows directly from Theorem 2:
J = Γ β ρ ω Z + Q ρ .
F subproblem : By fixing the other variables except F , the optimization problem becomes
min F T F = I T r ( F T L s F ) .
We can solve the problem in Equation (49) by calculating the eigenvectors corresponding to the smallest c eigenvalues of the Laplacian matrix L s .
α v subproblem : By fixing the other variables except α v , the optimization problem is equivalent to dealing with the model in Equation (50).
α v = 1 2 S Z ( v ) F 2 .
Q subproblem : The Lagrange multiplier Q can be updated as:
Q = Q + ρ ( Z J ) .
ρ subproblem : The penalty parameter ρ is updated as:
ρ ( t + 1 ) = μ · ρ ( t ) ,
where μ > 1 is employed to accelerate the convergence speed by gradually strengthening the penalty on constraint violations.

5. Experiments

This section evaluates the effectiveness of the proposed TCGL framework through extensive experiments on four real-world datasets. We benchmark its clustering performance against several state-of-the-art methods. Additionally, a comprehensive analysis is provided, including ablation studies, parameter sensitivity analysis, and an examination of empirical convergence. All experiments are implemented in MATLAB R2023b and run on two computing environments: a local workstation with a 12th Gen Intel Core i7-12700H CPU (2.30 GHz) and 16 GB RAM, and a rented server equipped with an Intel Xeon-class processor and 128 GB RAM. Large-scale datasets (Caltech7 and Animal) are evaluated on the server, while smaller datasets (ORL and UCI) are processed on the local workstation.
In our experiments, μ is selected via a simple grid search to achieve the most stable convergence behavior, and the initial value of ρ is set within the range ( 1 , 2 ] following common ADMM practices. This initialization–update scheme adopts a standard geometric growth strategy widely used in ADMM-based optimization to balance convergence stability and computational efficiency. Stopping tolerance 10 4 and k max = 30 are used throughout the experiments. The weight vector ω is initialized uniformly and remains fixed during optimization.

5.1. Experimental Settings

5.1.1. Datasets

(1)
ORL: The ORL dataset consists of 400 face images from 40 distinct individuals, with each individual having 10 images. The label of each individual serves as the ground truth class label. These images were captured under various conditions, including different lighting, times of day, facial expressions (such as smiling or not), and facial features (e.g., with or without glasses). In our experiments, three different feature views are employed: intensity-based features (view 1), Local Binary Pattern (LBP) features (view 2), and Gabor features (view 3).
(2)
UCI: The UCI dataset, sourced from the UCI machine learning repository, consists of 2000 handwritten digit images representing 10 different digits, with each digit having 200 samples. Each digit is treated as a distinct class label. For our experiments, three types of features are utilized: intensity-averaged features (view 1), Fourier coefficient-based features (view 2), and morphological features (view 3).
(3)
Caltech7: Caltech7 is a subset of the well-known Caltech101 dataset, which contains images from 101 categories, including objects such as animals, vehicles, and everyday items. Caltech101 includes 9144 images, making it a widely used benchmark in object recognition research. Caltech7 includes 1474 images from seven selected categories. For our experiments, six types of features are extracted from each image and utilized for classification.
(4)
Animal: The Animal dataset is a comprehensive collection of images featuring a diverse range of animal species. It contains 30,475 images across 50 different animal categories. The dataset captures various animals in different poses and environments, making it suitable for tasks such as image classification and object recognition. We utilize a subset of this dataset, which includes 11,673 images from the first 20 animal categories, with four different views for each animal. The dataset description is shown in Table 2.
Table 2. Performance Comparison on Animal Dataset.

5.1.2. Compared Methods

In the experiments, our TCGL is compared with nine state-of-the-art methods: BSV [], MIC [], DAIMC [], OMVC [], OPIMC [], MKKM-IK-MKC [], PIC [], and UEAF []. We briefly describe the methods not mentioned in previous sections as follows:
(1)
AGC-IMC []: Adaptive Graph Completion and Consensus Representation Learning, which enhances clustering performance through intra-view consistency, inter-view reasoning, and an adaptive weighting mechanism, addressing view imbalance.
(2)
TMBSD []: Tensor Nuclear Norm Regularization and High-order Consistency, maintaining a block-diagonal structure across views, optimizing for efficient clustering.
(3)
DIMVC []: A method without padding or fusion, utilizing deep feature embedding and clustering models to mine complementary information across views, enhancing clustering consistency.
(4)
IMVC-CBG []: A flexible bipartite graph framework that integrates multi-view anchor point learning and incomplete bipartite graph modeling, suitable for large-scale IMVC tasks.
(5)
LSIMVC []: Combines sparse constraints and local graph embedding, alleviating view imbalance and improving consensus representation and clustering performance.
(6)
TCIMC []: Tensor Schatten p-norm completion, combining connectivity constraints and low-rank structures to enhance the utilization of complementary information across views.
(7)
ETLSRR []: Enhanced Tensor Low-rank and Sparse Representation Recovery, using non-convex regularization to alleviate approximation bias and improve the discriminability of similarity matrices.
(8)
HCLS_CGL []: Consensus Graph Learning, simplifying and efficiently implementing incomplete multi-view clustering.
(9)
JTIV-LRR []: Low-rank constraints and inter-view correlation modeling, addressing the issue of neglecting cross-view relationships in traditional methods.
(10)
IMVTSC-MVI []: Infers missing views in the feature space and jointly learns similarity graphs in the manifold space, while incorporating a low-rank tensor constraint to capture high-order cross-view correlations and enhance clustering performance on incomplete multi-view data.
All competing methods are implemented using their recommended parameters or by conducting a parameter search to optimize performance.

5.1.3. Evaluation Metrics

Following previous works [,,], clustering accuracy (ACC), normalized mutual information (NMI), and purity are adopted as metrics to evaluate the performance of the aforementioned methods.

5.1.4. Parameter Settings

In our experiments, all parameters are fixed across different datasets to ensure fairness and reproducibility. The numbers of nearest neighbors are set to k 1 = k 2 = 13 , following common practice in graph-based clustering, where a moderately sized neighborhood helps preserve the local manifold structure while avoiding over-connection. The trade-off parameters are set to λ 1 = 0.01 , λ 2 = 0.01 , and λ 3 = 0.001 . These values balance the contributions of structure preservation, tensor regularization, and confidence integration, and we find that the model performance is stable within a reasonable range around these defaults. Therefore, extensive parameter tuning is unnecessary.
For incomplete multi-view settings, the missing rate is controlled by randomly removing a proportion of samples in each view according to the target missing ratio. This strategy is widely adopted in prior works and ensures that different methods are evaluated under identical missing patterns.

5.2. Experimental Results and Analysis

Table 3Table 4, and Table 5 present the clustering results under missing rates of 10%, 30% and 50% respectively, while Table 6 shows the results under more challenging missing rates of 30%, 50% and 70% on the larger Animal dataset.
Table 3. Performance Comparison on Caltech7 Dataset.
Table 4. Performance Comparison on UCI Dataset.
Table 5. Performance Comparison on ORL Dataset.
Table 6. Performance Comparison on Animal Dataset.
Regarding the selection of missing rates: The 10%–30%–50% scheme follows the common practice in incomplete multi-view learning literature [,,] to ensure fair comparison. The additional 70% missing rate was specifically tested on the Animal dataset to evaluate our method’s robustness under extreme missing conditions, which is particularly meaningful for larger datasets where higher missing rates are more likely to occur in real applications.
During the model training, we observed that the singular values of each view could exhibit considerable differences in the t-SVD process. To fully leverage the structural information from different views, TCGL does not simply explore high-order correlations between views through tensors. Instead, it accounts for the variations in singular values across different views by assigning different auxiliary multipliers during training. Overall, from these three tables, it can be seen that the HCLS_CGL method performs remarkably well on multiple datasets, especially on Caltech7 and Animal datasets. This suggests that leveraging local structural information to assist consensus graph learning is an effective strategy. However, this method is limited to extracting information from low-dimensional space and fails to capture deeper correlations between multi-view data points, thus lacking breakthroughs in clustering performance improvement.
The similarity graph of the consensus graph learned by our method on the Caltech7 dataset under 30% and 50% missing-view rates are presented in Figure 2. As observed in the left sub-figure, the learned graph exhibits a clear block-diagonal structure, effectively capturing the underlying cluster structure. More importantly, even under a high missing rate of 50%, the consensus graph maintains a pronounced block-diagonal pattern, demonstrating the robustness and effectiveness of our approach in handling incomplete multi-view data.
Figure 2. The consensus graph learned by our method from the Caltech7 with 30% and 50% missing views.

Runtime and Complexity Analysis

To further evaluate the efficiency of the proposed TCGL method, we report both empirical runtime statistics and a theoretical analysis of computational complexity. Table 6 summarizes the total running time and the average per-iteration time on four benchmark datasets under a 30% missing rate.
  • Empirical runtime analysis: As shown in Table 7, TCGL achieves competitive efficiency on small- and medium-scale datasets such as ORL, UCI, and Caltech101-7, with average iteration times all within a few seconds. On the larger Animal dataset, the computation becomes more demanding due to the tensor SVD step, which is theoretically the dominant cost of the algorithm. Nevertheless, TCGL remains practically executable and converges within a reasonable number of iterations (typically 20–30), making the overall runtime manageable even for larger data.
    Table 7. Runtime statistics of TCGL on four benchmark datasets.
  • Theoretical complexity analysis: According to the decomposition of Algorithm 1, the main computational costs arise from: (i) the update of S, requiring O ( n 2 log n ) due to row-wise simplex projection; (ii) the closed-form update of each Z ( v ) , with complexity O ( n v 2 ) per view; (iii) the computation of the c smallest eigenvectors for updating F, with cost O ( n 2 c ) ; and (iv) the tensor SVD for updating J, with complexity O V · ( n 1 n 2 min ( n 1 , n 2 ) ) , which dominates the per-iteration computation. By aggregating these components, the overall per-iteration complexity of TCGL can be expressed as O V · ( n 1 n 2 min ( n 1 , n 2 ) ) + n 2 log n + n 2 c .
Algorithm 1 Tensorized Consensus Graph Learning for Incomplete Multi-View Clustering with Confidence Integration
Require:
Multi-view dataset X = X ( v ) R d v × n v = 1 V , trade-off parameters α , β > 0 , weight vector ω > 0 (initialized as ω i = 1 / min ( n 1 , n 2 ) and kept fixed), and cluster number c.
1:
Initialize Q = J = 0 , and set ρ ( ρ ( 1 , 2 ] ).
2:
while F ( k ) F ( k 1 ) F F ( k 1 ) F > 10 4  and  k < 30   do
3:
     for  v = 1 , , V  do
4:
           Update S ( v ) by solving the problem in Equation (26).
5:
     end for
6:
     for  v = 1 , , V  do
7:
           Update Z ( v ) by solving the problem in Equation (14).
8:
     end for
9:
      Update J by solving the problem in Equation (34).
10:
    Update Q by Q = Q + ρ ( Z J ) .
11:
    Update F , which is formed by c eigenvectors of L s = D s S T + S 2 corresponding to the c smallest eigenvalues.
12:
    Update the weight α v by solving the problem in Equation (11).
13:
    Update ρ by ρ = μ ρ .
14:
end while
Ensure:
The learned consensus graph S and the clustering labels obtained by performing k-means on the spectral embedding F .
  • Discussion: Although tensor operations contribute significantly to the computational cost, TCGL benefits from fast convergence and maintains competitive runtime compared with existing tensor-based multi-view clustering methods. The results demonstrate that TCGL achieves a favorable balance between clustering accuracy and computational efficiency.

5.3. Parameter Sensitivity Analysis

From the experimental results shown in Figure 3a, it can be observed that the proposed method is sensitive to the parameter μ , where the best performance is achieved when μ = 2 and β = 0.15 . This indicates that selecting appropriate parameters for the tensor low-rank processing module is crucial, and it plays an important role in improving clustering performance. By combining the subfigures (a) and (b) in Figure 3, we can observe that the proposed TCGL model consistently demonstrates superior clustering results across different datasets.
Figure 3. Parameters analysis: The performances in terms of ACC when using different trade-off parameters μ and β on the Caltech7 and UCI datasets, respectively.

5.4. Visualization

As shown in Figure 4, we visualize the t-SNE results on the Caltech7 and UCI datasets. These visualizations clearly demonstrate that TCGL effectively reveals an underlying clustering structure.
Figure 4. t-SNE visualization results at 30% incomplete rate, where the numbers represent different classes. (ac) Original features, HCLS_CGL representations, and TCGL representations on the Caltech7 dataset. (df) Original features, HCLS_CGL representations, and TCGL representations on the UCI dataset.

5.5. Convergence Analysis

As previously mentioned, the complex objective function is decomposed into six convex subproblems. Minimizing each subproblem leads to a continuous decrease in the objective value. This implies that during the alternating update process of these variables, the loss of objective function (5) remains monotonically decreasing. In Figure 5, we present the variation of the objective function loss with the number of iterations on the Caltech7 and UCI datasets with a missing-view rate of 30%. From the figure, it can be observed that the objective loss consistently decreases and quickly converges to a stable state, which confirms the good convergence property of the proposed optimization algorithm.
Figure 5. The objective function value and clustering accuracy versus the number of iterations of the proposed method on the (a) Caltech7 dataset where the missing-view rate is set as 30% and the (b) UCI dataset where the missing-view rate is set as 10%.

6. Conclusions

In this paper, we propose a Tensorized Consensus Graph Learning with Confidence Integration (TCGL) method to address challenges in incomplete multi-view clustering, including inefficient local information utilization, noise sensitivity, and insufficient capture of higher-order cross-view correlations. By filtering missing samples to preserve valid data structure, employing tensor nuclear norm sparsification to model multi-view higher-order relationships, and integrating confidence-based probability matrices for noise robustness, TCGL enables effective clustering of incomplete multi-view data. Experimental evaluations across benchmark datasets demonstrate its superiority over state-of-the-art methods in capturing cross-view complementary information and handling data incompleteness. Notably, on the Caltech7 dataset with 30% missing rate, TCGL achieved the most significant improvements, outperforming the second-best method by 25.34% in ACC and 39.89% in NMI. Future research will explore optimizing TCGL for large-scale scenarios and integrating it with deep learning for complex non-linear data analysis.
Although TCGL achieves strong results, several directions remain open. The tensor and spectral operations may become computationally demanding on very large datasets, motivating more efficient large-scale approximations. The row-sum constraint on Z ( v ) may also be less suitable when data neighborhoods are highly irregular. Moreover, our experiments assume MCAR missingness, while behavior under MNAR settings requires further study. Extending TCGL toward neural clustering frameworks and learning adaptive view-specific weights are also promising directions for improving flexibility and modeling power.

Author Contributions

Conceptualization, G.J. and H.J.; Methodology, G.J., H.J., Z.C., and W.C.; Resources, G.J.; Data Curation, G.J., H.J., and W.C.; Writing—Original Draft Preparation, H.J., W.C., and G.J.; Writing—Review and Editing, G.J., H.J., Z.C., and W.C.; Funding Acquisition, G.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62306048; in part by the Changzhou Applied Basic Research Fund Project under Grant CQ20230092 and CJ20235036.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, H.; Chen, Y.; Yao, M.; Liu, W.; Peng, J.; Fu, X. Tensor Completion Framework by Graph Refinement for Incomplete Multi-view Clustering. IEEE Trans. Multimed. 2025, 1–14. [Google Scholar] [CrossRef]
  2. Zhang, D.; Li, H.; He, D.; Liu, N.; Cheng, L.; Wang, J.; Han, J. Unsupervised Pre-training with Language-Vision Prompts for Low-Data Instance Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 8642–8657. [Google Scholar] [CrossRef] [PubMed]
  3. Peng, J.; Li, M.; Wang, B.; Wang, H. Omni contextual aggregation networks for high-fidelity image inpainting. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 6129–6144. [Google Scholar] [CrossRef]
  4. Liu, Y.; Li, C.; Xu, S.; Han, J. Part-whole relational fusion towards multi-modal scene understanding. Int. J. Comput. Vis. 2025, 133, 4483–4503. [Google Scholar] [CrossRef]
  5. Jiang, G.; Peng, J.; Wang, H.; Mi, Z.; Fu, X. Tensorial multi-view clustering via low-rank constrained high-order graph learning. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 5307–5318. [Google Scholar] [CrossRef]
  6. Yao, M.; Wang, H.; Chen, Y.; Fu, X. Between/within view information completing for tensorial incomplete multi-view clustering. IEEE Trans. Multimed. 2024, 27, 1538–1550. [Google Scholar] [CrossRef]
  7. Zhang, D.; Cheng, L.; Liu, Y.; Wang, X.; Han, J. Mamba capsule routing towards part-whole relational camouflaged object detection. Int. J. Comput. Vis. 2025, 133, 7201–7221. [Google Scholar] [CrossRef]
  8. Liu, Y.; Cheng, D.; Zhang, D.; Xu, S.; Han, J. Capsule Networks With Residual Pose Routing. IEEE Trans. Neural Netw. Learn. Syst. 2024, 36, 2648–2661. [Google Scholar] [CrossRef]
  9. Feng, L.; Wang, H.; Jin, B.; Li, H.; Xue, M.; Wang, L. Learning a Distance Metric by Balancing KL-Divergence for Imbalanced Datasets. IEEE Trans. Syst. Man Cybern. Syst. 2019, 49, 2384–2395. [Google Scholar] [CrossRef]
  10. Wang, H.; Yao, M.; Jiang, G.; Mi, Z.; Fu, X. Graph-Collaborated Auto-Encoder Hashing for Multiview Binary Clustering. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 10121–10133. [Google Scholar] [CrossRef]
  11. Hu, Z.; Nie, F.; Chang, W.; Hao, S.; Wang, R.; Li, X. Multi-view spectral clustering via sparse graph learning. Neurocomputing 2020, 384, 1–10. [Google Scholar] [CrossRef]
  12. Cai, B.; Wang, H.; Yao, M.; Fu, X. Focus more on what? Guiding multi-task training for end-to-end person search. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 7266–7278. [Google Scholar] [CrossRef]
  13. Liu, X.; Li, M.; Tang, C.; Xia, J.; Xiong, J.; Liu, L.; Kloft, M.; Zhu, E. Efficient and effective regularized incomplete multi-view clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2634–2646. [Google Scholar] [CrossRef] [PubMed]
  14. Wen, J.; Liu, C.; Xu, G.; Wu, Z.; Huang, C.; Fei, L.; Xu, Y. Highly confident local structure based consensus graph learning for incomplete multi-view clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 15712–15721. [Google Scholar]
  15. Liu, J.; Wang, C.; Gao, J.; Han, J. Multi-view clustering via joint nonnegative matrix factorization. In Proceedings of the 2013 SIAM International Conference on Data Mining, SIAM, Austin, TX, USA, 2–4 May 2013; pp. 252–260. [Google Scholar]
  16. Khan, M.A.; Khan, G.A.; Khan, J.; Khan, M.R.; Atoum, I.; Ahmad, N.; Shahid, M.; Ishrat, M.; Alghamdi, A.A. Multi-view clustering based on multiple manifold regularized non-negative sparse matrix factorization. IEEE Access 2022, 10, 113249–113259. [Google Scholar] [CrossRef]
  17. Luong, K.; Nayak, R.; Balasubramaniam, T.; Bashar, M.A. Multi-layer manifold learning for deep non-negative matrix factorization-based multi-view clustering. Pattern Recognit. 2022, 131, 108815. [Google Scholar] [CrossRef]
  18. Kumar, S.; Ying, J.; Cardoso, J.V.d.M.; Palomar, D.P. A unified framework for structured graph learning via spectral constraints. J. Mach. Learn. Res. 2020, 21, 1–60. [Google Scholar]
  19. Wang, H.; Yao, M.; Chen, Y.; Xu, Y.; Liu, H.; Jia, W.; Fu, X.; Wang, Y. Manifold-based incomplete multi-view clustering via bi-consistency guidance. IEEE Trans. Multimed. 2024, 26, 10001–10014. [Google Scholar] [CrossRef]
  20. Chen, M.S.; Wang, C.D.; Lai, J.H. Low-rank tensor based proximity learning for multi-view clustering. IEEE Trans. Knowl. Data Eng. 2022, 35, 5076–5090. [Google Scholar] [CrossRef]
  21. Wang, H.; Jiang, G.; Peng, J.; Deng, R.; Fu, X. Towards adaptive consensus graph: Multi-view clustering via graph collaboration. IEEE Trans. Multimed. 2022, 25, 6629–6641. [Google Scholar] [CrossRef]
  22. Kilmer, M.E.; Braman, K.; Hao, N.; Hoover, R.C. Third-order tensors as operators on matrices: A theoretical and computational framework with applications in imaging. SIAM J. Matrix Anal. Appl. 2013, 34, 148–172. [Google Scholar] [CrossRef]
  23. Kilmer, M.E.; Martin, C.D. Factorization strategies for third-order tensors. Linear Algebra Its Appl. 2011, 435, 641–658. [Google Scholar] [CrossRef]
  24. Zhang, Z.; Ely, G.; Aeron, S.; Hao, N.; Kilmer, M. Novel methods for multilinear data completion and de-noising based on tensor-SVD. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3842–3849. [Google Scholar]
  25. Wen, J.; Zhang, Z.; Xu, Y.; Zhang, B.; Fei, L.; Liu, H. Unified embedding alignment with missing views inferring for incomplete multi-view clustering. AAAI Conf. Artif. Intell. 2019, 33, 5393–5400. [Google Scholar] [CrossRef]
  26. Ng, A.; Jordan, M.; Weiss, Y. On spectral clustering: Analysis and an algorithm. Adv. Neural Inf. Process. Syst. 2001, 14, 1–8. [Google Scholar]
  27. Von Luxburg, U. A tutorial on spectral clustering. Stat. Comput. 2007, 17, 395–416. [Google Scholar] [CrossRef]
  28. Feng, J.; Lin, Z.; Xu, H.; Yan, S. Robust subspace segmentation with block-diagonal prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3818–3825. [Google Scholar]
  29. Kang, Z.; Peng, C.; Cheng, Q. Clustering with adaptive manifold structure learning. In Proceedings of the 2017 IEEE 33rd International Conference on Data Engineering (ICDE), Diego, CA, USA, 19–22 April 2017; pp. 79–82. [Google Scholar]
  30. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; Ma, Y. Robust recovery of subspace structures by low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 171–184. [Google Scholar] [CrossRef]
  31. Nie, F.; Wang, X.; Jordan, M.; Huang, H. The constrained laplacian rank algorithm for graph-based clustering. AAAI Conf. Artif. Intell. 2016, 30, 1969–1976. [Google Scholar] [CrossRef]
  32. Nie, F.; Li, J.; Li, X. Self-weighted multiview clustering with multiple graphs. In Proceedings of the IJCAI, Melbourne, Australia, 19–25 August 2017; pp. 2564–2570. [Google Scholar]
  33. Zhan, K.; Zhang, C.; Guan, J.; Wang, J. Graph learning for multiview clustering. IEEE Trans. Cybern. 2017, 48, 2887–2895. [Google Scholar] [CrossRef] [PubMed]
  34. Wu, J.; Xie, X.; Nie, L.; Lin, Z.; Zha, H. Unified graph and low-rank tensor learning for multi-view clustering. AAAI Conf. Artif. Intell. 2020, 34, 6388–6395. [Google Scholar] [CrossRef]
  35. Wang, H.; Yang, Y.; Liu, B. GMC: Graph-based multi-view clustering. IEEE Trans. Knowl. Data Eng. 2019, 32, 1116–1129. [Google Scholar] [CrossRef]
  36. Duchi, J.; Shalev-Shwartz, S.; Singer, Y.; Chandra, T. Efficient projections onto the l 1-ball for learning in high dimensions. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 272–279. [Google Scholar]
  37. Zhao, H.; Liu, H.; Fu, Y. Incomplete multi-modal visual data grouping. In Proceedings of the IJCAI, New York, NY, USA, 9–15 July 2016; pp. 2392–2398. [Google Scholar]
  38. Shao, W.; He, L.; Yu, P.S. Multiple incomplete views clustering via weighted nonnegative matrix factorization with regularization. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Porto, Portugal, 7–11 September 2015; pp. 318–334. [Google Scholar]
  39. Hu, M.; Chen, S. Doubly aligned incomplete multi-view clustering. arXiv 2019, arXiv:1903.02785. [Google Scholar] [CrossRef]
  40. Shao, W.; He, L.; Lu, C.T.; Yu, P.S. Online multi-view clustering with incomplete views. In Proceedings of the 2016 IEEE International Conference on Big Data (Big Data), Washington, DC, USA, 5–8 December 2016; pp. 1012–1017. [Google Scholar]
  41. Hu, M.; Chen, S. One-pass incomplete multi-view clustering. AAAI Conf. Artif. Intell. 2019, 33, 3838–3845. [Google Scholar] [CrossRef]
  42. Liu, X.; Zhu, X.; Li, M.; Wang, L.; Zhu, E.; Liu, T.; Kloft, M.; Shen, D.; Yin, J.; Gao, W. Multiple kernel k k-means with incomplete kernels. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 1191–1204. [Google Scholar] [CrossRef] [PubMed]
  43. Wang, H.; Zong, L.; Liu, B.; Yang, Y.; Zhou, W. Spectral perturbation meets incomplete multi-view data. arXiv 2019, arXiv:1906.00098. [Google Scholar] [CrossRef]
  44. Wen, J.; Yan, K.; Zhang, Z.; Xu, Y.; Wang, J.; Fei, L.; Zhang, B. Adaptive graph completion based incomplete multi-view clustering. IEEE Trans. Multimed. 2020, 23, 2493–2504. [Google Scholar] [CrossRef]
  45. Li, Z.; Tang, C.; Liu, X.; Zheng, X.; Zhang, W.; Zhu, E. Tensor-based multi-view block-diagonal structure diffusion for clustering incomplete multi-view data. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  46. Xu, J.; Li, C.; Ren, Y.; Peng, L.; Mo, Y.; Shi, X.; Zhu, X. Deep Incomplete Multi-View Clustering via Mining Cluster Complementarity. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Online, 22 February–1 March 2022; pp. 8761–8769. [Google Scholar]
  47. Wang, S.; Liu, X.; Liu, L.; Tu, W.; Zhu, X.; Liu, J.; Zhou, S.; Zhu, E. Highly-efficient incomplete large-scale multi-view clustering with consensus bipartite graph. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 9776–9785. [Google Scholar]
  48. Liu, C.; Wu, Z.; Wen, J.; Xu, Y.; Huang, C. Localized sparse incomplete multi-view clustering. IEEE Trans. Multimed. 2022, 25, 5539–5551. [Google Scholar] [CrossRef]
  49. Xia, W.; Gao, Q.; Wang, Q.; Gao, X. Tensor Completion-Based Incomplete Multiview Clustering. IEEE Trans. Cybern. 2022, 52, 13635–13644. [Google Scholar] [CrossRef] [PubMed]
  50. Zhang, C.; Li, H.; Lv, W.; Huang, Z.; Gao, Y.; Chen, C. Enhanced tensor low-rank and sparse representation recovery for incomplete multi-view clustering. AAAI Conf. Artif. Intell. 2023, 37, 11174–11182. [Google Scholar] [CrossRef]
  51. Wang, J.; Zhao, Z.; Dobigeon, N.; Chen, J. Joint Tensor and Inter-View Low-Rank Recovery for Incomplete Multiview Clustering. arXiv 2025, arXiv:2503.02449. [Google Scholar]
  52. Wen, J.; Zhang, Z.; Zhang, Z.; Zhu, L.; Fei, L.; Zhang, B.; Xu, Y. Unified tensor framework for incomplete multi-view clustering and missing-view inferring. AAAI Conf. Artif. Intell. 2021, 35, 10273–10281. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.