1. Introduction
In the field of pattern recognition, subspace clustering (SC) is a very important research topic [
1,
2,
3,
4]. In recent decades, scholars have created many subspace clustering algorithms, among which spectral-type methods have shown good performance.
Sparse subspace clustering (SSC) [
5] and low-rank representation (LRR) [
6], which have achieved significant success, are two typical self-representation subspace clustering methods. Block diagonal structure is a matrix form where non-zero elements are confined to square blocks along the main diagonal, with all elements outside these blocks being zero. Recent studies [
7,
8] have shown that the block diagonal structure within the learned low-dimensional subspace projection serves the purpose of obtaining the correct clustering results. However, SSC and LRR pursue the block diagonal representation (BDR) matrix indirectly since they only impose nuclear-norm and L1-norm on the subspace representation, respectively. Furthermore, Feng et al. [
8] imposed a block diagonal prior on the subspace representation matrices obtained using SSC and LRR, and their clustering performance was improved. However, it is difficult to optimize Feng’s method since the rank constraint is NP-hard. To tackle this problem, Lu et al. [
7] developed a simple BDR that relaxes the rank constraint. Compared with Feng’s method, BDR is more easily optimized since it is smooth. Xu et al. [
9] developed a learning projective model for BDR to deal with the large-scale subspace clustering problem. Xing et al. [
10] proposed an enhanced version of DBSCAN, which is a highly prevalent algorithm in data mining, to improve the clustering process using the block diagonal property of affinity matrices. Meanwhile, Guo et al. [
11] put forward a spectral clustering algorithm with BDR for large-scale datasets.
The above-mentioned approaches are single-view-based since they assume that there is only one data source. However, in fact, data are generally sourced from various origins. For instance, one event can be represented by images, text, videos. As such, multi-view clustering (MVC) methods, which often demonstrate a better clustering performance than single-view methods [
12,
13,
14,
15,
16,
17,
18,
19], are becoming increasingly popular.
The authors of [
20] proposed co-training learning to flux the multi-view features. Additionally, the study in [
21] investigated the key factors contributing to the success of the co-training method. The co-training learning model is not robust enough against noise pollution, which can lead to error exaggeration. Kumar et al. [
17] proposed an MSC framework, in which the clustering hypotheses among views is co-regularized. Graph-based methods are another category MSC methods, which generally use the multiple graph fusion strategy to utilize the information among different views. Sa [
22] developed a two-view clustering method, which utilizes different information between two views by constructing bipartite graphs. Moreover, the authors of [
13] developed a multi-view spectral clustering algorithm with the help of low-rank and sparse decomposition (RMSC), achieving encouraging success in relation to several real datasets. In the work of Cao et al. [
15], the diversity-induced MSC (DiMSC) was presented, leveraging the Hilbert–Schmidt Independence Criterion (HSIC), which plays a key role in utilizing complementary information among different views to enhance the clustering.
By assuming that the different views of an object come from a potential subspace, subspace learning MVC methods can be developed to capture the shared potential subspace. Blaschko and Lampert [
23] introduced a novel spectral clustering technique that utilizes canonical correlation analysis (CCA) in its linear and kernel forms for dimensionality reduction. In [
24], a low-rank common subspace (LRCS) MVC method is proposed, which can obtain compatible intrinsic information among views by using a common low-rank projection.
These MVC methods shows promising performance in clustering applications; however, they only use paired associations between different views, and may overlook the higher-order associations hidden in multi-view data [
12,
14,
19,
25,
26,
27]. Zhang et al. [
12,
19] developed a novel multi-view spectral clustering method named LTMSC, incorporating low-rank tensor constraints. In the method, the subspace representations are constructed into a single tensor. It is possible to explore higher-order relationships hidden in the multi-view data. Lu et al. [
28] introduced an MSC method with hyper-Laplacian regularization and low-rank tensor constraints (HLR-MSCLRT), which can uncover the local information hidden in the data on the manifold. Nevertheless, the tensor norm employed in both LTMSC and HLR-MSCLRT lacks a clear physical interpretation.
Zhang et al. recently introduced the TNN [
29] leveraging the tensor singular value decomposition (t-SVD). The TNN, defined as the summation of singular values, provides a rigorous measure of tensor data low-rankness. In [
14], Xie et al. developed a t-SVD based MSC model, namely t-SVD-MSC, which preserves the low-rank property through TNN. With the use of TNN, t-SVD-MSC can more effectively explore the complementary information among all the views [
30,
31,
32,
33,
34]. Furthermore, in [
18], an essential tensor learning method for MSC using a TNN constraint, known as ETLMSC, is proposed. Pan et al. [
35] proposed a non-negative non-convex low-rank tensor kernel function in an MSC model (NLRTGC) to reduce the bias from rank. To exploit high-dimensional hidden information, Pan et al. [
36] proposed a low-rank fuzzy MSC learning algorithm with the TNN constraint (LRTGFL). Peng et al. [
37] designed log-based non-convex functions to approximate tensor rank and tensor sparsity in the Finger-MVC model; these are more precise than the convex ones. Wang et al. [
38] integrate noise elimination and subspace learning into a unified MSC framework, holding high-order associations of views constrained by the TNN. Du et al. [
39] proposed a robust t-SVD-based multi-view clustering which simultaneously uses low rank and local smooth priors. Luo et al. [
40] used an adaptively weighted tensor Schatten-p norm with an adjustable
p-value to eliminate the biased estimate of rank.
The optimized BDR structure in affinity matrices inherently encodes cluster information, thereby substantially enhancing clustering efficacy. The low-rank tensor representations intrinsically capture latent high-order correlations across multi-view data through subspace embeddings, resulting in statistically significant clustering improvements. In this paper, inspired by the optimized BDR structure and low-rank tensor representations, we propose a novel MSC method called TMSC-TNNBDR, which integrates the advantages of TNN and BDR. The proposed model imposes BDR constraints on each subspace representation matrix, and all affinity matrices are combined into a tensor regularized by TNN. Finally, an efficient optimization algorithm based on ALM is developed.
The primary contributions of our work are as follows:
The proposed TMSC-TNNBDR incorporates a BDR regularizer, which promotes a more pronounced block diagonal structure and improves clustering robustness.
In the TMSC-TNNBDR model, the optimized architecture encodes a TNN constraint, under which TMSC-TNNBDR captures the global structure across all views, thereby effectively exploiting latent complementary information and high-order interactions among views.
We proposed an ALM optimizer for TMSC-TNNBDR. This approach demonstrates superior clustering performance over comparative algorithms while maintaining competitive computational efficiency.
The remainder of this work is organized as follows. In
Section 2, we summarize the notations used and some preliminary definitions. In
Section 3, we briefly review two methods, namely the LRR [
6] and the BDR [
7]. Then, we propose the TMSC-TNNBDR and a solving procedure for TMSC-TNNBDR in
Section 4. Subsequently we documented the experimental findings in
Section 5. Ultimately, we conclude our work in
Section 6.
4. The Proposed TMSC-TNNBDR
In this section, we introduce the TMSC-TNNBDR framework that extends classical LRR and BDR approaches. Subsequently, we derive an ALM-based optimization scheme to solve the resulting non-convex problem.
4.1. Problem Formulation
Let
and
be, respectively, the feature matrix and subspace coefficient for the
vth view. The loss function of TMSC-TNNBDR is demonstrated as follows:
where
represents an function that stacks all
(
v = 1, 2, …,
V) into a tensor in
then applying a rotation transformation to
.
In Equation (10), is the self-representation reconstruction error, denotes the BDR constraint to , denotes TNN low-rank constraint to , and can be seen as a Robust PCA term to remove the noise contained in the H(v). Moreover, , , and are tunable hyperparameters.
4.2. Optimization
The loss function of TMSC-TNNBDR, i.e., Equation (10), can be optimized through the ALM. The theorem relating to is described as follows:
Theorem 1 ([
41])
. Suppose , where L is semi-positive; then, the following holds: In accordance with Theorem 1, Equation (10) can be rewritten as Equation (12):
To solve Equation (12), an auxiliary tensor variable
is introduced to replace
. Then, the loss function of TMSC-TNNBDR is converted into the following:
Equation (13) will be converted to the augmented Lagrangian formula, as follows:
where
denotes the Lagrange multiplier;
is actually the penalty parameter.
We get the resolutions to , , , and by solving each variable alternately in Equation (14). The steps are described as follows:
-subproblem: For computing
, we fix the other variables and tackle the following problem:
Differentiating by
, we can obtain the following:
-subproblem:
will be computed as follows:
For Equation (17),
, where
is a matrix concatenated from
k eigenvectors that correspond to the
k smallest eigenvalues of
[
7].
-subproblem:
can be computed as follows:
Equation (18) can be converted into the following:
The theorem in [
7] enables the solution of Equation (19).
-subproblem: We fixed the other variables and update
as follows:
The solution to Equation (20) can be obtained using the theorem in [
14,
29].
-subproblem: the Lagrange multiplier
can be updated as follows:
Finally, the TMSC-TNNBDR procedure is outlined in Algorithm 1.
Algorithm 1 TMSC-TNNBDR |
Input: , , , and cluster number k; Output: Clustering result Initialize: , ,, , , , , , While not converged do for do Update in accordance with Equation (16); Update in accordance with Equation (17); Update in accordance with Equation (19) end Update in accordance with Equation (20); Update in accordance with Equation (21); Update by ; Check the convergence conditions: . end Let ; Perform spectral clustering on S.
|
4.3. Computational Complexity and Convergence
To calculate , it involves matrix multiplication and matrix inversion, whose complexities are and , respectively. For computing , its complexity is because the main computational burdens are eigenvalue decomposition and matrix product. The computation of is since it mainly depends on matrix multiplication. As for computing , the computational complexity is . Thereafter, the total complexity of TMSC-TNNBDR is .
The procedure of TMSC-TNNBDR is non-convex, which means it cannot achieve a global optimal solution. Nevertheless, TMSC-TNNBDR can converge to a local optimal point. In fact, each variable in Algorithm 1 has a closed-form solution. Following this, the value of the loss function decreases monotonically and remains bounded below. Clustering experiments are performed on some classic datasets, and the results showed that the TMSC-TNNBDR could converge stably.