Next Article in Journal
False Information Mitigation Using Pattern-Based Anomaly Detection for Secure Vehicular Networks
Previous Article in Journal
Enhancing Multi-Factor Authentication for Mobile Devices Through Cryptographic Zero-Knowledge Protocols
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Commonness and Inconsistency Learning with Structure Constrained Adaptive Loss Minimization for Multi-View Clustering

1
College of Computer Science and Technology, Qingdao University, Qingdao 266071, China
2
Engineering Training Center, Chongqing University of Science and Technology, Chongqing 401331, China
3
Faculty of Information Science and Engineering, Ocean University of China, Qingdao 266000, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2025, 14(9), 1847; https://doi.org/10.3390/electronics14091847
Submission received: 11 February 2025 / Revised: 23 March 2025 / Accepted: 29 March 2025 / Published: 1 May 2025

Abstract

:
Subspace clustering has emerged as a prominent research focus, demonstrating remarkable potential in handling multi-view data by effectively harnessing their diverse and information-rich features. In this study, we present a novel framework for multi-view subspace clustering that addresses several critical aspects of the problem. Our approach introduces three key innovations: First, we propose a dual-component representation model that simultaneously considers both consistent and inconsistent elements across different views. The consistent component is designed to capture shared structural patterns with robust commonality, while the inconsistent component effectively models view-specific variations through sparsity constraints across multiple modes. Second, we implement cross-mode sparsity constraints that enable the inconsistent component to efficiently extract high-order information from the data. This design not only enhances the representation capability of the inconsistent component but also facilitates the consistent component in revealing high-order structural relationships within the data. Third, we develop an adaptive loss function that offers greater flexibility in handling noise and outliers, thereby significantly improving the model’s robustness in real-world applications. Through extensive experimentation, we demonstrate that our proposed method consistently outperforms existing approaches, achieving superior clustering performance across various benchmark datasets. The experimental results comprehensively validate the effectiveness and advantages of our approach in terms of clustering accuracy, robustness, and computational efficiency.

1. Introduction

As a basic approach in machine learning, clustering has proven to be highly effective in diverse areas, including image segmentation [1], face clustering [2], community detection [3], cancer biology [4], text mining [5], etc. High-dimensional data often contain redundant or irrelevant features that can obscure meaningful patterns, making traditional clustering methods less effective. By reducing dimensionality, innovative techniques help to reveal the underlying structure of the data, improving both the efficiency and accuracy of clustering algorithms. Thus, there has been an ongoing requirement to develop innovative techniques, such as low-dimensional representations [6,7,8], for clustering that can elucidate the latent structures within high-dimensional data.
Subspace clustering has been well studied during the last decades with great success in constructing low-dimensional representations for efficient clustering [1,7,9,10,11]. Among them, the subspace clustering methods built upon spectral clustering principles have attracted considerable interest and become the focus of recent developments in subspace clustering methods [6,7,12]. The sparse subspace clustering [13] and low-rank representation [7] approaches are among the most typical methods. Various algorithms have been developed with improved clustering performance [12,14]. However, in practical applications, data often originate from various sources [15]. As an example, we can utilize various features such as texts, images, or other types of characteristics to depict an object [15]. These features are considered as multi-view representations of the object, providing a comprehensive and diverse perspective on the data and resulting in multi-view data. Although multi-view data offer richer and more informative information than single-view data, single-view methods are unable to effectively utilize the features available across multiple views and thus they may have suboptimal learning performance.
To tackle this challenge, multi-view clustering techniques have been thoroughly investigated over the past decade, yielding promising clustering performance [16,17,18,19,20]. Existing methods can be broadly categorized into two types based on their approach to constructing the affinity matrix [21], which are briefly summarized in Figure 1. The first type constructs the affinity matrix directly from similarity matrices [16,17,18,22], while the second type, known as multi-view subspace clustering, builds on spectral clustering-based subspace clustering techniques [15,19,23]. For instance, LT-MSC extends the traditional low-rank representation framework by incorporating a tensor nuclear norm constraint to explore cross-view correlations and global structures [24]. Similarly, CSMSC emphasizes both consistency and specificity in representations, enabling the identification of shared and view-specific characteristics [25]. Another notable method, MLRR, improves low-rank representation by applying symmetry constraints and diversity regularization to preserve angular consistency and capture view-specific information [26]. MVCIL tackles catastrophic forgetting and concept interference in open-ended incremental learning via random representation, orthogonal fusion, and selective weight consolidation [27]. The EOMSC-CA method integrates anchor learning and graph construction, and by imposing graph connectivity constraints, it directly outputs clustering labels [28].
Despite these advancements, existing methods still face several critical limitations. First, many approaches fail to effectively exploit the cross-view high-order information and spatial structure of the data, resulting in incomplete representations that do not well capture the underlying relationships across views. Second, while consistency and diversity across views are crucial for multi-view clustering, current methods often struggle to balance these aspects, leading to either over-emphasis on shared structures or neglect of view-specific characteristics. Third, the robustness of existing methods to noise and outliers remains a challenge, as many approaches rely on rigid loss functions that are sensitive to data imperfections.
To address these limitations, we propose a novel multi-view clustering method that integrates consistency and diversity across views while capturing high-order information and enhancing robustness. Our approach introduces the following key innovations: (1) We adopt an adaptive loss function that combines the benefits of various matrix norms. This strategy effectively mitigates the effects of noise and outliers, and thus makes the proposed model more robust. (2) We separate the representation matrix of each view into two parts that correspond to consistency and inconsistency of the data, respectively. In particular, the consistency part exhibits the desired structural property with strict commonality and block diagonal structural constraints, which enhances the affinity matrix to explicitly identify the grouping pattern of the data. Meanwhile, the inconsistency part captures high-order information of the data through sparsity constraints imposed on different modes. This strategy effectively helps reveal latent cross-view structures of the data. (3) The proposed model involves several terms, which are complicated and not straightforward to solve. We design an effective optimization algorithm for the proposed method, which generalizes the practical applicability of the proposed techniques. (4) Through extensive experimental results, the effectiveness and superiority of our method have been effectively demonstrated, which further improves the learning performance of the MVC algorithms.
The rest of this paper is organized as follows: Section 2 introduces the essential background information. In Section 3, we introduce the proposed method, followed by the development of an optimization algorithm in Section 4. In Section 5, we perform comprehensive experiments to confirm the efficacy of the proposed method. Finally, we summarize the findings and conclude the paper in Section 6.

2. Background Knowledge

Herein, we introduce the required notations and offer a concise review of related methods. Given multi-view data X , we denote its number of samples by n, number of views by V, number of clusters by k, dimension of each view by d v with v = 1 , , V , and the largest dimension is d = M a x v = 1 V { d v } . We represent the feature matrix for each view as X ( v ) R d v × n . For ease of notation, the superscript ( · ) ( v ) is also used to denote the v-th frontal slice of the tensors. Subsequently, a self-expression-based multi-view subspace clustering model can be universally formulated as follows:
min Ƶ ( 1 ) , , Ƶ ( V ) G ( X ( v ) , X ( v ) Ƶ ( v ) ) + τ ψ ( Ƶ ( v ) ) ,
where tensor Ƶ R n × n × V consists of representation matrices in each frontal slice, G ( · ) serves as the loss function of Ƶ ( v ) , τ 0 serves as a balancing parameter, and ψ ( · ) functions as a regularizer of Ƶ ( v ) . When V = 1 , the aforementioned model simplifies to a traditional single-view subspace clustering method like the tensor nuclear norm and low-rank representation, while employing their respective regularizers.

3. Proposed Method

In general, utilizing the self-expressive property allows us to formulate the multi-view clustering model as
min Ƶ v = 1 V X ( v ) X ( v ) Ƶ ( v ) F 2 + f ( Ƶ ) ,
where f ( Ƶ ) denotes a proper regularizer for the affinity tensor. This regularizer enforces constraints on the structural characteristics of the low-dimensional representation within each individual view. Generally, when a model is fitted using the samples, most of the samples lead to a small loss while a big loss is generated by only a few samples [29].
To alleviate the loss function’s sensitivity to outliers, as well as to marry the advantages of different norms, for enhanced robustness and accurate information preservation from the data, we adopt an adaptive loss function [29]. In particular, for an arbitrary matrix A R a × b , the adaptive loss function can be formulated as follows:
A σ = i ( 1 + σ ) A i 2 2 A i 2 + σ
where the i-th row of the matrix A is denoted by A i and σ > 0 is a parameter. The adaptive loss function is closely related to various matrix norms, of which the properties are summarized in lemmaadaptive.
Lemma 1
([29]). For matrix A R a × b , the following are the characteristics of the adaptive loss function:
1.
The norm A σ is nonnegative and convex.
2.
The norm A σ is twice differentiable.
3.
If i { 1 , , a } , A i σ σ , then A σ ( 1 + σ ) σ A F 2 .
4.
If i { 1 , , a } , A i σ σ , then A σ ( 1 + σ ) A 2 , 1 , where A 2 , 1 = i = 1 m J = 1 n A i J 2 1 / 2 .
5.
lim σ 0 A σ A 2 , 1 .
6.
lim σ X σ X F 2 .
Thus, with the adaptive loss function, we may develop Equation (2) into
min Ƶ v = 1 V X ( v ) X ( v ) Ƶ ( v ) σ + f ( Ƶ ) .
In general, different views of the data exhibit both consistency and inconsistency, where the consistency reflects the shared information across all views, while the inconsistency captures the unique diversity information of each view. To effectively capture the consistency and inconsistency information of the representation matrices among different views, we follow and decompose the representation tensor Ƶ into consistent and inconsistent parts as Ƶ = C + E [30], which are captured by C R n × n × V and E R n × n × V , respectively. In this paper, we aim to assist in capturing the strong commonalities among different views, and we propose to extract a common part from the multiple representation matrices, which is denoted as C R n × n . Thus, each view’s representation matrix can be factorized as Ƶ ( v ) = C + E ( v ) with a hard commonness constraint. Therefore, we may further develop Equation (4) into
min C , E v = 1 V X ( v ) X ( v ) ( C + E ( v ) ) σ + g ( C , E ) ,
where g ( C , E ) serves as an appropriate regularizer to enforce the desired structural properties of both C and E . In general, it is expected that the inconsistency of different views should not be significantly large since different views have close connections, which suggests that E is sparse. To enhance the sparse structures of the inconsistency part, we consider the following two cases. First, the sample-wise sparsity has been widely considered, which leads to mode-2 sparsity of E . Second, for each sample, the representation vectors in different views should be similar, which leads to a mode-3 sparse structure of E . Thus, to ensure that E has a joint mode-2 and mode-3 sparse structure, we further develop Equation (5) into
min C , E v = 1 V X ( v ) X ( v ) ( C + E ( v ) ) σ + α i j v E i j v 2 1 / 2 + j v i E i j v 2 1 / 2 ,
where α is a balancing parameter. Although the sparsity of E helps capture the commonness across different views to construct C , we still need to exploit the structural constraint for C for the desired structural properties of the common representation. First, it is natural to impose structural constraints on the consensus affinity matrix, which gives rise to
min C , E v = 1 V X ( v ) X ( v ) ( C + E ( v ) ) σ + α i j v E i j v 2 1 / 2 + j v i E i j v 2 1 / 2 , s . t . diag ( C ) = 0 , C 0 , C = C ,
where diag ( · ) extracts the diagonal elements of the input matrix and arranges them into a column vector. The constraints imposed in Equation (7) require that C is nonnegative and symmetric and avoid a trivial interpretation of C that the coefficient of a sample itself is the largest by restricting diag ( C ) = 0 , which enforces the natural and desired property for an affinity matrix. To further exploit the local and nonlinear structures inherent in the data, we introduce the graph Laplacian to our model, which leads to
min C , E v = 1 V X ( v ) X ( v ) ( C + E ( v ) ) σ + α i j v E i j v 2 1 / 2 + j v i E i j v 2 1 / 2 + β v = 1 V Tr ( C L ( v ) C ) , s . t . diag ( C ) = 0 , C 0 , C = C ,
where β serves the role of a balancing parameter and L R n × n × V denotes a tensor that consists of graph Laplacian matrices of different views in frontal slices. It is seen that the manifold learning embedded in the third term of Equation (8) provides a consensus local geometric structure of C across different views, which is important in revealing the structures of the data [31]. The structural constraints naturally treat C as a graph matrix, which measures neighbor relationships of the data. For such data, the connected components are k, each corresponding to a distinct cluster.
To guarantee such property, we refer to the theorem stated below.
Theorem 1 
([32]). For any C 0 and C = C , let L ( C ) be its Laplacian matrix. The number of connected components present in C is reflected by the multiplicity k of the eigenvalue 0 of L C .
Theorem 1 inspires us to restrict that L ( C ) has exactly k zero eigenvalues for the desired connectivity property of C . It is straightforward to impose constraint Rank ( L ( C ) ) into Equation (7) to restrict the rank of L ( C ) , with Rank ( · ) being the rank operator. However, it is widely recognized that this problem, which involves rank minimization, is NP-hard [32], making it challenging to solve. Thus, we refer to the following k-block diagonal regularizer for a relaxed rank constraint.
Definition 1 
([14]). Given an affinity matrix B R n × n , a k-block diagonal regularizer can be defined as
B k = i = n k + 1 n λ i ( L ( B ) ) ,
where λ i ( · ) denotes the matrix singular value of the input and is the i-th largest.
It is seen that the k-block diagonal regularizer can be treated as a relaxed rank constraint, by minimizing which we may ideally restrict the desired rank for the target graph Laplacian matrix. Then, C ideally contains k connected sub-graphs, which correspond to the k clusters of the data and thus reveal the group structure of the data. Thus, with the connectivity constraint of C , we further develop our model as
min C , E v = 1 V X ( v ) X ( v ) ( C + E ( v ) ) σ + β v = 1 V Tr ( C L ( v ) C ) + α i j v E i j v 2 1 / 2 + j v i E i j v 2 1 / 2 + γ C k , s . t . diag ( C ) = 0 , C 0 , C = C ,
where γ is a balancing parameter. The aforementioned model is named Commonness and Inconsistency Learning with Structure Constrained Adaptive Loss Minimization for Multi-view Clustering (CISCAL-MVC). We summarize the flow diagram of the CISCAL-MVC in Figure 2. We will present an effective optimization algorithm to address Equation (10) in the next section.

4. Optimization

In this section, we will develop an effective optimization algorithm for Equation (10) by employing the augmented Lagrange multiplier method (ALM). First, we introduce several auxiliary variables to facilitate the optimization and construct the augmented Lagrange function. Then, our algorithm efficiently solves the optimization problem by decomposing it into several manageable subproblems. For all the subproblems, we provide their corresponding closed-form solutions using various techniques, such as the Sylvester equation, shrinkage operator, etc. We update each of the variables while fixing the others till convergence. The detailed optimization procedure is presented as follows:
Since the fourth term is highly nonlinear and difficult to solve, we convert it to an equivalent form to facilitate the optimization. First, we may refer to the following Theorem 2 as a useful tool.
Theorem 2 
([33]). L R n × n and L 0 , the following equality holds:
i = n k + 1 n λ i ( L ) = inf L , W | W R n × n , 0 W I , Tr ( W ) = k ,
where · denotes the inner product of matrices.
Then, according to Definition 1 and Theorem 2, we may rewrite C k as the following convex programming problem:
min W R n × n , 0 W I , Tr ( W ) = k L ( C ) , W .
Thus, we may convert the original objective into an equivalent form as follows:
min C , E , W v = 1 V X ( v ) X ( v ) ( C + E ( v ) ) σ + α i j v E i j v 2 1 / 2 + j v i E i j v 2 1 / 2 , + β v = 1 V Tr ( C L ( v ) C ) + γ Diag ( C 1 ) C ) , W s . t . diag ( C ) = 0 , C 0 , C = C , 0 W I , Tr ( W ) = k ,
where Diag ( · ) denotes an operator that transforms the input vector into a diagonal matrix. To solve Equation (13), we first introduce auxiliary variables J = C R n × n and S = F = E R n × n × V . Then, the augmented Lagrange function of equation Equation (13) can be expressed as
L ( J , C , E , F , S ) = v = 1 V X ( v ) X ( v ) ( J + S ( v ) ) σ + α i j v E i j v 2 1 / 2 + j v i F i j v 2 1 / 2 + β v = 1 V Tr ( J L ( v ) J ) + γ Diag ( C 1 ) C ) , W + μ 2 J C + Y / μ F 2 + μ 2 S ( v ) E ( v ) + Y 1 ( v ) / μ F 2 + μ 2 F ( v ) E ( v ) + Y 2 ( v ) / μ F 2 , s . t . diag ( C ) = 0 , C 0 , C = C , 0 W I , Tr ( W ) = k ,
where the Lagrange multipliers are given by Y R n × n , Y 1 R n × n × V , and Y 2 R n × n × V , and the penalty parameter of the ALM is given by μ , respectively.

4.1. J -Subproblem

The subproblem connected with J is
min J v = 1 V X ( v ) X ( v ) ( J + S ( v ) ) σ + β v = 1 V Tr ( J L ( v ) J ) + μ 2 J ( ) C + Y / μ F 2 .
Define H R n × n × V and D R n × n × V with H ( v ) = S ( v ) I   and   D i J ( v ) = I { i = J } · ( 1 + σ ) · ( X ( v ) ) i ( J + H ( v ) ) 2 + 2 σ 2 ( ( X ( v ) ) i ( J + H ( v ) ) 2 + σ ) 2 , then Equation (15) can be restructured as:
min J v = 1 V Tr D ( v ) X ( v ) ( H ( v ) + J ( H ( v ) + J ( X ( v ) ) + β v = 1 V Tr ( J L ( v ) J ) + μ 2 J C + Y / μ F 2 .
The above problem appears complicated in J and is not straightforward to solve. To facilitate the optimization, we first reformulate the above problem into the following form:
min J Tr ( J v = 1 V ( X ( v ) ) D ( v ) X ( v ) J ) + 2 Tr J v = 1 V ( X ( v ) ) D ( v ) X ( v ) H ( v ) + β Tr J v = 1 V L ( v ) J + μ 2 J C + Y / μ F 2 .
To ensure that the above problem has a minimizer, we need to show that the above problem is convex. For the first term, it is seen that ν R n   and   ν 0 , the following inequality holds:
ν v = 1 V ( X ( v ) ) D ( v ) X ( v ) ν = v = 1 V ν ( X ( v ) ) D ( v ) X ( v ) ν = v = 1 V ν ( X ( v ) ) ( ( D ( v ) ) 1 / 2 ) ( D ( v ) ) 1 / 2 X ( v ) ν = v = 1 V ( D ( v ) ) 1 / 2 X ( v ) ν 2 2 0 ,
which confirms the convexity of the first term of Equation (17). Since L ( v ) is the graph Laplacian matrix constructed from X ( v ) , we have L ( v ) 0 and thus v = 1 V L ( v ) 0 as well [31]. Moreover, it is clear that the second term of Equation (17) is linear while the last term is quadratic. Then, it is clear that Equation (16) is quadratic and convex, which can be solved by the first-order optimality condition:
v = 1 V 2 ( X ( v ) ) D ( v ) X ( v ) · J + J · 2 β v = 1 V L ( v ) + v = 1 V 2 ( X ( v ) ) D ( v ) X ( v ) H ( v ) + μ J + Y μ C = 0 .
The above equation is essentially a Sylvester equation. To be clearer, we define:
Θ 1 : = 2 v = 1 V ( X ( v ) ) D ( v ) X ( v ) + μ I , Θ 2 : = 2 β v = 1 V L ( v ) , Θ 3 : = v = 1 V 2 ( X ( v ) ) D ( v ) X ( v ) H ( v ) μ C + Y .
Then, by plugging Equation (20) into Equation (19), we obtain a standard Sylvester equation as follows:
Θ 1 J + J Θ 2 + Θ 3 = 0 .
For the Sylvester equation in Equation (21), we may adopt various algorithms such as the Bartels–Stewart [34] or the Hessenberg–Schur [35] algorithm to efficiently resolve the problem, which generally requires O ( n 3 ) complexity. It is noted that we may also apply other algorithms, such as the gradient descent, to the original optimization problem of Equation (17). By using the gradient, we need to obtain the gradient of the objective at each iteration, which requires O ( n 3 ) complexity. The Sylvester equation solver directly addresses the first-order optimality conditions, typically achieving an exact solution in a finite number of steps, making it highly efficient and numerically stable for small to medium-scale problems. In contrast, gradient descent relies on iterative updates, often exhibiting slower convergence, particularly for ill-conditioned problems, leading to potential robustness issues. Without loss of generality, we adopt the Sylvester equation solver and summarize the above steps as the following operator:
J = Sylvester ( Θ 1 , Θ 2 , Θ 3 ) .

4.2. C -Subproblem

The subproblem for C is presented as follows:
min C γ Diag ( C 1 ) C ) , W + μ 2 J C + Y / μ F 2 s . t . diag ( C ) = 0 , C 0 , C = C .
With straightforward algebra, the aforementioned problem can be reformulated equivalently as
min C C J Y γ · diag ( W ) 1 + γ W μ F 2 , s . t . diag ( C ) = 0 , C 0 , C = C .
For ease of notation, we define
C ¯ : = J + Y γ · diag ( W ) 1 + γ W μ .
Then, according to [14], Equation (24) admits a closed-form solution:
C = C ¯ + C ¯ 2 Diag ( diag ( C ¯ ) ) + ,
where ( · ) + denotes the nonnegative projection.

4.3. W -Subproblem

The subproblem of W is as follows:
min 0 W I , Tr ( W ) = k Diag ( C 1 ) C ) , W .
According to [36], Equation (26) admits a closed-form solution:
W = PP ,
where the eigenvectors of L ( C ) corresponding to the smallest k eigenvalues are represented by P .

4.4. S -Subproblem

The subproblem associated with S is as follows:
min S v = 1 V X ( v ) X ( v ) ( J + S ( v ) ) σ + μ 2 S E + Y 1 / μ F 2 .
In a way similar to Equation (16), we may rewrite the above problem into:
min S v = 1 V Tr ( J I + S ( v ) ) ( X ( v ) ) D ( v ) X ( v ) ( J I + S ( v ) ) + μ 2 S E + Y 1 / μ F 2 .
It is evident that Equation (29) can be solved by slice. For each S ( v ) , it is clear that the corresponding problem is quadratic and convex, and a closed-form solution can be obtained by the first-order optimality condition:
2 ( X ( v ) ) D ( v ) X ( v ) ( J I + S ( v ) ) + μ ( S ( v ) E ( v ) + Y 1 ( v ) / μ ) = 0 .
Thus, S can be obtained by slice as
S ( v ) = 2 ( X ( v ) ) D ( v ) X ( v ) + μ I 1 · μ E ( v ) Y 1 ( v ) 2 ( X ( v ) ) D ( v ) X ( v ) ( J I ) .

4.5. E -Subproblem

For E , the subproblem is:
min E α i j v E i j v 2 1 / 2 + μ 2 S ( v ) E ( v ) + Y 1 ( v ) / μ F 2 + μ 2 F ( v ) E ( v ) + Y 2 ( v ) / μ F 2 .
The above problem is essentially a 2 , 1 -shrinkage problem of E along the mode-3 fiber, which can be clearly seen by a simplified standard form in the following:
min E α 2 μ i j v E i j v 2 1 / 2 + 1 2 E ( v ) S ( v ) + F ( v ) + Y 1 ( v ) / μ + Y 2 ( v ) / μ 2 F 2 .
Then, according to [37], Equation (33) admits a closed-form solution along the mode-3 fiber as follows:
E i j v = ( v = 1 V E ^ i j v 2 ) 1 / 2 α 2 μ ( v = 1 V E ^ i j v 2 ) 1 / 2 · E ^ i j v · I { ( v = 1 V E ^ i j v 2 ) 1 / 2 > α 2 μ } ,
where E ^ = S + F + Y 1 / μ + Y 2 / μ 2 , and I { · } is a function that returns 0 if it does not satisfy the condition in the subscript and returns 0 if it does.

4.6. F -Subproblem

For F , the subproblem is
min F α j v i F i j v 2 1 / 2 + μ 2 F E + Y 2 / μ F 2 .
Similarly to Equation (32), the aforementioned problem can be solved using the 2 , 1 -shrinkage operator; so, we have the following closed-form solution:
F i j v = ( i = 1 n F ^ i j v 2 ) 1 / 2 α μ ( i = 1 n F ^ i j v 2 ) 1 / 2 · F ^ i j v · I { ( i = 1 n F ^ i j v 2 ) 1 / 2 > α μ } ,
where F = E Y 2 / μ .

4.7. Updating of Y , Y 1 , Y 2 , and μ

We update the Lagrange multiplier and penalty parameters using classical methods, as outlined below:
Y = Y + μ ( J C ) , Y 1 = Y 1 + μ ( S E ) , Y 2 = Y 2 + μ ( F E ) , μ = μ κ ,
where κ > 1 keeps that μ is increasing. We iterate the above steps until convergence is achieved, then we may obtain a unified view-commonness matrix C . Finally, we apply standard spectral clustering techniques to C to achieve the final clustering result.

4.8. Time Complexity

We analyze the time complexity of our method as follows. In particular, we analyze the dominant operations that determine the complexity of our algorithm at each iteration. For the J-subproblem, it needs to solve a Sylvester equation, which requires a time complexity of O ( V n 2 d + n 3 ) by utilizing the Bartels–Stewart algorithm [34]. For the C-subproblem, it requires a time complexity O ( n 3 ) . The time complexity for constructing graph Laplacian matrices is O ( n 3 ) . Similarly to J, the W-subproblem requires a time complexity of O ( V n 2 + n 2 n ) for calculating the matrix multiplications and solving the Sylvester equations. For the S -subproblem, it has a similar structure to J and requires a time complexity of O ( V n 3 + V n 2 d v ) . Regarding the subproblems of E and F , both entail l2,1 norm contraction operations, leading to a time complexity of O ( V n 2 ) for each. In summary, it requires a total time complexity of O ( V n 3 + V n 2 d + n 2 k ) .

5. Experiments

In this section, experiments are carried out to validate the effectiveness of our method. Especially, we compare our method with nine baseline methods, which will be introduced in the following subsection. Six datasets are utilized, including the 3Sources [38], ORL [39], WebKB-Texas [40], COIL-20 [41], Caltech-7 [42], and Yale [43] datasets. We adopt four evaluation metrics—the clustering accuracy (ACC) as a measure of cluster assignment correctness, normalized mutual information (NMI) as an indication of clustering quality, purity (PUR) as a metric for cluster homogeneity, and the adjusted rand index (ARI) as a measure of clustering agreement—whose details can be found in [44,45,46,47,48]. Basically, the first three metrics values range within [0, 1] while the last one is within [−1, 1], where the larger value indicates better performance. All experiments in this paper are conducted using MATLAB 2018a on a workstation equipped with a 6-core Intel i5-8500T CPU running at 2.10 GHz and 16 GB of memory. The detailed experimental setup, along with the corresponding results, is presented in the subsequent parts of this section. The code for our method can be accessed on 28 March 2025, at https://www.researchgate.net/profile/Chong-Peng-8/publications.

5.1. Methods in Comparison

In the experiment, we conduct a comparison between our approach and nine other baseline methods: (1) SC [49] is a traditional clustering technique for single-view data. In our experiment, we execute SC independently on each view and present the optimal results; (2) MVSpec [50] extends the SC for multi-view data, which uses a weighted combination of kernel matrices of different views for clustering; (3) RMSC [22] constructs probability matrices from different views and learns a unified probability matrix with low-rank constraint for multi-view clustering; (4) MVGL [17] creates a unified graph by integrating affinity graphs from various views under a low-rank constraint, which is utilized for multi-view clustering; (5) BMVC [46] performs binary learning for representation and clustering; (6) SMVSC [11] performs anchor learning collaboratively and builds a unified graph to represent the underlying structures of the data; (7) MVSC [51] constructs a unified bipartite graph for clustering based on cross-view sample-anchor bipartite graphs; (8) FPMVS-CAG [52] introduces a parameter-free method that aims to obtain an anchor representation by combining anchor selection with graph construction for clustering; (9) JSMC [30] concurrently integrates the cross-view commonness and inconsistencies into the learning of subspace representations. (10) MvDSCN [53] learns a multi-view self-representation matrix in an end-to-end manner by integrating multiple backbones and combining convolutional autoencoders with self-representation.

5.2. Datasets

In our experiment, we utilize six benchmark datasets. The main features of these datasets are outlined in Table 1, and we briefly introduce them as follows: (1) 3Sources [38] comprises data from 948 news articles. Each article presents multiple points of view and is drawn from three online news platforms: BBC, Reuters, and the Guardian. We only select the instances that have complete information across all views. The feature dimensions of the three news platforms are 3068d, 3631d, and 3560d, respectively. (2) ORL [39] contains 400 facial images of 40 individuals. The images were captured under various lighting conditions, showcasing a variety of facial expressions and intricate details. Three types of features, including the 4096d intensity, 6750d Gabor, and 3304d LBP, are preserved in this dataset. (3) WebKB-Texas [40] consists of 187 documents from the University of Texas Department of Computer Science homepage. These documents belong to five categories, including student, project, course, staff, and faculty. Each document has two types of features, including 187d citation and 1703d content. (4) COIL-20 [41] consists of 1440 grayscale images of 20 objects. For each object, 72 images are captured from different angles. Three types of features are preserved in this dataset, including the 1024d intensity, 3304d LBP, and 6750d Gabor. (5) Caltech-7 [42] selects 7 out of a total number of 101 classes from the Caltech-101 dataset, including Faces, Garfield, Snoopy, StopSign, Windsor-Chair, Dollar Bill, and Motorbikes. This dataset contains three types of features, including 512d GIST, 1984d HOG, and 928d LBP. (6) Yale [43] consists of 165 facial images of 15 persons. These images are captured with varying facial expressions under different lighting conditions. There are three types of features in this dataset, including 4096d intensity, 6750d Gabor, and 3304d LBP. We visually show some examples of the image datasets in Figure 3.

5.3. Clustering Performance

For all methods in the comparison, we utilize a grid search strategy for parameter tuning in the experiment. For our method, we adjust the balancing parameters α , β , and γ from the set S 1 = { 10 3 , 10 2 , , 10 3 } , with the parameter σ within the set S 2 = { 1 , 2 , 3 , , 10 } , respectively. For the baseline methods, if not otherwise clarified in the original paper, we tune the balancing parameters within S 1 as well. For each method, we enumerate all possible combinations of parameter values on each dataset and record the highest result for each evaluation metric, respectively. The clustering results are presented in detail in Table 2.
The experimental results demonstrate that the proposed CISCAL-MVC consistently achieves the best clustering performance across all benchmark datasets. In general, it secures the first-place results in 21 out of 24 cases and the second-best results in 3 cases, which is highly promising. Among the baseline methods, the JSMC and MvDSCN emerge as the most competitive, achieving at least the top three results in 24 and 14 out of 24 cases, respectively. Compared to other baseline methods, JSMC demonstrates substantial improvements. Based on the results, the following key observations can be made:
(1) On the 3Sources dataset, the CISCAL-MVC achieves the highest scores in all metrics, improving the second-best performance by 1.46% in ACC, 2.54% in NMI, 0.73% in ARI, and 2.96% in PUR, respectively. Notably, the MvDSCN, JSMC, and CISCAL-MVC are the only three methods that achieve results higher than 60% across all metrics, highlighting their robustness in handling multi-view data.
(2) On the ORL dataset, the CISCAL-MVC dominates with 89.25% ACC, 93.59% NMI, 83.58% ARI, and 90.50% PUR, which is highly promising. Among the baseline methods, the MvDSCN and JSMC are the most competitive, securing the second and third-best results, respectively. This demonstrates CISCAL-MVC’s ability to effectively capture both consistency and diversity across views.
(3) On the WebKB-Texas dataset, the CISCAL-MVC again achieves the best performance, with 77.83% ACC, 42.24% NMI, 44.84% ARI, and 77.83% PUR. While the MvDSCN shows competitive performance on the 3Sources and ORL datasets, it fails to rank among the top three on this dataset. This observation further emphasizes the robustness of CISCAL-MVC in handling diverse and challenging data structures. Moreover, the WebKB-Texas and 3Sources datasets are collected from a real-world news platform and academic website, thus the effectiveness of the proposed method on these datasets demonstrates the potential applicability in real-world scenarios.
(4) On the COIL-20 dataset, the CISCAL-MVC achieves perfect scores (100%) across all metrics, demonstrating its exceptional ability to effectively capture the underlying data structure. Among the baseline methods, the JSMC and MVGL are the most competitive, but their results are lower than CISCAL-MVC by approximately 16.04% in ACC, 6.20% in NMI, 18.72% in ARI, and 11.32% in PUR. Although MvDSCN is inferior to MVGL on this dataset, it is among the only three methods that consistently achieve results higher than 80%, showcasing the potential of deep learning approaches.
(5) On the Caltech-7 dataset, the CISCAL-MVC achieves the highest results, improving the second-best performance by 9.79% in ACC and 8.60% in ARI. While the JSMC and MvDSCN show competitive performance in certain metrics, they lack consistency across all metrics. For instance, the third-best results are obtained by different methods, such as the MvDSCN and MVGL, indicating their instability. In contrast, the CISCAL-MVC demonstrates superior stability and performance, making it a more reliable choice.
(6) On the Yale dataset, the CISCAL-MVC achieves the best result in purity (74.72%) and the second-best results in the other metrics. On this dataset, the CISCAL-MVC, MvDSCN, and JSMC show comparable performance, with at least 10.81% improvement in ACC, 8.73% in NMI, and 10.71% in ARI compared to other baseline methods. This further highlights the effectiveness of the CISCAL-MVC in leveraging high-order information and cross-view relationships.
To further validate the superiority of the proposed method from a statistical perspective, we apply the Wilcoxon signed-rank test to compare the clustering performance of the proposed method against baseline methods. Specifically, we aggregate the results across multiple datasets and evaluation metrics for each method. The Wilcoxon signed-rank test is then conducted on the combined results, with the statistical outcomes presented in Table 3. The results reveal that the p-value is consistently less than 0.001 in all cases, providing strong statistical evidence that the proposed method significantly outperforms the baseline methods.
In summary, the CISCAL-MVC consistently outperforms all baseline methods across different datasets and metrics. While the MvDSCN and JSMC show competitive performance in certain cases, they are unable to match the overall superiority of CISCAL-MVC. Notably, the JSMC demonstrates robustness across datasets, but other baseline methods fail to achieve consistent performance. Compared to individual baseline methods, the improvements achieved by the CISCAL-MVC are more significant. These observations confirm the effectiveness, as well as the potential real-world applicapability, of the proposed method.

5.4. Ablation Study

To assess the importance of incorporating various components in our model, we conduct an ablation study for evaluation. We conduct experiments comparing our model with three of its variants. In particular, these variants are developed by removing a specific term from our model, which is obtained by setting the corresponding parameter, i.e., β , α , or γ to zero, respectively. For the variant, we adhere to the experimental settings detailed in Section 5.3 to adjust the remaining parameters. We report the results of the CISCAL-MVC, as well as its variants, in Table 4.
Based on the results we obtained, it can be observed that the overall model always generates superior performance to the variants. For example, compared with these variants, the CISCAL-MVC shows about 20.98–65.74%, 9.26–41.88%, and 8.19–30.23% improvements across all datasets in purity, respectively, which is quite notable. In other metrics, we consistently observe that the performance of the model is significantly degraded when any single component is removed from the overall architecture. This consistent pattern across multiple evaluations underscores the critical role each component plays in the model’s functionality and overall effectiveness. Thus, it becomes evident that the various components used in our model are not only essential but also synergistically contribute to its robustness and superior performance.

5.5. Visual Illustration of Data Distribution

In this test, we visually show how the consistent and inconsistent parts of the representation behave in the learning process by showing the distribution of the data. Without loss of generality, we show the results using the ORL dataset from two perspectives.
First, we visually present C and the different views of E in Figure 4. The results demonstrate that the consistent part effectively preserves the strong group structure of the data, while the inconsistent part captures cross-view inconsistencies without exhibiting clear structural patterns. Furthermore, we visualize the distributions of the features recovered by the consistent and inconsistent representations using t-SNE in Figure 5. The results reveal that the features derived from the consistent representation exhibit distinct cluster patterns, whereas those from the inconsistent representation lack any discernible clustering structure. These observations provide a clear visual illustration of how the consistent and inconsistent components of the representation behave during the clustering process. They also validate the effectiveness of CISCAL-MVC in uncovering meaningful cluster patterns within the data.

5.6. Convergence Analysis

In this convergence analysis test, we carry out experiments to provide empirical evidence of the convergence behavior exhibited by the proposed method. Through these experiments, we aim to provide robust empirical evidence that substantiates the efficiency and reliability of the method’s convergence properties. Without compromising generality, we plot the error curves of the 3Sources, ORL, and COIL-20 datasets, where the parameters are fixed at β = 1 , α = 100 , γ = 0.1 , and σ = 9 . For ease of representation, we use ( t ) in the subscript to denote the number of iterations. Since C is used for the final clustering, we focus on the convergent behavior of { C ( t ) } . In this test, we cancel the other termination conditions and run our algorithm for 200 iterations. Then, we plot the first 200 values of the sequence { C ( t + 1 ) C ( t ) F } that are generated on the 3Sources, ORL, and COIL-20 datasets in Figure 6. Based on the experimental results, we find that the CISCAL-MVC converges in approximately 50 iterations, which is quite significant.
Next, we undertake additional experiments to delve deeper into the convergent behavior of the CISCAL-MVC. In particular, we treat C ( 200 ) as the convergent solution, which is denoted as C ( ) for clarity. Then, we plot the first 50 values of the sequence { C ( t + 1 ) C ( ) F / C ( t ) C ( ) F } that are generated on the 3Sources, ORL, and COIL-20 datasets in Figure 7, respectively. From the results, we may observe that the values of { C ( t + 1 ) C ( ) F / C ( t ) C ( ) F } are in the range of [ 0 , 1 ] ; this means that CISCAL-MVC has about a linear convergent rate.
These observations confirm the convergent behavior of the CISCAL-MVC, as well as its efficiency, that is essentially important for learning algorithms for guaranteed performance.

5.7. Comparison of Time Cost

In this test, we compare the time cost of our method with the baselines. Among the baseline methods, the MvDSCN and JSMC show the second and third best performance. Among the remaining baseline methods, the RMSC shows relatively better performance. Since the clustering result is essential, for clarity of presentation, we only compare our method with the above three baseline methods that show superior performance among these methods.
First, we compare the time complexity of these methods. In particular, the MvDSCN, RMSC and JSMC have time complexity of O ( 2 v = 1 V ( n 3 + 2 l = 1 D M l 2 K l 2 Q l v Q l ) ) , O ( V n 2 + n 3 + n k ) , O ( V n 3 ) , respectively, Where M represents the spatial dimension of the input data, l represents the l-th convolutional layer, Q represents the number of convolutional kernels in that layer, and K represents the size of each convolutional kernel. It can be seen that the non-deep learning methods generally require O ( n 3 ) complexity, which are comparable. Next, we empirically compare the time cost of these methods in Figure 8. From the results, it can be seen that the CISCAL-MVC is several times faster than the RMSC and JMSC on the COIL-20 and Caltech-7 datasets. On other datasets, such as the 3Sources, WebKB-Texas, and Yale, all methods terminate within a few seconds. These observations confirm that the CISCAL-MVC is at least comparable to, if not faster than, the baseline methods from both theoretical time complexity and empirical time cost perspectives. Considering the superiority of the CISCAL-MVC in clustering performance, such time cost is acceptable.

5.8. Parametric Sensitivity

In this experiment, we empirically investigate the impact of key parameters on the learning performance of the proposed method. Without loss of generality, we use the ORL dataset to present the results in Figure 9, with fixed values of μ = ? and κ = ? . To systematically evaluate parameter sensitivity, we vary the parameters over the following ranges: α { 0.001 , 0.01 , , 1000 } , β { 0.001 , 0.01 , , 1000 } , γ { 0.001 , 0.01 , , 1000 } , and σ { 1 , 2 , , 10 } . For each parameter, we analyze its individual impact on the performance of CISCAL-MVC while keeping the other parameters optimized. The experimental results reveal that the performance of CISCAL-MVC exhibits minimal variation across different values of α and σ , indicating robustness to these parameters. Furthermore, CISCAL-MVC achieves superior performance when smaller values are selected for β and γ . Notably, the method demonstrates competitive performance across a wide range of small values for β and γ These findings suggest that CISCAL-MVC is generally insensitive to its parameter settings. For practical implementation, we recommend selecting small values for β and γ , while adopting moderate values for α and σ .

5.9. Data Reconstruction

In this test, we conduct experiments to empirically showcase how the consistency and inconsistency parts of the representation capture features from the data. For illustrative purposes, and without compromising generality, we use the Yale and ORL datasets. Since the Yale dataset does not contain pixel features, we additionally introduce the original pixel features to the dataset as the 4-th view and use the augmented data X to learn C and E , which provides a clearer visualization of the learning effects. Then, we visually show some of the original and reconstructed features in Figure 10, respectively. Based on the results, we make the following key observations: First, the features reconstructed by the overall representation matrix capture well the information from the original data, where the recovered features are visually close to the original ones. Second, the features reconstructed by the consistency part well capture the key features of the data, and the reconstructed features of different samples within the same group are visually similar with strong commonness. Third, the features captured by the inconsistency part show the diversity of the samples. These findings highlight the importance of both the consistency and inconsistency components of the representation in capturing different types of features.

6. Conclusions

In this paper, a novel approach for multi-view subspace clustering, named CISCAL-MVC, is presented. First, the novel method utilizes an adaptive loss function, which is more flexible to account for noise effects. Moreover, both the consistency and inconsistency parts of the representation matrices are exploited across different views. In particular, the consistency part exhibits hard commonness with the desired structural property, while the inconsistency part captures cross-mode sparsity to account for cross-mode differences. Then, the cross-mode sparsity constraints promote the inconsistency part to effectively exploit high-order information, thereby enhancing the consistency part to reveal high-order information of the data as well. The effectiveness and superiority of the proposed method are confirmed by extensive experiments.
Despite the effectiveness of CISCAL-MVC, there are still several aspects that could be further improved. First, the high time complexity of CISCAL-MVC limits its scalability to large-scale datasets. One potential solution to enhance efficiency is to adopt the anchor-based strategy, which approximates the representation of samples using a selected set of anchor points. This approach can be formulated as: X ( v ) A ( v ) ( C ˜ + E ˜ ( v ) ) , where A R d v × m are the anchors, m is the number of anchor points, C ˜ R m × n denotes the consistency representation, and E ˜ R m × n × V is the inconsistent part of the representation. Since m n , the approach may significantly reduce the time complexity of the CISCAL-MVC. Second, while the current formulation of CISCAL-MVC incorporates the local geometric structure of the data on a manifold, it may not effectively capture the complex high-order neighbor relationships inherent in the data [54]. To address this limitation, inspired by recent advancements in graph-based learning [55], a high-order neighbor graph could be constructed to better preserve the intricate relationships among samples. This would allow the method to leverage more sophisticated structural information, potentially leading to improved clustering performance. Third, the proposed method requires that the dataset contains complete entries. For incomplete multi-view data, we present a potential approach as follows. We may first learn incomplete affinity matrices of the data using the observed entries. Then, we may adopt a low-rank tensor recovery technique to complete the affinity tensor for the subsequent learning procedure. However, since incomplete multi-view clustering is often considered as a different task from the general multi-view clustering [56], we do not fully expand this line of research in our current work.
The primary objective of this work is to develop a novel multi-view clustering framework that effectively utilizes cross-view high-order information, spatial structure, and both consistency and diversity across views. While the aforementioned improvements, such as anchor-based efficiency enhancement and high-order graph construction, could further advance the performance and scalability of CISCAL-MVC, they are beyond the scope of this paper. These directions, however, present promising avenues for future research.

Author Contributions

Conceptualization, K.Z., K.K. and C.P.; methodology, K.Z., K.K. and C.P.; software, K.Z.; validation, Y.B. and C.P.; formal analysis, K.Z. and K.K.; investigation, Y.B.; resources, K.Z. and K.K.; data curation, K.Z. and K.K.; writing—original draft preparation, K.Z. and K.K.; writing—review and editing, C.P.; visualization, K.Z. and K.K.; supervision K.Z., K.K., Y.B. and C.P.; project administration, K.Z., K.K., Y.B. and C.P.; funding acquisition, C.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Shandong Province Colleges and Universities Youth Innovation Technology Plan Innovation Team Project under Grant 2022KJ149 and the National Natural Science Foundation of China (NSFC) under Grant 62276147.

Data Availability Statement

We did not create new data, and the data sets we used were all existing mainstream data. However, due to privacy, we have chosen not to make the dataset public.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviation

The following abbreviation is used in this manuscript:
CISCAL-MVCCommonness and Inconsistency Learning with Structure Constrained Adaptive
Loss Minimization for Multi-view Clustering

References

  1. Lei, T.; Liu, P.; Jia, X.; Zhang, X.; Meng, H.; Nandi, A.K. Automatic fuzzy clustering framework for image segmentation. IEEE Trans. Fuzzy Syst. 2019, 28, 2078–2092. [Google Scholar]
  2. Peng, C.; Zhang, J.; Chen, Y.; Xing, X.; Chen, C.; Kang, Z.; Guo, L.; Cheng, Q. Preserving bilateral view structural information for subspace clustering. Knowl.-Based Syst. 2022, 258, 109915. [Google Scholar] [CrossRef]
  3. Park, N.; Rossi, R.; Koh, E.; Burhanuddin, I.A.; Kim, S.; Du, F.; Ahmed, N.; Faloutsos, C. Cgc: Contrastive graph clustering forcommunity detection and tracking. In Proceedings of the ACM Web Conference 2022, Lyon, France, 25–29 April 2022; pp. 1115–1126. [Google Scholar]
  4. Gönen, M.; Margolin, A.A. Localized data fusion for kernel k-means clustering with application to cancer biology. In Proceedings of the Advances in Neural Information Processing Systems, Quebec, QC, Canada, 8–13 December 2014; pp. 1305–1313. [Google Scholar]
  5. Guan, R.; Zhang, H.; Liang, Y.; Giunchiglia, F.; Huang, L.; Feng, X. Deep feature-based text clustering and its explanation. IEEE Trans. Knowl. Data Eng. 2020, 34, 3669–3680. [Google Scholar]
  6. Peng, C.; Kang, Z.; Li, H.; Cheng, Q. Subspace clustering using logdeterminant rank approximation. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015; pp. 925–934. [Google Scholar]
  7. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; Ma, Y. Robust recovery of subspace structures by low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 171–184. [Google Scholar]
  8. Peng, C.; Zhang, Y.; Chen, Y.; Kang, Z.; Chen, C.; Cheng, Q. Log-based sparse nonnegative matrix factorization for data representation. Knowl.-Based Syst. 2022, 251, 109127. [Google Scholar]
  9. Vidal, R.; Ma, Y.; Sastry, S. Generalized principal component analysis (gpca). IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1945–1959. [Google Scholar] [CrossRef]
  10. Rao, S.; Tron, R.; Vidal, R.; Ma, Y. Motion segmentation in the presence of outlying, incomplete, or corrupted trajectories. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 1832–1845. [Google Scholar]
  11. Sun, M.; Zhang, P.; Wang, S.; Zhou, S.; Tu, W.; Liu, X.; Zhu, E.; Wang, C. Scalable multi-view subspace clustering with unified anchors. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 3528–3536. [Google Scholar]
  12. Peng, C.; Kang, Z.; Cheng, Q. Subspace clustering via variance regularized ridge regression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2931–2940. [Google Scholar]
  13. Elhamifar, E.; Vidal, R. Sparse subspace clustering: Algorithm, theory, and applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar]
  14. Wang, F.; Chen, C.; Peng, C. Essential low-rank sample learning for group-Aware subspace clustering. IEEE Signal Process. Lett. 2023, 30, 1537–1541. [Google Scholar]
  15. Zhang, C.; Fu, H.; Hu, Q.; Cao, X.; Xie, Y.; Tao, D.; Xu, D. Generalized latent multi-view subspace clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 86–99. [Google Scholar]
  16. Wang, H.; Yang, Y.; Liu, B. Gmc: Graph-based multi-view clustering. IEEE Trans. Knowl. Data Eng. 2019, 32, 1116–1129. [Google Scholar] [CrossRef]
  17. Zhan, K.; Zhang, C.; Guan, J.; Wang, J. Graph learning for multiview clustering. IEEE Trans. Cybern. 2017, 48, 288–2895. [Google Scholar] [CrossRef]
  18. Kumar, A.; Rai, P.; Daume, H. Co-regularized multi-view spectral clustering. In Proceedings of the Advances in Neural Information Processing Systems, Granada, Spain, 12–17 December 2011; pp. 1413–1421. [Google Scholar]
  19. Wang, X.; Guo, X.; Lei, Z.; Zhang, C.; Li, S.Z. Exclusivity-consistency regularized multi-view subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 923–931. [Google Scholar]
  20. Bai, R.; Huang, R.; Qin, Y.; Chen, Y.; Xu, Y. A structural consensus representation learning framework for multi-view clustering. Knowl.-Based Syst. 2024, 283, 111132. [Google Scholar] [CrossRef]
  21. Wu, J.; Lin, Z.; Zha, H. Essential tensor learning for multi-view spectral clustering. IEEE Trans. Image Process. 2019, 28, 5910–5922. [Google Scholar] [PubMed]
  22. Xia, R.; Pan, Y.; Du, L.; Yin, J. Robust multi-view spectral clustering via low-rank and sparse decomposition. In Proceedings of the AAAI Conference on Artificial Intelligence, Québec, QC, Canada, 27–31 July 2014; pp. 2149–2155. [Google Scholar]
  23. Li, R.; Zhang, C.; Fu, H.; Peng, X.; Zhou, T.; Hu, Q. Reciprocal multi-layer subspace learning for multi-view clustering. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8172–8180. [Google Scholar]
  24. Zhang, C.; Fu, H.; Liu, S.; Liu, G.; Cao, X. Low-rank tensor constrained multiview subspace clustering. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1582–1590. [Google Scholar]
  25. Luo, S.; Zhang, C.; Zhang, W.; Cao, X. Consistent and specific multi-view subspace clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 3730–3737. [Google Scholar]
  26. Chen, J.; Yang, S.; Mao, H.; Fahy, C. Multiview subspace clustering using low-rank representation. IEEE Trans. Cybern. 2021, 52, 12364–12378. [Google Scholar] [CrossRef]
  27. Li, D.; Wang, T.; Chen, J.; Lian, C.; Zeng, Z. Multi-view class incremental learning. Inf. Fusion 2024, 102, 102021. [Google Scholar] [CrossRef]
  28. Liu, S.; Wang, S.; Zhang, P.; Xu, K.; Liu, X.; Zhang, C.; Gao, F. Efficient One-Pass Multi-View Subspace Clustering with Consensus Anchors. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; pp. 7576–7584. [Google Scholar]
  29. Nie, F.; Wang, H.; Huang, H.; Ding, C. Adaptive loss minimization for semisupervised elastic embedding. In Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence, Beijing, China, 3–9 August 2013; pp. 1565–1571. [Google Scholar]
  30. Cai, X.; Huang, D.; Zhang, G.Y.; Wang, C.D. Seeking commonness and inconsistencies: A jointly smoothed approach to multi-view subspace clustering. Inf. Fusion. 2023, 91, 364–375. [Google Scholar]
  31. Cai, D.; He, X.; Han, J. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1548–1560. [Google Scholar]
  32. Lu, C.; Feng, J.; Lin, Z.; Mei, T.; Yan, S. Subspace clustering by block diagonal representation. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 48–501. [Google Scholar] [CrossRef]
  33. Dattorro, J. Convex Optimization & Euclidean Distance Geometry, 2nd ed.; Meboo Publishing: Palo Alto, CA, USA, 2018. [Google Scholar]
  34. Bartels, R.H.; Stewart, G.W. Solution of the matrix equation AX + XB = C. Commun. ACM 1972, 15, 820–826. [Google Scholar] [CrossRef]
  35. Golub, G.; Nash, S.; Van, C. A Hessenberg-Schur method for the problem AX + XB= C. IEEE Trans. Autom. Control. 1979, 24, 909–913. [Google Scholar]
  36. Xie, X.; Guo, X.; Liu, G.; Wang, J. Implicit block diagonal low-rank representation. IEEE Trans. Image Process. 2017, 27, 477–489. [Google Scholar] [CrossRef] [PubMed]
  37. McCoy, M.B.; Tropp, J.A. Two proposals for robust pca using semidefinite programming. Electron. J. Stat. 2010, 5, 1123–1160. [Google Scholar] [CrossRef]
  38. Wang, H.; Li, H.; Fu, X. Auto-weighted mutli-view sparse reconstructive embedding. Multimed. Tools Appl. 2019, 78, 30959–30973. [Google Scholar] [CrossRef]
  39. Hu, Z.; Nie, F.; Wang, R.; Li, X. Multi-view spectral clustering via integrating nonnegative embedding and spectral embedding. Inf. Fusion 2020, 55, 251–259. [Google Scholar] [CrossRef]
  40. Craven, M.; DiPasquo, D.; Freitag, D.; McCallum, A.; Mitchell, T.; Nigam, K.; Slattery, S. Learning to extract symbolic knowledge from the world wide web. In Proceedings of the AAAI Conference on Artificial Intelligence, Madison, WI, USA, 26–30 July 1998; pp. 509–516. [Google Scholar]
  41. Nene, S.A.; Nayar, S.K.; Murase, H. Columbia Object Image Library (Coil-20); Department of Computer Science, Columbia University: New York, NY, USA, 1996. [Google Scholar]
  42. Li, F.-F.; Fergus, R.; Perona, P. Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. In Proceedings of the 2004 Conference on Computer Vision and Pattern Recognition Workshop, Washington, DC, USA, 27 June–2 July 2004; p. 178. [Google Scholar]
  43. Chen, M.S.; Huang, L.; Wang, C.D.; Huang, D.; Lai, J.H. Relaxed multiview clustering in latent embedding space. Inf. Fusion 2021, 68, 8–21. [Google Scholar] [CrossRef]
  44. Liu, X.; Zhu, X.; Li, M.; Wang, L.; Tang, C.; Yin, J.; Shen, D.; Wang, H.; Gao, W. Late fusion incomplete multi-view clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2410–2423. [Google Scholar] [CrossRef]
  45. Strehl, A.; Ghosh, J. Cluster ensembles—A knowledge reuse framework for combining multiple partitions. J. Mach. Learn. Res. 2002, 3, 583–617. [Google Scholar]
  46. Zhang, Z.; Liu, L.; Shen, F.; Shen, H.T.; Shao, L. Binary multi-view clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1774–1782. [Google Scholar] [CrossRef]
  47. Huang, D.; Wang, C.D.; Peng, H.; Lai, J.; Kwoh, C.K. Enhanced ensemble clustering via fast propagation of cluster-wise similarities. IEEE Trans. Syst. Man. Cybern. Syst. 2018, 51, 508–520. [Google Scholar] [CrossRef]
  48. Sharma, K.; Seal, A.; Yazidi, A.; Krejcar, O. S-Divergence-Based Internal Clustering Validation Index. Int. J. Interact. Multimedia Artif. Intell. 2023, 8, 127–139. [Google Scholar]
  49. Ng, A.; Jordan, M.; Weiss, Y. On spectral clustering: Analysis and an algorithm. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 3–8 December 2001; pp. 849–856. [Google Scholar]
  50. Tzortzis, G.; Likas, A. Kernel-based weighted multi-view clustering. In Proceedings of the 2012 IEEE 12th international conference on data mining, Brussels, Belgium, 10–13 December 2012; pp. 675–684. [Google Scholar]
  51. Li, Y.; Nie, F.; Huang, H.; Huang, J. Large-scale multi-view spectral clustering via bipartite graph. In Proceedings of the AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; pp. 2750–2756. [Google Scholar]
  52. Wang, S.; Liu, X.; Zhu, X.; Zhang, P.; Zhang, Y.; Gao, F.; Zhu, E. Fast parameter-free multi-view subspace clustering with consensus anchor guidance. IEEE Trans. Image Process. 2021, 31, 556–568. [Google Scholar] [CrossRef] [PubMed]
  53. Zhu, P.; Yao, X.; Wang, Y.; Hui, B.; Du, D.; Hu, Q. Multiview deep subspace clustering networks. IEEE Trans. Cybern. 2024, 54, 4280–4293. [Google Scholar]
  54. Peng, C.; Kang, K.; Chen, Y. Fine-grained essential tensor learning for robust multi-view spectral clustering. IEEE Trans. Image Process. 2024, 33, 3145–3160. [Google Scholar]
  55. Peng, C.; Zhang, K.; Chen, Y. Cross-view diversity embedded consensus learning for multi-view clustering. In Proceedings of the the Thirty-Third International Joint Conference on Artificial Intelligence, Jeju Island, Republic of Korea, 3–9 August 2024; pp. 4788–4796. [Google Scholar]
  56. Wen, J.; Zhang, Z.; Zhang, B.; Xu, Y.; Zhang, Z. A Survey on Incomplete Multiview Clustering. IEEE Trans. Cybern. 2023, 53, 1136–1149. [Google Scholar]
Figure 1. Flowdiagram illustrating graph-based and subspace clustering-based algorithms in the field of multi-view clustering. * denotes multiplication.
Figure 1. Flowdiagram illustrating graph-based and subspace clustering-based algorithms in the field of multi-view clustering. * denotes multiplication.
Electronics 14 01847 g001
Figure 2. Flow diagram of the proposed method. * denotes multiplication.
Figure 2. Flow diagram of the proposed method. * denotes multiplication.
Electronics 14 01847 g002
Figure 3. Visual examples of the image datasets. From top to bottom are ten samples selected from the ORL, COIL-20, Yale, and Caltech-7 datasets, respectively.
Figure 3. Visual examples of the image datasets. From top to bottom are ten samples selected from the ORL, COIL-20, Yale, and Caltech-7 datasets, respectively.
Electronics 14 01847 g003
Figure 4. Visual illustration of the overall, consistent, and inconsistent representation matrices of different views of the ORL dataset. In particular, we show the results of the second and third views. On the diagonal of the figures, clear group structures can be observed in the consistent representation.
Figure 4. Visual illustration of the overall, consistent, and inconsistent representation matrices of different views of the ORL dataset. In particular, we show the results of the second and third views. On the diagonal of the figures, clear group structures can be observed in the consistent representation.
Electronics 14 01847 g004
Figure 5. T−SNE illustration of the distributions of the original and recovered features of the ORL dataset. In particular, we show the results of the second and third views. The features recovered by the consistent representation have a clearer cluster structure than the original features of the data with t−SNE visualization.
Figure 5. T−SNE illustration of the distributions of the original and recovered features of the ORL dataset. In particular, we show the results of the second and third views. The features recovered by the consistent representation have a clearer cluster structure than the original features of the data with t−SNE visualization.
Electronics 14 01847 g005
Figure 6. Plot of the first 200 values of the sequence { C ( t + 1 ) C ( t ) F } generated by the CISCAL-MVC on 3Sources, ORL, and COIL-20 datasets, respectively. C ( t + 1 ) C ( t ) F represents the absolute difference of two consecutive updates, which measures the convergent behavior of the proposed algorithm. As observed, the sequences generally converge within about 50 iterations, which shows the fast convergent property of the CISCAL-MVC.
Figure 6. Plot of the first 200 values of the sequence { C ( t + 1 ) C ( t ) F } generated by the CISCAL-MVC on 3Sources, ORL, and COIL-20 datasets, respectively. C ( t + 1 ) C ( t ) F represents the absolute difference of two consecutive updates, which measures the convergent behavior of the proposed algorithm. As observed, the sequences generally converge within about 50 iterations, which shows the fast convergent property of the CISCAL-MVC.
Electronics 14 01847 g006
Figure 7. Plot of the first 50 values of the sequence { C ( t + 1 ) C ( ) F / C ( t ) C ( ) F } generated by the CISCAL-MVC on the 3Sources, ORL, and COIL-20 datasets, respectively. C ( t + 1 ) C ( ) F / C ( t ) C ( ) F represents the relative difference of two consecutive updates, which measures the convergent rate of the proposed algorithm. As observed, the values are generally within (0, 1), which empirically suggests that the CISCAL-MVC has a linear convergence rate on these datasets.
Figure 7. Plot of the first 50 values of the sequence { C ( t + 1 ) C ( ) F / C ( t ) C ( ) F } generated by the CISCAL-MVC on the 3Sources, ORL, and COIL-20 datasets, respectively. C ( t + 1 ) C ( ) F / C ( t ) C ( ) F represents the relative difference of two consecutive updates, which measures the convergent rate of the proposed algorithm. As observed, the values are generally within (0, 1), which empirically suggests that the CISCAL-MVC has a linear convergence rate on these datasets.
Electronics 14 01847 g007
Figure 8. Time cost of different methods for comparison. These methods are selected for comparison due to their superior clustering performance as reported in Section 5.3.
Figure 8. Time cost of different methods for comparison. These methods are selected for comparison due to their superior clustering performance as reported in Section 5.3.
Electronics 14 01847 g008
Figure 9. Performance changes of the CISCAL-MVC with respect to different values of the parameters on the ORL dataset. In particular, in each figure, the other parameters are tuned to the best. As can be seen, the CISCAL-MVC is insensitive to α and σ and achieves superior results with a broad selection of small values for γ and β .
Figure 9. Performance changes of the CISCAL-MVC with respect to different values of the parameters on the ORL dataset. In particular, in each figure, the other parameters are tuned to the best. As can be seen, the CISCAL-MVC is insensitive to α and σ and achieves superior results with a broad selection of small values for γ and β .
Electronics 14 01847 g009
Figure 10. Visualization of original and reconstructed features on the Yale dataset, where we use the pixel features X ( 1 ) for illustration. As observed from the figures, the recovered features X ( 1 ) ( C + E ( 1 ) ) well capture the information from the original features. Moreover, the consistent features X ( 1 ) C well capture the key features of the data, where the features from the same cluster have high visual similarities. In contrast, the inconsistent features X ( 1 ) E ( 1 ) are sparse and capture diverse features from the data.
Figure 10. Visualization of original and reconstructed features on the Yale dataset, where we use the pixel features X ( 1 ) for illustration. As observed from the figures, the recovered features X ( 1 ) ( C + E ( 1 ) ) well capture the information from the original features. Moreover, the consistent features X ( 1 ) C well capture the key features of the data, where the features from the same cluster have high visual similarities. In contrast, the inconsistent features X ( 1 ) E ( 1 ) are sparse and capture diverse features from the data.
Electronics 14 01847 g010
Table 1. Overview of benchmark datasets.
Table 1. Overview of benchmark datasets.
DatasetskVnDim. of Each View ( d v )
View 1 View 2 View 3
3Sources63169306836313560
ORL403400409667503304
WebKB-Texas421871871703
COIL-202031440102433046750
Caltech-77314745121984928
Yale153165409633046750
Table 2. Comparison of clustering performance across various methods on benchmark datasets.
Table 2. Comparison of clustering performance across various methods on benchmark datasets.
Dataset3SourcesORLWebKB-Texas
MethodACCNMIARIPURACCNMIARIPURACCNMIARIPUR
SC48.5238.6920.8159.1777.2589.1269.7580.0057.7529.3623.9169.52
MVSpec25.446.61000.8539.6419.5039.421.98021.2532.0910.15−4.64055.08
RMSC62.5458.5648.8776.3370.2384.7961.4274.3148.6426.0118.5168.80
MVGL35.508.070−0.72040.2472.0082.9638.7977.7551.876.5502.48057.22
BMVC65.6858.6954.3273.3716.7539.2900.6118.0056.6824.7325.1868.98
SMVSC31.368.9204.00044.3856.2575.2642.3960.2556.2521.4229.0167.91
MVSC75.8566.4858.3880.8972.3684.6956.5977.0555.131.9800.19056.10
FPMVS-CAG26.046.59042.601.01056.0074.6339.9960.0057.7522.7629.8165.78
JSMC77.5169.5265.1782.2582.0091.4677.7984.5073.3338.0432.2670.05
MvDSCN77.2468.2961.3581.2187.251.0182.9881.0657.6228.1428.1968.86
CISCAL-MVC78.7072.0665.9085.2189.2593.5983.5890.5077.8342.2444.8477.83
DatasetCOIL-20Caltech-7Yale
MethodACCNMIARIPURACCNMIARIPURACCNMIARIPUR
SC72.3680.8066.7974.7246.4038.2430.5582.2963.6464.9445.0662.24
MVSpec15.7624.364.48013.1938.8710.79−1.7355.1621.8225.140.26022.42
RMSC70.8480.4266.1472.5140.3942.1132.0548.1361.2166.3645.6562.36
MVGL81.3993.8078.2186.2566.2855.9841.9985.5364.8567.1541.5364.85
BMVC40.6350.6927.8340.7647.1547.2936.3484.9422.4227.571.77024.24
SMVSC60.5673.4851.1461.6746.6847.7938.2187.3156.9761.6837.5656.97
MVSC61.6675.6851.9564.6963.6854.4647.9282.9159.0364.1040.1560.33
FPMVS-CAG63.7574.6353.2165.4254.5547.3241.3468.7444.2449.7625.3046.67
JSMC83.9692.9881.2888.6870.9063.2060.1092.0673.3376.3857.5873.94
MvDSCN81.0689.2680.7285.9868.4551.2840.3480.8775.4774.2354.6372.51
CISCAL-MVC100.0100.0100.0100.080.6963.6068.7092.5974.2475.8756.29 74.72
The first three results are shown in red, blue, and orange for each dataset, respectively.
Table 3. Comparison results of Wilcoxon signed rank tests on a baseline dataset.
Table 3. Comparison results of Wilcoxon signed rank tests on a baseline dataset.
methodours vs. SCours vs. MVspecours vs. RMSCours vs. MVGLours vs. BMVC
p-value1.82 × 10 5 1.82 × 10 5 1.82 × 10 5 1.82 × 10 5 1.82 × 10 5
methodours vs. SMVSCours vs. MVSCours vs. FPMVS-CAGours vs. JSMCours vs. MvDSCN
p-value1.82 × 10 5 1.82 × 10 5 1.82 × 10 5 4.97 × 10 5 3.88 × 10 5
Table 4. Comparison of the CISCAL-MVC and its variants.
Table 4. Comparison of the CISCAL-MVC and its variants.
DatasetACCNMI
our α = 0 β = 0 γ = 0 our α = 0 β = 0 γ = 0
3sources78.7046.98 +31.7251.36 +27.3462.18 +16.5272.0646.46 +25.6038.24 +33.8257.42 +14.64
ORL89.2582.47 +6.78080.17 +9.0882.48 +6.76593.5992.42 +1.16991.47 +2.11192.31 +1.276
WebKB-Texas74.8363.30 +11.5340.54 +34.2960.39 +14.4438.9024.28 +14.6214.65 +24.2528.47 +10.43
COIL-20100.083.89 +16.1149.76 +50.2484.73 +15.27100.093.93 +6.07067.48 +32.5294.82 +5.177
Caltech-780.6456.69 +23.9555.63 +25.0159.97 +20.6763.6059.88 +3.71644.70 +18.9060.01 +3.585
Yale74.1170.60 +3.50467.57 +6.53573.69 +0.41374.8773.03 +1.83568.95 +5.91874.30 +0.565
Avg.score82.9167.32 +15.5957.50 +25.4170.57 +12.3473.8365.00 +8.83554.24 +19.5967.89 +5.945
DatasetPURARI
our α = 0 β = 0 γ = 0 our α = 0 β = 0 γ = 0
3sources85.2143.33 +41.8843.76 +41.4562.73 +22.4865.9026.29 +39.6131.18 +34.7248.09 +17.81
ORL90.5074.93 +15.5769.52 +20.9873.28 +11.2283.5878.19 +5.39374.31 +9.27577.07 +6.516
WebKB-Texas77.3463.28 +14.0641.62 +35.7261.18 +16.1641.6433.88 +7.775412.56 +29.0830.36 +11.28
COIL-20100.075.07 +24.9334.26 +65.7473.99 +26.01100.081.37 +18.6339.03 +60.9781.72 +18.28
Caltech-792.5983.32 +9.26468.20 +24.3984.39 +8.19358.2841.39 +16.8938.38 +19.9042.53 +15.75
Yale74.7052.07 +22.6344.52 +30.1844.47 +30.2352.2950.37 +1.91244.72 +7.56847.27 +5.012
Avg.score86.7665.33 +21.2750.31 +36.2966.67 +19.9366.9551.91 +15.0440.03 +26.9154.51 +12.44
Bolding is used to emphasize the performance of the overall model. For the variants, the result is reported as x+y, where x denotes the result of the variant and y denotes the improvement made by the overall model, accordingly.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, K.; Kang, K.; Bai, Y.; Peng, C. Commonness and Inconsistency Learning with Structure Constrained Adaptive Loss Minimization for Multi-View Clustering. Electronics 2025, 14, 1847. https://doi.org/10.3390/electronics14091847

AMA Style

Zhang K, Kang K, Bai Y, Peng C. Commonness and Inconsistency Learning with Structure Constrained Adaptive Loss Minimization for Multi-View Clustering. Electronics. 2025; 14(9):1847. https://doi.org/10.3390/electronics14091847

Chicago/Turabian Style

Zhang, Kai, Kehan Kang, Yang Bai, and Chong Peng. 2025. "Commonness and Inconsistency Learning with Structure Constrained Adaptive Loss Minimization for Multi-View Clustering" Electronics 14, no. 9: 1847. https://doi.org/10.3390/electronics14091847

APA Style

Zhang, K., Kang, K., Bai, Y., & Peng, C. (2025). Commonness and Inconsistency Learning with Structure Constrained Adaptive Loss Minimization for Multi-View Clustering. Electronics, 14(9), 1847. https://doi.org/10.3390/electronics14091847

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop