Next Article in Journal
REO: Revisiting Erase Operation for Improving Lifetime and Performance of Modern NAND Flash-Based SSDs
Previous Article in Journal
Privacy-Preserving Hierarchical Top-k Nearest Keyword Search on Graphs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fast Multi-View Subspace Clustering Based on Flexible Anchor Fusion

1
School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi 214122, China
2
Engineering Research Center of Intelligent Technology for Healthcare, Ministry of Education, Wuxi 214122, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(4), 737; https://doi.org/10.3390/electronics14040737
Submission received: 19 January 2025 / Revised: 4 February 2025 / Accepted: 6 February 2025 / Published: 13 February 2025

Abstract

:
Multi-view subspace clustering enhances clustering performance by optimizing and integrating structural information from multiple views. Recently, anchor-based methods have made notable progress in large-scale clustering scenarios by leveraging anchor points to capture data distribution across different views. Although these methods improve efficiency, a common limitation is that they typically select an equal number of anchor points from each view. Additionally, during the graph fusion stage, most existing frameworks use simple linear weighting to construct the final consensus graph, overlooking the inherent structural relationships between the data. To address these issues, we propose a novel and flexible anchor graph fusion framework which selects an appropriate number of anchor points for each view based on its data space, creating suitable anchor graphs. In the graph fusion stage, we introduce a regularization term which adaptively and flexibly combines anchor graphs of varying sizes. Moreover, our approach incorporates both global and local information between views, enabling a more accurate capture of the cluster structure within the data. Furthermore, our method operates with linear time complexity, making it well suited for large-scale datasets. Extensive experiments on multiple datasets demonstrate the superior performance of our proposed algorithm.

1. Introduction

With the rapid advancement of information technology, diverse, high-dimensional data from multiple sources are increasingly being generated in real-world scenarios [1,2]. Such data are often referred to as multi-view data, as they are collected from different perspectives or modalities, including various sensors, feature descriptors, or languages [3,4,5,6,7]. For example, an image can be described by different features like local binary patterns (LBPs), histograms of oriented gradients (HOGs), and Gabor descriptors [3], while a news article might be available in different languages like Chinese, English, and Spanish [8]. Similarly, a video clip can be broken down into visual frames, audio tracks, and textual descriptions, each collected from distinct sources [9,10]. The challenge of effectively leveraging the complementary information contained in these multiple views has driven the development of numerous multi-view clustering (MVC) approaches. Among them, subspace-based clustering techniques have gained significant interest due to their ability to efficiently reveal hidden data structures and uncover the intrinsic relationships among data pairs. Multi-view subspace clustering (MVSC), as a well-known unsupervised approach, has demonstrated great effectiveness in data mining and knowledge discovery, providing robust clustering results by integrating diverse and complementary information from multiple views [11,12,13,14].
Various MVSC methods often adopt the following framework: (1) in each view, learning a low-dimensional representation matrix or similarity matrix through subspace learning; (2) integrating and fusing all view-specific subspace representations to find a shared subspace which reflects the intrinsic structure of the data; and (3) using spectral clustering to derive the final clustering results. For example, Lan et al. [15] addressed feature degeneration by jointly learning consensus and view-specific subspaces, and Ma et al. [16] enhanced clustering quality through automatic neighbor discovery and an improved similarity matrix without hyperparameters.
Despite the improvements in MVSC methods above, these algorithms often suffer from high complexity, requiring significant computational and spatial resources [17]. This high complexity arises from two key steps. (1) In the single-view subspace construction and optimization phase, the learned low-dimensional representation matrix in each view is typically obtained using a self-representation strategy, leading to quadratic space complexity and cubic time complexity with respect to the number of samples. (2) In the fusion and clustering phase, after obtaining the fused subspace representation matrix, the clustering phase involves spectral clustering, which requires singular value decomposition (SVD) and takes a time complexity of O n 3 . For example, Maria et al. [18] proposed a multi-view subspace clustering method which learns joint subspace representation by constructing a shared association matrix. It combines a low rank constraint and sparse constraint to improve the clustering accuracy. Lu et al. [19] proposed a unified subspace clustering method by introducing block diagonal matrix regularization, directly pursuing the block’s diagonal structure to obtain clustering results. Although these methods have achieved good clustering effects, their time complexity of O n 2 or O n 3 hinders the application to large-scale scenarios.
In recent years, anchor graph algorithms have been proposed to accelerate multi-view subspace clustering, aiming to reduce complexity and extend its applicability to large-scale clustering tasks [20,21,22]. Typically, anchors are chosen or sampled independently in each view, using methods such as k-means clustering, random sampling, or DPP sampling. By independently sampling the anchor points for each view, the global matrix of d × n can be transformed into the anchor points matrix of d × m , where m represents the amount of selected anchor points and m < < n . After the single-view subspace construction phase, a weighted linear combination of the anchor graphs or similarity matrices from each view is typically formed to obtain a fused graph. Due to their efficiency, many MVSC methods based on anchor point strategies have been proposed. For example, Wang et al. [23] integrated anchor selection and graph construction to enhance efficiency and clustering quality. Zhou et al. [1] effectively extracted representative anchor points by combining the eigenvalues of each sample, which helps avoid instability in methods like k-means clustering, and then the clustering results can be obtained directly based on the connectivity of the fused graph, eliminating the instability of clustering results. Moreover, the time complexity and space complexity can be regarded as O n , which is suitable for large-scale clustering tasks.
However, anchor-based multi-view subspace clustering methods can still be further improved. First, many methods separate the anchor graph construction and graph fusion processes into two independent steps, which may result in the loss of consensus information during the final clustering, thereby reducing clustering performance. Secondly, many anchor-based methods focus on capturing the global structure of the data, specifically the relationships between the samples and all anchors. This often leads to a dense correspondence matrix. However, neglecting local structure information may result in the loss of complementary data between samples in the final fused consensus graph. Thirdly, current anchor-based methods require selection of the same number of anchors for each view, which is unrealistic due to the differing data distributions and representations across views. This discrepancy makes it challenging to determine an equal number of anchors which effectively represent the underlying data structure in all views. Furthermore, traditional weighted linear combination methods face difficulties in fusing anchor graphs of varying sizes. These limitations undermine the effectiveness and applicability of anchor point mechanisms in large-scale multi-view subspace clustering tasks.
To enhance the effectiveness and flexibility of anchor-based MVSC methods, we propose a novel framework in Figure 1 called FMVSC, based on the flexible fusion of anchor graphs. Specifically, to address the issues mentioned above, we first integrate construction and optimization of the anchor matrix and anchor graph into a unified framework, which is iteratively updated based on the consensus and complementary information. This approach allows the components to better align with the underlying data space. Additionally, we not only focus on capturing the global structure of the data but also explore the local structure. Integrating both the global and local structures helps enhance the final clustering performance. More importantly, unlike traditional anchor-based methods, which rely on an equal number of anchors and a simple linear combination, we carefully examine existing anchor-based frameworks and introduce structure regularization for anchor graph fusion, developing a new structural alignment framework. From the perspective of anchor point constraints, traditional frameworks only select an equal number of anchors for each view, whereas our framework can learn anchor graphs which align with the data space of individual views, integrating consensus and complementary information between individual views. From the perspective of fusion strategies, simple linear combination strategies can only fuse anchor graphs of the same size, whereas our framework more flexibly fuses anchor graphs of different sizes, aggregating more comprehensive information.
In addition, the time and space complexity of this method is linear, making it well suited for large-scale data applications. Comprehensive experiments on multiple benchmark datasets demonstrate that FMVSC outperforms current state-of-the-art methods in terms of both clustering performance and computational efficiency.
Overall, the primary contributions of this paper can be summarized in three key points:
(1)
Unlike existing frameworks, our proposed framework can automatically learn and update anchor points. We integrate the construction and optimization of the anchor matrix and anchor graph into a unified framework. By iteratively updating these components using consensus and complementary information, they can more accurately capture the underlying data structure.
(2)
We focus not only on capturing the global structure of the data but also exploring its local structure. Integrating both global and local structures helps to enhance the final clustering performance. Based on the global and local structures, we propose an alternative optimization method with linear complexity relative to the number of examples, and the experimental results demonstrate the effectiveness and efficiency of FMVSC.
(3)
Instead of using the same number of anchors, we believe that each view has its own unique underlying data space, and the number of anchors should be selected to match the characteristics of each view. Therefore, we propose a new framework which can learn anchor graphs that align with the data space of individual views and fuse anchor graphs of different sizes.
The remainder of this paper is organized as follows. Section 2 reviews the related work on multi-view subspace clustering and presents our proposed framework. Section 3 introduces the iterative optimization algorithm and discusses the time and space complexity, as well as the convergence issues of FMVSC. Section 4 presents the experimental results on benchmark datasets, comparing FMVSC with eight state-of-the-art multi-view clustering methods. Section 5 concludes the paper.

2. Related Work and Proposed Approach

2.1. Single-View Subspace Clustering

Subspace clustering assumes that high-dimensional data are distributed across several low-dimensional subspaces, with each class corresponding to a different subspace. Each data point can be represented by a linear combination of other points within the same subspace. For high-dimensional data, subspace clustering is an important unsupervised learning method. Based on the self-expression property of subspace clustering, many algorithms have been proposed to explore the underlying relationships among high-dimensional data. For example, in low-rank representation (LRR) [24], data recovery is achieved through low-rank representation. Sparse subspace clustering (SSC) [25] leverages the idea that sparse representation in high-dimensional subspaces better aligns with the structural model of the actual subspace. In a recent work, Nie et al. [26] improved subspace clustering by initializing the dictionary with partial samples, avoiding trivial solutions. The framework employs a bipartite graph to utilize the duality between samples and features, capturing subspace structures more effectively. Moreover, Nie introduced rank constraints and spare constraints to the framework, making it more effective and robust. The common single-view subspace clustering equation is shown in Equation (1) [27]:
min Z X X Z F 2 + α Z F 2
s . t .   Z 0 ,   Z 1 = 1
where Z R n × n denotes the nonnegative coefficient matrix. The second term is a regularization term, aiming to prevent trivial solutions.

2.2. Multi-View Subspace Clustering

Multi-view subspace clustering is the extension of single-view subspace clustering, which mainly improves the clustering accuracy by fully exploring the complementary and consensus information between views. Specifically, the MVSC approach typically adopts the following framework: Learn the low-dimensional representation matrix for each view, then learn a fused embedding matrix with consensus across all views. Finally, apply spectral clustering to obtain the final clustering result.
The common MVSC equation is shown in Equation (2) [28]:
min Z p p = 1 v X p X p Z p F 2 + λ f C , Z p
s . t .   Z p 0 ,   Z p T 1 = 1
where f · represents the regularization function,   Z p R n × n denotes the nonnegative self-representation matrix of each view, and C represents the spectral embedding matrix of all views.
In some recent studies, efforts have been made to enhance the precision of subspace clustering. For example, Shi et al. [29] enhanced latent multi-view subspace clustering by constructing an augmented data matrix with block-diagonal and non-diagonal entries, capturing complementary and consistent information. Chen et al. [13] proposed a multimodal subspace clustering algorithm based on spatiotemporal attention to improve clustering performance by applying cross-modal learning and attention mechanisms. But these methods usually have high time complexity, which limits the application of MVSC methods on large-scale datasets.

2.3. Anchor-Based Multi-View Subspace Clustering

Recently, in order to alleviate the high time complexity of traditional MVSC methods and make better use of the potential information between views, many scholars have begun to conduct multi-view subspace clustering research based on anchor points. For instance, Kang et al. [21] first applied the anchor point strategy to large-scale subspace clustering, which achieves linear time complexity for large datasets and is also suitable for single-view scenarios. Zhao et al. [30] combined bipartite graphs, rank constraints, and the anchor point strategy to propose a scalable structured graph learning framework, while also systematically addressing the out-of-sample problem for the first time. Li et al. [31] proposed a parameter-free multi-view subspace clustering framework based on bipartite graphs and anchor graphs. It obtains a fused graph through self-supervised weighting, and due to the applied connectivity constraints, the clustering results can be directly obtained without postprocessing. Yang et al. [32] combined the anchor point strategy and spectral embedding decomposition to propose a robust and efficient framework. It improves clustering performance by constructing an anchor graph, which approximates the sample graph, and then decomposing the anchor graph. Lao et al. [33] generated diversified anchor point sets for each view and learns a bipartite graph for each view based on them. These bipartite graphs are then learned and segmented, ultimately producing a unified bipartite graph for clustering. This method also has linear time complexity.
Although the mentioned MVSC methods have achieved reasonable results, they all use an equal number of anchor points for graph construction in each view, which is reasonable, as each view has its own unique data space.

2.4. Spectral Clustering

Spectral clustering (SC) is widely used in the field of unsupervised learning [25]. Traditional spectral clustering methods are relatively simple and can easily obtain clustering results by performing k-means clustering on the low-dimensional embedding matrix of the data matrix. For a similarity graph S R n × n , the traditional SC equation can be represented as shown in Equation (3) [34]:
max F T F = I T r F S F T
where F R k × n represents the low-dimensional spectral embedding matrix.

2.5. Proposed Approach

As mentioned above, most of the anchor-based multi-view clustering methods use k means clustering or random sampling to select the same number of anchor points in all views, carry out anchor graph learning, and finally obtain the consistent indicator matrix for clustering through simple weighted linear combination. Therefore, there is an underlying assumption in all these methods that an equal number of anchors can represent the latent data structure of each view in multi-view data. This is obviously irrational, as multi-view data are distributed and represented differently in each view, making it difficult to determine an equal number of anchor points to represent the latent data structure for all views. Moreover, facing the unique anchor graph built for each view, the traditional weighted linear combination method cannot fuse multiple anchor graphs of different sizes.
To solve these problems, we introduce a structural regularization term for fusing anchor graphs of different sizes in this paper. First of all, when reviewing Equation (3) in spectral clustering, we can notice that Equation (3) is equivalent to Equation (4) below:
min F T F = I S F T F F 2
In Equation (5), we provide the related proof:
  E q u a t i o n   4 = min F T r S S T 2 F S F T + F T F F F T = max F T r F S F T = E q u a t i o n   3
Assume that all views share the same underlying clustering structure, which means all views share a uniform F , even if they do not have the same anchor graph size. Noting that S in Equation (4) can be represented by Z T Z [21], in multi-view clustering methods, Equation (4) can be transformed into Equation (6):
m i n p = 1 v Z p T Z p F T F F 2    
In addition, in order to reduce the quadratic programming problems and high time complexity which may occur in the future, and to construct Z p such that it is more suitable for each data space, we introduce a rotation matrix R p R k × l p satisfying R p T R p = I l p . Then, Equation (6) can be rewritten as Equation (7) such that it is equivalent to Equation (7):
m i n p = 1 v Z p T R p T R p Z p F T F F 2 s . t .     Z p 0 ,   Z p T 1 = 1 ,   R p T R p = I l p ,   F F T = I k
In order to further simplify the subsequent iterative update process and reduce the time complexity, we replace Equation (7) with Equation (8):
m i n p = 1 v R p Z p F F 2 s . t .     Z p 0 ,   Z p T 1 = 1 ,   R p T R p = I l p ,   F F T = I k
Let C = R p Z p F . Then, we have proof in Equation (9) to make sure that the minimum value of Equation (7) can be obtained approximately by minimizing Equation (8), and F F = k 1 2 is a constant, with k being the number of classes:
                            Z p T R p T R p Z p F T F F = C + F T C + F F T F F = C T C + F T C + C T F F C T C F + 2 C T F F C F C F + 2 F F = C F ( C F + 2 k 1 2 )
The proof steps take advantage of the compatibility and subadditivity of the matrix norm. Then, we can minimize Equation (7) to approximate the upper bound of Equation (8).
Equation (8) is the final structure fusion regularization term in this paper. Compared with the existing structure fusion regularization terms for anchor graphs, our term can fuse different sizes of anchor graphs more flexibly, which makes our new framework more scalable to select the approximate number of anchor points for each view, and build the suitable anchor graph for latent data space in each view. In addition, our regularization term is simplified from Equation (7). Thus, it is more efficient and adaptable to large datasets than ordinary regularization terms.
Next, we will introduce the anchor learning term in our framework. As mentioned, heuristic sampling with fixed anchor points is not a good choice for complex multi-view data. Therefore, we added anchor points into the iterative update process so that the anchor points of each view could be adjusted adaptively as other parameters changed. Then, the anchor points can be dynamically updated and optimized using Equation (10):
min A p , Z p p = 1 v X p A p Z p F 2 s . t .     Z p 0 ,   Z p T 1 = 1 ,   A p T A p = I l p
where A p R d p × l p denotes the pth anchor point graph with l p anchors in the pth view and Z p R l p × n denotes the pth anchor graph. To make the learned anchor points more diverse and simplify the subsequent update process, we applied orthogonal constraints to the anchor points such that A p T A p = I l p . Equation (10) effectively learns the complementary and consensus information among multi-view data and explores the global information between the whole views well.
Furthermore, preserving local manifold structures also plays an important role in cluster analysis. Thus, we introduced Equation (11) to learn the local information between each view data:
min A p , Z p p = 1 v t r ( A p d i a g Z p 1 A p T ) s . t .     Z p 0 ,   Z p T 1 = 1 ,   A p T A p = I l p
where d i a g Z p 1 R l p × l p is a diagonal matrix. Equation (11) can be derived mathematically in a variety of ways [35,36]. With the local information regularization term, our final anchor graph learning regularization term becomes
min A p , Z p p = 1 v X p A p Z p F 2 + p = 1 v t r A p d i a g Z p 1 A p T s . t .     Z p 0 ,   Z p T 1 = 1 ,   A p T A p = I l p
Then, we added the structure fusion regularization term in Equation (8) to implement our framework. In general, we designed a scalable multi-view anchor graph clustering method based on structure fusion (FMVSC) as follows in Equation (13),
    min β , A p , Z p , F , R p p = 1 v β p 2 X p A p Z p F 2 g o b a l   i n f o r m a t i o n + t r A p d i a g Z p 1 A p T l o c a l   i n f o r m a t i o n a n c h o r   l e a r n i n g + λ R p Z p F F 2 s t r u c t u r e   f u s i o n s . t .     β T 1 = 1 ,   Z p 0 ,   Z p T 1 = 1 ,   A p T A p = I l p ,   R p T R p = I l p ,   F F T = I k
where A p R d p × l p denotes the pth anchor matrix, Z p R l p × n denotes the pth anchor graph, β R p × 1 is the weight factor, λ is the artificially set parameter, R p R k × l p denotes the pth rotation matrix for structure fusion, and F R k × n is the fused spectral embedding matrix. After finding the final F , we inserted it into k-means clustering to obtain clustering results. Additionally, we presented the mathematical symbols used in FMVSC along with their corresponding definitions in Table 1 below.

3. Optimization

When considering all variables together, the optimization problem in Equation (13) becomes a nonconvex problem. We developed an alternate iterative update method to break the problem into five sub-problems (i.e., we would fix four variables and solve for the other variable) [7].
For updating the anchor matrices { A p } p = 1 v , noting that { A p } p = 1 v are independent of each other, when the other four variables are fixed, Equation (13) can be broken into v sub-problems, and the formula for optimizing A p is as follows:
min A p   β p 2 X p A p Z p F 2 + t r A p d i a g Z p 1 A p T s . t .     A p T A p = I l p
We can convert Equation (14) to trace the form and remove the extraneous terms:
min A p   t r ( 2 β p 2 A p Z p X p T + A p ( β p 2 Z p Z p T + d i a g ( Z p 1 ) ) A p T ) s . t .     A p T A p = I l p
By taking the derivative, we can update A p with Equation (16):
A p = β p 2 X p Z p T [ β p 2 Z p Z p T + d i a g ( Z p 1 ) ] 1
For updating anchor graphs { Z p } p = 1 v , noting that { Z p } p = 1 v are independent of each other, with the other variables fixed, we can transform Equation (13) into its trace form and update Z p in turn by solving Equation (17):
min Z p   t r [ β p 2 + λ Z p T Z p Z p T 2 β p 2 A p T X p + 2 λ R p T F + A p d i a g Z p 1 A p T ] s . t .     Z p 0 ,   Z p T 1 = 1
In order to facilitate subsequent optimization, we rewrite t r [ A p d i a g Z p 1 A p T ] in Equation (18):
t r [ A p d i a g Z p 1 A p T ] = t r A p T A p d i a g Z p 1 = t r Z p T M p
where all elements of the ith row in M p are A p : , i T A p : , i and A p : , i denotes the ith column in A p . Then, Equation (17) can be written as Equation (19):
min Z p   t r [ β p 2 + λ Z p T Z p Z p T 2 β p 2 A p T X p + 2 λ R p T F M p ] s . t .     Z p 0 ,   Z p T 1 = 1
Noticing that ( 2 β p 2 A p T X p + 2 λ R p T H M p ) is a constant when other variables are fixed, Equation (19) can be rewritten as Equation (20):
min Z p Z p 1 β p 2 + λ ( β p 2 A p T X p + λ R p T F 1 2 M p ) F 2 s . t .     Z p 0 ,   Z p T 1 = 1
Let us set Z p = Y and 1 β p 2 + λ β p 2 A p T X p + λ R p T F 1 2 M p = K . Then, Equation (20) can be transformed into n optimization problems in Equation (21) with the form of
min Y : , j Y : , j K : , j F 2 s . t .     Y : , j 0 ,   Y : , j T 1 = 1
The Lagrangian function of Equation (21) is as follows in Equation (22):
L Y : , j , η , ζ j = Y : , j K : , j F 2 η Y : , j T 1 1 ζ j T Y : , j
where η and ζ j 0 are both Lagrangian multipliers. Based on the KKT condition, we can find the optimal solution for Y : , j in Equation (23):
Y i , j = max K i , j + η , 0
Obviously, the elements which are zeros of the optimal solution Y : , j correspond to the smaller component of K : , j , and to simplify the representation, we assume that the elements of the vector K : , j are ordered from largest to smallest, specifically K 1 , j K 2 , j K ρ , j K ρ + 1 , j K D , j , with D as the dimension of K : , j . Then, according to Equation (23) and the constraint Y : , j T 1 = 1 , we have Equation (24):
i = 1 ρ K i , j + η = 1 η = 1 ρ ( 1 i = 1 ρ K i , j )
where ρ = m a x { 1 i D :   K i , j + 1 i ( 1 i = 1 D K i , j ) > 0 } according to the solution given by Wang [37]. Then, we can update Z p column by column based on Equations (23) and (24).
For updating the rotation matrices { R p } p = 1 v , when fixing other variables, R p in the pth view can be updated by solving Equation (25):
max R p   t r R p T G s . t .     R p T R p = I l p
where G p = F Z p T . Suppose that G p can be decomposed by singular values into G p = S Σ V T . According to Wang [38], the optimal solution of R p is R p = S V T with S to be the left singular matrix and V to be the right matrix of G p .
By updating R p independently, in turn, we can optimize the individual rotation matrices in each view.
For updating the spectral embedding matrix F , F can be optimized by solving Equation (26):
max R p   t r F T Q s . t .    F T F = I k
where Q = R p Z p . Similar to the process of updating R p , the optimal solution of F is F = S V T , with S being the left singular matrix and V being the right matrix of Q .
For updating the weight parameter β , while fixing other variables, the optimization problem becomes Equation (27):
min β p p = 1 v β p 2 r p 2 s . t .     β T 1 = 1 , α 0
where r p = β p 2 X p A p Z p F , and according to the Cauchy inequality, Equation (27) can be transformed into Equation (28):
p = 1 v β p 2 r p 2 × p = 1 v 1 2 ( p = 1 v β p r p ) 2 s . t .     β T 1 = 1 ,   β 0
Equation (28) is equivalent when β p = r p p = 1 v 1 r p , and then Equation (27) takes the minimum value.
The entire process to solve Equation (13) is organized in Algorithm 1.
Algorithm 1 FMVSC
Input: Multi-view dataset { X p } p = 1 v , cluster number k , parameter λ
Initialization: Initialize A p ,   Z p ,   R p ,   F ,   β
1: Repeat
2:   Update A p in Equation (16)
3:   Update Z p in Equation (23)
4:   Update R p in Equation (25)
5:   Update F in Equation (26)
6:   Update β in Equation (28)
7: Until converge
8: Obtain the spectral embedding matrix F
Output: Perform k-means clustering on F to find the final results.
At the end of this section, we give our analysis of the time complexity, space complexity, and convergence.
The time complexity of our method consists of five update steps. In the process of updating the anchor matrices { A p } p = 1 v , they cost O ( p = 1 v ( l p 3 + d p l p 2 + d p l p n ) ) to perform matrix inversion and matrix multiplication. Then, the anchor graphs { Z p } p = 1 v cost O ( p = 1 v ( l p + k l p + d p l p + l p 2 ) n ) , and the rotation matrices { R p } p = 1 v take O ( p = 1 v ( k l p n + k 2 l p ) ) to perform SVD and matrix multiplication. Similar to { R p } p = 1 v , the spectral embedding matrix F costs O ( p = 1 v ( k l p n + k 2 n ) . As for the weight parameter B, it needs O ( p = 1 v d p l p n ) to update. On the whole, with t being the iteration number of the whole optimization process, our method takes O ( p = 1 v n t ( 3 d p l p + l p + 3 k l p + k 2 + l p 2 ) + p = 1 v t ( l p 3 + d p l p 2 + k 2 l p ) ) to perform the optimization process.
During the whole optimization process, we need O ( p = 1 v ( ( d p + n ) l p + k l p ) + n k ) of space to store the variables { A p } p = 1 v , { Z p } p = 1 v , { R p } p = 1 v and F .
The above analysis shows that the time and space complexity of our algorithm presents a linear relationship with the number of samples n. This shows that this algorithm has considerable advantages in terms of time and space and can be effectively applied to large-scale datasets.
In Equation (13), the objective function of our algorithm decreases monotonically in turn with iteration. At the same time, the lower bound of eq is zero. According to [39], our objective function can converge to the local minimum.

4. Experiments

In this section, we apply FMVSC to four artificial datasets and eight real-world datasets. The experimental results show that the FMVSC clustering algorithm performed well on most of the datasets. All of our experiments were run on a computer sourced in Wuxi, China with an i7-7700HQ CPU from Intel and 8 GB of RAM.

4.1. Dataset Descriptions

4.1.1. Artificial Datasets

We selected four manually generated datasets to test the clustering performance of FMVSC on artificial datasets:
(a)
Two-moon dataset: This dataset was generated randomly, where two sets of points were positioned next to each other in a moon shape and each set of points contained 500 random points.
(b)
Noisy two-moon dataset: This dataset was generated by adding 200 randomly generated noise points to the two-moon dataset.
(c)
Norm5 dataset: This randomly generated dataset consisted of five clusters, each containing 200 samples, distributed across five spherical regions.
(d)
Noisy norm5 dataset: This dataset was generated by adding 200 randomly generated noise points to the norm5 dataset.

4.1.2. Real-World Datasets

Notting-Hill [40]: This dataset contains 550 facial images in 5 categories, with three views and 110 images in each category.
WebKB [41]: This dataset consists of webpages which are described using contents and links. It has 1051 samples, two classes, and two views.
Wiki [3]: This dataset consists of 2866 images with 10 classes and two views.
CCV [17]: This dataset contains 20 categories extracted from YouTube videos, with 6773 samples and three views.
ALOI [3]: This dataset includes 10,800 images with 100 classes and four views.
Animal [1]: This dataset includes 50 animal species with 11,673 animal images, and it has four features representing four different views.
NUSWIDE [1]: This dataset contains 30,000 images of 31 categories with five views.
YoutubeFace [28]: The face video dataset contains 101,499 images in 31 categories.
More detailed information can be seen in Table 2.

4.2. Compared Methods

To present the performance of the FMVSC method more intuitively, we compared it with eight state-of-the-art methods, which are as follows:
MLRSSC (2018) [18]: This model constructs an affinity matrix with low-rank and sparse constraints.
MSC_IAS (2019) [42]: This model attempts to construct the affinity matrix using the Hilbert–Schmidt independence criterion.
LMVSC (2020) [21]: This model adopts the anchor point strategy to handle large datasets in linear time.
MSGL (2022) [30]: This model attempts to construct an affinity matrix using consensus anchor points and connectivity constraints.
SMVSC (2022) [17]: This model combines anchor graph learning with subspace construction to obtain a clustering indicator matrix.
SFMC (2022) [31]: This model is designed to construct a joint graph in a parameter-free, self-supervised weighted manner.
RAMCSF (2023) [32]: This model attempts to approximate the full sample graph through spectral embedding decomposition.
OMVCDR (2024) [43]: This model addresses large-scale clustering problems through matrix factorization and projection strategies.

4.3. Results on Artificial Datasets

In this section, we explore the performance of FMVSC on artificial datasets. We applied FMVSC and three baseline algorithms to four artificial datasets and visually presented the clustering results in the form of images in Figure 2, Figure 3, Figure 4 and Figure 5, where each color represents a cluster and black points represent noise. It can be observed that MLRSSC failed to correctly distinguish the five clusters in the Norm5 dataset, indicating that MLRSSC performs poorly in multi-cluster scenarios. Additionally, MLRSSC failed to obtain correct clustering results for the Noisy two-moon and Noisy norm5 datasets, demonstrating its lack of robustness and resistance to noise. LMVSC performed poorly on the Two-moon, Noisy two-moon, and Noisy norm datasets, suggesting that LMVSC loses some critical information during the rapid clustering process, leading to poor performance on nonlinear and noisy data. MSGL performed well on the first three datasets but failed to correctly distinguish the Noisy norm dataset, indicating that MSGL lacks robustness and is easily affected by noise. Conversely, FMVSC achieved correct clustering results across all four datasets, demonstrating its strong adaptability to various datasets and superior robustness.

4.4. Experimental Results

We selected four commonly used cluster evaluation metrics, which are as follows:
Accuracy (ACC): ACC measures the proportion of instances which are correctly classified.
Normalized mutual information (NMI): NMI evaluates the degree of similarity between the true and predicted labels.
Purity: Purity assesses how homogenous each cluster is, based on the dominance of a single class within the cluster.
F score: The F score is the harmonic mean of the precision and recall, balancing both metrics.
We conducted extensive experiments on the aforementioned dataset, comparing FMVSC with eight baseline algorithms. The detailed experimental results are shown in Table 3, Table 4, Table 5 and Table 6. The best performance is highlighted in bold black font, while the second-best performance is underlined. “OM” indicates cases where memory limitations occurred. To mitigate the instability caused by randomized methods such as k-means clustering, we selected the optimal parameter settings and repeated the experiments 20 times, reporting the average results and standard deviations. In the experiments, we employed a grid search strategy, where the number of anchor points for each view was explored within the set [k, 3k, 5k] and the hyperparameter λ was explored within the set [1, 10, 100, 1000]. Based on the results in Table 3, Table 4, Table 5 and Table 6, the advantages of our method are primarily reflected in the following aspects:
(a)
Compared with other baseline algorithms, FMVSC demonstrated significant superiority in performance. Especially in terms of the ACC metric, FMVSC achieved the optimal value across all cases. Compared with the second-best algorithm, FMVSC improved the ACC scores by 0.44%, 1.4%, 4.9%, 3.1%, 2.5%, 2.1%, 2.0%, and 3.8%, respectively. In the other three metrics, FMVSC also achieved either the optimal or second-best results. This highlights the excellent clustering performance of FMVSC.
(b)
Compared with classic MVSC algorithms such as MLRSSC and MSC_IAS, FMVSC reduced both the time and space complexity. By leveraging a flexible anchor point sampling strategy, it is better suited for large-scale datasets.
(c)
FMVSC also achieved impressive results on smaller-scale datasets. When compared with fast MVSC methods such as LMVSC, MSGL, SMVSC, RAMCSF, and OMVCDR, FMVSC demonstrated clear performance advantages. This indicates that the flexible anchor graphs constructed for each view effectively captured the data space.
(d)
Compared with fast MVSC algorithms such as LMVSC, MSGL, SMVSC, and RAMCSF, the smaller standard deviation of FMVSC’s experimental results indicates that the variation in the results was low, suggesting that FMVSC is stable. Additionally, SFMC and OMVCDR do not rely on randomized methods like k-means clustering during the clustering process, which is why their experimental results had a standard deviation of zero.
Table 3. Accuracy comparison results (mean ± Std).
Table 3. Accuracy comparison results (mean ± Std).
DatasetMLRSSCMSC_IASLMVSCMSGLSMVSCSFMCRAMCSFOMVCDROurs
Year201820192020202220222022202320242024
Notting-Hill80.61 ± 4.3177.24 ± 5.4677.82 ± 6.4873.41 ± 5.7775.45 ± 6.3384.91 ± 0.0086.64 ± 7.1883.57 ± 0.0088.38 ± 4.29
WebKB90.77 ± 0.2277.07 ± 1.4868.34 ± 0.1393.03 ± 0.0090.87 ± 0.0094.76 ± 0.0094.19 ± 0.0095.37 ± 0.0095.81 ± 0.00
Wiki15.67 ± 0.4523.97 ± 1.2649.83 ± 2.3351.16 ± 1.9750.03 ± 4.1733.27 ± 0.0043.94 ± 2.9446.71 ± 0.0056.04 ± 2.14
CCV15.87 ± 0.3114.10 ± 0.2920.12 ± 0.5815.44 ± 0.4119.78 ± 0.4611.13 ± 0.0014.16 ± 0.8719.74 ± 0.0022.87 ± 0.34
ALOIOMOM42.18 ± 1.7116.93 ± 0.7452.36 ± 2.24OM55.75 ± 2.3763.84 ± 0.0066.36 ± 2.11
AnimalOMOM10.63 ± 0.1911.09 ± 0.2113.16 ± 0.22OM15.67 ± 0.4214.23 ± 0.0017.77 ± 0.33
NUSWIDEOMOM12.27 ± 0.2713.60 ± 0.6712.13 ± 0.37OM14.08 ± 0.5313.53 ± 0.0016.02 ± 0.39
YoutubeFaceOMOM14.25 ± 0.5316.71 ± 0.57OMOM17.64 ± 0.85OM21.39 ± 0.61
The best performance is highlighted in bold black font, while the second-best performance is underlined.
Table 4. Normalized mutual information comparison results (mean ± Std).
Table 4. Normalized mutual information comparison results (mean ± Std).
DatasetMLRSSCMSC_IASLMVSCMSGLSMVSCSFMCRAMCSFOMVCDROurs
Year201820192020202220222022202320242024
Notting-Hill67.33 ± 4.0971.93 ± 3.5266.33 ± 4.0459.14 ± 5.8172.04 ± 4.8581.62 ± 0.0077.23 ± 6.4776.15 ± 0.0079.86 ± 3.61
WebKB47.27 ± 0.3119.09 ± 0.1213.24 ± 0.0858.14 ± 0.0055.19 ± 0.0062.74 ± 0.0060.81 ± 0.0064.29 ± 0.0073.25 ± 0.00
Wiki12.31 ± 0.6818.65 ± 1.0345.39 ± 2.0748.63 ± 1.8346.38 ± 3.2531.29 ± 0.0037.62 ± 2.0341.94 ± 0.0052.40 ± 1.50
CCV11.98 ± 0.409.50 ± 0.3516.48 ± 0.4711.89 ± 0.5314.82 ± 0.613.13 ± 0.007.62 ± 0.9517.58 ± 0.0019.27 ± 0.41
ALOIOMOM51.23 ± 1.9322.54 ± 0.8161.47 ± 1.89OM64.73 ± 2.3972.70 ± 0.0080.65 ± 1.76
AnimalOMOM6.95 ± 0.176.47 ± 0.2712.63 ± 0.24OM14.21 ± 0.3213.29 ± 0.0014.07 ± 0.24
NUSWIDEOMOM8.13 ± 0.331.86 ± 0.606.39 ± 0.29OM9.27 ± 0.3813.01 ± 0.0015.10 ± 0.28
YoutubeFaceOMOM5.20 ± 0.373.24 ± 0.29OMOM15.19 ± 0.41OM17.3 ± 0.43
The best performance is highlighted in bold black font, while the second-best performance is underlined.
Table 5. Purity comparison results (mean ± Std).
Table 5. Purity comparison results (mean ± Std).
DatasetMLRSSCMSC_IASLMVSCMSGLSMVSCSFMCRAMCSFOMVCDROurs
Year201820192020202220222022202320242024
Notting-Hill81.86+3.8483.30 ± 4.0980.59 ± 6.2979.82 ± 5.3383.41 ± 5.6185.45 ± 0.0086.91 ± 6.6389.64 ± 0.0089.14 ± 4.17
WebKB90.77 ± 0.2278.12 ± 0.1968.34 ± 0.1393.03 ± 0.0090.87 ± 0.0094.77 ± 0.0094.20 ± 0.0095.37 ± 0.0095.81 ± 0.00
Wiki15.37 ± 0.8323.68 ± 1.2754.31 ± 2.4058.33 ± 2.0451.79 ± 4.5934.18 ± 0.0042.36 ± 2.5847.34 ± 0.0061.36 ± 2.55
CCV20.10 ± 0.3218.25 ± 0.2623.74 ± 0.4320.04 ± 0.3722.25 ± 0.4911.68 ± 0.0016.31 ± 0.8921.36 ± 0.0024.87 ± 0.33
ALOIOMOM42.32 ± 1.6422.15 ± 0.7749.45 ± 2.31OM59.86 ± 1.9264.77 ± 0.0068.63 ± 1.81
AnimalOMOM18.57 ± 0.218.84 ± 0.2614.21 ± 0.21OM15.17 ± 0.4422.06 ± 0.0021.76 ± 0.30
NUSWIDEOMOM24.32 ± 0.3125.18 ± 0.5123.31 ± 0.44OM12.53 ± 0.3223.57 ± 0.0027.75 ± 0.34
YoutubeFaceOMOM27.32 ± 0.2532.79 ± 0.33OMOM28.64 ± 0.55OM31.60 ± 0.38
The best performance is highlighted in bold black font, while the second-best performance is underlined.
Table 6. F score comparison results (mean ± Std).
Table 6. F score comparison results (mean ± Std).
DatasetMLRSSCMSC_IASLMVSCMSGLSMVSCSFMCRAMCSFOMVCDROurs
Year201820192020202220222022202320242024
Notting-Hill71.89 ± 3.7775.61 ± 4.3473.24 ± 5.5777.82 ± 4.1465.41 ± 5.5972.45 ± 0.0078.91 ± 7.0982.64 ± 0.0083.74 ± 3.95
WebKB86.94 ± 0.1778.39 ± 0.1562.80 ± 0.1189.89 ± 0.0086.58 ± 0.0092.52 ± 0.0091.95 ± 0.0093.56 ± 0.0093.70 ± 0.00
Wiki17.46 ± 0.4725.74 ± 0.7443.75 ± 2.0839.77 ± 1.9137.64 ± 3.4121.38 ± 0.0029.72 ± 2.2431.54 ± 0.0049.87 ± 2.24
CCV9.85 ± 0.429.22 ± 0.1211.70 ± 0.319.76 ± 0.4611.93 ± 0.5310.85 ± 0.009.51 ± 0.9311.78 ± 0.0013.28 ± 0.27
ALOIOMOM38.43 ± 1.5210.95 ± 0.7441.67 ± 1.73OM47.82 ± 1.9753.76 ± 0.0057.43 ± 1.58
AnimalOMOM9.54 ± 0.0811.18 ± 0.0710.37 ± 0.15OM10.97 ± 0.249.27 ± 0.0010.83 ± 0.16
NUSWIDEOMOM8.84 ± 0.0710.47 ± 0.4311.39 ± 0.39OM9.33 ± 0.4610.52 ± 0.0011.67 ± 0.22
YoutubeFaceOMOM10.72 ± 0.1913.81 ± 0.21OMOM11.31 ± 0.87OM14.13 ± 0.35
The best performance is highlighted in bold black font, while the second-best performance is underlined.
In addition, we applied the t-distributed stochastic neighbor embedding (t-SNE) algorithm to the learned consistent spectral embedding matrices. A commonly used dimensionality reduction technique, t-SNE is primarily used for visualizing high-dimensional data. It works by embedding high-dimensional data into a lower-dimensional space for visualization while preserving the relative distances between data points as much as possible. We selected the WebKB and Notting-Hill datasets for visualization. LMVSC was chosen as the comparison algorithm.
From Figure 6, it can be seen that LMVSC performed poorly on the WebKB dataset, as the distribution of the red and cyan points in its t-SNE plot appears disorganized, with the two clusters not well separated. In contrast, FMVSC performed well, as the red and cyan points are clearly clustered into two distinct regions with a noticeable gap between them, indicating that FMVSC effectively separated the two clusters. Additionally, the red points represent a cluster with 821 samples, while the cyan points represent a cluster with 230 samples. Thus, the red region is noticeably larger than the cyan region.
From Figure 7, it can be seen that LMVSC did not separate the red and yellow points well, indicating that LMVSC performed poorly in clustering these two clusters. In contrast, FMVSC successfully grouped the five different colored points into their respective regions, demonstrating better clustering performance. Additionally, since each of the five clusters in the Notting-Hill dataset contained 110 sample points, the five different colored regions in FMVSC were almost the same size.
These experimental results demonstrate that FMVSC can flexibly construct anchor graphs of varying sizes and integrate consistent spectral embedding matrices which effectively capture the structured information of the data.
Overall, the FMVSC algorithm proposed in this paper showed significant advantages in its clustering results.

4.5. Runtime Comparison

In this section, we recorded the runtimes of each algorithm on the datasets to evaluate the efficiency of FMVSC. Based on Table 7, we can draw the following conclusions:
(a)
FMVSC achieved optimal or near-optimal running efficiency on five datasets, indicating that FMVSC inherits the linear time complexity of the anchor point strategy, significantly reducing the runtime. This is because FMVSC effectively leverages the consensus and complementary information between views, allowing it to converge in fewer iterations. At the same time, our algorithm only requires constructing smaller-sized anchor graphs for each view to represent the current view space.
(b)
Unlike the anchor-based methods, MLRSSC, MSC_IAS, and SFMC construct a representation matrix of n × n for each view during the clustering process and then combine them into a consistent matrix of n × n . This leads to O n 2 for the space complexity and at least O n 2 for the time complexity. Therefore, MLRSSC, MSC_IAS, and SFMC may encounter out-of-memory (OM) issues when dealing with large datasets.
(c)
Although LMVSC runs faster on Animal, NUSWIDE, and YoutubeFace, its simple heuristic process does not take full advantage of the consensus and complementary information between views, resulting in poor clustering performance.
Table 7. Runtime comparison (in seconds).
Table 7. Runtime comparison (in seconds).
DatasetMLRSSCMSC_IASLMVSCMSGLSMVSCSFMCRAMCSFOMVCDROurs
Year201820192020202220222022202320242024
Notting-Hill25.0218.785.0218.849.0411.652.915.824.13
WebKB367.35120.558.625.437.8612.582.174.675.36
Wiki294.3393.087.896.2019.6335.064.216.123.74
CCV2352.44839.4662.7735.7665.86595.6923.5346.8716.84
ALOIOMOM140.61203.54712.91OM120.06241.1592.42
AnimalOMOM230.111417.36903.93OM311.27814.73444.65
NUSWIDEOMOM367.74687.331684.32OM538.091725.61589.64
YoutubeFaceOMOM353.932134.65OMOM612.30OM537.95
The best performance is highlighted in bold black font, while the second-best performance is underlined.
In summary, both theoretical and experimental results confirm the high efficiency and effectiveness of FMVSC.

4.6. Ablation Experiments

As mentioned earlier, most existing anchor-based multi-view clustering methods use the same number of anchor points for each view. In contrast, our FMVSC method flexibly selects an appropriate number of anchor points for each view. To demonstrate the superiority of FMVSC, we conducted an ablation experiment comparing the equal anchor point strategy with the flexible anchor point strategy. Specifically, the flexible anchor strategy was our proposed FMVSC, and the equal anchor strategy was to replace the anchor selection strategy in FMVSC with a strategy which selected the same number of anchor points per view. As shown in Table 8, the performance of the flexible anchor point strategy was significantly improved, highlighting the advantages of our approach.

4.7. Parameter Analysis

The hyperparameter λ is for fusing various anchor graphs. In this section, we conduct sensitivity analysis experiments on λ . The specific method involves determining the optimal number of anchors for each view using a grid search strategy, fixing the optimal value, and conducting traversal experiments for λ values within the range [1, 10, 50, 100, 200, 500, 1000]. We selected the Notting-Hill, Wiki, CCV, and ALOI datasets to demonstrate the results.
From Figure 8 and Figure 9, it can generally be observed that λ has a certain impact on clustering performance when its value is between 1 and 1000, typically being within 0.15. Moreover, we can observe the following based on heuristic experience:
(a)
When λ had a small value, such as 1–50, the performance of the ACC, NMI, and Purity datasets was relatively poor. This indicates that when λ is too low, the model may lack sufficient regularization or optimization strength, making it difficult to capture the underlying structure of the data.
(b)
As λ increased to a moderate range, such as 100–200, all indicators reached their peak performance. This suggests that at this range, the value of λ balanced the optimization of model performance with the prevention of overfitting.
(c)
When λ became excessively large, such as 500–1000, the performance of all indicators began to slightly decline. This indicates that excessive regularization may limit the model’s clustering capability, resulting in reduced performance.
Figure 8. The sensitivity analysis experiments of hyperparameter λ on the Notting-Hill and Wiki datasets.
Figure 8. The sensitivity analysis experiments of hyperparameter λ on the Notting-Hill and Wiki datasets.
Electronics 14 00737 g008
Figure 9. The sensitivity analysis experiments of hyperparameter λ on the CCV and ALOI datasets.
Figure 9. The sensitivity analysis experiments of hyperparameter λ on the CCV and ALOI datasets.
Electronics 14 00737 g009
In summary, when λ was set within the range of 100–200, FMVSC could effectively capture the latent structure among the data and achieve optimal clustering performance.

4.8. Convergence Analysis

To prove that FMVSC is a convergence algorithm, we carried out the convergence analysis proof in Section 3. Here, we will continue to prove the convergence of FMVSC experimentally. As shown in Figure 10, the horizontal coordinate is the number of iterations of the experiment, and the vertical coordinate is the function value of our objective function. The trend of the objective value decreasing is evident. In all of the plots, the objective value decreased rapidly in the first few iterations, and as the number of iterations increased, the rate of decrease gradually slowed down. This indicates that the algorithm converged quickly in the initial stages of optimization, while progress slowed down in the later stages. Specially, we can make the following observations:
(a)
There were differences in the convergence speeds across different datasets. The objective values of some datasets, like WebKB, Wiki, CCV, and ALOI, nearly reached convergence within the first five iterations, indicating that FMVSC can quickly find a good solution on these datasets. On the other hand, the objective values of datasets like Notting-Hill, Animal, NUSWIDE, and YoutubeFace required more iterations to reach convergence, indicating that the convergence process was more complex for these datasets, and more computation time was needed.
(b)
The objective values for all datasets stabilized after 10 iterations, indicating that our algorithm can reach a convergent state after only a few iterations across different datasets. This makes FMVSC particularly suitable for large-scale datasets.
Figure 10. The convergence graph of the objective function for each dataset.
Figure 10. The convergence graph of the objective function for each dataset.
Electronics 14 00737 g010
In summary, we proved the convergence of FMVSC theoretically and experimentally.

5. Conclusions

In this paper, we proposed a scalable and fast multi-view subspace clustering method based on flexible anchor fusion. Different from the existing anchor-point based method, the same number of anchor points was selected from each view, and then the fusion map was obtained by a simple linear weighted combination. Our proposed framework can automatically select the number of anchor points suitable for each data space view and then flexibly integrate anchor graphs with different sizes. Our framework also learns the global and local information between data, which allows our fusion graph to preserve the clustering structure between views to the maximum extent possible. In order to solve our proposed objective function, we proposed an optimization algorithm which can prove convergence both experimentally and theoretically. Through a large number of experiments, we can also see that the clustering accuracy and efficiency of our proposed algorithm were better than the current large-scale multi-view clustering methods. Of course, FMVSC also has some limitations. For example, we have not yet extended it to the field of incomplete multi-view clustering. In future research, we will focus on applying the flexible anchor graph fusion strategy to incomplete multi-view clustering.

Author Contributions

Conceptualization, Y.Z. and S.Z.; methodology, Y.Z. and S.Z.; software, Y.Z.; validation, Y.Z.; formal analysis, Y.Z. and S.Z.; investigation, Y.Z.; resources, Y.Z. and S.Z.; data curation, Y.Z., S.Z., and G.J.; writing—original draft preparation, Y.Z. and S.Z.; visualization, Y.Z.; supervision, S.Z.; project administration, S.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, S.; Yang, M.; Wang, X.; Song, W. Anchor-based scalable multi-view subspace clustering. Inf. Sci. 2024, 666, 120374. [Google Scholar] [CrossRef]
  2. Wang, S.; Liu, X.; Zhu, X.; Zhang, P.; Zhang, Y.; Gao, F.; Zhu, E. Fast parameter-free multi-view subspace clustering with consensus anchor guidance. IEEE Trans. Image Process. 2021, 31, 556–568. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, S.; Liu, X.; Liu, S.; Tu, W.; Zhu, E. Scalable and structural multi-view graph clustering with adaptive anchor fusion. IEEE Trans. Image Process. 2024, 33, 4627–4639. [Google Scholar] [CrossRef]
  4. Peng, X.; Zhu, H.; Feng, J.; Shen, C.; Zhang, H.; Zhou, J.T. Deep clustering with sample-assignment invariance prior. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 4857–4868. [Google Scholar] [CrossRef]
  5. Wang, Q.; Qin, Z.; Nie, F.; Li, X. Spectral embedded adaptive neighbors clustering. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 1265–1271. [Google Scholar] [CrossRef]
  6. Yang, G.; Li, Q.; Yun, Y.; Lei, Y.; You, J. Hypergraph Learning-Based Semi-Supervised Multi-View Spectral Clustering. Electronics 2023, 12, 4083. [Google Scholar] [CrossRef]
  7. Xie, D.; Li, Z.; Sun, Y.; Song, W. Robust Tensor Learning for Multi-View Spectral Clustering. Electronics 2024, 13, 2181. [Google Scholar] [CrossRef]
  8. Chen, J.; Yi, Z. Sparse representation for face recognition by discriminative low-rank matrix recovery. J. Vis. Commun. Image Represent. 2014, 25, 763–773. [Google Scholar] [CrossRef]
  9. Liu, M.; Liang, K.; Zhao, Y.; Tu, W.; Zhou, S.; Gan, X.; Liu, X.; He, K. Self-supervised temporal graph learning with temporal and structural intensity alignment. IEEE Trans. Neural Netw. Learn. Syst. 2024, 1–13. [Google Scholar] [CrossRef] [PubMed]
  10. Wan, X.; Liu, J.; Liang, W.; Liu, X.; Wen, Y.; Zhu, E. Continual multi-view clustering. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; pp. 3676–3684. [Google Scholar]
  11. Gao, H.; Nie, F.; Li, X.; Huang, H. Multi-view subspace clustering. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 4238–4246. [Google Scholar]
  12. Lan, M.; Meng, M.; Yu, J.; Wu, J. Generalized multi-view collaborative subspace clustering. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 3561–3574. [Google Scholar] [CrossRef]
  13. Chen, Y.; Xiao, X.; Peng, C.; Lu, G.; Zhou, Y. Low-rank tensor graph learning for multi-view subspace clustering. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 92–104. [Google Scholar] [CrossRef]
  14. Wang, S.; Wang, Y.; Lu, G.; Le, W. Mixed structure low-rank representation for multi-view subspace clustering. Appl. Intell. 2023, 53, 18470–18487. [Google Scholar] [CrossRef]
  15. Lan, W.; Yang, T.; Chen, Q.; Zhang, S.; Dong, Y.; Zhou, H.; Pan, Y. Multiview subspace clustering via low-rank symmetric affinity graph. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 11382–11395. [Google Scholar] [CrossRef]
  16. Ma, H.; Wang, S.; Zhang, J.; Yu, S.; Liu, S.; Liu, X.; He, K. Symmetric Multi-view Subspace Clustering with Automatic Neighbor Discovery. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 8766–8778. [Google Scholar] [CrossRef]
  17. Sun, M.; Zhang, P.; Wang, S.; Zhou, S.; Tu, W.; Liu, X.; Zhu, E.; Wang, C. Scalable multi-view subspace clustering with unified anchors. In Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, China, 20–24 October 2021; pp. 3528–3536. [Google Scholar]
  18. Brbić, M.; Kopriva, I. Multi-view low-rank sparse subspace clustering. Pattern Recognit. 2018, 73, 247–258. [Google Scholar] [CrossRef]
  19. Lu, C.; Feng, J.; Lin, Z.; Mei, T.; Yan, S. Subspace clustering by block diagonal representation. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 487–501. [Google Scholar] [CrossRef]
  20. Chen, X.; Cai, D. Large scale spectral clustering with landmark-based representation. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 7–11 August 2011; pp. 313–318. [Google Scholar]
  21. Kang, Z.; Zhou, W.; Zhao, Z.; Shao, J.; Han, M.; Xu, Z. Large-scale multi-view subspace clustering in linear time. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 4412–4419. [Google Scholar]
  22. Li, Y.; Nie, F.; Huang, H.; Huang, J. Large-scale multi-view spectral clustering via bipartite graph. In Proceedings of the AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  23. Liu, S.; Wang, S.; Zhang, P.; Xu, K.; Liu, X.; Zhang, C.; Gao, F. Efficient one-pass multi-view subspace clustering with consensus anchors. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 27 February–2 March 2022; pp. 7576–7584. [Google Scholar]
  24. Chen, J.; Yang, J. Robust subspace segmentation via low-rank representation. IEEE Trans. Cybern. 2013, 44, 1432–1445. [Google Scholar] [CrossRef]
  25. Elhamifar, E.; Vidal, R. Sparse subspace clustering: Algorithm, theory, and applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar] [CrossRef]
  26. Nie, F.; Chang, W.; Wang, R.; Li, X. Learning an optimal bipartite graph for subspace clustering via constrained Laplacian rank. IEEE Trans. Cybern. 2021, 53, 1235–1247. [Google Scholar] [CrossRef]
  27. Parsons, L.; Haque, E.; Liu, H. Subspace clustering for high dimensional data: A review. ACM Sigkdd Explor. Newsl. 2004, 6, 90–105. [Google Scholar] [CrossRef]
  28. Zhou, S.; Wang, X.; Yang, M.; Song, W. Multi-view clustering with adaptive anchor and bipartite graph learning. Neurocomputing 2025, 611, 128627. [Google Scholar] [CrossRef]
  29. Shi, L.; Cao, L.; Wang, J.; Chen, B. Enhanced latent multi-view subspace clustering. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 12480–12495. [Google Scholar] [CrossRef]
  30. Kang, Z.; Lin, Z.; Zhu, X.; Xu, W. Structured graph learning for scalable subspace clustering: From single view to multiview. IEEE Trans. Cybern. 2021, 52, 8976–8986. [Google Scholar] [CrossRef] [PubMed]
  31. Li, X.; Zhang, H.; Wang, R.; Nie, F. Multiview clustering: A scalable and parameter-free bipartite graph fusion method. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 330–344. [Google Scholar] [CrossRef]
  32. Yang, B.; Wu, J.; Zhang, X.; Lin, Z.; Nie, F.; Chen, B. Robust anchor-based multi-view clustering via spectral embedded concept factorization. Neurocomputing 2023, 528, 136–147. [Google Scholar] [CrossRef]
  33. Lao, J.; Huang, D.; Wang, C.-D.; Lai, J.-H. Towards scalable multi-view clustering via joint learning of many bipartite graphs. IEEE Trans. Big Data 2023, 10, 77–91. [Google Scholar] [CrossRef]
  34. Chan, P.K.; Schlag, M.D.; Zien, J.Y. Spectral k-way ratio-cut partitioning and clustering. In Proceedings of the 30th international Design Automation Conference, Dallas, TX, USA, 14–18 June 1993; pp. 749–754. [Google Scholar]
  35. Li, L.; He, H. Bipartite graph based multi-view clustering. IEEE Trans. Knowl. Data Eng. 2020, 34, 3111–3125. [Google Scholar] [CrossRef]
  36. Jiang, T.; Gao, Q.; Gao, X. Multiple graph learning for scalable multi-view clustering. arXiv 2021, arXiv:2106.15382. [Google Scholar]
  37. Wang, W.; Carreira-Perpinán, M.A. Projection onto the probability simplex: An efficient algorithm with a simple proof, and an application. arXiv 2013, arXiv:1309.1541. [Google Scholar]
  38. Wang, S.; Liu, X.; Zhu, E.; Tang, C.; Liu, J.; Hu, J.; Xia, J.; Yin, J. Multi-view clustering via late fusion alignment maximization. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; pp. 3778–3784. [Google Scholar]
  39. Bezdek, J.C.; Hathaway, R.J. Convergence of alternating optimization. Neural Parallel Sci. Comput. 2003, 11, 351–368. [Google Scholar]
  40. Zhang, Y.-F.; Xu, C.; Lu, H.; Huang, Y.-M. Character identification in feature-length films using global face-name matching. IEEE Trans. Multimed. 2009, 11, 1276–1288. [Google Scholar] [CrossRef]
  41. Cai, D.; He, X.; Wu, X.; Han, J. Non-negative matrix factorization on manifold. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 63–72. [Google Scholar]
  42. Wang, X.; Lei, Z.; Guo, X.; Zhang, C.; Shi, H.; Li, S.Z. Multi-view subspace clustering with intactness-aware similarity. Pattern Recognit. 2019, 88, 50–63. [Google Scholar] [CrossRef]
  43. Wan, X.; Liu, J.; Gan, X.; Liu, X.; Wang, S.; Wen, Y.; Wan, T.; Zhu, E. One-step multi-view clustering with diverse representation. IEEE Trans. Neural Netw. Learn. Syst. 2024, 1–13. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (left) Existing anchor graph fusion strategy with equally sized anchors. (right) Our proposed FMVSC framework.
Figure 1. (left) Existing anchor graph fusion strategy with equally sized anchors. (right) Our proposed FMVSC framework.
Electronics 14 00737 g001
Figure 2. Running results of MLRSSC on artificial datasets.
Figure 2. Running results of MLRSSC on artificial datasets.
Electronics 14 00737 g002
Figure 3. Running results of LMVSC on artificial datasets.
Figure 3. Running results of LMVSC on artificial datasets.
Electronics 14 00737 g003aElectronics 14 00737 g003b
Figure 4. Running results of MSGL on artificial datasets.
Figure 4. Running results of MSGL on artificial datasets.
Electronics 14 00737 g004
Figure 5. Running results of FMVSC on artificial datasets.
Figure 5. Running results of FMVSC on artificial datasets.
Electronics 14 00737 g005aElectronics 14 00737 g005b
Figure 6. Illustrations of the learned data distributions of the FMVSC and LMVSC datasets with the t-SNE algorithm on the WebKB database.
Figure 6. Illustrations of the learned data distributions of the FMVSC and LMVSC datasets with the t-SNE algorithm on the WebKB database.
Electronics 14 00737 g006
Figure 7. Illustrations of the learned data distribution of FMVSC and LMVSC with the t- SNE algorithm on the Notting-Hill dataset.
Figure 7. Illustrations of the learned data distribution of FMVSC and LMVSC with the t- SNE algorithm on the Notting-Hill dataset.
Electronics 14 00737 g007
Table 1. The mathematical symbols used in our article.
Table 1. The mathematical symbols used in our article.
SymbolDefinition
n Number of samples
k Number of classes
v Number of views
l p Number of anchors in the pth view
d p Dimensions of the pth view
β View coefficient vector
λ Balance parameter
X p R d p × n The pth view data matrix
A p R d p × l p The pth anchor points matrix
Z p R l p × n The pth anchor graph
R p R k × l p The pth rotation matrix
H R k × n The fused spectral embedding matrix
Table 2. Introduction of multi-view datasets.
Table 2. Introduction of multi-view datasets.
DatasetSamplesViewClassFeature
Notting-Hill550352000, 3304, 6750
WebKB1051221840, 3000
Wiki2866210128, 10
CCV677332020, 20, 20
ALOI10,800410077, 13, 64, 125
Animal11,6734502689, 2000, 2001, 2000
NUSWIDE30,00053165, 226, 145, 74, 129
YoutubeFace101,49953164, 512, 64, 647, 838
Table 8. Ablation experiment of equal number anchors versus our proposed flexible strategy.
Table 8. Ablation experiment of equal number anchors versus our proposed flexible strategy.
MetricsAnchor NumbersDatasets
Notting-HillWebKBWikiCCVALOIAnimalNUSWIDEYoutubeFace
ACCequal82.31 ± 3.1693.26 ± 0.0051.99 ± 3.0619.05 ± 0.4461.09 ± 2.5214.62 ± 0.3611.96 ± 0.3016.11 ± 0.65
flexible88.38 ± 4.2995.81 ± 0.0056.04 ± 2.1422.87 ± 0.3466.36 ± 2.1117.77 ± 0.3316.02 ± 0.3921.39 ± 0.61
NMIequal71.64 ± 3.4069.54 ± 0.0050.10 ± 1.6215.11 ± 0.3876.99 ± 1.9412.21 ± 0.5511.83 ± 0.1612.39 ± 0.29
flexible79.86 ± 3.6173.25 ± 0.0052.40 ± 1.5019.27 ± 0.4180.65 ± 1.7614.07 ± 0.2415.10 ± 0.2818.73 ± 0.43
Purityequal83.02 ± 3.1893.26 ± 0.0056.74 ± 3.6922.33 ± 0.3963.44 ± 2.0319.15 ± 0.3723.22 ± 0.2323.84 ± 0.52
flexible89.14 ± 4.1795.81 ± 0.0061.36 ± 2.5524.87 ± 0.3368.63 ± 1.8121.76 ± 0.3027.75 ± 0.3431.60 ± 0.38
F scoreequal77.64 ± 2.2390.87 ± 0.0045.26 ± 3.4311.13 ± 0.3151.33 ± 1.4709.53 ± 0.0907.92 ± 0.1710.37 ± 0.24
flexible83.74 ± 3.9593.70 ± 0.0049.87 ± 2.2413.28 ± 0.2757.43 ± 1.5810.83 ± 0.1611.67 ± 0.2214.13 ± 0.35
The bold black font indicates the better result between the two comparisons.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, Y.; Zhou, S.; Jin, G. Fast Multi-View Subspace Clustering Based on Flexible Anchor Fusion. Electronics 2025, 14, 737. https://doi.org/10.3390/electronics14040737

AMA Style

Zhu Y, Zhou S, Jin G. Fast Multi-View Subspace Clustering Based on Flexible Anchor Fusion. Electronics. 2025; 14(4):737. https://doi.org/10.3390/electronics14040737

Chicago/Turabian Style

Zhu, Yihao, Shibing Zhou, and Guoqing Jin. 2025. "Fast Multi-View Subspace Clustering Based on Flexible Anchor Fusion" Electronics 14, no. 4: 737. https://doi.org/10.3390/electronics14040737

APA Style

Zhu, Y., Zhou, S., & Jin, G. (2025). Fast Multi-View Subspace Clustering Based on Flexible Anchor Fusion. Electronics, 14(4), 737. https://doi.org/10.3390/electronics14040737

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop