Fusing Local and Global Information for One ‐ Step Multi ‐ View Subspace Clustering

: Multi ‐ view subspace clustering has drawn significant attention in the pattern recognition and machine learning research community. However, most of the existing multi ‐ view subspace clustering methods are still limited in two aspects. (1) The subspace representation yielded by the self ‐ expression reconstruction model ignores the local structure information of the data. (2) The con ‐ struction of subspace representation and clustering are used as two individual procedures, which ignores their interactions. To address these problems, we propose a novel multi ‐ view subspace clus ‐ tering method fusing local and global information for one ‐ step multi ‐ view clustering. Our contri ‐ bution lies in three aspects. First, we merge the graph learning into the self ‐ expression model to explore the local structure information for constructing the specific subspace representations of dif ‐ ferent views. Second, we consider the multi ‐ view information fusion by integrating these specific subspace representations into one common subspace representation. Third, we combine the sub ‐ space representation learning, multi ‐ view information fusion, and clustering into a joint optimiza ‐ tion model to realize the one ‐ step clustering. We also develop an effective optimization algorithm to solve the proposed method. Comprehensive experimental results on nine popular multi ‐ view data sets confirm the effectiveness and superiority of the proposed method by comparing it with many state ‐ of ‐ the ‐ art multi ‐ view clustering methods.


Introduction
Clustering is a fundamental unsupervised learning problem that is widely used in the tasks of machine learning [1], computer vision [2], and data mining [3]. It attempts to help to understand the structure of unlabeled data by dividing the entire unlabeled samples into clusters, where the samples in the same cluster are not similar to samples in the other clusters [4][5][6].
With the continuous development of information technology, different features of the object can be easily acquired by different feature extractors, data sources or sensors. For example, an image can be depicted by the color, texture, and edge features. A news report is usually composed of text descriptions and pictures. In the field of autonomous driving, an obstacle can be captured by different types of sensors. These different features can be viewed as multi-view data. Since each view commonly contains view-specific information about the object, using only one view for clustering may yield poor results [7]. Therefore, it is reasonable and appropriate to fuse different views for clustering. It is known that multiple views come from the same object. Hence, multi-view data contain not only the consistency but also the diversity across views. How to reasonably utilize the consistency and diversity to find the underlying clustering structure of multi-view data has become an important research topic .
To deal with the multi-view data, a natural idea is to concatenate these different feature vectors into a new vector and then adopt some existing single-view clustering methods to group the multi-view data. Although this idea is intuitive and simple to deal with multi-view data, it ignores the consistency and complementary information across these views. To address this problem, lots of multi-view clustering methods have been developed to obtain the good clustering performance. For background reading, the reader can refer to the surveys on multi-view clustering [8][9][10]. In this paper, we mainly focus on the multi-view subspace clustering, which has received extensive attention due to its advanced clustering performance and good mathematical interpretability.
Multi-view subspace clustering attempts to construct an ideal subspace representation to describe the multiple linear subspace structure, and the clustering results are then obtained by utilizing the spectral clustering for this obtained subspace representation. The mechanism for computing the subspace representation is based on the self-expressive reconstruction model, where each sample is reconstructed by entire samples. Hence, subspace representation yielded by the self-expressive reconstruction model can exploit the global information but may ignore the local information of multi-view data. Nevertheless, exploring the local structure has been confirmed to improve the learning performance [11]. Moreover, most multi-view subspace clustering methods divide the learning subspace representation and clustering into two individual procedures, which ignores their communications.
To address the above-mentioned issues, in this paper, we propose a novel subspace clustering method fusing local and global information for one-step multi-view subspace clustering (LGOMSC). The proposed method combines the procedures of constructing subspace representation, multi-view information fusion, and clustering into a unified optimization framework. In this framework, as shown in Figure 1, to exploit the local and global information of multi-view data, we integrate graph learning into the self-expressive reconstruction model by adaptively exploring the local structure information for the construction of subspace representation. To capture latent consistency information across views, the proposed method adopts a multi-view information fusion to learn the common subspace representation from these specific subspace representations of different views. Meanwhile, in graph learning, a rank constraint is applied to the Laplacian matrix yielded by the common subspace representation to directly produce the clustering result. Therefore, the proposed method is a one-step multi-view subspace clustering method. The main contributions of the work are summarized as follows:  A novel one-step multi-view subspace clustering method is proposed, which fuses the subspace representation (exploring local and global information), multi-view information fusion (constructing a common subspace representation by fusing different view-specific subspace representations), and clustering (imposing rank constraint on the Laplacian matrix from the common subspace representation) as a unified optimization framework to realize the end-to-end clustering.  We develop an effective optimization algorithm to solve the proposed method. Comprehensive experiments on nine popular multi-view data sets confirm the effectiveness and superiority of the proposed method by comparing it with some state-of-theart multi-view clustering methods. The rest of this paper is organized as follows. In Section 2, we review the related works. In Section 3, we introduce the formulation of the proposed LGOMSC method. In Section 4, we provide the optimization algorithm to solve the proposed LGOMSC method, including the analysis of the convergence and computation complexity. In Section 5, we conduct the experiments on nine popular multi-view data sets and analyze the experimental results. Finally, we provide the conclusion in Section 6.

Related Work
Multi-view clustering is a very powerful data analysis tools for unsupervised learning of data with heterogeneous features. In the past two decades, many multi-view clustering methods have been proposed to achieve robust clustering performance. In the following, we will briefly introduce several multi-view clustering methods from different perspectives.
Subspace-based methods have recently become the mainstay of multi-view clustering research, aiming to discover potential subspace structures across different views. For example, Gao et al. [12] propose a multi-view subspace clustering method that utilizes a common cluster structure to exploit the consistency information across multiple views. Cao et al. [13] propose a diversity-induced multi-view subspace clustering method that adopts the Hilbert-Schmidt independence criterion as the diversity term to explore the complementary information of multi-view data. Luo et al. [14] propose a multi-view subspace clustering method that simultaneously considers the consistency and specificity for learning the subspace representation. Wang et al. [15] propose a multi-view subspace clustering method that considers the complementarity of multi-view data by adopting a position-aware exclusivity term. Guo et al. [16] propose a rank consistency induced multiview subspace clustering model that learns a consistent subspace structure. Brbić and Kopriva [17] propose a multi-view subspace clustering method that adopts an agreement term to ensure the consistency among these subspace representations. To capture the high-order correlations underlying multi-view data, the tensor technique is adopted to exploit the complementary information among different views. For example, Zhang et al. [18] propose a low-rank tensor constrained multi-view subspace clustering model that adopts a low-rank tensor constraint for the obtained subspace representations. Xie et al. [19] utilize the subspace representations of multiple views as a tensor data and then utilize the tensor-singular value decomposition on the rotated tensor to guarantee the consensus among different views. Zhang et al. [20] propose a tensorized multi-view subspace representation learning that adopts a low-rank constraint model for the subspace representation tensor. Yin et al. [21] propose a multi-view subspace clustering model by organizing the multi-view data as tensorial data, and the tensorial data can be represented by a tlinear combination with sparse and low-rank penalty. Recently, researchers considered partition-level multi-view information fusion and proposed a partition-based clustering model to construct joint optimization of multi-view subspace clustering. For example, Kang et al. [22] propose a unified multi-view subspace clustering model that implements the graph construction, the generation of basic partitions, and the fusion of consensus clustering in an interactive way. Lv et al. [23] propose a partition fusion-based multi-view subspace clustering method that utilizes the different partitions to find a shared partition. Zhang et al. [24] develop a consensus one-step multi-view subspace clustering method that fuses the subspace representation learning, partition learning, and clustering into a whole to iteratively optimize. Kang et al. [25] propose to integrate multi-view information in the partition space and obtain clustering results by assigning each partition with a respective rotation matrix. Furthermore, each view is assigned a weight to consider the differences in the clustering capacity of the views. The anchor-based model is proposed to fit for the large-scale multi-view data. Kang et al. [26] propose a large-scale multi-view subspace clustering method by integrating the anchor graphs from different views for spectral clustering. Wang et al. [27] propose a fast parameter-free multi-view subspace clustering by adaptively learning the anchors and graph structure. Sun et al. [28] propose to combine anchor learning and graph construction into a unified optimization framework, allowing the learned anchors to represent the actual latent data distribution more accurately, leading to a more discriminative clustering structure.
Matrix factorization-based methods refer to obtaining consistent latent representations through matrix factorization. Specifically, a given data matrix can be represented by the product of two or more low-dimensional matrices. Liu et al. [29] extended the traditional single-view non-negative matrix factorization algorithm to multi-view application scenarios and proposed a multi-view clustering algorithm based on non-negative matrix factorization. Guo et al. [30] propose to exploit group sparsity inducing norm in a matrix factorization framework to learn shared sparse subspace representations. Recently, Wang et al. [31] proposed a diversity non-negative matrix factorization multi-view clustering method by introducing a new diversity term to increase the diversity among multi-view representations and linearize the running time. Nie et al. [32] propose a new joint clustering method named Fast Multi-view Matrix Tri-Factorization to reduce the information loss in the matrix factorization process, while reducing the computational complexity and improving the operational efficiency. Liu et al. [33] propose a novel multi-view matrix factorization-based clustering method, which proposes to consider the higher-order relationships among features using an optimal graph regularization strategy and introduces the Hilbert-Schmidt independence criterion (HSIC) to fully explore the complementary information in different views. In addition, researchers have extended matrix factorization from the perspective of intact space learning [34]. For example, Zhang et al. [35] propose a latent multi-view subspace clustering that utilizes the latent representation for subspace clustering. Li et al. [36] propose a flexible multi-view representation learning that utilizes the kernel dependence measure to obtain a latent representation from different views for subspace clustering. Xie et al. [37] propose a multi-view subspace clustering method that fuses graph learning, latent representation, and clustering into a unified optimization framework.
Graph-based methods provides an effective way to solve the nonlinearly separated problems. For example, Tang et al. [38] propose a fusion process using linked matrix factorization to fuse the graph matrices corresponding to all views with multiple sources of information. Nie et al. [39] propose a multi-view graph clustering method based on the idea of manifold learning that can perform local structure learning and multi-view clustering at the same time and can also adaptively learn the weights corresponding to each view. Meanwhile, Nie et al. [40] propose an automatic weighting method to fuse a series of view-specific low-quality graphs into a high-quality unified graph, while extending the Laplacian rank approach to multi-view learning. Similarly, Zhan et al. [41] further design a notable clustering method based on twostep multiple graph fusion strategy. Recently, Zhan et al. [42] proposed a method to learn a consensus graph matrix by all views by minimizing disagreement between different views and constraining the rank of the Laplacian matrix. Wang et al. [43] propose another graph-based multi-view clustering method that automatically fuses multiple graph matrices to generate a unified graph matrix. The learned unified graph matrix can help the graph matrices of all views and gives the clustering indicator matrix. Recently, Zhao et al. [44] proposed to minimize the divergence between graphs using tensor Schatten p-norm regularization and integrate the tensor Schatten p-norm regularization and the manifold learning regularization into a unified framework to learn a shared common graph.
Although most of existing multi-view subspace clustering methods have achieved good clustering performance, they still have some limitations. First, the subspace representation generated by the self-expression reconstruction model usually ignores the local structure of the data set. Second, most multi-view subspace clustering methods usually divide the subspace representation learning process and the subsequent clustering task into two separate processes, ignoring the interactions between them. To address these issues, in this paper we propose an LGOMSC method that considers adding graph learning to explore local information adaptively for obtaining subspace representations. Moreover, LGOMSC performs multi-view information fusion directly on the subspace representation and introduces rank constraints on the Laplacian matrix of the common subspace representation matrix, which helps to naturally partition the data points into the desired number of clusters. Our approach integrates similarity learning, multi-view information fusion and clustering as a unified framework to achieve multi-view clustering in an endto-end manner.
Duan et al. [45] propose a multi-view subspace clustering (MVSCLG) that also utilizes the local and global information to achieve the end-to-end clustering. The main differences between MVSCLG and LGOMSC include: (1) to explore the consistency between different views, MVSCLG adopts the spectral matrix fusion, but LGOMSC adopts the graph matrix fusion; (2) to achieve end-to-end clustering, MVSCLG adopts a rotation matrix to map the common spectral matrix to the final cluster label matrix, but LGOMSC adopts a rank constraint on the common Laplacian matrix to directly achieve clustering. Moreover, compared with MVSCLG, LGOMSC has some advantages. Firstly, MVSCLG involves many singular value decomposition procedures and contains many variables, which leads to longer running times and more memory usage than LGOMSC. Secondly, LGOMSC contains fewer hyperparameters than MVSCLG, which is more suitable for practical applications.

Notations
For convenience, we list important mathematic notations that are used throughout the paper in Table 1. Matrices are represented in bold uppercase, while vectors are represented in bold lowercase.  The rank of a matrix

Formulation
In this section, we provide the detailed modeling process of the proposed LGOMSC method. For a multi-view data set with V views, let 1 , , V  X X be the data matrices of the V views and mensionality of the v-th view, and n is the number of data points. Since each view contains view-specific information about the object, we respectively compute the view-specific subspace representation of each view to capture the diversity across views. The objective function of the self-expression model for the multi-view data can be formulated as:  (1) ignores the local structure to construct the subspace representations. In other words, two closed data samples should have similar subspace representations. Hence, to exploit the local information of multi-view data, we integrate the graph learning into Model (1) to compute the subspace representation. Hence, the objection function can be formulated as: Since multi-view data come from the same object, they should have latent consistency. To characterize this consistency, we adopt a multi-view information fusion term to obtain a common subspace representation matrix S . This term can be represented as: Through minimizing Model (3), this common subspace representation matrix U can make these the subspace representation matrices

S S S
to have latent consistency. Hence, we add this multi-view information fusion term into Model (2) as: After obtaining the common subspace structure, we can get the affinity matrix and perform spectral clustering on such a subspace affinity matrix. However, the constructions of subspace representation and clustering are divided into two individual procedures, which ignore their interactions. To address this problem, we consider introducing a rank constraint [46] on the Laplacian matrix If the common subspace representation matrix U is non-negative, then the Laplacian matrix has the following theorem.  (4), the common subspace representation matrix U will have the ideal property. Therefore, we can directly obtain the cluster result from U without discretization.
However, it is difficult to directly solve the rank constraint where 0   is a parameter,  (5), we integrate subspace representation learning, multi-view information fusion, and clustering into a unified framework. The aim is to exploit the internal relationships of the three procedures to obtain a good clustering performance.

Optimization
There are several variables and constraints in the proposed method. To effectively solve these variables from LGOMSC, we developed an alternate optimization algorithm.

4.1.
Update v S When U and F are fixed, the objective function about v S becomes: In this paper, we adopt a two-step approximation strategy [25] to optimize v S .
Firstly, we ignore the constraints in Model (6) to solve v S as: Through making the derivative of Model (7) of v S as zero, we have: is the ij-th element of Model (9) is a constrained quadratic optimization problem, which can be effectively solved by the iterative algorithm in the work [49].

Update U
When v S and F are fixed, the objective function about U is represented as: Model (12) is effectively optimized by an iterative algorithm referring to the work [50].

Update F
When U and v S are fixed, the objective function about F is represented as: The optimal solution F yielded by Model (13) is formed by the c eigenvectors of U L corresponding to the c smallest eigenvalues. Finally, the procedure for optimizing Model (5) is described in Algorithm 1.

Convergence Analysis
In this paper, we adopt an alternate updating algorithm (Algorithm 1) to solve the objective function in Model (5). Since  is changed during the iteration to accelerate the procedure in the experiment, the objective function of Model (5) is varied during each iteration. Hence, it is difficult to guarantee convergence theoretically. However, in the experiments, the results show that Algorithm 1 for optimizing Model (5) has good convergence.

Computational Complexity Analysis
According to the optimization process described in Algorithm , where T is the total number of iterations.
Initialize U and F based on 1 2 , ... V S S S .

Repeat
Update v S by model (9). Update U by model (12). Update F by model (13).

Experiments
In this section, we use nine popular multi-view data sets to assess the clustering performance of LGOMSC.
We conduct these comparison methods from corresponding open-source codes and follow their papers to set the optimal parameters. LGOMSC contains two parameters  and 1  . For  , it is first set with a proper value in a heuristic way, then in each iteration,  is divided by two if the number of zero eigenvalues of U L is greater than c and multiplied by two if it is smaller than c. For 1  , we adopt the grid search method to empirically choose it in the range of {1,10, 20,...,100} . In this paper, we use L2 norm for data normalization, i.e.,

Experiment Results and Analysis
In the experiments, four popular metrics including accuracy (ACC), normalized mutual information (NMI), F-score and adjusted Rand index (ARI) are utilized to assess the clustering result. These evaluation metrics reflect different natures of the clustering results, thus providing a comprehensive analysis from multiple perspectives. For all four of the evaluation metrics, higher values indicate better results. The comparison results are shown in Tables 3-6. The best result is highlighted in red font, and the second best result is reported in blue font.  Tables 3-6, one can see that our proposed method can obtain the best results on all multi-view data sets except Caltech101. For the Caltech101 data set, LGOMSC is lower than FeatConcate and RMKMC in terms of NMI. However, in terms of ACC, our proposed method exceeds the second best results on the data sets including 3sources, 100leaves, BBC, Caltech101, COIL-20, NottingHill, Webkb, Cornell, and Wikipedia by 6.51%, 7.93%, 2.19%, 2.31%, 4.51%, 8.07%, 2.76%, 8.72% and 1.59%, respectively. For NMI, our proposed method is 8.93% lower than the best method, RMKMC, on the Caltech101 data set. For F-score, our proposed method exceeds the second best method by 1.41%, 14.45%, 4.39%, 3.89%, 6.55%, 11.23%, 3.75%, 7.75% and 0.63% for the corresponding data sets, respectively. In terms of ARI, our proposed method exceeds the second best method by 0.84%, 14.59%, 5.31%, 7.37%, 6.97%, 14.33%, 12.11%, 2.54% and 0.16% for the corresponding data set, respectively. The above results demonstrate the effectiveness and superiority of the proposed method. Hence, our LGOMSC method is a valuable multi-view subspace clustering method.
From the results in these tables, one can see that the baseline method (i.e., FeatConcate) sometimes exhibits comparable performance to the multi-view subspace method and even exceeds some multi-view clustering methods. However, in most cases, this baseline method still has a big gap in comparison with multi-view clustering methods. It confirms that multi-view clustering methods that consider the consistency or complementary information can obtain a good multi-view clustering performance. However, in some data sets, e.g., 3sources and 100leaves, the multi-view k-means method RMKMC, produces even worse results than FeatConcate. This phenomenon has been observed by some previous researchers [18,54].
Multi-view subspace clustering based on intact space learning methods like LMSC and FMR performs clustering on the latent representation space. However, there is a large gap between these methods and our approach, probably because they separate the representation learning and clustering processes, leading to suboptimal clustering results.
Compared with similar multi-view information fusion methods like GMC and LMVSC, our approach achieves a more impressive performance. This is mainly because we fuse graph learning into the self-expression model to jointly explore local and global structural information in the data. More information is used to serve the clustering task and therefore better performance is obtained.
Compared with the partition-based models to construct multi-view subspace clustering for joint optimization methods like COMVSC, mPAC, and PMSC, our approach achieves more impressive performance. This is mainly because we make the clustering structure of the multi-view data revealed while generating the common subspace representation under the rank constraint of Laplacian matrix. Thus, the end-to-end clustering approach facilitates a better clustering performance.
Compared with the MVSCLG method, the overall results from the nine data sets show that our method achieves the best results on all evaluation criteria, except for the Caltech101, Cornell and Wikipedia data sets. It states that using rank constraint can obtain an ideal graph matrix fitting for direct clustering. Our method also has superiorities in terms of running time and memory usage, which will be discussed in the next subsection. Thus, our method is more effective than MVSCLG.
Compared with these state-of-the-art clustering methods, the proposed method can achieve a more impressive clustering performance. The main reason is that it considers the local and global information from the original multi-view data to learn the subspace representation. Moreover, the proposed LGOMSC is an end-to-end model, which fuses the construction of subspace representation, multi-view information fusion, and clustering into a seamless whole. The purpose for this is to dig into their potential correlations.
Statistical Analysis. To demonstrate the statistical properties of our proposed method, we conducted the Friedman test and Nemenyi post-hoc test.
The Friedman test assumes that all the k compared methods hold the same performance on H data sets. Specifically, this model performance evaluation consists of the following two main steps. In the first step, first, sort all methods on each data set from high to low according to the clustering performance index and assign corresponding ordinal values (e.g., 1, 2, ...), and then calculate each method on all data sets average rank. In particular, the ordinal values are averaged if the performance of the two methods is the same.
in R programming language). If the hypothesis is rejected, this indicates a significant difference in the performance of the compared methods. A further Nemenyi post-hoc test is then required to further distinguish between the methods.
In the second step, the Nemenyi post-hoc test calculates the critical distance by Equation (15) to reflect the difference between the average ordinal results of various methods.
( 1) 6 k k CD q H    (15) where q  can be calculate by In our case, the number of compared methods, k, equals 13 and H equals 9. We sort the ACC, NMI, F-score, and ARI of the compared methods from high to low and obtain the average ranking of each method in terms of all data sets.
When  = 0.05, the threshold for the Friedman test was 1.8544. According to Equation (14), the F  values can be calculated for different clustering evaluation metrics (ACC, NMI, F-score, and ARI), which are 8.4348, 13.0826, 5.9892 and 11.3655, respectively.
These F  values are all greater than the threshold of the Friedman test, which rejects the hypothesis that all the methods being compared hold the same performance. Then, we perform the Nemenyi post-hoc test to further distinguish multiple methods. After obtaining the critical distance, CD = 6.2982, according to Equation (15), we can draw the Friedman test chart as Figure 2. For each method, the blue dot marks its average rank. The horizontal lines with the dot at the center indicate the critical distance, CD. If the lines do not have overlapping areas this indicates a significant difference in the comparison methods. From the figure, we can see that there are significant differences between the method in this paper and Co_reg_c, Co_reg_p, MLRSSC, and PMSC, and that the other methods do not differ from one another significantly. Compared with other methods, our method has the best average ranking regardless of the clustering evaluation index. In summary, our proposed method holds statistical advantages.

Running Time Analysis
We used MATLAB 2018b to run each clustering method independently and recorded the running time of each clustering method under one hyperparameter combination on each data set in Figure 3. From these results, one can see that FeatConcate is the fastest among most of multi-view data sets. To explore the consistency or complementary information of the multi-view data, these multi-view clustering methods generally need a relatively long running time to produce the final clustering. Our LGOMSC method is faster than Co_reg_c, Co_reg_p, LMSC, FMR, MLRSSC, RMKMC, mPAC, PMSC, and COMVSC on most data sets. On the Caltech101 data set, our LGOMSC method is 4.35 s slower than MLRSSC, and on the Wikipedia data set, our LGOMSC method is 1.89 s slower than MLRSSC and 0.21 s slower than RMKMC. On these data sets, LMVSC and GMC are the fastest, and they are more suitable for solving large-scale clustering problems. They are more concerned with efficiency rather than effectiveness. Therefore, the clustering performance is relatively poor. Although our proposed method is slower than the methods designed specifically for large-scale scenarios, our method is still comparable to LMVSC and GMC for the NottingHill data set. In Table 7, MVSCLG costs more running times than our method. In the NottingHill data set, LGOMSC costs 4744.49 s under one hyperparameter combination. However, since MVSCLG contains three hyperparameters, we need to conduct the grid search strategy to select the optimal one from 420 hyperparameter combinations, which may need 553.52 h. Hence, we ignore it.

Memory Usage Analysis
We use MATLAB 2018b to run each clustering method independently and recorded the memory usage of each clustering method. Specifically, we recorded the current matlab memory usage once before we ran the program. After the program has finished, we recorded the current matlab memory usage again. Subtracting the first reading from the second reading gives us the memory usage of the method. For example, on the 3sources data set, as can be seen in Figure 4, the memory usage of our method is small compared with the LMVSC, PMSC, COMVSC and MVSCLG methods. The memory usage of our method is slightly larger than the other remaining methods, but there is no significant difference between our method and the other remaining methods in terms of memory usage by an order of magnitude. Therefore, our method has an appropriate space complexity.

Convergence Study
The objective function of LGOMSC has multiple variables and constraints. We have developed an effective iterative optimization algorithm to solve the proposed method. We conduct the convergence experiment in Figure 5, which provides the convergence curves of LGOMSC on nine multi-view data sets. The x-axis displays the number of iterations, and the y-axis displays the corresponding objective function value. From these results, one can see that the proposed method is well convergent and converges quickly. Within 10 iterations, the objective function value can converge to a stable value.

Parameter Tuning
We conducted an experiment to analyze the hyperparameter 1  in this paper. In the experiment, we tuned 1  from a candidate set of {1,10, 20,...,100} . As shown in Figure   6, we give the parameter tuning of LGOMSC on nine multi-view data sets and show the clustering performance under different values. One can see that the proposed method keeps a relatively robust clustering performance under a large range of 1  . This helps us to easily select a proper parameter.

Conclusions
In this paper, we propose a novel one-step multi-view subspace clustering method, which integrates the self-expression model and graph learning to simultaneously exploit the local and global information of subspace representations from multi-view data. Moreover, to further exploit the hidden relationships between different steps to achieve an endto-end clustering, our method integrates the subspace representation learning, multi-view information fusion, and clustering tasks into a joint framework. Experimental results on nine popular multi-view data sets confirm the effectiveness of our method by comparing with many baseline methods.
In the future, we have two directions to improve our method. First, since our method is a linear model, we will consider expanding our model to non-linear cases to deal with complex multi-view data. Second, for better fitting for large-scale multi-view subspace clustering, we will adopt anchor-based ideas to improve our method.