Abstract
Kernel fuzzy c-means (KFCM) is a significantly improved version of fuzzy c-means (FCM) for processing linearly inseparable datasets. However, for fuzzification parameter , the problem of KFCM (kernel fuzzy c-means) cannot be solved by Lagrangian optimization. To solve this problem, an equivalent model, called kernel probabilistic k-means (KPKM), is proposed here. The novel model relates KFCM to kernel k-means (KKM) in a unified mathematic framework. Moreover, the proposed KPKM can be addressed by the active gradient projection (AGP) method, which is a nonlinear programming technique with constraints of linear equalities and linear inequalities. To accelerate the AGP method, a fast AGP (FAGP) algorithm was designed. The proposed FAGP uses a maximum-step strategy to estimate the step length, and uses an iterative method to update the projection matrix. Experiments demonstrated the effectiveness of the proposed method through a performance comparison of KPKM with KFCM, KKM, FCM and k-means. Experiments showed that the proposed KPKM is able to find nonlinearly separable structures in synthetic datasets. Ten real UCI datasets were used in this study, and KPKM had better clustering performance on at least six datsets. The proposed fast AGP requires less running time than the original AGP, and it reduced running time by 76–95% on real datasets.
1. Introduction
Clustering is an important unsupervised method, and the purpose of clustering is to divide a dataset into multiple clusters (or classes) with high intra-cluster similarity and low inter-cluster similarity. There have been many clustering algorithms, such as k-means (KM) and its variants [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16]. Others are based on minimal spanning trees [17,18,19], density analysis [20,21,22,23,24,25], spectral analysis [26,27], subspace clustering [28,29], etc.
Generally, k-means minimizes the sum of squared Euclidean distances between each sample point and its nearest clustering center [1]. One variant of k-means is kernel k-means (KKM) [30,31,32,33], which is able to find nonlinearly separable structures by using the kernel function method. Another variant of k-means is fuzzy c-means (FCM) [2], which determines partitions by computing the membership degree of each data point to each cluster. The higher the membership degree, the greater the possibility of the data point belonging to the cluster. Although FCM is more flexible in applications [11,12,13,14,15,16], it is primarily suitable for linearly separable datasets. Kernel fuzzy c-means (KFCM) [34] is a significantly improved version of fuzzy c-means for clustering linearly inseparable datasets. However, the problem of KFCM with fuzzification parameter cannot be solved by existing methods.
To solve the special case of KFCM for , a novel model called kernel probabilistic k-means (KPKM) is proposed. In fact, KPKM is a nonlinear programming model constrained on linear equalities and linear inequalities, and it is equivalent to KFCM at m = 1. In theory, the proposed KPKM can be solved by the active gradient projection (AGP) method [35,36]. Since the AGP method may take too much time on large datasets, we further propose a fast AGP (FAGP) algorithm to solve KPKM more efficiently. Moreover, we report experiments demonstrating its effectiveness compared with KFCM, KKM, FCM, and KM.
2. Background and Related Work
There has been a lot of work related to this paper, mainly including k-means, fuzzy c-means, kernel k-means and kernel fuzzy c-means.
K-means minimizes the sum of squared Euclidean distances between each sample point and its nearest cluster center. K-means first chooses initial clustering centers randomly or manually, and then partitions a dataset into several clusters (a data point belongs to the cluster whose clustering center is nearest to the data point), and computes the mean of a cluster as the clustering center. K-means repeatedly updates clustering centers and clusters until convergence. FCM has the same ideal as k-means. FCM introduces a membership degree and a fuzzy coefficient m into the objective function. The higher is, the greater possibility of the i-th data point belonging to the j-th cluster. K-means and FCM belong to partition-based clustering algorithms, and partition-based clustering algorithms usually are not able to cluster linearly inseparable datasets. Kernel method maps a linearly inseparable dataset into a linearly separable space, so kernel k-means (and FCM) using a kernel function can cluster linearly inseparable datasets.
2.1. K-Means and Fuzzy C-Means
Let represent a dataset. K-means divides it into k clusters. If denotes the j-th cluster, we have and . Using to stand for the number of elements in with the center of , k-means can be described as minimizing the following objective function:
where
Let . By using c instead of k and assigning membership degree to the i-th data point in the j-th cluster for and , the fuzzy c-means clustering model can be formulated as follows:
where , and are computed alternately below [2].
2.2. Kernel K-Means and Kernel Fuzzy C-Means
To improve performance of k-means and fuzzy c-means in linearly inseparable datasets, we may develop their kernel versions by choosing a feature mapping from data points to kernel Hilbert space [37]. Though is usually unknown, it must satisfy
where is a kernel function. Commonly used kernel functions are presented in Table 1.
Table 1.
Commonly used kernel functions.
In Table 1, denotes the inner product of and , and are parameters of the kernel. The objective function of kernel k-means is defined as
where
Let represent the membership degree of the i-th point belonging to the j-th class. Likewise, we can define the kernel FCM clustering model as follows.
where
The membership degree is computed via
3. Kernel Probabilistic K-Means
In this section, the kernel probabilistic k-means (KPKM) are proposed.
We first review the problem. When , the KFCM model gets into a special case, namely,
where
This special case cannot be solved by Lagrangian optimization for , because (12) cannot be computed when (when , cannot be computed).
In this paper, we use the optimization methods to solve this problem, but the partial derivative of (13) with respect to is , and it does not contain .
In order to solve the problem, we introduce (14) into (13), and redefine the member degrees as probability for and . Finally, we have
where probability vector
(16) is the proposed kernel probabilistic k-means.
The proposed KPKM is a soft clustering method. The is the probability of the i-th data point belonging to the j-th cluster. The higher is, the greater possibility of the i-th data point belonging to the j-th cluster. In KKM, the membership degree has only two values (0 and 1). In the proposed KPKM, (although final ).
4. A Fast Solution to KPKM
KPKM is a nonlinear programming problem with linear equalities and linear inequalities constraints, and it is able to be solved by active gradient projection [35,36] theoretically. In this section, on the basis of AGP, we first calculate the gradient of of the objection function of KPKM, and then design a fast AGP algorithm to solve the KPKM model.
The fast AGP has two advantages: iteratively updating the projection matrix and estimating the maximum step length.
4.1. Gradient Calculation
For convenience, we define as
Using the chain rule on (16), we obtain
According to (17), we can further derive
4.2. Fast AGP
In the constraints of KPKM, there are L linear equalities and K×L linear inequalities,
Let . Let be the identity matrix of size . Define two matrices, inequality matrix and equality matrix , where
Note that each row of corresponds to one and only one inequality in (23), and each row of corresponds to one and only one equality in (22). Accordingly, the KPKM’s constraints can be simply expressed as
where . Let stand for a randomly initialized probability vector, and for the probability vector at iteration k. The rows of inequality matrix can be broken into two groups: one is active; the other is inactive. The active group is composed of all inequalities that must work exactly as an equality at , whereas the inactive group is composed of the left inequalities. If and respectively denote the active group and the inactive group, we have
At iteration k, the active matrix is defined as
When is not a square matrix, we can construct its projection matrix and the corresponding orthogonal projection matrix as follows:
Suppose that from is the active row vector at iteration k, satisfying and . According to matrix theory [38], we can more efficiently compute and by
Furthermore, we can compute the projected gradient by
Using , we update the probability vector as
where is the step length. Usually, is chosen as a small number. For fast convergence, we estimate the maximum step length as follows.
- (1)
- Let for ;
- (2)
- Compute for and ;
- (3)
- .
As shown in Figure 1, maximum step length is compared with small step length.
Figure 1.
Maximum step length and small step length.
When is a square matrix, it must be invertible. In this case, we actually have , and . Thus, is not a feasible descent direction. A new descent direction can be computed as follows:
- (1)
- Compute a new vector,
- (2)
- Break into two parts and , namely,where the size of is the number of rows of , and that of is the number of rows of .
- (3)
4.3. Analysis of Complexity
In this section, the computational complexities for the traditional AGP algorithm and the proposed FAGP algorithm are analyzed. Let represent the iteration number of AGP. Let represent the iteration number of FAGP. In computations of all matrices, the computational complexity of the projection matrix is the highest. In the AGP algorithm, computing via (30) requires , so the total computational complexity for AGP algorithm is . In the FAGP algorithm, is computed via (32), and it does not require to compute the inverse of matrices (i.e., ), and computing via (32) requires , so the total computational complexity for FAGP algorithm (i.e., Algorithm 1) is .
| Algorithm 1 Fast active gradient projection (FAGP). |
| Input: X and K |
| Do: |
| (1) Let , ,, initialize , go to (4); |
| (2) Find meeting and ; |
| (3) Compute by (32); |
| (4) Compute by (31), and by (33); |
| (5) Compute by (35), and by (34); |
| (6) Let , if , go to (2); |
| (7) Construct by (29), and by (36); |
| (8) Break into and by (37); |
| (9) If all elements of are more than 0, stop; |
| (10) Choose any element less than 0 from , and delete the corresponding row of ; |
| (11) Reconstruct by (29); |
| (12) Compute by (30), and by (31); |
| (13) Compute by (33), and by (35); |
| (14) Compute by (34); |
| (15) Let , and go to (7). |
| Output: the probability vector |
5. Experimental Results
In order to evaluate performance of the proposed KPKM model (equivalent to KFCM at m = 1) solved by the FAGP algorithm, we have conducted a lot of experiments on one synthetic dataset, ten UCI datasets (http://archive.ics.uci.edu/ml (accessed on 8 March 2021)) and the MNIST dataset (http://yann.lecun.com/exdb/mnist/ (accessed on 8 March 2021)). These datasets are detailed in Section 5.1, Section 5.2 and Section 5.3. In Section 5.1, we use a synthetic dataset to compare KPKM using a Gaussian kernel with KPKM using a linear kernel. In Section 5.2 and Section 5.3, we compare the proposed KPKM with KFCM, KKM, FCM, and KM to evaluate the performance of KPKM solved by the proposed fast AGP. Moreover, the Section 5.4 and Section 5.5 evaluate the descent stability and convergence speed of the proposed FAGP, respectively. In the experiments, we implemented our own MATLAB code for KPKM, KFCM, and KKM with two build-in functions, sparse and full, employed for matrix optimization. Moreover, we called MATLAB’s build-in functions fcm and kmeans for FCM and KM, respectively.
All experiments were carried out on a PC with an Intel(R) Core(TM) i7-4790 CPU at 3.60 GHz, 8.00 G RAM, running Windows 7 and MATLAB 2015a.
5.1. The Experiment on the Synthetic Dataset
In this experiment, we analyzed the influences of the Gaussian radial basis function kernel and the linear kernel on KPKM when clustering one synthetic dataset, which is shown in Figure 2. The synthetic dataset contained 300 points, with 100 and 200 being from two linearly inseparable classes, disc and ring, respectively. The result is displayed in Figure 3. Obviously, Gaussian radial basis function KPKM can cluster perfectly, whereas linear KPKM cannot.
Figure 2.
The synthetic dataset, which is composed of two linearly inseparable classes: disc and ring.
Figure 3.
The results of the synthetic dataset clustered by Gaussian radial basis function kernel probabilistic k-means (KPKM) (a) and linear KPKM (b).
5.2. Experiment on Ten UCI Datasets
In this experiment, we compare the clustering performance of KPKM with the performances of KFCM, KKM, FCM, and KM in terms of three measures: normalized mutual information (NMI), adjusted Rand index (ARI), and v-measure (VM) [39,40]. NMI is a normalization of the mutual information score to scale the results between 0 (no mutual information) and 1 (perfect correlation). The NMI is defined as
where is the entropy of U; MI is given by
where is the number of the samples in cluster . ARI is an adjusted similarity measure between two clusters. The ARI is defined as
where RI is the ratio of data points clustered correctly to all data points, and E(RI) is the expectation of RI. VM is a harmonic mean between homogeneity and completeness, where homogeneity means that each cluster is a subset of a single class, and completeness means that each class is a subset of a single cluster. The VM is defined as
The higher the NMI, ARI and VM, the better the performance.
We describe the ten UCI datasets represented in Table 2 by name, code, number of instances, number of classes and number of dimensions. A set of experiential parameters was selected. We set m = 2 for KFCM, and m = 1.3 for FCM. With the Gaussian radial basis function kernel, we also selected appropriate (shown in Table 3), and report the clustering results in Table 4. The better results in each case is highlighted in bold. (The performance of many a method depends on parameters [41]. By experiments we found that for different algorithms with the same kernel function, the most appropriate parameters are usually different, so we used different settings to make the methods have the best clustering performances they could. We also tried to select the most appropriate parameter by using our experience.) We ran every algorithm 10 times, and average results were calculated.
Table 2.
The ten UCI datasets used.
Table 3.
Parameter selected for kernel clustering methods.
Table 4.
Comparisons of KPKM with kernel fuzzy c-means (KFCM), kernel k-means (KKM), fuzzy c-means (FCM), and k-means (KM) on ten UCI datasets (DS) in terms of normalized mutual information (NMI), adjusted Rand index (ARI), and v-measure (VM).
From Table 4, we can see that
- (1)
- For NMI, KPKM had the best clustering results on nine datasets, so the clustering performance of KPKM was the best for NMI.
- (2)
- For ARI, KPKM had the best clustering results on six datasets. KFCM, KKM, FCM, and KM performed the best on 2, 1, 0, and 1 datasets, respectively, so KPKM performed the best for ARI.
- (3)
- For VM, KPKM, KFCM, KKM, FCM, and KM had the best clustering results on 7, 0, 2, 1, and 0 datasets, respectively. Thus, KPKM had the best clustering performance for VM.
Overall, KPKM is better than the other models.
5.3. Experiment on the MNIST Dataset
In this experiment, we used the MNIST dataset to compare KPKM with KFCM and KKM when clustering digital images. The MNIST dataset was composed of handwritten digits, with 60,000 examples for training and 10,000 examples for testing. All digits were size-normalized and centered in fixed-size images. To evaluate the clustering performances of KPKM, KFCM, and KKM, we randomly chose 1000 training examples; 100 are shown in Figure 4.
Figure 4.
One-hundred digital examples, 10 per class.
Moreover, we defined a CWSS kernel based on complex wavelet structural similarity (CWSS) [42]. CWSS may be regarded as a coefficient that does not change the structural content of an image. By using to denote the similarity between two images and , we can express the CWSS kernel as
where is set for KFCM, and for both KPKM and KKM. Additionally, m = 1.3 is set for KFCM. We present the results in Table 5, with the examples of Figure 4 correspondingly being displayed in Figure 5. From Table 5, we can see that KPKM outperformed KFCM and KKM in terms of NMI, ARI, and VM. From Figure 5, we can observe that both KPKM and KKM found ten clusters, whereas KFCM found only seven clusters, although the number of clusters was set to ten.
Table 5.
Comparisons of KPKM with KFCM and KKM on the MNIST dataset in terms of NMI, ARI, and VM.
Figure 5.
The results of the examples of Figure 2 clustered by KPKM (a), KFCM (b), and KKM (c).
5.4. Experiment for Descent Stability
In this experiment, we used 10 UCI datasets to evaluate descent stability of FAGP. AGP selects a small step length to converge. If AGP selects a large step length, the objective function value may descend with oscillation, and even does not converge. FAGP iteratively estimates a maximum step length at each iteration for speeding up its convergence. However, does it have any serious influence on the convergence? we ran the proposed FAGP on 10 UCI datasets to demonstrate the descent stability of FAGP. As shown in Figure 6, we can see that FAGP descends stably without oscillation at each iteration.
Figure 6.
Descent stability of FAGP on 6 UCI datasets: Ionosphere (a), Iris (b), Seeds (c), Glass (d), Segmentation (e), Dermatology (f), Breast (g), Natural (h), Yeast (i), and Waveform (j). The x-axis shows the iteration number. The y-axis shows the objective function value.
5.5. Convergence Speed Comparison between FAGP and AGP
In this experiment, we compared the convergence speeds of FAGP and AGP by using running time on 10 UCI datasets. is the ratio of FAGP’s running time to AGP’s. FAGP and AGP used the same initializations in each case. The results are presented in Table 6 and Table 7 (“–” means the running time was too long to obtain the final clustering results), and we can observe the proposed FAGP ran faster than AGP on all the 6 datasets. For Iris and Seeds datasets, FAGP used less than 10% of running time of AGP. FAGP also required fewer iterations than AGP. For large datasets, the running time of AGP was too long, but the proposed fast AGP could obtain the final clustering results.
Table 6.
Running time (s) comparison of FAGP and AGP.
Table 7.
Comparison of the numbers of iterations used by FAGP and AGP.
6. Conclusions
In this paper, a novel clustering model (i.e., KPKM) was proposed. The proposed KPKM solves the problem of KFCM for m = 1, and this problem cannot be solved by existing method. The traditional AGP method can solve the proposed KPKM, but the efficiency of AGP is low. A fast AGP was proposed to speed up the AGP. The proposed fast AGP uses a maximum step length strategy to reduce the iteration number and uses an iterative method to update the projection matrix. The experimental results demonstrated that the fast AGP is able to solve the KPKM and the fast AGP requires less running time than AGP (the proposed FAGP requires 4.22–27.90% of the running time of AGP on real UCI datasets). The convergence of the proposed method was also analyzed by experiments. Additionally, in the experiments, the KPKM model could produce overall better clustering results than the other models, including KFCM, KKM, FCM, and KM. The proposed KPKM obtained the best clustering results on at least 6 real UCI datasets (a total of 10 real UCI datasets were used).
As future work, the proposed KPKM using other kernels will be evaluated in a variety of applications. For large datasets, the proposed method still has some disadvantages, so the next project will include speeding up the AGP on large datasets.
Author Contributions
Data curation, Z.L. and Z.Z.; writing—original draft, B.L.; writing—review and editing, T.Z. and Y.L. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by National Natural Science Foundation of China (61876010, 61806013, and 61906005).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Conflicts of Interest
All authors declare no conflict of interest.
Abbreviations
The following abbreviations are used in this manuscript:
| dataset | |
| i-th data points of | |
| feature mapping from data points to kernel Hilbert space | |
| Laplace Radial Basis Function kernel | |
| Gaussian Radial Basis Function kernel | |
| polynomial kernel | |
| sigmoid kernel | |
| j-th cluster | |
| L | number of elements in |
| number of elements in | |
| center of | |
| K and C | number of clustering center |
| membership degree matrix | |
| membership degree to the i-th data point in the j-th cluster | |
| inequality matrix | |
| equality matrix | |
| probability matrix | |
| identity matrix | |
| probability to the i-th data point in the j-th cluster | |
| projection matrix | |
| orthogonal projection matrix | |
| active matrix | |
| row vector of | |
| projected gradient | |
| t | step length |
References
- Lloyd, S. Least Squares Quantization in PCM. IEEE Trans. Inf. Theory 1982, 28, 129–137. [Google Scholar] [CrossRef]
- Bezdek, J.C. Pattern Recognition with Fuzzy Objective Function Algorithms. Adv. Appl. Pattern Recognit. 1981, 22, 203–239. [Google Scholar]
- Jing, L.; Ng, M.K.; Huang, J.Z. An entropy weighting k-means algorithm for subspace clustering of high-dimensional sparse data. IEEE Trans. Knowl. Data Eng. 2007, 19, 1026–1041. [Google Scholar] [CrossRef]
- Xenaki, S.; Koutroumbas, K.; Rontogiannis, A. Sparsity-aware possibilistic clustering algorithms. IEEE Trans. Fuzzy Syst. 2016, 24, 1611–1626. [Google Scholar] [CrossRef]
- Yang, M.S.; Nataliani, Y. A feature-reduction fuzzy clustering algorithm based on feature-weighted entropy. IEEE Trans. Fuzzy Syst. 2018, 26, 817–835. [Google Scholar] [CrossRef]
- Gu, J.; Jiao, L.; Yang, S.; Liu, F. Fuzzy double c-means clustering based on sparse self-representation. IEEE Trans. Fuzzy Syst. 2018, 26, 612–626. [Google Scholar] [CrossRef]
- Hamasuna, Y.; Endo, Y.; Miyamoto, S. On tolerant fuzzy c-means clustering and tolerant possibilistic clustering. Soft Comput. 2010, 14, 487–494. [Google Scholar] [CrossRef]
- Li, Y.; Yang, G.; He, H. A study of large-scale data clustering based on fuzzy clustering. Soft Comput. 2016, 20, 3231–3242. [Google Scholar] [CrossRef]
- Zhu, L.F.; Wang, J.S.; Wang, H.Y. A Novel Clustering Validity Function of FCM Clustering Algorithm. IEEE Access 2019, 7, 152289–152315. [Google Scholar] [CrossRef]
- Sinaga, K.P.; Yang, M. Unsupervised K-Means Clustering Algorithm. IEEE Access 2020, 8, 80716–80727. [Google Scholar] [CrossRef]
- Wang, C.; Pedrycz, W.; Yang, J.; Zhou, M.; Li, Z. Wavelet Frame-Based Fuzzy C-Means Clustering for Segmenting Images on Graphs. IEEE Trans. Cybern. 2020, 50, 3938–3949. [Google Scholar] [CrossRef]
- Wang, C.; Pedrycz, W.; Li, Z.; Zhou, M.; Ge, S.S. G-image Segmentation: Similarity-preserving Fuzzy C-Means with Spatial Information Constraint in Wavelet Space. IEEE Trans. Fuzzy Syst. 2020. [Google Scholar] [CrossRef]
- Zhang, R.; Li, X.; Zhang, H.; Nie, F. Deep Fuzzy K-Means With Adaptive Loss and Entropy Regularization. IEEE Trans. Fuzzy Syst. 2020, 28, 2814–2824. [Google Scholar] [CrossRef]
- Wang, C.; Pedrycz, W.; Zhou, M.; Li, Z. Sparse Regularization-Based Fuzzy C-Means Clustering Incorporating Morphological Grayscale Reconstruction and Wavelet Frames. IEEE Trans. Fuzzy Syst. 2020. [Google Scholar] [CrossRef]
- Wang, C.; Pedrycz, W.; Li, Z.; Zhou, M.; Zhao, J. Residual-sparse Fuzzy C-Means Clustering Incorporating Morphological Reconstruction and Wavelet frame. IEEE Trans. Fuzzy Syst. 2020. [Google Scholar] [CrossRef]
- Zhang, R.; Nie, F.; Guo, M.; Wei, X.; Li, X. Joint Learning of Fuzzy k-Means and Nonnegative Spectral Clustering with Side Information. IEEE Trans. Image Process. 2019, 28, 2152–2162. [Google Scholar] [CrossRef] [PubMed]
- Wang, X.; Yu, F.; Pedrycz, W.; Wang, J. Hierarchical clustering of unequal-length time series with area-based shape distance. Soft Comput. 2019, 23, 6331–6343. [Google Scholar] [CrossRef]
- Li, Y. A Clustering Algorithm Based on Maximal θ-Distant Subtrees. Pattern Recognit. 2007, 40, 1425–1431. [Google Scholar]
- Esgario, G.M.; Krohling, R.A. Clustering with Minimum Spanning Tree using TOPSIS with Multi-Criteria Information. In Proceedings of the IEEE International Conference on Fuzzy Systems, Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–7. [Google Scholar]
- Rodriguez, A.; Laio, A. Clustering by Fast Search and Find of Density Peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef] [PubMed]
- Bryant, A.; Cios, K. RNN-DBSCAN: A density-based clustering algorithm using reverse nearest neighbor density estimates. IEEE Trans. Knowl. Data Eng. 2018, 30, 1109–1121. [Google Scholar] [CrossRef]
- Xu, X.; Ding, S.; Xu, H.; Liao, H.; Xue, Y. A feasible density peaks clustering algorithm with a merging strategy. Soft Comput. 2019, 23, 5171–5183. [Google Scholar] [CrossRef]
- Wang, X.; Zhang, Y.; Xie, J. A density-core-based clustering algorithm with local resultant force. Soft Comput. 2020, 24, 6571–6590. [Google Scholar] [CrossRef]
- Wu, C.; Lee, J.; Isokawa, T.; Yao, J.; Xia, Y. Efficient Clustering Method Based on Density Peaks with Symmetric Neighborhood Relationship. IEEE Access 2019, 7, 60684–60696. [Google Scholar] [CrossRef]
- Liu, T.; Li, H.; Zhao, X. Clustering by Search in Descending Order and Automatic Find of Density Peaks. IEEE Access 2019, 7, 133772–133780. [Google Scholar] [CrossRef]
- Luxburg, U.V. A Tutorial on Spectral Clustering. Stat. Comput. 2007, 17, 395–416. [Google Scholar] [CrossRef]
- Chen, J.; Li, Z. Huang B. Linear Spectral Clustering Superpixel. IEEE Trans. Image Process. 2017, 26, 3317–3330. [Google Scholar] [CrossRef]
- Elhamifar, E.; Vidal, R. Sparse Subspace Clustering: Algorithm, Theory, and Applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar] [CrossRef]
- Lu, C.; Feng, J.; Lin, Z.; Mei, T.; Yan, S. Subspace Clustering by Block Diagonal Representation. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 487–501. [Google Scholar] [CrossRef]
- Gu, Y.; Chanussot, J.; Jia, X.; Benediktsson, J.A. Multiple Kernel Learning for Hyperspectral Image Classification: A Review. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6547–6565. [Google Scholar] [CrossRef]
- Nguyen, B.; Baets, B.D. Kernel-Based Distance Metric Learning for Supervised k-Means Clustering. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3084–3095. [Google Scholar] [CrossRef]
- Liu, X.; Zhu, X.; Li, M.; Wang, L.; Zhu, E.; Liu, T. Multiple Kernel K-means with Incomplete Kernels. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 1191–1204. [Google Scholar] [CrossRef] [PubMed]
- Marin, D.; Tang, M.; Ben, A.I.; Boykov, Y.Y. Kernel Clustering: Density Biases and Solutions. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 136–147. [Google Scholar] [CrossRef]
- Huang, H.C.; Chuang, Y.Y.; Chen, C.S. Multiple Kernel Fuzzy Clustering. IEEE Trans. Fuzzy Syst. 2012, 20, 120–134. [Google Scholar] [CrossRef]
- Rosen, J.B. The Gradient Projection Method for Nonlinear Programming. Part I. Linear Constraints. J. Soc. Ind. Appl. Math. 1961, 9, 514–532. [Google Scholar] [CrossRef]
- Goldfarb, D.; Lapidus, L. Conjugate Gradient Method for Nonlinear Programming Problems with Linear Constraints. Ind. Eng. Chem. Fundam. 1968, 7, 142–151. [Google Scholar] [CrossRef]
- Girolami, M. Mercer Kernel-Based Clustering in Feature Space. IEEE Trans. Neural Netw. 2002, 13, 780–784. [Google Scholar] [CrossRef]
- Honig, M.; Madhow, U.; Verdu, S. Blind Adaptive Multiuser Detection. IEEE Trans. Inf. Theory 1995, 41, 944–960. [Google Scholar] [CrossRef]
- Vinh, N.X.; Epps, J.; Bailey, J. Information theoretic measures for clusterings comparison: Variants, properties, normalization and correction for chance. J. Mach. Learn. Res. 2010, 11, 2837–2854. [Google Scholar]
- Rosenberg, A.; Hirschberg, J. V-measure: A conditional entropy-based external cluster evaluation measure. In Proceedings of the Empirical Methods in Natural Language Processing and Computational Natural Language Learning, Prague, Czech Republic, 28–30 June 2007; pp. 410–420. [Google Scholar]
- Gao, S.; Zhou, M.; Wang, Y.; Cheng, J.; Yachi, H.; Wang, J. Dendritic Neuron Model With Effective Learning Algorithms for Classification, Approximation, and Prediction. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 601–614. [Google Scholar] [CrossRef]
- Sampat, M.P.; Wang, Z.; Gupta, S.; Bovik, A.C.; Markey, M.K. Complex Wavelet Structural Similarity: A New Image Similarity Index. IEEE Trans. Image Process. 2009, 18, 2385–2401. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).





