You are currently on the new version of our website. Access the old version .
MathematicsMathematics
  • Article
  • Open Access

14 May 2024

Sparse Clustering Algorithm Based on Multi-Domain Dimensionality Reduction Autoencoder

,
,
,
and
1
College of Science, China University of Petroleum, Qingdao 266580, China
2
College of Computer, China University of Petroleum, Qingdao 266580, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.

Abstract

The key to high-dimensional clustering lies in discovering the intrinsic structures and patterns in data to provide valuable information. However, high-dimensional clustering faces enormous challenges such as dimensionality disaster, increased data sparsity, and reduced reliability of the clustering results. In order to address these issues, we propose a sparse clustering algorithm based on a multi-domain dimensionality reduction model. This method achieves high-dimensional clustering by integrating the sparse reconstruction process and sparse L1 regularization into a deep autoencoder model. A sparse reconstruction module is designed based on the L1 sparse reconstruction of features under different domains to reconstruct the data. The proposed method mainly contributes in two aspects. Firstly, the spatial and frequency domains are combined by taking into account the spatial distribution and frequency characteristics of the data to provide multiple perspectives and choices for data analysis and processing. Then, a neural network-based clustering model with sparsity is conducted by projecting data points onto multi-domains and implementing adaptive regularization penalty terms to the weight matrix. The experimental results demonstrate superior performance of the proposed method in handling clustering problems on high-dimensional datasets.
MSC:
62H30; 68T07; 68T09

1. Introduction

Clustering is one of the fundamental topics in machine learning to explore correlations within data by partitioning a set of samples into a number of homogeneous clusters [1]. It is intensively applied to security, text mining, image processing, bioinformatics and other fields [2]. Under the situation of no prior knowledge or a small amount of prior knowledge, clustering models usually classify data according to the similarity or dissimilarity measure between data, where the aim is to make intra-class data tend to be highly similar while inter-class data tend to be highly different [3]. To discover useful information hidden in a large amount of data, it helps to divide the data into several different groups with similar features. There is currently a large amount of research on clustering, such as centroid-based clustering, hierarchical clustering, and spectral clustering, to provide better understanding of the structure and discovers patterns and trends in the data. However, clustering on high-dimensional data remains a challenging research topic.
High-dimensional data are widely used in fields such as computer vision. Compared to low-dimensional data, high-dimensional data can provide richer information, help to gain a deeper understanding of the data and discover hidden complex relationships, leading to more accurate predictions. However, in high-dimensional data, there is often a large amount of redundant information, which poses great challenges to data processing and even leads to the problem of the “curse of dimensionality”. This also makes a predicament in clustering analysis, causing the failure of traditional clustering methods in dealing with high-dimensional data. As the dimensionality of data increases, the difficulty of clustering also increases. Therefore, continuing to explore new methods of high-dimensional clustering to address the challenges of big data is of great significance.
The high-dimensional clustering mainly confronts the following challenges:
  • As the dimension of data increases, a large number of irrelevant attributes emerge in high-dimensional datasets, which cannot form obvious data clusters in all dimensions. In high-dimensional space, the distribution of data points becomes sparser, and it is hard to find the correct clustering structure. Different distance measurement methods may generate different clustering results.
  • In high-dimensional data, data points becomes increasingly similar and the distance between them becomes more blurred. In this situation, traditional clustering methods often fail to fully capture the intrinsic structure of data, making it difficult to effectively identify and distinguish different data categories.
To address these problems, researchers have proposed various methods including ensemble clustering clustering, subspace clustering, and dimensionality reduction-based clustering. Huang et al. [4] introduced a new multi-directional integrated clustering method to solve the curse of dimensionality in high-dimensional data by creating a large number of diverse metrics and utilizing consistency functions to explore clustering diversity. Jia et al. [5] proposed a unified numerical and classification attribute weighting scheme based on subspace clustering, which quantifies the contribution of attributes to clustering by considering inter-cluster differences and intra-cluster similarity, thereby improving the processing ability of high-dimensional data. Hou et al. [6] proposed a new framework called discriminative embedding clustering, which iteratively alternates dimensionality reduction and clustering to improve the ability to efficiently process and manage large-scale high-dimensional data.
Clustering algorithms such as k-means and hierarchical clustering typically rely on manually designed features or similarity measures, which is limited when dealing with complex high-dimensional data. In recent years, deep learning has been used to improve clustering performance significantly. It can discover intricate patterns in data by learning feature representations and improve clustering performance. In 2022, Peng et al. [7] directly established interpretable neural layers to achieve end-to-end clustering learning. Inspired by this, we propose a sparse clustering model based on a multi-domain dimensionality reduction architecture (SMDR). Furthermore, the proposed model accounts for features within both the frequency domain and spatial domain information to facilitate a more comprehensive data analysis and foster improvements in clustering efficiency. On the one hand, the spatial part directly extracts features from data to obtain intrinsic law with intuitive characteristics of the high-dimensional data. On the other hand, frequency features mirror the contributions of each frequency component through decomposing data into different frequency components to capture more refined and complex information in high-dimensional data.
Our contributions of this work are summarized as follows:
  • We present a sparse learning-based architecture with adaptive regularization to achieve clustering in high-dimensional data. An automatic encoder model is designed to generate sparsity in weights and reduce the complexity of clustering based on the multi-domain dimensionality reduction.
  • The spatial information and frequency characteristics of data are combined to learn domain-invariant feature representation. This combination improves noise resistance, extracts rich features, achieves dimensionality reduction and compression, and provides more information and choices for data analysis and processing.
Compared with the TELL model [7], the proposed method amalgamates insights from both spatial and frequency domains to enhance clustering accuracy and robustness. This integration aims to mitigate inaccuracies and incompleteness in clustering outcomes, thereby facilitating a more nuanced description and analysis of data. SMDR enhances sparsity by introducing an adaptive regularization penalty term in the weight matrix to learn domain-invariant feature representation from both the spatial and frequency domains.
The remainder of the paper is organized as follows. Section 2 briefly introduces the related works for high-dimensional clustering. In Section 3, we detail the SDMR framework. Section 4 reports the experimental results on three benchmark datasets. Section 5 concludes the paper.

3. Sparse Clustering Algorithm Based on Multi-Domain Dimensionality Reduction Autoencoder

3.1. Multi-Domain Autoencoder

In this paper, we aim to achieve high-dimensional clustering in an end-to-end model. The sample data are encoded and mapped to a low-dimensional representation space from two domains. A multi-domain joint loss function is designed to guide the model to capture valuable information for clustering. The cluster loss part measures the importance of features between each cluster and the reconstruct loss part measures the reconstruction ability of data. The regulation term is used to inspire sparsity.
In its spatial form, the loss function is:
L c l u 1 W 1 = n = 1 N u = 1 U L u X n 1 ; W X n 1 x u 1 W 1 2 2 X ^ n 1 W 1 = g W 1 f W 1 X n 1 L r e c 1 W 1 = n = 1 N X n 1 X ^ n 1 W 1 2 2 L 1 W 1 = λ L c l u 1 W 1 + L r e c 1 W 1
where W 1 is the weight matrix of the clustering layer, L c l u 1 W 1 is the clustering loss in the spatial domain, n is the index of the sample data, N is the number of sample data, u is the index of clusters, U is the number of clusters, L u X n 1 ; W 1 is the attention of the sample data in the uth spatial domain in the uth cluster, X n 1 is the sample data, x u 1 W 1 is the clustering center of the uth cluster, X ^ n 1 W 1 represents the reconstructed sample data, f W 1 · represents the encoder, g W 1 · represents the decoder, L r e c 1 W 1 represents the reconstruction loss in the spatial domain, λ = 0.01 weights the two losses, and L 1 W 1 represents the total loss in the spatial domain.
Then, we convert the sample data to the frequency domain by wavelet transform. The inner products of the wavelet transform matrix and wavelet coefficient matrix obtain new sample data at different scales and displacements. During the clustering process, frequency domain information is captured, and noises are separated. For X 1 , the discrete form of wavelet transform can be represented as
X n 2 = Q C C = Q T X n 1 Q = ψ j , k P = 2 j / 2 ϕ 2 j P k
where X n 2 represents the sample in the frequency domain after wavelet transform; Q is the wavelet transform matrix, which is also the value of the wavelet basis function at each pixel point; C is the wavelet coefficient matrix; X n 1 represents the data in the spatial domain; ψ j , k represents the wavelet basis function; j represents the scale of the wavelet function; P is the position matrix under the uth data; ϕ is the mother wavelet function with scale 1 and translation 0; and 2 j / 2 is a scaling factor used to adjust the scale of wavelet functions. We use the above transformations to obtain sample data X 2 in the frequency domain. The loss function in the frequency domain is:
L c l u 2 W 2 = n = 1 N u = 1 U L u X n 2 ; W 2 X n 2 x u 2 W 2 2 2 X ^ n 2 W 2 = g W 2 f W 2 X n 2 L r e c 2 W 2 = n = 1 N X n 2 X ^ n 2 W 2 2 2 L 2 W 2 = λ L c l u 2 W 2 + L r e c 2 W 2
where W 2 is the weight matrix of the clustering layer, L c l u 2 W 2 is the clustering loss in the frequency domain, L u X n 2 , W 2 generates sparsity in the frequency domain in the uth cluster, x u 2 W 2 is the clustering center in the frequency domain of the uth cluster, X ^ n 2 represents the reconstructed sample data in the frequency domain, f W 2 · and g W 2 · are the same encoder and decoder as before but the parameter changes to W 2 , and L r e c 2 W 2 is the reconstruction loss in the frequency domain, while L 2 W 2 is the total loss under the frequency domain.
The frequency domain and spatial domains are integrated to construct an autoencoder in different domains. The specific workflow diagram is shown in Figure 1.
Figure 1. Sparse clustering algorithm flowchart under multi-domain autoencoder.
Each loss function part is as follows:
L c l u W = L c l u 1 W 1 + L c l u 2 W 2 = n = 1 N u = 1 U L u X n 1 ; W 1 X n 1 x u 1 W 1 2 2 + L u X n 2 ; W 2 X n 2 x u 2 W 2 2 2 L r e c W = L r e c 1 W 1 + L r e c 2 W 2 = n = 1 N X n 1 X ^ n 1 W 1 2 2 + X n 2 X ^ n 2 W 2 2 2 L M D W = λ L c l u W + L r e c W
where L c l u W is the clustering loss, L r e c W is the reconstruction loss, and L M D W is the total loss under multi-domains. At this point, we have obtained the self-encoder in multi-domains.

3.2. Sparse L1 Regularization Model

We propose a sparse L1 regularization to generate sparsity in the model-training process. This integrated approach implements an adaptive penalty term on the weight matrix:
ρ W = σ W I W 1 + W 1
where σ is the sparsity-induced function, taking the sigmoid function; ⊙ is the Hadamard product; · 1 represents the L1 norm of weight matrix; and I is a all-one matrix with the same dimensions as W. The W I W 1 term penalizes the elements in the W matrix that are close to 0.5 , that is, W and 1 W will approach 0 at these positions. W 1 encourages the elements in W to approach 0, thereby achieving sparsity, and adjusts the degree of regularization accordingly based on the input parameters. Considering that we are studying a multi-domain problem where there are multi-network weight matrices, we rewrite ρ W as ρ M W
ρ M W = m = 1 M σ W m I W m 1 + W m 1
where m is the index of the domain, M is the number of domains (here M = 2 ), and the weight matrices on the clustering layer are W 1 and W 2 , W = W 1 , W 2 . Apply it to the loss function and obtain the final loss function L t o t a l W :
L t o t a l W = L c l u W + L r e c W + ρ 2 W
We employ the L1 norm to generate sparse solutions, optimizing the 0 1 property of the function, and then utilize σ to control the sparsity of function, thereby optimizing the calculation of clustering analysis errors. The flowchart is shown in Figure 2.
Figure 2. Algorithm flowchart for adding sparse L1 regularization terms.
As shown in Figure 2, we add the regularization term ρ 2 W to the loss function of the autoencoder in multi-domains. In the optimization process, some elements in the weight matrix of the clustering layer are penalized to become or approach 0. This also makes the proposed model robust to irrelevant or meaningless noise. This reduces the influence of noise features, making clustering more concentrated and accurate.
The algorithm is as shown in Algorithm 1. As for the complexity of the algorithm, the analysis is mainly divided into the following three parts. For the first part, the encoding and decoding process, the complexity is O ( P ) , and P is the dimension of the input sample data. For the second part, calculate the loss function, which includes calculating the time complexity of soft labels as O ( U ) , where U is the number of cluster centers; calculating the time complexity of hard labels as O ( b a t c h _ s i z e ) , where b a t c h _ s i z e is the number of samples in each batch; and calculating the time complexity of delta as O ( b a t c h _ s i z e × U ) . Thus, the reconstruction loss time complexity can be calculated as O ( b a t c h _ s i z e × U ) and the clustering loss time complexity as O ( P ) , so the time complexity of the loss function is O ( b a t c h _ s i z e × U + P ) . For the third part, parameter updates and gradient updates have a time complexity of O ( P ) . The time complexity of updating cluster layer weights is O ( U ) , so the time complexity of parameter updates is O ( P + U ) . Consequently, the complexity of the proposed algorithm in each iteration is O ( 3 P + b a t c h _ s i z e × U ) .
Algorithm 1: Sparse clustering algorithm based on multi-domain autoencoder.
Data: Random 80 % of training dataset X.
Result: Clustered loss L c l u W , reconstruction loss L r e c W , N M I , A R I , A C C , A M I , H o m o g e n e i t y , C o m p l e t e .
Mathematics 12 01526 i029

4. Experimental Results and Analysis

4.1. Experimental Setup and Datasets

In this experiment, we implement all the results on a 3-node, 90-core ThinkSystem SD630 306 V2 with Python 3.9 and PyTorch 2.0.1. The proposed method is evaluated using three datasets from the torchvision library: MNIST, CIFAR-10, and CIFAR-100 datasets. Specifically, the MNIST dataset consists of a widely used collection of handwritten digit images, including digits 0 to 9, with a total of 70,000 images divided into 10 classes. Each image is a 28 × 28 pixel grayscale image represented by a handwritten digit. The CIFAR-10 dataset is comprised of 60,000 RGB images, each measuring 32 × 32 pixels, and assigned to one of ten different classes. Similar to CIFAR-10, the CIFAR-100 dataset contains a set of 60,000 RGB images of size 32 × 32 × 3 from 100 fine-grained classes, which belong to 20 coarse-grained super-classes. The corresponding description is given in Table 1. It is worth noting that we conduct experiments on the 20 super-classes of CIFAR-100 and compare the performance on 20 super-classes.
Table 1. The used datasets.
The SMDR model utilizes a convolutional autoencoder to derive 10-dimensional features from the input data, which are subsequently employed for clustering purposes at the clustering layer. The encoder is made up of four convolutional layers conv(1,16,3,1,1)-conv(16,32,3,2,1)-conv(32,32,3,1,1)-conv(32,16,3,2,1), and then is followed by two fully connected layers fc(256)-fc(10). Here, conv(1,16,3,1,1) denotes the configuration of a convolutional layer with 1 input channel, 16 output channels, a convolution kernel size of 3, a stride of 1, and a padding of 1. Additionally, the notation fc(256) denotes a fully connected layer comprising 256 neurons. Batch normalization is implemented following each convolutional layer, with ReLU activation applied after each layer except for the final fully connected layer. The decoder component functions in an inverse manner to the encoder. The clustering layer is a fully connected layer that allows for the selection of the desired number of clusters. Both the convolutional autoencoder and the clustering layer are initialized randomly with Kaiming uniform [32], and then they are trained together for 3000 epochs using the default Adadelta [33] optimizer. In addition, there are no hyper-parameters that need to be tuned in our proposed method.

4.2. Experimental Results and Analysis

We adopt seven evaluation metrics to assess the experimental results, namely, Accuracy ( A C C ), Adjusted Rand Index ( A R I ), Normalized Mutual Information ( N M I ), Adjusted Mutual Information ( A M I ), Homogeneity ( H o m o . ), Completeness ( C o m p . ), and V-Measure ( V _ M e a . ). An important point to note is that higher values are preferable for all seven metrics. A C C is the proportion of accurately predicted instances compared to the total number of instances. A R I reflects the degree of overlap between two partitions. The Rand Index ( R I ) is defined as
R I = a + b n 2 n 2 = C n 2 ,
where a represents the number of instance pairs that are clustered in the same category C and in the same cluster K, and b is defined as the number of instance pairs that are categorized differently in C and clustered differently in K. Based on this, we define A R I as
A R I = R I E R I max R I E R I .
N M I is used to measure the similarity of the clustering results. Mutual Information ( M I ) is defined as
M I X ; Y = x y P x , y log P x , y P x P y ,
where the joint distribution of two random variables ( X , Y ) is denoted as P ( x , y ) , and their marginal distributions are represented as P ( x ) and p ( y ) , respectively:
N M I X ; Y = M I X , Y F H X , H Y ,
H ( x ) represents the entropy of information, and F ( x , y ) can be the min, max functions, geometric mean function, or arithmetic mean function. A M I is denoted as
A M I X , Y = M I X , Y E M I X , Y F H X , H Y E M I X , Y .
H o m o . measures whether a clustering algorithm divides the data into homogeneous groups, meaning clusters consist solely of samples from the same category:
H o m o . = 1 H C | K H C .
C o m p . is used to measure whether a clustering algorithm’s partitioning of data is complete, meaning whether samples from the same category are assigned to the same cluster:
C o m p . = 1 H K | C H K .
V _ M e a . is a comprehensive evaluation metric that takes into account both homogeneity and completeness, and it can be used to assess the quality of a clustering algorithm’s partitioning of data:
V _ M e a . = 1 + β × H o m o . × C o m p . β × H o m o . + C o m p . .

4.2.1. Experimental Comparisons

At the beginning of the experiment, we compare SMDR with seven popular clustering methods on the MNIST dataset. These methods include one deep learning-based clustering method (TELL [7]), two large-scale clustering methods (SLRR [34], LSC [35]), three subspace clustering methods (SC [36], LRR [37], LSR [38]), and vanilla k-means. It should be noted that we adjusted the hyperparameters of the methods being compared in the experiments. The results of the comparison are displayed in Table 2.
Table 2. Clustering performance on MNIST dataset.
The clustering performance on the MNIST dataset displayed in Table 2, where bold numbers indicate data in which this model exceeds others, shows that SMDR outperforms the other clustering methods in all three evaluation metrics, demonstrating superior clustering performance on MNIST. Since TELL outperforms other clustering methods in the evaluation metrics, a more in-depth comparison with TELL will be conducted on the MNIST, CIFAR10, and CIFAR100 datasets to better understand the performance of SMDR.

4.2.2. The Comparison Results of SMDR and TELL on the MNIST Dataset

The comparison results of the clustering centers of SMDR and TELL on the MNIST dataset are shown in Table 3. During the training phase, both models are trained to identify the clustering centers of the MNIST dataset. Nevertheless, the TELL model exhibits an inability to accurately capture the entirety of these clustering centers. For instance, it is shown that at Epoch = 50, the model TELL encounters difficulty in accurately distinguishing between the digits “1”, “4”, and “9”, while the model SMDR demonstrates only a slight differentiation issue between the digits “4” and “9”. Although SMDR initially fails to separate “4” and “9”, it successfully distinguishes them when Epoch = 1000. During the training phase, both models are trained to identify the clustering centers of the MNIST dataset. Nevertheless, the TELL model demonstrates an inability to accurately capture the entirety of these clustering centers. For instance, it is shown that at Epoch = 50, the model TELL encounters difficulty in accurately distinguishing between the digits “1”, “4”, and “9”, while the model SMDR exhibits only a slight differentiation issue between the digits “4” and “9”. Notably, in this situation, TELL still has not managed to differentiate “1”, “4”, and “9”. Therefore, the proposed model achieves the global optimal faster compared to TELL.
Table 3. Comparison of clustering centers for SMDR and TELL [7] on the MNIST dataset.
Table 4 presents a comparison of the clustering performance achieved by SMDR and TELL algorithms on the MNIST dataset. Both models successfully partition the dataset into ten clusters. However, it is shown that after multiple iterations, the SMDR clustering performance plot shows larger distances between clusters and tighter intra-cluster cohesion. In this situation, SMDR outperforms TELL in terms of clustering quality.
Table 4. Comparison of clustering performance for SMDR and TELL on the MNIST dataset.
The evaluation indicators of SMDR on MNIST are all higher than those of TELL as shown in the data in Table 5. Although the increase in some indicators is not obvious, the value of other indicators is greatly improved. As far as ARI is concerned, SMDR exceeds TELL by 8.3%, which means that the clustering results are more consistent with the labels, indicating that our proposed clustering algorithm has better performance. In particular, ACC is 10.9% higher than TELL, and the prediction accuracy is greatly improved, which can prove the superiority of the SMDR model.
Table 5. Clustering performance for SMDR and TELL on the MNIST dataset. Bold numbers indicate data in which this model outperforms other models.

4.2.3. The Comparison Results of SMDR and TELL on CIFAR Dataset

The clustering performance of SMDR and TELL on the CIFAR-10 dataset is presented in Table 6. It can be observed from Table 6 that the clustering evaluation metrics of SMDR and TELL on CIFAR-10 are not significantly different, with SMDR only surpassing TELL in terms of accuracy. For CIFAR-100, which has higher data complexity, we conduct clustering on its 20 super-classes, and the comparison of clustering accuracy is shown in Table 7. It is not difficult to see that SMDR has higher accuracy than TELL on super-classes such as S6 and S7. However, the difference in accuracy between the two methods is not significant in other super-classes. This demonstrates the excellent performance of SMDR on complex datasets.
Table 6. Clustering performance of SMDR and TELL on CIFAR-10 dataset. Bold numbers indicate data in which this model outperforms other models.
Table 7. Clustering accuracy on the 20 super-classes of the CIFAR-100 dataset. Bold numbers indicate data in which this model outperforms other models.

4.3. Ablation Study

To gain a more comprehensive understanding of the effectiveness and robustness of the method we proposed, we conduct three ablation studies based on MNIST. These studies involve evaluating the performance of SMDR with different model structures, loss functions, and optimizers.

4.3.1. Influence of Model Architecture

In previous experiments, SMDR relied on a convolutional autoencoder for feature extraction. To verify the robustness of our proposed method, we conduct comparative experiments between SMDR based on a fully connected autoencoder and vanilla k-means. The structure of the fully connected autoencoder is fc(500)-fc(500)-fc(2000)-fc(10), with ReLU activation used for each fully connected layer except the last one. The decoder is the reverse of the encoder.
Table 8 shows the comparison of clustering results of SMDR on MNIST based on different model structures. It is apparent that both the FCN-based SMDR and the CNN-based SMDR outperform their corresponding vanilla k-means in terms of evaluation metrics. Furthermore, due to the powerful feature extraction capability of the convolutional autoencoder, the CNN-based SMDR performs better in clustering than the FCN-based SMDR. This further highlights the advantage of SMDR in feature extraction.
Table 8. Influence of model architecture on the MNIST dataset. Bold numbers indicate data in which this model outperforms other models.

4.3.2. Impact of Loss Function

In order to showcase the effectiveness of the sparse L1 regularization method that was previously proposed, we compare the performance of SMDR on the loss function with and without incorporating sparse L1 regularization. From Table 9, it is clear that the role of sparse L1 regularization in clustering is undeniable. This also proves that sparse L1 regularization can make some cluster weights in the clustering layer approach 0, thus making the clustering results more accurate.
Table 9. Impact of loss function on the MNIST dataset. Bold numbers indicate data in which this model outperforms other models.

4.3.3. Influence of Optimizer

In addition to the two ablation experiments mentioned above, we implement further evaluations of SMDR by employing different optimizers. In the experiment, we compare two popular optimizers (Adadelta [33], Adam [39]), both with default parameters. As shown in Table 10, it is indicated that SMDR using the Adadelta optimizer outperforms SMDR using the Adam optimizer.
Table 10. Influence of optimizer on the MNIST dataset. Bold numbers indicate data in which this model outperforms other models.

5. Conclusions

In this paper, we propose a sparse clustering model based on the multi-domain dimensionality reduction model to address the challenges in clustering high-dimensional data. The model synthetically learns the spatial and frequency domain invariant information in the data. Spatial information can provide useful information about the relationships between data points. Furthermore, we use discrete wavelet transform to project the spatial data into the frequency domain to provide local and frequency information. We integrate the sparse reconstruction process and sparse L1 regularization into the automatic encoder to complete the feature selection of high-dimensional data. After effective dimensionality reduction, the basic information of the original data is not lost, ensuring the accuracy of clustering. The sparse L1 regularization effectively increases the sparsity of the cluster layer weights, and improves the generalization performance and stability of the model. Finally, experimental results on MNIST, CIFAR10, and CIFAR100 database demonstrate the effectiveness of the proposed approach. In the future, we will focus on clustering on both high-dimensional and massive data.

Author Contributions

Conceptualization, H.Z.; Formal analysis and investigation, H.Z., Y.K., E.L., K.Z. and X.W.; writing—review and editing, H.Z.; supervision, H.Z.; methodology, Y.K., E.L., K.Z. and X.W.; writing—original draft preparation, Y.K., E.L., K.Z. and X.W. All authors agree to the final proofread version. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Natural Science Foundation of Shandong Province (Project No. ZR2023MF003).

Data Availability Statement

The data that support the findings of this study are available on request from the corresponding author, Huaqing Zhang, upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yu, Y.; Liu, J. SCM enables improved single-cell clustering by scoring consensus matrices. Mathematics 2023, 11, 3785. [Google Scholar] [CrossRef]
  2. Sun, C.; Shao, Q.; Zhou, Z.; Zhang, J. An enhanced FCM clustering method based on multi-strategy tuna swarm optimization. Mathematics 2024, 12, 453. [Google Scholar] [CrossRef]
  3. Di Nuzzo, C. Advancing spectral clustering for categorical and mixed-type data: Insights and applications. Mathematics 2024, 12, 508. [Google Scholar] [CrossRef]
  4. Huang, D.; Wang, C.D.; Lai, J.H.; Kwoh, C.K. Toward multidiversified ensemble clustering of high-dimensional data: From subspaces to metrics and beyond. IEEE Trans. Cybern. 2022, 52, 12231–12244. [Google Scholar] [CrossRef] [PubMed]
  5. Jia, H.; Cheung, Y.M. Subspace clustering of categorical and numerical data with an unknown number of clusters. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 3308–3325. [Google Scholar] [CrossRef] [PubMed]
  6. Hou, C.; Nie, F.; Yi, D.; Tao, D. Discriminative embedded clustering: A framework for grouping high-dimensional data. IEEE Trans. Neural Netw. Learn. Syst. 2015, 26, 1287–1299. [Google Scholar] [CrossRef] [PubMed]
  7. Peng, X.; Li, Y.; Tsang, I.W.; Zhu, H.; Lv, J.; Tianyi Zhou, J. XAI beyond classification: Interpretable neural clustering. J. Mach. Learn. Res. 2018, 23, 227–254. [Google Scholar] [CrossRef]
  8. Jing, L.; Ng, M.K.; Huang, J.Z. An entropy weighting k-means algorithm for subspace clustering of high-dimensional sparse data. IEEE Trans. Knowl. Data Eng. 2007, 19, 1026–1041. [Google Scholar] [CrossRef]
  9. Castelli, V.; Thomasian, A.; Chung-Sheng, L. CSVD: Clustering and singular value decomposition for approximate similarity search in high-dimensional spaces. IEEE Trans. Knowl. Data Eng. 2003, 15, 671–685. [Google Scholar] [CrossRef]
  10. Almalawi, A.M.; Fahad, A.; Tari, Z.; Cheema, M.A.; Khalil, I. k NNVWC: An efficient k -nearest neighbors approach based on various-widths clustering. IEEE Trans. Knowl. Data Eng. 2016, 28, 68–81. [Google Scholar] [CrossRef]
  11. Ordonez, C.; Omiecinski, E. Efficient disk-based k-means clustering for relational databases. IEEE Trans. Knowl. Data Eng. 2004, 16, 909–921. [Google Scholar] [CrossRef]
  12. Rathore, P.; Kumar, D.; Bezdek, J.C.; Rajasegarar, S.; Palaniswami, M. A rapid hybrid clustering algorithm for large volumes of high dimensional data. IEEE Trans. Knowl. Data Eng. 2019, 31, 641–654. [Google Scholar] [CrossRef]
  13. Guan, J.; Li, S.; Chen, X.; He, X.; Chen, J. DEMOS: Clustering by pruning a density-boosting cluster tree of density mounts. IEEE Trans. Knowl. Data Eng. 2023, 35, 10814–10830. [Google Scholar] [CrossRef]
  14. Zhao, X.; Nie, F.; Wang, R.; Li, X. Robust fuzzy k-means clustering with shrunk patterns learning. IEEE Trans. Knowl. Data Eng. 2023, 35, 3001–3013. [Google Scholar] [CrossRef]
  15. Yang, X.; Deng, C.; Dang, Z.; Tao, D. Deep multiview collaborative clustering. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 516–526. [Google Scholar] [CrossRef]
  16. Xu, J.; Ren, Y.; Tang, H.; Yang, Z.; Pan, L.; Yang, Y.; Pu, X.; Yu, P.S.; He, L. Self-supervised discriminative feature learning for deep multi-view clustering. IEEE Trans. Knowl. Data Eng. 2023, 35, 7470–7482. [Google Scholar] [CrossRef]
  17. Zhao, Y.; Li, X. Spectral clustering with adaptive neighbors for deep learning. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 2068–2078. [Google Scholar] [CrossRef]
  18. Guan, R.; Zhang, H.; Liang, Y.; Giunchiglia, F.; Huang, L.; Feng, X. Deep feature-based text clustering and its explanation. IEEE Trans. Knowl. Data Eng. 2022, 34, 3669–3680. [Google Scholar] [CrossRef]
  19. Li, S.; Liu, F.; Jiao, L.; Chen, P.; Li, L. Self-supervised self-organizing clustering network: A novel unsupervised representation learning method. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 1857–1871. [Google Scholar] [CrossRef]
  20. Huang, Z.; Chen, J.; Zhang, J.; Shan, H. Learning representation for clustering via prototype scattering and positive sampling. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 7509–7524. [Google Scholar] [CrossRef]
  21. Wang, Z.; Ni, Y.; Jing, B.; Wang, D.; Zhang, H.; Xing, E. DNB: A joint learning framework for deep bayesian nonparametric clustering. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 7610–7620. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, T.; Zhang, X.; Lan, L.; Luo, Z. Local-to-global deep clustering on approximate uniform manifold. IEEE Trans. Knowl. Data Eng. 2023, 35, 5035–5046. [Google Scholar] [CrossRef]
  23. Chang, J.; Meng, G.; Wang, L.; Xiang, S.; Pan, C. Deep self-evolution clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 809–823. [Google Scholar] [CrossRef] [PubMed]
  24. Ji, Q.; Sun, Y.; Gao, J.; Hu, Y.; Yin, B. A decoder-free variational deep embedding for unsupervised clustering. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 5681–5693. [Google Scholar] [CrossRef] [PubMed]
  25. Li, X.; Zhang, R.; Wang, Q.; Zhang, H. Autoencoder constrained clustering with adaptive neighbors. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 443–449. [Google Scholar] [CrossRef] [PubMed]
  26. Yang, L.; Fan, W.; Bouguila, N. Deep clustering analysis via dual variational autoencoder with spherical latent embeddings. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 6303–6312. [Google Scholar] [CrossRef] [PubMed]
  27. Wu, L.; Yuan, L.; Zhao, G.; Lin, H.; Li, S.Z. Deep clustering and visualization for end-to-end high-dimensional data analysis. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 8543–8554. [Google Scholar] [CrossRef] [PubMed]
  28. Wu, S.; Zheng, W.S. Semisupervised feature learning by deep entropy-sparsity subspace clustering. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 774–788. [Google Scholar] [CrossRef]
  29. Wang, S.; Chen, Z.; Du, S.; Lin, Z. Learning deep sparse regularizers with applications to multi-view clustering and semi-supervised classification. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 5042–5055. [Google Scholar] [CrossRef]
  30. Cheng-Ru, L.; Ken-Hao, L.; Ming-Syan, C. Dual clustering: Integrating data clustering over optimization and constraint domains. IEEE Trans. Knowl. Data Eng. 2005, 17, 628–637. [Google Scholar] [CrossRef]
  31. Werner, D.H.; Allard, R.J. The simultaneous interpolation of antenna radiation patterns in both the spatial and frequency domains using model-based parameter estimation. IEEE Trans. Antennas Propag. 2000, 48, 383–392. [Google Scholar] [CrossRef]
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imageNet classification. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar] [CrossRef]
  33. Zeiler, M.D. Adadelta: An adaptive learning rate method. arXiv 2012. [Google Scholar] [CrossRef]
  34. Peng, X.; Tang, H.; Zhang, L.; Yi, Z.; Xiao, S. A unified framework for representation-based subspace clustering of out-of-sample and large-scale data. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 2499–2512. [Google Scholar] [CrossRef] [PubMed]
  35. Cai, D.; Chen, X. Large scale spectral clustering via landmark-based sparse representation. IEEE Trans. Cybern. 2015, 45, 1669–1680. [Google Scholar] [CrossRef] [PubMed]
  36. Ng, A.; Jordan, M.; Weiss, Y. On spectral clustering: Analysis and an algorithm. Adv. Neural Inf. Process. Syst. 2001, 14, 849–856. [Google Scholar]
  37. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; Ma, Y. Robust recovery of subspace structures by low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 171–184. [Google Scholar] [CrossRef]
  38. Lu, C.Y.; Min, H.; Zhao, Z.Q.; Zhu, L.; Huang, D.S.; Yan, S. Robust and efficient subspace segmentation via least squares regression. In Proceedings of the Computer Vision-ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Part VII 12. pp. 347–360. [Google Scholar] [CrossRef]
  39. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.