Next Article in Journal
A K-Means Approach to Temperature Estimation in Non-Linear Power Inductors
Previous Article in Journal
Positional Component-Guided Hangul Font Image Generation via Deep Semantic Segmentation and Adversarial Style Transfer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Feature-Scheduling Aggregation Clustering Framework Based on Convolutional Neural Networks

1
School of Computer and Information, Anhui Normal University, Wuhu 241002, China
2
Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei 230088, China
*
Authors to whom correspondence should be addressed.
Electronics 2025, 14(13), 2700; https://doi.org/10.3390/electronics14132700
Submission received: 8 May 2025 / Revised: 25 June 2025 / Accepted: 2 July 2025 / Published: 4 July 2025

Abstract

This study presents a convolutional neural network (CNN)-based feature-scheduling aggregation clustering framework designed to address the limitations of conventional clustering algorithms in handling high-dimensional data structures. The proposed framework synergistically combines CNN’s automated feature extraction with adaptive feature-scheduling mechanisms and multi-level feature aggregation, enabling the effective capture of nonlinear data relationships and fine-grained pattern distinctions. The experimental results highlight the framework’s superior performance against existing algorithms, alongside its strong generalization capability. This framework advances cluster analysis methodology by providing an integrated deep learning architecture for complex data scenarios.

1. Introduction

In the era of big data, the exponential growth of multidimensional datasets (e.g., hyperspectral imagery, biomedical signals, and social networks) has created unprecedented challenges in discovering intrinsic data patterns and actionable insights. Traditional clustering methods, such as k-means and DBSCAN, often struggle with high-dimensional data due to their reliance on handcrafted similarity metrics and linear assumptions [1,2]. Cluster analysis, as a cornerstone of unsupervised learning, requires advanced techniques to address nonlinear separability and hierarchical feature interactions in complex data structures [3,4].
The advent of deep learning has revolutionized feature representation learning, particularly through convolutional neural networks (CNNs). CNNs excel in capturing spatial hierarchies and translation-invariant features, as evidenced by their success in image segmentation and time-series analysis [5]. Recent studies have integrated CNNs with clustering objectives, such as DeepCluster and Variational Deep Embedding, demonstrating superior performance in learning disentangled representations for unlabeled data [6,7]. However, these approaches often neglect dynamic feature prioritization and multi-scale aggregation, limiting their adaptability to heterogeneous datasets [8].
To address these gaps, this study proposes a clustering framework based on CNN for feature scheduling and aggregation, aiming to enhance the performance of clustering tasks using deep learning techniques. The framework integrates the feature extraction capabilities of CNN, the adaptive adjustment of feature scheduling, and the multi-band information integration of feature aggregation, in order to achieve better clustering results. Through experimental validation on multiple datasets, this paper demonstrates the superiority of the proposed framework compared to existing clustering algorithms and evaluates its generalization ability.
This paper is structured as follows: Section 2 introduces the related works. Section 3 delineates the overall framework architecture and the implementation details of the CNN backbone, feature scheduling mechanism, and feature aggregation module. Section 4 describes the experimental design, including datasets, evaluation metrics, and parameter configurations; following that, we present comparative results and provide an in-depth analysis of the findings. Finally, Section 5 concludes the paper and explores future research directions.

2. Related Works

The integration of deep learning with cluster analysis has emerged as a research frontier, stimulating numerous methodological innovations. Current approaches can be categorized based on their architectural paradigms, as follows.
Transformer-based methods dominate recent advancements. Clusterformer [9] implements iterative cluster center updates through cyclic attention mechanisms, enabling robust representation learning. Contrastive learning frameworks like DeepCluster [10] employ standard K-Means clustering on feature spaces, using partition results as supervisory signals for network optimization. The IIC (Invariant Information Clustering) [11] framework pioneers unsupervised neural classifiers that discover semantically meaningful clusters without labeled data. Geometric processing techniques such as PCT [12] enhance input embedding through farthest point sampling and nearest neighbor search, while SETR [13] establishes global context modeling via transformer layers coupled with dedicated decoders.
Parallel developments in computer vision introduce specialized architectures. SGSCN [14] achieves the joint optimization of feature representation and cluster assignment through iterative end-to-end learning, where spatially constrained loss functions enforce connectivity priors. For point cloud processing, DGCNN [15] proposes permutation-invariant operations preserving local geometric structures. CurveNet [16] models point clouds as undirected graphs, establishing long-range dependencies through curvilinear connections in shallow networks. Large-scale graph analysis benefits from Dink-Net [17], which efficiently captures complex structural patterns. The GDANet [18] architecture introduces geometric decoupling modules that dynamically separate 3D objects into contour (sharp) and planar (gradual) components, capturing complementary geometric semantics. CRM [19] captures invariant target information and suppresses the heterogenous differences in cross-drone information.
In multi-view clustering, MVEC [20] employs a multi-view deep embedding clustering framework designed for multi-density datasets, enhancing clustering accuracy and precision. CSMDC [21] proposes a deep multi-view document clustering framework that captures consistent contextual semantics to boost performance. FPGC [22] addresses graph clustering from a feature perspective through squeeze-and-excitation blocks that select clustering-relevant features, providing novel theoretical foundations for graph clustering research. MWDSC [23] employs multi-dimensional weighted deep subspace clustering to enhance the efficiency and effectiveness of feature extraction and clustering performance.
The MMOEA_DC_HR [24] algorithm, based on hierarchical clustering, conducts dual clustering within the decision and objective spaces. The multi-view C-means clustering algorithm [25] extends NCM clustering to handle multi-view data uncertainty while auto-determining view importance. By combining feature weight learning with the assessment of feature contributions, the ECM [26] enhances the accuracy and reliability of clustering outcomes. CGC-AF [27] enhances graph representation learning through contrastive frameworks and optimizes graph structures with adaptive filters.
MVC-DMLR [28] employs deep non-negative matrix factorization to learn multi-level representations, integrating feature learning and multi-view clustering for accuracy improvement. SPCNet [29] simplifies PCNet while enhancing robustness, providing new theoretical foundations for GNNs handling graphs with varying homophily levels. Co-MSE [30] optimizes graph embeddings by integrating multi-order correlations to address high-order dependency issues. DPSC [31] proposes a novel affinity learning framework to enhance diffusion process robustness. RGSL [32], a robust graph structure learning method for heterophilous graphs, offers new theoretical insights. SCDCMV [33] improves clustering performance by leveraging inter-view consistency while reducing data noise impacts.
Other related studies consist of PRANet [34], focusing on combining structural and relational learning for discriminative regional feature extraction, and a meta-heuristic [35] approach designed to partition datasets into distinct, non-overlapping clusters. This partitioning aims to minimize similarities within clusters and maximize differences between clusters. These works collectively advance cluster analysis through theoretical and methodological innovations.
However, cluster analysis for complex, high-dimensional data remains a significant challenge. Traditional machine learning approaches often struggle with scalability and performance when confronted with large-scale, high-dimensional datasets. While deep learning, particularly deep clustering methods utilizing autoencoders or specific architectures, has shown promise by automating feature learning, significant limitations persist. These methods face challenges when processing highly heterogeneous data where features exhibit diverse scales, types, and distributions.
Addressing these interconnected challenges, this study introduces a novel convolutional neural network (CNN)-based feature-scheduling aggregation clustering framework. Unlike traditional methods, the proposed framework leverages CNN’s strength in automated high-dimensional feature extraction directly from raw or complex structured data. Crucially, it integrates adaptive feature-scheduling mechanisms and multi-level feature aggregation modules within an end-to-end trainable architecture. This integrated design directly tackles the training instability inherent in alternating optimization by jointly optimizing feature learning and clustering objectives. The framework is engineered to inherently capture complex, nonlinear relationships and fine-grained pattern distinctions. Moreover, by its design and the aggregation mechanism, it implicitly promotes geometric consistency in the learned embedding space, better preserving data topology. The adaptive nature of the feature scheduling and aggregation also aims to provide enhanced multi-density adaptability, allowing the model to respond effectively to clusters with varying local densities without relying on fixed, global parameters.

3. Proposed Method

The convolutional neural network (CNN)-based feature-scheduling aggregation clustering framework comprises three core modules: (1) convolutional neural network (CNN), (2) feature scheduling, and (3) feature aggregation. The CNN initially extracts discriminative features from input data, which are then transmitted to the feature scheduling module. This secondary module performs sample-wise classification based on both raw data instances and extracted feature representations. In the final step, the feature aggregation module integrates high-frequency and low-frequency information within each data cluster to generate a unified and informative feature set. The complete architecture is visually summarized in Figure 1, which explicitly delineates inter-modular connections and functional relationships.

3.1. Convolutional Neural Networks

The convolutional layer adopts 3 × 3 kernels to maintain a balance between computational complexity and model performance. With stride = 1 and ‘same’ padding configuration, spatial dimensions of outputs are preserved relative to inputs. The initial layer contains 32 kernels, with quantity progressively increasing across network depth. Each convolutional operation is followed by ReLU activation to enhance nonlinear representation capabilities and helps mitigate gradient vanishing and exploding issues during training. Its expression is shown as follows:
f x = max 0 , x
Pooling Layer: Average pooling with 2 × 2 window (stride = 2, zero-padding) reduces feature map dimensionality while preserving critical information. This operation smooths activations through regional averaging, effectively controlling subsequent computational loads.
Model Training: Training involves forward propagation for prediction generation and loss computation, followed by backpropagation for parameter optimization. The mean-squared-error loss function quantifies discrepancies between predictions and ground truth. Cross-entropy loss is utilized as the loss function to optimize the model’s performance and its expression is shown as follows:
L = 1 N i y i · log p i + 1 y i · log 1 p i
Here, p i denotes the feature vector of the i-th input data point, y i denotes the predicted probability of point i, and N denotes the number of data points in batch.
To prevent overfitting in fully connected layers, dropout regularization (p = 0.35) randomly deactivates neurons during training, forcing robust feature learning. The Adam optimizer accelerates convergence through adaptive moment estimation, combining momentum and RMSprop principles for stable gradient updates.

3.2. Feature Scheduling

Feature scheduling constitutes an adaptive feature refinement mechanism. During clustering iterations, this module performs sample classification by evaluating cosine similarity between feature embedding and cluster centroids. For each feature embedding, the mechanism undertakes the following: (1) it computes similarity scores with all centroids, (2) it applies adaptive weighting to the embedding vector, and (3) it enhances sample–cluster relationships through geometric alignment, here referring to the optimization process that minimizes spatial discrepancies between feature distributions across different modalities. This dynamic optimization tightens intra-cluster cohesion while increasing inter-cluster separation in the latent space, ultimately improving clustering precision and noise robustness.
Following feature extraction, an adaptive weighting mechanism processes cluster patches based on their centroid similarity. This enables coherent structural understanding through context-aware feature recalibration. The updated patch embedding for each p i is calculated using the formula as follows:
    p i = p i + M L P ( 1 K k = 0 K s i m ( C k , p i ) C k )
Here, p i denotes the feature vector of the i-th input data point, K represents the total number of cluster centers, and s i m ( C k , p i ) denotes the similarity between the k-th cluster center and the input data point.
The scheduling mechanism refines feature embedding by integrating centroid-derived semantic information. Through iterative similarity-driven adjustments, the model develops enhanced comprehension of latent data structures and contextual relationships, enabling more effective capture of intrinsic data patterns.

3.3. Feature Aggregators

Feature aggregators implement local feature integration through a hybrid pooling strategy with parallel high-/low-frequency branches. When processing local feature patches, these modules extract complementary information via two distinct pathways:
  • High-frequency branch: Utilizes max pooling to capture salient features and accentuate data variations/detail patterns.
  • Low-frequency branch: Applies average pooling to derive global trend representations and background characteristics.
The concatenated outputs from both branches are subsequently processed by a multilayer perceptron (MLP) to generate final feature representations. This dual-branch architecture enables the comprehensive capture of spectral information while preserving inter-feature correlations, thereby enhancing the model’s capacity to handle data complexity in clustering tasks.
The hybrid pooling strategy improves upon conventional approaches through parallel processing rather than serial combination. Unlike traditional max pooling, which primarily aggregates local high-frequency components, our design maintains separate information pathways:
  • High-frequency signals are emphasized via max pooling;
  • Low-frequency components are distilled through average pooling.
High-frequency aggregation branch: To aggregate high-frequency information in the high-frequency aggregation branch, max pooling is utilized. The expression is presented as follows:
f h = R e s M L P M a x P o o l f g
Here, M a x P o o l stands for maximum pooling, and R e s M L P represents the residual MLP block.
Low-frequency aggregation branch: Within the low-frequency aggregation branch, average pooling is implemented to direct the module’s attention towards low-frequency information. The formulation for this branch is
f l = C o n T A v g P o o l f g
Through the high-frequency aggregation branch and the low-frequency aggregation branch, we can obtain the high-frequency and low-frequency information, respectively. Finally, the features from these two branches are concatenated and jointly input into the MLP to yield the output feature   f . The expression is as follows:
f = M L P f h , f l

4. Experiments

4.1. Datasets and Evaluation Metrics

Three benchmark datasets from the UCI Machine Learning Repository are selected for clustering evaluation:
  • Wisconsin Breast Cancer Dataset: Contains 569 instances with 30 features, classified into benign/malignant categories;
  • Heart Disease Dataset: Contains 303 instances containing 13 diagnostic features;
  • IRIS Dataset: Contains 150 samples with 4 morphological features across three species classes.
To fully measure the clustering performance of the model, two evaluation metrics are adopted.
Sum of Squared Errors (SSE): Quantifies intra-cluster cohesion through squared Euclidean distances to centroids (lower values preferred). The calculation formula is as follows:
S S E = i = 1 n j = 1 k w i j x i μ j 2
Here, n represents the total number of observations, k denotes the number of clusters, x i is the i-th observation, μ j is the centroid of the j-th cluster, and w i j takes a value of 1 if data point i is assigned to cluster j, and 0 otherwise.
Normalized Mutual Information (NMI): Measures label consistency between predicted clusters and ground truth [0, 1] range (higher values indicate better alignment). The calculation formulas are as follows:
N M I A , B = 2   I A ; B H A + H B
where I(A; B) denotes the mutual information between A and B, and H(A) and H(B) represent the entropy of A and B, respectively.

4.2. Experimental Parameter Setting

The experimental parameters, including the configuration of the model and the settings used during training and evaluation, are detailed in Table 1. This transparency in reporting is crucial for replicating the experiments and for understanding the context in which the results were achieved.

4.3. Comparative Experiments

Comparative experiments were conducted on three benchmark datasets (two biomedical collections and one botanical dataset) to rigorously evaluate the proposed model’s clustering performance against established algorithms. The biomedical datasets assess domain-specific performance in healthcare contexts, while the botanical dataset validates generalization capability in cross-domain scenarios. This multi-domain evaluation framework enables a systematic assessment of clustering efficacy across heterogeneous data types.
The comparative models were systematically selected based on recency, representativeness, and technical relevance, as shown in Table 2.
(1)
Breast Cancer Dataset Analysis
The experimental performance evaluation involves comparing the efficacy of various algorithmic models in classifying breast cancer subtypes. The detailed experimental results are presented in the table below.
As shown in Table 3, our method achieves state-of-the-art performance on this dataset:
  • SSE: 1.479 × 105 (lowest among compared methods);
  • NMI: 0.593 (superior feature-target correlation capture, as shown in Figure 2);
  • This tri-metric superiority demonstrates robust classification capability for breast cancer subtyping, particularly excelling in cluster cohesion (SSE) and label consistency (NMI).
(2)
Heart Disease Dataset Analysis
In our unsupervised experimental setup, we performed cluster analysis exclusively on feature space, excluding diagnostic labels to discover potential disease subtypes. The detailed experimental results are presented in the table below.
As detailed in Table 4, the proposed model demonstrates superior clustering performance:
  • SSE: 2.13 × 105 (lowest intra-cluster variance);
  • NMI: 0.185 (optimal feature-label correlation capture, as shown in Figure 3).
This tri-fold superiority confirms the model’s effectiveness in synthesizing spectral information (high-/low-frequency components) to exploit intrinsic data relationships, outperforming comparative methods in cardiovascular pattern discovery.
(3)
IRIS Dataset Analysis
To validate the cross-domain generalization capacity of our clustering framework, we conducted extended evaluations on the Iris Species Dataset—a benchmark botanical dataset containing 150 low-dimensional samples (sepal/petal length/width measurements). This controlled experiment tests model adaptability in simplified feature spaces. Table 4, below, provides the specific experimental results.
As detailed in Table 5, the proposed model demonstrates
  • SSE: 0.95436 (lowest intra-cluster variance);
  • NMI: 0.853 (superior cluster-label alignment, as shown in Figure 4).
Significant SSE/NMI improvements confirm the model’s robust pattern discovery capability across domains.

4.4. Ablation Study

To evaluate the necessity and effectiveness of the feature scheduling module and feature aggregator (including a high-frequency aggregation branch for capturing detailed features and a low-frequency aggregation branch for focusing on broader patterns) in the model, ablation experiments were conducted on the three datasets described above. The experimental results were assessed using SSE and NMI. The specific experimental models are as follows:
  • CNN-only: The baseline model only has the CNN module.
  • No-Feature Dispatching (No-FD): The model removes the feature scheduling module while retaining the CNN and feature aggregation module.
  • No-Feature Aggregator (No-FA): The model removes the feature aggregation module while retaining the CNN and feature scheduling module.
  • No-AvgPool: The model removes the low-frequency aggregation branch while retaining the CNN, feature scheduling, and high-frequency aggregation branch modules.
Given the critical role of CNNs and the high-frequency aggregation branch in the model, they were not evaluated individually in the ablation experiments. The results of the CNN-only, No-FD, No-FA, and No-AvgPool ablation experiments are presented in Table 6.
The results of the ablation experiments presented in the table demonstrate that the performance of each experimental model across the three datasets is comprehensively evaluated. The feature scheduling module, high-frequency aggregation branch, and low-frequency aggregation branch collectively influence the clustering performance of the model, with the feature aggregation module (FA) having the most significant impact on overall model performance. Specifically, the SSE and NMI metrics are the most affected, as shown in Figure 5.
  • In all datasets, the largest performance gap between No-FA and Our Method is observed in the SSE metric.
  • Except for the heart disease dataset, removing the feature aggregation module leads to a notable reduction in the NMI value.
In summary, the feature scheduling module and the high-frequency and low-frequency aggregation branches collectively enhance the model’s clustering performance, with the feature aggregation module playing the most critical role. These findings underscore the importance of integrating these components to optimize the model’s effectiveness.

5. Conclusions

The proposed CNN-based clustering framework introduces two novel components: (1) a feature scheduling module for similarity-driven adaptive feature alignment, and (2) a dual-branch spectral aggregation architecture combining max/average pooling operators. The scheduling module enhances intra-cluster coherence through dynamic feature-space optimization, while the spectral aggregator captures complementary patterns via parallel processing of high-frequency details (max pooling) and low-frequency trends (average pooling).
Extensive evaluations across medical (breast cancer and heart datasets) and botanical (Iris) datasets demonstrate the consistent superiority of the proposed model. Furthermore, complexity analysis has been performed to lay the groundwork for model simplification and optimization. This analysis not only unveils the model’s performance limitations but also offers insights into boosting computational efficiency, enhancing its applicability in real-world scenarios.
Furthermore, complexity analysis has been performed to lay the groundwork for model simplification and optimization. This analysis not only unveils the model’s performance limitations but also offers insights into boosting computational efficiency, enhancing its applicability in real-world scenarios.

Author Contributions

Conceptualization, Z.S.; Methodology, Y.J. and B.Y.; Validation, A.J. and Y.N.; Formal analysis, P.H.; Writing—review & editing, W.L.; Supervision, K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

The work in this paper is supported in part by the University Synergy Innovation Program of Anhui Province under GXXT-2022-054 and in part by the High-Level Scientific Research Platform Construction Project of Anhui Normal University under 2023GFXK180.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shin, Y.-M.; Kim, S.-W.; Shin, W.-Y. PAGE: Prototype-Based Model-Level Explanations for Graph Neural Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 6559–6576. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, H.; Wang, D.; Zhao, W.; Lu, Z.; Jiang, X. IMCSN: An improved neighborhood aggregation interaction strategy for multi-scale contrastive Siamese networks. Pattern Recognit. 2025, 158, 111052. [Google Scholar] [CrossRef]
  3. Deng, W.; Zhao, L.; Liao, Q.; Guo, D.; Kuang, G.; Hu, D.; Pietikainen, M.; Liu, L. Informative Feature Disentanglement for Unsupervised Domain Adaptation. IEEE Trans. Multimed. 2022, 24, 2407–2421. [Google Scholar] [CrossRef]
  4. Liu, D.; Zhang, C.; Song, Y.; Huang, H.; Wang, C.; Barnett, M.; Cai, W. Decompose to Adapt: Cross-Domain Object Detection Via Feature Disentanglement. IEEE Trans. Multimed. 2023, 25, 1333–1344. [Google Scholar] [CrossRef]
  5. Ding, Y.; Zhang, Z.; Zhao, X.; Cai, Y.; Li, S.; Deng, B.; Cai, W. Self-Supervised Locality Preserving Low-Pass Graph Convolutional Embedding for Large-Scale Hyperspectral Image Clustering. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  6. Xue, Y.; Jin, G.; Shen, T.; Tan, L.; Wang, N.; Gao, J.; Wang, L. SmallTrack: Wavelet Pooling and Graph Enhanced Classification for UAV Small Object Tracking. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  7. Malhi, U.S.; Zhou, J.; Yan, C.; Rasool, A.; Siddeeq, S.; Du, M. Unsupervised Deep Embedded Clustering for High-Dimensional Visual Features of Fashion Images. Appl. Sci. 2023, 13, 2828. [Google Scholar] [CrossRef]
  8. Zhou, W.; Song, W.; Zhang, Z.; Zhang, F.; Teng, Z.; Tian, Z. Hierarchical feature aggregation with mixed attention mechanism for single-cell RNA-seq analysis. Expert Syst. Appl. 2025, 260, 125340. [Google Scholar] [CrossRef]
  9. Liang, J.; Cui, Y.; Wang, Q.; Geng, T.; Wang, W.; Liu, D. ClusterFomer: Clustering As A Universal Visual Learner. Adv. Neural Inf. Process. Syst. 2023, 36, 64029–64042. [Google Scholar]
  10. Caron, M.; Bojanowski, P.; Joulin, A.; Douze, M. Deep Clustering for Unsupervised Learning of Visual Features. In Proceedings of the Computer Vision—ECCV 2018: 15th European Conference, Proceedings, Part XIV, Munich, Germany, 8–14 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 139–156. [Google Scholar]
  11. Ji, X.; Henriques, J.F.; Vedaldi, A. Invariant Information Clustering for Unsupervised Image Classification and Segmentation. arXiv 2018, arXiv:1807.06653. [Google Scholar]
  12. Guo, M.-H.; Cai, J.-X.; Liu, Z.-N.; Mu, T.-J.; Martin, R.R.; Hu, S.-M. PCT: Point cloud transformer. Comput. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
  13. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H.S.; et al. Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers. arXiv 2020, arXiv:2012.15840. [Google Scholar]
  14. de Bruijne, M.; Cattin, P.C.; Cotin, S.; Padoy, N.; Speidel, S.; Zheng, Y.; Essert, C. A Spatial Guided Self-supervised Clustering Network for Medical Image Segmentation. In Proceedings of theMedical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, 27 Septembe–1 October 2021; Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 379–388. [Google Scholar]
  15. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic Graph CNN for Learning on Point Clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef]
  16. Xiang, T.; Zhang, C.; Song, Y.; Yu, J.; Cai, W. Walk in the Cloud: Learning Curves for Point Clouds Shape Analysis. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 915–924. [Google Scholar]
  17. Liu, Y.; Liang, K.; Xia, J.; Zhou, S.; Yang, X.; Liu, X.; Li, S.Z. Dink-Net: Neural Clustering on Large Graphs. In Proceedings of the 40th International Conference on Machine Learning, Proceedings of Machine Learning Research, Honolulu, HI, USA, 23–29 July 2023; pp. 21794–21812. [Google Scholar]
  18. Xu, M.; Zhang, J.; Peng, Z.; Xu, M.; Qi, X.; Qiao, Y. Learning Geometry-Disentangled Representation for Complementary Understanding of 3D Object Point Cloud. Proc. AAAI Conf. Artif. Intell. 2021, 35, 3056–3064. [Google Scholar] [CrossRef]
  19. Xue, Y.; Jin, G.; Shen, T.; Tan, L.; Wang, N.; Gao, J.; Wang, L. Consistent Representation Mining for Multi-Drone Single Object Tracking. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 10845–10859. [Google Scholar] [CrossRef]
  20. Kassem, H.; Hajjar, S.E.; Abdallah, F.; Omrani, H. Multi-view Deep Embedded Clustering: Exploring a new dimension of air pollution. Eng. Appl. Artif. Intell. 2025, 139, 18. [Google Scholar] [CrossRef]
  21. Bai, R.; Huang, R.; Xu, L.; Qin, Y. CSMDC: Exploring consistently context semantics for multi-view document clustering. Expert Syst. Appl. 2025, 261, 125386. [Google Scholar] [CrossRef]
  22. Xie, X.; Li, B.; Pan, E.; Guo, Z.; Kang, Z.; Chen, W. One Node One Model: Featuring the Missing-Half for Graph Clustering. In Proceedings of the AAAI Conference on Artificial IntelligenceArxiv, Pennsylvania, PA, USA, 25 February–4 March 2025; pp. 21688–21696. [Google Scholar]
  23. Chen, S.; Zhang, X.; He, Y.; Peng, L.; Ou, Y.; Xu, S. Multi-dimensional weighted deep subspace clustering with feature classification. Expert Syst. Appl. 2025, 260, 125375. [Google Scholar] [CrossRef]
  24. Gu, Q.; Niu, Y.; Hui, Z.; Wang, Q.; Xiong, N. A hierarchical clustering algorithm for addressing multi-modal multi-objective optimization problems. Expert Syst. Appl. 2025, 264, 125710. [Google Scholar] [CrossRef]
  25. Liu, Z.; Qiu, H.; Deveci, M.; Pedrycz, W.; Siarry, P. Multi-view neutrosophic c-means clustering algorithms. Expert Syst. Appl. 2025, 260, 125454. [Google Scholar] [CrossRef]
  26. Liu, Z.; Qiu, H.; Senapati, T.; Lin, M.; Abualigah, L.; Deveci, M. Enhancements of evidential c-means algorithms: A clustering framework via feature-weight learning. Expert Syst. Appl. 2025, 259, 125246. [Google Scholar] [CrossRef]
  27. Xie, X.; Chen, W.; Kang, Z.; Peng, C. Contrastive graph clustering with adaptive filter. Expert Syst. Appl. 2023, 219, 119645. [Google Scholar] [CrossRef]
  28. Dou, Z.; Peng, N.; Hou, W.; Xie, X.; Ma, X. Learning multi-level topology representation for multi-view clustering with deep non-negative matrix factorization. Neural Netw. 2025, 182, 106856. [Google Scholar] [CrossRef]
  29. Li, B.; Xie, X.; Lei, H.; Fang, R.; Kang, Z. Simplified PCNet with robustness. Neural Netw. 2025, 184, 107099. [Google Scholar] [CrossRef] [PubMed]
  30. Zhan, S.; Jiang, H.; Shen, D. Co-regularized optimal high-order graph embedding for multi-view clustering. Pattern Recognit. 2025, 157, 110892. [Google Scholar] [CrossRef]
  31. Zhu, Y.; Li, Q.; Liu, W.; Yin, C. Diffusion process with structural changes for subspace clustering. Pattern Recognit. 2025, 158, 111066. [Google Scholar] [CrossRef]
  32. Xie, X.; Chen, W.; Kang, Z. Robust graph structure learning under heterophily. Neural Netw. 2025, 185, 107206. [Google Scholar] [CrossRef]
  33. Wu, T.; Lu, G.-F. Specific and coupled double consistency multi-view subspace clustering with low-rank tensor learning. Signal Process. 2025, 229, 109803. [Google Scholar] [CrossRef]
  34. Cheng, S.; Chen, X.; He, X.; Liu, Z.; Bai, X. PRA-Net: Point Relation-Aware Network for 3D Point Cloud Analysis. IEEE Trans. Image Process. 2021, 30, 4436–4448. [Google Scholar] [CrossRef]
  35. Kaur, A.; Kumar, Y.; Sidhu, J. Exploring meta-heuristics for partitional clustering: Methods, metrics, datasets, and challenges. Artif. Intell. Rev. 2024, 57, 287. [Google Scholar] [CrossRef]
Figure 1. Architecture of the proposed framework.
Figure 1. Architecture of the proposed framework.
Electronics 14 02700 g001
Figure 2. Cross-clustering model NMI distribution on the breast cancer dataset.
Figure 2. Cross-clustering model NMI distribution on the breast cancer dataset.
Electronics 14 02700 g002
Figure 3. Cross-clustering model NMI distribution on the heart dataset.
Figure 3. Cross-clustering model NMI distribution on the heart dataset.
Electronics 14 02700 g003
Figure 4. Cross-clustering model NMI distribution on the IRIS dataset.
Figure 4. Cross-clustering model NMI distribution on the IRIS dataset.
Electronics 14 02700 g004
Figure 5. NMI distribution of the ablation study.
Figure 5. NMI distribution of the ablation study.
Electronics 14 02700 g005
Table 1. Experimental parameters.
Table 1. Experimental parameters.
Parameter NameParameter Value
Iterations100
Batch size300
OptimizerAdam
Momentum0.9
Weight0.0001
Learning Rate0.001
Warm-up10
Centers100
Table 2. Comparative model selection criteria.
Table 2. Comparative model selection criteria.
CategoryModelsSelection Rationale
Transformer-BasedClusterformer [9], DeepCluster [10], IIC [11]SOTA attention architectures for clustering
Point Cloud ProcessingPCT [12], CurveNet [16], GDANet [18]Dominant methods for geometric data
Graph-Based CVDGCNN [15], SGSCN [14], Dink-Net [17]Graph convolution benchmarks for structural data
Multi-View ClusteringFPGC [22], SPCNet [29]Top-cited multi-view frameworks (>300 citations)
Architectural InnovationsSETR [13], PRANet [34]Novel positional encoding and attention mechanisms
Table 3. Results of clustering analysis on the breast cancer dataset.
Table 3. Results of clustering analysis on the breast cancer dataset.
ModelSSENMI
Clusterformer [9]1.4983 × 1050.579
DeepCluster [10]1.5265 × 1050.423
IIC [11]1.5065 × 1050.382
PCT [12]1.5042 × 1050.368
SETR [13]1.5025 × 1050.582
SGSCN [14]1.5606 × 1050.262
DGCNN [15]1.5241 × 1050.423
CurveNet [16]1.5006 × 1050.365
Dink-Net [17]1.4991 × 1050.563
GDANet [18]1.4948 × 1050.459
FPGC [22]1.4987 × 1050.561
SPCNet [29]1.5012 × 1050.447
PRANet [34]1.4978 × 1050.517
Our method1.4790 × 1050.593
Table 4. Results of clustering analysis on the heart disease dataset.
Table 4. Results of clustering analysis on the heart disease dataset.
ModelSSENMI
Clusterformer [9]2.1432 × 1050.177
DeepCluster [10]2.2223 × 1050.151
IIC [11]2.2143 × 1050.161
PCT [12]2.1510 × 1050.153
SETR [13]2.1345 × 1050.175
SGSCN [14]2.3020 × 1050.148
DGCNN [15]2.1871 × 1050.145
CurveNet [16]2.4152 × 1050.152
Dink-Net [17]2.2371 × 1050.169
GDANet [18]2.1311 × 1050.168
FPGC [22]2.1732 × 1050.178
SPCNet [29]2.1985 × 1050.160
PRANet [34]2.1433 × 1050.179
Our method2.1283 × 1050.185
Table 5. Results of clustering analysis on the IRIS dataset.
Table 5. Results of clustering analysis on the IRIS dataset.
ModelSSENMI
Clusterformer [9]1.0021 × 1020.699
DeepCluster [10]1.0344 × 1020.694
IIC [11]1.0213 × 1020.625
PCT [12]9.7512 × 100.685
SETR [13]9.6769 × 100.723
SGSCN [14]9.7523 × 100.717
DGCNN [15]1.0122 × 1020.701
CurveNet [16]9.7001 × 100.675
Dink-Net [17]1.0173 × 1020.798
GDANet [18]9.6655 × 100.760
FPGC [22]1.0122 × 1020.785
SPCNet [29]1.0039 × 1020.710
PRANet [34]9.6523 × 100.735
Our method9.5436 × 100.853
Table 6. Results of the ablation study.
Table 6. Results of the ablation study.
DatasetModelsSSENMI
Breast CancerCNN-only3.1526 × 1050.297
No-FD2.5673 × 1050.335
No-FA2.8743 × 1050.274
No-AvgPool2.4542 × 1050.368
Our Method1.4790 × 1050.593
Heart DiseaseCNN-only3.3203 × 1050.157
No-FD2.7432 × 1050.168
No-FA2.9353 × 1050.174
No-AvgPool2.6742 × 1050.181
Our Method2.1283 × 1050.185
IRISCNN-only2.7103 × 1050.377
No-FD2.1342 × 1020.593
No-FA2.3487 × 1020.417
No-AvgPool2.1043 × 1020.601
Our Method9.5436 × 100.853
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shen, Z.; Jiao, Y.; Ji, A.; Ye, B.; Niu, Y.; Zuo, K.; Hu, P.; Li, W. A Novel Feature-Scheduling Aggregation Clustering Framework Based on Convolutional Neural Networks. Electronics 2025, 14, 2700. https://doi.org/10.3390/electronics14132700

AMA Style

Shen Z, Jiao Y, Ji A, Ye B, Niu Y, Zuo K, Hu P, Li W. A Novel Feature-Scheduling Aggregation Clustering Framework Based on Convolutional Neural Networks. Electronics. 2025; 14(13):2700. https://doi.org/10.3390/electronics14132700

Chicago/Turabian Style

Shen, Zhangyi, Yu Jiao, Aohan Ji, Bingqing Ye, Yunfei Niu, Kaizhong Zuo, Peng Hu, and Wenjie Li. 2025. "A Novel Feature-Scheduling Aggregation Clustering Framework Based on Convolutional Neural Networks" Electronics 14, no. 13: 2700. https://doi.org/10.3390/electronics14132700

APA Style

Shen, Z., Jiao, Y., Ji, A., Ye, B., Niu, Y., Zuo, K., Hu, P., & Li, W. (2025). A Novel Feature-Scheduling Aggregation Clustering Framework Based on Convolutional Neural Networks. Electronics, 14(13), 2700. https://doi.org/10.3390/electronics14132700

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop