Next Article in Journal
Edge Based Priority-Aware Dynamic Resource Allocation for Internet of Things Networks
Next Article in Special Issue
Efficient System for Delimitation of Benign and Malignant Breast Masses
Previous Article in Journal
Intelligent Reflecting Surface Assisted Secure Transmission in UAV-MIMO Communication Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Grid-Based Clustering Using Boundary Detection

School of Computer Science and Technology, Jiangsu Normal University, Xuzhou 221116, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(11), 1606; https://doi.org/10.3390/e24111606
Submission received: 14 September 2022 / Revised: 31 October 2022 / Accepted: 2 November 2022 / Published: 4 November 2022
(This article belongs to the Special Issue Pattern Recognition and Data Clustering in Information Theory)

Abstract

:
Clustering can be divided into five categories: partitioning, hierarchical, model-based, density-based, and grid-based algorithms. Among them, grid-based clustering is highly efficient in handling spatial data. However, the traditional grid-based clustering algorithms still face many problems: (1) Parameter tuning: density thresholds are difficult to adjust; (2) Data challenge: clusters with overlapping regions and varying densities are not well handled. We propose a new grid-based clustering algorithm named GCBD that can solve the above problems. Firstly, the density estimation of nodes is defined using the standard grid structure. Secondly, GCBD uses an iterative boundary detection strategy to distinguish core nodes from boundary nodes. Finally, two clustering strategies are combined to group core nodes and assign boundary nodes. Experiments on 18 datasets demonstrate that the proposed algorithm outperforms 6 grid-based competitors.

1. Introduction

Clustering is the process of grouping similar objects together [1]. As a fundamental data mining task, it can be used either independently or as a preprocessing step before other data mining tasks. Clustering plays an important role in many scientific fields [2], including earth sciences [3,4], biology [5,6,7], and economics [8,9].
Generally, clustering can be divided into five categories: partitioning [10,11], hierarchical [12,13], model-based [14,15], density-based [16,17,18], and grid-based algorithms. Partitioned clustering is designed to discover clusters in the data by optimizing a given objective function. Hierarchical clustering deals with the clustering problem by constructing a tree diagram. Model-based clustering uses a probabilistic methodology to optimize the match between some mathematical models and the data. Density-based and grid-based solutions are two closely related categories that attempt to explore the data space at a high level of granularity.
In recent decades, many grid-based clustering algorithms have been developed. In these algorithms, the data space is partitioned into a finite number of cells to form a grid structure. Clusters correspond to regions that are the connected cells with more density. As most grid-based clustering algorithms rely on calculations of cell density, these algorithms may be considered density-based. Even some grid-based clustering algorithms are developed by improving density-based clustering. Among density-based clustering algorithms, DBSCAN [19] and DPC [20] algorithms are the most widely used and have the most variants.
The computation cost of grid-based clustering is determined by the number of grid cells, independent of dataset size. Generally, grid-based clustering is more efficient than other clustering algorithms for large-scale spatial data since the number of cells is significantly smaller than the number of data points. Although grid-based clustering algorithms greatly improve computational efficiency, they still face some of the following problems.
  • They are sensitive to the parameter of density threshold, which may be difficult to obtain.
  • They may not be sufficient to achieve desired clustering results for data with varying densities.
  • They are not able to handle boundary regions between some adjacent clusters well.
We aim at alleviating the above-mentioned issues and propose a grid-based clustering using boundary detection, named GCBD. Firstly, the density estimation of nodes is defined using the standard grid structure. Secondly, GCBD uses an iterative boundary detection strategy to distinguish core nodes from boundary nodes. Finally, DBSCAN and DPC clustering strategies are combined to group core nodes and assign boundary nodes.
The rest of this paper is organized as follows. We survey the related work in Section 2. Our proposed GCBD algorithm is presented in Section 3. Section 4 presents the experimental results. Conclusions and suggestions for future work are given in Section 5.

2. Related Work

2.1. Grid-Based Clustering

In this subsection, we discuss some classical grid-based clustering algorithms. To our knowledge, Schikuta [21] introduces the first grid-based hierarchical clustering algorithm called GRIDCLUS. In GRIDCLUS, data points are designated to blocks in the grid structure such that their topological distributions are maintained. GRIDCLUS calculates the density for each block. The blocks are clustered iteratively in order of descending density to form a nested sequence of nonempty, disjoint clusters. Schikuta and Erhart [22] further propose the BANG algorithm to improve the inefficiency of GRIDCLUS in terms of grid structure size and neighbor search. Wang et al. [23] propose a statistical information grid-based clustering method (STING) to cluster spatial data. In contrast to GRIDCLUS and BANG, STING divides the spatial area into rectangular cells and uses a hierarchical grid structure for storing the cells. Using the hierarchical grid structure may generate rougher cluster boundaries, which reduces clustering quality. As a solution to the problem, Sheikholeslami et al. [24] propose WaveCluster, a grid-based and density-based clustering approach utilizing wavelet transforms. In this algorithm, wavelet transforms are applied to the spatial data feature space to detect arbitrary shape clusters at different scales.
A significant issue with grid-based algorithms is their scalability in higher dimensional data since the time complexity grows as the number of grids increases. To address the challenge, Agrawal et al. [25] invent the algorithm CLIQUE (clustering in quest). CLIQUE seeks dense rectangular cells in all subspaces with high density by applying a bottom-up scheme. Clusters are generated as the connected areas of dense cells. OptiGrid (optimal grid clustering) [26] significantly modifies CLIQUE. OptiGrid constructs the best cutting hyperplanes through a set of projections to obtain optimal grid-partitioning. The above algorithm is very sensitive to the density threshold and it is difficult to adjust this parameter.
In the grid-based clustering, variants of the DBSCAN algorithm [19] are the most closely related to our algorithm. Wu et al. [27] propose a density- and grid-based (DGB) clustering algorithm inspired by grid partitioning and DBSCAN. The DGB algorithm only needs to calculate distances between grid nodes instead of distances between data points. Therefore, the algorithm processes spatial data more efficiently. Like DBSCAN, however, the algorithm cannot recognize clusters with varying densities. Uncu et al. [28] propose a three-step clustering algorithm (called GRIDBSCAN) to address the issue. The first step ensures homogeneous density in each grid by selecting appropriate grids. In the second step, cells that have similar densities are merged. Lastly, the DBSCAN algorithm is executed. However, these algorithms are not able to handle boundary regions between some adjacent clusters well. The main difference between the two algorithms and ours is the way they divide core and boundary nodes (or cells). The two algorithms apply a fixed, global density threshold to classify core and boundary nodes (or cells). In contrast, our algorithm uses an iterative boundary detection strategy to divide core and boundary nodes.

2.2. Density-Based Clustering

This subsection discusses several density-based clustering algorithms that are most relevant to our algorithm.
DBSCAN [19] is one of the most important density-based techniques. In DBSCAN, clusters are assumed to be connected to high-density regions separated by low-density regions. In DBSCAN, the core points are points in the dense part of a cluster. The boundary points in DBSCAN are defined as points that are part of a cluster but are not surrounded by a dense neighborhood. The primary difference between DBSCAN and our algorithm is that DBSCAN is a density-based clustering while our algorithm is a grid-based clustering. Our algorithm is similar to DBSCAN as it utilizes the notions of reachability and connectivity to find the maximally connected components of nodes. Despite this, they differ in their definitions of connectivity. DBSCAN defines connectivity between points and then generates clustering for points (including core points and boundary points) according to connectivity. However, our algorithm defines the connectivity between core nodes and then generates cluster cores for the core nodes based on their connectivity.
Rodriguez and Laio [20] propose density peaks clustering (DPC), a density-based algorithm. The algorithm assumes that cluster centers are surrounded by neighbors with lower local densities and that they are at a relatively large distance from any points with a higher local density. Our algorithm and DPC are similar in how the nodes (or points) are assigned. However, they are quite different. Unlike DPC’s assignment mechanism, where each non-centered point is assigned to the same cluster as its nearest neighbor with higher density, our assignment mechanism is a multilevel-based approach, where each boundary node is assigned to the same cluster as the node with the highest density among its nearest neighbors in the inner layers.

3. Related Concepts

3.1. DBSCAN Algorithm

DBSCAN [19] is one of the most widely used density-based clustering algorithms. It can identify arbitrary-shaped clusters and clusters with noise (i.e., outliers). In DBCSAN, there are two key parameters:
  • ϵ (or eps): It is a distance threshold. Two points are considered to be neighbors if the distance between them is less than or equal to eps.
  • k (or minPts): It specifies the minimum number of neighbors within a given radius.
Based on these two parameters, DBSCAN makes several definitions:
  • Core point: A point is a core point if there are at least m i n P t s number of points (including the point itself) in its surrounding area with radius e p s .
  • Reachable: A point x q is directly reachable from x p if point x q is within distance e p s from core point x p .
  • Density-connected: Two points x p and x q are density-connected if x p is directly or transitively reachable from x q or x q is directly or transitively reachable from x p .
  • Boundary point: A point is a boundary point if it is reachable from a core point and there are less than m i n P t s number of points within its surrounding area.
  • Noise point (or outlier): A point is a noise point (or outlier) if it is not a core point and not reachable from any core points.
  • Cluster: A cluster w.r.t. e p s and m i n P t s is a non-empty maximal subset of the data set such that every pair of points in the cluster is density-connected.
To explain the notions of core point, boundary point, and noise point, we provide an example in Figure 1. Red points are core points because there are at least 4 points within their surrounding area with a radius of e p s . This area is shown with the circles in the figure. The green points are boundary points because they are reachable from a core point and have less than 4 points within their neighborhood. Reachable means being in the surrounding area of a core point. The points x 2 and x 3 have two points (including the point itself) within their neighborhood (i.e., the surrounding area with a radius of e p s ). Finally, x 4 is a noise point because it is not a core point and cannot be reached from a core point.

3.2. DPC Algorithm

Rodriguez and Laio [20] present density peaks clustering (DPC), an algorithm that combines density-based clustering algorithms with centroid-based clustering algorithms. This algorithm has its basis on the assumptions that cluster centers are surrounded by neighbors with lower local densities and that they are far away from points of higher densities. There are two important quantities to describe each point x i : its local density ρ i and its distance from the nearest larger density point δ i . The local density ρ i of x i is calculated as
ρ i = x j X χ ( d ( x i , x j ) d c ) , χ ( z ) = 1 , z < 0 0 , z 0
where d ( x i , x j ) is the distance between points x i and x j , and the “cutoff distance” d c is a user-specified parameter.
For point x i with the highest density, DPC defines δ i = m a x ( d ( x i , x j ) ) . For the other points, δ i is the minimum distance between point x i and another point x j whose ρ j is higher than ρ i . Its formula is as follows:
δ i = min x j : ρ j > ρ i ( d ( x i , x j ) )
where x j X . X denotes the whole data set.
Based on DPC’s center assumption, density peaks with large ρ - δ are manually selected as centers by observing through a decision graph (i.e., a ρ - δ plot). Subsequently, each non-center point is allocated to the same cluster as its nearest point with higher density.

4. Proposed Algorithm

In this section, we describe our proposed clustering algorithm in detail and analyze its computational complexity.

4.1. Standard Grid Structure

As with most grid-based clustering algorithms, the first step in GCBD is to create a grid structure that divides the data space into a finite number of cells. To simplify subsequent calculations, we create a standard grid structure.
Let X = { x 1 , x 2 , , x n } is a dataset consisting of n data points, where each data point has m features, i.e., x i = { x i 1 , x i 2 , , x i m } . x ¯ j and x ̲ j denote the maximum and minimum values of the j-th feature, respectively. Assume that each dimension should be divided into l equal intervals. Features in the original dataset are scaled to fall between 1 and l + 1 by using the scaling function Φ ( · ) .
Φ ( x i j ) = l · x i j x ̲ j x ¯ j x ̲ j + 1
Let A = A 1 × A 2 × × A m is the transformed feature space, where A 1 , A 2 , , A m are the domains of the features (dimensions) of A . We define the notion of a standard grid structure.
Definition 1
(Standard grid structure). A grid structure is called a standard grid structure, if the transformed feature space is divided into l intervals of equal length after each transformed feature is scaled by Equation (3).
Let v = { v 1 , v 2 , , v m } is a node in the standard grid structure. We will obtain the following property.
Property 1.
The nodes in the standard grid structure are only located with integer coordinates, i.e., for each v j , 1 j m , v j { 1 , 2 , , l + 1 } .
To explain some notions in the standard grid structure, Figure 2 provides an example in a two-dimensional space. Green-shaded rectangles are cells in the standard grid structure. Red intersection points in the standard grid are called nodes. Blue points are transformed data.

4.2. Density Estimation

In density-based clustering, the densities of data points are computed. The traditional grid-based clustering calculates the densities of cells. In the proposed algorithm, we focus on the nodes’ density.
At the t-th iteration, we define V ( t ) as the set of active nodes and X ( t ) as the set of active points. To estimate the densities of nodes, we calculate the similarity between nodes and data points. In j-th dimension, a local scaling function of the node v V ( t ) and the data point x i X ( t ) is given by
f j ( t ) ( v j ) = max ( 1 Φ ( x i j ) v j , 0 )
where v j is the coordinate of the grid node v V ( t ) in the j-th dimension and x i j is the coordinate of the data point x i X ( t ) in the j-th dimension.
Using f, the node’s density value at t-th iteration is given by
ρ v ( t ) = j = 1 m f j ( t ) ( v j )
It is worth noting that each data point only affects the densities of the vertices (nodes) of the cell in which it is located. There may be a large number of nodes with a density value of 0 in the standard grid. To reduce the computational efficiency, we use a sparse tensor to preserve the density of non-zero nodes.

4.3. Boundary Detection

Inspired by border-peeling clustering [29], we use an iterative boundary detection strategy to divide the core and boundary nodes. The GCBD algorithm will classify a portion of the nodes of V ( t ) as boundary nodes and assign an inactive status to them during every boundary detection iteration.
The inactive nodes of V ( t ) are defined using a specific cut-off value, as follows:
V U ( t ) = { v | ρ v ( t ) τ ( t ) }
As with [16], a percentile is used to indirectly provide a series of cut-off values. The inactive nodes are defined as nodes whose density values fall below the 10-th percentile.
The inactive data points of X ( t ) are given by
X U ( t ) = { x | d i s t C ( Φ ( x ) , v ) 1 , v V U ( t ) }
where d i s t C is the Chebyshev distance.
At the next iteration, the active nodes are given by
V ( t + 1 ) = V ( t ) \ V U ( t ) .
Similarly, the active nodes at the next iteration are given by
X ( t + 1 ) = X ( t ) \ X U ( t ) .
At the end of all iterations, the set of activated nodes and the set of active data points are V ( T + 1 ) and X ( T + 1 ) , where T is the number of iterations.
Definition 2
(Core node). A node v is called a core node if it belongs to the set V ( T + 1 ) , i.e., if  v V ( T + 1 ) .
Definition 3
(Core point). A point x is called a core point if it belongs to the set X ( T + 1 ) , i.e., if  x X ( T + 1 ) .
Definition 4
(Boundary node). A node v is called a boundary node if it belongs to the set V B ( T ) , where V B ( T ) = V U ( 1 ) V U ( 2 ) V U ( T ) , i.e., if v V U ( 1 ) V U ( 2 ) V U ( T ) .
Definition 5
(Boundary point). A point x is called a boundary point if it belongs to the set X B ( T ) , where X B ( T ) = X U ( 1 ) X U ( 2 ) X U ( T ) , i.e., if x V U ( 1 ) V U ( 2 ) V U ( T ) .

4.4. Connection Strategy

Next, we introduce the merging and assignment steps for nodes.

4.4.1. Merging Step

Inspired by DBSCAN [19], we devise a merging step to classify the core nodes. We define the following notions.
Definition 6
(Reachable). Two core nodes v p and v q are reachable if d i s t C ( v p , v q )   = 1 .
Definition 7
(Connected). Two core nodes v p and v q are connected if they are directly or transitively reachable.
Definition 8
(Cluster core). A Cluster core C ^ is a non-empty maximal subset of V ( T + 1 ) such that every pair of nodes in C ^ is connected.

4.4.2. Assignment Step

Inspired by DPC [20], we devise an assignment step to assign boundary nodes to clusters. For each node v p V U ( t ) , we find a node v q V ( t + 1 ) with the highest density among its nearest neighbors. We associate v p to v q and form the resulting clusters C = { C 1 , C 2 , , C k } .

4.4.3. Mapping of Points to Clusters

All points are clustered into the same group with their nearest nodes. To improve computational efficiency, we use the “round” function to find the matching node for the point, instead of using the distance function to calculate distances between the point and all nodes.

4.5. Algorithm Description and Complexity Analysis

The proposed algorithm is summarized in Algorithm 1.
Algorithm 1: GCBD algorithm
Input:    A set of points, X = { x 1 , , x n }
                The number of intervals, l
                The number of iterations, T
Output: The clustering result, C = { C 1 , C 2 , , C k }
1 Create a standard grid structure based on Definition 1.
2 Calculate the initial node density for each node according to Equation (5).
3 Categorize nodes as core nodes or boundary nodes based on an iterative boundary detection strategy.
4 Merge core nodes based on Definition 8.
5 Assign boundary nodes (see Section 4.4.2).
6 Map points to clusters (see Section 4.4.3).
We analyze the time complexity of GCBD. Assume that the number of points, the number of dimensions, and the number of intervals in each dimension are denoted by n, m, and l, respectively. The number of all nodes is g, and the number of sparse nodes is g . The number of iterations is T. A standard grid structure can be constructed in O ( n ) time. The time complexity of calculating the initial node density is O ( g n ) . In the worst case, the time complexity of the third step is O ( T ( g 2 + g n ) ) = O ( g 2 + g n ) , where T is a small constant representing the number of iterations. The merging step and assignment step can be completed in O ( g log ( g ) ) time. The time complexity of the last step is O ( n ) .The time complexity of GCBD is approximated to be O ( g n ) if g n , O ( g 2 ) otherwise.
We also give the time complexity of some existing algorithms. The time complexity of DGB [27] is O ( g 2 ) , where g is the number of nodes. The time complexity of WaveCluster [24] is O ( c ) O ( g ) , where c is the number of cells. The time complexity of CLIQUE [25] is the number of O ( c 2 ) O ( g 2 ) . The time complexity of BANG [22] is O ( n log ( n ) ) , where n is the number of points. The time complexity of GRIDBSCAN [28] is O ( c log ( c ) ) . The time complexity of OpiGrid [26] is O ( n log ( n ) ) .

5. Experiments

In the following, the performance of the proposed clustering is studied in comparison to several grid-based clustering algorithms on various datasets.

5.1. Experiment Setup

We compare the performance of GCBD with 6 grid-based clustering algorithms, including DGB [27], WaveCluster [24], CLIQUE [25], BANG [22], GRIDBSCAN [28] and OpiGrid [26]. According to the following parameter settings, we search for the best clustering result. GCBD, DGB, WaveCluster, CLIQUE, and OpiGrid all require a parameter l describing the number of intervals. Its range is between [ 5 , 50 ] . In GCBD, the parameter T indicates the number of iterations. This parameter value lies in the range [ 2 , 12 ] . DGB and WaveCluster have one parameter ε . We set ε ranging from 0.01 to 0.1 with step 0.01. In WaveCluster and BANG, the parameter h indicates the number of levels. Its value falls between [ 1 , 5 ] . CLIQUE, BANG, and GRIDBSCAN all require a parameter c which indicates the density threshold. Its range is between [ 0 , 5 ] . GRIDBSCAN has one parameter p. We set p ranging from 0.01 to 0.1 with step 0.01.
We evaluate performance using 12 synthetic datasets, and six real-world datasets. The synthetic datasets include Mickey, Gu, Jain, ThreeD, DiffD, Moons, Shape3, Handl, Yinyang, T4, T7, and SF. The real-world datasets include ORL, Dermatology, Control, Dig, Optdigits, and Satimage. These datasets come from benchmark clustering datasets (https://github.com/milaan9/Clustering-Datasets, accessed on 1 November 2022) UCI machine learning repository (http://archive.ics.uci.edu/ml/index.php, accessed on 1 November 2022). Table 1 provides detailed information about these datasets.
Due to available class labels for the selected data sets, we employ external evaluation measures including Adjusted Mutual Information (AMI) [30], Fowlkes-Mallows index (FMI) [31], and F1 score [32]. The three measures range from [ 0 , 1 ] , with 1 denoting perfect results and 0 denoting the worst ones.

5.2. Results on Synthetic Datasets

In Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14, we present Mickey, Gu, Jain, ThreeD, DiffD, Moons, Shape3, Handl, Yinyang, T4, T7, and SF as examples to illustrate the superiority of our algorithm. Each cluster is indicated by a different color solid dot. And black dots with hollow shapes indicate noisy data.
In Figure 3 and Figure 4, the first two rows correspond to the clustering results of Mickey and Gu, two unbalanced data sets. The Mickey dataset is perfectly clustered by DGB, CLIQUE, BANG, and our algorithm. Compared to the Mickey dataset, the two spherical clusters on the Gu dataset are closer together, which makes clustering more challenging. Our algorithm is the only one that can cluster this dataset perfectly.
Figure 5, Figure 6 and Figure 7 show clustering results for three datasets with different densities (Jain, ThreeD, and DiffD). A perfect clustering of the Jain dataset is achieved by DGB and our algorithm. On the ThreeD dataset, the results of GCBD and DGB are nearly perfectly right. On the DiffD dataset, only our algorithm can find the correct number of clusters. Some algorithms (DGB and WaveCluster) incorrectly classify low-density clusters as noise, and others (CLIQUE and Bang) merge two adjacent high-density clusters into one class.
Figure 8, Figure 9 and Figure 10 show clustering results for three datasets that have clusters with overlapping regions (Moons, Shape3 and Handl). On these three datasets, only our algorithm can discover the overall structure of the clusters. Almost all comparison algorithms produce false merges between adjacent clusters.
Figure 11, Figure 12, Figure 13 and Figure 14 correspond to the clustering results of four datasets with various shapes (Yinyang, T4, T7, and SF). The Yinyang dataset is perfectly clustered by DGB and our algorithm. On T4, T7, and SF, our algorithm outperforms the others.
Table 2 shows the quantitative results from these datasets. On all datasets, GCBD outperforms other algorithms in terms of all metrics. Experimental results show that our algorithm can identify clusters with different sizes, varying densities, overlapping regions, and arbitrary shapes.

5.3. Results on Real-World Datasets

We evaluate the clustering quality of all algorithms on 6 real-world datasets. The datasets are preprocessed the same way as in [33]. Table 3 illustrates that GCBD outperforms all comparison algorithms on most real-world datasets. For the high-dimensional dataset ORL, our algorithm improves F1 by an average of 0.30 compared to the others. Compared with DGB, WaveCluster, CLIQUE, BANG, GRIDBSCAN, and OptiGrid, GCBD has significantly better AMI and FMI. On the Dermatology dataset, our algorithm improves AMI and FMI by an average of 0.23 and 0.30 compared to the others. GCBD improves F1 by an average of 0.51 over WaveCluster, GRIDBSCAN, and OptiGrid. Furthermore, its F1 is far superior to DGB, CLIQUE, and BANG. On the Control dataset, GCBD has better AMI, FMI, and F1 than other algorithms. In particular, the improvement of F1 is more significant. On the Dig dataset, GCBD outperforms all comparison algorithms. On the Optdigits dataset, our algorithm has better AMI, FMI, and F1 than the others. Specifically, the F1 score of GCBD improves by an average of 0.36 over other algorithms. On the Satimage dataset, GCBD still achieves better clustering results. It improves the clustering AMI, FMI, and F1 by an average of 0.25, 0.30, and 0.32 compared to other algorithms. In summary, our algorithm improves AMI, FMI, and F1 by an average of 0.23, 0.27, and 0.29 over the others.

5.4. Running Time

This subsection compares the running times of GCBD and 6 competitors (DGB, WaveCluster, CLIQUE, BANG, GRIDBSCAN, and OptiGrid) on synthetic data sets with different sizes (n = 1000:1000:10,000). For a fair comparison, the number of intervals (a parameter common to GCBD, DGB, WaveCluster, CLIQUE, and OpiGrid) is kept at 20. To determine the running time, we use the average and standard deviation of 50 repeated experiments. We perform all experiments in the Matlab environment on a PC machine containing an Intel(R) Core(TM)-i7-9700F CPU and 32 GB RAM.
Figure 15 shows the average and standard deviation of the running times for GCBD, DGB, WaveCluster, CLIQUE, BANG, GRIDBSCAN, and OptiGrid. It is important to note that the y-axis is plotted using a base-10 log scale. We see that the BANG is significantly slower than other algorithms (GCBD, DGB, WaveCluster, CLIQUE, GRIDBSCAN, and OptiGrid). GCBD is the second-fastest in more than half of the cases. Scalability comparisons show our algorithm is competitive.

6. Conclusions

This paper presents a novel grid-based clustering algorithm for clusters with different sizes, varying densities, overlapping regions, and arbitrary shapes. Specifically, we define a density estimation of nodes based on a standard grid structure. We use an iterative boundary detection strategy to distinguish core nodes from boundary nodes. Therefore, the density threshold does not need to be specified by the user. In addition, the iterative density estimation and boundary detection can discover the boundary regions between adjacent clusters well, which facilitates the processing of clusters with varying densities and overlapping regions. Finally, the adopted connectivity strategy is beneficial for identifying clusters of arbitrary shapes.
This algorithm is mainly applied only to low-dimensional data. However, in the case of very high-dimensional data. To embed the data into a proper dimension, dimension reduction techniques such as Principal Component Analysis (PCA) and Uniform Manifold Approximation and Projection (UMAP) may be adopted. To be able to better cluster high-dimensional data, our future work is to introduce the idea of subspaces to alleviate the problem of the “curse of dimensionality”.

Author Contributions

Conceptualization, M.D.; methodology, M.D.; software, M.D. and F.W.; validation, M.D. and F.W.; formal analysis, M.D.; investigation, M.D. and F.W.; resources, M.D.; data curation, M.D. and F.W.; writing—original draft preparation, M.D.; writing—review and editing, M.D.; visualization, M.D. and F.W.; supervision, M.D.; project administration, M.D.; funding acquisition, M.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work research was funded by the National Natural Science Foundation of China, grant number 62006104, and the Natural Science Foundation of the Jiangsu Higher Education Institutions, grant number 20KJB520012.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All datasets used for this study have been deposited in the clustering benchmarks repository (https://github.com/milaan9/Clustering-Datasets, accessed on 1 November 2022) and UCI Machine learning Repository (http://archive.ics.uci.edu/ml/index.php, accessed on 1 November 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shao, C.; Du, X.; Yu, J.; Chen, J. Cluster-based improved isolation forest. Entropy 2022, 24, 611. [Google Scholar] [CrossRef] [PubMed]
  2. Shalileh, S.; Mirkin, B. Community partitioning over feature-rich networks using an extended k-means method. Entropy 2022, 24, 626. [Google Scholar] [CrossRef] [PubMed]
  3. Malzer, C.; Baum, M. Constraint-based hierarchical cluster selection in automotive radar data. Sensors 2021, 21, 3410. [Google Scholar] [CrossRef] [PubMed]
  4. Cuzzocrea, A.; Gaber, M.M.; Fadda, E.; Grasso, G.M. An innovative framework for supporting big atmospheric data analytics via clustering-based spatio-temporal analysis. J. Ambient. Intell. Humaniz. Comput. 2019, 10, 3383–3398. [Google Scholar] [CrossRef]
  5. Prieto Santamaría, L.; García del Valle, E.P.; Lagunes García, G.; Zanin, M.; Rodríguez González, A.; Menasalvas Ruiz, E.; Pérez Gallardo, Y.; Hernández Chan, G.S. Analysis of new nosological models from disease similarities using clustering. In Proceedings of the 33rd IEEE International Symposium on Computer-Based Medical Systems (CBMS), Rochester, MN, USA, 28–30 July 2020; pp. 183–188. [Google Scholar] [CrossRef]
  6. Gamino-Sánchez, F.; Hernández-Gutiérrez, I.V.; Rosales-Silva, A.J.; Gallegos-Funes, F.J.; Mújica-Vargas, D.; Ramos-Díaz, E.; Carvajal-Gámez, B.E.; Kinani, J.M.V. Block-Matching Fuzzy C-Means clustering algorithm for segmentation of color images degraded with Gaussian noise. Eng. Appl. Artif. Intell. 2018, 73, 31–49. [Google Scholar] [CrossRef]
  7. Vianney Kinani, J.M.; Rosales Silva, A.J.; Gallegos Funes, F.; Mújica Vargas, D.; Ramos Díaz, E.; Arellano, A. Medical imaging lesion detection based on unified gravitational fuzzy clustering. J. Healthc. Eng. 2017, 2017, 8536206. [Google Scholar] [CrossRef] [Green Version]
  8. Dzuba, S.; Krylov, D. Cluster analysis of financial strategies of companies. Mathematics 2021, 9, 3192. [Google Scholar] [CrossRef]
  9. Guo, H.; Wang, L.; Liu, X.; Pedrycz, W. Information granulation-based fuzzy clustering of time series. IEEE Trans. Cybern. 2021, 51, 6253–6261. [Google Scholar] [CrossRef]
  10. Gatto, B.B.; dos Santos, E.M.; Molinetti, M.A.F.; Fukui, K. Multilinear clustering via tensor Fukunaga-Koontz transform with Fisher eigenspectrum regularization. Appl. Soft Comput. 2021, 113, 107899. [Google Scholar] [CrossRef]
  11. Lu, M.; Zhao, X.J.; Zhang, L.; Li, F.Z. Semi-supervised concept factorization for document clustering. Inf. Sci. 2016, 331, 86–98. [Google Scholar] [CrossRef]
  12. Nedyalkova, M.; Sarbu, C.; Tobiszewski, M.; Simeonov, V. Fuzzy divisive hierarchical clustering of solvents according to their experimentally and theoretically predicted descriptors. Symmetry 2020, 12, 1763. [Google Scholar] [CrossRef]
  13. Choudhary, A.; Kumar, S.; Gupta, S.; Gong, M.; Mahanti, A. FEHCA: A fault-tolerant energy-efficient hierarchical clustering algorithm for wireless sensor networks. Energies 2021, 14, 3935. [Google Scholar] [CrossRef]
  14. Yao, P.; Zhu, Q.; Zhao, R. Gaussian mixture model and self-organizing map neural-network-based coverage for target search in curve-shape area. IEEE Trans. Cybern. 2022, 52, 3971–3983. [Google Scholar] [CrossRef]
  15. Smieja, M.; Hajto, K.; Tabor, J. Efficient mixture model for clustering of sparse high dimensional binary data. Data Min. Knowl. Discov. 2019, 33, 1583–1624. [Google Scholar] [CrossRef] [Green Version]
  16. Du, M.; Wang, R.; Ji, R.; Wang, X.; Dong, Y. ROBP a robust border-peeling clustering using Cauchy kernel. Inf. Sci. 2021, 571, 375–400. [Google Scholar] [CrossRef]
  17. Sieranoja, S.; Fränti, P. Fast and general density peaks clustering. Pattern Recognit. Lett. 2019, 128, 551–558. [Google Scholar] [CrossRef]
  18. Du, M.; Zhao, J.; Sun, J.; Dong, Y. M3W: Multistep Three-Way Clustering. IEEE Trans. Neural Networks Learn. Syst. 2022. [Google Scholar] [CrossRef]
  19. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the 2nd International Conference on Knowledge Discovery and Data Mining (KDD-96), Portland, OR, USA, 2–4 August 1996; pp. 226–231. [Google Scholar]
  20. Rodriguez, A.; Laio, A. Clustering by fast search and find of density peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef] [Green Version]
  21. Schikuta, E. Grid-clustering: An efficient hierarchical clustering method for very large data sets. In Proceedings of the 13th International Conference on Pattern Recognition (ICPR), Vienna, Austria, 25–19 August 1996; pp. 101–105. [Google Scholar] [CrossRef]
  22. Schikuta, E.; Erhart, M. The BANG-clustering system: Grid-based data analysis. In Proceedings of the 2nd International Symposium on Intelligent Data Analysis (IDA), London, UK, 4–6 August 1997; pp. 513–524. [Google Scholar] [CrossRef]
  23. Wang, W.; Yang, J.; Muntz, R. STING: A statistical information grid approach to spatial data mining. In Proceedings of the 23th International Conference on Very Large Data Bases (VLDB), Athens, Greece, 25–29 August 1997; pp. 186–195. [Google Scholar] [CrossRef] [Green Version]
  24. Sheikholeslami, G.; Chatterjee, S.; Zhang, A. Wavecluster: A multi-resolution clustering approach for very large spatial databases. In Proceedings of the 24th International Conference on Very Large Data Bases (VLDB), New York, NY, USA, 24–27 August 1998; pp. 428–439. [Google Scholar] [CrossRef]
  25. Agrawal, R.; Gehrke, J.; Gunopulos, D.; Raghavan, P. Automatic subspace clustering of high dimensional data for data mining applications. In Proceedings of the 1998 ACM International Conference on Management of Data (SIGMOD), Seattle, WA, USA, 2–4 June 1998; pp. 94–105. [Google Scholar] [CrossRef] [Green Version]
  26. Hinneburg, A.; Keim, D.A. Optimal grid-clustering: Towards breaking the curse of dimensionality in high-dimensional clustering. In Proceedings of the 25th International Conference on Very Large Data Bases (VLDB), Edinburgh, UK, 7–10 September 1999; pp. 506–517. [Google Scholar]
  27. Wu, B.; Wilamowski, B.M. A fast density and grid based clustering method for data with arbitrary shapes and noise. IEEE Trans. Ind. Inform. 2017, 13, 1620–1628. [Google Scholar] [CrossRef]
  28. Uncu, O.; Gruver, W.A.; Kotak, D.B.; Sabaz, D.; Alibhai, Z.; Ng, C. GRIDBSCAN: GRId density-based spatial clustering of applications with noise. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics (ICSMC), Taipei, China, 8–11 October 2006; pp. 2976–2981. [Google Scholar] [CrossRef]
  29. Averbuch-Elor, H.; Bar, N.; Cohen-Or, D. Border-peeling clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 1791–1797. [Google Scholar] [CrossRef] [Green Version]
  30. Nguyen, X.V.; Epps, J.; Bailey, J. Information theoretic measures for clusterings comparison: Variants, properties, normalization and correction for chance. J. Mach. Learn. Res. 2010, 11, 2837–2854. [Google Scholar] [CrossRef]
  31. Fowlkes, E.B.; Mallows, C.L. A method for comparing two hierarchical clusterings. J. Am. Stat. Assoc. 1983, 78, 576–579. [Google Scholar] [CrossRef]
  32. Banerjee, A.; Krumpelman, C.; Ghosh, J.; Basu, S.; Mooney, R.J. Model-based overlapping clustering. In Proceedings of the Eleventh ACM SIGKDD International Conference on Knowledge Discovery in Data Mining (SIGKDD), Chicago, IL, USA, 21–24 August 2005; pp. 532–537. [Google Scholar] [CrossRef]
  33. Peng, D.; Gui, Z.; Wang, D.; Ma, Y.; Huang, Z.; Zhou, Y.; Wu, H. Clustering by measuring local direction centrality for data with heterogeneous density and weak connectivity. Nat. Commun. 2022, 13, 5455. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Example of some notions in DBSCAN.
Figure 1. Example of some notions in DBSCAN.
Entropy 24 01606 g001
Figure 2. Example of some notions in GCBD.
Figure 2. Example of some notions in GCBD.
Entropy 24 01606 g002
Figure 3. Clustering results on Mickey. (a) True structure of the Mickey dataset: It consists of three spherical clusters with different sizes. (b) Result of GCBD: It can yield the optimal structure of clusters. (c) Result of DGB: It can yield the optimal structure of clusters. (d) Result of WaveCluster: Two clusters are wrongly marked as outliers. (e) Result of CLIQUE: It can yield the optimal structure of clusters. (f) Result of BANG: It can yield the optimal structure of clusters. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: A few points are erroneously clustered.
Figure 3. Clustering results on Mickey. (a) True structure of the Mickey dataset: It consists of three spherical clusters with different sizes. (b) Result of GCBD: It can yield the optimal structure of clusters. (c) Result of DGB: It can yield the optimal structure of clusters. (d) Result of WaveCluster: Two clusters are wrongly marked as outliers. (e) Result of CLIQUE: It can yield the optimal structure of clusters. (f) Result of BANG: It can yield the optimal structure of clusters. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: A few points are erroneously clustered.
Entropy 24 01606 g003
Figure 4. Clustering results on Gu. (a) True structure of the Gu dataset: It consists of two Gaussian clusters with different sizes. (b) Result of GCBD: It can yield the optimal structure of clusters. (c) Result of DGB: Very few points are erroneously clustered. (d) Result of WaveCluster: A few points are wrongly marked as outliers. (e) Result of CLIQUE: Two clusters are mistakenly merged into one. (f) Result of BANG: Two clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: Two clusters are mistakenly merged into one. (h) Result of OptiGrid: Very few points are erroneously clustered.
Figure 4. Clustering results on Gu. (a) True structure of the Gu dataset: It consists of two Gaussian clusters with different sizes. (b) Result of GCBD: It can yield the optimal structure of clusters. (c) Result of DGB: Very few points are erroneously clustered. (d) Result of WaveCluster: A few points are wrongly marked as outliers. (e) Result of CLIQUE: Two clusters are mistakenly merged into one. (f) Result of BANG: Two clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: Two clusters are mistakenly merged into one. (h) Result of OptiGrid: Very few points are erroneously clustered.
Entropy 24 01606 g004
Figure 5. Clustering results on Jain. (a) True structure of the Jain dataset: It consists of two crescent-shaped clusters with different densities. (b) Result of GCBD: It can yield the optimal structure of clusters. (c) Result of DGB: It can yield the optimal structure of clusters. (d) Result of WaveCluster: Points in the cluster below are wrongly marked as outliers. (e) Result of CLIQUE: It divides the upper cluster into several small sub-clusters. (f) Result of BANG: It yields over-partitioning. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Two clusters are divided incorrectly into several small sub-clusters.
Figure 5. Clustering results on Jain. (a) True structure of the Jain dataset: It consists of two crescent-shaped clusters with different densities. (b) Result of GCBD: It can yield the optimal structure of clusters. (c) Result of DGB: It can yield the optimal structure of clusters. (d) Result of WaveCluster: Points in the cluster below are wrongly marked as outliers. (e) Result of CLIQUE: It divides the upper cluster into several small sub-clusters. (f) Result of BANG: It yields over-partitioning. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Two clusters are divided incorrectly into several small sub-clusters.
Entropy 24 01606 g005
Figure 6. Clustering results on ThreeD. (a) True structure of the ThreeD dataset: It consists of three clusters with different densities. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: A few points are wrongly marked as outliers. (d) Result of WaveCluster: Two clusters are wrongly marked as outliers. (e) Result of CLIQUE: Very few points are erroneously clustered. (f) Result of BANG: A mass of points are erroneously clustered. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: All clusters are divided incorrectly into several small sub-clusters.
Figure 6. Clustering results on ThreeD. (a) True structure of the ThreeD dataset: It consists of three clusters with different densities. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: A few points are wrongly marked as outliers. (d) Result of WaveCluster: Two clusters are wrongly marked as outliers. (e) Result of CLIQUE: Very few points are erroneously clustered. (f) Result of BANG: A mass of points are erroneously clustered. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: All clusters are divided incorrectly into several small sub-clusters.
Entropy 24 01606 g006
Figure 7. Clustering results on DiffD. (a) True structure of the DiffD dataset: It consists of four Gaussian clusters with different sizes. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: Points in the cluster upper are wrongly marked as outliers. (d) Result of WaveCluster: A mass of points are wrongly marked as outliers. (e) Result of CLIQUE: Two clusters are mistakenly merged into one. (f) Result of BANG: Two clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Two clusters are divided incorrectly several small sub-clusters.
Figure 7. Clustering results on DiffD. (a) True structure of the DiffD dataset: It consists of four Gaussian clusters with different sizes. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: Points in the cluster upper are wrongly marked as outliers. (d) Result of WaveCluster: A mass of points are wrongly marked as outliers. (e) Result of CLIQUE: Two clusters are mistakenly merged into one. (f) Result of BANG: Two clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Two clusters are divided incorrectly several small sub-clusters.
Entropy 24 01606 g007
Figure 8. Clustering results on Moons. (a) True structure of the Moons dataset: It consists of two crescent-shaped clusters with different densities. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: Two clusters are mistakenly merged into one. (d) Result of WaveCluster: Points in the cluster upper are wrongly marked as outliers. (e) Result of CLIQUE: Two clusters are mistakenly merged into one. (f) Result of BANG: It yields over-partitioning. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Two clusters are divided incorrectly into several small sub-clusters.
Figure 8. Clustering results on Moons. (a) True structure of the Moons dataset: It consists of two crescent-shaped clusters with different densities. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: Two clusters are mistakenly merged into one. (d) Result of WaveCluster: Points in the cluster upper are wrongly marked as outliers. (e) Result of CLIQUE: Two clusters are mistakenly merged into one. (f) Result of BANG: It yields over-partitioning. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Two clusters are divided incorrectly into several small sub-clusters.
Entropy 24 01606 g008
Figure 9. Clustering results on Shape3. (a) True structure of the Shape3 dataset: It consists of three clusters with different densities. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: Two clusters are mistakenly merged into one. (d) Result of WaveCluster: One cluster is wrongly marked as an outlier. (e) Result of CLIQUE: Two clusters are mistakenly merged into one. (f) Result of BANG: Two clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Two clusters are divided incorrectly into several small sub-clusters.
Figure 9. Clustering results on Shape3. (a) True structure of the Shape3 dataset: It consists of three clusters with different densities. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: Two clusters are mistakenly merged into one. (d) Result of WaveCluster: One cluster is wrongly marked as an outlier. (e) Result of CLIQUE: Two clusters are mistakenly merged into one. (f) Result of BANG: Two clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Two clusters are divided incorrectly into several small sub-clusters.
Entropy 24 01606 g009
Figure 10. Clustering results on Handl. (a) True structure of the Handl dataset: It consists of three clusters with different densities. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: A few points are wrongly marked as outliers. (d) Result of WaveCluster: A mass of points are wrongly marked as outliers. (e) Result of CLIQUE: One cluster is divided incorrectly into several small sub-clusters. (f) Result of BANG: Three clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Two clusters are mistakenly merged into one.
Figure 10. Clustering results on Handl. (a) True structure of the Handl dataset: It consists of three clusters with different densities. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: A few points are wrongly marked as outliers. (d) Result of WaveCluster: A mass of points are wrongly marked as outliers. (e) Result of CLIQUE: One cluster is divided incorrectly into several small sub-clusters. (f) Result of BANG: Three clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Two clusters are mistakenly merged into one.
Entropy 24 01606 g010
Figure 11. Clustering results on Yinyang. (a) True structure of the Yinyang dataset: Its composition is like a yin-yang diagram. (b) Result of GCBD: It can yield the optimal structure of clusters. (c) Result of DGB: It can yield the optimal structure of clusters. (d) Result of WaveCluster: A few points are wrongly marked as outliers. (e) Result of CLIQUE: Very few points are erroneously clustered. (f) Result of BANG: Very few points are erroneously clustered. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Three clusters are divided incorrectly into several small sub-clusters.
Figure 11. Clustering results on Yinyang. (a) True structure of the Yinyang dataset: Its composition is like a yin-yang diagram. (b) Result of GCBD: It can yield the optimal structure of clusters. (c) Result of DGB: It can yield the optimal structure of clusters. (d) Result of WaveCluster: A few points are wrongly marked as outliers. (e) Result of CLIQUE: Very few points are erroneously clustered. (f) Result of BANG: Very few points are erroneously clustered. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Three clusters are divided incorrectly into several small sub-clusters.
Entropy 24 01606 g011
Figure 12. Clustering results on T4. (a) True structure of the T4 dataset: It consists of six intertwined or nested clusters. (b) Result of GCBD: It can yield the optimal structure of clusters. (c) Result of DGB: Two clusters are mistakenly merged into one. (d) Result of WaveCluster: A few points are wrongly marked as outliers. (e) Result of CLIQUE: Two clusters are mistakenly merged into one. (f) Result of BANG: Three clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Six clusters are divided incorrectly several small sub-clusters.
Figure 12. Clustering results on T4. (a) True structure of the T4 dataset: It consists of six intertwined or nested clusters. (b) Result of GCBD: It can yield the optimal structure of clusters. (c) Result of DGB: Two clusters are mistakenly merged into one. (d) Result of WaveCluster: A few points are wrongly marked as outliers. (e) Result of CLIQUE: Two clusters are mistakenly merged into one. (f) Result of BANG: Three clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Six clusters are divided incorrectly several small sub-clusters.
Entropy 24 01606 g012
Figure 13. Clustering results on T7. (a) True structure of the T7 dataset: It consists of nine intertwined or nested clusters. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: Seven clusters are mistakenly merged into one. (d) Result of WaveCluster: Three clusters are mistakenly merged into one. (e) Result of CLIQUE: Seven clusters are mistakenly merged into one. (f) Result of BANG: Three clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Five clusters are divided incorrectly several small sub-clusters.
Figure 13. Clustering results on T7. (a) True structure of the T7 dataset: It consists of nine intertwined or nested clusters. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: Seven clusters are mistakenly merged into one. (d) Result of WaveCluster: Three clusters are mistakenly merged into one. (e) Result of CLIQUE: Seven clusters are mistakenly merged into one. (f) Result of BANG: Three clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Five clusters are divided incorrectly several small sub-clusters.
Entropy 24 01606 g013
Figure 14. Clustering results on SF. (a) True structure of the SF dataset: These clusters look like smiling faces. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: Two clusters are mistakenly merged into one. (d) Result of WaveCluster: Two clusters are wrongly marked as outliers. (e) Result of CLIQUE: All clusters are mistakenly merged into one. (f) Result of BANG: All clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Three clusters are divided incorrectly several small sub-clusters.
Figure 14. Clustering results on SF. (a) True structure of the SF dataset: These clusters look like smiling faces. (b) Result of GCBD: It can correctly find the general shapes of each cluster. (c) Result of DGB: Two clusters are mistakenly merged into one. (d) Result of WaveCluster: Two clusters are wrongly marked as outliers. (e) Result of CLIQUE: All clusters are mistakenly merged into one. (f) Result of BANG: All clusters are mistakenly merged into one. (g) Result of GRIDBSCAN: It yields over-partitioning. (h) Result of OptiGrid: Three clusters are divided incorrectly several small sub-clusters.
Entropy 24 01606 g014
Figure 15. Running time comparison.
Figure 15. Running time comparison.
Entropy 24 01606 g015
Table 1. Datasets used in experiments.
Table 1. Datasets used in experiments.
Names#Instances#Features#Classes
Mickey120023
Gu105022
Jain37322
ThreeD130023
DiffD86324
Moons100022
Shape3225023
Handl71523
Yinyang320025
T4732626
T7920829
SF16,38424
ORL10010,30710
Dermatology366346
Control600606
Dig17976410
Optdigits56206410
Satimage6435366
Table 2. Performance comparison for 7 clustering algorithms on 12 synthetic datasets.
Table 2. Performance comparison for 7 clustering algorithms on 12 synthetic datasets.
AlgorithmAMIFMIF1Clu#ParameterAMIFMIF1Clu#Parameter
Mickey(3 classes)Gu(2 classes)
GCBD1.0001.0001.0003l = 13; T = 21.0001.0001.0002l = 29; T = 4
DGB1.0001.0001.0003l = 8; ε = 0.10.8160.9850.9852l = 15; ε = 0.05
WaveCluster0.6660.9110.9071l = 5; ε = 0.01; h = 40.3990.8500.8404l = 10; ε = 0.1; h = 1
CLIQUE1.0001.0001.0003l = 30; c = 0−0.0100.9430.94210l = 18; c = 4
BANG1.0001.0001.0003h = 10; c = 40.5010.9220.91949h = 12; c = 5
GRIDBSCAN0.0140.0900.0191091p = 0.01; c = 10.0170.8350.831137p = 0.02; c = 2
OptiGrid0.9320.9810.9815l = 50.9440.9980.9982l = 50
Jain(2 classes)ThreeD(3 classes)
GCBD1.0001.0001.0002l = 26; T = 90.9800.9960.9963l = 42; T = 4
DGB1.0001.0001.0002l = 18; ε = 0.10.9610.9940.9943l = 25; ε = 0.05
WaveCluster0.5860.8870.8841l = 5; ε = 0.01; h = 40.5950.8520.8451l = 5; ε = 0.01; h = 4
CLIQUE0.8400.9710.9715l = 18; c = 00.9590.9940.9947l = 25; c = 0
BANG0.5470.8500.83942h = 12; c = 00.8590.9850.98528h = 12; c = 5
GRIDBSCAN0.000 0.000373p = 0.08; c = 10.2830.8360.834281p = 0.03; c = 1
OptiGrid0.4100.5070.4536l = 500.5280.5500.5035l = 4
DiffD(4 classes)Moons(2 classes)
GCBD0.9930.9970.9974l = 44; T = 80.8940.9720.9722l = 30; T = 8
DGB0.9810.9930.9933l = 18; ε = 0.050.0260.6990.6613l = 18; ε = 0.1
WaveCluster0.5100.6070.60235l = 15; ε = 0.01; h = 10.2520.5100.50826l = 40; ε = 0.05; h = 3
CLIQUE0.6360.7160.68219l = 25; c = 10.0120.7050.6652l = 18; c = 3
BANG0.8620.9710.97143h = 12; c = 00.2050.0770.012512h = 15; c = 4
GRIDBSCAN0.1310.4070.364539p = 0.1; c = 10.0000.0550.012924p = 0.01; c = 1
OptiGrid0.6170.5160.4539l = 30.4640.4020.28513l = 8
Shape3(3 classes)Handl(3 classes)
GCBD0.9570.9840.9843l = 40; T = 90.9600.9880.9883l = 29 = 31; T = 11
DGB0.7240.7950.77814l = 50; ε = 0.050.9140.9780.9784l = 30; ε = 0.1
WaveCluster0.4470.5190.51916l = 25; ε = 0.1; h = 20.3920.5710.56810l = 20; ε = 0.01; h = 2
CLIQUE0.6830.7890.77639l = 50; c = 00.7500.9240.92431l = 30; c = 0
BANG0.7530.7990.7808h = 12; c = 00.6500.8450.84535h = 12; c = 4
GRIDBSCAN0.1810.5130.4251538p = 0.05; c = 10.1660.6550.655240p = 0.05; c = 1
OptiGrid0.5580.4830.41712l = 40.4910.7050.7004l = 8
Yinyang(5 classes)T4(6 classes)
GCBD1.0001.0001.0005l = 41; T = 21.0001.0001.0006l = 44; T = 3
DGB1.0001.0001.0005l = 30; ε = 0.050.9550.9430.9415l = 40; ε = 0.05
WaveCluster0.6720.7010.6886l = 15; ε = 0.01; h = 10.6950.6620.6625l = 12; ε = 0.1; h = 1
CLIQUE0.9991.0001.0006l = 30; c = 40.9550.9430.9416l = 50; c = 0
BANG0.9970.9990.9998h = 12; c = 40.9320.9630.963110h = 15; c = 4
GRIDBSCAN0.5300.8640.8591311p = 0.03; c = 10.0020.1000.0815619p = 0.01; c = 1
OptiGrid0.5680.4640.38710l = 10.7340.5650.51214l = 1
T7(9 classes)SF(4 classes)
GCBD0.9970.9990.9999l = 49; T = 50.9550.9860.9864l = 38; T = 12
DGB0.5980.5420.4544l = 50; ε = 0.10.0640.5240.45811l = 50; ε = 0.1
WaveCluster0.6230.6120.6105l = 15; ε = 0.05; h = 10.4310.4750.47539l = 50; ε = 0.1; h = 3
CLIQUE0.5270.5320.4412l = 50; c = 20.0220.5520.46914l = 50; c = 4
BANG0.9490.9340.93362h = 15; c = 40.6410.7520.723347h = 15; c = 0
GRIDBSCAN0.0020.0660.0417839p = 0.01; c = 10.0000.0220.0033940p = 0.01; c = 1
OptiGrid0.6110.4050.38012l = 50.5550.5090.46910l = 5
Table 3. Performance comparison for 7 clustering algorithms on 6 real-world datasets.
Table 3. Performance comparison for 7 clustering algorithms on 6 real-world datasets.
AlgorithmAMIFMIF1Clu#ParameterAMIFMIF1Clu#Parameter
ORL(10 classes)Dermatology(6 classes)
GCBD0.9550.9470.94612l = 46; T = 60.9300.9460.9466l = 38; T = 11
DGB0.9190.8920.88616l = 40;   ε = 0.050.9260.8960.8915l = 12;   ε = 0.1
WaveCluster0.4930.4100.2921l = 8;   ε = 0.01; h = 40.7450.7470.7393l = 5;   ε = 0.01; h = 3
CLIQUE0.9490.9360.93512l = 20; c = 00.9260.8960.8915l = 18; c = 0
BANG0.9000.8620.85217h = 12; c = 00.8710.7860.7866h = 12; c = 0
GRIDBSCAN0.1490.2610.17489p = 0.02; c = 1−0.0010.0000.000364p = 0.01; c = 1
OptiGrid0.8540.7770.75319l = 20.7470.6090.55015l = 1
Control(6 classes)Dig(10 classes)
GCBD0.8600.7730.7484l = 9; T = 20.9170.9050.90510l = 41; T = 8
DGB0.8600.7730.7484l = 8;   ε = 0.010.8990.8630.86311l = 50;   ε = 0.1
WaveCluster0.6210.5750.4971l = 5;   ε = 0.01; h = 30.6640.4590.44810l = 15;   ε = 0.01; h = 1
CLIQUE0.8600.7730.7484l = 40; c = 00.9030.8440.84110l = 40; c = 0
BANG0.8600.7730.7484h = 12; c = 00.8990.8620.86219h = 15; c = 0
GRIDBSCAN0.1960.3840.373409p = 0.05; c = 1−0.0010.0150.0051701p = 0.01; c = 1
OptiGrid0.5920.4350.41617l = 30.7510.6440.63810l = 2
Optdigits(10 classes)Satimage(6 classes)
GCBD0.9550.9630.96310l = 46; T = 80.7220.7720.7665l = 28; T = 9
DGB0.9220.8710.86810l = 30;   ε = 0.050.6580.6480.6094l = 30;   ε = 0.1
WaveCluster0.7120.5590.55711l = 20;   ε = 0.01; h = 10.4600.5130.4843l = 10;   ε = 0.1; h = 1
CLIQUE0.8590.7330.7108l = 40; c = 20.6130.6000.5403l = 40; c = 5
BANG0.8940.8190.81312h = 15; c = 00.6130.6000.5403h = 12; c = 0
GRIDBSCAN0.0000.0210.0085243p = 0.01; c = 10.0030.0340.0105999p = 0.01; c = 1
OptiGrid0.7960.6690.66817l = 50.5070.4700.47011l = 5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Du, M.; Wu, F. Grid-Based Clustering Using Boundary Detection. Entropy 2022, 24, 1606. https://doi.org/10.3390/e24111606

AMA Style

Du M, Wu F. Grid-Based Clustering Using Boundary Detection. Entropy. 2022; 24(11):1606. https://doi.org/10.3390/e24111606

Chicago/Turabian Style

Du, Mingjing, and Fuyu Wu. 2022. "Grid-Based Clustering Using Boundary Detection" Entropy 24, no. 11: 1606. https://doi.org/10.3390/e24111606

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop