Next Article in Journal
On the Descriptive Complexity of Color Coding
Next Article in Special Issue
A Feature Selection Algorithm Performance Metric for Comparative Analysis
Previous Article in Journal
An Integrated Neural Network and SEIR Model to Predict COVID-19
Previous Article in Special Issue
An Exploratory Landscape Analysis-Based Benchmark Suite
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Understanding Clustering Problems and Algorithms: An Instance Space Analysis

by
Luiz Henrique dos Santos Fernandes
1,*,
Ana Carolina Lorena
1 and
Kate Smith-Miles
2
1
Instituto Tecnológico de Aeronáutica (ITA), São José dos Campos, SP 12200-000, Brazil
2
School of Mathematics and Statistics, The University of Melbourne, Parkville, VIC 3010, Australia
*
Author to whom correspondence should be addressed.
Algorithms 2021, 14(3), 95; https://doi.org/10.3390/a14030095
Submission received: 30 November 2020 / Revised: 15 March 2021 / Accepted: 16 March 2021 / Published: 19 March 2021

Abstract

:
Various criteria and algorithms can be used for clustering, leading to very distinct outcomes and potential biases towards datasets with certain structures. More generally, the selection of the most effective algorithm to be applied for a given dataset, based on its characteristics, is a problem that has been largely studied in the field of meta-learning. Recent advances in the form of a new methodology known as Instance Space Analysis provide an opportunity to extend such meta-analyses to gain greater visual insights of the relationship between datasets’ characteristics and the performance of different algorithms. The aim of this study is to perform an Instance Space Analysis for the first time for clustering problems and algorithms. As a result, we are able to analyze the impact of the choice of the test instances employed, and the strengths and weaknesses of some popular clustering algorithms, for datasets with different structures.

1. Introduction

With increasing demand for techniques to automatically discover patterns in large datasets, the importance of clustering for data exploration and decomposing datasets into smaller and more homogeneous subsets has created a renewed interest in clustering algorithms [1]. Such algorithms stem from statistical approaches, optimization algorithms, fuzzy approaches, evolutionary algorithms and unsupervised machine learning (ML) methods. With so many algorithms available for clustering, it is an open question about which algorithm is best suited for a particular dataset. The selection of the most effective algorithm for solving a new problem instance based on its characteristics has been studied since the early 2000s in a branch of ML called meta-learning [2,3,4]. Typically, the meta-learning system receives as input a set of meta-examples representing learning problems and the performances of a portfolio of candidate algorithms used in their solution. Each meta-example stores a set of features that describe the problem, or meta-features and a target attribute that indicates either: the best evaluated algorithm for that problem; a ranking of the top performing algorithms; or the performance value achieved by one or more algorithms when solving the problem.
Studies in meta-learning and the algorithm selection problem have motivated a number of other meta-analyses’ proposals. One of them is the Instance Space Analysis (ISA) methodology, which has successfully been applied to various problems such as classification [5], forecasting [6] and anomaly detection [7], but not yet to clustering. The framework was originally developed to evaluate optimization problems [8,9] and to learn how the characteristics of problem instances affect the performance of algorithms. It maps the instances into a 2D space, which is built based on a set of meta-features and the performance of a set of algorithms. From this 2D instance space perspective, it is possible to visualize sets of easy and hard instances, as well as determine “footprints” of the algorithms, indicating their domains of competence and reliability. This provides rich information for revealing the best set of algorithms to be applied to a given instance to support automated algorithm selection. Finally, the sufficiency of existing benchmark test suites can be scrutinized, and new test instances can be generated at target locations in order to expand and diversify the instance space.
This paper applies ISA to a comprehensive study of clustering problems and algorithms for the first time. Our central focus is to study how measurable features of a clustering dataset (problem instance) affect a given clustering algorithm’s performance metric, and to objectively analyze the strengths and weaknesses of different clustering algorithms from the inspection of the instance space generated. The definition of a cluster in the literature is not unique, and each algorithm may adopt a different clustering criterion. The biases assumed by different algorithms will be better suited for the identification of some types of structures in a dataset. Nonetheless, a dataset may contain different types of structures and valid clusters within different granularities. The clustering problem has, therefore, several challenging aspects and may benefit from meta-analysis that relate the characteristics of the problem’s instances to the performance of the algorithms which may be used in their solution, such as those provided by the ISA framework. These are the insights that we seek in this study.
The challenges of clustering are naturally transmitted to the ISA instantiation we seek. Firstly, extracting meaningful meta-features from clustering problems has received little attention in the meta-learning community and has been the subject of recent pieces of work [10,11]. Here, we gather and introduce a set of meta-features aimed to reveal different types of structures on data. Since there are multiple clusters’ definitions, the selection of the clustering algorithms to be evaluated must be careful and include representatives with different biases, which we also tried to fulfill, despite focusing on popular partitional and hierarchical clustering algorithms. Finally, the lack of an expected response hampers the evaluation of the clustering algorithms’ results. There are various clustering validation indexes which can be used in this evaluation, but each one of them has a bias towards one particular cluster definition, which may favor some algorithms to the detriment of others. Therefore, we have combined multiple validation indexes in order to better assess the algorithms’ performances.
The remainder of this paper is organized as follows: In Section 2, we present a theoretical summary and previous studies on meta-learning and ISA. In Section 3, we describe the methodology of ISA for clustering. Section 4 presents the results of building an instance space for clustering from artificial datasets, as well as the discussions around the results. Section 5 summarizes the results obtained by including real datasets in the analysis. Conclusions and future work are presented in Section 6.

2. Background

Meta-learning can be interpreted as the exploration of meta-knowledge to obtain efficient models and solutions to new problems [12]. One of the first and seminal meta-learning contributions was developed by Rice [13]. In his work, the author deals with the problem of algorithm selection, i.e., the choice of an algorithm, among a set of options that is more appropriate to solve a particular instance of a problem [3]. Figure 1 illustrates Rice’s framework. Given a set of problem instances x P , from which measurable features f ( x ) F are extracted, the objective is to learn a selection mapping S ( f ( x ) ) to an algorithm α * from a pool of candidate algorithms A which has optimized performance when solving x, as measured by a performance measure y Y .
Smith-Miles et al. [8] have extended Rice’s framework to propose a methodology known as Instance Space Analysis. Initially, the methodology was used to evaluate the suitability of algorithms for graph colouring problems, and later generalized to ML classification algorithms [5] and other problems [6,7,14]. According to Figure 2, the framework is composed of several steps. In the center, I contains a subset of the problem space P for which computational results are available. In ML, this subset is usually formed by datasets collected from benchmark repositories. The feature space F contains multiple measures that can be used to characterize the properties of the instances in I. The algorithm space A is composed of a portfolio of algorithms that can be used to solve the instances. The performance space Y measures the performance of the algorithms in A, when evaluated on the solution of the instances in I. Through a computational process, for all instances in I and all algorithms in A, a meta-dataset containing an ordered quadruple ( I , F , A , Y ) is composed. One can then learn, through regression or another appropriate supervised learning method, the relationship between the features in F and the performance Y of the algorithms. This was the challenge of the Algorithm Selection Problem described by Rice [13] and shown in the inner box in Figure 2.
ISA methodology extends Rice’s framework to achieve more insights than mere algorithm selection. A 2D projection which captures relationships between the features values F and the performance measures Y can be generated. In this space, one may visualize the level of difficulty of different instances, the performance of different algorithms across the space and the relationship to instance features. The ultimate goal is more than automated algorithm selection, but greater insights into the strengths and weaknesses of algorithms, and explanations of how instance characteristics affect performance and reliability. Once the meta-data has been generated, the construction of the instance space involves a feature selection step in order to keep only the most informative meta-features and then projection of the instances into a 2D plane. Both of these steps require the solution of optimization problems: (i) to find the optimal subset of features that best explain variations in algorithm performance and “easy-to-hard” phase transitions; and (ii) to find the optimal linear transformation matrix from the selected feature space to 2D such that the distribution of algorithm performance and features in the 2D space exhibit strong linear trends for maximal visual interpretability.
The result is a projection of the instances into a 2D instance space that aims to elicit insights into why some algorithms might be more or less suited to certain types of instances. An ML model is applied to the instances in this space to learn algorithm “goodness”. Herewith, the regions where there is evidence of good performance are called the algorithm’s footprints, and represent a fundamental concept in ISA. The relative size of the footprints provide an objective measure of algorithm power, and their location indicates uniqueness and applicability to certain types of instances. The distribution of the instances also permits analysis of the diversity of the chosen benchmarks and opens up the possibility to generate new test instances that span underrepresented regions of the instance space. This was done in Muñoz and Smith-Miles [15] for generating custom classification datasets, through the fitting of a Gaussian Mixture Model and a set of generalized Bernoulli distributions, later refined in Muñoz et al. [5].
Lemke et al. [16] observe that, within the scope of ML, a considerable portion of the literature focuses on the use of meta-learning for algorithm recommendation. Regarding clustering problems, meta-learning has been explored as an alternative for clustering algorithm recommendation since the late 2000s [10,17,18,19,20,21]. Following a different perspective, Sáez and Corchado [22] developed a meta-learning system to determine the number of clusters in a dataset based on some quality indices. Pimentel and Carvalho [11] proposed a new data characterization method for selecting clustering algorithms, which was subsequently used in a meta-learning setup for recommending the number of clusters within a dataset [23]. Sallem and Gallagher [24] focus on landscape analysis of clustering problems viewed as black-box optimization problems, and solved using two well-known metaheuristics in clustering: Particle Swarm Optimization and Differential Evolution.
In a complementary meta-analysis perspective, the present work aims to analyse two aspects using ISA: (i) the diversity of clustering benchmarks; and (ii) the algorithmic power of some popular clustering algorithms.

3. Methodology of Instance Space Analysis for Clustering

The objective of clustering is to find structures in a dataset which summarize how the data points relate to each other in terms of their similarity of attributes. The resulting similarity structures or clusters can be found using statistical approaches, or unsupervised machine learning methods, and involve solving an optimization problem to allocate data points to clusters in a manner that maximises intra-cluster similarity and inter-cluster dissimilarity. There are several ways to define a cluster, resulting in different clustering criteria. In order to deal with such heterogeneity of approach, we first use different measures to characterize the clustering datasets, including meta-features used in previous works and also new measures that extract knowledge about the datasets’ structures according to different perspectives. In the evaluation of the algorithms, we combine various internal validation indexes in an average ranking. The objective of this combination is to embrace different clustering criteria, while also smoothing biases towards specific algorithms, which can happen between validation indexes and clustering algorithms employing similar clustering criteria. Furthermore, there are multiple challenges associated with clustering, mainly due to the lack of an expected output for the instances and the possible presence of multiple valid grouping structures on data. These challenges are also reflected in the construction of a suitable instance space for clustering instances, which must consider appropriate sets of instances, meta-features, algorithms, and evaluation measures. In this section, we describe the main steps followed in the generation of an instance space for clustering.

3.1. Clustering Instances and Algorithms

The first stage of the experiments used 380 synthetic clustering datasets from the literature to compose the instance set I. A total of 80 Gaussian datasets of low dimension and 80 ellipsoidal datasets of high dimension were collected from Handl and Knowles [25]. Another 220 datasets of various shapes and different number of attributes, clusters and instances from different sources were included [26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42]. In the selected datasets, the number of examples ranges from 100 to 5000, the number of attributes ranges from 2 to 1024 and the number of embedded clusters ranges from 1 to 40. Although we could include some labeled classification datasets into the analysis, as done in previous work [22,23], their structure is not always suitable for clustering purposes. For instance, there may be classes spanning different regions or manifolds of the input space. Additionally, a dataset might present more than one valid structure, which is not reflected in the assigned data labels. By employing synthetic datasets dedicated to the evaluation of clustering problems, we are able to circumvent these issues and build an instance space more suitable to clustering problems.
For composing the set of algorithms A, we analyzed 10 different partitional and hierarchical clustering algorithms widely used in the literature: K-Means (KME), K-Medians (KMD), K-Medoids (KMO), Mini-Batch K-Means (MBK), Fuzzy C-Means (FCM), Single Linkage (SL), Complete Linkage (CL), Average Linkage (AL), Closest Centroid (CC) and Ward’s Method (WM). The algorithms were executed using the implementations contained in the following R packages: stats, Gmedian, kmed, ClusterR and e1071. Appendix B presents the parameters used for each of the algorithms.

3.2. Meta-Data

An important step in the construction of meta-data are the definition of meta-features F that correspond to the characterization measures of the datasets. Table 1 lists the 20 meta-features used in this work. For each meta-feature, this table presents the main aspect it intends to measure, its acronym, description, asymptotic computational cost and reference from which the measure was extracted or based. The asymptotic cost is analyzed as a function of the number of points n each dataset has and its number of input features m. Most of the measures require building a distance matrix between all data points first, which dominates the asymptotic cost. Nonetheless, one must notice that this matrix needs to be computed only once and multiple measures can be extracted from it afterwards, so that the overall cost involved in extracting all meta-features is lower than that obtained by summing up each of the individual costs.
We have chosen a diverse set of meta-features for revealing different cluster structures while also allowing to draw meaningful conclusions regarding the clustering algorithm’s performances. Measures already tested in related works are included along with other measures that highlight other complementary aspects. The meta-features’ functions have been implemented in R, using the following packages as support: MVN and igraph. The code is available in an open access repository (see https://github.com/ml-research-clustering/meta-analysis, last accessed on 18 March 2021). They were categorized in this work in five groups, according to properties they are able to extract from the datasets: distribution-based, neighborhood-based, density-based, dimensionality and network centrality. Distribution-based, neighborhood-based and density-based measures are commonly employed as clustering criteria in the literature. Compact clusters are obtained whenever one maximizes the intra-cluster similarity, whilst minimizing inter-cluster similarity. Neighborhood-based measures favor placing neighborhood data items into the same group. Density captures how the data are distributed in the input space, so that dense regions separated by sparse regions can be regarded as suitable clusters. Dimensionality can be an issue to any clustering algorithm, as datasets with few data items and a high dimensionality tend to be sparse, which prevents finding an underlying clustering structure. The network centrality measures are based on building a proximity graph from data and measuring characteristics from this structure. The measures clust_coef and net_dens from the Density category and avg_nnd from the Neighborhood-based category are also extracted from this proximity graph. The graphs were built from the adjacency matrix, based on an approximation method that uses an ϵ -NN (nearest neighbor) function. As in related work [46], ϵ is defined as 15% of the number of examples the dataset has. Finally, in our implementation, the percentage of points of high distance corresponds to the proportion of points with a distance greater than the mean distance plus its standard deviation. Conversely, the percentage of points of low distance corresponds to the proportion of points with a distance lower than the mean distance minus its standard deviation.
Some meta-features listed in Table 1 have already been used in previous work to characterize test instances in clustering problems [10,17,18,20,44,45,47]. We have also included some meta-features that have been used in classification problems that can be directly or easily adapted to characterize clustering problems [46]. As a contribution of this work, we have proposed using multivariate normality skewness and kurtosis measures [43] and a meta-feature group based on well-known network centrality measures [48,49]. The hypothesis is that the measurement of the importance of the vertices of a graph generated from the dataset reflects the structure of the clusters in such a way that it is possible to discriminate instances according to the underlying network complexity.
For evaluating the clustering algorithms’ performance and compose the set Y, we have employed a set of internal validation measures, which use information inherent to the data structure to assess the quality of the clusters [50,51,52,53,54,55,56,57,58]. There are numerous measures in the literature, each with a specific bias. The measures used in this work are classified in six groups, as shown in Table 2 [58]. We included representatives from different categories as defined in the related literature, so that the ability of the algorithms to recover different structures from a dataset can be assessed. Their code was implemented in R, using the following packages: clusterCrit, clValid and fpc and are also available in supplementary material.
As a result of this step, we have a meta-dataset composed of 380 instances (clustering datasets), described by 20 meta-features and the performances of 10 clustering algorithms according to 12 clustering validation indexes.

3.3. Steps for Generating a 2D Instance Space for Clustering

The instance space is produced using the Matlab toolkit MATILDA (acronym for Melbourne Algorithm Test Instance Library with Data Analytics). MATILDA has been developed to be used in the analysis of optimization and ML problems (see https://matilda.unimelb.edu.au for a library of case studies, last accessed on 18 March 2021). A more detailed description of the methodology can be found in Muñoz et al. [5]. The tool consists of a data pipeline that integrates a pre-processing stage, four main steps and a post-processing stage.
A preliminary and relevant step to build an instance space is the consolidation of metadata. These are composed by the mentioned meta-features and performance measures of the algorithms under analysis. These actions are common in all kinds of problems, but in clustering some specific aspects had to be observed. The first is the need to determine an optimal number of clusters for each dataset. Considering the unsupervised nature of clustering problems, the number of clusters each dataset has is not known a priori. Therefore, we let the data structure speak for itself and the number of clusters is found according to the clustering algorithms’ results. We use the Silhouette Method [59] from the R language’s NbClust package to determine the number of clusters to be adopted for each dataset and clustering algorithm uniformly. It does not have the same bias of the internal validation criteria taken as performance measures. The inclusion of a new instance with or without information about the number of clusters must follow the same procedure.
Another aspect is the definition of a performance measurement. In this work, we chose to use internal validation measures because we did not want to consider the ground truths of the datasets. We know that there are countless measures and that each one has a specific bias and can favor a class of algorithms over others. To reduce this problem, we decided to select and combine measures adopting different clustering criteria. We grouped them into six groups according to Table 2. Each measure, or group of measures, has its own properties. To consider different biases, we aggregated them into a single value based on an average ranking of performance across the algorithm portfolio, in order to smooth or compensate their individual influence on algorithms which employ similar clustering criteria.
The first stage of the framework is the pre-processing of the meta-dataset. In this stage, the data were normalized using the Box-Cox and Z transformations. Next, a feature selection process is performed in order to find a subset of meta-features that are more related to the clustering algorithms’ performances, allowing for obtaining smoother 2D projections relating the properties of the instances and the performance of the clustering algorithms. MATILDA’s automated feature selection process defines the Pearson correlation coefficient as a similarity metric between features. K-means clustering is used to identify groups of similar features, with the number of groups defined by a Silhouette measure. K is a parameter that corresponds to a maximum number of clusters in the feature selection step. According to Appendix A, values were tested in the range 6–10. The value 10 was defined based on the tests. To select one feature per group, the algorithm first projects the subset of selected features into two dimensions using Principal Components Analysis (PCA) and Random Forests are used to predict whether an instance is easy or not for a given algorithm. Instances are labeled as easy or hard according to the performance threshold adopted for defining a good performance. In this work, we employed a threshold of 0.5, so that an instance for which an average ranking performance below 0.5 is achieved is classified as hard; otherwise, it is considered easy. Then, the subset of features that results in the most accurate Random Forest models is selected. This routine can be computationally intensive, since it tests all possible feature combinations if they are less than 1000 or uses the combination of a Genetic Algorithm and a look-up table otherwise.
Next, the main step for the generation of the instance space is the projection model. This algorithm was idealized as an optimization problem. It is a procedure of dimensionality reduction that projects the instances in a 2D space so that a linear trend is observed between the selected meta-features and the performance of the algorithms under analysis. The choice of projection to 2D is an experimental decision and any such dimensionality reduction produces a loss of information. Nevertheless, the use of 2D projection allows for visualizing better the distribution of the selected meta-features along the instance space and also of the algorithms’ performances variations. The 2D projection facilitates the computational requirements to establish the algorithm footprints too, corresponding to areas where the algorithms perform well. According to Muñoz et al. [5], given a (meta-)feature matrix F = [ f 1 f 2 f 3 ] R m × n and algorithm performance vector y R n (which can also be generalized to a matrix containing multiple evaluations), where m is the number of features and n is the number of problem instances, we achieve an ideal projection of the instances if we can find A r R 2 × m , B r R m × 2 and c r R 2 which minimizes the approximation error
| | F F ^ | | F 2 + | | y y ^ | | F 2
such that
Z = A r F
F ^ = B r Z
y ^ = c r Z .
The problem was solved numerically by MATILDA using BIPOP-CMA-ES (Bi-population Covariance Matrix Adaptation Evolution Strategy) [5].
The next step was the use of a supervised learning meta-model for algorithm selection. MATILDA trains a series of SVM models that predict whether an algorithm will perform well or not for a given instance, given a threshold of good and bad performance, which was set as 0.5 in this work. SVMs are trained using the selected meta-features as inputs and the performance measurement of each algorithm as output. The implementation of SVM uses a Gaussian kernel. Their hyper-parameters are tuned by Bayesian Optimization. The model uses a stratified k-fold cross-validation technique and gives a probability for each instance of belonging to a given class by fitting a posterior probability model ( k = 5 by default).
A central concept in ISA are the footprints, which represent the areas in the instance space where an algorithm is expected to perform well. The areas are calculated by Delaunay triangulation, considering the removal of contradictory evidence. For each of these areas’ measures of density, which would be the number of instances per unit area, and purity, which would be the number of instances of good performance per number of instances in the region, are extracted. When areas of two or more algorithms overlap, the purity measure is used to determine which algorithm should predominate.
Table A1 in Appendix A shows the MATILDA parameter values adopted in this work. Most of the values adopted are default. Others were varied and the resulting projections were inspected for choosing suitable values. This examination is necessary to find a projection allowing for discriminating better the algorithm’s performances, whilst also showing smooth trends concerning the meta-features and algorithms’ performance values across the instance space. The aim of ISA is to gain insights and, if the parameters are not set suitably, few insights are available. For example, if the goodness threshold is not carefully chosen, all algorithms could be labeled good, and there will be little opportunity to allow ISA to distinguish strengths and weaknesses. While the default parameter settings in MATILDA have been chosen through many years of testing a variety of problems (see matilda.unimelb.edu.au for the collection of library problems, last accessed on 18 March 2021), there is still sometimes a need to explore a range of parameters to ensure the most insightful results are produced. In this study, experiments were conducted exploring different thresholds related to algorithmic performance and the selection of meta-features to arrive at some insightful visualisations, but the chosen parameters are quite robust.

4. Building an Instance Space from Artificial Datasets

All off the pre-processing steps of MATILDA were performed, with the exception of outlier detection. Experimental tests revealed that the exclusion of outliers reduced the quality of the projection in terms of the distribution of the instances in the 2D space, impairing the discrimination of the algorithms’ performances. All codes were executed on an Intel Core i7-7500U CPU with 2.70 GHz (Santa Clara, CA, USA), 16 GB RAM, and the Microsoft Windows 10 operating system (Redmond, WA, USA). The clustering algorithms ran once using their default parameter values (The code and datasets can be found at https://github.com/ml-research-clustering/meta-analysis, last accessed on 18 March 2021), except for the assumed number of clusters, which was adjusted to each dataset according to the Silhouette Method [59] from the NbClust R package. There are two details of implementation that deserve to be mentioned, both related to parameters of the employed function. First, the number of clusters to be considered was limited between the minimum and maximum values of the actual numbers of clusters all datasets from a given group contain. These limits were imposed to reduce the search space and the algorithmic runtime, since we are dealing with a large amount of datasets. If ground truths are not available, the minimum and maximum values can be set from 2 to n 1 , where n is the size of the dataset. Secondly, the clustering algorithm used is varied among: K-Means, K-Medians, Closest Centroid, Single Linkage, Complete Linkage, Average Linkage, Ward, and McQuitty methods.

4.1. An Average Ranking for Performance Measures

As previously described, the internal validation indexes listed in Table 2 were aggregated into a single measure by taking an average ranking of performance of the algorithms (avg_rank). The idea is to combine multiple performance views of the clustering algorithms. The ranking uses the rank function of R. It is calculated as follows: (i) for each performance measure and dataset, a ranking of the algorithms is made, so that algorithms with better performance receive a better position in the rank compared to algorithms of lower performance; (ii) Then, the average and standard deviation of the ranking positions assigned to each algorithm are calculated. Standardization based on median and median deviation was applied to the ranking results, as well as Min-Max normalization. As a result, the ranking values range between 0 and 1, with larger values corresponding to better results.
All test instances were included in the analysis, divided according to the geometric shape of the clusters when projected in a 2D space. They were categorized into Gaussian, ellipsoidal or multiple shapes. The algorithm performance metric considered absolute performance, with a good performance identified when an algorithm reaches an avg_rank performance metric greater than or equal to 0.50 for the given instance. This value corresponds to the threshold of good performance (opts.perf.epsilon). According to Appendix A, this parameter was varied in the range of 0.30–0.80. Adjusting this parameter allows for obtaining a projection that discriminates the performance of the algorithms better and makes it possible to visualize their footprints, that is, the regions of the space of instances in which the algorithms achieve good performance.
As mentioned earlier, MATILDA has the option of using a clustering-based feature selection method. It uses Pearson’s correlation coefficient as a measure of similarity and the k-Means algorithm to identify groups of similar features. The routine also embeds a silhouette analysis to infer the number of clusters and, therefore, the number of features to keep afterwards. A threshold of 0.70 was used for the silhouette coefficient in this setting. According to Appendix A, this parameter was varied in the range 0.50–0.80. During execution, the MATILDA system itself recommends using a higher or lower value in order to obtain an optimized number of clusters in the feature selection process.
Table 3 presents the mean and standard deviation of avg_rank measure across all test instances. The CC algorithm obtained the best average performance, while the KMO algorithm obtained less favorable results. Figure 3 further reveals that the partitional clustering algorithms performed poorly in the group of ellipsoidal datasets when compared with the hierarchical algorithms. The evidence found in these descriptive summary statistical results suggests that we should mostly rely on CC for a given clustering dataset, but does this algorithm have weaknesses? In addition, does the weakest algorithm, KMO, have hidden strengths that a more nuanced analysis would reveal? We aim to explore how ISA can fill this gap to gain greater insights than statistical averages can offer.

4.2. Feature Selection and Projection in the Instance Space

Ten meta-features were selected from the feature space F by MATILDA to be employed in the ISA analysis. Equation (5) presents the projection matrix that allows the transformation of instances from the 10D feature space to the 2D instance space. We can notice the presence of meta-features from all categories except from Network Centrality presented in Table 1, stressing the need for considering multiple views in order to properly characterize the structure of diverse datasets. Although no network centrality measures are present in the final set of meta-features chosen, the clust_coef meta-feature is also extracted from a proximity graph built from data.
Z 1 Z 2 = 0.124 0.495 0.572 0.026 0.147 0.049 0.231 0.152 0.435 0.025 0.505 0.186 0.154 0.266 0.205 0.254 0.131 0.312 0.007 0.226 ratio _ ex _ ftr skewness kurtosis perc _ out ratio _ pca avg _ abs _ cor low _ dist clust _ coef intr _ dim avg _ nnd
Figure 4 illustrates the characteristics of the 380 datasets projected in the generated instance space colored according to group (namely ellipsoidal, Gaussian and multiple shapes), number of examples, number of attributes and number of clusters defined as ground truth in the original datasets. There is a clear distinction between dataset groups. The ellipsoidal and multiply shaped datasets are spread in different regions of the instance space. Gaussian datasets occupy a region below them, with some degree of overlap. Among the Gaussian datasets, there is a set that is uniformly arranged in the lower region of the instance space. Those are the G2 datasets [42] that have the same number of observations (n = 2048), the same number of clusters (k = 2) and the number of attributes ranging from 2 to 1024. They also have an overlap parameter varying from 0 to 100. These datasets model each cluster as a separate Gaussian, which may or not overlap depending on the spread of their distributions.
We can see from the figure that the numbers of examples, attributes and clusters also are decisive in the distribution of datasets across the instance space. These characteristics were not directly used to project the instances, although the number of examples and attributes are used to compute some of the dimensionality measures. Datasets with more input attributes are concentrated in the bottom of the instance space, and their dimensionalities reduce towards the upper right quadrant of the projected space. For the number of clusters, although there are some mixed results, in general datasets with a lower number of clusters are concentrated towards the right of the instance space. Concerning the number of examples, datasets with fewer examples mostly have a z 2 coordinate larger than zero.
The distribution of the selected meta-features in the instance space is shown by Figure 5. All values were scaled to [0, 1]. In this case, the following observations can be drawn:
  • avg_abs_cor: the values of this meta-feature increase from left to right in the instance space. This meta-feature computes the correlations between all input attributes a dataset has and averages their absolute values. Higher values are obtained for datasets with more redundant input attributes;
  • avg_nnd: the values of this meta-feature tend to increase from right to left in the instance space. The degree captures more precisely the effective level of cohesiveness and affinity due to the actual interaction strength between nodes of the proximity graph built from a dataset [44]. A higher degree is expected for datasets with more examples near each other, so that the proximity graph shows more connections;
  • clust_coef: the values of this meta-feature increase from bottom-left to the top-right of the instance space. The clustering coefficient of a vertex measures the ratio of the number of edges between its neighbors and the maximum number of edges that could possibly exist between them. A higher value is obtained for datasets with a more dense clustering structure;
  • intr_dim: this meta-feature is characterized by having higher values for the G2 datasets. It measures the squared average distance between all points divided by the variance of the distances [45]. G2 datasets have a certain degree of overlap between clusters, justifying the higher values achieved;
  • kurtosis: the values of this meta-feature tend to increase smoothly towards the bottom-left of the instance space. It calculates the Mardia’s multivariate kurtosis coefficient [43] of the dataset’s input attributes. A higher value reflects more complex datasets in terms of data distribution;
  • low_dist: the values of this meta-feature tend to increase towards the bottom-right of the instance space. This measure computes the percentage of points with a distance lower than the mean distance between all data points in a dataset, minus its standard deviation. Higher values tend to be obtained for datasets with a greater degree of overlap between clusters;
  • perc_out: the values of this meta-feature increase from bottom-right to the top-left of the instance space. In this work, outliers are considered points whose distance is less than q 1 1.5 d and higher than q 3 + 1.5 d , where q 1 and q 3 are the first and third quartiles, respectively, and d = q 3 q 1 . Essentially, this is an indicator of data quality that shows the ability of algorithms to deal with datasets that have non-uniform distribution;
  • ratio_ex_ftr: the values of this meta-feature increase from bottom-left to the top-right of the instance space. This measure reflects the ratio between the number of examples and the number of attributes. It gives an indication of data sparsity, with lower values obtained for sparse datasets;
  • ratio_pca: the values of this meta-feature tend to increase towards the up-right of the instance space. It is a rough measure of the proportion of relevant dimensions the dataset has, when compared to the number of principal components needed to represent 95% of data variability. High values of this measure indicate more complex relationships of the input variables [60];
  • skewness: the values of this meta-feature tend to increase sharply towards the bottom-left of the instance space. It calculates the Mardia’s multivariate skewness coefficient [43]. A higher value reflects more complex datasets in terms of data distribution.

4.3. Algorithm Selection in the Instance Space—Performance Prediction

MATILDA uses SVM-based meta-models to predict in which regions of the instance space a particular algorithm would be the best choice for a given instance. Herewith, a specific module trains one SVM classification model per algorithm that predicts whether the algorithm will perform well or not for a given instance (see liveDemoIS.mlx file available at https://github.com/andremun/InstanceSpace, last accessed on 18 March 2021). It does this using the MATLAB fitcsvm package. In this process, the following operations are performed: (i) it produces resubstitution results using a stratified K-fold cross-validation (CV); (ii) it gives a probability for each instance of belonging to a given class, by fitting a posterior probability model; (iii) it uses Bayesian Optimization to fine tune the SVM hyper-parameters { C , γ } that are the cost or regularization term and the radial-basis function kernel (RBF) hyper-parameter, respectively.
Table 4 shows the results obtained by the SVM predictors for each clustering algorithm (10 in total), as well as the results of an idealized oracle that always picks the right algorithm, and a selector that combines the decisions made by all the SVMs models to predict the likely best algorithm for each instance using the highest precision SVM to break ties. This assumes that the most precise SVM (CV-Precision) is the most reliable. The table includes the average and standard deviation of the performance measure across all instances (the average ranking obtained for all datasets), probability of good performance and 10-fold CV performance of the SVM models (accuracy, precision and recall). The SVM parameter values employed are also shown. The average ranking and its standard deviation are the expected performance if we were to use each algorithm across all instances. As already mentioned in the paper, the threshold for good performance defined in MATILDA was set as an avg_rank performance metric of at least 0.50 . The probability of good performance is based on this parameter and represents the fraction of instances for which the performance measure is greater than or equal to 0.50 . The clustering algorithm that presented the best performance on average was CC (probability of good 0.85), while the worst was KMO (probability of good 0.35). The results also indicate that the class of hierarchical algorithms in general presented better performance. The Selector is successful since it has a probability of 0.91 of selecting an algorithm with good performance and a CV-Precision of 91.30—better than always selecting the best on-average algorithm, which is CC (0.85). Nonetheless, the relatively low CV Recall of the Selector indicates that it could be further improved.

4.4. Footprint Analysis and Algorithm Portfolio

One of the most relevant and characteristic results of ISA is the presentation of the regions of the instance space in which the algorithms perform well, that is, the footprints of the algorithm, supporting objective evaluation of algorithmic power and highlighting its strengths and weaknesses. According to Muñoz et al. [5], the algorithm that generates the footprints uses a combination of two approaches that were used in previous works. In its first step, the algorithm measures the Euclidian distances between two instances, and marks one of them for elimination if the distance is less than a threshold, α . This is done to avoid numerical instability with the triangulation algorithm, which is applied as the second step. As a third step, the algorithm finds the concave hull, by removing any triangle with edges larger than another threshold, β . As a fourth and final step, the algorithm calculates the density and purity of each triangle in the concave hull. The density, as the number of instances per area unit, and purity, as the number of instances of good performance (considering an avg_rank performance metric of at least 0.50) per number of instances of the region. If a triangle does not fulfill the density and purity limits, it is removed from the hull. When areas of two or more algorithms overlap, the purity measure is used to determine which algorithm should predominate. While we are not able to experiment with these default parameters, we can confirm that the resulting footprints make visual sense, and the parameters appear to behave in the intended manner, for our application, as well as all for the MATILDA library case studies. Changes to these parameters would merely affect the formation of footprints and the calculation of their areas, which is not a core requirement for gaining the visual insights we seek.
Figure 6 shows the footprints of the algorithms in the instance space. An interesting aspect to be observed is that all hierarchical algorithms perform well in the ellipsoidal datasets (center-left region), which are more complex in terms of some basic characteristics such as number of attributes and clusters. This was not observed for any of the partitional algorithms. We can also observe that the CC algorithm has the largest footprint in terms of coverage area across the instance space. On the other hand, the KMO algorithm has the smallest footprint area Δ . Some other interesting observations from Figure 6 are:
  • there is a large variation on the partitional algorithms’ performances, although all of them algorithms tended to present better results for datasets towards the bottom-right of the instance space, with exception from a small area in the case of KME. KMO and MBK present a large area of bad performance. The FCM variant is able to handle better a larger portion of the datasets;
  • the hierarchical clustering variants tended to present bad results for some datasets of the G2 group;
  • the WM algorithm presents a performance mostly complementary to that of the partitional algorithms.
Table 5 presents the results for the footprint analysis. The areas Δ , as well as their density ρ and purity ϕ measures, are detailed. Usually, the obtainment of large normalized areas (30% onwards) with density about one and purity above 70–80% is evidence that the algorithm is strong in the instance space. The CC algorithm showed the best performance ( Δ = 0.87 ), followed by the SL ( Δ = 0.82 ) and FCM ( Δ = 0.82 ). CC performed well on ellipsoidal, multiple shapes and Gaussian datasets, with the exception of a region formed by G2 datasets. SL obtained a similar result. FCM performed well in Gaussian datasets and multiple shapes, but did not obtain a good performance in ellipsoidal datasets. The worst performances were obtained by the KMO ( Δ = 0.19 ), WM ( Δ = 0.27 ) and MBK ( Δ = 0.35 ) algorithms. The hierarchical algorithms in general showed better performance, which can be evidenced by the measures of area and density.
An important aspect in relation to ISA is the possibility of establishing a visual relationship between features and footprints, in order to highlight the strengths and weaknesses of the algorithms. A simultaneous analysis of Figure 5 and Figure 6 allows us to make the following statements about the relationships between datasets, meta-features and the performance of the algorithms:
  • partitional algorithms results are directly related to ratio_pca, low_dist and avg_abs_cor, that is, those meta-features show where the algorithms perform well. On the other hand, their weaknesses are related to perc_out;
  • regarding hierarchical algorithms, perc_out has a direct relationship with good performance, while avg_abs_cor and low_dist show the most difficult instances;
  • similar relationships can also be verified when considering only the Gaussian datasets;
  • in the group of ellipsoidal datasets, features avg_nnd and skewness show the strength of hierarchical algorithms, while features avg_abs_cor and clust_coef show their weaknesses;
  • partitional algorithms do not have in general a uniform behavior for ellipsoidal datasets;
  • in the group of multiple shapes, features ratio_pca, ratio_ex_ftr and clust_coef show the strength of hierarchical algorithms. KME, KMD and FCM present a direct relationship with those features, while KMO and MBK present an inverse relationship with perc_out;
  • the best algorithm, CC, has its strengths evidenced by features clust_coef and ratio_ex_ftr, and weaknesses by features ratio_pca and low_dist;
  • the worst algorithm, KMO, has its strengths evidenced by features low_dist and avg_abs_cor, and weaknesses by features perc_out and kurtosis.
We can conclude, therefore, that for the instances under study, the partitional and hierarchical algorithms in general tend to present inverse behaviors in relation to performance across the instance space.
Figure 7 shows the footprints of the algorithm portfolio. The ISA’s results are consistent with those from the simple average performance analysis of algorithms, but with more nuanced insight possible. The SVM results recommend the CC algorithm for most instances of the three groups of datasets, but not all. There is also a recommendation for the SL and FCM algorithms in a small region of the Gaussian datasets.

5. Plotting Real Datasets in Instance Space

The instance space presented in the previous section was generated from artificial datasets with known clustering properties and structures. The adoption of this restriction resides in the fact that synthetic data allow a clearer and more objective understanding of the study results. They allow a more effective control over eventual relationships between selected meta-features and performance measures. However, reaching more comprehensive insights requires the use of datasets that reflect typical difficulties of real cases. We present in this section an ISA that incorporates real data, in order to verify in which regions of the instance space these datasets would be located and whether they influence the results obtained previously. This allows us to assess the difficulty level of real datasets and to examine the real-world-likeness of synthetic datasets. Only the main findings are highlighted, given that the details of the methodology employed are the same and have already been described in Section 3.

5.1. Generating an Instance Space from Combined Artificial and Real Datasets

A new instance space was generated from a total of 599 datasets, where 380 are the same artificial datasets used in Section 3 and another 219 real datasets were collected from the OpenML repository and used as benchmark for clustering algorithms in [11]. In these datasets, the number of examples ranges from 100 to 5000, the number of attributes ranges from 2 to 168, and the number of embedded clusters ranges from 2 to 30. One must notice that these datasets are representatives of classification problems. In classification problems, the objective is to find a decision model able to partition the data items into some previously defined categories. The datasets are then labeled and the model is trained to somewhat reproduce such labels. A same class can have data points spanning different regions of the input space. This differs from clustering, in which there are no pre-defined categories and the groups are usually based on their proximity. Whilst a problem for which the instances are distributed uniformly all over the input space with a clear linear decision border between the classes can be considered quite easy for classification, this is not the case for clustering (in fact, there is no cluster structure for data points randomly distributed in the input space). Therefore, although the usage of classification datasets in the evaluation of clustering algorithms is largely employed, some care must be taken in the interpretation of the obtained results.
The ground truths of the real datasets were not considered in the analysis. The same procedure used with artificial datasets was employed: the clusters were defined by the Silhouette method, using the implementation of the NbClust package from R. The categorization of artificial datasets described in the previous section has remained unchanged (namely ellipsoidal, Gaussian and multiple shapes). Equation (6) presents the new projection matrix. We can see that 10 meta-features were selected. A total of six remained in the projection matrix after the inclusion of the real datasets: kurtosis, perc_out, ratio_pca, avg_abs_cor, low_dist and avg_nnd. This indicates that these measures have a relevant influence in the characterization of both types of datasets.
Z 1 Z 2 = 0.110 0.249 0.084 0.195 0.201 0.094 0.270 0.157 0.071 0.287 0.178 0.224 0.197 0.019 0.197 0.013 0.163 0.023 0.077 0.132 number _ ftr kurtosis multi _ norm perc _ out ratio _ pca avg _ abs _ cor low _ dist contrast hub _ score avg _ nnd
Figure 8 shows the characteristics of the datasets projected in the instance space. The number of clusters for the real datasets correspond to their number of classes. Figure 9 and Figure 10 show the distribution of selected meta-features and algorithms footprints in the instance space, respectively.
Figure 8 reveals the region of the instance space where the real datasets were plotted. They are concentrated in the central and upper left region of the instance space, partially overlapping the less complex Gaussian datasets and the multiple shapes datasets. They also fill a central and lower left area not occupied by artificial datasets, which is characterized by a reduced number of examples and intermediate number of attributes and clusters. With very few exceptions, we can verify that, in general, they are not located in the areas occupied by the ellipsoidal and the G2 datasets, that is, the most complex ones. This finding highlights the importance that artificial datasets can have in experiments and that artificial and real datasets can act as complementary instances, providing a more comprehensive analysis of clustering problems.

5.2. Performance Prediction, Footprint Analysis and Algorithm Portfolio

Table 6 shows the results for each algorithm, plus the results of an idealized oracle that always picks the right algorithm, and our selector that combines the decisions made by the 10 SVMs (one per clustering algorithm). The columns of this table have the same interpretation as those in Table 4. CC presented the best performance on average (probability of good 0.86), while MBK was the worst (probability of good 0.39). The results also confirm what was obtained previously regarding the best average performance of the hierarchical algorithms. The AL algorithm obtained an average performance lower than that achieved by the best algorithm (probability of good 0.76). However, it was the best algorithm regarding the results of the cross-validation measures (accuracy, precision and recall). The Selector is successful since it has a probability of 0.95 of selecting an algorithm with good performance—better than always selecting the best on-average algorithm CC. The results show that the Selector has better performance, both in terms of precision and average ranking than using always CC, which is the best performing algorithm. However, the difference in the performance to the Oracle, the low CV performance for the KMO algorithm and a relatively low CV. Recall again indicate that the Selector could be further improved.
Through the analysis of Figure 9 and Figure 10, we can verify that partitional algorithms good results are directly related to ratio_pca, low_dist and avg_abs_cor. On the other hand, their weaknesses are related to perc_out. Regarding hierarchical algorithms, perc_out has a direct relationship with good performance, while avg_abs_cor and low_dist show the most difficult instances. We can conclude that, for the instances under study, the partitional and hierarchical algorithms in general tend to present inverse behaviors in relation to performance across the instance space. The insertion of the real datasets, therefore, maintained the main findings verified in the preliminary analysis. Figure 10 shows that all hierarchical algorithms perform well in the ellipsoidal datasets (lower region), as found in the analysis restricted to artificial datasets. This was not observed for partitional algorithms, with the exception of KMO. We see in Table 7, which presents the area Δ , density ρ , and purity ϕ of the algorithms’ footprints, that the AL algorithm has the largest footprint in terms of coverage area across the instance space. On the other hand, the KMO algorithm has the smallest footprint area. An interesting finding is that the hierarchical algorithms performed well when applied to real datasets, in contrast to the partitional algorithms. Nonetheless, we must keep in mind that the real datasets used are originally from classification and not clustering problems and this may have affected some of the results.
Figure 11 shows the footprints of the algorithm portfolio. In this case, the SVM results recommend the AL algorithm for most instances of the four groups of datasets, but not all. There are also recommendations for the CC and MBK algorithms.

6. Conclusions

In this paper, we have performed the first instance space analysis of the performance of well-known clustering algorithms on a representative sample of clustering datasets. We considered a new set of features that aim to reveal the presence of different types of structures in a dataset. We defined as performance measure a combined ranking of the performance of the algorithms as measured by internal validation indexes of different biases.
One of the limitations of this study resides in the need to pre-define the number of clusters to be used as a parameter in the clustering algorithms. We opted to monitor a Silhouette metric in the evaluation of the clustering algorithms’ results. It is also important to comment that the adequate definition of the good performance threshold plays a major role in the generation of an instance space that discriminates the algorithms’ performances. We have experimentally varied the values of the most sensitive parameters in order to obtain consistent and robust insights about the distribution of the instances in the space and to reveal subtleties on algorithmic performance.
An interesting finding was the good performance of the hierarchical algorithms in relation to the partitional algorithms when applied to ellipsoidal datasets. It should also be noted that, for the instances under study, the partitional and hierarchical algorithms in general tend to present inverse behaviors in relation to performance across the instance space. Partitional algorithms had their strengths evidenced by measures of dimensionality and neighborhood-based measures. Its weaknesses were revealed by measures of density. The opposite occurred for hierarchical algorithms. Different behaviors were observed when the analysis was segregated into groups of datasets. In general, hierarchical algorithms tend to perform better in more complex datasets. Finally, we can conclude that the inclusion of real datasets employed in related work maintained the essence of the results achieved when considering only artificial datasets, showing consistency in the methodology used. However, the main finding of this work is the observation that ISA goes further and indicates, in addition to the best algorithm for a given instance, which algorithm would be recommended for an instance whose characteristics lie in a certain region of the instance space. We can conclude, therefore, that ISA offers a unique perspective in relation to the methodologies of clustering analysis widely used in the literature.
In future works, we will expand the feature space, aiming for a more accurate and comprehensive characterization of clustering datasets. The results of the MATILDA’s feature selection showed that measures based on network centrality were not relevant in this process and this should be better investigated. Some characterization measures based on distance and entropy will be added to the study too, as well as other clustering algorithms. Some algorithms with different characteristics are already being tested, such as Affinity Propagation, Spectral Clustering, Bagged Clustering and model-based algorithms. Finally, a method will be developed to generate clustering datasets at controlled locations of the instance space. This is a natural follow-up in research using the ISA framework. The objective is to create instances that are located in regions not occupied by the test datasets, in order to diversify the instance space. Consequently, algorithms can be challenged with datasets of different difficulty levels from those commonly used as benchmarks.

Author Contributions

Conceptualization, A.C.L. and K.S.-M.; methodology, L.H.d.S.F., A.C.L. and K.S.-M.; software, L.H.d.S.F. and K.S.-M.; validation, L.H.d.S.F., A.C.L. and K.S.-M.; formal analysis, A.C.L. and K.S.-M.; investigation, L.H.d.S.F.; resources, L.H.d.S.F., A.C.L. and K.S.-M.; data curation, L.H.d.S.F.; writing—original draft preparation, L.H.d.S.F., A.C.L. and K.S.-M.; writing—review and editing, L.H.d.S.F., A.C.L. and K.S.-M.; visualization, L.H.d.S.F.; supervision, A.C.L. and K.S.-M.; project administration, A.C.L. and K.S.-M.; funding acquisition, A.C.L. and K.S.-M. All authors have read and agreed to the published version of the manuscript.

Funding

This study was financed in part by the Coordenação de Aperfeiçoamento de Pessoal de Nível Superior—Brazil (CAPES)—Finance Code 001.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data and tools used in this study are openly available in a dedicated GitHub repository (https://github.com/ml-research-clustering/meta-analysis, last accessed on 18 March 2021). The original sources are detailed in Section 3.

Acknowledgments

The authors would like to thank the Australian Research Council for their support under the Laureate Fellowship scheme (FL140100012).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. MATILDA Parameters

Table A1 shows the MATILDA parameters adopted in this study.
Table A1. MATILDA parameters.
Table A1. MATILDA parameters.
ParameterDescriptionTest RangeFinal Value
opts.perf.MaxPerfTrue if Y is a performance measure to maximize, false to minimizedefaulttrue
opts.perf.AbsPerfTrue if an absolute performance measure, false if relativedefaulttrue
opts.perf.epsilonThreshold of good performance0.30–0.800.50
opts.general.betaThresholdBeta-easy threshold0.45–0.750.50
opts.auto.preprocAutomatic preprocessing ondefaulttrue
opts.bound.flagBound the outlierstrue/falsefalse
opts.norm.flagNormalize/Standardize the datatrue/falsetrue
opts.auto.featselAutomatic feature selection ondefaulttrue
opts.corr.flagRun feature selection by correlation between performance and featuresdefaulttrue
opts.corr.thresholdTop N features (by correlation) per algorithm that are selecteddefault10
opts.clust.flagRun feature selection by clusteringdefaulttrue
opts.clust.KDEFAULTDefault maximum number of clusters6–1010
opts.clust.SILTHRESHOLDMinimum accepted value for the average silhouette value0.50–0.800.70
opts.clust.NTREESNumber of trees for the Random Forestdefault50
opts.clust.MaxIterMaximum number of iterationsdefault1000
opts.clust.ReplicatesNumber of replicatesdefault100
opts.pilot.analyticCalculate the analytical or numerical solutiondefaultfalse
opts.pilot.ntriesNumber of attempts carried out by PBLDRdefault10
opts.cloister.pvalMaximum correlation coefficientdefault0.05
opts.cloister.cthresThreshold of non-correlationdefault0.70
opts.pythia.cvfoldsnumber of folds of the stratified cross-validation partitiondefault5
opts.pythia.useweightsweighted (true) or unweighted classification is performeddefaultfalse
opts.trace.usesimUse the actual or simulated data to calculate the footprintsdefaulttrue
opts.trace.RHODensity thresholddefault5
opts.trace.PIPurity thresholddefault0.55
opts.selvars.smallscaleflagTrue if you want to do a small scale experimentdefaultfalse
opts.selvars.smallscalePercentage of instances to be kept for a small scale experimentdefault0.50
opts.selvars.fileidxflagActivates selection by indexingdefaultfalse
opts.selvars.fileidxName of a .csv file that contains in one column the indexes of the instances to be takendefault“ ”
opts.outputs.webOutput files employed to draw the figures in MATILDA’s web toolsdefaultfalse
opts.outputs.pngOutput as figures files for post-processing and analysisdefaulttrue

Appendix B. Parameters for Clustering Algorithms

Table A2 presents the parameters values of the clustering algorithms adopted in this work.
Table A2. Parameters for clustering algorithms.
Table A2. Parameters for clustering algorithms.
KME (Stats Package R, kmeans Function)
ParameterDescriptionTest rangeFinal value
iter.maxThe maximum number of iterations allowed.default10
nstartHow many random sets should be chosen.default1
algorithmAlgorithm implementation.default“Hartigan–Wong”
KMD (Gmedian package R, kGmedian function)
ParameterDescriptionTest rangeFinal value
gammaValue of the constant controlling the descent steps.default1
alphaRate of decrease of the descent steps.default0.75
nstartNumber of times the algorithm is ran, with random sets of initialization centers chosen among the observations.default10
nstartkmeansNumber of initialization points in the kmeans function for starting point of kGmedian.default10
iter.maxMaximum number of iterations considered in the kmeans function for starting point of kGmedian.default20
KMO (kmed package R, fastkmed function)
ParameterDescriptionTest rangeFinal value
iterateA number of iterations for the clustering algorithm.default10
initA vector of initial objects as the cluster medoids.defaultNULL
Hierarchical algorithms: SL, CL, AL, CC, WM (stats package R, hclust function)
ParameterDescriptionTest rangeFinal value
dA dissimilarity structure as produced by dist.distance matrixinput
methodThe agglomeration method to be used.--
membersNULL or a vector with length size of d.defaultNULL
MBK (ClusterR package R, MiniBatchKmeans function)
ParameterDescriptionTest rangeFinal value
batch_sizeThe size of the mini batches.default10
num_initNumber of times the algorithm will be run with different centroid seeds.default1
max_itersThe maximum number of clustering iterations.default100
init_fractionPercentage of data to use for the initialization centroids.default1
initializerThe method of initialization.default“kmeans++”
FCM (e1071 package R, cmeans function)
ParameterDescriptionTest rangeFinal value
iter.maxMaximum number of iterations.default100
distDistance: “euclidean” or “manhattan”.default“euclidean”
mA number greater than 1 giving the degree of fuzzification.default2
rate.parA number between 0 and 1 giving the parameter of the learning rate for the online variant.default0.3

References

  1. Calvetti, D.; Somersalo, E. Mathematics of Data Science: A Computational Approach to Clustering and Classification; SIAM: Philadelphia, PA, USA, 2020; Volume 1. [Google Scholar]
  2. Vilalta, R.; Drissi, Y. A perspective view and survey of meta-learning. Artif. Intell. Rev. 2002, 18, 77–95. [Google Scholar] [CrossRef]
  3. Smith-Miles, K.A. Cross-disciplinary perspectives on meta-learning for algorithm selection. ACM Comput. Surv. (CSUR) 2009, 41, 6. [Google Scholar] [CrossRef]
  4. Vanschoren, J. Meta-learning: A survey. arXiv 2018, arXiv:1810.03548. [Google Scholar]
  5. Munoz, M.A.; Villanova, L.; Baatar, D.; Smith-Miles, K. Instance spaces for machine learning classification. Mach. Learn. 2018, 107, 109–147. [Google Scholar] [CrossRef] [Green Version]
  6. Kang, Y.; Hyndman, R.J.; Smith-Miles, K. Visualising forecasting algorithm performance using time series instance spaces. Int. J. Forecast. 2017, 33, 345–358. [Google Scholar] [CrossRef] [Green Version]
  7. Kandanaarachchi, S.; Muñoz, M.A.; Hyndman, R.J.; Smith-Miles, K. On normalization and algorithm selection for unsupervised outlier detection. Data Min. Knowl. Discov. 2020, 34, 309–354. [Google Scholar] [CrossRef] [Green Version]
  8. Smith-Miles, K.; Baatar, D.; Wreford, B.; Lewis, R. Towards objective measures of algorithm performance across instance space. Comput. Oper. Res. 2014, 45, 12–24. [Google Scholar] [CrossRef]
  9. Muñoz, M.A.; Smith-Miles, K.A. Performance analysis of continuous black-box optimization algorithms via footprints in instance space. Evol. Comput. 2017, 25, 529–554. [Google Scholar] [CrossRef] [PubMed]
  10. Ferrari, D.G.; De Castro, L.N. Clustering algorithm selection by meta-learning systems: A new distance-based problem characterization and ranking combination methods. Inf. Sci. 2015, 301, 181–194. [Google Scholar] [CrossRef]
  11. Pimentel, B.A.; de Carvalho, A.C. A new data characterization for selecting clustering algorithms using meta-learning. Inf. Sci. 2019, 477, 203–219. [Google Scholar] [CrossRef]
  12. Brazdil, P.; Carrier, C.G.; Soares, C.; Vilalta, R. Metalearning: Applications to Data Mining; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  13. Rice, J.R. The algorithm selection problem. In Advances in Computers; Elsevier: Amsterdam, The Netherlands, 1976; Volume 15, pp. 65–118. [Google Scholar]
  14. Muñoz, M.A.; Smith-Miles, K. Generating new space-filling test instances for continuous black-box optimization. Evol. Comput. 2020, 28, 379–404. [Google Scholar] [CrossRef]
  15. Muñoz, M.A.; Smith-Miles, K. Generating custom classification datasets by targeting the instance space. In Proceedings of the Genetic and Evolutionary Computation Conference Companion, Berlin, Germany, 15–19 July 2017; ACM: New York, NY, USA, 2017; pp. 1582–1588. [Google Scholar]
  16. Lemke, C.; Budka, M.; Gabrys, B. Metalearning: A survey of trends and technologies. Artif. Intell. Rev. 2015, 44, 117–130. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. De Souto, M.C.; Prudencio, R.B.; Soares, R.G.; De Araujo, D.S.; Costa, I.G.; Ludermir, T.B.; Schliep, A. Ranking and selecting clustering algorithms using a meta-learning approach. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks, Hong Kong, China, 1–8 June 2008; pp. 3729–3735. [Google Scholar]
  18. Soares, R.G.; Ludermir, T.B.; De Carvalho, F.A. An analysis of meta-learning techniques for ranking clustering algorithms applied to artificial data. In Proceedings of the International Conference on Artificial Neural Networks, Limassol, Cyprus, 14–17 September 2009; Springer: Berlin/Heidelberg, Germany, 2009; pp. 131–140. [Google Scholar]
  19. Nascimento, A.C.; Prudêncio, R.B.; De Souto, M.C.; Costa, I.G. Mining rules for the automatic selection process of clustering methods applied to cancer gene expression data. In Proceedings of the International Conference on Artificial Neural Networks, Limassol, Cyprus, 14–17 September 2009; pp. 20–29. [Google Scholar]
  20. Ferrari, D.G.; de Castro, L.N. Clustering algorithm recommendation: A meta-learning approach. In Proceedings of the International Conference on Swarm, Evolutionary, and Memetic Computing, Bhubaneswar, India, 20–22 December 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 143–150. [Google Scholar]
  21. Vukicevic, M.; Radovanovic, S.; Delibasic, B.; Suknovic, M. Extending meta-learning framework for clustering gene expression data with component-based algorithm design and internal evaluation measures. Int. J. Data Min. Bioinform. 2016, 14, 101–119. [Google Scholar] [CrossRef]
  22. Sáez, J.A.; Corchado, E. A Meta-Learning Recommendation System for Characterizing Unsupervised Problems: On Using Quality Indices to Describe Data Conformations. IEEE Access 2019, 7, 63247–63263. [Google Scholar] [CrossRef]
  23. Pimentel, B.A.; de Carvalho, A.C. A Meta-learning approach for recommending the number of clusters for clustering algorithms. Knowl. Based Syst. 2020, 195, 105682. [Google Scholar] [CrossRef]
  24. Saleem, S.; Gallagher, M. Exploratory Analysis of Clustering Problems Using a Comparison of Particle Swarm Optimization and Differential Evolution. In Proceedings of the Australasian Conference on Artificial Life and Computational Intelligence, Geelong, Australia, 31 January–2 February 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 314–325. [Google Scholar]
  25. Handl, J.; Knowles, J. Cluster Generators for Large High-Dimensional Data Sets with Large Numbers of Clusters. 2005. Available online: http://dbkgroup.org/handl/generators (accessed on 19 July 2019).
  26. Zahn, C.T. Graph-theoretical methods for detecting and describing gestalt clusters. IEEE Trans. Comput. 1971, 100, 68–86. [Google Scholar] [CrossRef] [Green Version]
  27. Kärkkäinen, I.; Fränti, P. Dynamic Local Search Algorithm for the Clustering Problem; Technical Report A-2002-6; Department of Computer Science, University of Joensuu: Joensuu, Finland, 2002; Available online: http://cs.uef.fi/sipu/pub/A-2002-6.pdf (accessed on 10 September 2019).
  28. Veenman, C.J.; Reinders, M.J.T.; Backer, E. A maximum variance cluster algorithm. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 1273–1280. [Google Scholar] [CrossRef] [Green Version]
  29. Salvador, S.; Chan, P. Determining the number of clusters/segments in hierarchical clustering/segmentation algorithms. In Proceedings of the 16th IEEE International Conference on Tools with Artificial Intelligence, Boca Raton, FL, USA, 15–17 November 2004; pp. 576–584. [Google Scholar]
  30. Jain, A.K.; Law, M.H. Data clustering: A user’s dilemma. In Proceedings of the International Conference on Pattern Recognition and Machine Intelligence, Kolkata, India, 20–22 December 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 1–10. [Google Scholar]
  31. Su, M.C.; Chou, C.H.; Hsieh, C.C. Fuzzy C-means algorithm with a point symmetry distance. Int. J. Fuzzy Syst. 2005, 7, 175–181. [Google Scholar]
  32. Ultsch, A. Clustering with SOM: Uˆ* C. In Proceedings of the Workshop on Self-Organizing Maps, Paris, France, 5 September 2005; pp. 59–71. [Google Scholar]
  33. Zelnik-Manor, L.; Perona, P. Self-Tuning Spectral Clustering. 2004. Available online: https://papers.nips.cc/paper/2004/file/40173ea48d9567f1f393b20c855bb40b-Paper.pdf (accessed on 15 July 2020).
  34. Fränti, P.; Virmajoki, O. Iterative shrinking method for clustering problems. Pattern Recognit. 2006, 39, 761–765. [Google Scholar] [CrossRef]
  35. Fränti, P.; Virmajoki, O.; Hautamäki, V. Fast agglomerative clustering using a k-nearest neighbor graph. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1875–1881. [Google Scholar] [CrossRef]
  36. Fu, L.; Medico, E. FLAME, a novel fuzzy clustering method for the analysis of DNA microarray data. BMC Bioinform. 2007, 8, 3. [Google Scholar] [CrossRef] [PubMed]
  37. Gionis, A.; Mannila, H.; Tsaparas, P. Clustering aggregation. ACM Trans. Knowl. Discov. Data (TKDD) 2007, 1, 4. [Google Scholar] [CrossRef] [Green Version]
  38. Kärkkäinen, I.; Fränti, P. Gradual model generator for single-pass clustering. Pattern Recognit. 2007, 40, 784–795. [Google Scholar] [CrossRef]
  39. Chang, H.; Yeung, D.Y. Robust path-based spectral clustering. Pattern Recognit. 2008, 41, 191–203. [Google Scholar] [CrossRef]
  40. Piantoni, J.; Faceli, K.; Sakata, T.C.; Pereira, J.C.; de Souto, M.C. Impact of base partitions on multi-objective and traditional ensemble clustering algorithms. In Proceedings of the International Conference on Neural Information Processing, Taipei, Taiwan, 30 October–2 November 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 696–704. [Google Scholar]
  41. Faceli, K.; Sakata, T. Multiple Solutions in Cluster Analysis: Partitions x Clusters. Available online: https://dcomp.ufscar.br/wp-content/uploads/2016/05/DComp-TR-002.pdf (accessed on 17 July 2019).
  42. Fränti, P.; Mariescu-Istodor, R.; Zhong, C. XNN graph. In Joint IAPR International Workshops on Statistical Techniques in Pattern Recognition (SPR) and Structural and Syntactic Pattern Recognition (SSPR); Springer: Berlin/Heidelberg, Germany, 2016; pp. 207–217. [Google Scholar]
  43. Mardia, K.V. Measures of multivariate skewness and kurtosis with applications. Biometrika 1970, 57, 519–530. [Google Scholar] [CrossRef]
  44. Barrat, A.; Barthelemy, M.; Pastor-Satorras, R.; Vespignani, A. The architecture of complex weighted networks. Proc. Natl. Acad. Sci. USA 2004, 101, 3747–3752. [Google Scholar] [CrossRef] [Green Version]
  45. Fränti, P.; Sieranoja, S. K-means properties on six clustering benchmark datasets. Appl. Intell. 2018, 48, 4743–4759. [Google Scholar] [CrossRef]
  46. Lorena, A.C.; Garcia, L.P.; Lehmann, J.; Souto, M.C.; Ho, T.K. How Complex is your classification problem? A survey on measuring classification complexity. ACM Comput. Surv. (CSUR) 2019, 52, 1–34. [Google Scholar] [CrossRef] [Green Version]
  47. Ho, T.K.; Basu, M. Complexity measures of supervised classification problems. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 289–300. [Google Scholar]
  48. Bonacich, P. Power and centrality: A family of measures. Am. J. Sociol. 1987, 92, 1170–1182. [Google Scholar] [CrossRef]
  49. Kleinberg, J.M. Authoritative sources in a hyperlinked environment. J. ACM 1999, 46, 604–632. [Google Scholar] [CrossRef]
  50. Ball, G.H.; Hall, D.J. ISODATA, a Novel Method of Data Analysis and Pattern Classification; Technical Report; Stanford Research Institute: Menlo Park, CA, USA, 1965; Available online: https://apps.dtic.mil/dtic/tr/fulltext/u2/699616.pdf (accessed on 5 August 2019).
  51. Caliński, T.; Harabasz, J. A dendrite method for cluster analysis. Commun. Stat. Theory Methods 1974, 3, 1–27. [Google Scholar] [CrossRef]
  52. Dunn, J.C. Well-separated clusters and optimal fuzzy partitions. J. Cybern. 1974, 4, 95–104. [Google Scholar] [CrossRef]
  53. Hubert, L.; Schultz, J. Quadratic assignment as a general data analysis strategy. Br. J. Math. Stat. Psychol. 1976, 29, 190–241. [Google Scholar] [CrossRef]
  54. Ratkowsky, D.; Lance, G. Criterion for determining the number of groups in a classification. Aust. Comput. J. 1978, 10, 115–117. [Google Scholar]
  55. Davies, D.L.; Bouldin, D.W. A cluster separation measure. IEEE Trans. Pattern Anal. Mach. Intell. 1979, 2, 224–227. [Google Scholar] [CrossRef]
  56. Halkidi, M.; Batistakis, Y.; Vazirgiannis, M. On clustering validation techniques. J. Intell. Inf. Syst. 2001, 17, 107–145. [Google Scholar] [CrossRef]
  57. Datta, S.; Datta, S. Comparisons and validation of statistical clustering techniques for microarray gene expression data. Bioinformatics 2003, 19, 459–466. [Google Scholar] [CrossRef]
  58. Handl, J.; Knowles, J.; Kell, D.B. Computational cluster validation in post-genomic data analysis. Bioinformatics 2005, 21, 3201–3212. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. Rousseeuw, P.J. Silhouettes: A graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 1987, 20, 53–65. [Google Scholar] [CrossRef] [Green Version]
  60. Lorena, A.C.; Costa, I.G.; Spolaôr, N.; De Souto, M.C. Analysis of complexity indices for classification problems: Cancer gene expression data. Neurocomputing 2012, 75, 33–42. [Google Scholar] [CrossRef]
Figure 1. Rice’s algorithm selection framework.
Figure 1. Rice’s algorithm selection framework.
Algorithms 14 00095 g001
Figure 2. Instance Space Analysis (ISA) framework, building upon Rice’s algorithm selection framework.
Figure 2. Instance Space Analysis (ISA) framework, building upon Rice’s algorithm selection framework.
Algorithms 14 00095 g002
Figure 3. Average performance ranking of the algorithms for ellipsoidal datasets.
Figure 3. Average performance ranking of the algorithms for ellipsoidal datasets.
Algorithms 14 00095 g003
Figure 4. Characteristics of the datasets projected in the instance space, color coded according to (a) distribution of the datasets by groups, (b) number of examples, (c) number of attributes and (d) number of clusters represented by ground truth. The characteristics have been l o g 10 –scaled.
Figure 4. Characteristics of the datasets projected in the instance space, color coded according to (a) distribution of the datasets by groups, (b) number of examples, (c) number of attributes and (d) number of clusters represented by ground truth. The characteristics have been l o g 10 –scaled.
Algorithms 14 00095 g004
Figure 5. Distribution of selected meta–features on the projected instance space.
Figure 5. Distribution of selected meta–features on the projected instance space.
Algorithms 14 00095 g005
Figure 6. Footprints of the algorithms in the instance space.
Figure 6. Footprints of the algorithms in the instance space.
Algorithms 14 00095 g006
Figure 7. Footprints of the algorithm portfolio predicted by SVM.
Figure 7. Footprints of the algorithm portfolio predicted by SVM.
Algorithms 14 00095 g007
Figure 8. Characteristics of the datasets projected in the instance space after the inclusion of real datasets, color coded according to (a) distribution of the datasets by groups, (b) number of examples, (c) number of attributes and (d) number of clusters represented by ground truth. The characteristics have been l o g 10 –scaled.
Figure 8. Characteristics of the datasets projected in the instance space after the inclusion of real datasets, color coded according to (a) distribution of the datasets by groups, (b) number of examples, (c) number of attributes and (d) number of clusters represented by ground truth. The characteristics have been l o g 10 –scaled.
Algorithms 14 00095 g008
Figure 9. Distribution of selected meta–features on the projected instance space after the inclusion of real datasets.
Figure 9. Distribution of selected meta–features on the projected instance space after the inclusion of real datasets.
Algorithms 14 00095 g009
Figure 10. Footprints of the algorithms in the instance space after the inclusion of real datasets.
Figure 10. Footprints of the algorithms in the instance space after the inclusion of real datasets.
Algorithms 14 00095 g010
Figure 11. Footprints of the algorithm portfolio predicted by SVM after the inclusion of real datasets.
Figure 11. Footprints of the algorithm portfolio predicted by SVM after the inclusion of real datasets.
Algorithms 14 00095 g011
Table 1. Meta-features classified by the main properties they measure. In the Asymp. column, n stands for the number of data items a dataset has and m corresponds to its number of input features.
Table 1. Meta-features classified by the main properties they measure. In the Asymp. column, n stands for the number of data items a dataset has and m corresponds to its number of input features.
Meta-FeatureDescriptionAsymp.Reference
(1) Distribution-based:
multi_normMultivariate normality O ( m · n + n 2 ) [17]
skewnessMultivariate normality skewness O ( m · n + n 2 ) [43]
kurtosisMultivariate normality kurtosis O ( m · n + n 2 ) [43]
(2) Neighborhood-based:
avg_nndAverage nearest neighbor degree O ( m · n 2 ) [44]
contrastContrast O ( n 2 ) [45]
high_distPercentage of points of high distance O ( n 2 ) [10]
low_distPercentage of points of low distance O ( n 2 ) [10]
(3) Density:
clust_coefClustering coefficient O ( m · n 2 ) [46]
net_densNetwork density O ( m · n 2 ) [46]
perc_outPercentage of outliers O ( n 2 ) [17]
(4) Dimensionality:
number_ex log 10 number of examples O ( n ) [17]
number_ftr log 10 number of attributes O ( m ) [18]
ratio_ex_ftrRatio number of examples to attributes O ( m + n ) [47]
avg_abs_corAverage absolute correlation O ( n ) [20]
intr_dimIntrinsic dimensionality O ( n 2 ) [45]
avg_pcaAverage number of points per PCA dimension O ( m 2 · n + m 3 ) [46]
ratio_pcaRatio PCA to the original dimension O ( m 2 · n + m 3 ) [46]
(5) Network centrality:
power_centBonacich power centrality O ( n 3 ) [48]
eigen_centEigenvalue centrality of minimum spanning tree O ( n 2 ) [48]
hub_scoreKleinberg’s hub centrality O ( n 3 ) [49]
Table 2. Performance measures classified by properties.
Table 2. Performance measures classified by properties.
AcronymPerformance MeasureReference
(1) Compactness or Homogeneity:
BHBall-Hall[50]
(2) Connectedness:
CConnectivity[58]
(3) Separation:
RLRatkowsky–Lance[54]
(4) Combinations:
CHCalinski–Harabasz[51]
DBDavies–Bouldin[55]
DDunn[52]
SDSD[56]
(5) Stability:
APNAverage proportion of non-overlap[57]
ADAverage distance[57]
ADMAverage distance between means[57]
FOMFigure of merit[57]
(6) Compliance between a partitioning and distance information:
HUBPearson correlation version of Hubert’s statistic[53]
Table 3. Mean, standard deviation and rank of each clustering algorithm.
Table 3. Mean, standard deviation and rank of each clustering algorithm.
AlgorithmMeanStdRank
CC0.6380.1811
AL0.6290.1742
SL0.6130.1573
FCM0.5590.2204
KME0.5510.1775
CL0.5390.1866
WM0.5220.2097
KMD0.5210.2038
MBK0.4760.1829
KMO0.4030.19310
Table 4. Results of SVM prediction models, where Avg. Perf. is the Average Performance, SD Perf. is the Standard Deviation Performance, CV Accuracy, CV Precision and CV Recall are cross-validated performance metrics, C is the cost in the regularization term and γ is the RBF hyper-parameter.
Table 4. Results of SVM prediction models, where Avg. Perf. is the Average Performance, SD Perf. is the Standard Deviation Performance, CV Accuracy, CV Precision and CV Recall are cross-validated performance metrics, C is the cost in the regularization term and γ is the RBF hyper-parameter.
AlgorithmAvg. Perf.SD Perf.Prob. GoodCV AccuracyCV PrecisionCV RecallC γ
KME0.550.180.6274.7075.9086.30206.991.34
KMD0.520.200.5978.7078.9087.603.820.28
KMO0.400.190.3571.8072.6033.30554.324.53
MBK0.480.180.4373.2085.4046.1079.394.56
FCM0.560.220.6790.0087.7098.8097.260.45
SL0.610.160.8288.4088.0099.401.120.34
CL0.540.190.6375.0077.3085.405.230.83
AL0.630.170.7681.1082.6095.2044.550.99
WM0.520.210.5580.3080.1085.100.440.38
CC0.640.180.8591.1090.7099.700.840.92
Oracle0.810.091.00-----
Selector0.670.140.91-91.3047.70--
Table 5. Footprint analysis of the algorithms, where Δ is the area, ρ is the density and ϕ is the purity.
Table 5. Footprint analysis of the algorithms, where Δ is the area, ρ is the density and ϕ is the purity.
Algorithm Δ ρ ϕ
KME0.710.971.00
KMD0.700.890.89
KMO0.190.801.00
MBK0.350.661.00
FCM0.780.960.98
SL0.821.140.98
CL0.541.261.00
AL0.651.331.00
WM0.271.980.99
CC0.871.081.00
Table 6. Results of SVM prediction models after inclusion of real datasets, where Avg. Perf. is the Average Performance, SD Perf. is the Standard Deviation Performance, CV Accuracy, CV Precision and CV Recall are cross-validated performance metrics, C is the cost in the regularization term and γ is the RBF hyper-parameter.
Table 6. Results of SVM prediction models after inclusion of real datasets, where Avg. Perf. is the Average Performance, SD Perf. is the Standard Deviation Performance, CV Accuracy, CV Precision and CV Recall are cross-validated performance metrics, C is the cost in the regularization term and γ is the RBF hyper-parameter.
AlgorithmAvg. Perf.SD Perf.Prob. GoodCV AccuracyCV PrecisionCV RecallC γ
KME0.590.180.6980.8085.0087.7033.380.52
KMD0.510.190.5569.1075.9063.604.892.13
KMO0.440.180.4058.1036.508.000.013.00
MBK0.450.180.3971.0091.9028.802.780.93
FCM0.580.210.6977.5082.4085.7019.672.36
SL0.630.150.8584.6085.6098.6059.201.88
CL0.550.180.6777.0084.8080.207.010.69
AL0.610.170.7699.8099.80100.000.380.00
WM0.580.180.6471.6069.8098.40998.900.48
CC0.640.160.8689.8089.8099.402.160.92
Oracle0.810.091.00-----
Selector0.660.130.95-95.3048.80--
Table 7. Footprint analysis of the algorithms after the inclusion of real datasets, where Δ is the area, ρ is the density and ϕ is the purity.
Table 7. Footprint analysis of the algorithms after the inclusion of real datasets, where Δ is the area, ρ is the density and ϕ is the purity.
Algorithm Δ ρ ϕ
KME0.651.130.92
KMD0.500.881.00
KMO0.021.831.00
MBK0.091.181.00
FCM0.611.151.00
SL0.961.031.00
CL0.581.030.95
AL1.010.960.79
WM0.831.070.95
CC0.941.011.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Fernandes, L.H.d.S.; Lorena, A.C.; Smith-Miles, K. Towards Understanding Clustering Problems and Algorithms: An Instance Space Analysis. Algorithms 2021, 14, 95. https://doi.org/10.3390/a14030095

AMA Style

Fernandes LHdS, Lorena AC, Smith-Miles K. Towards Understanding Clustering Problems and Algorithms: An Instance Space Analysis. Algorithms. 2021; 14(3):95. https://doi.org/10.3390/a14030095

Chicago/Turabian Style

Fernandes, Luiz Henrique dos Santos, Ana Carolina Lorena, and Kate Smith-Miles. 2021. "Towards Understanding Clustering Problems and Algorithms: An Instance Space Analysis" Algorithms 14, no. 3: 95. https://doi.org/10.3390/a14030095

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop