Next Article in Journal
Differential Quadrature Method for Bending Analysis of Asymmetric Circular Organic Solar Cells Resting on Kerr Foundation in Hygrothermal Environment
Previous Article in Journal
Joint Event Density and Curvature Within Spatio-Temporal Neighborhoods-Based Event Camera Noise Reduction and Pose Estimation Method for Underground Coal Mine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hubness Information-Based k-Nearest Neighbor Approach for Multi-Label Learning

1
College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
2
College of Computer Science and Engineering, Northeastern University, Shenyang 110169, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(7), 1202; https://doi.org/10.3390/math13071202
Submission received: 17 February 2025 / Revised: 29 March 2025 / Accepted: 1 April 2025 / Published: 5 April 2025

Abstract

:
Multi-label classification (MLC) plays a crucial role in various real-world scenarios. Prediction with nearest neighbors has achieved competitive performance in MLC. Hubness, a phenomenon in which a few points appear in the k-nearest neighbor (kNN) lists of many points in high-dimensional spaces, may significantly impact machine learning applications and has recently attracted extensive attention. However, it has not been adequately addressed in developing MLC algorithms. To address this issue, we propose a hubness-aware kNN-based MLC algorithm in this paper, named multi-label hubness information-based k-nearest neighbor (MLHiKNN). Specifically, we introduce a fuzzy measure of label relevance and employ a weighted kNN scheme. The hubness information is used to compute each training example’s membership in relevance and irrelevance to each label and calculate weights for the nearest neighbors of a query point. Then, MLHiKNN exploits high-order label correlations by training a logistic regression model for each label using the kNN voting results with respect to all possible labels. Experimental results on 28 benchmark datasets demonstrate that MLHiKNN is competitive among the compared methods, including nine well-established MLC algorithms and three commonly used hubness reduction techniques, in dealing with MLC problems.

1. Introduction

Multi-label learning solves one of the main categories of machine learning classification tasks, where each instance could belong to multiple labels simultaneously. For example, a news document could be labeled by multiple topics simultaneously, such as environment, health, and technology. An image scene could contain various concepts, such as buildings, cars, and pavement. Multi-label classification (MLC) problems widely exist in both daily life and specialized fields, such as social media tagging [1], product categorization [2], remote sensing image classification [3], and legal text mining [4]. This leads to a significant demand for efficient and accurate MLC algorithms.
A key difference in designing MLC algorithms compared to single-label classification (SLC) algorithms is that MLC algorithms can leverage label correlations, which are derived from the co-occurrence of labels in multi-label datasets (MLDs), to improve their performance. Many MLC algorithms across various domains have been proposed. Existing methods can be categorized into first-order, second-order, and high-order approaches, which consider no correlation, pairwise correlations, and correlations among multiple labels. Higher-order approaches can explore more complex relationships, which is helpful in improving the classifier performance. The k-nearest neighbor (kNN) search technique is employed in some popular MLC algorithms, such as first-order methods, BRkNN [5] and MLKNN [6], and high-order methods, DMLkNN [7] and IBLR [8]. These kNN-based approaches achieve competitive performance among existing MLC algorithms [9,10] and constitute one of the major branches in multi-label learning.
Recently, an issue related to the kNN search, hubness in high-dimensional data, has drawn extensive attention in various machine learning areas. It refers to the phenomenon where a small number of points appear in the k-neighborhoods of a large number of examples. Specifically, let the k-occurrence of point x, O k ( x ) , denote the number of times x appears among the k nearest neighbors of all other points in a dataset. The average k-occurrence of all points is k. Some popular examples may have k-occurrences several times larger than k, and some may be included in the k-neighborhoods of very few or no other points. These two types of points are called hubs and antihubs, respectively. Hubness is an inherent property of data distributions in high-dimensional space under widely used assumptions and is not caused by other effects such as finite sample size [11]. Existing studies have shown that the hubness problem can have a significant impact on the performance of machine learning algorithms involving kNN search in high dimensions [11,12]. Take the classification task, for example. Let the term bad k-occurrence represent that a point lies in the k-neighborhood of another point, but the two points’ labels do not match. Empirical results indicate that hubs are more prone to picking up bad k-occurrence in high dimensions than other points. Similarly, hubs are more likely to appear in the kNN lists of query points and belong to different classes from those of the query points. Existing studies show that dimensionality reduction may not have a significant effect on reducing hubness until intrinsic dimensionality is reached [11]. However, there may be a loss of information if the dimensionality is further reduced. Some techniques have been proposed to mitigate the negative effects of hubness in machine learning and have achieved significant improvements. On the other hand, existing hubness-aware approaches for classification tasks mainly focus on single-label cases.
To illustrate the hubness phenomenon in MLDs, we show the distributions of O k ( x ) and the k-occurrence skewness S k , which is typically used to measure the degree of hubness, of some benchmark MLDs with different dimensionalities in Figure 1. S k is the standardized third central moment of the k-occurrence distribution.
S k = E ( O k μ O k ) 3 σ O k 3 ,
where μ O k and σ O k are the mean and standard deviation of O k , respectively. S k characterizes the asymmetry of O k . It signifies skewness to the right (left) for S k with a positive (negative) value and no skewness for S k = 0 . Higher skewness to the right indicates a higher degree of hubness. As shown in Figure 1, in the six datasets with relatively higher dimensions, the highest probability density occurs at O k = 0 , corresponding to antihubs. Meanwhile, a few points have values of O k several tens of times greater than k, corresponding to hubs. In Figure 1, five multi-label datasets have the values of S k varying from 9.92 to 23.24. Note that among the 50 different machine learning datasets employed in an empirical comparison study of hubness reduction methods [12], 47 datasets have values of S k varying from −0.12 to 5.58. The other three datasets have S k values of 9.35, 10.71, and 15.52, respectively. By comparison, these MLDs exhibit high degrees of hubness. For that, in this paper, we propose a hubness-aware kNN-based MLC algorithm named MLHiKNN, i.e., multi-label hubness information-based k-nearest neighbor. To address the hubness issue, MLHiKNN introduces a fuzzy measure of label relevance for training examples and uses hubness weighting for different neighbors. Specifically, the degree of a training example’s support for a label is determined by its relevance to the label, its history k-occurrence, and the labels of its reverse nearest neighbors. Further, MLHiKNN trains a logistic regression model with the kNN voting results with respect to all possible labels. Thus, it is a high-order method with respect to label correlation. As far as we know, it is the first approach to address the hubness issue in MLC. We conducted experiments on twenty-eight benchmark MLDs from different domains to compare MLHiKNN with nine well-established MLC algorithms and three commonly used hubness reduction techniques. Experimental results show that MLHiKNN achieves the best performance in a considerable number of scenarios, indicating that it is highly competitive among the compared methods. The remainder of this paper is organized as follows: Section 2 reviews some MLC algorithms and hubness-aware techniques. Section 3 introduces the framework and details of the proposed method. Section 4 reports and discusses the results of experiments for method comparison, ablation analysis, and parameter analysis. Section 5 concludes this work.

2. Related Work

This section reviews the existing MLC algorithms, followed by the works addressing the hubness phenomenon in high-dimensional spaces.

2.1. Multi-Label Classification Algorithms

Let a multi-label dataset be denoted by D = { ( x i , y x i ) | 1 i n } , where the feature vector x i = ( x i 1 , x i 2 , , x i d ) T , and the label indicator vector y x i = ( y x i 1 , y x i 2 , , y x i q ) T . n is the number of examples in D, q is the number of possible labels, and d is the number of features. y x i j = 1 if example x i belongs to the j-th label and y x i j = 0 otherwise. Multi-label learning is learning to map from the feature space to the label space with a dataset like D, and then predict the label indicator vector for each query example with its features. Existing MLC algorithms can be divided into problem transformation methods and algorithm adaptation methods based on the strategy for making an MLC problem solvable with SLC algorithms. Further, many efforts have been made to exploit correlations among labels to improve the generalization ability of an MLC algorithm. Existing MLC algorithms can also be roughly categorized into first-order, second-order, and high-order approaches according to the order of correlations being considered [13].
Problem transformation methods fit data to the algorithm by transforming a multi-label dataset into multiple single-label datasets (SLDs). Then, a sub-problem is solved with SLC algorithms for each single-label dataset. BR [14] transforms the original MLC problem into q binary classification sub-problems, where each sub-problem corresponds to one possible label and takes the original features as inputs. BR belongs to first-order approaches, which do not consider label correlations. Second-order approaches consider the correlations between each pair of labels. These approaches usually generate a sub-problem for each pair of possible labels, such as CLR [15] and RPC [16], which generate q ( q 1 ) / 2 sub-problems in total. High-order approaches consider the correlations among multiple labels. For example, CC [17] constructs a chain of binary classifiers, where each classifier corresponds to one label. Different from BR, each binary classifier in CC considers the labels corresponding to the previous binary classifiers as additional features for prediction. However, a drawback of CC is that the pattern of correlations depends on the order of binary classifiers, which is predetermined. ECC [17], the ensemble version of CC, deals with this problem by building a set of CC models, where the order of binary classifiers is randomly generated for each model. Another representative high-order approach, LP [14], transforms the original MLC problem into a multi-class classification problem, where each class corresponds to a distinct label set that appeared in the dataset. LP has two major limitations. One is that LP cannot predict unseen label sets based on the training set. The other is that LP is inefficient since the number of classes in the transformed problem may be very large for a dataset with a large q. RAkEL [18], the ensemble version of LP, tackles the two drawbacks by employing many LP models, where each model only deals with several labels. These LP models overlap in labels and cover all possible labels together. On the other hand, RAkELd [18] splits all the possible labels to these LP models without overlap.
Algorithm adaptation methods fit the algorithm to the data by modifying the components of an SLC algorithm to deal with MLDs directly. For example, ML-DT [19] modifies the C4.5 approach by allowing multiple labels in the leaves to deal with multi-label data. Rank-SVM [20] adapts the maximum-margin strategy to learn from multi-label data. ML-RBF [21] adapts the Radial Basis Function for multi-label learning. kNN-based methods also constitute a major branch of this category of approaches. BRkNN [5] is conceptually equivalent to using BR as the problem transformation method and the KNN algorithm as the base classifier. In practice, BRkNN only needs to search for the k nearest neighbors (kNNs) of each example for a single time. BRkNNa [5] extends BRkNN by avoiding the empty output when predicting a label set. Unlike BRkNN, which compares the positive and negative votes from the kNNs to make predictions, MLKNN [6] follows a maximum a posteriori (MAP) principle based on the counts of positive neighbors to predict the relevance to each label. Thus, the information of training examples far from the query sample also contributes to the prediction in MLKNN. Meanwhile, MLKNN is less affected by class imbalance than BRkNN. ML-k’sNN [22] extends MLKNN by searching and applying the best parameter k, which is the number of nearest neighbors to take into account, for the prediction of each label. The above kNN-based MLC algorithms are all first-order approaches. The high-order approach DMLkNN [7] extends MLKNN by taking into consideration the occurrences of all other labels in the MAP model besides the label to be predicted. IBLR [8] exploits high-order correlations by augmenting the original features with the labels of the kNNs of an instance when training a logistic regression model for each label.
Recently, many works have attempted to address some more specific challenges in multi-label learning [23], such as class imbalance [24], the different types of supervision like MLC with missing labels [25,26] and partial multi-label learning [27,28], the extremely large number of labels [29,30], and online multi-label learning [31,32]. Nevertheless, as far as we know, the hubness problem has not been addressed in MLC. The following subsection will review some studies about the hubness phenomenon and existing techniques focusing on learning with hubs in SLC.

2.2. Methods Addressing the Hubness Problem

Hubness is one aspect of the curse of dimensionality when performing a search for nearest neighbors. The hubness problem was first noted in audio-based music recommendation [33]. Specifically, a small number of hub songs are recommended conspicuously frequently in nearest neighbor-based playlists. Radovanović et al. studied the causes and implications of hubness extensively from both theoretical and empirical perspectives [11]. The hubness phenomenon is related to data concentration, another aspect of the dimensionality curse. The phenomenon of data concentration is usually expressed as the ratio between some measure of spread and magnitude, such as the ratio between the standard deviation and the mean of the distances to an arbitrary reference point of all points. If this ratio converges to zero as the dimensionality approaches infinity, it is said that the distances concentrate, indicating that all points are almost at the same distance from the reference point. Theoretical analysis in existing works indicates that points in high dimensions are approximately lying on a hypersphere centered at the data mean. Radovanović et al. proposed the argument that the points closer to the data mean are more likely to become hubs, and points further away from the data mean are more likely to be antihubs in high dimensions [11]. The nearest-neighbor search is an essential component in many commonly used algorithms for a wide range of machine learning applications, such as classification, regression, clustering, and outlier detection. For these algorithms, the hubness problem can have a significant impact on their performance in high dimensions. There have been some studies addressing this issue across various application domains. We roughly divide existing techniques into distance transformation and algorithm adaptation methods.
Distance transformation methods can be incorporated into the nearest-neighbor search process of various algorithms. These methods reduce hubness by modifying the neighbor graph through alternative dissimilarity measures and distance metrics, re-scaling, centering, and other techniques. Alternative dissimilarity measure techniques directly operate on the position vectors of each point for pairwise dissimilarities. Then, the dissimilarities are used instead of the original distances in the search for nearest neighbors to reduce hubness. These hubness-aware measures typically consider the information of other points as an addition when computing the dissimilarity between each pair of points. DSL [34] flattens the density gradient to reduce spatial centrality in Euclidean spaces. For two examples, x 1 and x 2 , DSL calculates their dissimilarity by removing the sample-wise centrality from their distance,
DSL ( x 1 , x 2 ) = x 1 x 2 2 2 x 1 c k ( x 1 ) 2 2 x 2 c k ( x 2 ) 2 2 .
Let D k ( x ) be a set consisting of the kNNs of x. c k ( x ) = 1 k x D k ( x ) x is a local centroid with respect to x. Parameter k is the number of nearest neighbors considered to compute centroids. m p -dissimilarity [35] is calculated based on probability mass and follows the principle that two points in a sparse region are more similar than two points in a dense region. s i m h u b s IN [36] is based on the observation that a rarely shared neighbor (antihub) is more informative than a commonly shared neighbor (hub) in similarity calculation. Unlike the above three measures, which rely only on feature vectors, the following two measures incorporate class information for dissimilarity computation and are specifically for classification tasks. MADD [37] is based on the mean absolute difference between inter-point pairwise distances. s i m h u b s PUR [36] prefers the points whose reverse neighbors are almost exclusively from the same class compared to those with reverse neighbors from various classes when computing similarities. The incorporation of class information in the two techniques does not generally reduce hubness like other distance transformation methods but can be regarded as trying to preserve similarities with the hubs, which can provide helpful information for classification, and to reduce similarities with the hubs that offer no benefit for classification. Note that the two techniques are not applicable to multi-label scenarios. Alternative distance metric techniques do not utilize the information of other points to compute the distance between each pair of points but use hubness-aware metrics to reduce hubness. Aggarwal et al. showed that fractional norms are less prone to distance concentration than the Euclidean norm [38]. To be specific, despite all p norms exhibiting concentration, the degree of concentration is influenced by both the distribution of data and the value of p. Based on this observation, Flexer et al. define antihub occurrence as the percentage of antihubs in the dataset and hub occurrence as the percentage of hubs in all nearest-neighbor slots. Then, the authors reduce hubness by choosing the p norm, by which the corresponding antihub occurrence or hub occurrence is minimal [39].
Re-scaling techniques compute the pairwise distances using a well-established metric, such as the Euclidean distance, first. Then, they scale the distance between each pair of points based on other pairwise distances, considering a local or global perspective. Local scaling (LS) and mutual proximity (MP) [40,41] are two representative re-scaling techniques. They reduce hubness by repairing asymmetric neighbor relations by computing pairwise secondary distances. For two examples x 1 and x 2 , LS scales their original distance d x 1 , x 2 using the local distance information of the two examples,
LS ( x 1 , x 2 ) = 1 exp ( d x 1 , x 2 2 d k ( x 1 ) d k ( x 2 ) ) ,
where d k ( x ) is the distance between x and its k-th nearest neighbor. Unlike LS, which uses local distance as the scaling parameter, MP considers all pairwise distances in the dataset. MP calculates the proximity between x 1 and x 2 by the number of instances whose distances to x 1 and x 2 are both greater than d x 1 , x 2 ,
MP ( x 1 , x 2 ) = | { j | d x 1 , j > d x 1 , x 2 } { j | d x 2 , j > d x 2 , x 1 } | n 2 ,
where n is the number of examples in the dataset. Centering techniques were also shown to be effective in reducing hubness. CENT [42] shifts the origin space to the global data centroid and uses inner product similarities. LCENT [43] further considers local centroids for the datasets where the sample size is large and the dimension of vectors is not necessarily very high. Distance transformation hubness reduction techniques have been verified to be effective in improving algorithm performance across datasets from different domains and have been employed in various applications, such as knowledge graph embedding [44], single-cell omics profiling [45], and Sentence-BERT embedding [46].
Algorithm adaptation methods modify existing algorithms involving nearest-neighbor search to deal with the hubness problem. The kNN classifier has naturally become one of the most noted algorithms, with hubness-aware modifications being implemented in existing studies. hw-kNN [47] reduces the voting weights of bad hubs, which refer to the hub examples whose labels do not match those of their reverse nearest neighbors. Concretely, the weight assigned to a point x is based on the count of bad occurrences for x, where a bad occurrence is defined as a reverse nearest neighbor belonging to a different class from that of x. NHBNN [48] performs a Naive-Bayesian inference for each query point according to the kNNs. For each neighbor of a query point, a probability estimation is conducted based on the labels of its historical reverse nearest neighbors in the training set. h-FNN [49] modifies the class preference provided to query points of each training example by introducing fuzzy measures. The fuzzy membership of a training point x in each class is related to the percentage of samples in this class among all reverse nearest neighbors of x. Meanwhile, for points rarely appearing in the kNN lists of other points, h-FNN offers four alternative approaches to estimate their fuzziness from a global or local perspective. Distance weighting can also be incorporated into the h-FNN framework. HIKNN [50] utilizes the historical occurrence information to adjust both the weights and class preferences of training examples in kNN voting. Specifically, antihubs are more informative than hubs. The class preference of a training point depends on both its label and the labels of its reverse nearest neighbors.
There have also been some studies that focus on dealing with hubness in other machine learning tasks, such as regression, clustering, and outlier detection. For the case of regression, EWCkNN [51], on the one hand, performs error-based weighted k-nearest-neighbor regression, where the weight of a point x is inversely related to the normalized average absolute difference between the label of x and the labels of x’s k reverse nearest neighbors. On the other hand, it uses the corrected labels for prediction, where the corrected label of a training point x is defined as the average of the labels of x’s k reverse nearest neighbors. For the case of clustering, the K-hubs [52] approach uses hubs in the same way that centroids are typically used and converges to final clusters very quickly. GHPC [52] implements a squared hubness–proportional stochastic scheme and uses the simulated annealing approach for optimization. For the case of outlier detection, the AntiHub 2 [53] approach computes an outlier score for each point x by considering the occurrence numbers of both x and x’s neighbors in the kNN lists of other points. PEHS [54] constructs a digraph based on the neighborhood relations of points and performs a tailored Markov random walk process on the digraph to derive a hubness score for each point. Then, it brings together the hubness score and the potential energy of a point, which is derived from the concept of potential of particles in physics, for an importance degree. Points with small degrees of importance are considered outliers.
Besides the works mentioned above, the hubness issue has also been addressed in some specific applications. For example, zero-shot learning involves nearest-neighbor searches in a learned high-dimensional embedding space as the last step and thus suffers from the hubness issue. It is also shown that ridge regression tends to promote hubness in regression-based zero-shot learning [55]. Different techniques have been pursued to deal with the hubness issue in zero-shot learning, such as considering the global distribution of test data in a transductive way [56], using the visual feature space of a CNN subnet as the embedding space [57], modifying the objective function to maximize the scatter between different classes besides the attribute regression accuracy [58], and learning a latent space preserving the semantic relationship and discriminating information [59]. Besides the works in zero-shot learning, HLM [60] extends the Bipartite Local Model (BLM), a leading method for drug–target interaction prediction, by incorporating the ECkNN approach, a hubness-aware regression technique that corrects the label of a training point by the labels of its reverse nearest neighbors. Wang et al. proposed a hybrid similarity measurement that combines rating-based and structural similarity measurements to deal with the hubness problem in memory-based collaborative filtering, a recommendation system method for user’s rating or preference prediction [61]. Tian et al. proposed a novel retrieval measuring metric that includes the hubness degree as a penalty factor to alleviate the hubness issue in similarity-based translation memory retrieval for ontology translation [62]. The proposed method, MLHiKNN, can be categorized into the algorithm adaptation category for dealing with the hubness issue. The details will be introduced in the following section.

3. The Proposed Method: MLHiKNN

The major components in the model of MLHiKNN include the label relevance score (Section 3.1) used to leverage the hubness information of nearest neighbors and a label predictor (Section 3.4), where the label relevance scores are considered as the features of a query instance. The elements used for computing label relevance scores are introduced in Section 3.2 and Section 3.3. The complete steps of MLHiKNN are introduced in Section 3.4.

3.1. Label Relevance Score for Query Points

For each query instance t whose label set is to be estimated, we first introduce its positive and negative voting scores with respect to label l, v 1 l ( t ) and v 0 l ( t ) . Let D k ( t ) be the set consisting of the kNNs of t. v 1 l ( t ) and v 0 l ( t ) are the results of weighted voting by considering the label and hubness information of all instances in D k ( t ) as follows:
v b l ( t ) = x D k ( t ) w d ( t , x ) · w h ( t , x ) · u b l ( x ) , b { 0 , 1 } ,
where u b l ( x ) is the fuzzy membership of x with respect to label l (Section 3.2). w d ( t , x ) and w h ( t , x ) (Section 3.3) are two factors to adjust the voting weights among the kNNs of t. Then, the label relevance score of t with respect to label l is obtained as follows:
r l ( t ) = v 1 l ( t ) v 0 l ( t ) + v 1 l ( t ) .
Instance t is more likely to be relevant to label l if it has a larger r l ( t ) . The computation of the label relevance score involves hubness information in both the fuzzy membership u b l ( x ) and the weight factor w h . The details are introduced in the following subsections.

3.2. Fuzzy Measure of Label Relevance for Training Points

For a point x, its degree of hubness is measured by its k-occurrence O k ( x ) , which is the number of times x appears among the kNNs of all other points in the dataset. This is also known as the size of the k reverse neighborhood of x. The label hubness of x with respect to label l is measured by its label k-occurrence, O k l ( x ) , which refers to the number of instances relevant to label l in the k reverse neighborhood of x. In this work, we append each instance to its own k-neighborhood (resulting in k + 1 points in the k-neighborhood) when computing k-occurrence and label k-occurrence in case k-occurrence is zero for some points, leading to tedious formulas in the proposed approach. The procedures of counting k-occurrence and label k-occurrence are shown in Algorithm 1.
Algorithm 1 k-occurrence counting.
Input:  
  D: training set;
  k: number of nearest neighbors to take into account.
Output:   
   O k ( x ) , O k l ( x ) : the k-occurrence and label k-occurrence with respect to each possible label l of each instance x in D.
  1:
q number of possible labels in D
  2:
find the k-neighborhood D k ( x ) for each instance x in D
  3:
for each instance x in D do
  4:
      O k ( x ) 0
  5:
     for  l 1   to   q  do
  6:
          O k l ( x ) 0
  7:
     end for
  8:
     for each instance x  in D do
  9:
         if  x D k ( x )  then
10:
             O k ( x ) O k ( x ) + 1
11:
            for  l 1   to   q  do
12:
                if  y x l = 1  then
13:
                     O k l ( x ) O k l ( x ) + 1
14:
                end if
15:
            end for
16:
         end if
17:
     end for
18:
end for
19:
return { O k ( x ) , O k l ( x ) | x D , 1 l q }
Given the condition that x is in the k-neighborhood of an unlabeled instance t, the probability that t is relevant to label l can be obtained based on O k ( x ) and O k l ( x ) as follows:
P ( y t l = 1 | x D k ( t ) ) = O k l ( x ) + s O k ( x ) + s × 2 ,
where s is a smoothing parameter. In this work, s is set to be 1 (Laplace smoothing). Let p 1 l ( x ) denote the above conditional probability and p 0 l ( x ) denote such a conditional probability that t is irrelevant to label l.
p 0 l ( x ) = P ( y t l = 0 | x D k ( t ) ) = 1 p 1 l ( x ) .
From the perspective of probability estimation, t is more likely to be relevant to label l if x has a greater p 1 l ( x ) . Estimation with p 1 l ( x ) is appropriate for x with a large O k ( x ) according to the law of large numbers. For x with a small O k ( x ) , an appropriate alternative, following the principle in the KNN classifier, is to assume that t is more likely to have label l if x has label l. The conclusions drawn from the two clues might be inconsistent for hubs that are more prone to picking up bad k-occurrence in high dimensions. Accordingly, we introduce the fuzzy measure of label relevance to consider both the label and hubness information of an instance in the training set. Let u 1 l ( x ) and u 0 l ( x ) denote the degrees of membership of instance x in terms of relevance and irrelevance to label l, respectively.
u 1 l ( x ) = g ( x ) · y x l + ( 1 g ( x ) ) · p 1 l ( x ) ,
u 0 l ( x ) = g ( x ) · ( 1 y x l ) + ( 1 g ( x ) ) · p 0 l ( x ) ,
where g ( x ) is a factor that weighs the label and hubness information of a labeled instance x.
For an appropriate function g ( x ) , the following conditions for the generation of fuzzy sets need to be satisfied first. For each label l,
b { 0 , 1 } u b l ( x ) = 1 ,
0 < x D u b l ( x ) < | D | ,
u b l ( x ) [ 0 , 1 ] .
Equation (11) always holds according to the function of u b l ( x ) in Equations (9) and (10). Equation (12) is to avoid the situation where all points certainly belong to the same fuzzy set. Equation (13) is to ensure the reasonable range of membership.
Second, we set g ( x ) to be in [ 0 , 1 ] as a weight factor. It reflects the intuition that the predictive importance of x’s own labels should decrease as x appears more frequently in the k-neighborhoods of other training instances. When a training instance x is rarely selected as a neighbor by other training instances (i.e., O k ( x ) is small), its own labels are essential for making predictions for a query point. In the extreme case where x does not appear in the k-neighborhood of any other training point, only its own labels are available for label prediction, and hence, we define g ( x ) = 1 .
g ( x ) = 1 , if O k ( x ) = 1 .
As O k ( x ) increases, it becomes possible to estimate the query label based on the labels of other instances for which x is in the k-neighborhoods. Thus, the weight assigned to x’s own label set should gradually decrease. The hubness information contributes more to the fuzzy membership u b l ( x ) . We define a threshold λ k , where λ is a preset parameter, such that when O k ( x ) λ k , the contribution of x’s own label set is reduced to zero, i.e., g ( x ) = 0 . At this point, we consider that the hubness information provides significantly more informative value than the labels of x. The threshold is proportional to k since a larger k usually leads to a larger O k for each instance. The parameter λ determines how the threshold changes with respect to k. The use of λ k is also motivated by the observation in previous research that an example is typically identified as a hub if its k-occurrence is larger than 2 k . In such cases, a point’s frequent appearance in k-neighborhoods is often caused by the hubness phenomenon rather than label similarity, making its label less reliable for prediction. Empirically, we found that setting λ = 2 yields good average performance, in line with the threshold commonly adopted in the literature for hub definition. A detailed parameter analysis is provided in Section 4.5.
g ( x ) = 0 , if O k ( x ) λ k .
For x with O k ( x ) between 1 and λ k , g ( x ) is in ( 0 , 1 ) .
0 < g ( x ) < 1 , if 1 < O k ( x ) < λ k .
Third, for x with O k ( x ) in the range of ( 1 , λ k ) , the derivative of g ( x ) with respect to O k ( x ) is expected to be less than 0,
d g ( x ) d O k ( x ) < 0 , if 1 < O k ( x ) < λ k ,
so that as O k ( x ) increases, the weight of p b l ( x ) , b { 0 , 1 } , increases, and the weight of y x l decreases. Meanwhile, the second-order derivative of g ( x ) with respect to O k ( x ) is expected to be larger than 0 for x with O k ( x ) in ( 1 , λ k ) ,
d 2 g ( x ) d O k ( x ) 2 > 0 , if 1 < O k ( x ) < λ k .
For example, assume four points x 1 , x 2 , x 3 , and x 4 with k-occurrences of 1, 10, 51, and 60. The difference in k-occurrence between x 1 and x 2 is equal to that between x 3 and x 4 , O k ( x 4 ) O k ( x 3 ) = O k ( x 2 ) O k ( x 1 ) . On the other hand, the difference in importance when calculating the membership of label relevance between p b l ( x 1 ) and p b l ( x 2 ) is expected to be greater than that between p b l ( x 3 ) and p b l ( x 4 ) , due to the rate of change in k-occurrence being larger from x 1 to x 2 than from x 3 to x 4 . This results in a corresponding difference in their weights. Specifically, ( 1 g ( x 2 ) ) ( 1 g ( x 1 ) ) > ( 1 g ( x 4 ) ) ( 1 g ( x 3 ) ) , which simplifies to g ( x 4 ) g ( x 3 ) > g ( x 2 ) g ( x 1 ) . Based on the above considerations in Equations (11)–(18), we define the function of g ( x ) as follows:
g ( x ) = log γ λ k log γ min ( O k ( x ) , λ k ) log γ λ k = 1 log λ k min ( O k ( x ) , λ k ) .

3.3. Distance and Hubness Weighting

This subsection introduces the weighting factors w d and w h for calculating label relevance scores. As in Section 3.1, the label relevance score is computed based on kNN voting. For a query instance t, the importance of different points in its k-neighborhood may not be the same. One common consideration is that a neighbor is more important if it is closer to the query point. Following this principle, we introduce w d for distance weighting by employing an existing scheme as follows:
w d ( t , x ) = t x 2 / ( β 1 ) x D k ( t ) t x 2 / ( β 1 ) ,
where β is a preset parameter to determine how heavily the distance is weighted.
For a point x in the training set, besides its distance to the query point, its k-occurrence also influences its relative importance among the k nearest neighbors. For example, assume that O k ( x ) = | D | . That is, x is in the k-neighborhood of every sample in the training set. Thus, neighboring x is useless neighboring information for a query point t. For that, we introduce w h for hubness weighting. As one of the k nearest neighbors of query point t, the weight of x is denoted by w h ( t , x ) . Following the above example,
w h ( t , x ) = 0 , if O k ( x ) = | D | .
Similarly, x is more important as a neighbor of the query point if it appears less frequently in the k-neighborhoods of other points in the training set. Thus, for query point t,
d w h ( t , x ) d O k ( x ) < 0 , if 1 < O k ( x ) < | D | .
Based on the similar consideration for g ( x ) in Section 3.2, the second-order derivative of w h ( t , x ) with respect to O k ( x ) is supposed to be larger than 0 for x with O k ( x ) in ( 1 , | D | ) ,
d 2 w h ( t , x ) d O k ( x ) 2 > 0 , if 1 < O k ( x ) < | D | .
To balance the heaviness of distance weighting and hubness weighting, for each query point t, the sum of w h is expected to be equal to that of w d .
x D k ( t ) w h ( t , x ) = x D k ( t ) w d ( t , x ) .
Based on the above considerations in Equations (21)–(24), we define the function of w h ( t , x ) as follows:
w h ( t , x ) = log γ | D | log γ O k ( x ) x D k ( t ) log γ | D | log γ O k ( x ) = log | D | k x D k ( t ) O k ( x ) | D | O k ( x ) .

3.4. Learning with Label Relevance Scores

In the previous subsections, we have introduced the mathematical procedures of label relevance score with respect to label l, r l ( t ) . Note that 0 r l ( t ) < 0.5 if the positive voting score with respect to label l, v 1 l ( t ) , is lower than the negative voting score with respect to label l, v 0 l ( t ) . And 0.5 < r l ( t ) 1 if v 1 l ( t ) is greater than v 0 l ( t ) . For a query instance t, an intuitive way to predict its relevance to label l is to assign l to t if and only if r l ( t ) 0.5 . This scheme is straightforward but ignores label correlations. To explore and utilize the correlations among labels, for each label, we train a logistic regression model using r l ( t ) , l { 1 , 2 , , q } , as inputs and the label indicator as output, where q is the number of possible labels in the training set. Thus, the label relevance scores with respect to all other labels are also considered while predicting the relevance to label l, resulting in high-order label correlation exploration.
The training and prediction processes of MLHiKNN are shown in Algorithms 2 and 3, respectively. There are two phases in the training process. The first phase (Algorithm 2, lines 4–10) computes membership for each example in the training set. The second phase (Algorithm 2, lines 11–23) trains a logistic regression model with label relevance scores. In this phase, for each instance x in the training set, the label relevance score with respect to each possible label is first computed with its neighborhood D k ( x ) . Note that for x, its membership u b l ( x ) does not participate in the computation of its label relevance score r l ( x ) as in Algorithm 2, line 17. In the prediction process, the label relevance scores of t with respect to all possible labels r l ( t ) , 1 l q , are first computed (Algorithm 3, lines 3–10). Then, the prediction is obtained by the logistic regression model with r l ( t ) , 1 l q . Our proposed method has four parameters. s is set to 1 as part of Laplace smoothing. k is the number of nearest neighbors to take into account. λ is a threshold parameter. β is a distance weighting parameter. For k , λ , and β , their recommended values will be introduced in Section 4.5.
Algorithm 2 The training process of MLHiKNN.
Input:  
   D: training set;
   k: number of nearest neighbors to take into account;
    λ : threshold parameter for computing g ( · ) ;
    β : distance weighting parameter for computing w d ( · ) ;
   s: smoothing parameter for computing p 1 l ( · ) .
Output:   
   trained MLHiKNN classifier;
  1:
q number of possible labels in D
  2:
find the k-neighborhood D k ( x ) for each instance x in D
  3:
count k-occurrence O k ( x ) and label k-occurrence O k l ( x ) for each instance x in D by Algorithm 1
  4:
for each instance x in D do
  5:
        g ( x ) ← Equation (19) with O k ( x ) , λ , and k
  6:
       for  l 1   to   q  do
  7:
              p b l ( x ) , b { 0 , 1 } ← Equation (7) and Equation (8) with O k ( x ) , O k l ( x ) , and s
  8:
              u b l ( x ) , b { 0 , 1 } ← Equation (9) and Equation (10) with g ( x ) , p b l ( x ) , and y x l
  9:
       end for
10:
end for
11:
for each instance x in D do
12:
       for each instance x  in  D k ( x )  do
13:
              w d ( x , x ) ← Equation (20) with x , D k ( x ) , and β
14:
              w h ( x , x ) ← Equation (25) with { O k ( x ) | x D k ( x ) } and | D |
15:
       end for
16:
       for  l 1   to   q  do
17:
              v b l ( x ) , b { 0 , 1 } ← Equation (5) with { w d ( x , x ) , w h ( x , x ) , u b l ( x ) | x D k ( x ) }
18:
              r l ( x ) ← Equation (6) with v 1 l ( x ) and v 0 l ( x )
19:
       end for
20:
end for
21:
for  l 1   to   q  do
22:
      l train a binary logistic regression model with { r l ( x ) , y x l | 1 l q , x D } , where for each instance x, { r l ( x ) | 1 l q } is the input, and y x l is the output
23:
end for
24:
return trained MLHiKNN classifier ← { D , u b l , l , β | 1 l q , b { 0 , 1 } }
Algorithm 3 The prediction process of MLHiKNN.
Input:  
  trained MLHiKNN classifier: { D , u b l , l , β } ;
  t: query instance.
Output:   
   y ^ t : label set prediction of instance t.
  1:
q number of possible labels in D
  2:
find the k-neighborhood D k ( t ) for query instance t
  3:
for each instance x in D k ( t )  do
  4:
       w d ( t , x ) ← Equation (20) with t , D k ( t ) , and β
  5:
       w h ( t , x ) ← Equation (25) with { O k ( x ) | x D k ( t ) } and | D |
  6:
end for
  7:
for  l 1   to   q  do
  8:
       v b l ( t ) , b { 0 , 1 } ← Equation (5) with { w d ( t , x ) , w h ( t , x ) , u b l ( x ) | x D k ( t ) }
  9:
       r l ( t ) ← Equation (6) with v 1 l ( t ) and v 0 l ( t )
10:
end for
11:
for  l 1   to   q  do
12:
       y ^ t l l with { r l ( t ) | 1 l q }
13:
end for
14:
return y ^ t { y ^ t l | 1 l q }

3.5. Complexity Analysis

Let n be the number of samples in the training set, d be the dimension of a feature vector, q be the number of possible labels, and k be the number of nearest neighbors to take into account. First, for the training process, the cost of searching for nearest neighbors is O ( n 2 d ) . The cost of counting k-occurrence and label k-occurrence is O ( n 2 q ) . The complexity of computing membership is O ( n q ) . The cost of computing the weighting factors w d and w h is O ( n k ) . The cost of computing the label relevance score is O ( n q k ) . The complexity of training the logistic regression model is O ( n q 2 ) . Accordingly, the overall complexity of the training process of MLHiKNN is O ( n 2 ( d + q ) + n q 2 + n q k ) . For the prediction of a query instance, the cost of searching for nearest neighbors is O ( n d ) . The cost of computing the weighting factors w d and w h is O ( k ) . The cost of computing the label relevance score is O ( q k ) . The complexity of prediction with the logistic regression model is O ( q 2 ) . So, the overall complexity of MLHiKNN for predicting a query example is O ( n d + q 2 + q k ) .

4. Experimental Results and Discussions

We conducted experiments on 28 benchmark datasets to validate the effectiveness of the proposed approach. In Section 4.2, the results of the experiments comparing MLHiKNN with nine state-of-the-art MLC algorithms are first reported, followed by those comparing MLHiKNN with four commonly used kNN-based MLC algorithms under five different settings of parameter k. The proposed approach is also compared with three hubness reduction techniques, which are incorporated into both BRkNNa and MLKNN. How these hubness reduction techniques work with the two MLC algorithms is introduced in Section 4.1.2. The experimental results are reported in Section 4.3. Afterward, some experimental results for further analysis of MLHiKNN are reported. In Section 4.4, we report the results of experiments for analyzing the effectiveness of each component in the proposed approach. In Section 4.5, the results of parameter analysis experiments are reported to provide recommendations for the parameters k , λ , and β in MLHiKNN.

4.1. Experimental Setup

4.1.1. Datasets and Metrics

Twenty-eight benchmark datasets from domains including audio, music, biology, text, image, and video were used in the experiments and are listed in Table 1 ordered by sample size. Each dataset is available at the MULAN [63] or MEKA [64] Library. These datasets vary from 327 to 42,177 in sample size, from 53 to 1186 in the dimensionality of feature space, and from 6 to 215 in the number of possible labels. Table 1 also lists some common indicators for measuring properties of a multi-label dataset. LCard denotes l a b e l   c a r d i n a l i t y , the average number of labels per instance. It measures the degree of multi-labeledness. L D e n denotes l a b e l   d e n s i t y , which is the normalization of label cardinality by the number of possible labels. L D i v denotes l a b e l   d i v e r s i t y , which is the number of distinct label sets that have appeared in the dataset. P L D i v denotes the p r o p o r t i o n   o f   d i s t i n c t   l a b e l   s e t s , which is the normalization of label diversity by the number of examples. Some measures for characterizing the properties of the hubness of a dataset are also shown in this table. The proportion of nearest-neighbor slots occupied by hubs, denoted as H o c c k , is defined in Equation (26):
H o c c k = x D h u b O k ( x ) n k ,
where D h u b is the set consisting of all hubs. n is the number of samples in the dataset. Similarly, the proportion of antihubs in the dataset, A o c c k , is given by Equation (27):
A o c c k = | { x | O k ( x ) = 0 } | n .
This equation represents the fraction of data points not selected as one of the k nearest neighbors by any other sample. In Table 1, the hubness measures were computed with k = 10 . Meanwhile, an example was identified as a hub if its k-occurrence was larger than 2 k and an antihub if its k-occurrence was zero. To reduce the influence of extremely infrequent labels, those with fewer than 20 instances were removed from each dataset. Similar label-frequency thresholds have been adopted in previous work to mitigate the impact of rare labels [65,66]. The chi-squared test was used to reduce the dimensionality of feature space for text datasets. In each experiment, 2-fold cross-validation was performed five times. Specifically, each dataset was randomly split into two equal-sized parts, with each part used once as the training set and once as the testing set. This 2-fold cross-validation procedure was repeated five times with different random splits. The average results are reported.
Many metrics have been proposed to evaluate the generalization performance of MLC algorithms. Generally, existing multi-label metrics can be categorized into two groups, example-based and label-based metrics, according to the way they deal with multiple labels. Example-based metrics first perform evaluations on each test example by considering all possible labels and then return the average value over all test instances. In contrast, label-based metrics first perform evaluations on each class label and then return the averaged value across all possible labels. In addition, similar to single-label settings, multi-label metrics can also be categorized into classification and ranking metrics according to the perspective from which the performance is measured. Importantly, different metrics vary in their sensitivity to majority and minority labels. Example-based metrics tend to be more sensitive to majority labels. For those measured from the classification perspective, majority labels contribute more often to per-instance calculations, since these labels are more likely to appear in the true label set of each instance and are also more frequently predicted by models that do not aggressively optimize recall for minority labels. For these ranking-based metrics, such as Ranking Loss, majority labels participate more often in pairwise comparisons against irrelevant labels because they are more likely to be included in the label set of each instance. Label-based metrics can be computed using two different averaging strategies: micro-averaging and macro-averaging. Micro-averaging aggregates predictions across all labels and instances and is also more influenced by majority labels. For example, in classification metrics, majority labels typically have higher overall counts of true positives and false negatives. Meanwhile, minority labels often yield smaller numbers of true positives and false positives, particularly when the classifier does not aggressively optimize their recall. Ranking-based metric AUC micro compares all relevant and irrelevant instance–label pairs. Majority labels generate more relevant instance–label pairs and thus dominate the set of relevant and irrelevant comparisons. In contrast, macro-averaging computes metric values independently for each label and then returns the averaged value. This strategy provides greater sensitivity to the performance of minority labels than micro-averaging and example-based metrics. To ensure a comprehensive evaluation of the classifier performance, we employed six evaluation metrics in our experimental analysis, including AUC macro, AUC micro, Ranking Loss, F1 macro, F1 micro, and Hamming Loss. The categorization of these metrics is summarized in Table 2.

4.1.2. Compared Approaches

To validate the effectiveness of the proposed method, we experimentally compared MLHiKNN with some well-established approaches, including nine MLC algorithms and three hubness reduction techniques. The compared MLC algorithms include BR [14], CC [17], ECC [17], RAkEL [18], RAkELd [18], BRkNNa [5], MLKNN [6], DMLkNN [7], and IBLR [8]. We employed the implementations in the Scikit-multilearn [67] library for BRkNNa and MLKNN and the implementations in the MULAN [63] library for the other seven approaches. For BR, CC, ECC, RAkEL, and RAkELd, we used C4.5 as the base classifier and employed its implementation J48 on the WEKA [68] platform. For ECC, we followed the default parameter settings. Specifically, ECC employed sampling with replacement to create the data for each base model in the ensemble. Each model was trained on 100% of the instances. The number of models was set to 10. For RAkEL and RAkELd, the size of the label subsets was set to 3. The number of models was determined adaptively based on the number of labels and the subset size, following the default settings. For the kNN-based methods, BRkNNa, MLKNN, DMLkNN, IBLR, and MLHiKNN, the number of nearest neighbors to take into account, k, was set to 20 based on both the prior literature and empirical results. Previous studies typically adopted k values ranging from 8 to 12 for MLKNN, from 8 to 10 for DMLkNN, and from 1 to 30 for BRkNNa, and used k = 10 for IBLR. In our experiments, we further evaluated each method with k { 5 , 10 , 15 , 20 , 25 } and observed that the optimal performance tended to occur when k was between 15 and 25 for BRkNNa, DMLkNN, and IBLR, and between 5 and 20 for MLKNN, in terms of AUC macro. To ensure a fair and consistent comparison across methods, we fixed k = 20 for all methods and also report comparison results for these kNN-based methods under each k value in Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7. The smooth parameter was set to 1.0 for MLKNN and DMLkNN. Other parameters followed the default parameter settings. For the proposed approach, MLHiKNN, λ and β were set to 3 and 2, respectively.
The compared hubness reduction techniques include local dissimilarity (DSL) [34], local scaling (LS) [40,41], and mutual proximity (MP) [40,41]. For the three hubness reduction techniques, we employed their implementations in the Scikit-hubness [69] library. For DSL and LS, k was set to be 20. For MP, the Gaussian method was used to model the distance distributions. Other parameters all followed the default settings.
To compare the hubness reduction techniques with MLHiKNN, we incorporated these techniques into BRkNNa and MLKNN, respectively, resulting in two groups of methods. They are hubness-reduced BRkNNa approaches (Algorithm 4), including BRkNNa-dsl, BRkNNa-ls, BRkNNa-mp, and hubness-reduced MLKNN approaches (Algorithm 5), including MLKNN-dsl, MLKNN-ls, and MLKNN-mp. Compared with the hubness-reduced BRkNNa approaches, the hubness-reduced MLKNN approaches also need to search for the k nearest neighbors for each training example with hubness reduction techniques to compute the posterior probabilities in the training process.
Algorithm 4 Hubness-reduced BRkNNa.
Input:  
  D: training set;
   P h : parameters needed for BRkNNa and the hubness reduction technique;
   D t e s t : testing set.
Output:  
   Y ^ t e s t : label predictions of testing set.
1:
for each instance t in  D t e s t  do
2:
       G h b r ( t , D ) Compute the transformed pairwise distances or relations between t and all examples in D with a hubness reduction technique based on the dataset { t D }
3:
       D k ( t ) Identify the kNNs of t based on the relations in G h b r ( t , D )
4:
       y ^ t Predict the label set of t with BRkNNa taking D k ( t ) as the kNNs
5:
end for
6:
return Y ^ t e s t { y ^ t | t D t e s t }
Algorithm 5 Hubness-reduced MLKNN.
Input:  
  D: training set;
   P h : parameters needed for MLKNN and the hubness reduction technique;
   D t e s t : testing set.
Output:  
   Y ^ t e s t : label predictions of testing set.
  1:
Compute the prior probabilities as performed with MLKNN
  2:
for each instance x in D do
  3:
       G h b r ( x , D ) Compute the transformed pairwise distances or relations between x and all examples in D with a hubness reduction technique based on the training set D
  4:
       D k ( x ) Identify the kNNs of x based on the relations in G h b r ( x , D )
  5:
end for
  6:
Compute the posterior probabilities as performed by MLKNN based on the neighboring relations in { D k ( x ) | x D }
  7:
for each instance t in  D t e s t  do
  8:
       G h b r ( t , D ) Compute the transformed pairwise distances or relations between t and all examples in D with a hubness reduction technique based on the dataset { t D }
  9:
       D k ( t ) Identify the kNNs of t based on the relations in G h b r ( t , D )
10:
       y ^ t Predict the label set of t with the prior and posterior probabilities as performed by MLKNN, taking D k ( t ) as the kNNs
11:
end for
12:
return Y ^ t e s t { y ^ t | t D t e s t }

4.2. Comparisons with MLC Algorithms

In this subsection, we first report the experimental results of MLHiKNN and the nine MLC algorithms under the parameter settings mentioned in the previous subsection. Then, we report the comparison results among MLHiKNN and the four kNN-based MLC algorithms, BRkNNa, MLKNN, DMLkNN, and IBLR, under different settings of the parameter k. Table 3 shows the performance in terms of AUC macro of the compared MLC algorithms on each dataset. Each result contains the mean performance and the standard deviation across ten testing sets. For each dataset, the rank on the mean performance of each algorithm is enclosed in parentheses. For each dataset, the performance of the top-ranked algorithm is marked with •. For each algorithm, the average rank across all the datasets is shown in the second line from the bottom, with the best average rank shown in bold. The last line shows the win, tie, and loss counts of each algorithm, which were obtained by a pairwise Wilcoxon signed-rank test between each pair of algorithms on each dataset across the performance on ten testing sets. The significance level was set to 0.05. Specifically, a win (loss) was counted for an algorithm if it performed significantly better (worse) than another algorithm on a dataset. Otherwise, a tie was recorded. For brevity, the experimental results of the compared MLC algorithms on the other five metrics, AUC micro, Ranking Loss, F1 macro, F1 micro, and Hamming Loss, are shown in Table A2, Table A3, Table A4, Table A5 and Table A6. Meanwhile, the average ranks of these algorithms on each metric are summarized in Table 4 for ease of viewing, with the best average rank in each case shown in bold. As Table 3 shows, the proposed method MLHiKNN is top-ranked on 26 out of 28 benchmark datasets on AUC macro. MLHiKNN also achieves the best average rank, the highest win counts, and the lowest loss counts on AUC macro. For the other five metrics, MLHiKNN achieves the best average rank on AUC micro, Ranking Loss, and Hamming Loss, shares the third-best average rank with RAkELd on F1 macro, and ranks fifth on average based on F1 micro. In terms of F1 macro, RAkEL and BR achieve the first- and second-best ranks. In terms of F1 micro, RAkEL, ECC, BR, and RAkELd obtain the highest four average ranks. Meanwhile, MLHiKNN outperforms the other four kNN-based MLC algorithms on all metrics.
To statistically analyze the MLC algorithms across all these datasets, the Friedman test was conducted at a significance level of 0.05. The results on each metric are shown in Table 5. Then, the pairwise Wilcoxon signed-rank test was employed for each pair of these algorithms across all the datasets as the post hoc test [70]. Meanwhile, the significant level of 0.05 was corrected by the Bonferroni method to control the family-wise Type-I error. The results are shown in Table 6, Table 7, Table 8, Table 9, Table 10 and Table 11. In each table, the i , j entry is the p-value of the pairwise Wilcoxon signed-rank test, whose alternative hypothesis is that the i-th method performs significantly better than the j-th method based on their performance on all the datasets. The p-values less than the corrected significant level are filled in gray. As shown in Table 5, the test statistic of the Friedman test is larger than the Critical Value for each metric. So, the null hypothesis of the Friedman test is rejected in terms of each metric. According to Table 6, Table 7, Table 8, Table 9, Table 10 and Table 11, MLHiKNN significantly outperforms all the other nine MLC algorithms in terms of AUC macro. For the other two ranking-based metrics, AUC micro and Ranking Loss, MLHiKNN performs significantly better than all compared algorithms, except for DMLkNN and IBLR on both metrics, and BRkNNa on AUC micro. For the three classification-based metrics, MLHiKNN significantly outperforms BRkNNa and DMLkNN in all cases. In terms of Hamming Loss, it also achieves significantly better performance than all other compared algorithms, except for ECC and RAkEL. Moreover, no algorithm significantly outperforms MLHiKNN on any of the evaluated metrics. These experimental results indicate that MLHiKNN is highly competitive among the compared MLC algorithms.
Further, since MLHiKNN is a k-nearest neighbor approach, we also compared MLHiKNN with the four kNN-based MLC algorithms, BRkNNa, MLKNN, DMLkNN, and IBLR, under different settings of k. Experiments were conducted on the five algorithms with parameter k set to be each value in { 5 , 10 , 15 , 20 , 25 } . Hence, there were a total of twenty-five combinations of algorithm and k. For each combination of algorithm and k, its average rank across all the datasets out of the twenty-five combinations is reported in Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7 in terms of each metric. Thus, it is convenient to observe the performance rank differences among the five algorithms under each setting of k and among the combinations of one algorithm with different ks. Then, we try to answer three questions based on the results of these experiments. First, how does parameter k influence the performance of the kNN-based MLC algorithms? Second, how does the proposed method MLHiKNN perform compared with the other four kNN-based algorithms? Third, how does the performance of these algorithms vary across different metrics? According to Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7, the following observations can be made to answer the above questions. First, in terms of the three ranking-based metrics, AUC macro, AUC micro, and Ranking Loss, the average performance of all algorithms generally improves at first and then tends to stabilize as k increases from 5 to 25 in steps of 5, except for MLKNN on all three metrics and DMLkNN on AUC micro. Naturally, considering more nearest neighbors can bring more information for predicting labels. However, the performance of these algorithms does not always exhibit the same tendency in terms of the three classification-based metrics as that in terms of the ranking-based metrics as k increases. One reason for that lies in the class imbalance. For MLKNN, the performance decreases as k increases from 5 to 25 in terms of Ranking Loss, F1 macro, and F1 micro. The reason is that MLKNN makes a prediction for each possible label l following an MAP principle based on the number of examples with label l among the kNNs. A larger k leads to more possible values of such a number. Thus, on average, there are fewer instances for each value of this number, since the total number of training examples is fixed. As a result, the posterior probability estimation may be less accurate for some values of this number due to the lack of examples. The second observation is that MLHiKNN achieves the best average rank on each value of k in terms of AUC macro and Hamming Loss. For AUC micro, MLHiKNN achieves the best average rank on two values of k and the second-best rank on three values of k. For Ranking Loss, MLHiKNN achieves the best average rank on one value of k and the second-best rank on four values of k. In terms of F1 macro and F1 micro, it achieves the best average rank on four values of k and the second-best rank on one value of k. These results indicate the competitive performance of MLHiKNN on different values of k compared with the other kNN-based MLC algorithms. Third, the performance of these algorithms varies greatly across different metrics. For example, MLKNN achieves better average ranks on F1 macro than on AUC macro. DMLkNN achieves better average ranks on Ranking Loss than on F1 macro. IBLR ranks better on average on F1 macro than on Hamming Loss. These differences in performance across the metrics show the different focuses of these algorithms. Specifically, minority classes have more influence on macro-averaging metrics than on micro-averaging metrics and example-based metrics. These metrics also have the difference of classification and ranking. Note that MLHiKNN always achieves the best or second-best average rank in terms of each metric. It shows the effectiveness of MLHiKNN across different focuses.

4.3. Comparisons with Hubness Reduction Techniques

In this subsection, we report the experimental results of the hubness-reduced MLC approaches, introduced in Section 4.1.2. The performance on each dataset is shown in Table 12, Table A7, Table A8, Table A9, Table A10 and Table A11. Meanwhile, the average ranks of the algorithms in terms of each metric are organized into Figure 8 for ease of viewing. For statistical analysis, the Friedman test and post hoc pairwise Wilcoxon signed-rank test were also employed. The significance level was set to 0.05. The results of the Friedman test are shown in Table A12. The p-values of the post hoc test are shown in Table 13, Table A13, Table A14, Table A15, Table A16 and Table A17. The representations of the contents in these tables follow those introduced in Section 4.2. We try to answer two questions based on the results reported in this subsection. First, do the hubness reduction techniques improve the performance of BRkNNa and MLKNN? Second, how does MLHiKNN perform compared with the hubness-reduced MLC approaches? For the first question, both BRkNNa-dsl and BRkNNa-ls achieve better average ranks than BRkNNa on all six metrics, indicating that incorporating the two hubness reduction techniques, DSL and LS, leads to performance improvements for BRkNNa. For MLKNN, the three variants, MLKNN-dsl, MLKNN-ls, and MLKNN-mp, all achieve worse average ranks than MLKNN across all these metrics, except for Hamming Loss. This is because MLKNN is less affected by the hubness phenomenon than BRkNNa, owing to its MAP strategy. It might not benefit much from hubness reduction. Instead, the distance transformation in the hubness reduction techniques, which is based on the distribution estimated using training examples, introduces the negative effects of the error between the estimated distribution and true distribution to MLKNN in its training and predicting processes. For the second question, MLHiKNN achieves the best average rank in terms of each metric. Meanwhile, it performs significantly better than all the compared approaches discussed in this subsection in terms of AUC macro, Ranking Loss, and Hamming Loss. For F1 macro and F1 micro, it also achieves significantly better performance than all compared approaches, except for MLKNN. In terms of AUC micro, MLHiKNN outperforms both MLKNN and its three variants, MLKNN-dsl, MLKNN-ls, and MLKNN-mp. It shows MLHiKNN is highly competitive compared with the hubness-reduced MLC approaches.

4.4. Ablation Analysis for MLHiKNN

To validate the effectiveness of each strategy in the proposed method, we conducted experiments on some variants of MLHiKNN for ablation analysis. MLHiKNN-g1: g ( x ) was fixed at 1 for computing Equations (9) and (10). Thus, for each neighbor of a query instance t, its membership was either 0 or 1 according to its label. When computing its membership, the hubness information was not considered. There was no fuzziness, actually. MLHiKNN-g0: g ( x ) was fixed at 0 for computing Equations (9) and (10). Thus, for each training example, its membership only depended on its hubness information. As a neighbor, x voted by its history k-occurrence and label k-occurrence information rather than its label. The labels of each neighbor were ignored in the kNN voting. MLHiKNN-h1: w h ( t , x ) was fixed at 1 for computing Equation (5). Thus, for the kNNs of a query instance, the history k-occurrence made no difference in their importance in kNN voting. The hubness weighting was removed. MLHiKNN-d1: w d ( t , x ) was fixed at 1 for computing Equation (5). Thus, for the kNNs of a query instance t, the distance to t made no difference in their importance in the kNN voting. The distance weighting was removed. MLHiKNN-fo: The label relevance score computed by Equation (6) was directly used for prediction without the logistic regression model. Specifically, label l was assigned to the query instance t if and only if r l ( t ) 0.5 . MLHiKNN-fo is a first-order approach, accordingly. The above approaches are summarized in Table 14. The performance of MLHiKNN and each variant approach on each dataset is shown in Table 15 and Table A18, Table A19, Table A20, Table A21 and Table A22. The average ranks across all the datasets are shown in Figure 9. As the figure shows, MLHiKNN achieves the second-best average rank on F1 macro and the best average rank on the remaining five metrics. It verifies the effectiveness of each strategy in MLHiKNN.

4.5. Parameter Analysis for MLHiKNN

Three parameters in MLHiKNN may affect the performance to some extent. k is the number of nearest neighbors to take into account. β determines how heavily the distance is weighted. λ is the threshold parameter for computing g ( · ) in Equation (19). For a proper range of each parameter, we conducted experiments with MLHiKNN using different parameters on all the datasets. The default values of k, λ , and β were 20, 3, and 2, respectively. In the experiments for one parameter, the other two were fixed at the default values. Note that the following two equations hold for λ , k, and g ( x ) , according to Equation (19),
g ( x ) = 0 , if 0 < λ 1 k ,
g ( x ) = 1 , if λ + .
Accordingly, λ was set to vary in { 0.05 , 0.5 , 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , + } , where the first value, 0.05, equals one over the default value of k. Parameter β was used for calculating w d ( t , x ) in Equation (20) and appeared in the exponent in the form of 2 / ( β 1 ) . β was set to vary in { 5 / 4 , 3 / 2 , 5 / 3 , 2 , 3 , 5 , 9 , 17 } in the experiments for parameter analysis. Thus, the exponent 2 / ( β 1 ) varied in { 8 , 4 , 3 , 2 , 1 , 1 / 2 , 1 / 4 , 1 / 8 } . Parameter k began at 3 and then varied from 5 to 40 with a step size of 5. Table A23, Table A24, Table A25, Table A26, Table A27, Table A28, Table A29, Table A30, Table A31, Table A32, Table A33, Table A34, Table A35, Table A36, Table A37, Table A38, Table A39 and Table A40 report the performance of MLHiKNN under varying values of k, λ , and β , respectively. The performance trends with respect to parameter changes varied across datasets and also differed depending on the evaluation metrics. The average ranks across all the datasets of MLHiKNN with different parameters are reported in Figure 10, Figure 11 and Figure 12. For parameter k, as it increases, the average rank first becomes higher and then becomes lower on four metrics. In terms of AUC micro and Ranking Loss, the average rank first becomes higher and then levels off. For parameter λ , as it increases, the average rank first becomes higher and then becomes lower on four metrics. In terms of F1 macro and F1 micro, the average rank generally keeps becoming lower. For parameter β , as it increases, the average rank first becomes higher and then becomes lower on six metrics. Generally, the best average ranks appear in the middle region of the parameter variation ranges in most cases. It is more likely for MLHiKNN to achieve good performance when parameter k ranges from 10 to 25, parameter λ from 2 to 5, and parameter β from 2 to 5, respectively.

5. Conclusions

In high-dimensional data, a small number of samples may appear in the k-neighborhoods of other examples much more frequently. This is known as the hubness phenomenon, which widely exists in the real world. To a certain extent, this phenomenon affects the performance of machine learning algorithms, which involves searching for nearest neighbors. However, the hubness phenomenon has not been tackled in MLC algorithms. Hence, in this work, we propose a novel MLC algorithm named MLHiKNN, which alleviates the negative effects of the hubness issue in a kNN-based multi-label learning framework. Specifically, MLHiKNN involves a kNN voting process. As a neighbor, the degree of its support for a label is determined by its relevance to this label, its history k-occurrence, and the labels of its reverse nearest neighbors. Meanwhile, in the voting process, each neighbor is weighted by information on its hubness and distance. Further, MLHiKNN exploits high-order label correlations by training a logistic regression model with the kNN voting results with respect to all possible labels. Experiments on twenty-eight benchmark datasets, nine commonly used MLC algorithms, and three hubness reduction techniques indicate that MLHiKNN is highly competitive among the well-established methods in dealing with MLC problems in both high- and low-dimensional cases.
The main difference between the hubness reduction techniques and MLHiKNN is that they address the hubness issue from data-level and algorithm-level perspectives, respectively. The hubness reduction techniques transform the original pairwise distance matrix to a hubness-reduced dissimilarity matrix, leading to changes in the neighbors graph. While MLHiKNN does not change the graph of kNNs for any point, it utilizes the occurrence information of the hubs to deal with the hubness problem. Our future work is to combine the two types of techniques to better handle multi-label datasets with hubs.

Author Contributions

Conceptualization, Z.T.; funding acquisition, M.H.; investigation, Z.T.; methodology, Z.T.; project administration, M.H.; software, Z.T.; supervision, M.H. and X.W.; validation, S.T.; writing—original draft, Z.T.; writing—review and editing, S.T., M.H., and X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China Key Supported Project of the Major Research Plan under grant number 92267206, the National Natural Science Foundation of China under grant number 62032013, and the Liaoning Revitalizing Talent Program under grant number XLYC2202045.

Data Availability Statement

The source code of the proposed method, MLHiKNN, is available at https://github.com/T-zeyuteng/MLHiKNN. The datasets in this study are available in MULAN at https://mulan.sourceforge.net/datasets-mlc.html (accessed on 16 February 2025), reference number [63]; MEKA at https://waikato.github.io/meka/datasets/(accessed on 16 February 2025), reference number [64].

Acknowledgments

We would like to thank Yonggui Zhang, Xianqiang Wang, Xinfeng Li, Huawu Zhang, and Guomin Peng for their valuable suggestions and support. We are also grateful to the editor and reviewers for their constructive comments and helpful feedback.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
MLCMulti-label classification
SLCSingle-label classification
MLDsMulti-label datasets
SLDsSingle-label datasets
kNNk-nearest neighbor
MLHiKNNMulti-label hubness information-based k-nearest neighbor
LSLocal scaling
MPMutual proximity
DSLLocal dissimilarity
D k ( t ) A set consisting of the k-nearest neighbors of example t
O k ( x ) The number of times x appears among the k nearest neighbors of all other points in a dataset
O k l ( x ) The number of instances relevant to label l in D k ( x )

Appendix A. Time Costs of the Proposed Method

The time costs of BRkNNa, MLKNN, MLHiKNN-fo, and MLHiKNN in the experiments are shown in Table A1. For BRkNNa and MLKNN, we employed the implementations in the Scikit-multilearn [67] library (version 0.2.0), based on Python 3.8. MLHiKNN and MLHiKNN-fo were developed using Python. The CPU used was an AMD 5950X, with a RAM size of 96 GB.
Table A1. Average time costs across the 5 × 2-fold training set and testing set in seconds.
Table A1. Average time costs across the 5 × 2-fold training set and testing set in seconds.
DatasetBRkNNaMLKNNMLHiKNN-foMLHiKNN
TrainTestTrainTestTrainTestTrainTest
birds 0.020.050.050.590.591.140.59
CAL500 0.030.600.300.110.1213.010.17
emotions 0.030.070.090.140.150.430.16
genbase 0.040.110.120.230.240.720.26
LLOG 0.081.040.570.040.063.820.10
enron 0.100.560.450.910.924.010.98
scene 0.120.250.359.399.3910.029.68
yeast 0.120.440.421.131.172.641.23
Slashdot 0.130.410.460.020.070.950.09
corel5k 0.295.813.370.070.1657.890.71
rcv1subset1 0.353.892.520.110.2323.560.40
rcv1subset2 0.353.982.570.100.2223.640.36
rcv1subset3 0.353.992.570.100.2223.590.36
rcv1subset4 0.373.942.550.090.2222.930.35
rcv1subset5 0.354.022.590.090.2223.770.35
bibtex 0.428.955.160.120.2763.460.81
Arts 0.441.811.650.130.283.720.38
Health 0.542.031.920.420.623.780.63
Business 0.672.782.470.240.464.770.57
Education 0.763.182.780.680.965.860.87
Computers 0.823.332.920.751.029.091.02
Entertainment 0.842.472.510.901.184.051.20
Recreation 0.832.942.760.510.764.370.76
Society 1.013.813.381.001.319.491.29
eurlex-dc-l 1.1529.2116.561.261.5490.642.11
eurlex-sm 1.3320.0611.821.792.2274.253.35
tmc2007-500 2.597.126.538.479.3415.647.77
mediamill 4.5233.9020.46282.61285.35395.58275.31

Appendix B. Experimental Results of Compared MLC Algorithms

Table A2. The performance of nine compared MLC algorithms and MLHiKNN on AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A2. The performance of nine compared MLC algorithms and MLHiKNN on AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Micro
BRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
birds0.666(8) ± 0.0390.660(9) ± 0.0280.785(1) ± 0.015 •0.761(2) ± 0.0110.676(7) ± 0.0250.736(5) ± 0.0140.659(10) ± 0.0130.730(6) ± 0.0160.736(4) ± 0.0140.746(3) ± 0.019
CAL5000.628(7) ± 0.0110.589(9) ± 0.0100.731(4) ± 0.0040.688(6) ± 0.0020.601(8) ± 0.0050.749(3) ± 0.0030.713(5) ± 0.0030.755(2) ± 0.0010.558(10) ± 0.0070.757(1) ± 0.004 •
emotions0.693(9) ± 0.0220.688(10) ± 0.0190.837(4) ± 0.0080.812(6) ± 0.0090.697(8) ± 0.0140.852(3) ± 0.0050.747(7) ± 0.0110.828(5) ± 0.0070.856(2) ± 0.0070.862(1) ± 0.005 •
genbase0.996(7) ± 0.0060.995(8) ± 0.0070.997(1) ± 0.003 •0.997(3) ± 0.0040.996(6) ± 0.0060.993(9) ± 0.0040.970(10) ± 0.0120.997(2) ± 0.0020.997(4) ± 0.0030.996(5) ± 0.004
LLOG0.771(5) ± 0.0020.670(8) ± 0.0210.682(7) ± 0.0270.659(9) ± 0.0080.618(10) ± 0.0230.788(4) ± 0.0050.791(3) ± 0.0020.805(1) ± 0.003 •0.805(2) ± 0.0030.767(6) ± 0.002
enron0.764(8) ± 0.0070.760(9) ± 0.0110.851(4) ± 0.0050.823(6) ± 0.0050.751(10) ± 0.0090.801(7) ± 0.0120.825(5) ± 0.0040.863(1) ± 0.003 •0.855(3) ± 0.0050.860(2) ± 0.002
scene0.747(9) ± 0.0190.744(10) ± 0.0200.928(5) ± 0.0050.902(6) ± 0.0030.755(8) ± 0.0120.934(4) ± 0.0030.885(7) ± 0.0040.940(3) ± 0.0040.945(2) ± 0.0030.950(1) ± 0.002 •
yeast0.683(8) ± 0.0100.652(10) ± 0.0130.816(5) ± 0.0040.791(6) ± 0.0030.679(9) ± 0.0070.837(3) ± 0.0030.766(7) ± 0.0040.832(4) ± 0.0030.840(2) ± 0.0040.842(1) ± 0.004 •
Slashdot0.937(5) ± 0.0040.940(4) ± 0.0030.892(10) ± 0.0040.892(9) ± 0.0040.918(7) ± 0.0080.932(6) ± 0.0080.912(8) ± 0.0090.951(1) ± 0.004 •0.948(2) ± 0.0060.945(3) ± 0.004
corel5k0.776(3) ± 0.0040.744(6) ± 0.0030.573(9) ± 0.0080.558(10) ± 0.0030.714(7) ± 0.0060.713(8) ± 0.0080.772(5) ± 0.0040.802(1) ± 0.003 •0.775(4) ± 0.0050.796(2) ± 0.004
rcv1subset10.788(7) ± 0.0060.727(10) ± 0.0080.814(6) ± 0.0040.748(8) ± 0.0050.746(9) ± 0.0080.894(3) ± 0.0040.866(5) ± 0.0030.913(1) ± 0.002 •0.901(2) ± 0.0040.874(4) ± 0.009
rcv1subset20.803(6) ± 0.0070.736(9) ± 0.0090.800(7) ± 0.0040.730(10) ± 0.0050.762(8) ± 0.0090.895(2) ± 0.0030.874(4) ± 0.0030.914(1) ± 0.003 •0.893(3) ± 0.0050.865(5) ± 0.010
rcv1subset30.797(7) ± 0.0050.712(10) ± 0.0090.808(6) ± 0.0040.727(9) ± 0.0050.763(8) ± 0.0090.893(3) ± 0.0030.871(4) ± 0.0030.912(1) ± 0.003 •0.896(2) ± 0.0020.861(5) ± 0.007
rcv1subset40.818(7) ± 0.0070.730(10) ± 0.0100.819(6) ± 0.0040.732(9) ± 0.0050.783(8) ± 0.0060.906(2) ± 0.0040.884(4) ± 0.0020.922(1) ± 0.002 •0.904(3) ± 0.0030.884(5) ± 0.010
rcv1subset50.809(6) ± 0.0090.756(9) ± 0.0090.807(7) ± 0.0050.743(10) ± 0.0040.768(8) ± 0.0060.891(3) ± 0.0030.871(4) ± 0.0020.914(1) ± 0.002 •0.896(2) ± 0.0030.868(5) ± 0.009
bibtex0.788(6) ± 0.0070.776(7) ± 0.0070.709(9) ± 0.0030.667(10) ± 0.0040.739(8) ± 0.0080.838(3) ± 0.0030.802(5) ± 0.0040.863(1) ± 0.003 •0.821(4) ± 0.0070.863(2) ± 0.006
Arts0.789(5) ± 0.0060.787(6) ± 0.0050.765(7) ± 0.0040.689(10) ± 0.0040.722(8) ± 0.0230.818(4) ± 0.0030.713(9) ± 0.0040.840(3) ± 0.0030.845(2) ± 0.0040.851(1) ± 0.003 •
Health0.864(5) ± 0.0050.862(6) ± 0.0080.857(7) ± 0.0040.819(9) ± 0.0030.817(10) ± 0.0100.893(4) ± 0.0040.827(8) ± 0.0040.910(3) ± 0.0030.912(2) ± 0.0030.919(1) ± 0.002 •
Business0.908(5) ± 0.0040.895(7) ± 0.0020.869(9) ± 0.0040.853(10) ± 0.0030.882(8) ± 0.0060.927(4) ± 0.0020.908(6) ± 0.0020.945(2) ± 0.0020.943(3) ± 0.0010.947(1) ± 0.001 •
Education0.830(6) ± 0.0040.840(5) ± 0.0050.823(7) ± 0.0040.746(10) ± 0.0030.757(9) ± 0.0220.876(4) ± 0.0020.822(8) ± 0.0030.900(3) ± 0.0020.900(2) ± 0.0020.906(1) ± 0.002 •
Computers0.827(5) ± 0.0030.817(8) ± 0.0070.822(6) ± 0.0040.776(10) ± 0.0030.779(9) ± 0.0130.862(4) ± 0.0030.820(7) ± 0.0030.888(3) ± 0.0030.888(2) ± 0.0030.896(1) ± 0.003 •
Entertainment0.821(6) ± 0.0050.805(7) ± 0.0050.847(5) ± 0.0060.775(9) ± 0.0030.786(8) ± 0.0100.863(4) ± 0.0020.754(10) ± 0.0040.879(3) ± 0.0020.880(2) ± 0.0010.891(1) ± 0.002 •
Recreation0.810(5) ± 0.0040.796(7) ± 0.0040.800(6) ± 0.0040.717(9) ± 0.0040.756(8) ± 0.0190.839(4) ± 0.0050.697(10) ± 0.0040.859(3) ± 0.0040.863(2) ± 0.0040.874(1) ± 0.004 •
Society0.805(6) ± 0.0030.812(5) ± 0.0030.772(7) ± 0.0040.722(10) ± 0.0020.738(9) ± 0.0180.829(4) ± 0.0030.753(8) ± 0.0040.858(3) ± 0.0020.860(2) ± 0.0020.869(1) ± 0.002 •
eurlex-dc-l0.837(6) ± 0.0030.827(7) ± 0.0040.758(9) ± 0.0040.712(10) ± 0.0020.799(8) ± 0.0050.882(3) ± 0.0040.875(4) ± 0.0030.895(2) ± 0.0030.863(5) ± 0.0030.919(1) ± 0.005 •
eurlex-sm0.871(6) ± 0.0030.866(7) ± 0.0040.837(8) ± 0.0020.806(10) ± 0.0010.831(9) ± 0.0050.919(4) ± 0.0020.917(5) ± 0.0020.936(2) ± 0.0020.932(3) ± 0.0020.952(1) ± 0.002 •
tmc2007-5000.842(9) ± 0.0030.845(8) ± 0.0040.920(5) ± 0.0010.907(6) ± 0.0010.832(10) ± 0.0040.928(4) ± 0.0010.891(7) ± 0.0020.942(3) ± 0.0010.942(2) ± 0.0010.958(1) ± 0.001 •
mediamill0.815(8) ± 0.0040.780(9) ± 0.0030.879(6) ± 0.0020.857(7) ± 0.0010.763(10) ± 0.0040.930(5) ± 0.0010.930(4) ± 0.0000.950(2) ± 0.0010.949(3) ± 0.0010.957(1) ± 0.001 •
average rank6.437.866.188.188.394.146.392.292.892.25
win/tie/loss86/24/14255/22/17597/14/14141/16/19532/22/198161/9/8297/13/142208/14/30189/18/45211/14/27
Table A3. The performance of nine compared MLC algorithms and MLHiKNN on Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A3. The performance of nine compared MLC algorithms and MLHiKNN on Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetRanking Loss
BRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
birds0.362(7) ± 0.0390.372(9) ± 0.0310.317(5) ± 0.0230.370(8) ± 0.0260.391(10) ± 0.0280.317(4) ± 0.0200.345(6) ± 0.0180.271(3) ± 0.0180.264(2) ± 0.0200.251(1) ± 0.020 •
CAL5000.403(7) ± 0.0130.450(8) ± 0.0120.337(5) ± 0.0040.381(6) ± 0.0040.485(10) ± 0.0100.282(3) ± 0.0030.284(4) ± 0.0030.242(2) ± 0.0020.456(9) ± 0.0070.240(1) ± 0.005 •
emotions0.385(9) ± 0.0210.385(8) ± 0.0240.213(5) ± 0.0110.230(6) ± 0.0100.432(10) ± 0.0140.184(3) ± 0.0110.241(7) ± 0.0150.185(4) ± 0.0110.158(2) ± 0.0070.153(1) ± 0.007 •
genbase0.007(6) ± 0.0110.007(5) ± 0.0110.005(3) ± 0.0040.007(4) ± 0.0080.008(7) ± 0.0110.016(9) ± 0.0060.026(10) ± 0.0100.010(8) ± 0.0050.002(1) ± 0.002 •0.004(2) ± 0.003
LLOG0.186(4) ± 0.0020.283(7) ± 0.0230.481(9) ± 0.0600.536(10) ± 0.0200.368(8) ± 0.0280.229(6) ± 0.0050.186(3) ± 0.0020.186(1) ± 0.002 •0.186(2) ± 0.0020.187(5) ± 0.002
enron0.225(5) ± 0.0090.231(6) ± 0.0130.243(7) ± 0.0100.301(10) ± 0.0100.260(9) ± 0.0110.260(8) ± 0.0180.165(4) ± 0.0050.130(1) ± 0.003 •0.135(3) ± 0.0040.135(2) ± 0.003
scene0.321(9) ± 0.0180.302(8) ± 0.0210.115(5) ± 0.0070.138(7) ± 0.0040.360(10) ± 0.0180.102(4) ± 0.0050.121(6) ± 0.0040.078(3) ± 0.0040.076(2) ± 0.0040.071(1) ± 0.003 •
yeast0.364(8) ± 0.0180.409(9) ± 0.0110.225(5) ± 0.0040.245(7) ± 0.0030.417(10) ± 0.0080.191(4) ± 0.0030.234(6) ± 0.0040.175(3) ± 0.0030.169(2) ± 0.0030.168(1) ± 0.003 •
Slashdot0.049(5) ± 0.0040.047(4) ± 0.0040.155(10) ± 0.0090.153(9) ± 0.0080.081(7) ± 0.0090.084(8) ± 0.0110.066(6) ± 0.0080.042(1) ± 0.004 •0.043(2) ± 0.0040.047(3) ± 0.005
corel5k0.227(4) ± 0.0040.256(6) ± 0.0030.848(9) ± 0.0170.879(10) ± 0.0060.293(7) ± 0.0060.452(8) ± 0.0110.228(5) ± 0.0030.199(1) ± 0.003 •0.221(3) ± 0.0060.206(2) ± 0.004
rcv1subset10.205(6) ± 0.0060.265(8) ± 0.0090.315(9) ± 0.0090.457(10) ± 0.0110.254(7) ± 0.0080.143(5) ± 0.0050.120(4) ± 0.0030.077(1) ± 0.002 •0.087(2) ± 0.0030.118(3) ± 0.008
rcv1subset20.182(6) ± 0.0090.257(8) ± 0.0110.326(9) ± 0.0080.479(10) ± 0.0100.231(7) ± 0.0120.141(5) ± 0.0050.112(3) ± 0.0020.072(1) ± 0.003 •0.085(2) ± 0.0030.118(4) ± 0.009
rcv1subset30.188(6) ± 0.0080.277(8) ± 0.0080.314(9) ± 0.0070.481(10) ± 0.0100.231(7) ± 0.0100.147(5) ± 0.0040.115(3) ± 0.0010.075(1) ± 0.002 •0.085(2) ± 0.0020.122(4) ± 0.005
rcv1subset40.163(6) ± 0.0080.250(8) ± 0.0110.284(9) ± 0.0070.469(10) ± 0.0080.206(7) ± 0.0070.125(5) ± 0.0060.100(3) ± 0.0010.064(1) ± 0.002 •0.077(2) ± 0.0020.100(4) ± 0.009
rcv1subset50.177(6) ± 0.0110.239(8) ± 0.0120.322(9) ± 0.0100.461(10) ± 0.0110.225(7) ± 0.0070.150(5) ± 0.0060.115(3) ± 0.0020.072(1) ± 0.001 •0.084(2) ± 0.0020.117(4) ± 0.009
bibtex0.202(5) ± 0.0060.204(6) ± 0.0060.566(9) ± 0.0050.638(10) ± 0.0090.257(7) ± 0.0090.282(8) ± 0.0060.197(4) ± 0.0050.122(1) ± 0.003 •0.182(3) ± 0.0060.147(2) ± 0.004
Arts0.189(5) ± 0.0070.186(4) ± 0.0040.330(9) ± 0.0110.545(10) ± 0.0090.285(8) ± 0.0220.206(6) ± 0.0040.266(7) ± 0.0040.136(3) ± 0.0030.135(2) ± 0.0030.129(1) ± 0.003 •
Health0.119(5) ± 0.0050.118(4) ± 0.0070.209(9) ± 0.0070.285(10) ± 0.0060.185(8) ± 0.0110.125(6) ± 0.0040.146(7) ± 0.0050.076(3) ± 0.0020.076(2) ± 0.0020.070(1) ± 0.002 •
Business0.066(5) ± 0.0030.076(6) ± 0.0020.174(9) ± 0.0060.198(10) ± 0.0040.098(8) ± 0.0060.081(7) ± 0.0030.066(4) ± 0.0010.037(2) ± 0.0010.038(3) ± 0.0010.037(1) ± 0.001 •
Education0.167(7) ± 0.0040.156(5) ± 0.0040.261(8) ± 0.0060.441(10) ± 0.0070.265(9) ± 0.0220.156(4) ± 0.0030.162(6) ± 0.0020.093(2) ± 0.0020.094(3) ± 0.0020.089(1) ± 0.002 •
Computers0.146(5) ± 0.0030.148(6) ± 0.0050.267(9) ± 0.0050.378(10) ± 0.0050.214(8) ± 0.0130.166(7) ± 0.0030.144(4) ± 0.0040.089(2) ± 0.0020.090(3) ± 0.0020.086(1) ± 0.002 •
Entertainment0.169(4) ± 0.0060.181(6) ± 0.0040.227(8) ± 0.0090.387(10) ± 0.0070.244(9) ± 0.0130.174(5) ± 0.0030.224(7) ± 0.0030.122(3) ± 0.0020.121(2) ± 0.0020.111(1) ± 0.002 •
Recreation0.164(4) ± 0.0060.172(5) ± 0.0040.283(9) ± 0.0080.499(10) ± 0.0090.255(7) ± 0.0200.197(6) ± 0.0060.277(8) ± 0.0050.127(3) ± 0.0030.126(2) ± 0.0040.115(1) ± 0.004 •
Society0.168(5) ± 0.0040.152(4) ± 0.0030.326(9) ± 0.0080.448(10) ± 0.0070.248(8) ± 0.0170.189(6) ± 0.0040.216(7) ± 0.0040.112(2) ± 0.0020.112(3) ± 0.0020.105(1) ± 0.002 •
eurlex-dc-l0.150(5) ± 0.0030.156(6) ± 0.0030.455(9) ± 0.0080.547(10) ± 0.0050.188(7) ± 0.0050.206(8) ± 0.0060.118(3) ± 0.0030.074(1) ± 0.002 •0.119(4) ± 0.0030.078(2) ± 0.005
eurlex-sm0.127(5) ± 0.0030.130(6) ± 0.0030.307(9) ± 0.0030.372(10) ± 0.0030.169(8) ± 0.0060.138(7) ± 0.0030.084(4) ± 0.0010.053(2) ± 0.0010.064(3) ± 0.0010.050(1) ± 0.002 •
tmc2007-5000.148(9) ± 0.0040.145(8) ± 0.0030.121(6) ± 0.0030.144(7) ± 0.0020.179(10) ± 0.0040.092(4) ± 0.0010.105(5) ± 0.0010.059(3) ± 0.0010.057(2) ± 0.0010.044(1) ± 0.001 •
mediamill0.181(6) ± 0.0050.224(8) ± 0.0040.187(7) ± 0.0020.233(9) ± 0.0020.240(10) ± 0.0040.092(5) ± 0.0010.059(4) ± 0.0000.043(2) ± 0.0000.044(3) ± 0.0000.038(1) ± 0.001 •
average rank5.866.577.799.048.215.755.112.182.611.89
win/tie/loss108/18/12690/19/14357/9/18620/10/22248/10/194115/5/132127/12/113218/8/26203/11/38214/18/20
Table A4. The performance of nine compared MLC algorithms and MLHiKNN on F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A4. The performance of nine compared MLC algorithms and MLHiKNN on F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Macro
BRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
birds0.368(4) ± 0.0400.367(5) ± 0.0300.385(2) ± 0.0300.409(1) ± 0.035 •0.382(3) ± 0.0170.034(10) ± 0.0190.157(8) ± 0.0270.117(9) ± 0.0290.242(7) ± 0.0240.258(6) ± 0.069
CAL5000.170(5) ± 0.0070.195(3) ± 0.0060.141(6) ± 0.0060.191(4) ± 0.0050.215(2) ± 0.0030.083(10) ± 0.0020.103(7) ± 0.0030.084(8) ± 0.0020.229(1) ± 0.007 •0.083(9) ± 0.003
emotions0.570(8) ± 0.0110.551(10) ± 0.0230.639(2) ± 0.0130.614(4) ± 0.0130.555(9) ± 0.0180.608(5) ± 0.0110.602(6) ± 0.0230.595(7) ± 0.0170.638(3) ± 0.0140.651(1) ± 0.013 •
genbase0.980(5) ± 0.0120.977(6) ± 0.0180.983(3) ± 0.0070.984(2) ± 0.0120.984(1) ± 0.014 •0.639(10) ± 0.0580.913(8) ± 0.0300.820(9) ± 0.0800.980(4) ± 0.0090.954(7) ± 0.017
LLOG0.117(2) ± 0.0050.042(10) ± 0.0060.049(9) ± 0.0100.094(8) ± 0.0050.097(7) ± 0.0100.141(1) ± 0.012 •0.116(4) ± 0.0050.113(5) ± 0.0030.116(3) ± 0.0050.097(6) ± 0.016
enron0.215(5) ± 0.0070.218(2) ± 0.0060.217(4) ± 0.0070.224(1) ± 0.005 •0.217(3) ± 0.0110.067(10) ± 0.0080.131(8) ± 0.0070.103(9) ± 0.0100.170(6) ± 0.0100.131(7) ± 0.009
scene0.622(8) ± 0.0090.610(9) ± 0.0100.723(5) ± 0.0080.688(6) ± 0.0090.602(10) ± 0.0130.643(7) ± 0.0150.726(3) ± 0.0070.726(4) ± 0.0070.730(2) ± 0.0080.749(1) ± 0.008 •
yeast0.382(5) ± 0.0150.381(7) ± 0.0060.389(2) ± 0.0100.388(3) ± 0.0050.382(6) ± 0.0060.347(10) ± 0.0060.385(4) ± 0.0100.352(9) ± 0.0100.370(8) ± 0.0060.420(1) ± 0.009 •
Slashdot0.135(6) ± 0.0330.132(7) ± 0.0240.129(8) ± 0.0220.177(2) ± 0.0240.172(3) ± 0.0230.081(10) ± 0.0020.149(5) ± 0.0360.120(9) ± 0.0210.177(1) ± 0.023 •0.163(4) ± 0.025
corel5k0.029(4) ± 0.0030.029(6) ± 0.0040.008(8) ± 0.0020.029(5) ± 0.0020.031(3) ± 0.0030.001(10) ± 0.0010.038(2) ± 0.0060.003(9) ± 0.0020.055(1) ± 0.004 •0.026(7) ± 0.004
rcv1subset10.243(3) ± 0.0080.242(4) ± 0.0070.216(6) ± 0.0110.246(2) ± 0.0090.238(5) ± 0.0070.074(10) ± 0.0060.181(7) ± 0.0100.097(9) ± 0.0040.255(1) ± 0.008 •0.152(8) ± 0.014
rcv1subset20.231(3) ± 0.0080.237(2) ± 0.0130.197(6) ± 0.0080.230(4) ± 0.0070.227(5) ± 0.0050.067(10) ± 0.0030.173(7) ± 0.0110.081(9) ± 0.0030.246(1) ± 0.007 •0.133(8) ± 0.015
rcv1subset30.217(3) ± 0.0070.221(2) ± 0.0080.191(6) ± 0.0070.214(5) ± 0.0090.215(4) ± 0.0070.060(10) ± 0.0010.164(7) ± 0.0060.076(9) ± 0.0060.239(1) ± 0.008 •0.120(8) ± 0.010
rcv1subset40.231(3) ± 0.0140.244(2) ± 0.0080.197(6) ± 0.0070.228(4) ± 0.0110.227(5) ± 0.0100.081(10) ± 0.0050.178(7) ± 0.0080.091(9) ± 0.0040.253(1) ± 0.007 •0.149(8) ± 0.013
rcv1subset50.228(2) ± 0.0090.226(3) ± 0.0070.192(6) ± 0.0070.223(4) ± 0.0080.221(5) ± 0.0090.069(10) ± 0.0030.163(7) ± 0.0060.082(9) ± 0.0080.243(1) ± 0.008 •0.134(8) ± 0.017
bibtex0.213(1) ± 0.007 •0.202(4) ± 0.0050.191(5) ± 0.0060.204(3) ± 0.0080.204(2) ± 0.0050.060(9) ± 0.0020.148(8) ± 0.0060.052(10) ± 0.0030.173(7) ± 0.0030.187(6) ± 0.006
Arts0.267(2) ± 0.0070.256(6) ± 0.0080.257(5) ± 0.0110.274(1) ± 0.008 •0.266(3) ± 0.0110.140(10) ± 0.0100.223(7) ± 0.0130.182(9) ± 0.0140.218(8) ± 0.0190.263(4) ± 0.015
Health0.445(4) ± 0.0150.446(3) ± 0.0120.450(2) ± 0.0100.459(1) ± 0.011 •0.435(5) ± 0.0130.269(10) ± 0.0160.360(8) ± 0.0190.338(9) ± 0.0170.360(7) ± 0.0160.404(6) ± 0.011
Business0.272(1) ± 0.020 •0.245(5) ± 0.0140.237(8) ± 0.0120.269(2) ± 0.0190.265(4) ± 0.0140.141(10) ± 0.0080.239(7) ± 0.0100.194(9) ± 0.0090.244(6) ± 0.0140.267(3) ± 0.005
Education0.260(4) ± 0.0120.264(3) ± 0.0120.258(5) ± 0.0090.269(2) ± 0.0100.257(6) ± 0.0100.145(10) ± 0.0080.220(8) ± 0.0090.179(9) ± 0.0150.235(7) ± 0.0120.278(1) ± 0.015 •
Computers0.303(2) ± 0.0150.289(5) ± 0.0130.275(6) ± 0.0150.306(1) ± 0.014 •0.296(3) ± 0.0140.142(10) ± 0.0110.255(8) ± 0.0140.211(9) ± 0.0110.257(7) ± 0.0110.294(4) ± 0.010
Entertainment0.390(5) ± 0.0130.378(6) ± 0.0140.400(2) ± 0.0120.404(1) ± 0.008 •0.393(3) ± 0.0120.257(10) ± 0.0080.347(7) ± 0.0120.304(9) ± 0.0120.327(8) ± 0.0120.390(4) ± 0.008
Recreation0.369(4) ± 0.0100.361(6) ± 0.0100.373(3) ± 0.0150.385(2) ± 0.0110.367(5) ± 0.0100.240(10) ± 0.0220.352(7) ± 0.0140.316(9) ± 0.0130.352(8) ± 0.0150.390(1) ± 0.016 •
Society0.261(4) ± 0.0100.261(5) ± 0.0100.250(6) ± 0.0080.271(2) ± 0.0060.265(3) ± 0.0090.158(10) ± 0.0080.246(7) ± 0.0100.220(9) ± 0.0080.243(8) ± 0.0120.291(1) ± 0.012 •
eurlex-dc-l0.258(2) ± 0.0080.258(3) ± 0.0080.237(6) ± 0.0040.253(4) ± 0.0070.251(5) ± 0.0060.148(9) ± 0.0030.177(7) ± 0.0050.068(10) ± 0.0030.169(8) ± 0.0040.267(1) ± 0.010 •
eurlex-sm0.377(3) ± 0.0040.370(6) ± 0.0060.372(5) ± 0.0060.377(2) ± 0.0050.372(4) ± 0.0070.252(9) ± 0.0050.296(8) ± 0.0070.200(10) ± 0.0060.307(7) ± 0.0070.388(1) ± 0.009 •
tmc2007-5000.560(4) ± 0.0060.558(5) ± 0.0070.582(2) ± 0.0050.578(3) ± 0.0050.544(6) ± 0.0040.239(10) ± 0.0040.464(7) ± 0.0100.425(9) ± 0.0090.452(8) ± 0.0060.603(1) ± 0.007 •
mediamill0.172(4) ± 0.0060.156(5) ± 0.0040.128(8) ± 0.0050.196(2) ± 0.0040.180(3) ± 0.0050.085(10) ± 0.0020.137(7) ± 0.0060.088(9) ± 0.0020.155(6) ± 0.0050.293(1) ± 0.005 •
average rank3.795.005.072.894.399.296.648.684.864.39
win/tie/loss140/55/57124/45/83116/39/97175/41/36135/49/6819/5/22880/27/14533/11/208134/23/95143/27/82
Table A5. The performance of nine compared MLC algorithms and MLHiKNN on F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A5. The performance of nine compared MLC algorithms and MLHiKNN on F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Micro
BRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
birds0.452(3) ± 0.0440.448(4) ± 0.0350.497(1) ± 0.021 •0.481(2) ± 0.0240.438(5) ± 0.0210.074(10) ± 0.0420.230(8) ± 0.0500.208(9) ± 0.0390.327(7) ± 0.0270.358(6) ± 0.061
CAL5000.353(4) ± 0.0110.362(3) ± 0.0110.386(1) ± 0.007 •0.383(2) ± 0.0040.351(5) ± 0.0050.321(10) ± 0.0060.338(8) ± 0.0090.331(9) ± 0.0050.338(7) ± 0.0080.342(6) ± 0.014
emotions0.579(8) ± 0.0120.561(10) ± 0.0220.655(3) ± 0.0120.627(6) ± 0.0120.564(9) ± 0.0180.635(4) ± 0.0130.629(5) ± 0.0200.616(7) ± 0.0180.662(2) ± 0.0110.672(1) ± 0.011 •
genbase0.987(4) ± 0.0080.986(6) ± 0.0100.990(2) ± 0.0050.990(1) ± 0.007 •0.989(3) ± 0.0100.839(10) ± 0.0400.949(8) ± 0.0200.909(9) ± 0.0360.986(5) ± 0.0070.965(7) ± 0.014
LLOG0.429(2) ± 0.0130.100(10) ± 0.0350.145(9) ± 0.0610.367(7) ± 0.0160.371(6) ± 0.0330.466(1) ± 0.023 •0.427(3) ± 0.0130.421(5) ± 0.0080.427(4) ± 0.0120.363(8) ± 0.059
enron0.522(4) ± 0.0090.524(3) ± 0.0120.584(1) ± 0.007 •0.565(2) ± 0.0090.514(5) ± 0.0080.219(10) ± 0.0220.476(6) ± 0.0110.450(9) ± 0.0140.454(8) ± 0.0150.460(7) ± 0.013
scene0.610(8) ± 0.0090.594(9) ± 0.0120.715(5) ± 0.0090.677(6) ± 0.0090.591(10) ± 0.0120.649(7) ± 0.0110.720(3) ± 0.0070.720(4) ± 0.0080.725(2) ± 0.0080.743(1) ± 0.008 •
yeast0.573(8) ± 0.0110.539(9) ± 0.0100.637(3) ± 0.0060.609(7) ± 0.0020.538(10) ± 0.0050.628(6) ± 0.0040.635(4) ± 0.0060.628(5) ± 0.0060.641(2) ± 0.0030.643(1) ± 0.006 •
Slashdot0.845(5) ± 0.0070.846(4) ± 0.0070.846(3) ± 0.0070.849(1) ± 0.008 •0.849(2) ± 0.0080.843(7) ± 0.0070.843(6) ± 0.0110.842(10) ± 0.0080.843(9) ± 0.0060.843(8) ± 0.008
corel5k0.093(4) ± 0.0080.094(2) ± 0.0080.029(8) ± 0.0060.081(5) ± 0.0060.095(1) ± 0.007 •0.003(10) ± 0.0030.094(3) ± 0.0130.007(9) ± 0.0040.075(6) ± 0.0040.050(7) ± 0.011
rcv1subset10.369(3) ± 0.0090.357(6) ± 0.0040.365(5) ± 0.0090.387(1) ± 0.005 •0.366(4) ± 0.0060.203(10) ± 0.0070.345(7) ± 0.0110.261(9) ± 0.0070.374(2) ± 0.0070.296(8) ± 0.021
rcv1subset20.383(3) ± 0.0050.369(6) ± 0.0090.377(5) ± 0.0110.388(1) ± 0.007 •0.382(4) ± 0.0060.275(10) ± 0.0080.367(7) ± 0.0160.288(9) ± 0.0070.385(2) ± 0.0070.334(8) ± 0.012
rcv1subset30.383(3) ± 0.0120.371(6) ± 0.0090.372(5) ± 0.0070.387(1) ± 0.012 •0.381(4) ± 0.0060.258(10) ± 0.0070.362(7) ± 0.0130.279(9) ± 0.0090.386(2) ± 0.0060.333(8) ± 0.009
rcv1subset40.427(4) ± 0.0080.423(6) ± 0.0120.420(7) ± 0.0080.437(1) ± 0.006 •0.426(5) ± 0.0090.336(10) ± 0.0070.430(3) ± 0.0070.362(9) ± 0.0050.434(2) ± 0.0070.394(8) ± 0.019
rcv1subset5 0.403(2) ± 0.0040.377(6) ± 0.0070.384(5) ± 0.0080.411(1) ± 0.006 •0.398(3) ± 0.0080.284(10) ± 0.0100.369(7) ± 0.0090.303(9) ± 0.0100.397(4) ± 0.0050.348(8) ± 0.014
bibtex0.393(1) ± 0.008 •0.386(5) ± 0.0060.387(4) ± 0.0060.392(2) ± 0.0070.389(3) ± 0.0050.241(9) ± 0.0060.328(7) ± 0.0040.231(10) ± 0.0060.256(8) ± 0.0060.367(6) ± 0.006
Arts0.393(5) ± 0.0050.393(3) ± 0.0080.419(1) ± 0.007 •0.410(2) ± 0.0080.393(4) ± 0.0100.263(10) ± 0.0160.340(7) ± 0.0170.296(9) ± 0.0220.316(8) ± 0.0180.369(6) ± 0.014
Health0.611(4) ± 0.0060.619(3) ± 0.0060.654(1) ± 0.006 •0.634(2) ± 0.0060.608(5) ± 0.0080.491(10) ± 0.0120.531(7) ± 0.0170.519(9) ± 0.0160.530(8) ± 0.0130.574(6) ± 0.011
Business0.720(4) ± 0.0040.712(8) ± 0.0050.732(1) ± 0.004 •0.730(2) ± 0.0030.717(5) ± 0.0050.705(10) ± 0.0050.714(7) ± 0.0060.707(9) ± 0.0040.716(6) ± 0.0040.722(3) ± 0.004
Education0.421(4) ± 0.0070.426(3) ± 0.0050.460(1) ± 0.006 •0.440(2) ± 0.0030.414(5) ± 0.0050.288(10) ± 0.0120.357(7) ± 0.0130.307(9) ± 0.0190.321(8) ± 0.0140.378(6) ± 0.014
Computers0.514(3) ± 0.0040.502(6) ± 0.0070.551(1) ± 0.004 •0.532(2) ± 0.0030.504(5) ± 0.0040.468(10) ± 0.0040.490(7) ± 0.0080.478(9) ± 0.0130.487(8) ± 0.0040.511(4) ± 0.005
Entertainment0.534(4) ± 0.0060.507(5) ± 0.0090.572(1) ± 0.007 •0.562(2) ± 0.0060.534(3) ± 0.0060.400(10) ± 0.0090.472(7) ± 0.0140.441(9) ± 0.0110.453(8) ± 0.0110.505(6) ± 0.008
Recreation0.468(3) ± 0.0060.422(7) ± 0.0070.489(1) ± 0.005 •0.484(2) ± 0.0070.466(4) ± 0.0060.345(10) ± 0.0180.428(6) ± 0.0170.402(9) ± 0.0170.406(8) ± 0.0150.454(5) ± 0.012
Society0.444(6) ± 0.0050.488(2) ± 0.0050.494(1) ± 0.004 •0.470(3) ± 0.0020.447(5) ± 0.0030.385(10) ± 0.0140.433(7) ± 0.0100.415(9) ± 0.0200.432(8) ± 0.0090.468(4) ± 0.010
eurlex-dc-l0.483(4) ± 0.0050.484(3) ± 0.0060.494(1) ± 0.007 •0.489(2) ± 0.0040.478(6) ± 0.0050.390(8) ± 0.0030.401(7) ± 0.0060.288(9) ± 0.0060.278(10) ± 0.0030.479(5) ± 0.007
eurlex-sm0.599(3) ± 0.0030.594(5) ± 0.0040.624(1) ± 0.005 •0.616(2) ± 0.0040.596(4) ± 0.0040.504(9) ± 0.0060.533(7) ± 0.0060.481(10) ± 0.0060.510(8) ± 0.0060.589(6) ± 0.007
tmc2007-5000.668(4) ± 0.0030.665(5) ± 0.0030.716(2) ± 0.0030.707(3) ± 0.0020.657(6) ± 0.0030.577(10) ± 0.0030.634(9) ± 0.0060.638(8) ± 0.0020.640(7) ± 0.0040.726(1) ± 0.003 •
mediamill0.553(8) ± 0.0030.536(10) ± 0.0020.597(3) ± 0.0020.598(2) ± 0.0020.541(9) ± 0.0030.573(6) ± 0.0020.583(4) ± 0.0040.568(7) ± 0.0020.580(5) ± 0.0020.635(1) ± 0.001 •
average rank4.215.542.932.575.008.826.218.465.865.39
win/tie/loss140/41/71108/32/112184/21/47195/20/37121/40/9128/8/21690/30/13241/14/197101/39/112121/17/114
Table A6. The performance of nine compared MLC algorithms and MLHiKNN on Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A6. The performance of nine compared MLC algorithms and MLHiKNN on Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetHamming Loss
BRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
birds0.143(4) ± 0.0130.145(7) ± 0.0120.119(1) ± 0.007 •0.134(2) ± 0.0080.157(10) ± 0.0110.144(5) ± 0.0070.145(6) ± 0.0070.146(8) ± 0.0060.151(9) ± 0.0060.137(3) ± 0.008
CAL5000.228(7) ± 0.0030.244(8) ± 0.0050.198(5) ± 0.0020.227(6) ± 0.0020.263(9) ± 0.0030.187(2) ± 0.0020.192(4) ± 0.0010.189(3) ± 0.0010.300(10) ± 0.0030.187(1) ± 0.001 •
emotions0.260(8) ± 0.0100.270(9) ± 0.0110.209(4) ± 0.0070.225(7) ± 0.0080.271(10) ± 0.0110.198(2) ± 0.0070.211(5) ± 0.0060.216(6) ± 0.0070.199(3) ± 0.0060.190(1) ± 0.005 •
genbase0.002(4) ± 0.0010.002(5) ± 0.0020.002(2) ± 0.0010.002(1) ± 0.001 •0.002(3) ± 0.0020.026(10) ± 0.0070.009(8) ± 0.0030.014(9) ± 0.0050.002(6) ± 0.0010.006(7) ± 0.002
LLOG0.185(1) ± 0.004 •0.206(10) ± 0.0070.202(9) ± 0.0090.187(5) ± 0.0040.187(6) ± 0.0050.192(8) ± 0.0040.185(3) ± 0.0040.185(2) ± 0.0040.185(4) ± 0.0040.188(7) ± 0.006
enron0.080(5) ± 0.0010.080(7) ± 0.0020.070(1) ± 0.001 •0.074(2) ± 0.0010.083(9) ± 0.0020.090(10) ± 0.0010.080(6) ± 0.0010.078(4) ± 0.0010.082(8) ± 0.0020.075(3) ± 0.001
scene0.138(8) ± 0.0030.147(10) ± 0.0050.095(5) ± 0.0030.108(7) ± 0.0030.145(9) ± 0.0040.103(6) ± 0.0030.092(3) ± 0.0020.093(4) ± 0.0030.089(2) ± 0.0030.084(1) ± 0.003 •
yeast0.255(8) ± 0.0070.277(9) ± 0.0050.212(6) ± 0.0030.234(7) ± 0.0030.280(10) ± 0.0040.199(2) ± 0.0030.204(5) ± 0.0020.202(4) ± 0.0020.198(1) ± 0.002 •0.200(3) ± 0.003
Slashdot0.028(5) ± 0.0010.028(4) ± 0.0010.028(3) ± 0.0010.027(1) ± 0.001 •0.027(2) ± 0.0020.028(8) ± 0.0010.028(6) ± 0.0020.028(7) ± 0.0010.028(9) ± 0.0010.029(10) ± 0.001
corel5k0.022(6) ± 0.0000.022(7) ± 0.0000.021(4) ± 0.0000.021(5) ± 0.0000.022(8) ± 0.0000.021(3) ± 0.0000.022(9) ± 0.0000.021(2) ± 0.0000.038(10) ± 0.0010.021(1) ± 0.000 •
rcv1subset10.036(7) ± 0.0000.039(10) ± 0.0010.034(4) ± 0.0000.036(6) ± 0.0000.037(8) ± 0.0000.034(5) ± 0.0000.033(3) ± 0.0000.033(2) ± 0.0000.037(9) ± 0.0010.032(1) ± 0.000 •
rcv1subset20.031(7) ± 0.0000.032(9) ± 0.0010.028(5) ± 0.0000.030(6) ± 0.0000.031(8) ± 0.0000.028(3) ± 0.0000.028(4) ± 0.0000.028(2) ± 0.0000.033(10) ± 0.0010.027(1) ± 0.000 •
rcv1subset30.031(7) ± 0.0010.033(10) ± 0.0010.028(5) ± 0.0000.030(6) ± 0.0010.031(8) ± 0.0000.028(3) ± 0.0000.028(4) ± 0.0000.028(2) ± 0.0000.032(9) ± 0.0010.027(1) ± 0.000 •
rcv1subset40.028(7) ± 0.0000.030(10) ± 0.0010.026(5) ± 0.0000.027(6) ± 0.0000.028(8) ± 0.0000.025(4) ± 0.0000.025(3) ± 0.0000.025(2) ± 0.0000.030(9) ± 0.0000.024(1) ± 0.000 •
rcv1subset50.030(7) ± 0.0000.033(10) ± 0.0010.028(5) ± 0.0000.029(6) ± 0.0000.030(8) ± 0.0000.027(3) ± 0.0000.028(4) ± 0.0000.027(2) ± 0.0000.032(9) ± 0.0000.027(1) ± 0.000 •
bibtex0.013(4) ± 0.0000.013(7) ± 0.0000.013(1) ± 0.000 •0.013(3) ± 0.0000.013(6) ± 0.0000.013(8) ± 0.0000.013(5) ± 0.0000.013(9) ± 0.0000.026(10) ± 0.0010.013(2) ± 0.000
Arts0.061(7) ± 0.0010.073(10) ± 0.0010.058(1) ± 0.001 •0.059(2) ± 0.0010.063(9) ± 0.0010.061(6) ± 0.0010.061(5) ± 0.0000.061(8) ± 0.0000.060(4) ± 0.0000.059(3) ± 0.001
Health0.053(3) ± 0.0000.054(5) ± 0.0010.049(1) ± 0.001 •0.050(2) ± 0.0010.054(4) ± 0.0010.061(10) ± 0.0010.059(9) ± 0.0010.059(8) ± 0.0010.057(7) ± 0.0010.055(6) ± 0.001
Business0.032(8) ± 0.0000.033(10) ± 0.0010.030(1) ± 0.001 •0.031(2) ± 0.0000.032(9) ± 0.0010.032(7) ± 0.0010.031(5) ± 0.0010.032(6) ± 0.0010.031(4) ± 0.0010.031(3) ± 0.000
Education0.049(8) ± 0.0000.055(10) ± 0.0010.045(1) ± 0.001 •0.046(2) ± 0.0000.050(9) ± 0.0010.047(6) ± 0.0010.047(7) ± 0.0010.047(5) ± 0.0010.047(4) ± 0.0010.046(3) ± 0.001
Computers0.045(8) ± 0.0000.049(10) ± 0.0010.042(1) ± 0.000 •0.042(2) ± 0.0000.046(9) ± 0.0000.045(7) ± 0.0000.044(5) ± 0.0010.044(6) ± 0.0000.044(4) ± 0.0000.043(3) ± 0.000
Entertainment0.062(4) ± 0.0010.077(10) ± 0.0010.058(1) ± 0.002 •0.058(2) ± 0.0010.064(6) ± 0.0010.064(9) ± 0.0010.064(7) ± 0.0010.064(8) ± 0.0010.063(5) ± 0.0010.061(3) ± 0.001
Recreation0.052(5) ± 0.0010.068(10) ± 0.0010.049(1) ± 0.001 •0.050(2) ± 0.0010.053(9) ± 0.0010.053(8) ± 0.0010.052(6) ± 0.0010.052(7) ± 0.0010.051(4) ± 0.0010.050(3) ± 0.001
Society0.054(8) ± 0.0000.056(10) ± 0.0010.052(3) ± 0.0000.052(4) ± 0.0000.055(9) ± 0.0010.054(7) ± 0.0010.053(5) ± 0.0010.053(6) ± 0.0010.052(2) ± 0.0010.051(1) ± 0.001 •
eurlex-dc-l0.005(4) ± 0.0000.005(6) ± 0.0000.004(1) ± 0.000 •0.005(3) ± 0.0000.005(5) ± 0.0000.005(8) ± 0.0000.005(7) ± 0.0000.005(9) ± 0.0000.010(10) ± 0.0000.004(2) ± 0.000
eurlex-sm0.011(4) ± 0.0000.011(6) ± 0.0000.010(1) ± 0.000 •0.010(2) ± 0.0000.011(5) ± 0.0000.012(8) ± 0.0000.012(7) ± 0.0000.012(9) ± 0.0000.013(10) ± 0.0000.011(3) ± 0.000
tmc2007-5000.064(4) ± 0.0000.065(5) ± 0.0010.055(2) ± 0.0010.056(3) ± 0.0000.066(8) ± 0.0000.070(10) ± 0.0010.067(9) ± 0.0000.066(7) ± 0.0000.065(6) ± 0.0010.051(1) ± 0.000 •
mediamill0.035(8) ± 0.0000.037(9) ± 0.0000.030(2) ± 0.0000.031(7) ± 0.0000.038(10) ± 0.0000.031(3) ± 0.0000.031(5) ± 0.0000.031(6) ± 0.0000.031(4) ± 0.0000.028(1) ± 0.000 •
average rank5.938.322.893.897.646.115.545.466.502.71
win/tie/loss92/32/12834/28/190185/21/46156/20/7655/25/17296/28/128110/34/108113/30/10990/25/137197/21/34

Appendix C. Experimental Results of Compared Hubness-Reduced MLC Approaches

Table A7. The performance of the algorithms for hubness reduction technique comparisons on AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A7. The performance of the algorithms for hubness reduction technique comparisons on AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Micro
BRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
birds0.736(4) ± 0.0140.724(5) ± 0.0210.754(1) ± 0.011 •0.738(3) ± 0.0090.659(6) ± 0.0130.565(7) ± 0.0290.545(8) ± 0.0140.532(9) ± 0.0240.746(2) ± 0.019
CAL5000.749(2) ± 0.0030.746(5) ± 0.0030.749(4) ± 0.0020.749(3) ± 0.0030.713(6) ± 0.0030.703(7) ± 0.0030.703(8) ± 0.0030.702(9) ± 0.0030.757(1) ± 0.004 •
emotions0.852(4) ± 0.0050.841(5) ± 0.0080.861(2) ± 0.0070.859(3) ± 0.0070.747(6) ± 0.0110.731(9) ± 0.0150.734(7) ± 0.0130.731(8) ± 0.0130.862(1) ± 0.005 •
genbase0.993(4) ± 0.0040.993(3) ± 0.0040.994(2) ± 0.0040.986(5) ± 0.0040.970(6) ± 0.0120.832(7) ± 0.0430.831(8) ± 0.0530.721(9) ± 0.0420.996(1) ± 0.004 •
LLOG0.788(2) ± 0.0050.782(3) ± 0.0060.781(5) ± 0.0040.782(3) ± 0.0060.791(1) ± 0.002 •0.592(9) ± 0.0400.604(8) ± 0.0410.607(7) ± 0.0370.767(6) ± 0.002
enron0.801(5) ± 0.0120.828(3) ± 0.0050.830(2) ± 0.0050.801(6) ± 0.0070.825(4) ± 0.0040.714(7) ± 0.0050.711(8) ± 0.0050.708(9) ± 0.0060.860(1) ± 0.002 •
scene0.934(5) ± 0.0030.943(4) ± 0.0020.948(2) ± 0.0020.948(3) ± 0.0030.885(6) ± 0.0040.868(9) ± 0.0090.871(7) ± 0.0090.870(8) ± 0.0090.950(1) ± 0.002 •
yeast0.837(4) ± 0.0030.827(5) ± 0.0040.842(2) ± 0.0030.842(3) ± 0.0030.766(6) ± 0.0040.734(9) ± 0.0040.740(8) ± 0.0050.740(7) ± 0.0050.842(1) ± 0.004 •
Slashdot0.932(5) ± 0.0080.935(3) ± 0.0080.938(2) ± 0.0040.935(4) ± 0.0080.912(6) ± 0.0090.880(9) ± 0.0090.887(8) ± 0.0120.890(7) ± 0.0090.945(1) ± 0.004 •
corel5k0.713(6) ± 0.0080.747(5) ± 0.0080.752(4) ± 0.0080.752(3) ± 0.0090.772(2) ± 0.0040.668(8) ± 0.0050.668(7) ± 0.0050.666(9) ± 0.0040.796(1) ± 0.004 •
rcv1subset10.894(4) ± 0.0040.911(2) ± 0.0030.912(1) ± 0.003 •0.899(3) ± 0.0040.866(6) ± 0.0030.689(7) ± 0.0060.687(8) ± 0.0060.656(9) ± 0.0080.874(5) ± 0.009
rcv1subset20.895(4) ± 0.0030.912(1) ± 0.003 •0.909(2) ± 0.0040.898(3) ± 0.0030.874(5) ± 0.0030.672(8) ± 0.0120.683(7) ± 0.0150.636(9) ± 0.0120.865(6) ± 0.010
rcv1subset30.893(3) ± 0.0030.908(2) ± 0.0030.909(1) ± 0.006 •0.893(4) ± 0.0040.871(5) ± 0.0030.671(7) ± 0.0060.670(8) ± 0.0150.639(9) ± 0.0120.861(6) ± 0.007
rcv1subset40.906(4) ± 0.0040.918(1) ± 0.003 •0.917(2) ± 0.0040.909(3) ± 0.0040.884(5) ± 0.0020.703(8) ± 0.0060.707(7) ± 0.0060.671(9) ± 0.0070.884(6) ± 0.010
rcv1subset5 0.891(4) ± 0.0030.907(1) ± 0.002 •0.904(2) ± 0.0020.893(3) ± 0.0030.871(5) ± 0.0020.665(8) ± 0.0060.690(7) ± 0.0060.635(9) ± 0.0070.868(6) ± 0.009
bibtex0.838(5) ± 0.0030.853(4) ± 0.0030.864(1) ± 0.003 •0.853(3) ± 0.0030.802(6) ± 0.0040.489(9) ± 0.0070.524(8) ± 0.0060.541(7) ± 0.0050.863(2) ± 0.006
Arts0.818(4) ± 0.0030.822(2) ± 0.0040.819(3) ± 0.0030.810(5) ± 0.0040.713(6) ± 0.0040.592(8) ± 0.0060.603(7) ± 0.0040.511(9) ± 0.0150.851(1) ± 0.003 •
Health0.893(4) ± 0.0040.896(2) ± 0.0030.895(3) ± 0.0020.884(5) ± 0.0040.827(6) ± 0.0040.717(8) ± 0.0070.724(7) ± 0.0030.546(9) ± 0.0230.919(1) ± 0.002 •
Business0.927(4) ± 0.0020.929(3) ± 0.0020.930(2) ± 0.0020.925(5) ± 0.0020.908(6) ± 0.0020.824(8) ± 0.0050.825(7) ± 0.0050.797(9) ± 0.0070.947(1) ± 0.001 •
Education0.876(3) ± 0.0020.876(2) ± 0.0020.876(4) ± 0.0020.868(5) ± 0.0020.822(6) ± 0.0030.744(9) ± 0.0040.747(7) ± 0.0030.746(8) ± 0.0040.906(1) ± 0.002 •
Computers0.862(4) ± 0.0030.865(2) ± 0.0020.865(3) ± 0.0030.853(5) ± 0.0030.820(6) ± 0.0030.695(8) ± 0.0050.700(7) ± 0.0040.582(9) ± 0.0050.896(1) ± 0.003 •
Entertainment0.863(3) ± 0.0020.865(2) ± 0.0020.861(4) ± 0.0030.854(5) ± 0.0030.754(6) ± 0.0040.637(8) ± 0.0050.647(7) ± 0.0060.481(9) ± 0.0110.891(1) ± 0.002 •
Recreation0.839(4) ± 0.0050.840(2) ± 0.0050.839(3) ± 0.0040.824(5) ± 0.0040.697(6) ± 0.0040.534(8) ± 0.0100.541(7) ± 0.0070.443(9) ± 0.0130.874(1) ± 0.004 •
Society0.829(4) ± 0.0030.830(2) ± 0.0030.830(3) ± 0.0020.823(5) ± 0.0040.753(6) ± 0.0040.646(7) ± 0.0040.645(8) ± 0.0040.553(9) ± 0.0120.869(1) ± 0.002 •
eurlex-dc-l0.882(4) ± 0.0040.884(2) ± 0.0040.883(3) ± 0.0030.872(6) ± 0.0030.875(5) ± 0.0030.603(7) ± 0.0060.603(8) ± 0.0050.267(9) ± 0.0080.919(1) ± 0.005 •
eurlex-sm0.919(4) ± 0.0020.921(2) ± 0.0020.920(3) ± 0.0020.914(6) ± 0.0020.917(5) ± 0.0020.704(7) ± 0.0050.699(8) ± 0.0050.692(9) ± 0.0040.952(1) ± 0.002 •
tmc2007-5000.928(4) ± 0.0010.939(2) ± 0.0010.937(3) ± 0.0010.925(5) ± 0.0010.891(6) ± 0.0020.783(7) ± 0.0040.775(8) ± 0.0040.733(9) ± 0.0110.958(1) ± 0.001 •
mediamill0.930(6) ± 0.0010.932(3) ± 0.0010.933(2) ± 0.0010.932(4) ± 0.0010.930(5) ± 0.0000.862(7) ± 0.0010.862(8) ± 0.0010.861(9) ± 0.0010.957(1) ± 0.001 •
average rank4.042.802.544.165.367.867.548.612.11
win/tie/loss132/16/76162/16/46171/24/29129/19/7699/3/12225/18/18129/21/1744/13/207187/10/27
Table A8. The performance of the algorithms for hubness reduction technique comparisons on Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A8. The performance of the algorithms for hubness reduction technique comparisons on Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetRanking Loss
BRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
birds0.317(4) ± 0.0200.325(5) ± 0.0220.292(2) ± 0.0160.310(3) ± 0.0140.345(6) ± 0.0180.441(7) ± 0.0300.462(8) ± 0.0170.469(9) ± 0.0260.251(1) ± 0.020 •
CAL5000.282(2) ± 0.0030.285(6) ± 0.0040.282(4) ± 0.0030.282(3) ± 0.0030.284(5) ± 0.0030.293(7) ± 0.0030.294(8) ± 0.0020.294(9) ± 0.0030.240(1) ± 0.005 •
emotions0.184(4) ± 0.0110.201(5) ± 0.0140.175(2) ± 0.0090.176(3) ± 0.0090.241(6) ± 0.0150.256(9) ± 0.0180.252(7) ± 0.0140.255(8) ± 0.0180.153(1) ± 0.007 •
genbase0.016(4) ± 0.0060.015(2) ± 0.0050.015(3) ± 0.0050.028(6) ± 0.0060.026(5) ± 0.0100.155(7) ± 0.0450.155(8) ± 0.0530.245(9) ± 0.0440.004(1) ± 0.003 •
LLOG0.229(3) ± 0.0050.235(4) ± 0.0080.235(6) ± 0.0090.235(4) ± 0.0080.186(1) ± 0.002 •0.388(9) ± 0.0450.336(7) ± 0.0270.343(8) ± 0.0320.187(2) ± 0.002
enron0.260(6) ± 0.0180.218(4) ± 0.0070.217(3) ± 0.0080.255(5) ± 0.0090.165(2) ± 0.0050.285(7) ± 0.0030.287(8) ± 0.0060.290(9) ± 0.0060.135(1) ± 0.003 •
scene0.102(5) ± 0.0050.091(4) ± 0.0040.085(2) ± 0.0040.085(3) ± 0.0050.121(6) ± 0.0040.138(8) ± 0.0110.138(9) ± 0.0110.138(7) ± 0.0100.071(1) ± 0.003 •
yeast0.191(4) ± 0.0030.198(5) ± 0.0040.186(3) ± 0.0020.186(2) ± 0.0030.234(6) ± 0.0040.262(8) ± 0.0050.261(7) ± 0.0050.262(9) ± 0.0050.168(1) ± 0.003 •
Slashdot0.084(6) ± 0.0110.076(4) ± 0.0130.071(3) ± 0.0070.077(5) ± 0.0130.066(2) ± 0.0080.091(9) ± 0.0090.087(7) ± 0.0100.087(8) ± 0.0060.047(1) ± 0.005 •
corel5k0.452(9) ± 0.0110.408(8) ± 0.0150.400(7) ± 0.0140.399(6) ± 0.0170.228(2) ± 0.0030.331(4) ± 0.0050.330(3) ± 0.0050.332(5) ± 0.0040.206(1) ± 0.004 •
rcv1subset10.143(6) ± 0.0050.119(3) ± 0.0040.119(2) ± 0.0040.137(5) ± 0.0060.120(4) ± 0.0030.301(7) ± 0.0080.304(8) ± 0.0070.342(9) ± 0.0080.118(1) ± 0.008 •
rcv1subset20.141(6) ± 0.0050.117(2) ± 0.0040.122(4) ± 0.0060.136(5) ± 0.0040.112(1) ± 0.002 •0.314(8) ± 0.0150.302(7) ± 0.0180.354(9) ± 0.0160.118(3) ± 0.009
rcv1subset30.147(6) ± 0.0040.124(4) ± 0.0040.124(3) ± 0.0080.145(5) ± 0.0040.115(1) ± 0.001 •0.313(7) ± 0.0070.314(8) ± 0.0120.347(9) ± 0.0120.122(2) ± 0.005
rcv1subset40.125(6) ± 0.0060.108(3) ± 0.0050.110(4) ± 0.0060.121(5) ± 0.0060.100(1) ± 0.001 •0.267(8) ± 0.0070.262(7) ± 0.0070.299(9) ± 0.0070.100(2) ± 0.009
rcv1subset50.150(6) ± 0.0060.127(3) ± 0.0040.131(4) ± 0.0040.146(5) ± 0.0050.115(1) ± 0.002 •0.317(8) ± 0.0060.291(7) ± 0.0060.349(9) ± 0.0070.117(2) ± 0.009
bibtex0.282(6) ± 0.0060.256(4) ± 0.0070.240(3) ± 0.0070.257(5) ± 0.0070.197(2) ± 0.0050.493(9) ± 0.0070.456(8) ± 0.0060.447(7) ± 0.0060.147(1) ± 0.004 •
Arts0.206(4) ± 0.0040.202(2) ± 0.0040.204(3) ± 0.0040.216(5) ± 0.0040.266(6) ± 0.0040.407(8) ± 0.0080.396(7) ± 0.0050.506(9) ± 0.0170.129(1) ± 0.003 •
Health0.125(4) ± 0.0040.121(2) ± 0.0030.122(3) ± 0.0020.137(5) ± 0.0040.146(6) ± 0.0050.249(8) ± 0.0070.242(7) ± 0.0030.460(9) ± 0.0260.070(1) ± 0.002 •
Business0.081(5) ± 0.0030.079(4) ± 0.0030.076(3) ± 0.0020.084(6) ± 0.0040.066(2) ± 0.0010.126(8) ± 0.0040.123(7) ± 0.0040.146(9) ± 0.0050.037(1) ± 0.001 •
Education0.156(3) ± 0.0030.156(2) ± 0.0020.156(4) ± 0.0020.167(6) ± 0.0030.162(5) ± 0.0020.251(9) ± 0.0040.249(8) ± 0.0030.249(7) ± 0.0040.089(1) ± 0.002 •
Computers0.166(5) ± 0.0030.160(3) ± 0.0030.162(4) ± 0.0030.174(6) ± 0.0050.144(2) ± 0.0040.253(8) ± 0.0050.245(7) ± 0.0040.390(9) ± 0.0060.086(1) ± 0.002 •
Entertainment0.174(3) ± 0.0030.171(2) ± 0.0020.175(4) ± 0.0040.183(5) ± 0.0050.224(6) ± 0.0030.345(8) ± 0.0050.338(7) ± 0.0050.530(9) ± 0.0110.111(1) ± 0.002 •
Recreation0.197(4) ± 0.0060.194(2) ± 0.0060.194(3) ± 0.0050.212(5) ± 0.0050.277(6) ± 0.0050.462(8) ± 0.0130.453(7) ± 0.0080.577(9) ± 0.0140.115(1) ± 0.004 •
Society0.189(4) ± 0.0040.187(2) ± 0.0040.188(3) ± 0.0040.196(5) ± 0.0050.216(6) ± 0.0040.335(8) ± 0.0040.334(7) ± 0.0060.457(9) ± 0.0130.105(1) ± 0.002 •
eurlex-dc-l0.206(5) ± 0.0060.204(3) ± 0.0060.206(4) ± 0.0050.224(6) ± 0.0050.118(2) ± 0.0030.377(7) ± 0.0050.378(8) ± 0.0040.735(9) ± 0.0090.078(1) ± 0.005 •
eurlex-sm0.138(5) ± 0.0030.136(3) ± 0.0030.137(4) ± 0.0030.146(6) ± 0.0030.084(2) ± 0.0010.317(7) ± 0.0050.323(8) ± 0.0040.330(9) ± 0.0040.050(1) ± 0.002 •
tmc2007-5000.092(4) ± 0.0010.077(2) ± 0.0010.081(3) ± 0.0010.095(5) ± 0.0010.105(6) ± 0.0010.209(7) ± 0.0030.217(8) ± 0.0040.255(9) ± 0.0100.044(1) ± 0.001 •
mediamill0.092(6) ± 0.0010.087(3) ± 0.0010.088(4) ± 0.0010.088(5) ± 0.0010.059(2) ± 0.0000.120(7) ± 0.0020.120(8) ± 0.0010.121(9) ± 0.0010.038(1) ± 0.001 •
average rank4.823.453.464.843.647.687.368.541.21
win/tie/loss106/21/97141/24/59143/29/52105/20/99143/12/6927/20/17734/24/1668/17/199212/11/1
Table A9. The performance of the algorithms for hubness reduction technique comparisons on F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A9. The performance of the algorithms for hubness reduction technique comparisons on F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Macro
BRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
birds0.034(9) ± 0.0190.058(7) ± 0.0240.072(6) ± 0.0190.052(8) ± 0.0180.157(2) ± 0.0270.118(3) ± 0.0220.092(4) ± 0.0340.082(5) ± 0.0270.258(1) ± 0.069 •
CAL5000.083(5) ± 0.0020.075(9) ± 0.0040.083(2) ± 0.0020.083(4) ± 0.0030.103(1) ± 0.003 •0.080(6) ± 0.0050.080(8) ± 0.0050.080(7) ± 0.0040.083(3) ± 0.003
emotions0.608(4) ± 0.0110.544(9) ± 0.0090.626(2) ± 0.0110.624(3) ± 0.0130.602(7) ± 0.0230.604(5) ± 0.0230.595(8) ± 0.0240.603(6) ± 0.0200.651(1) ± 0.013 •
genbase0.639(8) ± 0.0580.648(7) ± 0.0620.649(6) ± 0.0590.567(9) ± 0.0540.913(2) ± 0.0300.743(3) ± 0.0630.740(4) ± 0.0830.701(5) ± 0.0970.954(1) ± 0.017 •
LLOG0.141(1) ± 0.012 •0.115(5) ± 0.0240.118(3) ± 0.0280.115(5) ± 0.0240.116(4) ± 0.0050.124(2) ± 0.0130.095(9) ± 0.0170.103(7) ± 0.0170.097(8) ± 0.016
enron0.067(8) ± 0.0080.084(7) ± 0.0080.090(6) ± 0.0060.067(9) ± 0.0080.131(2) ± 0.0070.100(4) ± 0.0090.101(3) ± 0.0050.092(5) ± 0.0070.131(1) ± 0.009 •
scene0.643(9) ± 0.0150.664(8) ± 0.0100.696(7) ± 0.0100.699(6) ± 0.0110.726(5) ± 0.0070.738(3) ± 0.0120.739(2) ± 0.0120.734(4) ± 0.0090.749(1) ± 0.008 •
yeast0.347(6) ± 0.0060.285(9) ± 0.0090.344(8) ± 0.0060.344(7) ± 0.0060.385(2) ± 0.0100.356(3) ± 0.0100.356(5) ± 0.0100.356(4) ± 0.0080.420(1) ± 0.009 •
Slashdot0.081(8) ± 0.0020.090(7) ± 0.0140.097(6) ± 0.0170.080(9) ± 0.0010.149(2) ± 0.0360.115(5) ± 0.0260.117(4) ± 0.0260.126(3) ± 0.0160.163(1) ± 0.025 •
corel5k0.001(9) ± 0.0010.006(8) ± 0.0010.007(7) ± 0.0010.007(6) ± 0.0010.038(1) ± 0.006 •0.016(4) ± 0.0030.016(5) ± 0.0030.017(3) ± 0.0030.026(2) ± 0.004
rcv1subset10.074(9) ± 0.0060.091(7) ± 0.0050.097(6) ± 0.0060.078(8) ± 0.0040.181(1) ± 0.010 •0.145(4) ± 0.0060.145(3) ± 0.0040.136(5) ± 0.0060.152(2) ± 0.014
rcv1subset20.067(9) ± 0.0030.077(7) ± 0.0030.083(6) ± 0.0030.073(8) ± 0.0020.173(1) ± 0.011 •0.125(3) ± 0.0050.123(4) ± 0.0080.111(5) ± 0.0060.133(2) ± 0.015
rcv1subset30.060(9) ± 0.0010.074(7) ± 0.0020.081(6) ± 0.0100.067(8) ± 0.0030.164(1) ± 0.006 •0.120(3) ± 0.0080.124(2) ± 0.0120.111(5) ± 0.0080.120(4) ± 0.010
rcv1subset40.081(9) ± 0.0050.094(7) ± 0.0060.099(6) ± 0.0060.086(8) ± 0.0050.178(1) ± 0.008 •0.135(3) ± 0.0080.132(4) ± 0.0100.128(5) ± 0.0090.149(2) ± 0.013
rcv1subset5 0.069(9) ± 0.0030.075(7) ± 0.0040.081(6) ± 0.0030.072(8) ± 0.0040.163(1) ± 0.006 •0.117(3) ± 0.0110.107(5) ± 0.0030.107(4) ± 0.0080.134(2) ± 0.017
bibtex0.060(9) ± 0.0020.069(8) ± 0.0040.090(6) ± 0.0050.079(7) ± 0.0040.148(2) ± 0.0060.122(4) ± 0.0060.131(3) ± 0.0050.112(5) ± 0.0030.187(1) ± 0.006 •
Arts0.140(7) ± 0.0100.140(6) ± 0.0110.148(5) ± 0.0100.112(8) ± 0.0110.223(2) ± 0.0130.188(3) ± 0.0090.178(4) ± 0.0150.004(9) ± 0.0060.263(1) ± 0.015 •
Health0.269(7) ± 0.0160.278(5) ± 0.0150.278(6) ± 0.0110.216(8) ± 0.0120.360(2) ± 0.0190.336(3) ± 0.0090.311(4) ± 0.0170.026(9) ± 0.0030.404(1) ± 0.011 •
Business0.141(7) ± 0.0080.147(6) ± 0.0100.154(5) ± 0.0100.124(8) ± 0.0070.239(2) ± 0.0100.192(3) ± 0.0100.172(4) ± 0.0110.039(9) ± 0.0010.267(1) ± 0.005 •
Education0.145(6) ± 0.0080.146(5) ± 0.0080.144(7) ± 0.0080.114(9) ± 0.0120.220(2) ± 0.0090.167(3) ± 0.0120.164(4) ± 0.0140.137(8) ± 0.0090.278(1) ± 0.015 •
Computers0.142(7) ± 0.0110.144(6) ± 0.0120.160(5) ± 0.0080.121(8) ± 0.0110.255(2) ± 0.0140.201(3) ± 0.0150.182(4) ± 0.0120.023(9) ± 0.0020.294(1) ± 0.010 •
Entertainment0.257(7) ± 0.0080.260(6) ± 0.0080.263(5) ± 0.0070.215(8) ± 0.0110.347(2) ± 0.0120.304(3) ± 0.0130.285(4) ± 0.0120.006(9) ± 0.0080.390(1) ± 0.008 •
Recreation0.240(7) ± 0.0220.247(6) ± 0.0180.254(5) ± 0.0180.191(8) ± 0.0160.352(2) ± 0.0140.315(3) ± 0.0150.297(4) ± 0.0240.002(9) ± 0.0030.390(1) ± 0.016 •
Society0.158(7) ± 0.0080.161(6) ± 0.0080.164(5) ± 0.0080.117(8) ± 0.0130.246(2) ± 0.0100.230(3) ± 0.0130.216(4) ± 0.0130.021(9) ± 0.0030.291(1) ± 0.012 •
eurlex-dc-l0.148(3) ± 0.0030.147(4) ± 0.0040.145(5) ± 0.0030.118(8) ± 0.0050.177(2) ± 0.0050.140(6) ± 0.0040.132(7) ± 0.0040.001(9) ± 0.0010.267(1) ± 0.010 •
eurlex-sm0.252(5) ± 0.0050.253(4) ± 0.0050.250(6) ± 0.0050.234(8) ± 0.0040.296(2) ± 0.0070.259(3) ± 0.0040.248(7) ± 0.0060.232(9) ± 0.0060.388(1) ± 0.009 •
tmc2007-5000.239(9) ± 0.0040.289(7) ± 0.0020.291(6) ± 0.0040.248(8) ± 0.0060.464(2) ± 0.0100.443(3) ± 0.0090.429(4) ± 0.0050.395(5) ± 0.0080.603(1) ± 0.007 •
mediamill0.085(6) ± 0.0020.055(9) ± 0.0020.084(8) ± 0.0010.085(7) ± 0.0020.137(2) ± 0.0060.103(3) ± 0.0050.100(4) ± 0.0030.095(5) ± 0.0030.293(1) ± 0.005 •
average rank7.116.775.617.482.113.464.546.321.61
win/tie/loss41/22/16153/20/15178/31/11531/22/171190/10/24137/30/57111/39/7458/34/132196/18/10
Table A10. The performance of the algorithms for hubness reduction technique comparisons on F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A10. The performance of the algorithms for hubness reduction technique comparisons on F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 micro
BRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
birds0.074(9) ± 0.0420.130(7) ± 0.0580.159(5) ± 0.0340.112(8) ± 0.0370.230(2) ± 0.0500.205(3) ± 0.0350.164(4) ± 0.0540.146(6) ± 0.0450.358(1) ± 0.061 •
CAL5000.321(6) ± 0.0060.311(9) ± 0.0070.316(8) ± 0.0050.317(7) ± 0.0060.338(2) ± 0.0090.330(3) ± 0.0090.327(5) ± 0.0090.330(4) ± 0.0090.342(1) ± 0.014 •
emotions0.635(6) ± 0.0130.592(9) ± 0.0130.654(2) ± 0.0130.652(3) ± 0.0140.629(8) ± 0.0200.641(4) ± 0.0190.629(7) ± 0.0280.639(5) ± 0.0150.672(1) ± 0.011 •
genbase0.839(7) ± 0.0400.846(6) ± 0.0400.850(5) ± 0.0360.779(9) ± 0.0350.949(2) ± 0.0200.851(4) ± 0.0370.855(3) ± 0.0460.819(8) ± 0.0470.965(1) ± 0.014 •
LLOG0.466(1) ± 0.023 •0.394(4) ± 0.0620.407(3) ± 0.0630.394(4) ± 0.0620.427(2) ± 0.0130.360(7) ± 0.0280.327(9) ± 0.0530.351(8) ± 0.0390.363(6) ± 0.059
enron0.219(8) ± 0.0220.314(7) ± 0.0160.321(6) ± 0.0140.219(9) ± 0.0270.476(1) ± 0.011 •0.458(3) ± 0.0170.453(4) ± 0.0130.414(5) ± 0.0210.460(2) ± 0.013
scene0.649(9) ± 0.0110.663(8) ± 0.0090.696(7) ± 0.0090.700(6) ± 0.0090.720(5) ± 0.0070.734(3) ± 0.0110.736(2) ± 0.0110.730(4) ± 0.0080.743(1) ± 0.008 •
yeast0.628(6) ± 0.0040.585(9) ± 0.0070.627(8) ± 0.0040.627(7) ± 0.0050.635(5) ± 0.0060.638(2) ± 0.0040.637(4) ± 0.0050.637(3) ± 0.0050.643(1) ± 0.006 •
Slashdot0.843(4) ± 0.0070.843(3) ± 0.0070.844(1) ± 0.007 •0.843(6) ± 0.0080.843(2) ± 0.0110.833(8) ± 0.0140.832(9) ± 0.0120.835(7) ± 0.0090.843(5) ± 0.008
corel5k0.003(9) ± 0.0030.020(8) ± 0.0070.026(7) ± 0.0070.027(6) ± 0.0100.094(1) ± 0.013 •0.031(3) ± 0.0080.028(5) ± 0.0070.030(4) ± 0.0050.050(2) ± 0.011
rcv1subset10.203(9) ± 0.0070.240(7) ± 0.0040.246(6) ± 0.0050.209(8) ± 0.0050.345(1) ± 0.011 •0.299(3) ± 0.0140.306(2) ± 0.0110.286(5) ± 0.0100.296(4) ± 0.021
rcv1subset20.275(8) ± 0.0080.288(7) ± 0.0070.295(6) ± 0.0060.274(9) ± 0.0050.367(1) ± 0.016 •0.329(3) ± 0.0110.327(4) ± 0.0120.314(5) ± 0.0120.334(2) ± 0.012
rcv1subset30.258(9) ± 0.0070.281(7) ± 0.0070.290(6) ± 0.0160.263(8) ± 0.0090.362(1) ± 0.013 •0.324(4) ± 0.0100.326(3) ± 0.0190.318(5) ± 0.0140.333(2) ± 0.009
rcv1subset40.336(9) ± 0.0070.355(7) ± 0.0060.359(6) ± 0.0060.339(8) ± 0.0050.430(1) ± 0.007 •0.396(2) ± 0.0100.392(4) ± 0.0150.387(5) ± 0.0120.394(3) ± 0.019
rcv1subset50.284(8) ± 0.0100.297(7) ± 0.0090.299(6) ± 0.0080.280(9) ± 0.0070.369(1) ± 0.009 •0.341(3) ± 0.0140.329(4) ± 0.0110.321(5) ± 0.0110.348(2) ± 0.014
bibtex0.241(9) ± 0.0060.252(8) ± 0.0060.295(5) ± 0.0060.272(7) ± 0.0060.328(2) ± 0.0040.312(4) ± 0.0040.323(3) ± 0.0080.292(6) ± 0.0070.367(1) ± 0.006 •
Arts0.263(7) ± 0.0160.267(5) ± 0.0170.267(6) ± 0.0160.234(8) ± 0.0200.340(2) ± 0.0170.310(3) ± 0.0150.299(4) ± 0.0210.012(9) ± 0.0200.369(1) ± 0.014 •
Health0.491(7) ± 0.0120.502(5) ± 0.0130.502(6) ± 0.0120.463(8) ± 0.0130.531(2) ± 0.0170.525(3) ± 0.0110.519(4) ± 0.0140.258(9) ± 0.0590.574(1) ± 0.011 •
Business0.705(7) ± 0.0050.707(4) ± 0.0050.705(6) ± 0.0040.696(8) ± 0.0050.714(2) ± 0.0060.708(3) ± 0.0040.705(5) ± 0.0040.670(9) ± 0.0040.722(1) ± 0.004 •
Education0.288(7) ± 0.0120.290(5) ± 0.0110.290(6) ± 0.0110.254(9) ± 0.0120.357(2) ± 0.0130.310(3) ± 0.0140.301(4) ± 0.0160.267(8) ± 0.0180.378(1) ± 0.014 •
Computers0.468(4) ± 0.0040.463(7) ± 0.0040.463(6) ± 0.0030.446(8) ± 0.0130.490(2) ± 0.0080.480(3) ± 0.0060.466(5) ± 0.0090.362(9) ± 0.0330.511(1) ± 0.005 •
Entertainment0.400(7) ± 0.0090.407(5) ± 0.0100.405(6) ± 0.0090.349(8) ± 0.0110.472(2) ± 0.0140.457(3) ± 0.0110.435(4) ± 0.0080.021(9) ± 0.0320.505(1) ± 0.008 •
Recreation0.345(7) ± 0.0180.350(6) ± 0.0180.357(5) ± 0.0160.285(8) ± 0.0180.428(2) ± 0.0170.411(3) ± 0.0150.399(4) ± 0.0160.005(9) ± 0.0090.454(1) ± 0.012 •
Society0.385(7) ± 0.0140.389(5) ± 0.0120.386(6) ± 0.0110.354(8) ± 0.0160.433(2) ± 0.0100.430(3) ± 0.0140.412(4) ± 0.0140.236(9) ± 0.0470.468(1) ± 0.010 •
eurlex-dc-l0.390(3) ± 0.0030.390(4) ± 0.0030.389(5) ± 0.0040.333(8) ± 0.0090.401(2) ± 0.0060.370(6) ± 0.0050.359(7) ± 0.0050.026(9) ± 0.0270.479(1) ± 0.007 •
eurlex-sm0.504(6) ± 0.0060.506(4) ± 0.0060.505(5) ± 0.0050.490(8) ± 0.0030.533(2) ± 0.0060.516(3) ± 0.0050.502(7) ± 0.0070.489(9) ± 0.0050.589(1) ± 0.007 •
tmc2007-5000.577(8) ± 0.0030.606(7) ± 0.0030.607(6) ± 0.0030.574(9) ± 0.0060.634(5) ± 0.0060.658(2) ± 0.0040.652(3) ± 0.0050.635(4) ± 0.0060.726(1) ± 0.003 •
mediamill0.573(6) ± 0.0020.548(9) ± 0.0020.571(7) ± 0.0020.571(8) ± 0.0020.583(2) ± 0.0040.577(3) ± 0.0030.575(4) ± 0.0030.574(5) ± 0.0050.635(1) ± 0.001 •
average rank6.896.415.617.592.293.464.546.541.68
win/tie/loss46/25/15359/26/13978/36/11030/26/168175/23/26129/44/51103/47/7453/38/133190/25/9
Table A11. The performance of the algorithms for hubness reduction technique comparisons on Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A11. The performance of the algorithms for hubness reduction technique comparisons on Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetHamming Loss
BRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
birds0.144(7) ± 0.0070.141(3) ± 0.0060.139(2) ± 0.0060.142(5) ± 0.0060.145(8) ± 0.0070.141(4) ± 0.0070.144(6) ± 0.0090.146(9) ± 0.0090.137(1) ± 0.008 •
CAL5000.187(3) ± 0.0020.187(2) ± 0.0010.188(5) ± 0.0010.188(4) ± 0.0010.192(9) ± 0.0010.189(7) ± 0.0010.189(6) ± 0.0010.189(8) ± 0.0020.187(1) ± 0.001 •
emotions0.198(4) ± 0.0070.216(9) ± 0.0070.192(2) ± 0.0070.193(3) ± 0.0090.211(8) ± 0.0060.203(5) ± 0.0100.207(7) ± 0.0100.203(6) ± 0.0060.190(1) ± 0.005 •
genbase0.026(7) ± 0.0070.024(6) ± 0.0060.024(5) ± 0.0060.035(9) ± 0.0060.009(2) ± 0.0030.023(4) ± 0.0050.022(3) ± 0.0060.027(8) ± 0.0060.006(1) ± 0.002 •
LLOG0.192(3) ± 0.0040.195(5) ± 0.0050.193(4) ± 0.0060.195(5) ± 0.0050.185(1) ± 0.004 •0.253(9) ± 0.0370.215(8) ± 0.0210.213(7) ± 0.0210.188(2) ± 0.006
enron0.090(8) ± 0.0010.085(7) ± 0.0020.085(6) ± 0.0020.090(9) ± 0.0010.080(4) ± 0.0010.079(2) ± 0.0010.079(3) ± 0.0010.082(5) ± 0.0010.075(1) ± 0.001 •
scene0.103(9) ± 0.0030.097(8) ± 0.0020.091(6) ± 0.0020.090(5) ± 0.0030.092(7) ± 0.0020.087(3) ± 0.0030.086(2) ± 0.0020.087(4) ± 0.0030.084(1) ± 0.003 •
yeast0.199(6) ± 0.0030.208(9) ± 0.0020.197(1) ± 0.002 •0.197(3) ± 0.0030.204(8) ± 0.0020.197(2) ± 0.0030.198(4) ± 0.0020.198(5) ± 0.0020.200(7) ± 0.003
Slashdot0.028(4) ± 0.0010.028(3) ± 0.0010.028(1) ± 0.001 •0.028(5) ± 0.0010.028(2) ± 0.0020.030(7) ± 0.0030.030(9) ± 0.0030.030(8) ± 0.0020.029(6) ± 0.001
corel5k0.021(4) ± 0.0000.021(6) ± 0.0000.021(8) ± 0.0000.021(7) ± 0.0000.022(9) ± 0.0000.021(3) ± 0.0000.021(2) ± 0.0000.021(5) ± 0.0000.021(1) ± 0.000 •
rcv1subset10.034(9) ± 0.0000.033(2) ± 0.0000.033(5) ± 0.0000.034(8) ± 0.0000.033(6) ± 0.0000.033(3) ± 0.0000.033(4) ± 0.0000.033(7) ± 0.0000.032(1) ± 0.000 •
rcv1subset20.028(7) ± 0.0000.028(3) ± 0.0000.028(2) ± 0.0000.028(5) ± 0.0000.028(9) ± 0.0000.028(6) ± 0.0000.028(4) ± 0.0000.028(8) ± 0.0000.027(1) ± 0.000 •
rcv1subset30.028(8) ± 0.0000.027(3) ± 0.0000.027(2) ± 0.0000.028(4) ± 0.0000.028(9) ± 0.0000.028(6) ± 0.0000.028(5) ± 0.0000.028(7) ± 0.0000.027(1) ± 0.000 •
rcv1subset40.025(9) ± 0.0000.025(2) ± 0.0000.025(3) ± 0.0000.025(7) ± 0.0000.025(8) ± 0.0000.025(4) ± 0.0000.025(6) ± 0.0000.025(5) ± 0.0000.024(1) ± 0.000 •
rcv1subset50.027(7) ± 0.0000.027(2) ± 0.0000.027(3) ± 0.0000.027(4) ± 0.0000.028(9) ± 0.0000.027(5) ± 0.0000.027(6) ± 0.0000.027(8) ± 0.0000.027(1) ± 0.000 •
bibtex0.013(9) ± 0.0000.013(8) ± 0.0000.013(1) ± 0.000 •0.013(5) ± 0.0000.013(6) ± 0.0000.013(4) ± 0.0000.013(3) ± 0.0000.013(7) ± 0.0000.013(2) ± 0.000
Arts0.061(5) ± 0.0010.061(4) ± 0.0010.061(7) ± 0.0010.063(8) ± 0.0010.061(3) ± 0.0000.060(2) ± 0.0010.061(5) ± 0.0010.069(9) ± 0.0010.059(1) ± 0.001 •
Health0.061(7) ± 0.0010.060(6) ± 0.0010.060(5) ± 0.0010.065(8) ± 0.0010.059(4) ± 0.0010.058(2) ± 0.0010.058(3) ± 0.0010.079(9) ± 0.0020.055(1) ± 0.001 •
Business0.032(7) ± 0.0010.031(5) ± 0.0010.031(4) ± 0.0000.033(8) ± 0.0010.031(2) ± 0.0010.031(3) ± 0.0010.032(6) ± 0.0010.034(9) ± 0.0010.031(1) ± 0.000 •
Education0.047(6) ± 0.0010.047(4) ± 0.0000.047(5) ± 0.0000.049(9) ± 0.0010.047(7) ± 0.0010.047(2) ± 0.0000.047(3) ± 0.0010.048(8) ± 0.0000.046(1) ± 0.001 •
Computers0.045(7) ± 0.0000.045(6) ± 0.0000.044(4) ± 0.0000.046(8) ± 0.0000.044(3) ± 0.0010.044(2) ± 0.0000.045(5) ± 0.0000.055(9) ± 0.0020.043(1) ± 0.000 •
Entertainment0.064(5) ± 0.0010.064(4) ± 0.0010.065(6) ± 0.0010.067(8) ± 0.0010.064(3) ± 0.0010.063(2) ± 0.0010.065(7) ± 0.0010.083(9) ± 0.0020.061(1) ± 0.001 •
Recreation0.053(7) ± 0.0010.053(6) ± 0.0010.053(5) ± 0.0010.055(8) ± 0.0010.052(3) ± 0.0010.051(2) ± 0.0010.052(4) ± 0.0010.065(9) ± 0.0000.050(1) ± 0.001 •
Society0.054(7) ± 0.0010.053(5) ± 0.0010.053(6) ± 0.0000.055(8) ± 0.0010.053(3) ± 0.0010.052(2) ± 0.0000.053(4) ± 0.0010.063(9) ± 0.0020.051(1) ± 0.001 •
eurlex-dc-l0.005(4) ± 0.0000.005(3) ± 0.0000.005(5) ± 0.0000.005(8) ± 0.0000.005(2) ± 0.0000.005(6) ± 0.0000.005(7) ± 0.0000.006(9) ± 0.0000.004(1) ± 0.000 •
eurlex-sm0.012(6) ± 0.0000.012(4) ± 0.0000.012(5) ± 0.0000.012(8) ± 0.0000.012(3) ± 0.0000.012(2) ± 0.0000.012(7) ± 0.0000.012(9) ± 0.0000.011(1) ± 0.000 •
tmc2007-5000.070(8) ± 0.0010.066(4) ± 0.0000.067(5) ± 0.0010.070(9) ± 0.0010.067(7) ± 0.0000.063(2) ± 0.0010.064(3) ± 0.0000.067(6) ± 0.0010.051(1) ± 0.000 •
mediamill0.031(7) ± 0.0000.031(9) ± 0.0000.030(4) ± 0.0000.030(6) ± 0.0000.031(8) ± 0.0000.030(2) ± 0.0000.030(3) ± 0.0000.030(5) ± 0.0000.028(1) ± 0.000 •
average rank6.454.954.186.595.463.684.847.391.46
win/tie/loss50/46/12892/44/88109/47/6853/36/13580/34/110118/49/5789/56/7934/38/152200/16/8
Table A12. The Friedman test statistics F F for comparisons among nine algorithms BRkNNa, BRkNNa-dsl, BRkNNa-ls, BRkNNa-mp, MLKNN, MLKNN-dsl, MLKNN-ls, MLKNN-mp, and MLHiKNN. The critical value was calculated at a significance level of 0.05. # Algorithms and # Datasets denote the number of algorithms and the number of datasets, respectively.
Table A12. The Friedman test statistics F F for comparisons among nine algorithms BRkNNa, BRkNNa-dsl, BRkNNa-ls, BRkNNa-mp, MLKNN, MLKNN-dsl, MLKNN-ls, MLKNN-mp, and MLHiKNN. The critical value was calculated at a significance level of 0.05. # Algorithms and # Datasets denote the number of algorithms and the number of datasets, respectively.
Metric F F # Algorithms# DatasetsCritical Value
AUC macro203.4019281.981
AUC micro181.614
Ranking Loss172.663
F1 macro143.012
F1 micro134.357
Hamming Loss95.295
Table A13. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of AUC micro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table A13. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of AUC micro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
AUC microBRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
BRkNNa-0.9981.0000.2191.83 × 10 5 3.73 × 10 9 3.73 × 10 9 3.73 × 10 9 0.998
BRkNNa-dsl0.002-0.8680.0084.10 × 10 7 3.73 × 10 9 3.73 × 10 9 3.73 × 10 9 0.959
BRkNNa-ls6.47 × 10 6 0.137-1.23 × 10 7 2.05 × 10 7 3.73 × 10 9 3.73 × 10 9 3.73 × 10 9 0.925
BRkNNa-mp0.7880.9921.000-5.52 × 10 6 3.73 × 10 9 3.73 × 10 9 3.73 × 10 9 0.998
MLKNN1.0001.0001.0001.000-3.73 × 10 9 3.73 × 10 9 3.73 × 10 9 1.000
MLKNN-dsl1.0001.0001.0001.0001.000-0.9842.63 × 10 4 1.000
MLKNN-ls1.0001.0001.0001.0001.0000.017-1.03 × 10 5 1.000
MLKNN-mp1.0001.0001.0001.0001.0001.0001.000-1.000
MLHiKNN0.0020.0430.0780.0021.14 × 10 6 3.73 × 10 9 3.73 × 10 9 3.73 × 10 9 -
Table A14. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of Ranking Loss. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table A14. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of Ranking Loss. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Ranking LossBRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
BRkNNa-0.9991.0000.3150.8422.61 × 10 7 3.28 × 10 7 2.05 × 10 7 1.000
BRkNNa-dsl0.001-0.5930.0020.5581.23 × 10 7 1.23 × 10 7 1.23 × 10 7 1.000
BRkNNa-ls3.37 × 10 6 0.416-5.10 × 10 7 0.4787.08 × 10 8 7.08 × 10 8 3.73 × 10 8 1.000
BRkNNa-mp0.6930.9981.000-0.8367.08 × 10 8 7.08 × 10 8 7.08 × 10 8 1.000
MLKNN0.1640.4510.5310.169-3.73 × 10 9 3.73 × 10 9 3.73 × 10 9 1.000
MLKNN-dsl1.0001.0001.0001.0001.000-0.9842.63 × 10 4 1.000
MLKNN-ls1.0001.0001.0001.0001.0000.017-1.14 × 10 6 1.000
MLKNN-mp1.0001.0001.0001.0001.0001.0001.000-1.000
MLHiKNN3.73 × 10 9 1.86 × 10 8 3.73 × 10 9 3.73 × 10 9 7.71 × 10 7 3.73 × 10 9 3.73 × 10 9 3.73 × 10 9 -
Table A15. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of F1 macro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table A15. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of F1 macro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
F1 MacroBRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
BRkNNa-0.9761.0000.0661.0001.0001.0000.4071.000
BRkNNa-dsl0.025-1.0000.0161.0001.0001.0000.4421.000
BRkNNa-ls4.66 × 10 5 2.00 × 10 6 -3.28 × 10 7 1.0001.0001.0000.2841.000
BRkNNa-mp0.9370.9841.000-1.0001.0001.0000.5131.000
MLKNN2.61 × 10 8 3.73 × 10 9 2.61 × 10 8 1.86 × 10 8 -5.22 × 10 8 1.12 × 10 8 1.86 × 10 8 0.999
MLKNN-dsl4.10 × 10 7 1.86 × 10 8 3.28 × 10 7 7.08 × 10 8 1.000-5.29 × 10 5 6.30 × 10 7 1.000
MLKNN-ls8.83 × 10 6 3.28 × 10 7 2.76 × 10 5 6.30 × 10 7 1.0001.000-2.76 × 10 5 1.000
MLKNN-mp0.6020.5670.7240.4961.0001.0001.000-1.000
MLHiKNN2.61 × 10 8 1.12 × 10 8 2.61 × 10 8 1.12 × 10 8 0.0011.60 × 10 7 1.86 × 10 8 1.12 × 10 8 -
Table A16. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of F1 micro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table A16. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of F1 micro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
F1 MicroBRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
BRkNNa-0.9721.0000.0371.0001.0001.0000.2991.000
BRkNNa-dsl0.030-1.0000.0101.0001.0001.0000.3151.000
BRkNNa-ls5.32 × 10 4 1.37 × 10 4 -5.10 × 10 7 1.0001.0000.9990.0661.000
BRkNNa-mp0.9650.9901.000-1.0001.0001.0000.4331.000
MLKNN2.61 × 10 7 3.73 × 10 9 1.60 × 10 7 7.08 × 10 8 -1.83 × 10 5 2.00 × 10 6 6.30 × 10 7 0.965
MLKNN-dsl3.60 × 10 5 3.37 × 10 6 9.68 × 10 5 9.42 × 10 7 1.000-1.09 × 10 4 2.61 × 10 8 1.000
MLKNN-ls2.92 × 10 4 1.09 × 10 4 9.34 × 10 4 2.11 × 10 5 1.0001.000-2.76 × 10 5 1.000
MLKNN-mp0.7090.6930.9370.5761.0001.0001.000-1.000
MLHiKNN1.67 × 10 6 5.22 × 10 8 1.60 × 10 7 7.08 × 10 8 0.0373.73 × 10 8 7.08 × 10 8 3.73 × 10 9 -
Table A17. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of Hamming Loss. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table A17. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of Hamming Loss. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Hamming LossBRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
BRkNNa-0.9941.0000.1050.9400.9970.9730.0141.000
BRkNNa-dsl0.007-0.9550.0080.9030.9820.7940.0071.000
BRkNNa-ls1.91 × 10 4 0.047-3.37 × 10 6 0.2610.8360.3232.42 × 10 5 1.000
BRkNNa-mp0.8990.9921.000-0.8770.9960.9750.0101.000
MLKNN0.0630.1010.7460.127-0.9980.9070.0301.000
MLKNN-dsl0.0040.0190.1690.0040.002-0.0095.52 × 10 6 1.000
MLKNN-ls0.0280.2120.6850.0270.0970.992-1.09 × 10 4 1.000
MLKNN-mp0.9870.9941.0000.9900.9721.0001.000-1.000
MLHiKNN7.08 × 10 8 1.86 × 10 8 5.52 × 10 6 9.42 × 10 7 2.38 × 10 6 2.00 × 10 6 4.10 × 10 7 2.05 × 10 7 -

Appendix D. Results of Ablation Analysis Experiments

Table A18. The performance of MLHiKNN and the variant approaches in terms of AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A18. The performance of MLHiKNN and the variant approaches in terms of AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Micro
MLHiKNN-g1MLHiKNN-g0MLHiKNN-h1MLHiKNN-d1MLHiKNN-foMLHiKNN
birds0.737(5) ± 0.0220.743(4) ± 0.0250.744(3) ± 0.0180.737(6) ± 0.0150.784(1) ± 0.019 •0.746(2) ± 0.019
CAL5000.751(5) ± 0.0070.722(6) ± 0.0100.759(2) ± 0.0040.758(3) ± 0.0030.764(1) ± 0.002 •0.757(4) ± 0.004
emotions0.862(1) ± 0.005 •0.854(6) ± 0.0060.862(2) ± 0.0050.859(4) ± 0.0070.858(5) ± 0.0060.862(3) ± 0.005
genbase0.995(6) ± 0.0040.996(5) ± 0.0040.997(3) ± 0.0030.997(2) ± 0.0040.999(1) ± 0.001 •0.996(4) ± 0.004
LLOG0.764(6) ± 0.0050.767(5) ± 0.0020.767(2) ± 0.0020.767(4) ± 0.0020.798(1) ± 0.004 •0.767(3) ± 0.002
enron0.859(4) ± 0.0030.847(6) ± 0.0050.857(5) ± 0.0020.859(3) ± 0.0040.868(1) ± 0.004 •0.860(2) ± 0.002
scene0.950(1) ± 0.003 •0.948(6) ± 0.0020.948(4) ± 0.0020.949(3) ± 0.0020.948(5) ± 0.0020.950(2) ± 0.002
yeast0.841(4) ± 0.0040.829(6) ± 0.0050.842(3) ± 0.0040.840(5) ± 0.0030.843(1) ± 0.003 •0.842(2) ± 0.004
Slashdot0.938(6) ± 0.0060.944(4) ± 0.0030.944(5) ± 0.0040.948(2) ± 0.0040.949(1) ± 0.005 •0.945(3) ± 0.004
corel5k0.791(5) ± 0.0040.781(6) ± 0.0050.796(3) ± 0.0030.795(4) ± 0.0030.802(1) ± 0.004 •0.796(2) ± 0.004
rcv1subset10.859(6) ± 0.0070.905(2) ± 0.0030.871(5) ± 0.0080.872(4) ± 0.0090.931(1) ± 0.002 •0.874(3) ± 0.009
rcv1subset20.843(6) ± 0.0080.906(2) ± 0.0040.863(5) ± 0.0080.870(3) ± 0.0120.933(1) ± 0.004 •0.865(4) ± 0.010
rcv1subset30.835(6) ± 0.0060.897(2) ± 0.0050.860(5) ± 0.0090.864(3) ± 0.0080.930(1) ± 0.002 •0.861(4) ± 0.007
rcv1subset40.866(6) ± 0.0070.917(2) ± 0.0030.886(4) ± 0.0050.889(3) ± 0.0090.938(1) ± 0.002 •0.884(5) ± 0.010
rcv1subset50.847(6) ± 0.0070.902(2) ± 0.0030.866(5) ± 0.0080.871(3) ± 0.0070.931(1) ± 0.002 •0.868(4) ± 0.009
bibtex0.820(6) ± 0.0080.869(2) ± 0.0040.860(4) ± 0.0060.859(5) ± 0.0050.883(1) ± 0.004 •0.863(3) ± 0.006
Arts0.838(6) ± 0.0030.849(4) ± 0.0030.851(3) ± 0.0030.851(2) ± 0.0030.849(5) ± 0.0030.851(1) ± 0.003 •
Health0.912(6) ± 0.0020.916(5) ± 0.0020.918(3) ± 0.0020.918(2) ± 0.0020.917(4) ± 0.0020.919(1) ± 0.002 •
Business0.939(6) ± 0.0020.946(4) ± 0.0020.947(2) ± 0.0010.948(1) ± 0.001 •0.945(5) ± 0.0010.947(3) ± 0.001
Education0.899(6) ± 0.0020.905(4) ± 0.0020.906(3) ± 0.0020.906(2) ± 0.0020.903(5) ± 0.0020.906(1) ± 0.002 •
Computers0.884(6) ± 0.0050.892(4) ± 0.0040.895(3) ± 0.0030.895(2) ± 0.0030.891(5) ± 0.0020.896(1) ± 0.003 •
Entertainment0.880(6) ± 0.0010.889(4) ± 0.0020.890(3) ± 0.0020.890(2) ± 0.0020.886(5) ± 0.0020.891(1) ± 0.002 •
Recreation0.854(6) ± 0.0040.873(3) ± 0.0040.873(4) ± 0.0040.875(1) ± 0.004 •0.868(5) ± 0.0040.874(2) ± 0.004
Society0.861(5) ± 0.0020.864(4) ± 0.0040.868(2) ± 0.0020.868(3) ± 0.0020.860(6) ± 0.0020.869(1) ± 0.002 •
eurlex-dc-l0.904(6) ± 0.0050.915(3) ± 0.0040.919(1) ± 0.005 •0.905(5) ± 0.0060.909(4) ± 0.0030.919(2) ± 0.005
eurlex-sm0.948(3) ± 0.0010.945(5) ± 0.0020.952(2) ± 0.0020.946(4) ± 0.0020.943(6) ± 0.0020.952(1) ± 0.002 •
tmc2007-5000.956(3) ± 0.0010.949(6) ± 0.0010.956(4) ± 0.0010.952(5) ± 0.0010.956(2) ± 0.0010.958(1) ± 0.001 •
mediamill0.958(1) ± 0.001 •0.948(6) ± 0.0010.957(3) ± 0.0000.951(5) ± 0.0010.954(4) ± 0.0010.957(2) ± 0.001
average rank4.964.213.323.252.862.39
win/tie/loss19/23/9842/22/7655/46/3954/41/4581/13/4676/41/23
Table A19. The performance of MLHiKNN and the variant approaches in terms of Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A19. The performance of MLHiKNN and the variant approaches in terms of Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetRanking Loss
MLHiKNN-g1MLHiKNN-g0MLHiKNN-h1MLHiKNN-d1MLHiKNN-foMLHiKNN
birds0.269(6) ± 0.0250.253(3) ± 0.0290.256(4) ± 0.0190.262(5) ± 0.0180.217(1) ± 0.023 •0.251(2) ± 0.020
CAL5000.246(5) ± 0.0080.275(6) ± 0.0100.239(2) ± 0.0040.239(3) ± 0.0030.233(1) ± 0.002 •0.240(4) ± 0.005
emotions0.154(3) ± 0.0070.160(5) ± 0.0070.152(1) ± 0.007 •0.156(4) ± 0.0100.161(6) ± 0.0090.153(2) ± 0.007
genbase0.005(5) ± 0.0030.006(6) ± 0.0030.004(1) ± 0.002 •0.004(2) ± 0.0030.004(3) ± 0.0020.004(4) ± 0.003
LLOG0.190(5) ± 0.0050.187(4) ± 0.0030.187(1) ± 0.002 •0.187(3) ± 0.0020.195(6) ± 0.0030.187(2) ± 0.002
enron0.136(4) ± 0.0040.143(6) ± 0.0050.139(5) ± 0.0030.135(3) ± 0.0050.124(1) ± 0.003 •0.135(2) ± 0.003
scene0.071(1) ± 0.004 •0.074(5) ± 0.0020.073(4) ± 0.0030.072(3) ± 0.0030.074(6) ± 0.0030.071(2) ± 0.003
yeast0.168(2) ± 0.0030.180(6) ± 0.0040.168(4) ± 0.0030.170(5) ± 0.0030.167(1) ± 0.003 •0.168(3) ± 0.003
Slashdot0.051(6) ± 0.0050.048(4) ± 0.0030.048(5) ± 0.0040.044(2) ± 0.0040.042(1) ± 0.005 •0.047(3) ± 0.005
corel5k0.211(5) ± 0.0040.218(6) ± 0.0050.206(3) ± 0.0040.207(4) ± 0.0040.173(1) ± 0.004 •0.206(2) ± 0.004
rcv1subset10.132(6) ± 0.0070.087(2) ± 0.0030.121(5) ± 0.0070.119(4) ± 0.0080.058(1) ± 0.001 •0.118(3) ± 0.008
rcv1subset20.139(6) ± 0.0070.081(2) ± 0.0040.120(5) ± 0.0080.115(3) ± 0.0100.054(1) ± 0.002 •0.118(4) ± 0.009
rcv1subset30.145(6) ± 0.0040.089(2) ± 0.0030.123(5) ± 0.0070.120(3) ± 0.0080.056(1) ± 0.001 •0.122(4) ± 0.005
rcv1subset40.113(6) ± 0.0060.071(2) ± 0.0020.099(4) ± 0.0050.095(3) ± 0.0080.048(1) ± 0.002 •0.100(5) ± 0.009
rcv1subset50.135(6) ± 0.0080.087(2) ± 0.0020.118(5) ± 0.0080.114(3) ± 0.0060.055(1) ± 0.001 •0.117(4) ± 0.009
bibtex0.184(6) ± 0.0060.136(2) ± 0.0040.150(4) ± 0.0040.150(5) ± 0.0030.092(1) ± 0.002 •0.147(3) ± 0.004
Arts0.140(6) ± 0.0030.131(5) ± 0.0030.130(4) ± 0.0030.129(3) ± 0.0030.128(1) ± 0.002 •0.129(2) ± 0.003
Health0.077(6) ± 0.0010.072(5) ± 0.0020.071(4) ± 0.0020.070(2) ± 0.0020.070(3) ± 0.0020.070(1) ± 0.002 •
Business0.042(6) ± 0.0010.037(4) ± 0.0010.037(2) ± 0.0010.036(1) ± 0.001 •0.038(5) ± 0.0010.037(3) ± 0.001
Education0.095(6) ± 0.0020.091(5) ± 0.0020.090(3) ± 0.0020.089(1) ± 0.002 •0.090(4) ± 0.0020.089(2) ± 0.002
Computers0.093(6) ± 0.0030.089(5) ± 0.0020.086(2) ± 0.0020.087(3) ± 0.0020.087(4) ± 0.0020.086(1) ± 0.002 •
Entertainment0.119(6) ± 0.0020.114(4) ± 0.0030.113(3) ± 0.0030.112(2) ± 0.0020.114(5) ± 0.0020.111(1) ± 0.002 •
Recreation0.129(6) ± 0.0040.116(3) ± 0.0040.117(4) ± 0.0040.115(1) ± 0.004 •0.118(5) ± 0.0040.115(2) ± 0.004
Society0.112(6) ± 0.0030.110(4) ± 0.0040.106(2) ± 0.0020.107(3) ± 0.0020.110(5) ± 0.0020.105(1) ± 0.002 •
eurlex-dc-l0.092(6) ± 0.0040.081(3) ± 0.0040.078(1) ± 0.005 •0.090(4) ± 0.0050.090(5) ± 0.0040.078(2) ± 0.005
eurlex-sm0.055(3) ± 0.0010.056(5) ± 0.0020.050(2) ± 0.0020.056(6) ± 0.0020.056(4) ± 0.0010.050(1) ± 0.002 •
tmc2007-5000.048(5) ± 0.0010.050(6) ± 0.0010.046(2) ± 0.0010.047(4) ± 0.0010.047(3) ± 0.0010.044(1) ± 0.001 •
mediamill0.037(1) ± 0.001 •0.044(6) ± 0.0010.038(3) ± 0.0010.043(5) ± 0.0000.040(4) ± 0.0000.038(2) ± 0.001
average rank5.044.213.213.212.892.43
win/tie/loss17/23/10040/20/8053/46/4156/42/4278/23/3978/42/20
Table A20. The performance of MLHiKNN and the variant approaches in terms of F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A20. The performance of MLHiKNN and the variant approaches in terms of F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Macro
MLHiKNN-g1MLHiKNN-g0MLHiKNN-h1MLHiKNN-d1MLHiKNN-foMLHiKNN
birds0.217(5) ± 0.0740.348(1) ± 0.043 •0.269(2) ± 0.0520.236(4) ± 0.0660.161(6) ± 0.0180.258(3) ± 0.069
CAL5000.071(6) ± 0.0040.129(1) ± 0.006 •0.081(4) ± 0.0040.083(2) ± 0.0040.075(5) ± 0.0030.083(3) ± 0.003
emotions0.641(5) ± 0.0120.643(4) ± 0.0170.646(2) ± 0.0170.645(3) ± 0.0180.617(6) ± 0.0100.651(1) ± 0.013 •
genbase0.969(2) ± 0.0140.927(6) ± 0.0180.957(3) ± 0.0190.970(1) ± 0.013 •0.950(5) ± 0.0080.954(4) ± 0.017
LLOG0.091(6) ± 0.0160.094(4) ± 0.0130.102(2) ± 0.0130.094(5) ± 0.0180.114(1) ± 0.015 •0.097(3) ± 0.016
enron0.125(5) ± 0.0140.163(1) ± 0.008 •0.120(6) ± 0.0060.128(4) ± 0.0080.155(2) ± 0.0090.131(3) ± 0.009
scene0.748(2) ± 0.0090.744(5) ± 0.0060.744(4) ± 0.0070.746(3) ± 0.0090.602(6) ± 0.0150.749(1) ± 0.008 •
yeast0.371(5) ± 0.0100.435(1) ± 0.008 •0.418(3) ± 0.0080.416(4) ± 0.0060.322(6) ± 0.0040.420(2) ± 0.009
Slashdot0.156(4) ± 0.0280.144(6) ± 0.0190.152(5) ± 0.0270.179(2) ± 0.0320.182(1) ± 0.017 •0.163(3) ± 0.025
corel5k0.020(5) ± 0.0030.064(1) ± 0.006 •0.028(2) ± 0.0040.025(4) ± 0.0040.016(6) ± 0.0040.026(3) ± 0.004
rcv1subset10.122(5) ± 0.0110.260(1) ± 0.005 •0.149(4) ± 0.0110.154(2) ± 0.0120.114(6) ± 0.0080.152(3) ± 0.014
rcv1subset20.100(6) ± 0.0090.259(1) ± 0.007 •0.129(4) ± 0.0130.143(2) ± 0.0220.112(5) ± 0.0050.133(3) ± 0.015
rcv1subset30.085(6) ± 0.0070.215(1) ± 0.013 •0.114(4) ± 0.0100.126(2) ± 0.0080.097(5) ± 0.0050.120(3) ± 0.010
rcv1subset40.109(6) ± 0.0080.249(1) ± 0.011 •0.147(4) ± 0.0090.158(2) ± 0.0130.118(5) ± 0.0050.149(3) ± 0.013
rcv1subset50.096(6) ± 0.0130.237(1) ± 0.008 •0.130(4) ± 0.0150.140(2) ± 0.0120.108(5) ± 0.0030.134(3) ± 0.017
bibtex0.139(5) ± 0.0060.234(1) ± 0.006 •0.182(4) ± 0.0060.189(2) ± 0.0050.084(6) ± 0.0030.187(3) ± 0.006
Arts0.193(5) ± 0.0140.292(1) ± 0.011 •0.260(4) ± 0.0180.271(2) ± 0.0130.170(6) ± 0.0130.263(3) ± 0.015
Health0.353(5) ± 0.0120.414(1) ± 0.011 •0.399(4) ± 0.0110.407(2) ± 0.0150.309(6) ± 0.0120.404(3) ± 0.011
Business0.198(5) ± 0.0140.303(1) ± 0.010 •0.266(4) ± 0.0050.289(2) ± 0.0110.180(6) ± 0.0080.267(3) ± 0.005
Education0.210(5) ± 0.0160.305(1) ± 0.012 •0.274(4) ± 0.0160.283(2) ± 0.0110.182(6) ± 0.0110.278(3) ± 0.015
Computers0.228(5) ± 0.0180.322(1) ± 0.015 •0.292(4) ± 0.0120.307(2) ± 0.0150.217(6) ± 0.0120.294(3) ± 0.010
Entertainment0.325(5) ± 0.0100.398(1) ± 0.011 •0.382(4) ± 0.0090.385(3) ± 0.0100.297(6) ± 0.0060.390(2) ± 0.008
Recreation0.309(6) ± 0.0140.414(1) ± 0.013 •0.382(4) ± 0.0180.407(2) ± 0.0140.310(5) ± 0.0170.390(3) ± 0.016
Society0.232(5) ± 0.0100.305(1) ± 0.009 •0.285(4) ± 0.0120.298(2) ± 0.0130.204(6) ± 0.0110.291(3) ± 0.012
eurlex-dc-l0.240(4) ± 0.0060.266(3) ± 0.0080.267(2) ± 0.0110.233(5) ± 0.0080.194(6) ± 0.0030.267(1) ± 0.010 •
eurlex-sm0.367(4) ± 0.0070.377(3) ± 0.0070.387(2) ± 0.0070.353(5) ± 0.0070.303(6) ± 0.0060.388(1) ± 0.009 •
tmc2007-5000.601(2) ± 0.0060.526(5) ± 0.0080.590(3) ± 0.0090.553(4) ± 0.0070.484(6) ± 0.0080.603(1) ± 0.007 •
mediamill0.285(3) ± 0.0120.235(4) ± 0.0080.294(1) ± 0.004 •0.184(5) ± 0.0050.117(6) ± 0.0030.293(2) ± 0.005
average rank4.752.113.462.865.252.57
win/tie/loss29/21/90105/14/2156/32/5271/26/4319/4/11775/33/32
Table A21. The performance of MLHiKNN and the variant approaches in terms of F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A21. The performance of MLHiKNN and the variant approaches in terms of F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Micro
MLHiKNN-g1MLHiKNN-g0MLHiKNN-h1MLHiKNN-d1MLHiKNN-foMLHiKNN
birds0.317(5) ± 0.0910.414(1) ± 0.035 •0.361(2) ± 0.0720.323(4) ± 0.0800.245(6) ± 0.0240.358(3) ± 0.061
CAL5000.329(6) ± 0.0150.356(1) ± 0.008 •0.336(4) ± 0.0140.342(3) ± 0.0140.330(5) ± 0.0080.342(2) ± 0.014
emotions0.666(4) ± 0.0130.661(5) ± 0.0130.668(2) ± 0.0150.666(3) ± 0.0150.650(6) ± 0.0110.672(1) ± 0.011 •
genbase0.976(1) ± 0.011 •0.947(6) ± 0.0100.968(3) ± 0.0150.974(2) ± 0.0090.963(5) ± 0.0120.965(4) ± 0.014
LLOG0.343(6) ± 0.0640.357(4) ± 0.0510.382(2) ± 0.0430.350(5) ± 0.0680.415(1) ± 0.037 •0.363(3) ± 0.059
enron0.457(4) ± 0.0130.488(1) ± 0.010 •0.446(5) ± 0.0130.472(2) ± 0.0100.437(6) ± 0.0050.460(3) ± 0.013
scene0.743(2) ± 0.0090.737(5) ± 0.0070.738(4) ± 0.0070.740(3) ± 0.0090.613(6) ± 0.0120.743(1) ± 0.008 •
yeast0.644(1) ± 0.004 •0.630(5) ± 0.0070.641(3) ± 0.0070.639(4) ± 0.0060.625(6) ± 0.0040.643(2) ± 0.006
Slashdot0.839(6) ± 0.0100.844(2) ± 0.0080.843(3) ± 0.0070.847(1) ± 0.007 •0.839(5) ± 0.0080.843(4) ± 0.008
corel5k0.031(6) ± 0.0040.138(1) ± 0.008 •0.056(2) ± 0.0070.049(4) ± 0.0090.032(5) ± 0.0090.050(3) ± 0.011
rcv1subset10.244(6) ± 0.0100.386(1) ± 0.005 •0.290(4) ± 0.0170.300(2) ± 0.0220.245(5) ± 0.0090.296(3) ± 0.021
rcv1subset20.293(6) ± 0.0180.417(1) ± 0.009 •0.326(4) ± 0.0100.338(2) ± 0.0170.305(5) ± 0.0060.334(3) ± 0.012
rcv1subset30.290(6) ± 0.0120.403(1) ± 0.005 •0.325(4) ± 0.0110.337(2) ± 0.0150.291(5) ± 0.0080.333(3) ± 0.009
rcv1subset40.362(6) ± 0.0100.469(1) ± 0.006 •0.394(4) ± 0.0170.400(2) ± 0.0180.368(5) ± 0.0060.394(3) ± 0.019
rcv1subset50.286(6) ± 0.0130.425(1) ± 0.005 •0.346(4) ± 0.0190.357(2) ± 0.0220.320(5) ± 0.0060.348(3) ± 0.014
bibtex0.327(5) ± 0.0090.373(1) ± 0.004 •0.361(4) ± 0.0040.371(2) ± 0.0050.265(6) ± 0.0060.367(3) ± 0.006
Arts0.299(5) ± 0.0200.380(1) ± 0.013 •0.365(4) ± 0.0150.372(2) ± 0.0130.285(6) ± 0.0180.369(3) ± 0.014
Health0.538(5) ± 0.0130.569(3) ± 0.0110.568(4) ± 0.0110.574(2) ± 0.0100.521(6) ± 0.0120.574(1) ± 0.011 •
Business0.704(6) ± 0.0040.721(4) ± 0.0040.721(3) ± 0.0030.725(1) ± 0.004 •0.711(5) ± 0.0040.722(2) ± 0.004
Education0.315(5) ± 0.0170.393(1) ± 0.012 •0.374(4) ± 0.0130.379(2) ± 0.0140.299(6) ± 0.0140.378(3) ± 0.014
Computers0.475(6) ± 0.0070.513(2) ± 0.0050.508(4) ± 0.0050.514(1) ± 0.005 •0.487(5) ± 0.0060.511(3) ± 0.005
Entertainment0.454(5) ± 0.0100.508(1) ± 0.006 •0.498(4) ± 0.0070.504(3) ± 0.0070.436(6) ± 0.0080.505(2) ± 0.008
Recreation0.395(5) ± 0.0170.463(1) ± 0.011 •0.447(4) ± 0.0140.455(2) ± 0.0100.392(6) ± 0.0190.454(3) ± 0.012
Society0.431(5) ± 0.0120.470(1) ± 0.008 •0.465(4) ± 0.0110.467(3) ± 0.0090.420(6) ± 0.0130.468(2) ± 0.010
eurlex-dc-l0.466(3) ± 0.0070.453(4) ± 0.0060.479(2) ± 0.0080.447(5) ± 0.0040.420(6) ± 0.0050.479(1) ± 0.007 •
eurlex-sm0.580(3) ± 0.0060.564(5) ± 0.0060.586(2) ± 0.0070.565(4) ± 0.0070.537(6) ± 0.0050.589(1) ± 0.007 •
tmc2007-5000.732(1) ± 0.002 •0.679(5) ± 0.0030.721(3) ± 0.0030.679(6) ± 0.0020.687(4) ± 0.0030.726(2) ± 0.003
mediamill0.643(1) ± 0.002 •0.598(4) ± 0.0020.634(3) ± 0.0020.596(5) ± 0.0020.591(6) ± 0.0020.635(2) ± 0.001
average rank4.502.463.392.825.362.46
win/tie/loss33/26/8193/16/3163/22/5570/30/4012/9/11983/29/28
Table A22. The performance of MLHiKNN and the variant approaches in terms of Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A22. The performance of MLHiKNN and the variant approaches in terms of Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetHamming Loss
MLHiKNN-g1MLHiKNN-g0MLHiKNN-h1MLHiKNN-d1MLHiKNN-foMLHiKNN
birds0.135(2) ± 0.0090.144(6) ± 0.0090.139(4) ± 0.0080.140(5) ± 0.0070.133(1) ± 0.004 •0.137(3) ± 0.008
CAL5000.186(2) ± 0.0010.200(6) ± 0.0020.187(4) ± 0.0010.187(5) ± 0.0010.186(1) ± 0.001 •0.187(3) ± 0.001
emotions0.193(2) ± 0.0070.199(6) ± 0.0060.193(3) ± 0.0070.193(5) ± 0.0070.193(4) ± 0.0060.190(1) ± 0.005 •
genbase0.004(1) ± 0.002 •0.009(6) ± 0.0020.006(3) ± 0.0030.004(2) ± 0.0020.006(5) ± 0.0020.006(4) ± 0.002
LLOG0.190(6) ± 0.0070.188(4) ± 0.0050.187(2) ± 0.0050.189(5) ± 0.0060.187(1) ± 0.004 •0.188(3) ± 0.006
enron0.076(3) ± 0.0010.078(6) ± 0.0010.076(4) ± 0.0010.075(1) ± 0.001 •0.077(5) ± 0.0010.075(2) ± 0.001
scene0.083(1) ± 0.003 •0.087(5) ± 0.0020.086(4) ± 0.0020.085(3) ± 0.0030.104(6) ± 0.0030.084(2) ± 0.003
yeast0.195(1) ± 0.003 •0.211(6) ± 0.0040.200(4) ± 0.0040.201(5) ± 0.0040.197(2) ± 0.0030.200(3) ± 0.003
Slashdot0.029(5) ± 0.0020.028(2) ± 0.0020.029(3) ± 0.0010.028(1) ± 0.001 •0.030(6) ± 0.0010.029(4) ± 0.001
corel5k0.021(1) ± 0.000 •0.025(6) ± 0.0010.021(3) ± 0.0000.021(5) ± 0.0000.021(2) ± 0.0000.021(4) ± 0.000
rcv1subset10.033(4) ± 0.0000.035(6) ± 0.0010.032(2) ± 0.0000.033(3) ± 0.0000.033(5) ± 0.0000.032(1) ± 0.000 •
rcv1subset20.027(5) ± 0.0000.029(6) ± 0.0000.027(3) ± 0.0000.027(4) ± 0.0000.027(1) ± 0.000 •0.027(2) ± 0.000
rcv1subset30.027(4) ± 0.0000.029(6) ± 0.0000.027(3) ± 0.0000.027(2) ± 0.0000.027(5) ± 0.0000.027(1) ± 0.000 •
rcv1subset40.025(5) ± 0.0000.026(6) ± 0.0000.024(3) ± 0.0000.024(2) ± 0.0000.025(4) ± 0.0000.024(1) ± 0.000 •
rcv1subset50.027(5) ± 0.0000.028(6) ± 0.0000.027(4) ± 0.0000.027(3) ± 0.0000.027(2) ± 0.0000.027(1) ± 0.000 •
bibtex0.013(5) ± 0.0000.015(6) ± 0.0000.013(3) ± 0.0000.013(1) ± 0.000 •0.013(4) ± 0.0000.013(2) ± 0.000
Arts0.062(6) ± 0.0010.060(4) ± 0.0010.059(3) ± 0.0000.059(1) ± 0.000 •0.060(5) ± 0.0010.059(2) ± 0.001
Health0.057(5) ± 0.0010.056(4) ± 0.0010.055(3) ± 0.0010.055(2) ± 0.0010.058(6) ± 0.0010.055(1) ± 0.001 •
Business0.032(6) ± 0.0000.031(5) ± 0.0000.031(3) ± 0.0000.031(1) ± 0.000 •0.031(4) ± 0.0000.031(2) ± 0.000
Education0.048(6) ± 0.0010.046(5) ± 0.0010.046(3) ± 0.0010.046(1) ± 0.001 •0.046(4) ± 0.0010.046(2) ± 0.001
Computers0.044(6) ± 0.0010.043(5) ± 0.0000.043(3) ± 0.0000.043(1) ± 0.000 •0.043(4) ± 0.0000.043(2) ± 0.000
Entertainment0.064(6) ± 0.0010.061(4) ± 0.0010.061(3) ± 0.0010.061(2) ± 0.0010.062(5) ± 0.0010.061(1) ± 0.001 •
Recreation0.053(6) ± 0.0010.050(3) ± 0.0010.050(4) ± 0.0010.050(1) ± 0.001 •0.051(5) ± 0.0010.050(2) ± 0.001
Society0.053(6) ± 0.0010.051(4) ± 0.0010.051(3) ± 0.0010.051(1) ± 0.001 •0.052(5) ± 0.0010.051(2) ± 0.001
eurlex-dc-l0.005(3) ± 0.0000.005(6) ± 0.0000.005(2) ± 0.0000.005(5) ± 0.0000.005(4) ± 0.0000.004(1) ± 0.000 •
eurlex-sm0.011(3) ± 0.0000.011(6) ± 0.0000.011(2) ± 0.0000.011(5) ± 0.0000.011(4) ± 0.0000.011(1) ± 0.000 •
tmc2007-5000.050(1) ± 0.000 •0.060(6) ± 0.0000.052(3) ± 0.0000.060(5) ± 0.0000.055(4) ± 0.0000.051(2) ± 0.000
mediamill0.027(1) ± 0.000 •0.031(6) ± 0.0000.028(3) ± 0.0000.030(5) ± 0.0000.029(4) ± 0.0000.028(2) ± 0.000
average rank3.825.253.112.933.862.04
win/tie/loss47/24/6918/11/11163/36/4176/29/3548/24/6890/33/17

Appendix E. Results of Parameter Analysis Experiments for MLHiKNN

Table A23. The performance of MLHiKNN across different values of k in terms of AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A23. The performance of MLHiKNN across different values of k in terms of AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Macro
k = 3k = 5k = 10k = 15k = 20k = 25k = 30k = 35k = 40
birds0.735(1) ± 0.020 •0.732(3) ± 0.0200.732(2) ± 0.0250.730(4) ± 0.0260.728(7) ± 0.0300.729(6) ± 0.0280.729(5) ± 0.0260.720(8) ± 0.0240.715(9) ± 0.023
CAL5000.551(9) ± 0.0040.558(8) ± 0.0060.566(7) ± 0.0080.570(6) ± 0.0080.571(5) ± 0.0060.572(4) ± 0.0070.573(1) ± 0.007 •0.572(2) ± 0.0070.572(3) ± 0.007
emotions0.832(9) ± 0.0100.841(8) ± 0.0080.843(7) ± 0.0060.849(1) ± 0.006 •0.846(5) ± 0.0050.848(4) ± 0.0070.849(2) ± 0.0070.848(3) ± 0.0070.846(6) ± 0.006
genbase0.992(9) ± 0.0050.995(8) ± 0.0040.996(7) ± 0.0040.997(3) ± 0.0030.997(2) ± 0.0030.997(1) ± 0.003 •0.997(4) ± 0.0030.996(5) ± 0.0030.996(6) ± 0.003
LLOG0.603(8) ± 0.0060.603(9) ± 0.0060.603(2) ± 0.0060.603(7) ± 0.0060.603(6) ± 0.0060.603(1) ± 0.006 •0.603(4) ± 0.0060.603(3) ± 0.0060.603(5) ± 0.006
enron0.661(9) ± 0.0110.669(8) ± 0.0120.672(7) ± 0.0150.674(3) ± 0.0110.676(1) ± 0.012 •0.675(2) ± 0.0100.674(4) ± 0.0110.673(5) ± 0.0090.673(6) ± 0.009
scene0.925(9) ± 0.0040.935(8) ± 0.0040.942(7) ± 0.0020.943(5) ± 0.0030.944(1) ± 0.002 •0.944(2) ± 0.0020.943(3) ± 0.0020.943(4) ± 0.0020.943(6) ± 0.002
yeast0.682(9) ± 0.0090.695(8) ± 0.0090.707(7) ± 0.0080.711(2) ± 0.0070.711(1) ± 0.006 •0.710(3) ± 0.0060.710(4) ± 0.0060.708(5) ± 0.0050.707(6) ± 0.006
Slashdot0.720(8) ± 0.0150.713(9) ± 0.0170.722(6) ± 0.0200.724(2) ± 0.0160.723(3) ± 0.0210.723(4) ± 0.0200.732(1) ± 0.019 •0.722(5) ± 0.0240.722(7) ± 0.016
corel5k0.605(9) ± 0.0090.624(8) ± 0.0130.656(7) ± 0.0110.677(6) ± 0.0070.690(5) ± 0.0050.698(4) ± 0.0030.703(3) ± 0.0040.709(2) ± 0.0040.713(1) ± 0.004 •
rcv1subset10.880(9) ± 0.0040.889(5) ± 0.0030.892(1) ± 0.003 •0.892(2) ± 0.0050.891(3) ± 0.0040.890(4) ± 0.0040.888(6) ± 0.0040.887(7) ± 0.0040.886(8) ± 0.004
rcv1subset20.876(9) ± 0.0060.886(4) ± 0.0060.891(1) ± 0.005 •0.890(2) ± 0.0050.888(3) ± 0.0050.886(5) ± 0.0060.885(6) ± 0.0060.885(7) ± 0.0050.884(8) ± 0.005
rcv1subset30.870(9) ± 0.0050.879(4) ± 0.0040.882(1) ± 0.005 •0.882(2) ± 0.0040.880(3) ± 0.0040.877(5) ± 0.0030.876(6) ± 0.0030.875(7) ± 0.0040.874(8) ± 0.003
rcv1subset40.875(9) ± 0.0040.886(4) ± 0.0070.889(1) ± 0.006 •0.888(2) ± 0.0080.887(3) ± 0.0070.885(5) ± 0.0080.884(6) ± 0.0070.883(7) ± 0.0070.883(8) ± 0.008
rcv1subset50.862(9) ± 0.0050.873(3) ± 0.0050.876(1) ± 0.005 •0.874(2) ± 0.0050.873(4) ± 0.0050.870(5) ± 0.0050.869(6) ± 0.0050.869(7) ± 0.0050.868(8) ± 0.005
bibtex0.863(9) ± 0.0020.876(8) ± 0.0020.883(7) ± 0.0020.886(2) ± 0.0010.886(1) ± 0.001 •0.886(3) ± 0.0020.885(4) ± 0.0020.885(5) ± 0.0020.884(6) ± 0.001
Arts0.692(9) ± 0.0110.707(8) ± 0.0100.722(7) ± 0.0120.728(6) ± 0.0100.730(5) ± 0.0090.735(4) ± 0.0090.738(3) ± 0.0090.739(1) ± 0.008 •0.739(2) ± 0.008
Health0.738(9) ± 0.0070.755(8) ± 0.0100.769(7) ± 0.0110.775(6) ± 0.0090.778(5) ± 0.0090.779(3) ± 0.0080.779(4) ± 0.0110.782(2) ± 0.0100.782(1) ± 0.010 •
Business0.700(9) ± 0.0100.720(8) ± 0.0100.741(7) ± 0.0090.749(6) ± 0.0080.752(5) ± 0.0070.755(4) ± 0.0070.757(3) ± 0.0080.759(1) ± 0.008 •0.758(2) ± 0.006
Education0.700(9) ± 0.0140.715(8) ± 0.0150.733(7) ± 0.0150.744(6) ± 0.0120.748(5) ± 0.0140.750(4) ± 0.0140.753(3) ± 0.0140.755(2) ± 0.0110.756(1) ± 0.011 •
Computers0.706(9) ± 0.0080.720(8) ± 0.0080.736(7) ± 0.0050.744(6) ± 0.0050.750(5) ± 0.0050.752(4) ± 0.0060.753(3) ± 0.0050.755(2) ± 0.0060.755(1) ± 0.004 •
Entertainment0.744(9) ± 0.0080.757(8) ± 0.0070.770(7) ± 0.0060.777(6) ± 0.0030.782(5) ± 0.0040.784(4) ± 0.0040.785(3) ± 0.0050.787(2) ± 0.0040.789(1) ± 0.005 •
Recreation0.752(9) ± 0.0120.765(8) ± 0.0110.779(7) ± 0.0080.785(6) ± 0.0080.788(5) ± 0.0090.790(4) ± 0.0090.791(3) ± 0.0080.792(2) ± 0.0100.793(1) ± 0.009 •
Society0.676(9) ± 0.0070.687(8) ± 0.0070.697(7) ± 0.0050.702(6) ± 0.0050.707(5) ± 0.0060.709(4) ± 0.0060.711(3) ± 0.0060.713(2) ± 0.0050.714(1) ± 0.007 •
eurlex-dc-l0.865(9) ± 0.0050.882(8) ± 0.0070.895(7) ± 0.0050.899(6) ± 0.0050.900(5) ± 0.0040.903(4) ± 0.0040.904(3) ± 0.0040.904(2) ± 0.0040.905(1) ± 0.004 •
eurlex-sm0.894(9) ± 0.0030.903(8) ± 0.0020.907(7) ± 0.0030.909(2) ± 0.0020.908(4) ± 0.0030.909(1) ± 0.003 •0.909(3) ± 0.0030.908(5) ± 0.0030.907(6) ± 0.003
tmc2007-5000.893(9) ± 0.0030.904(8) ± 0.0020.915(7) ± 0.0020.919(6) ± 0.0020.920(5) ± 0.0020.922(4) ± 0.0020.924(3) ± 0.0020.924(2) ± 0.0020.925(1) ± 0.002 •
mediamill0.801(9) ± 0.0030.813(8) ± 0.0020.824(4) ± 0.0020.824(3) ± 0.0030.824(2) ± 0.0020.824(1) ± 0.003 •0.823(5) ± 0.0030.821(6) ± 0.0030.819(7) ± 0.004
average rank8.647.185.434.143.893.543.713.964.50
win/tie/loss1/22/20136/40/14874/54/9693/72/59104/74/46110/70/44109/72/43107/78/39102/62/60
Table A24. The performance of MLHiKNN across different values of k in terms of AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A24. The performance of MLHiKNN across different values of k in terms of AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Micro
k = 3k = 5k = 10k = 15k = 20k = 25k = 30k = 35k = 40
birds0.752(2) ± 0.0170.754(1) ± 0.014 •0.750(3) ± 0.0180.748(5) ± 0.0190.746(6) ± 0.0190.748(4) ± 0.0160.745(7) ± 0.0220.737(8) ± 0.0180.730(9) ± 0.025
CAL5000.754(8) ± 0.0030.754(9) ± 0.0050.757(7) ± 0.0040.758(1) ± 0.005 •0.757(4) ± 0.0040.757(5) ± 0.0060.758(2) ± 0.0020.757(6) ± 0.0050.757(3) ± 0.004
emotions0.845(9) ± 0.0090.855(8) ± 0.0070.859(7) ± 0.0060.862(4) ± 0.0070.862(5) ± 0.0050.863(3) ± 0.0050.863(1) ± 0.006 •0.863(2) ± 0.0060.861(6) ± 0.005
genbase0.991(9) ± 0.0050.995(8) ± 0.0030.996(3) ± 0.0040.997(1) ± 0.003 •0.996(2) ± 0.0040.996(4) ± 0.0030.995(5) ± 0.0040.995(6) ± 0.0040.995(7) ± 0.004
LLOG0.770(2) ± 0.0020.765(9) ± 0.0020.767(8) ± 0.0020.770(1) ± 0.001 •0.767(4) ± 0.0020.767(5) ± 0.0030.767(6) ± 0.0020.767(3) ± 0.0020.767(7) ± 0.002
enron0.858(7) ± 0.0020.860(1) ± 0.003 •0.860(3) ± 0.0040.859(4) ± 0.0030.860(2) ± 0.0020.859(5) ± 0.0030.858(6) ± 0.0030.857(9) ± 0.0020.857(8) ± 0.003
scene0.931(9) ± 0.0040.941(8) ± 0.0040.948(7) ± 0.0020.949(4) ± 0.0020.950(1) ± 0.002 •0.949(2) ± 0.0020.949(3) ± 0.0020.949(5) ± 0.0020.948(6) ± 0.002
yeast0.827(9) ± 0.0040.833(8) ± 0.0050.840(7) ± 0.0040.842(4) ± 0.0040.842(1) ± 0.004 •0.842(3) ± 0.0040.842(2) ± 0.0030.841(6) ± 0.0030.841(5) ± 0.003
Slashdot0.944(5) ± 0.0040.945(2) ± 0.0030.944(9) ± 0.0040.945(4) ± 0.0040.945(1) ± 0.004 •0.944(8) ± 0.0040.945(3) ± 0.0040.944(6) ± 0.0040.944(7) ± 0.004
corel5k0.785(9) ± 0.0030.787(8) ± 0.0030.791(7) ± 0.0040.794(6) ± 0.0030.796(4) ± 0.0040.796(5) ± 0.0030.798(3) ± 0.0040.799(2) ± 0.0040.801(1) ± 0.004 •
rcv1subset10.872(9) ± 0.0070.875(1) ± 0.005 •0.875(2) ± 0.0090.874(3) ± 0.0090.874(6) ± 0.0090.874(5) ± 0.0060.872(8) ± 0.0090.873(7) ± 0.0070.874(4) ± 0.008
rcv1subset20.870(3) ± 0.0100.874(1) ± 0.011 •0.871(2) ± 0.0100.868(4) ± 0.0090.865(5) ± 0.0100.863(6) ± 0.0110.861(8) ± 0.0100.862(7) ± 0.0090.860(9) ± 0.010
rcv1subset30.858(5) ± 0.0060.861(2) ± 0.0050.862(1) ± 0.007 •0.859(4) ± 0.0100.861(3) ± 0.0070.858(6) ± 0.0100.857(9) ± 0.0080.857(7) ± 0.0080.857(8) ± 0.007
rcv1subset40.886(4) ± 0.0080.888(2) ± 0.0100.888(1) ± 0.008 •0.888(3) ± 0.0100.884(8) ± 0.0100.885(5) ± 0.0090.885(7) ± 0.0080.883(9) ± 0.0090.885(6) ± 0.006
rcv1subset50.869(4) ± 0.0060.876(1) ± 0.007 •0.871(3) ± 0.0060.868(7) ± 0.0080.868(8) ± 0.0090.868(6) ± 0.0080.867(9) ± 0.0070.872(2) ± 0.0070.869(5) ± 0.008
bibtex0.838(9) ± 0.0050.850(8) ± 0.0080.859(6) ± 0.0090.862(3) ± 0.0040.863(2) ± 0.0060.862(4) ± 0.0070.863(1) ± 0.005 •0.860(5) ± 0.0040.859(7) ± 0.006
Arts0.836(9) ± 0.0030.842(8) ± 0.0030.847(7) ± 0.0030.850(6) ± 0.0030.851(5) ± 0.0030.853(4) ± 0.0030.854(2) ± 0.0030.854(3) ± 0.0030.854(1) ± 0.003 •
Health0.910(9) ± 0.0030.914(8) ± 0.0020.917(7) ± 0.0020.919(6) ± 0.0020.919(5) ± 0.0020.919(4) ± 0.0020.919(3) ± 0.0020.920(2) ± 0.0020.920(1) ± 0.002 •
Business0.942(9) ± 0.0010.944(8) ± 0.0010.946(7) ± 0.0010.947(6) ± 0.0010.947(5) ± 0.0010.948(4) ± 0.0010.948(3) ± 0.0010.948(2) ± 0.0010.949(1) ± 0.001 •
Education0.896(9) ± 0.0020.900(8) ± 0.0020.904(7) ± 0.0020.906(6) ± 0.0020.906(5) ± 0.0020.907(4) ± 0.0020.907(3) ± 0.0020.908(2) ± 0.0020.908(1) ± 0.002 •
Computers0.884(9) ± 0.0030.888(8) ± 0.0030.892(7) ± 0.0030.894(6) ± 0.0020.896(5) ± 0.0030.896(4) ± 0.0030.896(3) ± 0.0030.897(1) ± 0.003 •0.896(2) ± 0.003
Entertainment0.876(9) ± 0.0020.881(8) ± 0.0020.887(7) ± 0.0020.890(6) ± 0.0020.891(5) ± 0.0020.892(4) ± 0.0020.893(3) ± 0.0020.893(2) ± 0.0020.894(1) ± 0.002 •
Recreation0.857(9) ± 0.0050.863(8) ± 0.0050.870(7) ± 0.0040.873(6) ± 0.0040.874(5) ± 0.0040.875(4) ± 0.0040.875(3) ± 0.0040.875(2) ± 0.0040.875(1) ± 0.004 •
Society0.858(9) ± 0.0030.862(8) ± 0.0020.866(7) ± 0.0020.868(6) ± 0.0020.869(5) ± 0.0020.870(4) ± 0.0020.870(3) ± 0.0020.871(2) ± 0.0020.871(1) ± 0.002 •
eurlex-dc-l0.906(9) ± 0.0050.913(8) ± 0.0060.917(7) ± 0.0050.920(5) ± 0.0050.919(6) ± 0.0050.920(4) ± 0.0050.921(1) ± 0.004 •0.921(3) ± 0.0040.921(2) ± 0.005
eurlex-sm0.946(9) ± 0.0010.950(8) ± 0.0010.951(7) ± 0.0010.952(5) ± 0.0020.952(6) ± 0.0020.953(1) ± 0.002 •0.953(2) ± 0.0020.952(4) ± 0.0020.952(3) ± 0.002
tmc2007-5000.947(9) ± 0.0010.952(8) ± 0.0010.956(7) ± 0.0010.957(6) ± 0.0010.958(5) ± 0.0010.959(4) ± 0.0010.959(3) ± 0.0010.959(1) ± 0.001 •0.959(2) ± 0.001
mediamill0.955(9) ± 0.0010.957(4) ± 0.0010.958(1) ± 0.001 •0.958(2) ± 0.0010.957(3) ± 0.0010.957(5) ± 0.0010.957(6) ± 0.0010.956(7) ± 0.0000.956(8) ± 0.001
average rank7.546.045.504.434.364.364.114.324.36
win/tie/loss13/59/15243/56/12555/81/8882/89/5384/95/4574/111/3985/106/3386/102/3689/95/40
Table A25. The performance of MLHiKNN across different values of k in terms of Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A25. The performance of MLHiKNN across different values of k in terms of Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetRanking Loss
k = 3k = 5k = 10k = 15k = 20k = 25k = 30k = 35k = 40
birds0.250(2) ± 0.0210.248(1) ± 0.016 •0.253(5) ± 0.0210.257(7) ± 0.0230.251(3) ± 0.0200.251(4) ± 0.0180.255(6) ± 0.0290.261(8) ± 0.0200.270(9) ± 0.029
CAL5000.243(8) ± 0.0030.243(9) ± 0.0050.241(6) ± 0.0040.239(1) ± 0.005 •0.240(4) ± 0.0050.241(5) ± 0.0060.240(2) ± 0.0020.241(7) ± 0.0050.240(3) ± 0.004
emotions0.171(9) ± 0.0100.164(8) ± 0.0100.158(7) ± 0.0090.154(5) ± 0.0110.153(4) ± 0.0070.152(3) ± 0.0050.151(1) ± 0.005 •0.152(2) ± 0.0070.155(6) ± 0.007
genbase0.010(9) ± 0.0050.006(8) ± 0.0030.004(3) ± 0.0030.004(1) ± 0.003 •0.004(2) ± 0.0030.005(4) ± 0.0030.006(6) ± 0.0030.006(7) ± 0.0030.005(5) ± 0.003
LLOG0.186(1) ± 0.002 •0.189(9) ± 0.0030.187(7) ± 0.0020.186(2) ± 0.0020.187(4) ± 0.0020.187(5) ± 0.0020.187(6) ± 0.0020.187(3) ± 0.0020.188(8) ± 0.003
enron0.136(6) ± 0.0040.135(1) ± 0.004 •0.135(4) ± 0.0030.135(3) ± 0.0040.135(2) ± 0.0030.136(5) ± 0.0030.136(7) ± 0.0030.137(9) ± 0.0030.137(8) ± 0.004
scene0.092(9) ± 0.0040.082(8) ± 0.0050.074(7) ± 0.0030.073(6) ± 0.0030.071(2) ± 0.0030.072(3) ± 0.0030.071(1) ± 0.003 •0.072(5) ± 0.0020.072(4) ± 0.003
yeast0.183(9) ± 0.0040.178(8) ± 0.0040.172(7) ± 0.0040.169(6) ± 0.0040.168(3) ± 0.0030.168(4) ± 0.0040.168(1) ± 0.003 •0.168(5) ± 0.0030.168(2) ± 0.004
Slashdot0.048(6) ± 0.0040.047(3) ± 0.0040.048(9) ± 0.0040.048(7) ± 0.0050.047(1) ± 0.005 •0.048(8) ± 0.0040.047(2) ± 0.0050.047(4) ± 0.0040.048(5) ± 0.004
corel5k0.216(9) ± 0.0030.215(8) ± 0.0030.211(7) ± 0.0040.208(6) ± 0.0030.206(4) ± 0.0040.207(5) ± 0.0030.204(3) ± 0.0040.203(2) ± 0.0040.202(1) ± 0.004 •
rcv1subset10.117(4) ± 0.0070.115(1) ± 0.006 •0.116(2) ± 0.0090.117(3) ± 0.0090.118(5) ± 0.0080.118(7) ± 0.0060.119(9) ± 0.0090.119(8) ± 0.0070.118(6) ± 0.008
rcv1subset20.111(2) ± 0.0090.109(1) ± 0.009 •0.113(3) ± 0.0090.116(4) ± 0.0080.118(5) ± 0.0090.119(6) ± 0.0100.122(8) ± 0.0090.121(7) ± 0.0080.123(9) ± 0.009
rcv1subset30.123(4) ± 0.0060.121(2) ± 0.0050.120(1) ± 0.007 •0.125(5) ± 0.0090.122(3) ± 0.0050.125(6) ± 0.0070.125(9) ± 0.0050.125(7) ± 0.0060.125(8) ± 0.006
rcv1subset40.097(3) ± 0.0080.095(1) ± 0.008 •0.096(2) ± 0.0070.097(4) ± 0.0090.100(8) ± 0.0090.099(6) ± 0.0080.100(7) ± 0.0070.101(9) ± 0.0080.099(5) ± 0.006
rcv1subset50.114(4) ± 0.0070.109(1) ± 0.006 •0.114(3) ± 0.0070.117(7) ± 0.0080.117(8) ± 0.0090.117(6) ± 0.0080.118(9) ± 0.0070.114(2) ± 0.0070.116(5) ± 0.007
bibtex0.170(9) ± 0.0050.160(8) ± 0.0070.150(7) ± 0.0070.147(3) ± 0.0040.147(2) ± 0.0040.148(4) ± 0.0050.147(1) ± 0.002 •0.149(5) ± 0.0040.150(6) ± 0.006
Arts0.145(9) ± 0.0030.139(8) ± 0.0030.134(7) ± 0.0030.130(6) ± 0.0030.129(5) ± 0.0030.127(4) ± 0.0030.126(3) ± 0.0030.126(2) ± 0.0030.126(1) ± 0.003 •
Health0.078(9) ± 0.0020.075(8) ± 0.0020.072(7) ± 0.0020.070(6) ± 0.0020.070(5) ± 0.0020.070(4) ± 0.0020.069(3) ± 0.0020.069(2) ± 0.0020.069(1) ± 0.002 •
Business0.040(9) ± 0.0010.039(8) ± 0.0010.037(7) ± 0.0010.037(6) ± 0.0010.037(5) ± 0.0010.036(4) ± 0.0010.036(3) ± 0.0010.036(2) ± 0.0010.036(1) ± 0.001 •
Education0.099(9) ± 0.0010.096(8) ± 0.0020.092(7) ± 0.0020.090(6) ± 0.0020.089(5) ± 0.0020.089(4) ± 0.0020.088(3) ± 0.0020.088(2) ± 0.0020.088(1) ± 0.001 •
Computers0.094(9) ± 0.0020.091(8) ± 0.0020.089(7) ± 0.0020.087(6) ± 0.0020.086(5) ± 0.0020.086(4) ± 0.0020.085(3) ± 0.0020.085(1) ± 0.002 •0.085(2) ± 0.002
Entertainment0.124(9) ± 0.0020.120(8) ± 0.0030.116(7) ± 0.0020.113(6) ± 0.0030.111(5) ± 0.0020.111(4) ± 0.0020.110(3) ± 0.0030.109(2) ± 0.0030.109(1) ± 0.003 •
Recreation0.131(9) ± 0.0040.127(8) ± 0.0040.120(7) ± 0.0040.117(6) ± 0.0030.115(5) ± 0.0040.115(4) ± 0.0040.114(3) ± 0.0040.114(2) ± 0.0040.113(1) ± 0.004 •
Society0.115(9) ± 0.0030.112(8) ± 0.0030.108(7) ± 0.0020.106(6) ± 0.0030.105(5) ± 0.0020.104(4) ± 0.0020.104(3) ± 0.0020.104(2) ± 0.0020.104(1) ± 0.002 •
eurlex-dc-l0.090(9) ± 0.0050.084(8) ± 0.0050.080(7) ± 0.0040.078(5) ± 0.0050.078(6) ± 0.0050.077(4) ± 0.0040.076(1) ± 0.003 •0.076(3) ± 0.0030.076(2) ± 0.004
eurlex-sm0.057(9) ± 0.0020.053(8) ± 0.0010.051(7) ± 0.0020.050(6) ± 0.0020.050(5) ± 0.0020.049(1) ± 0.002 •0.049(3) ± 0.0020.049(4) ± 0.0020.049(2) ± 0.002
tmc2007-5000.055(9) ± 0.0010.050(8) ± 0.0010.047(7) ± 0.0010.045(6) ± 0.0010.044(5) ± 0.0010.044(4) ± 0.0010.043(3) ± 0.0010.043(2) ± 0.0010.043(1) ± 0.001 •
mediamill0.040(9) ± 0.0010.038(6) ± 0.0000.037(1) ± 0.000 •0.038(2) ± 0.0000.038(3) ± 0.0010.038(4) ± 0.0010.038(5) ± 0.0010.038(7) ± 0.0000.039(8) ± 0.000
average rank7.216.115.644.894.254.504.074.324.00
win/tie/loss13/62/14949/46/12965/70/8970/91/6376/100/4873/105/4687/100/3791/106/2794/100/30
Table A26. The performance of MLHiKNN across different values of k in terms of F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A26. The performance of MLHiKNN across different values of k in terms of F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Macro
k = 3k = 5k = 10k = 15k = 20k = 25k = 30k = 35k = 40
birds0.316(2) ± 0.0590.324(1) ± 0.046 •0.284(3) ± 0.0540.264(4) ± 0.0640.258(6) ± 0.0690.261(5) ± 0.0430.236(7) ± 0.0270.226(8) ± 0.0300.223(9) ± 0.041
CAL5000.087(1) ± 0.008 •0.086(2) ± 0.0060.082(4) ± 0.0070.082(5) ± 0.0050.083(3) ± 0.0030.080(7) ± 0.0040.079(9) ± 0.0040.081(6) ± 0.0040.080(8) ± 0.005
emotions0.640(9) ± 0.0230.656(1) ± 0.014 •0.651(4) ± 0.0140.650(8) ± 0.0140.651(6) ± 0.0130.652(3) ± 0.0140.654(2) ± 0.0150.651(5) ± 0.0110.650(7) ± 0.010
genbase0.953(9) ± 0.0120.960(2) ± 0.0110.961(1) ± 0.012 •0.957(4) ± 0.0170.954(8) ± 0.0170.955(7) ± 0.0130.956(5) ± 0.0160.956(6) ± 0.0170.957(3) ± 0.017
LLOG0.115(1) ± 0.005 •0.090(9) ± 0.0180.093(8) ± 0.0170.115(2) ± 0.0050.097(5) ± 0.0160.099(3) ± 0.0150.097(6) ± 0.0130.098(4) ± 0.0160.096(7) ± 0.013
enron0.150(1) ± 0.011 •0.145(2) ± 0.0120.136(3) ± 0.0130.131(4) ± 0.0110.131(5) ± 0.0090.128(6) ± 0.0070.126(7) ± 0.0080.122(8) ± 0.0110.121(9) ± 0.009
scene0.718(9) ± 0.0120.737(8) ± 0.0110.748(3) ± 0.0080.749(1) ± 0.007 •0.749(2) ± 0.0080.746(4) ± 0.0080.746(5) ± 0.0100.742(7) ± 0.0090.742(6) ± 0.007
yeast0.409(8) ± 0.0140.412(6) ± 0.0100.422(2) ± 0.0070.425(1) ± 0.006 •0.420(3) ± 0.0090.415(5) ± 0.0060.415(4) ± 0.0070.412(7) ± 0.0090.406(9) ± 0.012
Slashdot0.176(1) ± 0.034 •0.169(2) ± 0.0200.157(8) ± 0.0230.168(3) ± 0.0330.163(5) ± 0.0250.158(7) ± 0.0300.165(4) ± 0.0290.158(6) ± 0.0260.147(9) ± 0.037
corel5k0.019(9) ± 0.0030.021(8) ± 0.0030.022(7) ± 0.0030.027(5) ± 0.0030.026(6) ± 0.0040.027(4) ± 0.0030.030(3) ± 0.0050.031(2) ± 0.0020.031(1) ± 0.004 •
rcv1subset10.160(2) ± 0.0120.160(3) ± 0.0070.160(1) ± 0.013 •0.156(4) ± 0.0150.152(5) ± 0.0140.152(6) ± 0.0110.148(8) ± 0.0150.147(9) ± 0.0110.151(7) ± 0.013
rcv1subset20.150(2) ± 0.0180.152(1) ± 0.018 •0.143(3) ± 0.0150.139(4) ± 0.0170.133(5) ± 0.0150.128(6) ± 0.0180.124(8) ± 0.0180.125(7) ± 0.0130.122(9) ± 0.017
rcv1subset30.124(3) ± 0.0080.130(1) ± 0.010 •0.125(2) ± 0.0100.117(5) ± 0.0140.120(4) ± 0.0100.112(6) ± 0.0130.108(8) ± 0.0110.110(7) ± 0.0080.108(9) ± 0.010
rcv1subset40.167(1) ± 0.018 •0.164(2) ± 0.0190.156(3) ± 0.0160.154(4) ± 0.0200.149(5) ± 0.0130.147(6) ± 0.0140.143(7) ± 0.0120.141(8) ± 0.0120.140(9) ± 0.008
rcv1subset5 0.150(2) ± 0.0110.154(1) ± 0.008 •0.144(3) ± 0.0100.135(4) ± 0.0160.134(5) ± 0.0170.129(7) ± 0.0120.127(8) ± 0.0120.130(6) ± 0.0160.127(9) ± 0.015
bibtex0.172(9) ± 0.0060.182(7) ± 0.0060.189(1) ± 0.007 •0.189(2) ± 0.0050.187(5) ± 0.0060.188(4) ± 0.0050.188(3) ± 0.0050.186(6) ± 0.0040.182(8) ± 0.006
Arts0.252(9) ± 0.0140.264(3) ± 0.0130.265(2) ± 0.0100.267(1) ± 0.011 •0.263(4) ± 0.0150.262(5) ± 0.0110.261(6) ± 0.0130.255(8) ± 0.0120.256(7) ± 0.011
Health0.389(9) ± 0.0140.396(8) ± 0.0110.403(3) ± 0.0110.405(1) ± 0.011 •0.404(2) ± 0.0110.401(5) ± 0.0110.401(4) ± 0.0140.400(6) ± 0.0130.400(7) ± 0.014
Business0.266(5) ± 0.0130.271(3) ± 0.0080.274(2) ± 0.0100.275(1) ± 0.008 •0.267(4) ± 0.0050.266(7) ± 0.0050.265(8) ± 0.0090.263(9) ± 0.0080.266(6) ± 0.007
Education0.269(9) ± 0.0120.277(2) ± 0.0150.276(3) ± 0.0190.275(4) ± 0.0140.278(1) ± 0.015 •0.274(5) ± 0.0190.272(6) ± 0.0140.271(7) ± 0.0160.271(8) ± 0.015
Computers0.296(4) ± 0.0130.302(1) ± 0.012 •0.301(2) ± 0.0100.297(3) ± 0.0110.294(5) ± 0.0100.294(6) ± 0.0150.290(8) ± 0.0120.292(7) ± 0.0140.288(9) ± 0.012
Entertainment0.377(9) ± 0.0080.391(3) ± 0.0060.395(1) ± 0.008 •0.394(2) ± 0.0090.390(4) ± 0.0080.388(5) ± 0.0080.386(6) ± 0.0120.384(7) ± 0.0070.381(8) ± 0.009
Recreation0.383(6) ± 0.0120.390(4) ± 0.0160.397(2) ± 0.0140.398(1) ± 0.016 •0.390(3) ± 0.0160.388(5) ± 0.0150.382(7) ± 0.0140.380(8) ± 0.0140.377(9) ± 0.012
Society0.272(9) ± 0.0150.286(6) ± 0.0150.293(1) ± 0.012 •0.289(4) ± 0.0140.291(2) ± 0.0120.289(3) ± 0.0090.288(5) ± 0.0100.284(7) ± 0.0090.283(8) ± 0.011
eurlex-dc-l0.285(2) ± 0.0120.288(1) ± 0.011 •0.279(3) ± 0.0110.274(4) ± 0.0100.267(5) ± 0.0100.266(6) ± 0.0080.264(7) ± 0.0080.259(9) ± 0.0070.260(8) ± 0.006
eurlex-sm0.404(2) ± 0.0090.408(1) ± 0.006 •0.397(3) ± 0.0080.395(4) ± 0.0100.388(5) ± 0.0090.387(6) ± 0.0120.385(7) ± 0.0100.380(8) ± 0.0090.377(9) ± 0.008
tmc2007-5000.576(9) ± 0.0070.589(8) ± 0.0060.602(4) ± 0.0050.603(3) ± 0.0070.603(2) ± 0.0070.603(1) ± 0.008 •0.601(5) ± 0.0090.599(6) ± 0.0100.597(7) ± 0.007
mediamill0.326(1) ± 0.006 •0.326(2) ± 0.0080.314(3) ± 0.0070.303(4) ± 0.0030.293(5) ± 0.0050.287(6) ± 0.0060.284(7) ± 0.0060.276(8) ± 0.0040.266(9) ± 0.005
average rank5.113.503.043.294.295.216.076.867.64
win/tie/loss77/71/7698/89/37103/95/26101/95/2863/118/4349/120/5536/113/7520/105/9913/90/121
Table A27. The performance of MLHiKNN across different values of k in terms of F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A27. The performance of MLHiKNN across different values of k in terms of F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Micro
k = 3k = 5k = 10k = 15k = 20k = 25k = 30k = 35k = 40
birds0.402(2) ± 0.0510.412(1) ± 0.036 •0.384(3) ± 0.0440.362(5) ± 0.0610.358(6) ± 0.0610.367(4) ± 0.0370.337(8) ± 0.0360.338(7) ± 0.0310.332(9) ± 0.032
CAL5000.336(3) ± 0.0120.334(7) ± 0.0080.330(9) ± 0.0150.336(5) ± 0.0150.342(1) ± 0.014 •0.336(6) ± 0.0120.333(8) ± 0.0110.337(2) ± 0.0100.336(4) ± 0.015
emotions0.656(9) ± 0.0200.672(4) ± 0.0130.671(6) ± 0.0130.672(5) ± 0.0110.672(3) ± 0.0110.673(2) ± 0.0100.675(1) ± 0.013 •0.669(8) ± 0.0100.670(7) ± 0.007
genbase0.960(7) ± 0.0120.967(2) ± 0.0110.970(1) ± 0.011 •0.965(3) ± 0.0140.965(4) ± 0.0140.965(5) ± 0.0130.962(6) ± 0.0160.958(8) ± 0.0170.955(9) ± 0.020
LLOG0.425(1) ± 0.012 •0.342(9) ± 0.0710.351(8) ± 0.0660.425(2) ± 0.0120.363(6) ± 0.0590.369(3) ± 0.0490.365(5) ± 0.0500.367(4) ± 0.0550.361(7) ± 0.053
enron0.466(1) ± 0.009 •0.466(2) ± 0.0110.463(3) ± 0.0130.461(4) ± 0.0160.460(5) ± 0.0130.456(8) ± 0.0100.455(9) ± 0.0130.459(6) ± 0.0150.457(7) ± 0.013
scene0.710(9) ± 0.0120.731(8) ± 0.0100.742(3) ± 0.0080.743(2) ± 0.0080.743(1) ± 0.008 •0.740(4) ± 0.0080.740(5) ± 0.0100.736(7) ± 0.0090.736(6) ± 0.008
yeast0.631(9) ± 0.0080.638(8) ± 0.0090.643(6) ± 0.0080.643(5) ± 0.0080.643(3) ± 0.0060.642(7) ± 0.0070.644(2) ± 0.0080.645(1) ± 0.008 •0.643(4) ± 0.009
Slashdot0.844(1) ± 0.007 •0.844(2) ± 0.0070.840(8) ± 0.0070.839(9) ± 0.0090.843(4) ± 0.0080.841(7) ± 0.0090.843(3) ± 0.0090.842(5) ± 0.0070.841(6) ± 0.006
corel5k0.037(9) ± 0.0060.038(8) ± 0.0090.040(7) ± 0.0070.050(6) ± 0.0100.050(5) ± 0.0110.055(4) ± 0.0100.058(3) ± 0.0140.064(1) ± 0.008 •0.063(2) ± 0.014
rcv1subset10.313(2) ± 0.0170.314(1) ± 0.012 •0.307(3) ± 0.0170.300(4) ± 0.0200.296(5) ± 0.0210.286(8) ± 0.0230.290(6) ± 0.0230.286(7) ± 0.0210.285(9) ± 0.022
rcv1subset20.356(1) ± 0.011 •0.353(2) ± 0.0130.342(3) ± 0.0130.337(4) ± 0.0140.334(5) ± 0.0120.329(6) ± 0.0090.324(8) ± 0.0150.326(7) ± 0.0090.321(9) ± 0.014
rcv1subset30.349(1) ± 0.009 •0.348(2) ± 0.0140.342(3) ± 0.0140.333(5) ± 0.0160.333(4) ± 0.0090.326(8) ± 0.0120.327(6) ± 0.0090.326(7) ± 0.0120.324(9) ± 0.011
rcv1subset40.425(1) ± 0.017 •0.422(2) ± 0.0140.408(3) ± 0.0200.401(4) ± 0.0230.394(5) ± 0.0190.389(6) ± 0.0190.384(7) ± 0.0150.383(8) ± 0.0120.382(9) ± 0.012
rcv1subset50.364(2) ± 0.0140.369(1) ± 0.014 •0.364(3) ± 0.0130.351(4) ± 0.0130.348(5) ± 0.0140.342(7) ± 0.0200.339(8) ± 0.0140.344(6) ± 0.0250.336(9) ± 0.021
bibtex0.351(9) ± 0.0050.363(8) ± 0.0050.370(1) ± 0.004 •0.369(2) ± 0.0040.367(5) ± 0.0060.368(3) ± 0.0050.367(4) ± 0.0070.366(6) ± 0.0060.363(7) ± 0.006
Arts0.353(9) ± 0.0120.364(7) ± 0.0120.369(3) ± 0.0120.372(1) ± 0.014 •0.369(2) ± 0.0140.368(4) ± 0.0140.367(5) ± 0.0140.364(6) ± 0.0120.363(8) ± 0.012
Health0.557(9) ± 0.0130.567(8) ± 0.0100.576(1) ± 0.009 •0.574(3) ± 0.0120.574(2) ± 0.0110.574(4) ± 0.0100.571(5) ± 0.0110.571(6) ± 0.0110.571(7) ± 0.011
Business0.717(9) ± 0.0030.721(6) ± 0.0030.723(2) ± 0.0030.723(1) ± 0.003 •0.722(3) ± 0.0040.722(4) ± 0.0030.721(5) ± 0.0040.721(7) ± 0.0030.721(8) ± 0.003
Education0.368(9) ± 0.0120.380(3) ± 0.0110.383(1) ± 0.012 •0.382(2) ± 0.0130.378(4) ± 0.0140.378(5) ± 0.0140.374(6) ± 0.0100.372(7) ± 0.0130.370(8) ± 0.013
Computers0.498(9) ± 0.0070.505(8) ± 0.0050.511(4) ± 0.0050.511(3) ± 0.0060.511(1) ± 0.005 •0.511(2) ± 0.0070.510(5) ± 0.0060.510(6) ± 0.0060.509(7) ± 0.006
Entertainment0.483(9) ± 0.0100.495(8) ± 0.0070.503(7) ± 0.0060.505(3) ± 0.0070.505(6) ± 0.0080.506(1) ± 0.006 •0.505(2) ± 0.0070.505(5) ± 0.0070.505(4) ± 0.007
Recreation0.446(9) ± 0.0120.455(3) ± 0.0120.457(1) ± 0.012 •0.456(2) ± 0.0120.454(4) ± 0.0120.453(5) ± 0.0120.451(6) ± 0.0120.449(7) ± 0.0110.449(8) ± 0.011
Society0.450(9) ± 0.0110.461(8) ± 0.0100.469(1) ± 0.009 •0.468(3) ± 0.0100.468(2) ± 0.0100.467(4) ± 0.0090.466(5) ± 0.0100.466(6) ± 0.0110.462(7) ± 0.010
eurlex-dc-l0.493(2) ± 0.0060.494(1) ± 0.005 •0.488(3) ± 0.0060.483(4) ± 0.0070.479(5) ± 0.0070.477(6) ± 0.0040.475(7) ± 0.0060.473(9) ± 0.0060.474(8) ± 0.004
eurlex-sm0.603(2) ± 0.0070.606(1) ± 0.006 •0.598(3) ± 0.0070.594(4) ± 0.0070.589(5) ± 0.0070.586(6) ± 0.0080.584(7) ± 0.0070.582(8) ± 0.0060.580(9) ± 0.006
tmc2007-5000.701(9) ± 0.0030.713(8) ± 0.0030.722(7) ± 0.0030.725(5) ± 0.0030.726(4) ± 0.0030.727(1) ± 0.003 •0.727(2) ± 0.0030.726(3) ± 0.0030.725(6) ± 0.003
mediamill0.644(3) ± 0.0020.647(1) ± 0.002 •0.644(2) ± 0.0020.639(4) ± 0.0020.635(5) ± 0.0010.631(6) ± 0.0020.628(7) ± 0.0020.625(8) ± 0.0020.623(9) ± 0.002
average rank5.544.613.753.713.934.865.396.007.21
win/tie/loss71/44/10990/68/66102/89/3393/96/3575/107/4259/109/5647/111/6633/107/8423/99/102
Table A28. The performance of MLHiKNN across different values of k in terms of Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A28. The performance of MLHiKNN across different values of k in terms of Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetHamming Loss
k = 3k = 5k = 10k = 15k = 20k = 25k = 30k = 35k = 40
birds0.140(9) ± 0.0080.138(4) ± 0.0060.135(1) ± 0.005 •0.138(6) ± 0.0070.137(3) ± 0.0080.139(8) ± 0.0060.139(7) ± 0.0040.137(2) ± 0.0050.138(5) ± 0.004
CAL5000.189(8) ± 0.0010.189(9) ± 0.0020.188(7) ± 0.0020.187(5) ± 0.0010.187(1) ± 0.001 •0.187(3) ± 0.0020.187(2) ± 0.0020.187(4) ± 0.0020.187(6) ± 0.001
emotions0.202(9) ± 0.0100.192(7) ± 0.0060.192(5) ± 0.0070.191(4) ± 0.0060.190(2) ± 0.0050.191(3) ± 0.0040.189(1) ± 0.007 •0.193(8) ± 0.0050.192(6) ± 0.003
genbase0.007(7) ± 0.0020.006(2) ± 0.0020.005(1) ± 0.002 •0.006(3) ± 0.0020.006(4) ± 0.0020.006(5) ± 0.0020.007(6) ± 0.0030.007(8) ± 0.0030.008(9) ± 0.003
LLOG0.185(1) ± 0.004 •0.189(9) ± 0.0060.189(8) ± 0.0060.185(2) ± 0.0040.188(6) ± 0.0060.188(5) ± 0.0050.188(3) ± 0.0060.188(4) ± 0.0050.189(7) ± 0.006
enron0.077(9) ± 0.0010.077(8) ± 0.0010.076(7) ± 0.0010.076(6) ± 0.0010.075(4) ± 0.0010.075(3) ± 0.0010.075(5) ± 0.0010.075(1) ± 0.001 •0.075(2) ± 0.001
scene0.095(9) ± 0.0040.089(8) ± 0.0030.085(3) ± 0.0020.084(2) ± 0.0020.084(1) ± 0.003 •0.085(4) ± 0.0030.085(5) ± 0.0030.086(7) ± 0.0030.086(6) ± 0.002
yeast0.209(9) ± 0.0030.204(8) ± 0.0040.202(7) ± 0.0040.200(6) ± 0.0030.200(4) ± 0.0030.200(5) ± 0.0040.199(2) ± 0.0040.199(1) ± 0.003 •0.200(3) ± 0.005
Slashdot0.028(2) ± 0.0010.028(1) ± 0.001 •0.029(7) ± 0.0010.030(9) ± 0.0020.029(4) ± 0.0010.029(8) ± 0.0020.029(3) ± 0.0020.029(5) ± 0.0010.029(6) ± 0.001
corel5k0.021(8) ± 0.0000.021(9) ± 0.0000.021(7) ± 0.0000.021(6) ± 0.0000.021(5) ± 0.0000.021(4) ± 0.0000.021(2) ± 0.0000.021(3) ± 0.0000.021(1) ± 0.000 •
rcv1subset10.033(9) ± 0.0000.033(8) ± 0.0000.033(7) ± 0.0000.032(6) ± 0.0000.032(3) ± 0.0000.032(2) ± 0.0000.032(5) ± 0.0000.032(1) ± 0.000 •0.032(4) ± 0.000
rcv1subset20.028(9) ± 0.0000.027(8) ± 0.0000.027(6) ± 0.0000.027(1) ± 0.000 •0.027(2) ± 0.0000.027(4) ± 0.0000.027(7) ± 0.0000.027(3) ± 0.0000.027(5) ± 0.000
rcv1subset30.028(9) ± 0.0000.027(8) ± 0.0000.027(7) ± 0.0000.027(3) ± 0.0000.027(5) ± 0.0000.027(6) ± 0.0000.027(4) ± 0.0000.027(1) ± 0.000 •0.027(2) ± 0.000
rcv1subset40.025(9) ± 0.0000.024(4) ± 0.0000.024(2) ± 0.0000.024(1) ± 0.000 •0.024(3) ± 0.0000.024(5) ± 0.0000.024(6) ± 0.0000.024(7) ± 0.0000.024(8) ± 0.000
rcv1subset50.027(9) ± 0.0000.027(4) ± 0.0000.027(1) ± 0.000 •0.027(2) ± 0.0000.027(3) ± 0.0000.027(5) ± 0.0000.027(6) ± 0.0000.027(7) ± 0.0000.027(8) ± 0.000
bibtex0.013(9) ± 0.0000.013(8) ± 0.0000.013(7) ± 0.0000.013(1) ± 0.000 •0.013(2) ± 0.0000.013(3) ± 0.0000.013(4) ± 0.0000.013(5) ± 0.0000.013(6) ± 0.000
Arts0.063(9) ± 0.0010.061(8) ± 0.0010.060(7) ± 0.0010.059(6) ± 0.0010.059(5) ± 0.0010.059(1) ± 0.001 •0.059(2) ± 0.0000.059(3) ± 0.0000.059(4) ± 0.000
Health0.059(9) ± 0.0010.057(8) ± 0.0010.055(7) ± 0.0010.055(6) ± 0.0010.055(2) ± 0.0010.055(1) ± 0.001 •0.055(5) ± 0.0010.055(4) ± 0.0010.055(3) ± 0.001
Business0.032(9) ± 0.0000.031(8) ± 0.0000.031(2) ± 0.0000.031(1) ± 0.000 •0.031(3) ± 0.0000.031(6) ± 0.0000.031(4) ± 0.0000.031(5) ± 0.0000.031(7) ± 0.000
Education0.049(9) ± 0.0010.048(8) ± 0.0000.046(7) ± 0.0010.046(6) ± 0.0010.046(5) ± 0.0010.046(4) ± 0.0010.046(3) ± 0.0010.046(1) ± 0.001 •0.046(2) ± 0.001
Computers0.045(9) ± 0.0010.044(8) ± 0.0000.043(7) ± 0.0000.043(6) ± 0.0000.043(5) ± 0.0000.043(2) ± 0.0010.043(3) ± 0.0010.043(4) ± 0.0010.043(1) ± 0.001 •
Entertainment0.065(9) ± 0.0010.063(8) ± 0.0010.062(7) ± 0.0010.061(6) ± 0.0010.061(5) ± 0.0010.060(4) ± 0.0010.060(3) ± 0.0010.060(2) ± 0.0010.060(1) ± 0.001 •
Recreation0.053(9) ± 0.0010.051(8) ± 0.0010.050(7) ± 0.0010.050(5) ± 0.0010.050(1) ± 0.001 •0.050(3) ± 0.0010.050(2) ± 0.0010.050(6) ± 0.0010.050(4) ± 0.001
Society0.054(9) ± 0.0010.052(8) ± 0.0010.051(7) ± 0.0010.051(6) ± 0.0010.051(5) ± 0.0010.050(3) ± 0.0010.050(2) ± 0.0010.050(1) ± 0.001 •0.051(4) ± 0.001
eurlex-dc-l0.005(9) ± 0.0000.004(3) ± 0.0000.004(1) ± 0.000 •0.004(2) ± 0.0000.004(4) ± 0.0000.005(5) ± 0.0000.005(6) ± 0.0000.005(7) ± 0.0000.005(8) ± 0.000
eurlex-sm0.011(7) ± 0.0000.011(1) ± 0.000 •0.011(2) ± 0.0000.011(3) ± 0.0000.011(4) ± 0.0000.011(5) ± 0.0000.011(6) ± 0.0000.011(8) ± 0.0000.011(9) ± 0.000
tmc2007-5000.056(9) ± 0.0000.054(8) ± 0.0000.052(7) ± 0.0010.051(6) ± 0.0000.051(4) ± 0.0000.051(1) ± 0.000 •0.051(2) ± 0.0000.051(3) ± 0.0010.051(5) ± 0.001
mediamill0.028(6) ± 0.0000.028(2) ± 0.0000.028(1) ± 0.000 •0.028(3) ± 0.0000.028(4) ± 0.0000.028(5) ± 0.0000.028(7) ± 0.0000.028(8) ± 0.0000.029(9) ± 0.000
average rank8.146.545.184.253.544.044.044.255.04
win/tie/loss12/31/18145/48/13175/80/6989/96/3986/114/2473/124/2769/125/3068/125/3159/121/44
Table A29. The performance of MLHiKNN across different values of λ in terms of AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A29. The performance of MLHiKNN across different values of λ in terms of AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Macro
λ = 0.05λ = 0.5λ = 1λ = 2λ = 3λ = 4λ = 5λ = 6λ = 7λ = 8λ → +∞
birds0.723(9) ± 0.0330.722(11) ± 0.0270.722(10) ± 0.0310.727(8) ± 0.0310.728(6) ± 0.0300.730(3) ± 0.0290.729(5) ± 0.0300.729(4) ± 0.0290.730(1) ± 0.029 •0.730(2) ± 0.0290.728(7) ± 0.031
CAL5000.564(10) ± 0.0070.569(9) ± 0.0080.571(4) ± 0.0070.572(1) ± 0.006 •0.571(2) ± 0.0060.571(3) ± 0.0060.571(5) ± 0.0060.570(6) ± 0.0060.570(7) ± 0.0060.570(8) ± 0.0060.561(11) ± 0.006
emotions0.840(11) ± 0.0060.842(10) ± 0.0070.843(9) ± 0.0050.845(8) ± 0.0050.846(6) ± 0.0050.847(4) ± 0.0060.847(3) ± 0.0070.847(5) ± 0.0070.848(2) ± 0.0060.848(1) ± 0.006 •0.845(7) ± 0.009
genbase0.997(10) ± 0.0030.997(1) ± 0.002 •0.997(3) ± 0.0020.997(4) ± 0.0030.997(5) ± 0.0030.997(6) ± 0.0030.997(2) ± 0.0030.997(8) ± 0.0030.997(9) ± 0.0030.997(7) ± 0.0030.996(11) ± 0.004
LLOG0.603(2) ± 0.0060.603(5) ± 0.0060.603(1) ± 0.006 •0.603(3) ± 0.0060.603(7) ± 0.0060.603(10) ± 0.0060.603(4) ± 0.0060.603(8) ± 0.0060.603(6) ± 0.0060.603(11) ± 0.0060.603(9) ± 0.006
enron0.671(10) ± 0.0070.680(1) ± 0.012 •0.679(2) ± 0.0120.676(3) ± 0.0130.676(4) ± 0.0120.675(5) ± 0.0120.674(6) ± 0.0110.674(7) ± 0.0110.673(8) ± 0.0110.673(9) ± 0.0110.663(11) ± 0.011
scene0.942(11) ± 0.0020.942(10) ± 0.0020.943(9) ± 0.0020.944(8) ± 0.0020.944(7) ± 0.0020.945(6) ± 0.0020.945(5) ± 0.0020.945(3) ± 0.0020.945(2) ± 0.0020.945(1) ± 0.002 •0.945(4) ± 0.003
yeast0.690(11) ± 0.0100.692(10) ± 0.0100.698(9) ± 0.0100.709(7) ± 0.0070.711(5) ± 0.0060.712(4) ± 0.0060.712(3) ± 0.0060.712(1) ± 0.006 •0.712(2) ± 0.0060.711(6) ± 0.0070.699(8) ± 0.004
Slashdot0.719(7) ± 0.0210.716(10) ± 0.0190.718(9) ± 0.0210.723(3) ± 0.0180.723(2) ± 0.0210.724(1) ± 0.019 •0.722(4) ± 0.0190.720(5) ± 0.0190.719(6) ± 0.0200.719(8) ± 0.0190.686(11) ± 0.016
corel5k0.651(11) ± 0.0090.671(9) ± 0.0050.683(8) ± 0.0050.689(7) ± 0.0050.690(2) ± 0.0050.690(1) ± 0.005 •0.690(3) ± 0.0050.690(4) ± 0.0050.689(5) ± 0.0040.689(6) ± 0.0040.666(10) ± 0.004
rcv1subset10.877(11) ± 0.0050.892(3) ± 0.0040.893(1) ± 0.004 •0.893(2) ± 0.0040.891(4) ± 0.0040.890(5) ± 0.0040.890(6) ± 0.0040.890(7) ± 0.0040.889(9) ± 0.0040.889(8) ± 0.0030.882(10) ± 0.003
rcv1subset20.873(11) ± 0.0060.889(1) ± 0.005 •0.889(3) ± 0.0050.889(2) ± 0.0050.888(4) ± 0.0050.887(5) ± 0.0050.887(6) ± 0.0050.886(7) ± 0.0050.886(8) ± 0.0050.886(9) ± 0.0050.873(10) ± 0.007
rcv1subset30.868(10) ± 0.0030.880(4) ± 0.0040.880(1) ± 0.004 •0.879(5) ± 0.0030.880(2) ± 0.0040.880(3) ± 0.0040.879(6) ± 0.0040.879(7) ± 0.0040.878(8) ± 0.0040.878(9) ± 0.0040.863(11) ± 0.004
rcv1subset40.877(10) ± 0.0060.889(1) ± 0.006 •0.888(2) ± 0.0060.887(4) ± 0.0070.887(3) ± 0.0070.886(5) ± 0.0080.886(6) ± 0.0070.885(7) ± 0.0080.885(8) ± 0.0080.885(9) ± 0.0080.873(11) ± 0.006
rcv1subset50.864(10) ± 0.0050.875(1) ± 0.005 •0.875(2) ± 0.0050.874(3) ± 0.0050.873(4) ± 0.0050.872(5) ± 0.0050.872(6) ± 0.0050.872(7) ± 0.0050.871(8) ± 0.0050.871(9) ± 0.0050.857(11) ± 0.005
bibtex0.862(11) ± 0.0040.878(9) ± 0.0030.883(8) ± 0.0020.885(7) ± 0.0010.886(6) ± 0.0010.886(5) ± 0.0010.887(4) ± 0.0010.887(3) ± 0.0010.887(2) ± 0.0010.887(1) ± 0.001 •0.878(10) ± 0.002
Arts0.725(10) ± 0.0080.729(7) ± 0.0100.731(2) ± 0.0080.731(1) ± 0.009 •0.730(3) ± 0.0090.730(4) ± 0.0100.729(5) ± 0.0090.729(6) ± 0.0090.728(8) ± 0.0100.728(9) ± 0.0090.701(11) ± 0.009
Health0.771(10) ± 0.0100.780(1) ± 0.010 •0.780(2) ± 0.0090.779(3) ± 0.0100.778(6) ± 0.0090.778(4) ± 0.0100.778(5) ± 0.0100.777(7) ± 0.0100.777(8) ± 0.0090.776(9) ± 0.0100.749(11) ± 0.007
Business0.732(10) ± 0.0180.747(9) ± 0.0100.752(4) ± 0.0060.753(1) ± 0.008 •0.752(3) ± 0.0070.753(2) ± 0.0070.752(5) ± 0.0080.751(6) ± 0.0080.751(7) ± 0.0080.751(8) ± 0.0080.709(11) ± 0.007
Education0.736(10) ± 0.0110.746(5) ± 0.0120.747(3) ± 0.0120.748(1) ± 0.012 •0.748(2) ± 0.0140.747(4) ± 0.0140.745(6) ± 0.0130.745(7) ± 0.0130.743(8) ± 0.0140.743(9) ± 0.0130.711(11) ± 0.018
Computers0.743(10) ± 0.0070.748(9) ± 0.0070.750(3) ± 0.0060.750(2) ± 0.0060.750(1) ± 0.005 •0.750(4) ± 0.0060.749(7) ± 0.0060.749(5) ± 0.0060.749(6) ± 0.0060.748(8) ± 0.0060.716(11) ± 0.008
Entertainment0.775(10) ± 0.0090.781(3) ± 0.0060.781(5) ± 0.0050.782(1) ± 0.005 •0.782(2) ± 0.0040.781(4) ± 0.0040.780(6) ± 0.0040.780(7) ± 0.0030.780(8) ± 0.0030.779(9) ± 0.0040.757(11) ± 0.005
Recreation0.786(10) ± 0.0080.786(6) ± 0.0070.788(3) ± 0.0080.789(1) ± 0.008 •0.788(2) ± 0.0090.788(4) ± 0.0090.787(5) ± 0.0090.786(7) ± 0.0080.786(8) ± 0.0080.786(9) ± 0.0080.756(11) ± 0.009
Society0.704(10) ± 0.0040.705(9) ± 0.0040.706(7) ± 0.0050.706(4) ± 0.0060.707(1) ± 0.006 •0.706(5) ± 0.0060.706(8) ± 0.0060.706(3) ± 0.0070.707(2) ± 0.0070.706(6) ± 0.0070.685(11) ± 0.007
eurlex-dc-l0.874(11) ± 0.0030.881(10) ± 0.0060.892(8) ± 0.0060.899(7) ± 0.0050.900(5) ± 0.0040.900(3) ± 0.0050.900(4) ± 0.0050.900(2) ± 0.0050.901(1) ± 0.005 •0.900(6) ± 0.0050.883(9) ± 0.007
eurlex-sm0.896(11) ± 0.0040.900(9) ± 0.0040.905(8) ± 0.0030.908(7) ± 0.0030.908(6) ± 0.0030.908(2) ± 0.0030.909(1) ± 0.003 •0.908(3) ± 0.0030.908(5) ± 0.0030.908(4) ± 0.0030.899(10) ± 0.004
tmc2007-5000.907(11) ± 0.0020.913(10) ± 0.0020.917(8) ± 0.0020.919(7) ± 0.0020.920(6) ± 0.0020.921(5) ± 0.0020.922(4) ± 0.0020.922(3) ± 0.0020.923(1) ± 0.002 •0.923(2) ± 0.0020.914(9) ± 0.002
mediamill0.791(11) ± 0.0030.803(10) ± 0.0030.812(9) ± 0.0030.822(7) ± 0.0020.824(6) ± 0.0020.825(5) ± 0.0020.826(4) ± 0.0020.826(3) ± 0.0020.827(2) ± 0.0020.827(1) ± 0.002 •0.820(8) ± 0.003
average rank9.966.545.114.184.004.184.795.295.546.579.86
win/tie/loss13/64/20369/124/87106/103/71129/96/55124/119/37124/124/32109/122/4996/125/5992/114/7486/109/8516/52/212
Table A30. The performance of MLHiKNN across different values of λ in terms of AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A30. The performance of MLHiKNN across different values of λ in terms of AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Micr
λ = 0.05λ = 0.5λ = 1λ = 2λ = 3λ = 4λ = 5λ = 6λ = 7λ = 8λ → +∞
birds0.743(10) ± 0.0250.747(7) ± 0.0140.746(8) ± 0.0190.747(6) ± 0.0190.746(9) ± 0.0190.751(1) ± 0.017 •0.749(3) ± 0.0170.749(5) ± 0.0150.750(2) ± 0.0160.749(4) ± 0.0140.737(11) ± 0.022
CAL5000.722(11) ± 0.0100.745(10) ± 0.0040.754(8) ± 0.0030.759(1) ± 0.004 •0.757(3) ± 0.0040.758(2) ± 0.0040.757(4) ± 0.0040.756(7) ± 0.0050.756(6) ± 0.0050.757(5) ± 0.0050.751(9) ± 0.007
emotions0.854(11) ± 0.0060.856(10) ± 0.0060.859(9) ± 0.0050.860(8) ± 0.0060.862(6) ± 0.0050.862(5) ± 0.0050.863(2) ± 0.0060.862(7) ± 0.0060.862(4) ± 0.0060.863(1) ± 0.006 •0.862(3) ± 0.005
genbase0.996(7) ± 0.0040.996(10) ± 0.0030.996(9) ± 0.0030.996(4) ± 0.0040.996(2) ± 0.0040.997(1) ± 0.004 •0.996(3) ± 0.0040.996(6) ± 0.0040.996(5) ± 0.0040.996(8) ± 0.0040.995(11) ± 0.004
LLOG0.767(9) ± 0.0020.767(8) ± 0.0020.767(10) ± 0.0020.767(1) ± 0.002 •0.767(2) ± 0.0020.767(4) ± 0.0020.767(6) ± 0.0030.767(7) ± 0.0030.767(5) ± 0.0020.767(3) ± 0.0020.764(11) ± 0.005
enron0.847(11) ± 0.0050.859(10) ± 0.0030.860(5) ± 0.0020.860(1) ± 0.002 •0.860(2) ± 0.0020.860(3) ± 0.0020.860(4) ± 0.0020.860(6) ± 0.0020.860(7) ± 0.0020.859(8) ± 0.0020.859(9) ± 0.003
scene0.948(11) ± 0.0020.949(10) ± 0.0010.949(9) ± 0.0020.950(8) ± 0.0020.950(7) ± 0.0020.950(5) ± 0.0020.951(4) ± 0.0020.951(2) ± 0.0020.951(3) ± 0.0020.951(1) ± 0.002 •0.950(6) ± 0.003
yeast0.829(11) ± 0.0050.830(10) ± 0.0050.833(9) ± 0.0050.839(8) ± 0.0040.842(6) ± 0.0040.844(5) ± 0.0030.844(4) ± 0.0030.845(1) ± 0.004 •0.845(2) ± 0.0030.844(3) ± 0.0040.841(7) ± 0.004
Slashdot0.944(8) ± 0.0030.944(10) ± 0.0030.944(9) ± 0.0030.945(1) ± 0.004 •0.945(2) ± 0.0040.945(4) ± 0.0040.945(3) ± 0.0040.944(7) ± 0.0040.945(5) ± 0.0040.945(6) ± 0.0040.938(11) ± 0.006
corel5k0.781(11) ± 0.0050.799(1) ± 0.003 •0.797(2) ± 0.0040.796(5) ± 0.0040.796(3) ± 0.0040.796(4) ± 0.0040.795(7) ± 0.0030.795(6) ± 0.0030.795(8) ± 0.0030.795(9) ± 0.0030.791(10) ± 0.004
rcv1subset10.905(1) ± 0.003 •0.896(2) ± 0.0050.887(3) ± 0.0090.876(4) ± 0.0090.874(5) ± 0.0090.871(6) ± 0.0090.871(7) ± 0.0070.869(8) ± 0.0080.869(9) ± 0.0080.869(10) ± 0.0070.859(11) ± 0.007
rcv1subset20.906(1) ± 0.004 •0.894(2) ± 0.0080.878(3) ± 0.0120.869(4) ± 0.0120.865(5) ± 0.0100.861(7) ± 0.0080.861(6) ± 0.0080.860(8) ± 0.0080.859(10) ± 0.0080.859(9) ± 0.0070.843(11) ± 0.008
rcv1subset30.897(1) ± 0.005 •0.882(2) ± 0.0060.871(3) ± 0.0090.863(4) ± 0.0070.861(5) ± 0.0070.858(6) ± 0.0070.855(7) ± 0.0080.855(8) ± 0.0080.853(9) ± 0.0080.851(10) ± 0.0080.835(11) ± 0.006
rcv1subset40.917(1) ± 0.003 •0.908(2) ± 0.0080.897(3) ± 0.0090.888(4) ± 0.0090.884(5) ± 0.0100.884(6) ± 0.0080.883(7) ± 0.0080.883(8) ± 0.0080.883(10) ± 0.0070.883(9) ± 0.0070.866(11) ± 0.007
rcv1subset50.902(1) ± 0.003 •0.885(2) ± 0.0070.877(3) ± 0.0060.870(4) ± 0.0070.868(5) ± 0.0090.867(6) ± 0.0090.866(7) ± 0.0080.864(8) ± 0.0070.864(9) ± 0.0080.863(10) ± 0.0080.847(11) ± 0.007
bibtex0.869(3) ± 0.0040.877(1) ± 0.004 •0.872(2) ± 0.0050.866(4) ± 0.0060.863(5) ± 0.0060.860(6) ± 0.0040.859(7) ± 0.0040.858(8) ± 0.0040.856(9) ± 0.0040.856(10) ± 0.0040.820(11) ± 0.008
Arts0.849(10) ± 0.0030.850(9) ± 0.0040.850(8) ± 0.0040.852(1) ± 0.003 •0.851(2) ± 0.0030.851(5) ± 0.0040.851(4) ± 0.0040.851(6) ± 0.0040.851(3) ± 0.0030.851(7) ± 0.0040.838(11) ± 0.003
Health0.916(10) ± 0.0020.917(9) ± 0.0020.918(8) ± 0.0020.919(7) ± 0.0020.919(5) ± 0.0020.919(3) ± 0.0020.919(1) ± 0.002 •0.919(2) ± 0.0020.919(4) ± 0.0020.919(6) ± 0.0020.912(11) ± 0.002
Business0.946(10) ± 0.0020.947(9) ± 0.0010.948(1) ± 0.001 •0.947(5) ± 0.0010.947(3) ± 0.0010.947(4) ± 0.0000.947(2) ± 0.0010.947(6) ± 0.0010.947(8) ± 0.0010.947(7) ± 0.0010.939(11) ± 0.002
Education0.905(10) ± 0.0020.906(9) ± 0.0020.906(4) ± 0.0020.907(1) ± 0.002 •0.906(3) ± 0.0020.906(2) ± 0.0020.906(6) ± 0.0020.906(5) ± 0.0020.906(7) ± 0.0020.906(8) ± 0.0020.899(11) ± 0.002
Computers0.892(10) ± 0.0040.893(9) ± 0.0040.895(8) ± 0.0030.896(2) ± 0.0030.896(1) ± 0.003 •0.895(4) ± 0.0030.895(5) ± 0.0030.895(3) ± 0.0030.895(6) ± 0.0020.895(7) ± 0.0020.884(11) ± 0.005
Entertainment0.889(10) ± 0.0020.890(9) ± 0.0020.891(7) ± 0.0020.891(3) ± 0.0020.891(1) ± 0.002 •0.891(2) ± 0.0020.891(4) ± 0.0010.891(5) ± 0.0010.891(6) ± 0.0010.891(8) ± 0.0010.880(11) ± 0.001
Recreation0.873(7) ± 0.0040.874(4) ± 0.0040.875(2) ± 0.0040.875(1) ± 0.004 •0.874(3) ± 0.0040.874(5) ± 0.0040.873(6) ± 0.0040.873(8) ± 0.0040.873(9) ± 0.0040.872(10) ± 0.0040.854(11) ± 0.004
Society0.864(10) ± 0.0040.867(9) ± 0.0020.868(8) ± 0.0020.869(5) ± 0.0020.869(1) ± 0.002 •0.869(4) ± 0.0020.869(3) ± 0.0020.869(2) ± 0.0020.869(6) ± 0.0020.869(7) ± 0.0020.861(11) ± 0.002
eurlex-dc-l0.915(10) ± 0.0040.916(9) ± 0.0050.918(8) ± 0.0060.920(1) ± 0.005 •0.919(2) ± 0.0050.919(3) ± 0.0060.919(7) ± 0.0060.919(6) ± 0.0060.919(4) ± 0.0060.919(5) ± 0.0060.904(11) ± 0.005
eurlex-sm0.945(11) ± 0.0020.947(10) ± 0.0020.950(8) ± 0.0020.952(7) ± 0.0020.952(6) ± 0.0020.952(5) ± 0.0020.953(4) ± 0.0020.953(3) ± 0.0020.953(2) ± 0.0010.953(1) ± 0.001 •0.948(9) ± 0.001
tmc2007-5000.949(11) ± 0.0010.954(10) ± 0.0010.956(9) ± 0.0010.957(7) ± 0.0010.958(6) ± 0.0010.959(5) ± 0.0010.959(4) ± 0.0010.959(3) ± 0.0010.959(2) ± 0.0010.959(1) ± 0.001 •0.956(8) ± 0.001
mediamill0.948(11) ± 0.0010.951(10) ± 0.0010.954(9) ± 0.0010.956(8) ± 0.0010.957(7) ± 0.0010.958(5) ± 0.0010.958(4) ± 0.0010.958(3) ± 0.0010.958(2) ± 0.0010.958(1) ± 0.001 •0.958(6) ± 0.001
average rank8.147.256.254.114.004.214.685.505.796.219.86
win/tie/loss70/43/16795/68/117105/93/82111/108/61110/123/4799/136/4597/130/5390/134/5686/133/6183/132/6521/46/213
Table A31. The performance of MLHiKNN across different values of λ in terms of Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A31. The performance of MLHiKNN across different values of λ in terms of Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetRanking Loss
λ = 0.05λ = 0.5λ = 1λ = 2λ = 3λ = 4λ = 5λ = 6λ = 7λ = 8λ → +∞
birds0.253(9) ± 0.0290.252(7) ± 0.0200.254(10) ± 0.0250.251(4) ± 0.0210.251(5) ± 0.0200.248(1) ± 0.020 •0.250(3) ± 0.0230.251(6) ± 0.0210.250(2) ± 0.0200.253(8) ± 0.0180.269(11) ± 0.025
CAL5000.275(11) ± 0.0100.253(10) ± 0.0050.243(8) ± 0.0030.238(1) ± 0.004 •0.240(3) ± 0.0050.240(2) ± 0.0040.240(4) ± 0.0040.241(7) ± 0.0050.241(6) ± 0.0050.241(5) ± 0.0050.246(9) ± 0.008
emotions0.160(11) ± 0.0070.157(10) ± 0.0070.155(9) ± 0.0060.155(8) ± 0.0100.153(3) ± 0.0070.153(2) ± 0.0070.152(1) ± 0.008 •0.154(6) ± 0.0100.153(5) ± 0.0100.153(4) ± 0.0100.154(7) ± 0.007
genbase0.006(11) ± 0.0030.006(10) ± 0.0040.005(8) ± 0.0030.004(7) ± 0.0030.004(2) ± 0.0030.004(1) ± 0.003 •0.004(3) ± 0.0030.004(4) ± 0.0030.004(5) ± 0.0030.004(6) ± 0.0030.005(9) ± 0.003
LLOG0.187(5) ± 0.0030.187(6) ± 0.0020.187(8) ± 0.0020.187(1) ± 0.002 •0.187(2) ± 0.0020.187(4) ± 0.0030.187(7) ± 0.0020.187(10) ± 0.0030.187(9) ± 0.0020.187(3) ± 0.0020.190(11) ± 0.005
enron0.143(11) ± 0.0050.135(4) ± 0.0030.134(2) ± 0.0030.134(1) ± 0.003 •0.135(3) ± 0.0030.135(5) ± 0.0030.135(6) ± 0.0030.136(7) ± 0.0030.136(9) ± 0.0030.136(8) ± 0.0030.136(10) ± 0.004
scene0.074(11) ± 0.0020.073(10) ± 0.0030.073(9) ± 0.0030.072(8) ± 0.0040.071(7) ± 0.0030.071(5) ± 0.0030.070(1) ± 0.003 •0.071(3) ± 0.0030.071(4) ± 0.0030.070(2) ± 0.0040.071(6) ± 0.004
yeast0.180(11) ± 0.0040.180(10) ± 0.0040.177(9) ± 0.0040.171(8) ± 0.0040.168(7) ± 0.0030.167(5) ± 0.0030.166(4) ± 0.0030.166(1) ± 0.003 •0.166(2) ± 0.0030.166(3) ± 0.0030.168(6) ± 0.003
Slashdot0.048(8) ± 0.0030.048(10) ± 0.0030.048(9) ± 0.0040.047(1) ± 0.005 •0.047(2) ± 0.0050.047(4) ± 0.0050.047(3) ± 0.0040.048(7) ± 0.0040.047(5) ± 0.0040.047(6) ± 0.0040.051(11) ± 0.005
corel5k0.218(11) ± 0.0050.204(1) ± 0.003 •0.205(2) ± 0.0040.206(5) ± 0.0040.206(3) ± 0.0040.206(4) ± 0.0040.207(7) ± 0.0040.207(6) ± 0.0030.207(8) ± 0.0030.207(9) ± 0.0030.211(10) ± 0.004
rcv1subset10.087(1) ± 0.003 •0.097(2) ± 0.0050.105(3) ± 0.0080.115(4) ± 0.0090.118(5) ± 0.0080.120(7) ± 0.0080.120(6) ± 0.0070.122(8) ± 0.0080.122(9) ± 0.0070.122(10) ± 0.0060.132(11) ± 0.007
rcv1subset20.081(1) ± 0.004 •0.093(2) ± 0.0070.107(3) ± 0.0110.114(4) ± 0.0120.118(5) ± 0.0090.122(7) ± 0.0080.122(6) ± 0.0090.123(8) ± 0.0080.124(10) ± 0.0080.124(9) ± 0.0070.139(11) ± 0.007
rcv1subset30.089(1) ± 0.003 •0.103(2) ± 0.0050.112(3) ± 0.0070.120(4) ± 0.0050.122(5) ± 0.0050.125(6) ± 0.0050.127(7) ± 0.0060.127(8) ± 0.0060.129(9) ± 0.0060.130(10) ± 0.0060.145(11) ± 0.004
rcv1subset40.071(1) ± 0.002 •0.081(2) ± 0.0070.090(3) ± 0.0080.097(4) ± 0.0090.100(7) ± 0.0090.100(6) ± 0.0070.100(5) ± 0.0070.100(9) ± 0.0070.100(8) ± 0.0070.101(10) ± 0.0060.113(11) ± 0.006
rcv1subset5 0.087(1) ± 0.002 •0.102(2) ± 0.0060.109(3) ± 0.0060.116(4) ± 0.0070.117(5) ± 0.0090.118(6) ± 0.0090.118(7) ± 0.0080.120(8) ± 0.0070.120(9) ± 0.0080.121(10) ± 0.0080.135(11) ± 0.008
bibtex0.136(2) ± 0.0040.133(1) ± 0.003 •0.139(3) ± 0.0030.145(4) ± 0.0040.147(5) ± 0.0040.150(6) ± 0.0040.150(7) ± 0.0050.151(8) ± 0.0040.152(9) ± 0.0040.152(10) ± 0.0040.184(11) ± 0.006
Arts0.131(10) ± 0.0030.131(9) ± 0.0040.130(8) ± 0.0040.129(1) ± 0.003 •0.129(3) ± 0.0030.130(7) ± 0.0030.129(4) ± 0.0030.130(6) ± 0.0030.129(2) ± 0.0030.129(5) ± 0.0030.140(11) ± 0.003
Health0.072(10) ± 0.0020.071(9) ± 0.0020.070(8) ± 0.0020.070(6) ± 0.0020.070(4) ± 0.0020.070(3) ± 0.0020.070(1) ± 0.002 •0.070(2) ± 0.0020.070(5) ± 0.0020.070(7) ± 0.0020.077(11) ± 0.001
Business0.037(10) ± 0.0010.037(6) ± 0.0010.036(1) ± 0.001 •0.037(5) ± 0.0010.037(3) ± 0.0010.037(4) ± 0.0010.037(2) ± 0.0010.037(7) ± 0.0010.037(8) ± 0.0010.037(9) ± 0.0010.042(11) ± 0.001
Education0.091(10) ± 0.0020.090(9) ± 0.0020.090(6) ± 0.0020.089(2) ± 0.0020.089(3) ± 0.0020.089(1) ± 0.002 •0.090(5) ± 0.0020.090(4) ± 0.0020.090(7) ± 0.0020.090(8) ± 0.0020.095(11) ± 0.002
Computers0.089(10) ± 0.0020.087(9) ± 0.0020.087(8) ± 0.0020.086(7) ± 0.0020.086(2) ± 0.0020.086(3) ± 0.0020.086(6) ± 0.0020.086(1) ± 0.002 •0.086(4) ± 0.0020.086(5) ± 0.0020.093(11) ± 0.003
Entertainment0.114(10) ± 0.0030.113(9) ± 0.0030.112(8) ± 0.0030.112(3) ± 0.0030.111(2) ± 0.0020.111(1) ± 0.002 •0.112(4) ± 0.0020.112(5) ± 0.0020.112(6) ± 0.0020.112(7) ± 0.0020.119(11) ± 0.002
Recreation0.116(10) ± 0.0040.116(6) ± 0.0030.115(3) ± 0.0040.115(1) ± 0.004 •0.115(2) ± 0.0040.115(4) ± 0.0040.116(5) ± 0.0040.116(7) ± 0.0040.116(8) ± 0.0040.116(9) ± 0.0040.129(11) ± 0.004
Society0.110(10) ± 0.0040.107(9) ± 0.0030.106(8) ± 0.0030.105(4) ± 0.0020.105(1) ± 0.002 •0.106(5) ± 0.0020.105(2) ± 0.0020.105(3) ± 0.0020.106(6) ± 0.0020.106(7) ± 0.0020.112(11) ± 0.003
eurlex-dc-l0.081(10) ± 0.0040.081(9) ± 0.0040.079(8) ± 0.0050.077(1) ± 0.004 •0.078(3) ± 0.0050.078(2) ± 0.0050.078(5) ± 0.0050.078(6) ± 0.0050.078(4) ± 0.0050.078(7) ± 0.0050.092(11) ± 0.004
eurlex-sm0.056(11) ± 0.0020.054(9) ± 0.0020.052(8) ± 0.0020.050(7) ± 0.0020.050(6) ± 0.0020.050(5) ± 0.0020.049(3) ± 0.0020.050(4) ± 0.0020.049(2) ± 0.0020.049(1) ± 0.002 •0.055(10) ± 0.001
tmc2007-5000.050(11) ± 0.0010.047(9) ± 0.0010.046(8) ± 0.0010.045(7) ± 0.0010.044(6) ± 0.0010.044(5) ± 0.0010.044(4) ± 0.0010.044(3) ± 0.0010.044(2) ± 0.0010.044(1) ± 0.001 •0.048(10) ± 0.001
mediamill0.044(11) ± 0.0010.043(10) ± 0.0010.041(9) ± 0.0010.038(8) ± 0.0010.038(7) ± 0.0010.037(6) ± 0.0010.037(4) ± 0.0010.037(3) ± 0.0000.037(2) ± 0.0000.037(1) ± 0.000 •0.037(5) ± 0.001
average rank8.186.866.254.293.964.184.365.615.896.4310.0
win/tie/loss69/51/16088/78/114108/100/72112/118/5099/137/4492/141/4785/152/4377/153/5076/153/5173/153/5415/56/209
Table A32. The performance of MLHiKNN across different values of λ in terms of F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A32. The performance of MLHiKNN across different values of λ in terms of F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Macro
λ = 0.05λ = 0.5λ = 1λ = 2λ = 3λ = 4λ = 5λ = 6λ = 7λ = 8λ → +∞
birds0.348(1) ± 0.043 •0.299(2) ± 0.0280.271(4) ± 0.0530.269(5) ± 0.0580.258(10) ± 0.0690.271(3) ± 0.0640.263(8) ± 0.0640.265(7) ± 0.0660.265(6) ± 0.0650.259(9) ± 0.0600.217(11) ± 0.074
CAL5000.129(1) ± 0.006 •0.105(2) ± 0.0060.093(3) ± 0.0030.085(4) ± 0.0030.083(5) ± 0.0030.081(6) ± 0.0030.079(7) ± 0.0040.077(8) ± 0.0040.077(9) ± 0.0040.076(10) ± 0.0040.071(11) ± 0.004
emotions0.643(8) ± 0.0170.643(10) ± 0.0180.643(9) ± 0.0170.649(5) ± 0.0140.651(1) ± 0.013 •0.649(3) ± 0.0140.647(7) ± 0.0170.648(6) ± 0.0150.649(4) ± 0.0140.650(2) ± 0.0150.641(11) ± 0.012
genbase0.927(11) ± 0.0180.939(10) ± 0.0160.942(9) ± 0.0130.951(8) ± 0.0180.954(7) ± 0.0170.955(6) ± 0.0170.958(5) ± 0.0160.962(4) ± 0.0110.962(3) ± 0.0100.962(2) ± 0.0100.969(1) ± 0.014 •
LLOG0.094(8) ± 0.0130.094(9) ± 0.0130.093(10) ± 0.0140.096(2) ± 0.0140.097(1) ± 0.016 •0.095(4) ± 0.0180.095(3) ± 0.0180.095(6) ± 0.0190.095(6) ± 0.0190.095(6) ± 0.0190.091(11) ± 0.016
enron0.163(1) ± 0.008 •0.140(2) ± 0.0080.134(3) ± 0.0060.133(4) ± 0.0070.131(5) ± 0.0090.129(6) ± 0.0100.129(8) ± 0.0090.128(10) ± 0.0090.128(9) ± 0.0100.129(7) ± 0.0100.125(11) ± 0.014
scene0.744(11) ± 0.0060.748(8) ± 0.0060.748(10) ± 0.0070.749(6) ± 0.0070.749(5) ± 0.0080.749(7) ± 0.0080.750(4) ± 0.0080.751(2) ± 0.0080.750(3) ± 0.0080.751(1) ± 0.010 •0.748(9) ± 0.009
yeast0.435(2) ± 0.0080.435(1) ± 0.009 •0.432(3) ± 0.0100.426(4) ± 0.0090.420(5) ± 0.0090.416(6) ± 0.0090.415(7) ± 0.0070.409(8) ± 0.0100.408(9) ± 0.0090.405(10) ± 0.0080.371(11) ± 0.010
Slashdot0.144(11) ± 0.0190.148(10) ± 0.0210.154(9) ± 0.0260.163(6) ± 0.0290.163(7) ± 0.0250.164(4) ± 0.0260.163(5) ± 0.0240.166(2) ± 0.0230.168(1) ± 0.025 •0.166(3) ± 0.0230.156(8) ± 0.028
corel5k0.064(1) ± 0.006 •0.031(2) ± 0.0030.028(3) ± 0.0050.026(4) ± 0.0030.026(5) ± 0.0040.026(6) ± 0.0040.025(9) ± 0.0030.025(8) ± 0.0040.025(7) ± 0.0040.025(10) ± 0.0030.020(11) ± 0.003
rcv1subset10.260(1) ± 0.005 •0.203(2) ± 0.0130.179(3) ± 0.0170.158(4) ± 0.0170.152(5) ± 0.0140.149(6) ± 0.0130.149(7) ± 0.0130.147(8) ± 0.0140.145(10) ± 0.0140.146(9) ± 0.0140.122(11) ± 0.011
rcv1subset20.259(1) ± 0.007 •0.188(2) ± 0.0170.154(3) ± 0.0200.139(4) ± 0.0210.133(5) ± 0.0150.127(7) ± 0.0130.127(6) ± 0.0120.125(8) ± 0.0130.122(10) ± 0.0130.123(9) ± 0.0120.100(11) ± 0.009
rcv1subset30.215(1) ± 0.013 •0.149(2) ± 0.0090.132(3) ± 0.0130.120(4) ± 0.0110.120(5) ± 0.0100.116(6) ± 0.0100.111(7) ± 0.0100.110(8) ± 0.0080.108(9) ± 0.0110.106(10) ± 0.0100.085(11) ± 0.007
rcv1subset40.249(1) ± 0.011 •0.192(2) ± 0.0140.167(3) ± 0.0140.153(4) ± 0.0130.149(5) ± 0.0130.146(6) ± 0.0100.145(7) ± 0.0100.143(8) ± 0.0100.142(9) ± 0.0100.142(10) ± 0.0110.109(11) ± 0.008
rcv1subset50.237(1) ± 0.008 •0.166(2) ± 0.0130.150(3) ± 0.0110.139(4) ± 0.0140.134(5) ± 0.0170.131(6) ± 0.0180.130(7) ± 0.0180.126(8) ± 0.0140.126(9) ± 0.0140.125(10) ± 0.0140.096(11) ± 0.013
bibtex0.234(1) ± 0.006 •0.211(2) ± 0.0070.201(3) ± 0.0070.191(4) ± 0.0060.187(5) ± 0.0060.184(6) ± 0.0040.182(7) ± 0.0040.180(8) ± 0.0030.179(9) ± 0.0040.178(10) ± 0.0040.139(11) ± 0.006
Arts0.292(1) ± 0.011 •0.280(2) ± 0.0140.275(3) ± 0.0120.268(4) ± 0.0140.263(5) ± 0.0150.258(6) ± 0.0170.255(7) ± 0.0170.255(8) ± 0.0160.253(9) ± 0.0140.251(10) ± 0.0160.193(11) ± 0.014
Health0.414(1) ± 0.011 •0.412(2) ± 0.0130.409(3) ± 0.0120.406(4) ± 0.0100.404(5) ± 0.0110.404(6) ± 0.0110.402(7) ± 0.0110.402(8) ± 0.0110.398(9) ± 0.0090.398(10) ± 0.0090.353(11) ± 0.012
Business0.303(1) ± 0.010 •0.293(2) ± 0.0090.282(3) ± 0.0110.274(4) ± 0.0070.267(5) ± 0.0050.263(6) ± 0.0060.262(7) ± 0.0060.260(8) ± 0.0080.259(9) ± 0.0090.259(10) ± 0.0100.198(11) ± 0.014
Education0.305(1) ± 0.012 •0.302(2) ± 0.0150.293(3) ± 0.0170.283(4) ± 0.0170.278(5) ± 0.0150.276(6) ± 0.0140.271(7) ± 0.0100.270(8) ± 0.0120.267(9) ± 0.0120.266(10) ± 0.0110.210(11) ± 0.016
Computers0.322(1) ± 0.015 •0.321(2) ± 0.0140.312(3) ± 0.0110.300(4) ± 0.0130.294(5) ± 0.0100.291(6) ± 0.0120.290(7) ± 0.0120.288(8) ± 0.0110.285(9) ± 0.0100.284(10) ± 0.0090.228(11) ± 0.018
Entertainment0.398(3) ± 0.0110.400(1) ± 0.013 •0.398(2) ± 0.0090.392(4) ± 0.0090.390(5) ± 0.0080.390(6) ± 0.0070.388(7) ± 0.0080.387(8) ± 0.0080.387(9) ± 0.0090.386(10) ± 0.0090.325(11) ± 0.010
Recreation0.414(1) ± 0.013 •0.409(2) ± 0.0120.407(3) ± 0.0120.399(4) ± 0.0140.390(5) ± 0.0160.387(6) ± 0.0170.384(7) ± 0.0190.382(8) ± 0.0180.379(9) ± 0.0150.376(10) ± 0.0150.309(11) ± 0.014
Society0.305(1) ± 0.009 •0.303(2) ± 0.0100.300(3) ± 0.0100.295(4) ± 0.0090.291(5) ± 0.0120.287(6) ± 0.0130.285(7) ± 0.0110.284(8) ± 0.0100.281(10) ± 0.0090.281(9) ± 0.0100.232(11) ± 0.010
eurlex-dc-l0.266(8) ± 0.0080.257(10) ± 0.0100.259(9) ± 0.0090.267(6) ± 0.0090.267(7) ± 0.0100.269(2) ± 0.0120.269(3) ± 0.0100.269(1) ± 0.010 •0.268(4) ± 0.0100.268(5) ± 0.0090.240(11) ± 0.006
eurlex-sm0.377(8) ± 0.0070.375(10) ± 0.0080.376(9) ± 0.0090.383(7) ± 0.0100.388(6) ± 0.0090.390(5) ± 0.0090.390(4) ± 0.0080.390(3) ± 0.0090.391(1) ± 0.007 •0.390(2) ± 0.0070.367(11) ± 0.007
tmc2007-5000.526(11) ± 0.0080.572(10) ± 0.0070.587(9) ± 0.0060.598(8) ± 0.0080.603(6) ± 0.0070.606(5) ± 0.0090.609(4) ± 0.0080.610(3) ± 0.0080.610(2) ± 0.0080.612(1) ± 0.008 •0.601(7) ± 0.006
mediamill0.235(11) ± 0.0080.238(10) ± 0.0040.252(9) ± 0.0040.284(8) ± 0.0050.293(6) ± 0.0050.297(5) ± 0.0030.299(4) ± 0.0050.300(3) ± 0.0050.302(1) ± 0.005 •0.301(2) ± 0.0050.285(7) ± 0.012
average rank3.894.395.004.755.215.466.256.546.937.3910.18
win/tie/loss186/40/54167/53/60146/51/83136/66/78121/72/87100/84/9682/97/10169/105/10664/98/11859/87/13414/39/227
Table A33. The performance of MLHiKNN across different values of λ in terms of F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A33. The performance of MLHiKNN across different values of λ in terms of F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Micro
λ = 0.05λ = 0.5λ = 1λ = 2λ = 3λ = 4λ = 5λ = 6λ = 7λ = 8λ → +∞
birds0.414(1) ± 0.035 •0.388(2) ± 0.0320.361(7) ± 0.0580.363(4) ± 0.0560.358(8) ± 0.0610.366(3) ± 0.0570.357(9) ± 0.0560.362(6) ± 0.0570.363(5) ± 0.0560.357(10) ± 0.0530.317(11) ± 0.091
CAL5000.356(1) ± 0.008 •0.345(2) ± 0.0080.340(5) ± 0.0120.341(4) ± 0.0100.342(3) ± 0.0140.339(6) ± 0.0150.336(7) ± 0.0150.334(9) ± 0.0160.335(8) ± 0.0160.332(10) ± 0.0130.329(11) ± 0.015
emotions0.661(11) ± 0.0130.662(10) ± 0.0130.664(9) ± 0.0140.669(7) ± 0.0130.672(1) ± 0.011 •0.671(3) ± 0.0130.670(6) ± 0.0150.671(5) ± 0.0140.671(4) ± 0.0130.672(2) ± 0.0140.666(8) ± 0.013
genbase0.947(11) ± 0.0100.954(10) ± 0.0110.958(9) ± 0.0100.963(8) ± 0.0150.965(7) ± 0.0140.966(6) ± 0.0130.968(5) ± 0.0130.969(4) ± 0.0090.970(3) ± 0.0090.970(2) ± 0.0090.976(1) ± 0.011 •
LLOG0.357(4) ± 0.0510.357(3) ± 0.0500.351(10) ± 0.0510.362(2) ± 0.0510.363(1) ± 0.059 •0.355(6) ± 0.0680.355(5) ± 0.0680.352(8) ± 0.0730.352(8) ± 0.0730.352(8) ± 0.0730.343(11) ± 0.064
enron0.488(1) ± 0.010 •0.477(2) ± 0.0110.469(3) ± 0.0100.462(4) ± 0.0090.460(5) ± 0.0130.455(8) ± 0.0140.454(9) ± 0.0130.453(10) ± 0.0130.453(11) ± 0.0130.456(7) ± 0.0120.457(6) ± 0.013
scene0.737(11) ± 0.0070.742(9) ± 0.0070.742(10) ± 0.0080.743(7) ± 0.0070.743(5) ± 0.0080.743(6) ± 0.0080.744(4) ± 0.0090.745(2) ± 0.0090.745(3) ± 0.0080.745(1) ± 0.010 •0.743(8) ± 0.009
yeast0.630(10) ± 0.0070.629(11) ± 0.0080.632(9) ± 0.0080.639(8) ± 0.0080.643(7) ± 0.0060.645(5) ± 0.0060.647(4) ± 0.0050.648(3) ± 0.0060.648(2) ± 0.0060.650(1) ± 0.005 •0.644(6) ± 0.004
Slashdot0.844(1) ± 0.008 •0.844(3) ± 0.0080.843(4) ± 0.0080.844(2) ± 0.0080.843(5) ± 0.0080.843(6) ± 0.0070.841(9) ± 0.0070.842(7) ± 0.0080.842(8) ± 0.0080.841(10) ± 0.0080.839(11) ± 0.010
corel5k0.138(1) ± 0.008 •0.071(2) ± 0.0100.060(3) ± 0.0110.054(4) ± 0.0110.050(5) ± 0.0110.050(6) ± 0.0100.046(9) ± 0.0080.047(8) ± 0.0090.047(7) ± 0.0080.046(10) ± 0.0080.031(11) ± 0.004
rcv1subset10.386(1) ± 0.005 •0.364(2) ± 0.0060.342(3) ± 0.0140.300(4) ± 0.0200.296(5) ± 0.0210.289(6) ± 0.0230.284(7) ± 0.0170.282(8) ± 0.0180.282(9) ± 0.0170.282(10) ± 0.0180.244(11) ± 0.010
rcv1subset20.417(1) ± 0.009 •0.386(2) ± 0.0160.358(3) ± 0.0180.339(4) ± 0.0200.334(5) ± 0.0120.322(6) ± 0.0080.320(7) ± 0.0090.319(8) ± 0.0090.316(10) ± 0.0120.316(9) ± 0.0100.293(11) ± 0.018
rcv1subset30.403(1) ± 0.005 •0.370(2) ± 0.0070.352(3) ± 0.0160.338(4) ± 0.0090.333(5) ± 0.0090.328(6) ± 0.0120.323(8) ± 0.0140.324(7) ± 0.0110.321(9) ± 0.0150.317(10) ± 0.0130.290(11) ± 0.012
rcv1subset40.469(1) ± 0.006 •0.441(2) ± 0.0140.419(3) ± 0.0170.404(4) ± 0.0220.394(5) ± 0.0190.392(6) ± 0.0150.391(7) ± 0.0150.391(8) ± 0.0150.389(9) ± 0.0150.389(10) ± 0.0150.362(11) ± 0.010
rcv1subset50.425(1) ± 0.005 •0.390(2) ± 0.0240.371(3) ± 0.0200.357(4) ± 0.0170.348(5) ± 0.0140.345(6) ± 0.0160.344(7) ± 0.0140.339(8) ± 0.0130.337(9) ± 0.0150.333(10) ± 0.0120.286(11) ± 0.013
bibtex0.373(3) ± 0.0040.379(1) ± 0.005 •0.375(2) ± 0.0050.371(4) ± 0.0040.367(5) ± 0.0060.365(6) ± 0.0050.364(7) ± 0.0040.363(8) ± 0.0040.362(9) ± 0.0040.361(10) ± 0.0040.327(11) ± 0.009
Arts0.380(1) ± 0.013 •0.379(2) ± 0.0120.377(3) ± 0.0130.372(4) ± 0.0140.369(5) ± 0.0140.366(6) ± 0.0150.363(7) ± 0.0160.361(8) ± 0.0150.359(9) ± 0.0150.357(10) ± 0.0140.299(11) ± 0.020
Health0.569(10) ± 0.0110.571(9) ± 0.0110.573(4) ± 0.0110.574(3) ± 0.0120.574(1) ± 0.011 •0.574(2) ± 0.0120.573(5) ± 0.0120.573(6) ± 0.0120.572(7) ± 0.0120.572(8) ± 0.0120.538(11) ± 0.013
Business0.721(8) ± 0.0040.722(4) ± 0.0040.722(3) ± 0.0040.723(1) ± 0.004 •0.722(2) ± 0.0040.722(6) ± 0.0040.722(5) ± 0.0030.721(7) ± 0.0040.720(9) ± 0.0040.720(10) ± 0.0040.704(11) ± 0.004
Education0.393(1) ± 0.012 •0.392(2) ± 0.0120.388(3) ± 0.0120.383(4) ± 0.0130.378(5) ± 0.0140.376(6) ± 0.0130.375(7) ± 0.0140.373(8) ± 0.0140.372(9) ± 0.0140.371(10) ± 0.0140.315(11) ± 0.017
Computers0.513(4) ± 0.0050.515(1) ± 0.005 •0.515(2) ± 0.0060.513(3) ± 0.0050.511(5) ± 0.0050.510(6) ± 0.0050.509(7) ± 0.0050.509(8) ± 0.0050.508(9) ± 0.0050.507(10) ± 0.0050.475(11) ± 0.007
Entertainment0.508(3) ± 0.0060.510(1) ± 0.007 •0.508(2) ± 0.0080.507(4) ± 0.0070.505(5) ± 0.0080.503(6) ± 0.0070.503(7) ± 0.0070.502(8) ± 0.0060.502(9) ± 0.0060.501(10) ± 0.0060.454(11) ± 0.010
Recreation0.463(1) ± 0.011 •0.462(2) ± 0.0100.460(3) ± 0.0110.458(4) ± 0.0110.454(5) ± 0.0120.453(6) ± 0.0110.450(7) ± 0.0120.448(8) ± 0.0110.446(9) ± 0.0110.445(10) ± 0.0120.395(11) ± 0.017
Society0.470(3) ± 0.0080.471(1) ± 0.009 •0.471(2) ± 0.0090.469(4) ± 0.0100.468(5) ± 0.0100.468(6) ± 0.0100.467(7) ± 0.0100.466(8) ± 0.0110.464(9) ± 0.0120.463(10) ± 0.0120.431(11) ± 0.012
eurlex-dc-l0.453(11) ± 0.0060.455(10) ± 0.0060.461(9) ± 0.0060.472(7) ± 0.0050.479(6) ± 0.0070.482(5) ± 0.0070.483(4) ± 0.0070.484(3) ± 0.0070.485(1) ± 0.007 •0.484(2) ± 0.0070.466(8) ± 0.007
eurlex-sm0.564(11) ± 0.0060.568(10) ± 0.0070.574(9) ± 0.0070.585(7) ± 0.0070.589(6) ± 0.0070.591(5) ± 0.0070.592(4) ± 0.0070.592(3) ± 0.0070.593(2) ± 0.0070.593(1) ± 0.007 •0.580(8) ± 0.006
tmc2007-5000.679(11) ± 0.0030.705(10) ± 0.0020.715(9) ± 0.0020.723(8) ± 0.0030.726(7) ± 0.0030.729(6) ± 0.0030.730(5) ± 0.0030.732(4) ± 0.0030.732(2) ± 0.0030.733(1) ± 0.003 •0.732(3) ± 0.002
mediamill0.598(11) ± 0.0020.604(10) ± 0.0020.615(9) ± 0.0020.629(8) ± 0.0020.635(7) ± 0.0010.638(6) ± 0.0010.639(5) ± 0.0010.640(4) ± 0.0010.641(3) ± 0.0010.641(2) ± 0.0020.643(1) ± 0.002 •
average rank4.824.545.144.684.865.616.436.576.867.299.21
win/tie/loss152/35/93152/54/74134/66/80142/61/77114/78/8896/90/9486/97/9783/94/10374/99/10770/78/13238/46/196
Table A34. The performance of MLHiKNN across different values of λ in terms of Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A34. The performance of MLHiKNN across different values of λ in terms of Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetHamming Loss
λ = 0.05λ = 0.5λ = 1λ = 2λ = 3λ = 4λ = 5λ = 6λ = 7λ = 8λ → +∞
birds0.144(11) ± 0.0090.141(9) ± 0.0090.141(10) ± 0.0080.139(8) ± 0.0090.137(7) ± 0.0080.136(5) ± 0.0080.137(6) ± 0.0080.136(4) ± 0.0070.136(2) ± 0.0070.136(3) ± 0.0070.135(1) ± 0.009 •
CAL5000.200(11) ± 0.0020.192(10) ± 0.0020.189(9) ± 0.0010.187(8) ± 0.0010.187(7) ± 0.0010.186(2) ± 0.0010.186(5) ± 0.0020.186(4) ± 0.0010.186(1) ± 0.001 •0.186(3) ± 0.0020.186(6) ± 0.001
emotions0.199(11) ± 0.0060.197(10) ± 0.0060.196(9) ± 0.0070.192(7) ± 0.0060.190(1) ± 0.005 •0.191(5) ± 0.0060.191(6) ± 0.0070.191(4) ± 0.0070.190(3) ± 0.0070.190(2) ± 0.0070.193(8) ± 0.007
genbase0.009(11) ± 0.0020.008(10) ± 0.0020.007(9) ± 0.0020.006(8) ± 0.0030.006(7) ± 0.0020.006(6) ± 0.0020.006(5) ± 0.0020.005(4) ± 0.0020.005(3) ± 0.0020.005(2) ± 0.0020.004(1) ± 0.002 •
LLOG0.188(4) ± 0.0050.188(2) ± 0.0050.189(5) ± 0.0050.188(1) ± 0.005 •0.188(3) ± 0.0060.189(7) ± 0.0060.189(6) ± 0.0060.189(9) ± 0.0060.189(9) ± 0.0060.189(9) ± 0.0060.190(11) ± 0.007
enron0.078(11) ± 0.0010.075(1) ± 0.001 •0.075(2) ± 0.0010.075(3) ± 0.0010.075(4) ± 0.0010.075(5) ± 0.0010.076(7) ± 0.0010.076(10) ± 0.0010.076(9) ± 0.0010.076(8) ± 0.0010.076(6) ± 0.001
scene0.087(11) ± 0.0020.085(10) ± 0.0020.085(9) ± 0.0020.085(8) ± 0.0020.084(7) ± 0.0030.084(6) ± 0.0020.084(5) ± 0.0030.084(4) ± 0.0020.084(3) ± 0.0020.083(2) ± 0.0030.083(1) ± 0.003 •
yeast0.211(10) ± 0.0040.211(11) ± 0.0040.208(9) ± 0.0040.202(8) ± 0.0040.200(7) ± 0.0030.198(6) ± 0.0030.197(5) ± 0.0030.196(4) ± 0.0040.196(3) ± 0.0030.196(2) ± 0.0030.195(1) ± 0.003 •
Slashdot0.028(1) ± 0.002 •0.029(3) ± 0.0010.029(4) ± 0.0010.028(2) ± 0.0010.029(5) ± 0.0010.029(6) ± 0.0010.029(9) ± 0.0010.029(7) ± 0.0010.029(8) ± 0.0020.029(10) ± 0.0010.029(11) ± 0.002
corel5k0.025(11) ± 0.0010.021(10) ± 0.0000.021(9) ± 0.0000.021(8) ± 0.0000.021(7) ± 0.0000.021(6) ± 0.0000.021(5) ± 0.0000.021(4) ± 0.0000.021(3) ± 0.0000.021(2) ± 0.0000.021(1) ± 0.000 •
rcv1subset10.035(11) ± 0.0010.033(10) ± 0.0000.033(8) ± 0.0000.033(7) ± 0.0000.032(6) ± 0.0000.032(5) ± 0.0000.032(4) ± 0.0000.032(3) ± 0.0000.032(2) ± 0.0000.032(1) ± 0.000 •0.033(9) ± 0.000
rcv1subset20.029(11) ± 0.0000.028(10) ± 0.0000.027(8) ± 0.0000.027(1) ± 0.000 •0.027(2) ± 0.0000.027(7) ± 0.0000.027(5) ± 0.0000.027(6) ± 0.0000.027(4) ± 0.0000.027(3) ± 0.0000.027(9) ± 0.000
rcv1subset30.029(11) ± 0.0000.028(10) ± 0.0000.027(8) ± 0.0000.027(4) ± 0.0000.027(7) ± 0.0000.027(6) ± 0.0000.027(5) ± 0.0000.027(2) ± 0.0000.027(3) ± 0.0000.027(1) ± 0.000 •0.027(9) ± 0.000
rcv1subset40.026(11) ± 0.0000.025(9) ± 0.0000.024(8) ± 0.0000.024(1) ± 0.000 •0.024(6) ± 0.0000.024(7) ± 0.0000.024(4) ± 0.0000.024(2) ± 0.0000.024(5) ± 0.0000.024(3) ± 0.0000.025(10) ± 0.000
rcv1subset50.028(11) ± 0.0000.027(9) ± 0.0000.027(4) ± 0.0000.027(1) ± 0.000 •0.027(2) ± 0.0000.027(3) ± 0.0000.027(5) ± 0.0000.027(6) ± 0.0000.027(7) ± 0.0000.027(8) ± 0.0000.027(10) ± 0.000
bibtex0.015(11) ± 0.0000.013(9) ± 0.0000.013(8) ± 0.0000.013(6) ± 0.0000.013(4) ± 0.0000.013(1) ± 0.000 •0.013(7) ± 0.0000.013(3) ± 0.0000.013(2) ± 0.0000.013(5) ± 0.0000.013(10) ± 0.000
Arts0.060(10) ± 0.0010.059(6) ± 0.0010.059(3) ± 0.0000.059(1) ± 0.000 •0.059(2) ± 0.0010.059(4) ± 0.0010.059(5) ± 0.0010.059(7) ± 0.0000.060(8) ± 0.0010.060(9) ± 0.0010.062(11) ± 0.001
Health0.056(10) ± 0.0010.055(9) ± 0.0010.055(8) ± 0.0010.055(6) ± 0.0010.055(2) ± 0.0010.055(1) ± 0.001 •0.055(3) ± 0.0010.055(4) ± 0.0010.055(5) ± 0.0010.055(7) ± 0.0010.057(11) ± 0.001
Business0.031(10) ± 0.0000.031(9) ± 0.0000.031(7) ± 0.0000.031(2) ± 0.0000.031(1) ± 0.000 •0.031(4) ± 0.0000.031(3) ± 0.0000.031(5) ± 0.0000.031(6) ± 0.0000.031(8) ± 0.0010.032(11) ± 0.000
Education0.046(10) ± 0.0010.046(9) ± 0.0010.046(8) ± 0.0010.046(2) ± 0.0010.046(1) ± 0.001 •0.046(3) ± 0.0010.046(4) ± 0.0010.046(5) ± 0.0010.046(6) ± 0.0010.046(6) ± 0.0010.048(11) ± 0.001
Computers0.043(10) ± 0.0000.043(9) ± 0.0000.043(3) ± 0.0000.043(1) ± 0.000 •0.043(2) ± 0.0000.043(4) ± 0.0000.043(5) ± 0.0000.043(6) ± 0.0000.043(7) ± 0.0000.043(8) ± 0.0000.044(11) ± 0.001
Entertainment0.061(10) ± 0.0010.061(9) ± 0.0010.061(4) ± 0.0010.061(2) ± 0.0010.061(1) ± 0.001 •0.061(3) ± 0.0010.061(5) ± 0.0010.061(6) ± 0.0010.061(7) ± 0.0010.061(8) ± 0.0010.064(11) ± 0.001
Recreation0.050(5) ± 0.0010.050(3) ± 0.0010.050(1) ± 0.001 •0.050(2) ± 0.0010.050(4) ± 0.0010.050(6) ± 0.0010.050(7) ± 0.0010.050(8) ± 0.0010.051(9) ± 0.0010.051(10) ± 0.0010.053(11) ± 0.001
Society0.051(10) ± 0.0010.051(6) ± 0.0010.051(3) ± 0.0010.051(2) ± 0.0010.051(1) ± 0.001 •0.051(4) ± 0.0010.051(5) ± 0.0010.051(7) ± 0.0010.051(8) ± 0.0010.051(9) ± 0.0010.053(11) ± 0.001
eurlex-dc-l0.005(11) ± 0.0000.005(10) ± 0.0000.005(9) ± 0.0000.005(8) ± 0.0000.004(6) ± 0.0000.004(5) ± 0.0000.004(4) ± 0.0000.004(1) ± 0.000 •0.004(2) ± 0.0000.004(3) ± 0.0000.005(7) ± 0.000
eurlex-sm0.011(11) ± 0.0000.011(10) ± 0.0000.011(9) ± 0.0000.011(8) ± 0.0000.011(6) ± 0.0000.011(5) ± 0.0000.011(4) ± 0.0000.011(3) ± 0.0000.011(2) ± 0.0000.011(1) ± 0.000 •0.011(7) ± 0.000
tmc2007-5000.060(11) ± 0.0000.055(10) ± 0.0000.053(9) ± 0.0000.052(8) ± 0.0000.051(7) ± 0.0000.051(6) ± 0.0000.050(5) ± 0.0000.050(4) ± 0.0000.050(3) ± 0.0000.050(1) ± 0.000 •0.050(2) ± 0.000
mediamill0.031(11) ± 0.0000.031(10) ± 0.0000.030(9) ± 0.0000.028(8) ± 0.0000.028(7) ± 0.0000.028(6) ± 0.0000.028(5) ± 0.0000.028(4) ± 0.0000.027(3) ± 0.0000.027(2) ± 0.0000.027(1) ± 0.000 •
average rank9.898.326.864.684.364.795.144.864.774.887.46
win/tie/loss16/24/24040/61/17977/81/122119/98/63124/113/43120/113/47119/121/40119/124/37111/125/44112/113/5569/62/149
Table A35. The performance of MLHiKNN across different values of β in terms of AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A35. The performance of MLHiKNN across different values of β in terms of AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Macro
β = 5/4β = 3/2β = 5/3β = 2β = 3β = 5β = 9β = 17
birds0.734(1) ± 0.026 •0.731(2) ± 0.0290.730(3) ± 0.0290.728(4) ± 0.0300.723(6) ± 0.0310.724(5) ± 0.0300.722(7) ± 0.0280.721(8) ± 0.029
CAL5000.570(8) ± 0.0080.571(6) ± 0.0070.572(1) ± 0.007 •0.571(3) ± 0.0060.571(2) ± 0.0050.571(4) ± 0.0050.571(5) ± 0.0050.571(7) ± 0.005
emotions0.847(1) ± 0.006 •0.847(2) ± 0.0060.847(3) ± 0.0070.846(4) ± 0.0050.845(6) ± 0.0070.845(5) ± 0.0070.844(7) ± 0.0070.844(8) ± 0.007
genbase0.993(8) ± 0.0050.996(7) ± 0.0030.996(6) ± 0.0030.997(5) ± 0.0030.997(4) ± 0.0030.998(3) ± 0.0030.998(2) ± 0.0030.998(1) ± 0.003 •
LLOG0.603(7) ± 0.0060.603(4) ± 0.0060.603(1) ± 0.006 •0.603(4) ± 0.0060.603(1) ± 0.006 •0.603(7) ± 0.0060.603(4) ± 0.0060.603(4) ± 0.006
enron0.673(5) ± 0.0120.674(4) ± 0.0130.675(2) ± 0.0130.676(1) ± 0.012 •0.675(3) ± 0.0130.673(6) ± 0.0110.672(7) ± 0.0110.672(8) ± 0.011
scene0.946(1) ± 0.002 •0.945(2) ± 0.0020.945(3) ± 0.0020.944(4) ± 0.0020.944(5) ± 0.0020.944(6) ± 0.0020.943(8) ± 0.0020.943(7) ± 0.002
yeast0.718(1) ± 0.007 •0.715(2) ± 0.0070.713(3) ± 0.0060.711(4) ± 0.0060.709(5) ± 0.0060.708(6) ± 0.0060.708(7) ± 0.0050.707(8) ± 0.005
Slashdot0.713(8) ± 0.0260.721(7) ± 0.0220.722(6) ± 0.0220.723(5) ± 0.0210.724(4) ± 0.0200.724(3) ± 0.0190.724(2) ± 0.0200.724(1) ± 0.020 •
corel5k0.690(4) ± 0.0030.690(1) ± 0.004 •0.690(2) ± 0.0040.690(3) ± 0.0050.690(5) ± 0.0050.689(6) ± 0.0050.689(7) ± 0.0050.689(8) ± 0.005
rcv1subset10.891(4) ± 0.0030.892(2) ± 0.0030.892(1) ± 0.003 •0.891(3) ± 0.0040.890(5) ± 0.0040.889(6) ± 0.0040.888(7) ± 0.0040.888(8) ± 0.004
rcv1subset20.890(1) ± 0.005 •0.890(2) ± 0.0050.888(3) ± 0.0040.888(4) ± 0.0050.887(5) ± 0.0050.886(6) ± 0.0050.885(7) ± 0.0050.884(8) ± 0.006
rcv1subset30.880(2) ± 0.0030.881(1) ± 0.004 •0.880(3) ± 0.0040.880(4) ± 0.0040.879(5) ± 0.0040.879(6) ± 0.0040.878(7) ± 0.0020.877(8) ± 0.002
rcv1subset40.887(3) ± 0.0070.888(1) ± 0.007 •0.888(2) ± 0.0070.887(4) ± 0.0070.885(5) ± 0.0070.884(6) ± 0.0070.884(7) ± 0.0070.883(8) ± 0.007
rcv1subset50.872(6) ± 0.0050.874(2) ± 0.0050.874(1) ± 0.006 •0.873(3) ± 0.0050.872(5) ± 0.0050.872(4) ± 0.0060.871(7) ± 0.0050.871(8) ± 0.005
bibtex0.885(5) ± 0.0010.887(1) ± 0.001 •0.886(2) ± 0.0010.886(3) ± 0.0010.885(4) ± 0.0010.885(6) ± 0.0010.885(7) ± 0.0010.884(8) ± 0.002
Arts0.720(8) ± 0.0130.728(7) ± 0.0110.729(6) ± 0.0100.730(5) ± 0.0090.731(2) ± 0.0090.730(4) ± 0.0080.731(3) ± 0.0080.731(1) ± 0.008 •
Health0.771(8) ± 0.0090.776(7) ± 0.0090.777(6) ± 0.0090.778(3) ± 0.0090.780(1) ± 0.009 •0.779(2) ± 0.0100.778(5) ± 0.0100.778(4) ± 0.010
Business0.742(8) ± 0.0070.748(7) ± 0.0070.751(3) ± 0.0080.752(2) ± 0.0070.752(1) ± 0.006 •0.750(6) ± 0.0080.750(4) ± 0.0070.750(5) ± 0.008
Education0.737(8) ± 0.0140.745(7) ± 0.0130.747(4) ± 0.0130.748(2) ± 0.0140.749(1) ± 0.012 •0.747(6) ± 0.0120.747(3) ± 0.0140.747(5) ± 0.013
Computers0.742(8) ± 0.0060.749(7) ± 0.0050.749(5) ± 0.0050.750(2) ± 0.0050.751(1) ± 0.007 •0.750(3) ± 0.0070.749(6) ± 0.0070.749(4) ± 0.007
Entertainment0.773(8) ± 0.0050.778(7) ± 0.0050.780(6) ± 0.0040.782(1) ± 0.004 •0.781(3) ± 0.0050.781(2) ± 0.0050.781(4) ± 0.0050.781(5) ± 0.005
Recreation0.781(8) ± 0.0080.786(7) ± 0.0080.788(6) ± 0.0080.788(5) ± 0.0090.789(4) ± 0.0080.789(3) ± 0.0080.789(2) ± 0.0080.789(1) ± 0.008 •
Society0.700(8) ± 0.0070.705(7) ± 0.0060.706(6) ± 0.0060.707(3) ± 0.0060.707(1) ± 0.005 •0.707(2) ± 0.0050.706(4) ± 0.0050.706(5) ± 0.004
eurlex-dc-l0.896(6) ± 0.0050.900(3) ± 0.0050.900(2) ± 0.0050.900(1) ± 0.004 •0.899(4) ± 0.0040.896(5) ± 0.0040.896(7) ± 0.0040.895(8) ± 0.004
eurlex-sm0.907(4) ± 0.0030.910(1) ± 0.002 •0.909(2) ± 0.0030.908(3) ± 0.0030.906(5) ± 0.0020.905(6) ± 0.0020.904(7) ± 0.0030.904(8) ± 0.002
tmc2007-5000.921(1) ± 0.002 •0.921(6) ± 0.0020.920(8) ± 0.0020.920(7) ± 0.0020.921(2) ± 0.0020.921(4) ± 0.0020.921(3) ± 0.0020.921(5) ± 0.002
mediamill0.830(1) ± 0.002 •0.830(2) ± 0.0020.828(3) ± 0.0020.824(4) ± 0.0020.815(5) ± 0.0050.810(6) ± 0.0040.806(7) ± 0.0040.803(8) ± 0.005
average rank5.054.093.553.453.594.805.485.98
win/tie/loss51/62/8377/86/3382/92/2271/102/2364/89/4338/104/5426/93/7717/88/91
Table A36. The performance of MLHiKNN across different values of β in terms of AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A36. The performance of MLHiKNN across different values of β in terms of AUC micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Micro
β = 5/4β = 3/2β = 5/3β = 2β = 3β = 5β = 9β = 17
birds0.756(1) ± 0.022 •0.751(2) ± 0.0170.750(3) ± 0.0190.746(4) ± 0.0190.744(5) ± 0.0180.741(6) ± 0.0160.740(7) ± 0.0170.739(8) ± 0.018
CAL5000.757(7) ± 0.0040.756(8) ± 0.0030.758(5) ± 0.0030.757(6) ± 0.0040.760(1) ± 0.003 •0.759(4) ± 0.0030.759(3) ± 0.0030.759(2) ± 0.003
emotions0.862(2) ± 0.0040.862(1) ± 0.005 •0.862(4) ± 0.0060.862(3) ± 0.0050.859(7) ± 0.0060.860(5) ± 0.0070.860(6) ± 0.0060.859(8) ± 0.006
genbase0.992(8) ± 0.0060.995(7) ± 0.0030.996(6) ± 0.0030.996(4) ± 0.0040.996(5) ± 0.0040.997(2) ± 0.0040.997(3) ± 0.0040.997(1) ± 0.004 •
LLOG0.767(5) ± 0.0020.767(2) ± 0.002 •0.767(7) ± 0.0020.767(2) ± 0.002 •0.767(7) ± 0.0020.767(5) ± 0.0020.767(2) ± 0.002 •0.767(2) ± 0.002 •
enron0.858(6) ± 0.0030.858(8) ± 0.0040.859(3) ± 0.0040.860(1) ± 0.002 •0.860(2) ± 0.0020.859(4) ± 0.0030.858(5) ± 0.0030.858(7) ± 0.003
scene0.952(1) ± 0.002 •0.951(2) ± 0.0020.951(3) ± 0.0020.950(4) ± 0.0020.950(5) ± 0.0020.949(7) ± 0.0020.949(8) ± 0.0020.949(6) ± 0.002
yeast0.847(1) ± 0.004 •0.845(2) ± 0.0040.843(3) ± 0.0030.842(4) ± 0.0040.840(5) ± 0.0040.840(6) ± 0.0030.840(7) ± 0.0030.840(8) ± 0.003
Slashdot 0.945(2) ± 0.0040.945(4) ± 0.0040.945(1) ± 0.004 •0.945(3) ± 0.0040.945(6) ± 0.0040.945(8) ± 0.0040.945(5) ± 0.0040.945(7) ± 0.004
corel5k0.796(1) ± 0.002 •0.796(3) ± 0.0040.796(4) ± 0.0040.796(2) ± 0.0040.796(5) ± 0.0030.795(6) ± 0.0040.795(8) ± 0.0030.795(7) ± 0.003
rcv1subset10.867(8) ± 0.0070.872(7) ± 0.0070.873(6) ± 0.0080.874(2) ± 0.0090.875(1) ± 0.008 •0.873(3) ± 0.0080.873(5) ± 0.0090.873(4) ± 0.009
rcv1subset20.865(5) ± 0.0100.864(7) ± 0.0070.862(8) ± 0.0070.865(6) ± 0.0100.868(2) ± 0.0120.866(3) ± 0.0100.866(4) ± 0.0100.868(1) ± 0.010 •
rcv1subset30.850(8) ± 0.0070.856(7) ± 0.0100.856(6) ± 0.0080.861(5) ± 0.0070.864(4) ± 0.0080.865(2) ± 0.0090.866(1) ± 0.007 •0.865(3) ± 0.007
rcv1subset40.881(8) ± 0.0100.885(6) ± 0.0100.886(5) ± 0.0100.884(7) ± 0.0100.887(4) ± 0.0090.888(3) ± 0.0090.889(2) ± 0.0090.889(1) ± 0.009 •
rcv1subset50.862(8) ± 0.0060.868(5) ± 0.0060.870(4) ± 0.0080.868(7) ± 0.0090.868(6) ± 0.0090.871(3) ± 0.0090.872(1) ± 0.009 •0.871(2) ± 0.010
bibtex0.850(8) ± 0.0070.860(7) ± 0.0060.861(6) ± 0.0040.863(1) ± 0.006 •0.862(4) ± 0.0040.863(2) ± 0.0050.862(5) ± 0.0050.862(3) ± 0.005
Arts0.846(8) ± 0.0040.850(7) ± 0.0030.851(5) ± 0.0030.851(4) ± 0.0030.852(2) ± 0.0030.852(1) ± 0.003 •0.852(3) ± 0.0030.851(6) ± 0.003
Health0.917(8) ± 0.0020.919(4) ± 0.0020.919(3) ± 0.0020.919(2) ± 0.0020.919(1) ± 0.002 •0.919(5) ± 0.0020.918(6) ± 0.0020.918(7) ± 0.002
Business0.945(8) ± 0.0010.947(7) ± 0.0010.947(6) ± 0.0010.947(3) ± 0.0010.948(1) ± 0.001 •0.947(4) ± 0.0010.947(5) ± 0.0000.947(2) ± 0.001
Education0.904(8) ± 0.0020.906(7) ± 0.0020.906(6) ± 0.0020.906(4) ± 0.0020.907(1) ± 0.002 •0.907(2) ± 0.0020.906(3) ± 0.0020.906(5) ± 0.002
Computers0.891(8) ± 0.0030.894(7) ± 0.0030.895(3) ± 0.0030.896(1) ± 0.003 •0.895(2) ± 0.0030.895(5) ± 0.0030.895(6) ± 0.0030.895(4) ± 0.003
Entertainment0.886(8) ± 0.0020.890(7) ± 0.0020.891(6) ± 0.0020.891(2) ± 0.0020.891(1) ± 0.002 •0.891(3) ± 0.0020.891(4) ± 0.0020.891(5) ± 0.002
Recreation0.868(8) ± 0.0040.872(7) ± 0.0040.873(6) ± 0.0030.874(5) ± 0.0040.875(1) ± 0.004 •0.875(2) ± 0.0040.875(3) ± 0.0040.875(4) ± 0.004
Society0.866(8) ± 0.0020.868(4) ± 0.0020.869(1) ± 0.002 •0.869(2) ± 0.0020.869(3) ± 0.0020.868(5) ± 0.0020.868(6) ± 0.0020.868(7) ± 0.002
eurlex-dc-l0.914(7) ± 0.0060.918(3) ± 0.0070.920(1) ± 0.006 •0.919(2) ± 0.0050.918(4) ± 0.0050.915(5) ± 0.0050.914(6) ± 0.0050.913(8) ± 0.006
eurlex-sm0.951(4) ± 0.0020.953(1) ± 0.001 •0.953(2) ± 0.0020.952(3) ± 0.0020.951(5) ± 0.0020.950(6) ± 0.0020.949(7) ± 0.0020.949(8) ± 0.001
tmc2007-5000.958(1) ± 0.001 •0.958(8) ± 0.0010.958(7) ± 0.0010.958(5) ± 0.0010.958(2) ± 0.0010.958(4) ± 0.0010.958(3) ± 0.0010.958(6) ± 0.001
mediamill0.960(1) ± 0.000 •0.959(2) ± 0.0010.958(3) ± 0.0010.957(4) ± 0.0010.955(5) ± 0.0010.953(6) ± 0.0010.952(7) ± 0.0010.952(8) ± 0.001
average rank5.595.094.413.523.484.204.705.02
win/tie/loss41/49/10651/91/5456/106/3465/110/2170/103/2349/107/4034/108/5432/98/66
Table A37. The performance of MLHiKNN across different values of β in terms of Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A37. The performance of MLHiKNN across different values of β in terms of Ranking Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetRanking Loss
β = 5/4β = 3/2β = 5/3β = 2β = 3β = 5β = 9β = 17
birds0.245(1) ± 0.025 •0.248(3) ± 0.0210.248(2) ± 0.0220.251(4) ± 0.0200.256(5) ± 0.0220.259(6) ± 0.0190.260(7) ± 0.0200.260(8) ± 0.019
CAL5000.240(7) ± 0.0040.241(8) ± 0.0030.240(5) ± 0.0040.240(6) ± 0.0050.238(1) ± 0.003 •0.238(4) ± 0.0040.238(3) ± 0.0030.238(2) ± 0.003
emotions0.153(4) ± 0.0090.152(1) ± 0.007 •0.152(2) ± 0.0070.153(3) ± 0.0070.155(5) ± 0.0090.156(7) ± 0.0100.156(6) ± 0.0100.156(8) ± 0.010
genbase0.009(8) ± 0.0050.006(7) ± 0.0030.005(6) ± 0.0030.004(5) ± 0.0030.004(4) ± 0.0020.004(3) ± 0.0040.004(2) ± 0.0030.003(1) ± 0.003 •
LLOG0.187(3) ± 0.002 •0.187(3) ± 0.002 •0.187(7) ± 0.0020.187(3) ± 0.002 •0.187(7) ± 0.0020.187(3) ± 0.002 •0.187(3) ± 0.002 •0.187(3) ± 0.002 •
enron0.137(6) ± 0.0040.136(5) ± 0.0040.136(4) ± 0.0040.135(1) ± 0.003 •0.135(2) ± 0.0030.136(3) ± 0.0040.137(7) ± 0.0040.137(8) ± 0.004
scene0.070(1) ± 0.003 •0.070(2) ± 0.0030.071(3) ± 0.0030.071(4) ± 0.0030.072(5) ± 0.0030.072(6) ± 0.0030.072(7) ± 0.0030.072(8) ± 0.003
yeast0.165(1) ± 0.004 •0.167(2) ± 0.0040.168(3) ± 0.0030.168(4) ± 0.0030.169(6) ± 0.0030.169(5) ± 0.0030.170(7) ± 0.0030.170(8) ± 0.003
Slashdot0.047(3) ± 0.0050.047(2) ± 0.0050.047(1) ± 0.004 •0.047(4) ± 0.0050.047(7) ± 0.0040.047(8) ± 0.0050.047(5) ± 0.0040.047(6) ± 0.004
corel5k0.206(1) ± 0.003 •0.206(3) ± 0.0040.206(4) ± 0.0040.206(2) ± 0.0040.206(5) ± 0.0040.207(6) ± 0.0040.207(8) ± 0.0040.207(7) ± 0.004
rcv1subset10.124(8) ± 0.0080.119(7) ± 0.0070.119(6) ± 0.0080.118(3) ± 0.0080.116(1) ± 0.008 •0.117(2) ± 0.0080.118(5) ± 0.0090.118(4) ± 0.009
rcv1subset20.117(3) ± 0.0100.118(5) ± 0.0070.121(8) ± 0.0080.118(6) ± 0.0090.116(1) ± 0.012 •0.118(4) ± 0.0100.119(7) ± 0.0090.117(2) ± 0.009
rcv1subset30.132(8) ± 0.0080.126(6) ± 0.0090.126(7) ± 0.0070.122(5) ± 0.0050.119(4) ± 0.0070.118(2) ± 0.0080.118(1) ± 0.007 •0.118(3) ± 0.007
rcv1subset40.103(8) ± 0.0090.099(6) ± 0.0090.098(5) ± 0.0080.100(7) ± 0.0090.098(4) ± 0.0090.096(3) ± 0.0080.096(2) ± 0.0080.096(1) ± 0.008 •
rcv1subset50.123(8) ± 0.0060.117(5) ± 0.0060.115(4) ± 0.0080.117(7) ± 0.0090.117(6) ± 0.0070.114(3) ± 0.0080.113(1) ± 0.008 •0.114(2) ± 0.009
bibtex0.158(8) ± 0.0050.149(7) ± 0.0040.148(5) ± 0.0030.147(1) ± 0.004 •0.148(3) ± 0.0030.148(2) ± 0.0030.148(6) ± 0.0030.148(4) ± 0.004
Arts0.134(8) ± 0.0030.130(7) ± 0.0030.129(4) ± 0.0030.129(5) ± 0.0030.129(1) ± 0.003 •0.129(2) ± 0.0030.129(3) ± 0.0030.129(6) ± 0.003
Health0.072(8) ± 0.0020.070(5) ± 0.0020.070(3) ± 0.0020.070(2) ± 0.0020.070(1) ± 0.002 •0.070(4) ± 0.0020.070(6) ± 0.0010.070(7) ± 0.001
Business0.038(8) ± 0.0010.037(7) ± 0.0010.037(6) ± 0.0010.037(5) ± 0.0010.037(2) ± 0.0010.037(3) ± 0.0010.037(4) ± 0.0010.037(1) ± 0.001 •
Education0.092(8) ± 0.0020.090(7) ± 0.0020.090(6) ± 0.0020.089(4) ± 0.0020.089(1) ± 0.002 •0.089(2) ± 0.0020.090(5) ± 0.0020.089(3) ± 0.002
Computers0.088(8) ± 0.0030.086(4) ± 0.0020.086(2) ± 0.0020.086(1) ± 0.002 •0.086(3) ± 0.0020.087(5) ± 0.0020.087(7) ± 0.0020.087(6) ± 0.002
Entertainment0.115(8) ± 0.0030.112(7) ± 0.0030.112(3) ± 0.0020.111(1) ± 0.002 •0.112(2) ± 0.0030.112(5) ± 0.0030.112(4) ± 0.0020.112(6) ± 0.002
Recreation0.120(8) ± 0.0040.117(7) ± 0.0040.116(6) ± 0.0030.115(5) ± 0.0040.115(1) ± 0.003 •0.115(2) ± 0.0040.115(3) ± 0.0040.115(4) ± 0.004
Society0.108(8) ± 0.0030.106(4) ± 0.0020.105(1) ± 0.002 •0.105(2) ± 0.0020.106(3) ± 0.0020.106(5) ± 0.0020.106(6) ± 0.0020.107(7) ± 0.002
eurlex-dc-l0.083(8) ± 0.0050.079(4) ± 0.0060.077(1) ± 0.005 •0.078(2) ± 0.0050.078(3) ± 0.0040.081(5) ± 0.0040.082(6) ± 0.0040.083(7) ± 0.005
eurlex-sm0.051(5) ± 0.0020.049(1) ± 0.002 •0.049(2) ± 0.0020.050(3) ± 0.0020.051(4) ± 0.0020.052(6) ± 0.0020.053(7) ± 0.0020.053(8) ± 0.002
tmc2007-5000.044(1) ± 0.001 •0.044(2) ± 0.0010.044(3) ± 0.0010.044(4) ± 0.0010.044(5) ± 0.0010.044(6) ± 0.0010.044(7) ± 0.0010.044(8) ± 0.001
mediamill0.036(1) ± 0.000 •0.036(2) ± 0.0010.037(3) ± 0.0010.038(4) ± 0.0010.040(5) ± 0.0010.041(6) ± 0.0010.041(7) ± 0.0000.042(8) ± 0.001
average rank5.624.624.023.703.484.235.095.23
win/tie/loss37/53/10654/99/4358/109/2957/115/2462/110/2443/114/3928/117/5132/109/55
Table A38. The performance of MLHiKNN across different values of β in terms of F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A38. The performance of MLHiKNN across different values of β in terms of F1 macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Macro
β = 5/4β = 3/2β = 5/3β = 2β = 3β = 5β = 9β = 17
birds0.283(1) ± 0.039 •0.280(2) ± 0.0440.279(3) ± 0.0460.258(4) ± 0.0690.257(5) ± 0.0670.250(6) ± 0.0550.245(8) ± 0.0710.245(7) ± 0.071
CAL5000.080(8) ± 0.0020.081(7) ± 0.0040.081(6) ± 0.0030.083(4) ± 0.0030.083(5) ± 0.0030.084(3) ± 0.0040.084(2) ± 0.0040.084(1) ± 0.004 •
emotions0.657(1) ± 0.014 •0.653(2) ± 0.0150.650(4) ± 0.0150.651(3) ± 0.0130.646(5) ± 0.0180.645(6) ± 0.0190.644(7) ± 0.0190.644(8) ± 0.018
genbase0.956(7) ± 0.0130.957(5) ± 0.0120.956(6) ± 0.0150.954(8) ± 0.0170.961(4) ± 0.0120.967(3) ± 0.0110.969(2) ± 0.0100.971(1) ± 0.012 •
LLOG0.097(3) ± 0.016 •0.097(3) ± 0.016 •0.097(7) ± 0.0160.097(3) ± 0.016 •0.097(7) ± 0.0160.097(3) ± 0.016 •0.097(3) ± 0.016 •0.097(3) ± 0.016 •
enron0.135(1) ± 0.010 •0.131(4) ± 0.0110.131(3) ± 0.0090.131(2) ± 0.0090.127(5) ± 0.0090.125(6) ± 0.0090.122(7) ± 0.0100.121(8) ± 0.009
scene0.754(1) ± 0.009 •0.751(2) ± 0.0070.750(3) ± 0.0080.749(4) ± 0.0080.747(5) ± 0.0080.746(6) ± 0.0070.746(7) ± 0.0070.746(8) ± 0.007
yeast0.428(1) ± 0.008 •0.424(2) ± 0.0070.421(3) ± 0.0080.420(4) ± 0.0090.415(8) ± 0.0090.415(5) ± 0.0080.415(7) ± 0.0080.415(6) ± 0.007
Slashdot 0.161(6) ± 0.0230.163(2) ± 0.0270.163(3) ± 0.0260.163(4) ± 0.0250.166(1) ± 0.025 •0.162(5) ± 0.0320.160(8) ± 0.0300.160(7) ± 0.031
corel5k0.026(2) ± 0.0030.027(1) ± 0.004 •0.026(4) ± 0.0050.026(3) ± 0.0040.026(5) ± 0.0030.024(8) ± 0.0040.025(7) ± 0.0030.025(6) ± 0.003
rcv1subset10.142(8) ± 0.0110.152(7) ± 0.0130.154(5) ± 0.0140.152(6) ± 0.0140.156(1) ± 0.013 •0.155(4) ± 0.0130.156(2) ± 0.0120.155(3) ± 0.012
rcv1subset20.127(6) ± 0.0170.125(7) ± 0.0100.125(8) ± 0.0100.133(5) ± 0.0150.138(2) ± 0.0210.135(4) ± 0.0150.137(3) ± 0.0170.141(1) ± 0.019 •
rcv1subset30.099(8) ± 0.0070.109(7) ± 0.0120.113(6) ± 0.0100.120(5) ± 0.0100.122(4) ± 0.0090.124(3) ± 0.0110.126(1) ± 0.007 •0.125(2) ± 0.009
rcv1subset40.138(8) ± 0.0160.148(7) ± 0.0170.151(4) ± 0.0150.149(6) ± 0.0130.151(5) ± 0.0170.154(3) ± 0.0130.154(1) ± 0.012 •0.154(2) ± 0.012
rcv1subset50.128(8) ± 0.0130.132(7) ± 0.0130.135(4) ± 0.0170.134(5) ± 0.0170.132(6) ± 0.0140.138(2) ± 0.0150.138(1) ± 0.014 •0.137(3) ± 0.014
bibtex0.176(8) ± 0.0060.187(3) ± 0.0040.187(1) ± 0.005 •0.187(2) ± 0.0060.185(5) ± 0.0050.186(4) ± 0.0070.184(6) ± 0.0060.184(7) ± 0.006
Arts0.236(8) ± 0.0140.252(7) ± 0.0140.258(6) ± 0.0150.263(5) ± 0.0150.269(3) ± 0.0140.273(1) ± 0.014 •0.272(2) ± 0.0140.268(4) ± 0.014
Health0.382(8) ± 0.0100.394(7) ± 0.0110.397(6) ± 0.0110.404(4) ± 0.0110.407(2) ± 0.0110.407(1) ± 0.012 •0.405(3) ± 0.0110.403(5) ± 0.011
Business0.245(8) ± 0.0070.255(7) ± 0.0080.259(6) ± 0.0080.267(5) ± 0.0050.279(3) ± 0.0090.281(1) ± 0.009 •0.279(4) ± 0.0080.280(2) ± 0.010
Education0.254(8) ± 0.0130.268(7) ± 0.0130.274(6) ± 0.0140.278(5) ± 0.0150.286(1) ± 0.013 •0.283(2) ± 0.0130.279(4) ± 0.0120.281(3) ± 0.011
Computers0.271(8) ± 0.0120.286(7) ± 0.0100.292(6) ± 0.0090.294(5) ± 0.0100.303(1) ± 0.014 •0.301(4) ± 0.0150.302(2) ± 0.0170.302(3) ± 0.014
Entertainment0.372(8) ± 0.0050.383(7) ± 0.0080.386(6) ± 0.0090.390(2) ± 0.0080.391(1) ± 0.010 •0.389(3) ± 0.0100.388(4) ± 0.0110.387(5) ± 0.011
Recreation0.367(8) ± 0.0150.378(7) ± 0.0170.384(6) ± 0.0110.390(5) ± 0.0160.399(3) ± 0.0140.400(1) ± 0.015 •0.399(2) ± 0.0150.398(4) ± 0.014
Society0.266(8) ± 0.0120.278(7) ± 0.0110.284(6) ± 0.0120.291(5) ± 0.0120.297(1) ± 0.010 •0.296(2) ± 0.0110.295(3) ± 0.0120.295(4) ± 0.013
eurlex-dc-l0.261(5) ± 0.0100.266(3) ± 0.0120.269(1) ± 0.012 •0.267(2) ± 0.0100.265(4) ± 0.0120.257(6) ± 0.0110.251(7) ± 0.0100.247(8) ± 0.011
eurlex-sm0.385(4) ± 0.0070.392(1) ± 0.010 •0.390(2) ± 0.0090.388(3) ± 0.0090.383(5) ± 0.0120.373(6) ± 0.0100.365(7) ± 0.0100.363(8) ± 0.010
tmc2007-5000.604(4) ± 0.0080.602(8) ± 0.0080.602(7) ± 0.0080.603(6) ± 0.0070.604(5) ± 0.0080.604(1) ± 0.007 •0.604(2) ± 0.0070.604(3) ± 0.007
mediamill0.332(1) ± 0.005 •0.319(2) ± 0.0040.312(3) ± 0.0030.293(4) ± 0.0050.257(5) ± 0.0050.224(6) ± 0.0080.202(7) ± 0.0070.192(8) ± 0.007
average rank5.554.954.704.273.843.774.274.66
win/tie/loss42/55/9948/78/7054/83/5965/90/4162/110/2466/90/4052/96/4847/94/55
Table A39. The performance of MLHiKNN across different values of β in terms of F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A39. The performance of MLHiKNN across different values of β in terms of F1 micro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetF1 Micro
β = 5/4β = 3/2β = 5/3β = 2β = 3β = 5β = 9β = 17
birds0.366(3) ± 0.0370.376(1) ± 0.044 •0.369(2) ± 0.0500.358(4) ± 0.0610.352(6) ± 0.0620.352(5) ± 0.0530.335(7) ± 0.0850.332(8) ± 0.084
CAL5000.336(7) ± 0.0090.335(8) ± 0.0140.339(6) ± 0.0100.342(4) ± 0.0140.341(5) ± 0.0120.342(2) ± 0.0140.342(3) ± 0.0130.343(1) ± 0.013 •
emotions0.680(1) ± 0.015 •0.675(2) ± 0.0130.672(4) ± 0.0130.672(3) ± 0.0110.668(5) ± 0.0160.665(8) ± 0.0150.666(6) ± 0.0160.665(7) ± 0.015
genbase0.956(8) ± 0.0150.962(7) ± 0.0140.965(6) ± 0.0120.965(5) ± 0.0140.970(4) ± 0.0120.974(3) ± 0.0110.976(1) ± 0.010 •0.976(2) ± 0.010
LLOG0.363(3) ± 0.059 •0.363(3) ± 0.059 •0.360(7) ± 0.0570.363(3) ± 0.059 •0.360(7) ± 0.0570.363(3) ± 0.059 •0.363(3) ± 0.059 •0.363(3) ± 0.059 •
enron0.444(8) ± 0.0120.452(7) ± 0.0120.454(6) ± 0.0140.460(5) ± 0.0130.464(4) ± 0.0100.466(3) ± 0.0110.467(1) ± 0.012 •0.467(2) ± 0.011
scene0.748(1) ± 0.009 •0.746(2) ± 0.0070.744(3) ± 0.0090.743(4) ± 0.0080.741(5) ± 0.0090.741(6) ± 0.0080.741(7) ± 0.0070.740(8) ± 0.008
yeast0.652(1) ± 0.008 •0.647(2) ± 0.0080.645(3) ± 0.0060.643(4) ± 0.0060.641(6) ± 0.0070.641(5) ± 0.0070.640(7) ± 0.0060.640(8) ± 0.006
Slashdot0.844(1) ± 0.007 •0.844(2) ± 0.0070.843(4) ± 0.0080.843(8) ± 0.0080.843(7) ± 0.0070.843(3) ± 0.0070.843(5) ± 0.0070.843(6) ± 0.007
corel5k0.051(4) ± 0.0090.051(1) ± 0.011 •0.051(2) ± 0.0120.050(5) ± 0.0110.051(3) ± 0.0100.048(8) ± 0.0100.048(7) ± 0.0090.049(6) ± 0.009
rcv1subset10.277(8) ± 0.0190.288(7) ± 0.0170.291(6) ± 0.0200.296(5) ± 0.0210.298(4) ± 0.0200.300(3) ± 0.0210.302(2) ± 0.0230.302(1) ± 0.023 •
rcv1subset20.325(7) ± 0.0120.330(6) ± 0.0090.324(8) ± 0.0100.334(3) ± 0.0120.336(1) ± 0.017 •0.333(4) ± 0.0140.332(5) ± 0.0140.335(2) ± 0.016
rcv1subset30.313(8) ± 0.0100.323(7) ± 0.0160.326(6) ± 0.0100.333(5) ± 0.0090.334(3) ± 0.0140.334(4) ± 0.0140.335(2) ± 0.0100.335(1) ± 0.011 •
rcv1subset40.382(8) ± 0.0180.394(6) ± 0.0240.398(3) ± 0.0220.394(7) ± 0.0190.396(5) ± 0.0220.399(1) ± 0.019 •0.397(4) ± 0.0170.398(2) ± 0.018
rcv1subset50.340(8) ± 0.0160.347(7) ± 0.0160.348(6) ± 0.0150.348(5) ± 0.0140.350(4) ± 0.0160.357(2) ± 0.0200.359(1) ± 0.022 •0.357(3) ± 0.020
bibtex0.360(8) ± 0.0070.367(3) ± 0.0060.368(1) ± 0.006 •0.367(2) ± 0.0060.365(4) ± 0.0040.364(5) ± 0.0050.363(7) ± 0.0040.363(6) ± 0.004
Arts0.344(8) ± 0.0170.358(7) ± 0.0160.365(6) ± 0.0140.369(5) ± 0.0140.372(1) ± 0.014 •0.371(2) ± 0.0140.371(3) ± 0.0140.370(4) ± 0.014
Health0.565(8) ± 0.0110.570(7) ± 0.0100.572(6) ± 0.0110.574(2) ± 0.0110.575(1) ± 0.009 •0.573(4) ± 0.0110.573(3) ± 0.0100.573(5) ± 0.010
Business0.718(8) ± 0.0040.720(7) ± 0.0040.721(6) ± 0.0030.722(5) ± 0.0040.723(1) ± 0.003 •0.723(3) ± 0.0040.723(2) ± 0.0030.722(4) ± 0.004
Education0.358(8) ± 0.0130.371(7) ± 0.0140.376(6) ± 0.0130.378(5) ± 0.0140.381(1) ± 0.013 •0.379(2) ± 0.0140.379(4) ± 0.0120.379(3) ± 0.013
Computers0.501(8) ± 0.0040.507(7) ± 0.0050.509(6) ± 0.0050.511(5) ± 0.0050.513(1) ± 0.006 •0.512(2) ± 0.0060.512(3) ± 0.0060.512(4) ± 0.005
Entertainment0.489(8) ± 0.0080.499(7) ± 0.0070.502(6) ± 0.0060.505(4) ± 0.0080.506(2) ± 0.0070.506(1) ± 0.006 •0.505(3) ± 0.0060.504(5) ± 0.006
Recreation0.439(8) ± 0.0120.449(7) ± 0.0120.452(5) ± 0.0100.454(3) ± 0.0120.456(1) ± 0.012 •0.454(2) ± 0.0130.453(4) ± 0.0130.452(6) ± 0.012
Society0.455(8) ± 0.0120.464(7) ± 0.0110.467(4) ± 0.0100.468(3) ± 0.0100.470(1) ± 0.009 •0.468(2) ± 0.0100.466(6) ± 0.0100.466(5) ± 0.010
eurlex-dc-l0.475(4) ± 0.0080.479(3) ± 0.0070.481(1) ± 0.007 •0.479(2) ± 0.0070.474(5) ± 0.0050.467(6) ± 0.0050.462(7) ± 0.0060.459(8) ± 0.005
eurlex-sm0.593(2) ± 0.0060.594(1) ± 0.007 •0.592(3) ± 0.0070.589(4) ± 0.0070.582(5) ± 0.0080.576(6) ± 0.0080.572(7) ± 0.0080.570(8) ± 0.008
tmc2007-5000.728(1) ± 0.003 •0.726(2) ± 0.0030.726(3) ± 0.0030.726(4) ± 0.0030.726(5) ± 0.0030.726(6) ± 0.0030.726(7) ± 0.0030.726(8) ± 0.003
mediamill0.655(1) ± 0.001 •0.649(2) ± 0.0010.644(3) ± 0.0010.635(4) ± 0.0010.618(5) ± 0.0010.606(6) ± 0.0020.601(7) ± 0.0020.598(8) ± 0.002
average rank5.594.844.594.233.663.844.454.80
win/tie/loss47/38/11162/57/7761/74/6172/95/2976/88/3266/85/4548/91/5743/90/63
Table A40. The performance of MLHiKNN across different values of β in terms of Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table A40. The performance of MLHiKNN across different values of β in terms of Hamming Loss. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetHamming Loss
β = 5/4β = 3/2β = 5/3β = 2β = 3β = 5β = 9β = 17
birds0.139(7) ± 0.0050.139(6) ± 0.0070.139(5) ± 0.0070.137(1) ± 0.008 •0.137(2) ± 0.0060.137(3) ± 0.0070.138(4) ± 0.0070.139(8) ± 0.007
CAL5000.187(3) ± 0.0010.187(8) ± 0.0010.187(6) ± 0.0010.187(2) ± 0.0010.186(1) ± 0.001 •0.187(7) ± 0.0010.187(5) ± 0.0010.187(4) ± 0.001
emotions0.186(1) ± 0.008 •0.189(2) ± 0.0060.190(4) ± 0.0060.190(3) ± 0.0050.192(5) ± 0.0070.194(8) ± 0.0070.194(6) ± 0.0070.194(7) ± 0.007
genbase0.008(8) ± 0.0030.007(7) ± 0.0020.006(6) ± 0.0020.006(5) ± 0.0020.005(4) ± 0.0020.005(3) ± 0.0020.004(1) ± 0.002 •0.004(1) ± 0.002 •
LLOG0.188(3) ± 0.006 •0.188(3) ± 0.006 •0.188(7) ± 0.0050.188(3) ± 0.006 •0.188(7) ± 0.0050.188(3) ± 0.006 •0.188(3) ± 0.006 •0.188(3) ± 0.006 •
enron0.076(8) ± 0.0010.076(7) ± 0.0010.075(4) ± 0.0010.075(5) ± 0.0010.075(1) ± 0.001 •0.075(6) ± 0.0010.075(2) ± 0.0010.075(3) ± 0.001
scene0.083(1) ± 0.003 •0.084(2) ± 0.0020.084(3) ± 0.0030.084(4) ± 0.0030.085(5) ± 0.0030.085(6) ± 0.0020.085(7) ± 0.0020.085(8) ± 0.002
yeast0.195(1) ± 0.005 •0.198(2) ± 0.0040.199(3) ± 0.0030.200(4) ± 0.0030.201(5) ± 0.0040.201(6) ± 0.0040.201(7) ± 0.0040.201(8) ± 0.004
Slashdot 0.028(1) ± 0.001 •0.028(2) ± 0.0010.029(4) ± 0.0020.029(8) ± 0.0010.029(7) ± 0.0010.029(3) ± 0.0010.029(5) ± 0.0010.029(6) ± 0.001
corel5k0.021(1) ± 0.000 •0.021(2) ± 0.0000.021(7) ± 0.0000.021(4) ± 0.0000.021(3) ± 0.0000.021(6) ± 0.0000.021(5) ± 0.0000.021(8) ± 0.000
rcv1subset10.033(8) ± 0.0000.033(6) ± 0.0000.033(4) ± 0.0000.032(1) ± 0.000 •0.032(3) ± 0.0000.032(2) ± 0.0000.033(5) ± 0.0000.033(7) ± 0.000
rcv1subset20.027(6) ± 0.0000.027(1) ± 0.000 •0.027(3) ± 0.0000.027(2) ± 0.0000.027(5) ± 0.0000.027(4) ± 0.0000.027(7) ± 0.0000.027(8) ± 0.000
rcv1subset30.027(8) ± 0.0000.027(6) ± 0.0000.027(4) ± 0.0000.027(3) ± 0.0000.027(1) ± 0.000 •0.027(2) ± 0.0000.027(5) ± 0.0000.027(7) ± 0.000
rcv1subset40.024(7) ± 0.0000.024(4) ± 0.0000.024(1) ± 0.000 •0.024(2) ± 0.0000.024(5) ± 0.0000.024(3) ± 0.0000.024(8) ± 0.0000.024(5) ± 0.000
rcv1subset50.027(3) ± 0.0000.027(5) ± 0.0000.027(4) ± 0.0000.027(1) ± 0.000 •0.027(2) ± 0.0000.027(6) ± 0.0000.027(7) ± 0.0000.027(8) ± 0.000
bibtex0.013(1) ± 0.000 •0.013(3) ± 0.0000.013(2) ± 0.0000.013(4) ± 0.0000.013(5) ± 0.0000.013(6) ± 0.0000.013(7) ± 0.0000.013(8) ± 0.000
Arts0.060(8) ± 0.0010.060(7) ± 0.0010.059(6) ± 0.0010.059(4) ± 0.0010.059(1) ± 0.000 •0.059(2) ± 0.0000.059(3) ± 0.0000.059(5) ± 0.000
Health0.056(8) ± 0.0010.055(7) ± 0.0010.055(6) ± 0.0010.055(2) ± 0.0010.055(1) ± 0.001 •0.055(3) ± 0.0010.055(4) ± 0.0010.055(5) ± 0.001
Business0.031(8) ± 0.0000.031(6) ± 0.0000.031(4) ± 0.0000.031(1) ± 0.000 •0.031(2) ± 0.0000.031(3) ± 0.0000.031(5) ± 0.0000.031(7) ± 0.000
Education0.047(8) ± 0.0010.046(7) ± 0.0010.046(6) ± 0.0010.046(2) ± 0.0010.046(1) ± 0.001 •0.046(3) ± 0.0010.046(4) ± 0.0010.046(5) ± 0.001
Computers0.043(8) ± 0.0000.043(7) ± 0.0000.043(5) ± 0.0000.043(2) ± 0.0000.043(1) ± 0.001 •0.043(3) ± 0.0010.043(6) ± 0.0010.043(4) ± 0.001
Entertainment0.062(8) ± 0.0010.061(7) ± 0.0010.061(6) ± 0.0010.061(3) ± 0.0010.061(1) ± 0.001 •0.061(2) ± 0.0010.061(4) ± 0.0010.061(5) ± 0.001
Recreation0.051(8) ± 0.0010.050(7) ± 0.0010.050(6) ± 0.0010.050(3) ± 0.0010.050(1) ± 0.001 •0.050(2) ± 0.0010.050(4) ± 0.0010.050(5) ± 0.001
Society0.052(8) ± 0.0010.051(7) ± 0.0010.051(6) ± 0.0010.051(2) ± 0.0010.051(1) ± 0.001 •0.051(3) ± 0.0010.051(4) ± 0.0010.051(5) ± 0.001
eurlex-dc-l0.005(4) ± 0.0000.004(1) ± 0.000 •0.004(2) ± 0.0000.004(3) ± 0.0000.005(5) ± 0.0000.005(6) ± 0.0000.005(7) ± 0.0000.005(8) ± 0.000
eurlex-sm0.011(2) ± 0.0000.011(1) ± 0.000 •0.011(3) ± 0.0000.011(4) ± 0.0000.011(5) ± 0.0000.011(6) ± 0.0000.011(7) ± 0.0000.011(8) ± 0.000
tmc2007-5000.051(1) ± 0.000 •0.051(2) ± 0.0000.051(3) ± 0.0000.051(4) ± 0.0000.051(5) ± 0.0000.051(6) ± 0.0000.051(7) ± 0.0000.051(8) ± 0.000
mediamill0.027(1) ± 0.000 •0.027(2) ± 0.0000.027(3) ± 0.0000.028(4) ± 0.0000.029(5) ± 0.0000.030(6) ± 0.0000.030(7) ± 0.0000.030(8) ± 0.000
average rank4.984.554.413.093.254.275.256.20
win/tie/loss55/69/7253/90/5357/101/3865/108/2357/102/3749/108/3930/115/5123/98/75

References

  1. Huang, A.; Xu, R.; Chen, Y.; Guo, M. Research on multi-label user classification of social media based on ML-KNN algorithm. Technol. Forecast. Soc. Change 2023, 188, 122271. [Google Scholar] [CrossRef]
  2. George, M.; Floerkemeier, C. Recognizing Products: A Per-exemplar Multi-label Image Classification Approach. In Proceedings of the Computer Vision—ECCV 2014, Zurich, Switzerland, 6–12 September 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer: Cham, Switzerland, 2014; pp. 440–455. [Google Scholar] [CrossRef]
  3. Zhang, J.; Zhang, J.; Dai, T.; He, Z. Exploring Weighted Dual Graph Regularized Non-Negative Matrix Tri-Factorization Based Collaborative Filtering Framework for Multi-Label Annotation of Remote Sensing Images. Remote Sens. 2019, 11, 922. [Google Scholar] [CrossRef]
  4. Chalkidis, I.; Fergadiotis, E.; Malakasiotis, P.; Androutsopoulos, I. Large-Scale Multi-Label Text Classification on EU Legislation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Stroudsburg, PA, USA, 28 July–2 August 2019; pp. 6314–6322. [Google Scholar] [CrossRef]
  5. Spyromitros, E.; Tsoumakas, G.; Vlahavas, I. An Empirical Study of Lazy Multilabel Classification Algorithms. In Proceedings of the Artificial Intelligence: Theories, Models and Applications, Syros, Greece, 2–4 October 2008; Darzentas, J., Vouros, G.A., Vosinakis, S., Arnellos, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2008; pp. 401–406. [Google Scholar] [CrossRef]
  6. Zhang, M.L.; Zhou, Z.H. ML-KNN: A lazy learning approach to multi-label learning. Pattern Recognit. 2007, 40, 2038–2048. [Google Scholar] [CrossRef]
  7. Younes, Z.; Abdallah, F.; Denoeux, T. Multi-label classification algorithm derived from K-nearest neighbor rule with label dependencies. In Proceedings of the 2008 16th European Signal Processing Conference, Lausanne, Switzeerland, 25–19 August 2008; pp. 1–5. [Google Scholar]
  8. Cheng, W.; Hüllermeier, E. Combining instance-based learning and logistic regression for multilabel classification. Mach. Learn. 2009, 76, 211–225. [Google Scholar] [CrossRef]
  9. Madjarov, G.; Kocev, D.; Gjorgjevikj, D.; Džeroski, S. An extensive experimental comparison of methods for multi-label learning. Pattern Recognit. 2012, 45, 3084–3104. [Google Scholar] [CrossRef]
  10. Bogatinovski, J.; Todorovski, L.; Džeroski, S.; Kocev, D. Comprehensive comparative study of multi-label classification methods. Expert Syst. Appl. 2022, 203, 117215. [Google Scholar] [CrossRef]
  11. Radovanović, M.; Nanopoulos, A.; Ivanović, M. Hubs in space: Popular nearest neighbors in high-dimensional data. J. Mach. Learn. Res. 2010, 11, 2487–2531. [Google Scholar]
  12. Feldbauer, R.; Flexer, A. A comprehensive empirical comparison of hubness reduction in high-dimensional spaces. Knowl. Inf. Syst. 2019, 59, 137–166. [Google Scholar] [CrossRef]
  13. Zhang, M.L.; Zhou, Z.H. A Review on Multi-Label Learning Algorithms. IEEE Trans. Knowl. Data Eng. 2014, 26, 1819–1837. [Google Scholar] [CrossRef]
  14. Boutell, M.R.; Luo, J.; Shen, X.; Brown, C.M. Learning multi-label scene classification. Pattern Recognit. 2004, 37, 1757–1771. [Google Scholar] [CrossRef]
  15. Fürnkranz, J.; Hüllermeier, E.; Mencía, E.L.; Brinker, K. Multilabel classification via calibrated label ranking. Mach. Learn. 2008, 73, 133–153. [Google Scholar] [CrossRef]
  16. Hüllermeier, E.; Fürnkranz, J.; Cheng, W.; Brinker, K. Label ranking by learning pairwise preferences. Artif. Intell. 2008, 172, 1897–1916. [Google Scholar] [CrossRef]
  17. Read, J.; Pfahringer, B.; Holmes, G.; Frank, E. Classifier chains for multi-label classification. Mach. Learn. 2011, 85, 333–359. [Google Scholar] [CrossRef]
  18. Tsoumakas, G.; Katakis, I.; Vlahavas, I. Random k-Labelsets for Multilabel Classification. IEEE Trans. Knowl. Data Eng. 2011, 23, 1079–1089. [Google Scholar] [CrossRef]
  19. Clare, A.; King, R.D. Knowledge Discovery in Multi-label Phenotype Data. In Proceedings of the Principles of Data Mining and Knowledge Discovery, Freiburg, Germany, 3–5 September 2001; De Raedt, L., Siebes, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2001; pp. 42–53. [Google Scholar] [CrossRef]
  20. Elisseeff, A.; Weston, J. A Kernel Method for Multi-Labelled Classification. In Proceedings of the 14th International Conference on Neural Information Processing Systems: Natural and Synthetic, Cambridge, MA, USA, 3–8 December 2001; NIPS’01. pp. 681–687. [Google Scholar]
  21. Zhang, M.L. Ml-rbf: RBF Neural Networks for Multi-Label Learning. Neural Process. Lett. 2009, 29, 61–74. [Google Scholar] [CrossRef]
  22. Cuevas-Muñoz, J.M.; García-Pedrajas, N.E. ML-k’sNN: Label Dependent k Values for Multi-Label k-Nearest Neighbor Rule. Mathematics 2023, 11, 275. [Google Scholar] [CrossRef]
  23. Liu, W.; Wang, H.; Shen, X.; Tsang, I.W. The Emerging Trends of Multi-Label Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 7955–7974. [Google Scholar] [CrossRef]
  24. Tarekegn, A.N.; Giacobini, M.; Michalak, K. A review of methods for imbalanced multi-label classification. Pattern Recognit. 2021, 118, 107965. [Google Scholar] [CrossRef]
  25. Qian, K.; Min, X.Y.; Cheng, Y.; Min, F. Weight matrix sharing for multi-label learning. Pattern Recognit. 2023, 136, 109156. [Google Scholar] [CrossRef]
  26. Bakhshi, S.; Can, F. Balancing efficiency vs. effectiveness and providing missing label robustness in multi-label stream classification. Knowl.-Based Syst. 2024, 289, 111489. [Google Scholar] [CrossRef]
  27. Xie, M.K.; Huang, S.J. Partial multi-label learning. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, AAAI 2018, New Orleans, LA, USA, 2–7 February 2018; pp. 4302–4309. [Google Scholar] [CrossRef]
  28. Hu, Y.; Fang, X.; Kang, P.; Chen, Y.; Fang, Y.; Xie, S. Dual Noise Elimination and Dynamic Label Correlation Guided Partial Multi-Label Learning. IEEE Trans. Multimed. 2024, 26, 5641–5656. [Google Scholar] [CrossRef]
  29. Xu, C.; Tao, D.; Xu, C. Robust Extreme Multi-label Learning. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 13–17 August 2016; KDD ’16. pp. 1275–1284. [Google Scholar] [CrossRef]
  30. Mittal, A.; Sachdeva, N.; Agrawal, S.; Agarwal, S.; Kar, P.; Varma, M. ECLARE: Extreme Classification with Label Graph Correlations. In Proceedings of the Web Conference 2021, New York, NY, USA, 19–23 April 2021; WWW ’21. pp. 3721–3732. [Google Scholar] [CrossRef]
  31. Venkatesan, R.; Er, M.J.; Wu, S.; Pratama, M. A novel online real-time classifier for multi-label data streams. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 1833–1840. [Google Scholar] [CrossRef]
  32. Esteban, A.; Cano, A.; Zafra, A.; Ventura, S. Hoeffding adaptive trees for multi-label classification on data streams. Knowl.-Based Syst. 2024, 304, 112561. [Google Scholar] [CrossRef]
  33. Aucouturier, J.J.; Pachet, F. Improving timbre similarity: How high is the sky. J. Negat. Results Speech Audio Sci. 2004, 1, 1–13. [Google Scholar]
  34. Hara, K.; Suzuki, I.; Kobayashi, K.; Fukumizu, K.; Radovanovíc, M. Flattening the density gradient for eliminating spatial centrality to reduce hubness. In Proceedings of the 30th AAAI Conference on Artificial Intelligence, AAAI 2016, Phoenix, AZ, USA, 12–17 February 2016; pp. 1659–1665. [Google Scholar] [CrossRef]
  35. Aryal, S.; Ting, K.M.; Washio, T.; Haffari, G. Data-dependent dissimilarity measure: An effective alternative to geometric distance measures. Knowl. Inf. Syst. 2017, 53, 479–506. [Google Scholar] [CrossRef]
  36. Tomašev, N.; Mladenić, D. Hubness-aware shared neighbor distances for high-dimensional k-nearest neighbor classification. Knowl. Inf. Syst. 2014, 39, 89–122. [Google Scholar] [CrossRef]
  37. Pal, A.K.; Mondal, P.K.; Ghosh, A.K. High dimensional nearest neighbor classification based on mean absolute differences of inter-point distances. Pattern Recognit. Lett. 2016, 74, 1–8. [Google Scholar] [CrossRef]
  38. Aggarwal, C.C.; Hinneburg, A.; Keim, D.A. On the Surprising Behavior of Distance Metrics in High Dimensional Space. In Proceedings of the Database Theory—ICDT 2001, London, UK, 4–6 January 2001; Van den Bussche, J., Vianu, V., Eds.; Springer: Berlin/Heidelberg, Germany, 2001; pp. 420–434. [Google Scholar] [CrossRef]
  39. Flexer, A.; Schnitzer, D. Choosing ℓp norms in high-dimensional spaces based on hub analysis. Neurocomputing 2015, 169, 281–287. [Google Scholar] [CrossRef]
  40. Schnitzer, D.; Flexer, A.; Schedl, M.; Widmer, G. Local and Global Scaling Reduce Hubs in Space. J. Mach. Learn. Res. 2012, 13, 2871–2902. [Google Scholar]
  41. Feldbauer, R.; Leodolter, M.; Plant, C.; Flexer, A. Fast Approximate Hubness Reduction for Large High-Dimensional Data. In Proceedings of the 2018 IEEE International Conference on Big Knowledge (ICBK), Singapore, 17–18 November 2018; pp. 358–367. [Google Scholar] [CrossRef]
  42. Suzuki, I.; Hara, K.; Shimbo, M.; Saerens, M.; Fukumizu, K. Centering similarity measures to reduce hubs. In Proceedings of the EMNLP 2013—2013 Conference on Empirical Methods in Natural Language Processing, Seattle, WA, USA, 18–21 October 2013; Proceedings of the Conference. pp. 613–623. [Google Scholar]
  43. Hara, K.; Suzuki, I.; Shimbo, M.; Kobayashi, K.; Fukumizu, K.; Radovanovic, M. Localized centering: Reducing hubness in large-sample data. Proc. Natl. Conf. Artif. Intell. 2015, 4, 2645–2651. [Google Scholar] [CrossRef]
  44. Obraczka, D.; Rahm, E. An Evaluation of Hubness Reduction Methods for Entity Alignment with Knowledge Graph Embeddings. Int. Jt. Conf. Knowl. Discov. Knowl. Eng. Knowl. Manag. IC3K Proc. 2021, 2, 28–39. [Google Scholar] [CrossRef]
  45. Amblard, E.; Bac, J.; Chervov, A.; Soumelis, V.; Zinovyev, A. Hubness reduction improves clustering and trajectory inference in single-cell transcriptomic data. Bioinformatics 2022, 38, 1045–1051. [Google Scholar] [CrossRef] [PubMed]
  46. Nielsen, B.M.; Hansen, L.K. Hubness Reduction Improves Sentence-BERT Semantic Spaces. Proc. Mach. Learn. Res. 2024, 233. [Google Scholar]
  47. Radovanović, M.; Nanopoulos, A.; Ivanović, M. Nearest neighbors in high-dimensional data: The emergence and influence of hubs. In Proceedings of the 26th Annual International Conference on Machine Learning, New York, NY, USA, 14–18 June 2009; ICML ’09. pp. 865–872. [Google Scholar] [CrossRef]
  48. Tomasev, N.; Radovanović, M.; Mladenić, D.; Ivanović, M. A probabilistic approach to nearest-neighbor classification: Naive hubness bayesian kNN. In Proceedings of the 20th ACM International Conference on Information and Knowledge Management, New York, NY, USA, 24–28 October 2011; CIKM ’11. pp. 2173–2176. [Google Scholar] [CrossRef]
  49. Tomašev, N.; Radovanović, M.; Mladenić, D.; Ivanović, M. Hubness-based fuzzy measures for high-dimensional k-nearest neighbor classification. Int. J. Mach. Learn. Cybern. 2014, 5, 445–458. [Google Scholar]
  50. Tomašev, N.; Mladenić, D. Nearest neighbor voting in high dimensional data: Learning from past occurrences. Comput. Sci. Inf. Syst. 2012, 9, 691–712. [Google Scholar] [CrossRef]
  51. Buza, K.; Nanopoulos, A.; Nagy, G. Nearest neighbor regression in the presence of bad hubs. Knowl.-Based Syst. 2015, 86, 250–260. [Google Scholar] [CrossRef]
  52. Tomasev, N.; Radovanovic, M.; Mladenic, D.; Ivanovic, M. The Role of Hubness in Clustering High-Dimensional Data. IEEE Trans. Knowl. Data Eng. 2014, 26, 739–751. [Google Scholar] [CrossRef]
  53. Radovanović, M.; Nanopoulos, A.; Ivanović, M. Reverse Nearest Neighbors in Unsupervised Distance-Based Outlier Detection. IEEE Trans. Knowl. Data Eng. 2015, 27, 1369–1382. [Google Scholar] [CrossRef]
  54. Liu, H.; Zhang, S.; Wu, Z.; Li, X. Outlier detection using local density and global structure. Pattern Recognit. 2025, 157, 110947. [Google Scholar] [CrossRef]
  55. Shigeto, Y.; Suzuki, I.; Hara, K.; Shimbo, M.; Matsumoto, Y. Ridge Regression, Hubness, and Zero-Shot Learning. In Proceedings of the Machine Learning and Knowledge Discovery in Databases, Porto, Portugal, 7–11 September 2015; Appice, A., Rodrigues, P.P., Santos Costa, V., Soares, C., Gama, J., Jorge, A., Eds.; Springer: Cham, Switzerland, 2015; pp. 135–151. [Google Scholar] [CrossRef]
  56. Dinu, G.; Lazaridou, A.; Baroni, M. Improving zero-shot learning by mitigating the hubness problem. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Workshop Track Proceedings, San Diego, CA, USA, 7–9 May 2015; pp. 1–10. [Google Scholar]
  57. Zhang, L.; Xiang, T.; Gong, S. Learning a Deep Embedding Model for Zero-Shot Learning. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3010–3019. [Google Scholar] [CrossRef]
  58. Luo, C.; Li, Z.; Huang, K.; Feng, J.; Wang, M. Zero-Shot Learning via Attribute Regression and Class Prototype Rectification. IEEE Trans. Image Process. 2018, 27, 637–648. [Google Scholar] [CrossRef]
  59. Paul, A.; Krishnan, N.C.; Munjal, P. Semantically Aligned Bias Reducing Zero Shot Learning. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7049–7058. [Google Scholar] [CrossRef]
  60. Buza, K.; Peška, L. Drug–target interaction prediction with Bipartite Local Models and hubness-aware regression. Neurocomputing 2017, 260, 284–293. [Google Scholar] [CrossRef]
  61. Wang, D.; Yih, Y.; Ventresca, M. Improving neighbor-based collaborative filtering by using a hybrid similarity measurement. Expert Syst. Appl. 2020, 160, 113651. [Google Scholar] [CrossRef]
  62. Tian, M.; Giunchiglia, F.; Song, R.; Xu, H. Guiding ontology translation with hubness-aware translation memory. Expert Syst. Appl. 2025, 264, 125650. [Google Scholar] [CrossRef]
  63. Tsoumakas, G.; Spyromitros-Xioufis, E.; Vilcek, J.; Vlahavas, I. MULAN: A Java Library for Multi-Label Learning. J. Mach. Learn. Res. 2011, 12, 2411–2414. [Google Scholar]
  64. Read, J.; Reutemann, P.; Pfahringer, B.; Holmes, G. MEKA: A Multi-label/Multi-target Extension to WEKA. J. Mach. Learn. Res. 2016, 17, 1–5. [Google Scholar]
  65. Sun, L.; Ji, S.; Ye, J. Hypergraph spectral learning for multi-label classification. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2008; KDD ’08. pp. 668–676. [Google Scholar] [CrossRef]
  66. Zhang, M.L.; Li, Y.K.; Yang, H.; Liu, X.Y. Towards Class-Imbalance Aware Multi-Label Learning. IEEE Trans. Cybern. 2022, 52, 4459–4471. [Google Scholar] [CrossRef] [PubMed]
  67. Szymański, P.; Kajdanowicz, T. A scikit-based Python environment for performing multi-label classification. arXiv 2017, arXiv:1702.01460. [Google Scholar]
  68. Hall, M.; Frank, E.; Holmes, G.; Pfahringer, B.; Reutemann, P.; Witten, I.H. The WEKA Data Mining Software: An Update. SIGKDD Explor. Newsl. 2009, 11, 10–18. [Google Scholar] [CrossRef]
  69. Feldbauer, R.; Rattei, T.; Flexer, A. scikit-hubness: Hubness Reduction and Approximate Neighbor Search. J. Open Source Softw. 2020, 5, 1957. [Google Scholar] [CrossRef]
  70. Benavoli, A.; Corani, G.; Mangili, F. Should We Really Use Post-Hoc Tests Based on Mean-Ranks? J. Mach. Learn. Res. 2016, 17, 152–161. [Google Scholar]
Figure 1. Empirical distributions of O k and the k-occurrence skewness (for the datasets Slashdot, bibtex, and eurlex-dc-leaves, O k and S k are computed based on the dimensionality-reduced feature space), where k = 10 , of nine MLDs with different dimensionalities.
Figure 1. Empirical distributions of O k and the k-occurrence skewness (for the datasets Slashdot, bibtex, and eurlex-dc-leaves, O k and S k are computed based on the dimensionality-reduced feature space), where k = 10 , of nine MLDs with different dimensionalities.
Mathematics 13 01202 g001
Figure 2. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of AUC macro.
Figure 2. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of AUC macro.
Mathematics 13 01202 g002
Figure 3. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of AUC micro.
Figure 3. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of AUC micro.
Mathematics 13 01202 g003
Figure 4. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of Ranking Loss.
Figure 4. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of Ranking Loss.
Mathematics 13 01202 g004
Figure 5. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of F1 macro.
Figure 5. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of F1 macro.
Mathematics 13 01202 g005
Figure 6. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of F1 micro.
Figure 6. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of F1 micro.
Mathematics 13 01202 g006
Figure 7. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of Hamming Loss.
Figure 7. The average rank across all the datasets of each combination of algorithm and k out of twenty-five combinations in terms of Hamming Loss.
Mathematics 13 01202 g007
Figure 8. The average rank across all the datasets of each algorithm for hubness reduction technique comparisons in terms of each metric.
Figure 8. The average rank across all the datasets of each algorithm for hubness reduction technique comparisons in terms of each metric.
Mathematics 13 01202 g008
Figure 9. The average ranks across all the datasets of MLHiKNN and variant approaches in terms of each metric.
Figure 9. The average ranks across all the datasets of MLHiKNN and variant approaches in terms of each metric.
Mathematics 13 01202 g009
Figure 10. Results of k analysis experiments for MLHiKNN.
Figure 10. Results of k analysis experiments for MLHiKNN.
Mathematics 13 01202 g010
Figure 11. Results of λ analysis experiments for MLHiKNN.
Figure 11. Results of λ analysis experiments for MLHiKNN.
Mathematics 13 01202 g011
Figure 12. Results of β analysis experiments for MLHiKNN.
Figure 12. Results of β analysis experiments for MLHiKNN.
Mathematics 13 01202 g012
Table 1. Datasets used in the experiments. k is set to be 10 for computing S k , H o c c k , and A o c c k . #Attr denotes the number of attributes.
Table 1. Datasets used in the experiments. k is set to be 10 for computing S k , H o c c k , and A o c c k . #Attr denotes the number of attributes.
Dataset#Samples#Attr#LabelsLCardLDenLDivPLDiv S k H occ k A occ k Domain
birds327260121.760.15920.28-0.220.010.0000audio
CAL5005026812425.060.205021.000.580.120.0000music
emotions5937261.870.31270.050.190.010.0000music
genbase6621186131.130.09140.02−0.060.000.0000biology
LLOG14291007516.280.2212130.8511.381.000.9685text
enron17021001353.280.096750.409.920.650.0000text
scene240729461.070.18150.014.480.520.0000image
yeast2417103144.240.301980.081.550.260.0000biology
Slashdot277153121.180.10520.0215.780.890.7694text
corel5k49854991533.180.0226260.5324.640.790.0000images
rcv1subset16000472802.840.049770.168.210.420.0010text
rcv1subset26000472832.600.039030.155.280.410.0047text
rcv1subset36000472832.580.038990.156.170.410.0048text
rcv1subset46000472822.460.037620.135.420.400.0053text
rcv1subset56000472842.610.039020.156.020.410.0097text
bibtex73951841592.400.0228560.3912.080.520.0368text
Arts7484462241.650.075950.0817.880.440.0345text
Health9205612211.640.083140.0314.840.440.0214text
Business11,204438251.600.062210.0226.530.540.1252text
Education12,016550271.460.054880.0412.240.430.0137text
Computers12,434681271.500.064030.0318.040.460.0314text
Entertainment12,730640171.410.083170.0223.590.450.0324text
Recreation12,828606221.430.065300.0428.080.460.0677text
Society14,512636261.670.0610530.0718.500.450.0300text
eurlex-dc-l (In the tables of this paper, the dataset eurlex-dc-leaves is denoted by eurlex-dc-l for short)18,5932502151.260.0111150.0623.240.320.0575text
eurlex-sm19,2822501362.200.0223140.123.680.280.0200text
tmc2007-50028,596500222.220.1011720.0414.160.620.0002text
mediamill42,1771201014.560.0565540.160.980.160.0000video
Table 2. Categories of the metrics used in the evaluation.
Table 2. Categories of the metrics used in the evaluation.
PerspectiveMacro-AveragingMicro-AveragingExample-Based
RankingAUC macroAUC microRanking Loss
ClassificationF1 macroF1 microHamming Loss
Table 3. The performance of nine compared MLC algorithms and MLHiKNN on AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table 3. The performance of nine compared MLC algorithms and MLHiKNN on AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Macro
BR CC ECC RAkEL RAkELd BRkNNa MLKNN DMLkNN IBLR MLHiKNN
birds0.624(9) ± 0.0300.625(8) ± 0.0200.753(1) ± 0.017 •0.723(3) ± 0.0160.651(7) ± 0.0140.702(4) ± 0.0160.606(10) ± 0.0190.684(6) ± 0.0210.691(5) ± 0.0170.728(2) ± 0.030
CAL5000.511(9) ± 0.0050.508(10) ± 0.0050.537(3) ± 0.0060.527(5) ± 0.0040.513(8) ± 0.0030.543(2) ± 0.0060.515(7) ± 0.0040.530(4) ± 0.0030.516(6) ± 0.0080.571(1) ± 0.006 •
emotions0.690(8) ± 0.0180.680(10) ± 0.0160.818(4) ± 0.0080.793(6) ± 0.0090.689(9) ± 0.0140.834(3) ± 0.0050.723(7) ± 0.0130.805(5) ± 0.0090.836(2) ± 0.0100.846(1) ± 0.005 •
genbase0.989(8) ± 0.0100.987(9) ± 0.0140.996(4) ± 0.0050.997(2) ± 0.0040.994(7) ± 0.0080.996(5) ± 0.0040.963(10) ± 0.0130.998(1) ± 0.001 •0.995(6) ± 0.0060.997(3) ± 0.003
LLOG0.526(9) ± 0.0020.525(10) ± 0.0040.539(6) ± 0.0060.531(8) ± 0.0020.536(7) ± 0.0050.598(4) ± 0.0060.588(5) ± 0.0060.602(3) ± 0.0060.602(2) ± 0.0060.603(1) ± 0.006 •
enron0.580(9) ± 0.0080.576(10) ± 0.0040.646(2) ± 0.0050.631(5) ± 0.0050.585(8) ± 0.0090.627(6) ± 0.0120.596(7) ± 0.0110.634(4) ± 0.0140.641(3) ± 0.0170.676(1) ± 0.012 •
scene0.761(8) ± 0.0150.752(10) ± 0.0140.921(5) ± 0.0040.894(7) ± 0.0030.760(9) ± 0.0110.937(3) ± 0.0030.898(6) ± 0.0070.932(4) ± 0.0040.940(2) ± 0.0030.944(1) ± 0.002 •
yeast0.569(8) ± 0.0060.562(10) ± 0.0070.641(5) ± 0.0050.625(6) ± 0.0040.562(9) ± 0.0100.685(3) ± 0.0050.583(7) ± 0.0060.671(4) ± 0.0060.692(2) ± 0.0060.711(1) ± 0.006 •
Slashdot0.560(8) ± 0.0190.581(6) ± 0.0210.535(10) ± 0.0140.550(9) ± 0.0100.595(5) ± 0.0190.680(3) ± 0.0170.563(7) ± 0.0330.671(4) ± 0.0340.682(2) ± 0.0350.723(1) ± 0.021 •
corel5k0.522(6) ± 0.0020.516(10) ± 0.0020.520(8) ± 0.0020.516(9) ± 0.0010.522(7) ± 0.0020.582(4) ± 0.0040.569(5) ± 0.0040.587(3) ± 0.0070.603(2) ± 0.0080.690(1) ± 0.005 •
rcv1subset10.642(9) ± 0.0080.614(10) ± 0.0050.728(6) ± 0.0040.661(7) ± 0.0050.649(8) ± 0.0110.819(4) ± 0.0060.766(5) ± 0.0060.835(2) ± 0.0050.834(3) ± 0.0050.891(1) ± 0.004 •
rcv1subset20.645(9) ± 0.0070.595(10) ± 0.0090.721(6) ± 0.0050.652(8) ± 0.0030.653(7) ± 0.0060.819(3) ± 0.0080.770(5) ± 0.0070.841(2) ± 0.0060.810(4) ± 0.0070.888(1) ± 0.005 •
rcv1subset30.635(9) ± 0.0070.584(10) ± 0.0080.718(6) ± 0.0050.646(8) ± 0.0040.648(7) ± 0.0060.816(3) ± 0.0070.767(5) ± 0.0070.834(2) ± 0.0080.816(4) ± 0.0050.880(1) ± 0.004 •
rcv1subset40.643(8) ± 0.0120.600(10) ± 0.0070.714(6) ± 0.0040.642(9) ± 0.0050.652(7) ± 0.0100.816(3) ± 0.0060.769(5) ± 0.0060.840(2) ± 0.0070.811(4) ± 0.0070.887(1) ± 0.007 •
rcv1subset50.640(9) ± 0.0100.594(10) ± 0.0070.714(6) ± 0.0050.647(8) ± 0.0030.649(7) ± 0.0080.810(4) ± 0.0060.758(5) ± 0.0050.834(2) ± 0.0060.812(3) ± 0.0080.873(1) ± 0.005 •
bibtex0.639(7) ± 0.0070.634(8) ± 0.0070.661(6) ± 0.0030.617(10) ± 0.0040.626(9) ± 0.0060.794(4) ± 0.0030.754(5) ± 0.0040.807(2) ± 0.0030.806(3) ± 0.0050.886(1) ± 0.001 •
Arts0.598(8) ± 0.0050.594(9) ± 0.0070.650(5) ± 0.0040.631(6) ± 0.0040.601(7) ± 0.0040.695(3) ± 0.0090.545(10) ± 0.0070.685(4) ± 0.0090.704(2) ± 0.0130.730(1) ± 0.009 •
Health0.687(8) ± 0.0080.686(9) ± 0.0090.731(5) ± 0.0060.713(6) ± 0.0050.687(7) ± 0.0110.745(3) ± 0.0080.606(10) ± 0.0080.737(4) ± 0.0100.754(2) ± 0.0110.778(1) ± 0.009 •
Business0.609(7) ± 0.0100.597(9) ± 0.0070.641(5) ± 0.0090.628(6) ± 0.0050.609(8) ± 0.0100.701(3) ± 0.0090.589(10) ± 0.0080.704(2) ± 0.0090.699(4) ± 0.0110.752(1) ± 0.007 •
Education0.601(9) ± 0.0110.607(7) ± 0.0110.648(5) ± 0.0050.635(6) ± 0.0060.606(8) ± 0.0150.698(2) ± 0.0090.595(10) ± 0.0060.683(4) ± 0.0100.696(3) ± 0.0140.748(1) ± 0.014 •
Computers0.618(8) ± 0.0070.602(10) ± 0.0090.674(5) ± 0.0080.657(6) ± 0.0070.620(7) ± 0.0100.707(4) ± 0.0060.611(9) ± 0.0040.710(3) ± 0.0060.715(2) ± 0.0090.750(1) ± 0.005 •
Entertainment0.654(9) ± 0.0100.661(8) ± 0.0110.724(5) ± 0.0070.694(6) ± 0.0050.677(7) ± 0.0130.743(4) ± 0.0050.581(10) ± 0.0080.749(3) ± 0.0040.754(2) ± 0.0060.782(1) ± 0.004 •
Recreation0.648(8) ± 0.0080.647(9) ± 0.0090.709(5) ± 0.0060.678(6) ± 0.0060.658(7) ± 0.0110.752(3) ± 0.0060.559(10) ± 0.0120.751(4) ± 0.0100.762(2) ± 0.0090.788(1) ± 0.009 •
Society0.593(8) ± 0.0060.585(9) ± 0.0040.636(5) ± 0.0040.622(6) ± 0.0040.594(7) ± 0.0060.675(3) ± 0.0050.554(10) ± 0.0070.673(4) ± 0.0050.683(2) ± 0.0060.707(1) ± 0.006 •
eurlex-dc-l0.635(8) ± 0.0050.630(10) ± 0.0060.676(6) ± 0.0020.637(7) ± 0.0040.634(9) ± 0.0050.820(2) ± 0.0050.785(4) ± 0.0050.798(3) ± 0.0050.774(5) ± 0.0060.900(1) ± 0.004 •
eurlex-sm0.686(8) ± 0.0050.683(10) ± 0.0070.736(6) ± 0.0030.707(7) ± 0.0040.685(9) ± 0.0060.852(2) ± 0.0050.814(5) ± 0.0040.842(3) ± 0.0060.841(4) ± 0.0060.908(1) ± 0.003 •
tmc2007-5000.761(10) ± 0.0050.765(9) ± 0.0060.847(5) ± 0.0030.837(6) ± 0.0030.770(8) ± 0.0100.863(4) ± 0.0020.799(7) ± 0.0030.888(3) ± 0.0020.891(2) ± 0.0020.920(1) ± 0.002 •
mediamill0.596(8) ± 0.0040.577(10) ± 0.0040.638(7) ± 0.0030.647(6) ± 0.0030.584(9) ± 0.0060.747(4) ± 0.0030.704(5) ± 0.0030.781(2) ± 0.0030.761(3) ± 0.0050.824(1) ± 0.002 •
average rank8.299.295.296.577.643.397.183.183.071.11
win/tie/loss29/30/19317/20/215128/8/11688/18/14649/32/171175/19/5875/8/169181/16/55184/22/46244/7/1
Table 4. The average ranks across all the datasets of nine compared MLC algorithms and MLHiKNN on each metric. The best average rank in each case is shown in bold.
Table 4. The average ranks across all the datasets of nine compared MLC algorithms and MLHiKNN on each metric. The best average rank in each case is shown in bold.
MetricBRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
AUC macro8.299.295.296.577.643.397.183.183.071.11
AUC micro6.437.866.188.188.394.146.392.292.892.25
Ranking Loss5.866.577.799.048.215.755.112.182.611.89
F1 macro3.795.005.072.894.399.296.648.684.864.39
F1 micro4.215.542.932.575.008.826.218.465.865.39
Hamming Loss5.938.322.893.897.646.115.545.466.502.71
Table 5. The Friedman test statistics F F for comparisons among ten MLC algorithms BR, CC, ECC, RAkEL, RAkELd, BRkNNa, MLKNN, DMLkNN, IBLR, and MLHiKNN. The critical value is calculated at a significance level of 0.05. # Algorithms and # Datasets denote the number of algorithms and the number of datasets, respectively.
Table 5. The Friedman test statistics F F for comparisons among ten MLC algorithms BR, CC, ECC, RAkEL, RAkELd, BRkNNa, MLKNN, DMLkNN, IBLR, and MLHiKNN. The critical value is calculated at a significance level of 0.05. # Algorithms and # Datasets denote the number of algorithms and the number of datasets, respectively.
Metric F F # Algorithms# DatasetsCritical Value
AUC macro200.74310281.919
AUC micro161.135
Ranking Loss180.210
F1 macro118.442
F1 micro114.732
Hamming Loss95.447
Table 6. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of AUC macro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table 6. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of AUC macro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
AUC MacroBRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
BR-3.23  × 10 4 1.0001.0000.9961.0000.9861.0001.0001.000
CC1.000-1.0001.0001.0001.0000.9921.0001.0001.000
ECC5.22  × 10 8 1.23  × 10 7 -5.10  × 10 7 9.42  × 10 7 1.0000.1131.0001.0001.000
RAkEL4.70  × 10 6 5.10  × 10 7 1.000-2.36  × 10 4 1.0000.6851.0001.0001.000
RAkELd0.0043.28  × 10 7 1.0001.000-1.0000.9771.0001.0001.000
BRkNNa3.73  × 10 9 3.73  × 10 9 2.84  × 10 6 7.08  × 10 8 3.73  × 10 9 -3.73  × 10 9 0.6850.9591.000
MLKNN0.0150.0090.8910.3230.0241.000-1.0001.0001.000
DMLkNN3.73  × 10 9 3.73  × 10 9 1.03  × 10 5 7.08  × 10 8 3.73  × 10 9 0.3233.73  × 10 9 -0.6691.000
IBLR3.73  × 10 9 3.73  × 10 9 2.84  × 10 6 9.31  × 10 8 3.73  × 10 9 0.0431.12  × 10 8 0.339-1.000
MLHiKNN3.73  × 10 9 3.73  × 10 9 1.86  × 10 8 7.45  × 10 9 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 1.12  × 10 8 3.73  × 10 9 -
Table 7. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of AUC micro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table 7. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of AUC micro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
AUC MicroBRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
BR-1.03  × 10 5 0.6190.0453.28  × 10 7 1.0000.9671.0001.0001.000
CC1.000-0.9800.3230.0241.0000.9971.0001.0001.000
ECC0.3900.021-1.12  × 10 8 1.09  × 10 4 1.0000.7601.0001.0001.000
RAkEL0.9570.6851.000-0.4781.0000.9991.0001.0001.000
RAkELd1.0000.9771.0000.531-1.0001.0001.0001.0001.000
BRkNNa1.14  × 10 6 9.31  × 10 8 3.98  × 10 6 9.31  × 10 8 1.86  × 10 8 -1.83  × 10 5 1.0000.9990.998
MLKNN0.0350.0040.2470.0011.19  × 10 5 1.000-1.0001.0001.000
DMLkNN3.73  × 10 9 3.73  × 10 9 3.28  × 10 7 1.86  × 10 8 3.73  × 10 9 6.47  × 10 6 3.73  × 10 9 -0.0550.746
IBLR1.14  × 10 6 2.61  × 10 8 1.83  × 10 5 9.42  × 10 7 1.86  × 10 8 7.08  × 10 4 6.47  × 10 6 0.948-0.937
MLHiKNN1.12  × 10 8 3.73  × 10 9 1.23  × 10 7 1.86  × 10 8 3.73  × 10 9 0.0021.14  × 10 6 0.2610.066-
Table 8. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of Ranking Loss. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table 8. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of Ranking Loss. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Ranking LossBRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
BR-0.0010.0022.11  × 10 5 3.73  × 10 9 0.8130.9901.0001.0001.000
CC0.999-0.0103.60  × 10 5 8.61  × 10 5 0.9570.9981.0001.0001.000
ECC0.9980.990-1.12  × 10 8 0.9401.0001.0001.0001.0001.000
RAkEL1.0001.0001.000-0.9991.0001.0001.0001.0001.000
RAkELd1.0001.0000.0636.45  × 10 4 -1.0001.0001.0001.0001.000
BRkNNa0.1930.0455.22  × 10 8 7.45  × 10 9 3.16  × 10 5 -0.8421.0001.0001.000
MLKNN0.0100.0024.70  × 10 6 5.22  × 10 8 9.31  × 10 8 0.164-1.0001.0001.000
DMLkNN1.12  × 10 8 7.45  × 10 9 7.45  × 10 9 7.45  × 10 9 7.45  × 10 9 7.45  × 10 9 3.73  × 10 9 -0.0310.558
IBLR2.05  × 10 7 1.86  × 10 8 1.60  × 10 7 2.61  × 10 8 3.73  × 10 9 4.70  × 10 6 7.57  × 10 6 0.970-0.950
MLHiKNN7.45  × 10 9 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 7.71  × 10 7 0.4510.052-
Table 9. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of F1 macro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table 9. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of F1 macro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
F1 MacroBRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
BR-0.0130.0240.9980.0639.31  × 10 8 4.66  × 10 5 9.42  × 10 7 0.0600.187
CC0.988-0.1371.0000.8994.10  × 10 7 9.34  × 10 4 8.83  × 10 6 0.1690.451
ECC0.9770.868-1.0000.9721.60  × 10 7 1.22  × 10 4 3.28  × 10 7 0.3230.339
RAkEL0.0021.22  × 10 4 6.77  × 10 5 -6.77  × 10 5 3.73  × 10 8 5.10  × 10 7 7.08  × 10 8 0.0070.055
RAkELd0.9400.1050.0301.000-2.61  × 10 7 8.61  × 10 5 2.00  × 10 6 0.0600.219
BRkNNa1.0001.0001.0001.0001.000-1.0000.9991.0001.000
MLKNN1.0000.9991.0001.0001.0002.61  × 10 8 -3.73  × 10 9 1.0000.999
DMLkNN1.0001.0001.0001.0001.0005.86  × 10 4 1.000-1.0001.000
IBLR0.9430.8360.6850.9930.9431.86  × 10 8 4.83  × 10 4 3.73  × 10 9 -0.701
MLHiKNN0.8190.5580.6690.9480.7882.61  × 10 8 0.0011.86  × 10 8 0.307-
Table 10. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of F1 micro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table 10. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of F1 micro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
F1 MicroBRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
BR-0.0030.9961.0004.66  × 10 5 8.83  × 10 6 0.0036.77  × 10 5 0.0100.063
CC0.997-1.0001.0000.7941.71  × 10 4 0.0500.0030.1270.276
ECC0.0046.47  × 10 6 -0.1530.0033.98  × 10 6 1.91  × 10 4 6.47  × 10 6 0.0060.001
RAkEL2.42  × 10 5 6.30  × 10 7 0.853-5.10  × 10 7 7.71  × 10 7 7.64  × 10 5 9.42  × 10 7 2.12  × 10 4 0.001
RAkELd1.0000.2120.9971.000-5.29  × 10 5 0.0204.83  × 10 4 0.0430.101
BRkNNa1.0001.0001.0001.0001.000-1.0000.9971.0001.000
MLKNN0.9980.9531.0001.0000.9812.61  × 10 7 -1.86  × 10 8 0.4780.965
DMLkNN1.0000.9971.0001.0001.0000.0031.000-1.0001.000
IBLR0.9910.8770.9951.0000.9591.59  × 10 5 0.5311.60  × 10 7 -0.953
MLHiKNN0.9400.7320.9990.9990.9031.67  × 10 6 0.0371.14  × 10 6 0.050-
Table 11. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of Hamming Loss. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table 11. p-values of the post hoc tests for the comparison among ten MLC algorithms in terms of Hamming Loss. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 5.56 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Hamming LossBRCCECCRAkELRAkELdBRkNNaMLKNNDMLkNNIBLRMLHiKNN
BR-2.61  × 10 8 1.0001.0001.38  × 10 6 0.7880.9340.9480.1321.000
CC1.000-1.0001.0000.9630.9971.0001.0000.8951.000
ECC2.38  × 10 6 3.73  × 10 9 -1.19  × 10 5 3.98  × 10 6 0.0130.0130.0160.0040.753
RAkEL6.30  × 10 7 3.73  × 10 9 1.000-3.73  × 10 9 0.1990.4510.4070.0060.899
RAkELd1.0000.0391.0001.000-0.9670.9990.9990.7601.000
BRkNNa0.2190.0030.9880.8070.035-0.9400.9400.3811.000
MLKNN0.0693.96  × 10 4 0.9880.5588.52  × 10 4 0.063-0.3730.1581.000
DMLkNN0.0552.92  × 10 4 0.9850.6026.45  × 10 4 0.0630.636-0.1581.000
IBLR0.8730.1090.9960.9940.2470.6270.8470.847-1.000
MLHiKNN5.29  × 10 5 7.71  × 10 7 0.2540.1058.83  × 10 6 7.08  × 10 8 2.38  × 10 6 2.84  × 10 6 2.11  × 10 5 -
Table 12. The performance of the algorithms for hubness reduction technique comparisons on AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table 12. The performance of the algorithms for hubness reduction technique comparisons on AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Macro
BRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
birds0.702(4) ± 0.0160.701(5) ± 0.0280.728(1) ± 0.009 •0.708(3) ± 0.0100.606(6) ± 0.0190.490(7) ± 0.0250.472(8) ± 0.0250.452(9) ± 0.0320.728(2) ± 0.030
CAL5000.543(4) ± 0.0060.539(5) ± 0.0070.545(3) ± 0.0040.546(2) ± 0.0060.515(6) ± 0.0040.494(7) ± 0.0050.494(8) ± 0.0040.491(9) ± 0.0040.571(1) ± 0.006 •
emotions0.834(4) ± 0.0050.830(5) ± 0.0080.845(2) ± 0.0070.842(3) ± 0.0060.723(6) ± 0.0130.707(7) ± 0.0150.705(8) ± 0.0160.701(9) ± 0.0140.846(1) ± 0.005 •
genbase0.996(4) ± 0.0040.996(2) ± 0.0040.996(3) ± 0.0040.990(5) ± 0.0060.963(6) ± 0.0130.773(7) ± 0.0420.771(8) ± 0.0520.627(9) ± 0.0800.997(1) ± 0.003 •
LLOG0.598(2) ± 0.0060.595(3) ± 0.0060.594(5) ± 0.0050.595(3) ± 0.0060.588(6) ± 0.0060.549(7) ± 0.0200.512(9) ± 0.0260.520(8) ± 0.0190.603(1) ± 0.006 •
enron0.627(4) ± 0.0120.638(3) ± 0.0120.644(2) ± 0.0110.616(5) ± 0.0090.596(6) ± 0.0110.439(7) ± 0.0150.434(8) ± 0.0120.431(9) ± 0.0100.676(1) ± 0.012 •
scene0.937(5) ± 0.0030.940(4) ± 0.0020.945(1) ± 0.003 •0.944(2) ± 0.0030.898(6) ± 0.0070.879(7) ± 0.0160.879(8) ± 0.0140.877(9) ± 0.0180.944(3) ± 0.002
yeast0.685(4) ± 0.0050.663(5) ± 0.0060.689(2) ± 0.0040.689(3) ± 0.0040.583(6) ± 0.0060.522(9) ± 0.0070.529(8) ± 0.0110.531(7) ± 0.0120.711(1) ± 0.006 •
Slashdot0.680(5) ± 0.0170.690(3) ± 0.0200.700(2) ± 0.0120.688(4) ± 0.0150.563(6) ± 0.0330.405(9) ± 0.0310.409(8) ± 0.0380.433(7) ± 0.0290.723(1) ± 0.021 •
corel5k0.582(5) ± 0.0040.626(2) ± 0.0070.625(4) ± 0.0070.626(3) ± 0.0080.569(6) ± 0.0040.415(7) ± 0.0070.414(8) ± 0.0070.412(9) ± 0.0060.690(1) ± 0.005 •
rcv1subset10.819(5) ± 0.0060.844(3) ± 0.0050.845(2) ± 0.0040.823(4) ± 0.0050.766(6) ± 0.0060.396(7) ± 0.0160.393(8) ± 0.0160.367(9) ± 0.0070.891(1) ± 0.004 •
rcv1subset20.819(5) ± 0.0080.843(2) ± 0.0070.840(3) ± 0.0070.823(4) ± 0.0060.770(6) ± 0.0070.363(8) ± 0.0140.376(7) ± 0.0120.328(9) ± 0.0160.888(1) ± 0.005 •
rcv1subset30.816(4) ± 0.0070.838(3) ± 0.0050.840(2) ± 0.0110.815(5) ± 0.0070.767(6) ± 0.0070.360(7) ± 0.0150.358(8) ± 0.0230.336(9) ± 0.0210.880(1) ± 0.004 •
rcv1subset40.816(5) ± 0.0060.837(2) ± 0.0070.837(3) ± 0.0070.823(4) ± 0.0090.769(6) ± 0.0060.380(8) ± 0.0100.387(7) ± 0.0100.344(9) ± 0.0100.887(1) ± 0.007 •
rcv1subset50.810(4) ± 0.0060.830(2) ± 0.0040.827(3) ± 0.0060.809(5) ± 0.0060.758(6) ± 0.0050.324(8) ± 0.0140.370(7) ± 0.0210.316(9) ± 0.0150.873(1) ± 0.005 •
bibtex0.794(5) ± 0.0030.813(4) ± 0.0030.826(2) ± 0.0020.814(3) ± 0.0030.754(6) ± 0.0040.352(9) ± 0.0120.400(8) ± 0.0070.432(7) ± 0.0070.886(1) ± 0.001 •
Arts0.695(4) ± 0.0090.699(2) ± 0.0100.696(3) ± 0.0080.677(5) ± 0.0070.545(6) ± 0.0070.339(9) ± 0.0080.343(8) ± 0.0080.357(7) ± 0.0070.730(1) ± 0.009 •
Health0.745(4) ± 0.0080.750(2) ± 0.0080.746(3) ± 0.0040.714(5) ± 0.0080.606(6) ± 0.0080.324(9) ± 0.0130.346(7) ± 0.0070.326(8) ± 0.0070.778(1) ± 0.009 •
Business0.701(4) ± 0.0090.705(3) ± 0.0080.718(2) ± 0.0060.679(5) ± 0.0070.589(6) ± 0.0080.328(8) ± 0.0070.317(9) ± 0.0050.342(7) ± 0.0070.752(1) ± 0.007 •
Education0.698(4) ± 0.0090.701(2) ± 0.0090.701(3) ± 0.0080.675(5) ± 0.0100.595(6) ± 0.0060.331(9) ± 0.0110.336(8) ± 0.0140.352(7) ± 0.0110.748(1) ± 0.014 •
Computers0.707(4) ± 0.0060.711(3) ± 0.0050.715(2) ± 0.0060.685(5) ± 0.0060.611(6) ± 0.0040.319(9) ± 0.0060.321(8) ± 0.0040.336(7) ± 0.0050.750(1) ± 0.005 •
Entertainment0.743(4) ± 0.0050.746(2) ± 0.0050.744(3) ± 0.0060.726(5) ± 0.0080.581(6) ± 0.0080.374(8) ± 0.0060.379(7) ± 0.0050.329(9) ± 0.0080.782(1) ± 0.004 •
Recreation0.752(4) ± 0.0060.754(2) ± 0.0060.752(3) ± 0.0060.728(5) ± 0.0050.559(6) ± 0.0120.296(9) ± 0.0060.301(8) ± 0.0070.312(7) ± 0.0060.788(1) ± 0.009 •
Society0.675(4) ± 0.0050.679(2) ± 0.0050.677(3) ± 0.0060.661(5) ± 0.0060.554(6) ± 0.0070.348(9) ± 0.0050.350(8) ± 0.0060.363(7) ± 0.0060.707(1) ± 0.006 •
eurlex-dc-l0.820(4) ± 0.0050.822(2) ± 0.0050.820(3) ± 0.0050.805(5) ± 0.0040.785(6) ± 0.0050.296(8) ± 0.0140.303(7) ± 0.0120.197(9) ± 0.0040.900(1) ± 0.004 •
eurlex-sm0.852(3) ± 0.0050.853(2) ± 0.0040.852(4) ± 0.0040.842(5) ± 0.0040.814(6) ± 0.0040.302(9) ± 0.0100.303(7) ± 0.0110.302(8) ± 0.0080.908(1) ± 0.003 •
tmc2007-5000.863(4) ± 0.0020.885(2) ± 0.0010.881(3) ± 0.0020.858(5) ± 0.0020.799(6) ± 0.0030.440(7) ± 0.0090.433(8) ± 0.0120.374(9) ± 0.0170.920(1) ± 0.002 •
mediamill0.747(4) ± 0.0030.747(5) ± 0.0030.752(2) ± 0.0030.752(3) ± 0.0040.704(6) ± 0.0030.326(7) ± 0.0050.323(8) ± 0.0050.322(9) ± 0.0070.824(1) ± 0.002 •
average rank4.142.952.644.166.007.937.828.251.11
win/tie/loss129/21/74158/21/45164/27/33127/16/8184/0/14019/16/18923/21/18016/15/193216/7/1
Table 13. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of AUC macro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
Table 13. p-values of the post hoc tests for the comparisons with hubness reduction techniques in terms of AUC macro. The i , j entry corresponds to the test whose alternative hypothesis is that the i-th method performs significantly better than the j-th method. The corrected significance level is 6.94 × 10 4 . The p-values less than the corrected significant level are filled in gray.
AUC MacroBRkNNaBRkNNa-dslBRkNNa-lsBRkNNa-mpMLKNNMLKNN-dslMLKNN-lsMLKNN-mpMLHiKNN
BRkNNa-1.0001.0000.0893.73  × 10 9 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 1.000
BRkNNa-dsl5.32  × 10 4 -0.9378.06  × 10 4 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 1.000
BRkNNa-ls4.10  × 10 7 0.066-2.05  × 10 7 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 1.000
BRkNNa-mp0.9150.9991.000-3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 1.000
MLKNN1.0001.0001.0001.000-3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 1.000
MLKNN-dsl1.0001.0001.0001.0001.000-0.8870.0721.000
MLKNN-ls1.0001.0001.0001.0001.0000.118-0.0501.000
MLKNN-mp1.0001.0001.0001.0001.0000.9310.953-1.000
MLHiKNN3.73  × 10 9 3.73  × 10 9 3.73  × 10 8 7.45  × 10 9 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 3.73  × 10 9 -
Table 14. The variant approaches for ablation analysis.
Table 14. The variant approaches for ablation analysis.
ApproachExplanation
MLHiKNN-g1 g ( x ) is fixed at 1 in Equations (9) and (10)
MLHiKNN-g0 g ( x ) is fixed at 0 in Equations (9) and (10)
MLHiKNN-h1 w h ( t , x ) is fixed at 1 in Equation (5)
MLHiKNN-d1 w d ( t , x ) is fixed at 1 in Equation (5)
MLHiKNN-fo y ^ t l = 1 if r l ( t ) 0.5 0 otherwise
Table 15. The performance of MLHiKNN and the variant approaches in terms of AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
Table 15. The performance of MLHiKNN and the variant approaches in terms of AUC macro. The best average rank is shown in bold. For each dataset, the performance of the top-ranked algorithm is marked with •.
DatasetAUC Macro
MLHiKNN-g1MLHiKNN-g0MLHiKNN-h1MLHiKNN-d1MLHiKNN-foMLHiKNN
birds0.728(3) ± 0.0310.723(4) ± 0.0330.718(6) ± 0.0320.719(5) ± 0.0280.755(1) ± 0.019 •0.728(2) ± 0.030
CAL5000.561(5) ± 0.0060.564(4) ± 0.0070.573(1) ± 0.006 •0.571(3) ± 0.0050.558(6) ± 0.0040.571(2) ± 0.006
emotions0.845(2) ± 0.0090.840(6) ± 0.0060.845(3) ± 0.0070.844(4) ± 0.0070.843(5) ± 0.0060.846(1) ± 0.005 •
genbase0.996(6) ± 0.0040.997(5) ± 0.0030.997(3) ± 0.0030.998(2) ± 0.0030.998(1) ± 0.002 •0.997(4) ± 0.003
LLOG0.603(5) ± 0.0060.603(1) ± 0.006 •0.603(2) ± 0.0060.603(3) ± 0.0060.603(6) ± 0.0060.603(4) ± 0.006
enron0.663(6) ± 0.0110.671(4) ± 0.0070.670(5) ± 0.0100.673(2) ± 0.0120.672(3) ± 0.0110.676(1) ± 0.012 •
scene0.945(1) ± 0.003 •0.942(6) ± 0.0020.943(4) ± 0.0020.943(3) ± 0.0020.943(5) ± 0.0010.944(2) ± 0.002
yeast0.699(5) ± 0.0040.690(6) ± 0.0100.710(2) ± 0.0060.707(3) ± 0.0050.706(4) ± 0.0060.711(1) ± 0.006 •
Slashdot0.686(6) ± 0.0160.719(5) ± 0.0210.721(4) ± 0.0200.730(2) ± 0.0140.743(1) ± 0.014 •0.723(3) ± 0.021
corel5k0.666(4) ± 0.0040.651(5) ± 0.0090.689(2) ± 0.0050.689(3) ± 0.0050.635(6) ± 0.0070.690(1) ± 0.005 •
rcv1subset10.882(4) ± 0.0030.877(5) ± 0.0050.886(3) ± 0.0040.887(2) ± 0.0040.863(6) ± 0.0040.891(1) ± 0.004 •
rcv1subset20.873(4) ± 0.0070.873(6) ± 0.0060.881(3) ± 0.0040.884(2) ± 0.0050.873(5) ± 0.0090.888(1) ± 0.005 •
rcv1subset30.863(6) ± 0.0040.868(4) ± 0.0030.874(3) ± 0.0030.876(2) ± 0.0020.868(5) ± 0.0040.880(1) ± 0.004 •
rcv1subset40.873(5) ± 0.0060.877(4) ± 0.0060.881(3) ± 0.0070.882(2) ± 0.0070.863(6) ± 0.0050.887(1) ± 0.007 •
rcv1subset50.857(6) ± 0.0050.864(5) ± 0.0050.869(4) ± 0.0060.871(2) ± 0.0050.869(3) ± 0.0050.873(1) ± 0.005 •
bibtex0.878(4) ± 0.0020.862(5) ± 0.0040.882(3) ± 0.0020.885(2) ± 0.0020.841(6) ± 0.0050.886(1) ± 0.001 •
Arts0.701(6) ± 0.0090.725(4) ± 0.0080.730(2) ± 0.0090.731(1) ± 0.009 •0.718(5) ± 0.0080.730(3) ± 0.009
Health0.749(6) ± 0.0070.771(5) ± 0.0100.775(3) ± 0.0090.779(1) ± 0.011 •0.771(4) ± 0.0070.778(2) ± 0.009
Business0.709(6) ± 0.0070.732(5) ± 0.0180.751(3) ± 0.0070.751(4) ± 0.0080.766(1) ± 0.007 •0.752(2) ± 0.007
Education0.711(6) ± 0.0180.736(4) ± 0.0110.748(2) ± 0.0130.746(3) ± 0.0130.734(5) ± 0.0100.748(1) ± 0.014 •
Computers0.716(6) ± 0.0080.743(5) ± 0.0070.749(4) ± 0.0060.750(3) ± 0.0070.750(2) ± 0.0050.750(1) ± 0.005 •
Entertainment0.757(5) ± 0.0050.775(4) ± 0.0090.780(2) ± 0.0040.780(3) ± 0.0060.756(6) ± 0.0080.782(1) ± 0.004 •
Recreation0.756(6) ± 0.0090.786(4) ± 0.0080.787(3) ± 0.0080.790(1) ± 0.008 •0.777(5) ± 0.0070.788(2) ± 0.009
Society0.685(6) ± 0.0070.704(4) ± 0.0040.707(2) ± 0.0060.705(3) ± 0.0040.693(5) ± 0.0050.707(1) ± 0.006 •
eurlex-dc-l0.883(4) ± 0.0070.874(5) ± 0.0030.900(2) ± 0.0040.892(3) ± 0.0050.856(6) ± 0.0050.900(1) ± 0.004 •
eurlex-sm0.899(4) ± 0.0040.896(5) ± 0.0040.907(2) ± 0.0030.900(3) ± 0.0020.881(6) ± 0.0060.908(1) ± 0.003 •
tmc2007-5000.914(4) ± 0.0020.907(6) ± 0.0020.918(3) ± 0.0020.912(5) ± 0.0020.919(2) ± 0.0020.920(1) ± 0.002 •
mediamill0.820(3) ± 0.0030.791(6) ± 0.0030.824(1) ± 0.002 •0.801(4) ± 0.0050.796(5) ± 0.0030.824(2) ± 0.002
average rank4.794.712.862.714.321.61
win/tie/loss22/25/9324/32/8466/46/2866/46/2836/32/7297/37/6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Teng, Z.; Tang, S.; Huang, M.; Wang, X. A Hubness Information-Based k-Nearest Neighbor Approach for Multi-Label Learning. Mathematics 2025, 13, 1202. https://doi.org/10.3390/math13071202

AMA Style

Teng Z, Tang S, Huang M, Wang X. A Hubness Information-Based k-Nearest Neighbor Approach for Multi-Label Learning. Mathematics. 2025; 13(7):1202. https://doi.org/10.3390/math13071202

Chicago/Turabian Style

Teng, Zeyu, Shanshan Tang, Min Huang, and Xingwei Wang. 2025. "A Hubness Information-Based k-Nearest Neighbor Approach for Multi-Label Learning" Mathematics 13, no. 7: 1202. https://doi.org/10.3390/math13071202

APA Style

Teng, Z., Tang, S., Huang, M., & Wang, X. (2025). A Hubness Information-Based k-Nearest Neighbor Approach for Multi-Label Learning. Mathematics, 13(7), 1202. https://doi.org/10.3390/math13071202

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop