Next Article in Journal
Stochastic Operation of BESS and MVDC Link in Distribution Networks Under Uncertainty
Previous Article in Journal
LLM-as-a-Judge Approaches as Proxies for Mathematical Coherence in Narrative Extraction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Generative Learning from Semantically Confused Label Distribution via Auto-Encoding Variational Bayes

1
Zhongshan Power Supply Bureau, China Southern Power Grid Co., Ltd., Zhongshan 528400, China
2
Department of Automation, Shanghai Jiao Tong University, Shanghai 200240, China
3
Department of Computing, The Hong Kong Polytechnic University, Hong Kong 999077, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(13), 2736; https://doi.org/10.3390/electronics14132736
Submission received: 27 May 2025 / Revised: 26 June 2025 / Accepted: 4 July 2025 / Published: 7 July 2025
(This article belongs to the Special Issue Neural Networks: From Software to Hardware)

Abstract

Label Distribution Learning (LDL) has emerged as a powerful paradigm for addressing label ambiguity, offering a more nuanced quantification of the instance–label relationship compared to traditional single-label and multi-label learning approaches. This paper focuses on the challenge of noisy label distributions, which is ubiquitous in real-world applications due to the annotator subjectivity, algorithmic biases, and experimental errors. Existing related LDL algorithms often assume a linear combination of true and random label distributions when modeling the noisy label distributions, an oversimplification that fails to capture the practical generation processes of noisy label distributions. Therefore, this paper introduces an assumption that the noise in label distributions primarily arises from the semantic confusion between labels and proposes a novel generative label distribution learning algorithm to model the confusion-based generation process of both the feature data and the noisy label distribution data. The proposed model is inferred using variational methods and its effectiveness is demonstrated through extensive experiments across various real-world datasets, showcasing its superiority in handling noisy label distributions.

Graphical Abstract

1. Introduction

Label Distribution Learning (LDL) [1] is emerging as a potent paradigm for mitigating label ambiguity. In the context of LDL, the training instance is annotated with a label distribution vector, where each component signifies the degree to which each label describes the instance. Distinct from conventional single-label learning and multi-label learning, in which only the presence or absence of label relevance is indicated, LDL offers a more detailed quantification of the instance–label relationship. This nuanced approach enhances the applicability of LDL across various domains, including age estimation [2,3,4,5,6], affective analysis [7,8,9,10,11], and rating prediction [12,13,14,15].
Current works in LDL predominantly emphasize enhancing generalization performance. For instance, some research efforts have been directed towards developing more robust LDL models that effectively represent the mapping from features to label distributions [4,13,16]. Additionally, there studies have focused on devising regularized loss functions that encapsulate prior knowledge about the mapping from features to label distributions [17,18,19]. While these algorithms have demonstrated commendable performance across various datasets, they predominantly rely on high-quality training label distributions. However, the quality of training label distributions in many real-world applications is not assured. Human-annotated label distributions are particularly vulnerable to factors such as annotator subjectivity and expertise levels. Moreover, label distributions produced by non-human systems such as label smoothing algorithms and experimental apparatuses are prone to the limitations of algorithmic assumptions, data distribution biases, and intrinsic errors on the part of the experimental apparatus. These combined factors pose significant challenges in guaranteeing the quality of training label distributions, potentially leading to the unreliability of existing LDL algorithms. Although numerous efforts have been made to tackle this problem, they typically rely on the straightforward assumption that a noisy label distribution is a linear combination of the true label distribution and a random label distribution [20]. However, this simplified assumption falls short of accurately capturing the complex and nuanced processes of generating noisy label distributions in practical scenarios.
Therefore, in this paper we investigate the mechanisms underlying the generation of noisy label distributions and propose a novel generative label distribution learning algorithm. Specifically, we first propose that the noise in label distributions primarily arises from the semantic confusion between labels. Taking the image recognition task as an example, it is evident that cats and dogs can be accurately distinguished, whereas horses and donkeys are prone to being confused. Unlike the linear noise assumption, label noise based on semantic confusion more precisely captures the underlying noise generation mechanism between labels and is better able to model the complex uncertainties present in human-annotated label distributions. To address this noise generation mechanism, we construct a probabilistic graphical model to fundamentally model this process. The generation and inference processes of the proposed model are illustrated in Figure 1. During the generation process, the latent true label distribution is obtained by normalizing a normal vector through the softmax function. Meanwhile, we assume that the probabilities of semantic confusion between labels adhere to a Dirichlet distribution, allowing the complex dependencies between labels to be captured. Subsequently, in order to utilize the mutual confusion among labels, we assume that the value of each label in the noisy label distribution is affected by the weighted combination of the values of other labels, where the weights are the probabilities of other labels being misclassified as that label. Furthermore, the feature vector for each instance is assumed to be generated from a normal distribution, the parameters of which are determined by the nonlinear transformation of the true label distribution, thereby establishing the association between features and label distributions. In the inference process, we derive the posterior distribution of the true label distribution by variational methods and derive a differentiable variational lower bound through the aforementioned generation process. Finally, we conduct extensive experiments on a wide range of real-world datasets to validate the effectiveness and superiority of the proposed algorithm.

2. Related Work

To predict the label distributions for unseen instances, the primary approach involves learning a mapping from feature vectors to label distributions based on the feature vectors of training instances and their corresponding label distributions (a process termed LDL, i.e., label distribution learning) and subsequently utilizing this mapping to output the label distribution for test instances. The crucial challenge of LDL lies in the design of LDL algorithms with robust generalization performance. Existing research on this issue predominantly focuses on two aspects, namely, model representation and loss functions.
Studies on model representation have primarily concentrated on the manner in which the mapping from feature vectors to label distributions is represented. To date, the findings about this issue can be categorized into three distinct classes. The first class comprises maximum entropy LDL models [1]. Maximum entropy LDL models are derived from the principle of maximum entropy [21]. Formally, these models typically perform exponential normalization on the linear or nonlinear mapping of feature vectors to produce the label distribution. The second class is prototype-based models, wherein the mapping from features to label distributions is constructed through existing examples in the training set. For example, the AAkNN algorithm [1] extends the traditional k-nearest neighbors algorithm to accommodate the task of LDL. The GMML-kLDL algorithm [22] introduces a distance metric that captures label correlation and classification information, which is then used to identify the nearest neighbors of a test instance and leverage their label distribution information to predict the label distribution of the test instance. Furthermore, the LDL-LCR algorithm [23] seeks several nearest neighbors of a test instance within the training set and reconstructs the feature vector of the test instance based on these neighbors, ultimately combining the reconstruction coefficients with the neighbors’ label distributions to derive the label distribution of the test instance. The LDSVR algorithm [13] utilizes kernel techniques to assess the correlation between test and training instances, then combines these correlations as weights to generate the label distribution for the test instance. The third class is ensemble models, which aim to directly extend existing ensemble learning models to meet the requirements of LDL. Specifically, the LDLogitBoost algorithm [16] extends the LogitBoost algorithm [24] to LDL by introducing weighted regressors, resulting in a boosting algorithm for LDL. The LDLFs algorithm [3,4] applies the differentiable decision tree model to LDL with the goal of enhancing the model’s ability to fit the distribution form. In addition, the ENN-LDL algorithm [25] integrates ensemble neural networks into LDL through an ensemble architecture designed to improve the model’s capacity to represent the highly correlated label distributions. Similarly, the StructRF algorithm [26] introduces the structured random forest model into label distribution learning tasks, which aims to enhance the representation of highly correlated label distributions.
Studies on loss functions can be roughly divided into three categories. The first class focuses on mining and utilizing the label correlation or sample correlation within label distributions. For example, the LDLLC algorithm [27] constructs a distance metric to calculate the label correlation based on the training label distributions, then uses the distance between any two labels to regulate the distance between the corresponding column vectors in the parameter matrix. This ensures that the learning process of the label distribution model maintains the label correlation based on the training label distributions. The LALOT algorithm [28] treats the process of mining label correlation as a metric learning problem, employing the optimal transport distance to capture the geometric information of the underlying label space. The LDL-SCL algorithm [29] introduces the local label correlation assumption, positing that label correlation within label distributions varies across different sample clusters; accordingly, it constructs a local correlation vector as an additional feature for each sample to capture this characteristic. The EDL-LRL algorithm [9], also based on the local label correlation, employs a low-rank structure on local samples to mine local label correlation. The LDL-LCLR algorithm [30] simultaneously learns global and local label correlations, then mines them through low-rank approximation and sample clustering, respectively. The LDLSF algorithm [31] proposes a new label correlation assumption in the field of LDL, i.e., that the correlated labels should have similar corresponding components in the output label distribution. The LDL-LDM algorithm [17] utilizes global and local label correlation in a data-driven manner which first learns the manifold structure of label distributions and promotes the model output to also distribute within this manifold. The LDL-LRR [18] and LDL-DPA [19] algorithms propose regularizing the learning process by using the label rankings within the label distributions. The second class explores how to improve the generalization ability of models through label embedding. For instance, the MSLP algorithm [32] achieves label embedding for label distribution learning via multi-scale location preservation. The BC-LDL algorithm [33] integrates the label distribution information into the binary coding process to produce high-quality binary codes. The DBC-LDL algorithm [34] considers an efficient discrete coding framework to learn the binary coding of instances. The third class designs loss functions for different variants of LDL task. For example, the IncomLDL algorithm [35] proposes that the label distribution matrix is of low rank and proposes an optimization objective based on trace norm minimization to address incomplete LDL scenarios where the training label distributions have missing values. The IncomLDL-LR algorithm [36] assumes that the label distribution of each example can be linearly reconstructed based on its neighbors; as such, it recovers the missing label distribution information through the label distributions of nearest neighbor examples based on feature vectors. The GRME [37] and IncomLDL-LCD [38] algorithms recover the missing label distribution values by mining the correlation between labels. GLDL [39] constructs a graph-structured LDL framework which explicitly models the instance relation and label relation. In addition, some LDL studies have focused on weakly supervised LDL; for example, certain LDL algorithms are specifically designed to address the noise in LDL [20,40,41]. It should be noted that although the GCIA algorithm [41] utilized confusion matrix to describe the noise generation process, which is similar to our work, these two approaches have two significant differences. First, GCIA implicitly captures the dependency between the ground-truth label distribution and the sample features using a latent clustering variable. In contrast, our method directly conditions the sample feature variable on the ground-truth label distribution. By modeling this dependency explicitly, our method aims to better exploit the relationship between features and labels. Second, GCIA treats the confusion matrix as a learnable parameter; however, this approach does not account for the inherent uncertainty of the confusion matrix in real-world scenarios. In contrast, our method models the confusion matrix as a set of Dirichlet random vectors to more effectively capture the uncertainty and variability of the semantic confusion. Finally, several LDL algorithms have proposed learning the label distributions by using more accessible label forms such as logical labels [42,43,44], ternary labels [45], or label rankings [46,47,48].

3. Methodology

The commonly used mathematical notation is shown in Table 1. We address training datasets that appear as data pairs { ( x n , y n ) } n = 1 N . Our goal is to learn a label distribution predictor f : X D Δ M (i.e., a mapping from the feature space to the label distribution space) using the training dataset { ( x n , y n ) } n = 1 N . The generation process of the observations can be formalized as follows.
  • Generate a sample of the latent logits u ( 0 , 1 ) M from a standard normal prior:
    u Norm u 0 , diag ( 1 ) .
  • Generate a sample of the confusion vector of each label ω m Δ M from a Dirichlet prior:
    { ω m } m = 1 M m = 1 M Diri ( ω m [ α I ( t = m ) + 1 ] t = 1 M )
    where the t-th element of ω m (i.e., ω m t ) denotes the probability of the t-th label being mis-annotated as the m-th label.
  • Generate a sample of the noisy label distribution from a Dirichlet distribution conditioned on the latent logits and the confusion vector:
    z u , { ω m } m = 1 M Diri z ω m , exp ( u ) m = 1 M .
  • Generate observations of the feature variables:
    x u Norm x H ( u ; Θ mean ( p ) ) , diag ( λ 1 1 )
    where λ is the precision of the normal distribution for generating the feature vector.
The joint probability distribution of the complete data can be decomposed as follows:
p z , x , u , { ω m } m = 1 M = p x | u p z | u , { ω m } m = 1 M p ( u ) p { ω m } m = 1 M .

4. Variational Inference

We aim to infer the posterior distribution of the latent variables for practical decision-making. However, due to the complicated dependencies among the variables, obtaining an exact solution for the posterior distribution of the latent variables is challenging. Consequently, we employ a parameterized variational distribution q ( u , { ω m } m = 1 M | x , z ) to approximate the true posterior distribution. Formally, we aim to minimize the Kullback–Leibler (KL) divergence between the variational posterior and the true posterior [49]:
arg min q ( u , { ω m } m = 1 M | x , z ) Q D KL q ( u , { ω m } m = 1 M | x , z ) p ( u , { ω m } m = 1 M | x , z )
where D KL denotes the KL divergence and Q denotes the variational posterior family. Because p ( u , { ω m } m = 1 M | x , z ) is intractable, we rewrite Equation (6) as Equation (7):
D KL q ( u , { ω m } m = 1 M | x , z ) p ( u , { ω m } m = 1 M | x , z ) = log p ( x , z ) q ( u , { ω m } m = 1 M | x , z ) log p z , x , u , { ω m } m = 1 M q ( u , { ω m } m = 1 M | x , z ) d u d ω 1 d ω M = log p ( x , z ) ( E q ( u , { ω m } m = 1 M | x , z ) log p x , z | u , { ω m } m = 1 M D KL q u , { ω m } m = 1 M | x , z p u , { ω m } m = 1 M ) .
Because log p ( x , z ) is a constant that is independent of the inference process, the minimization process in Equation (6) is equivalent to the maximization process in Equation (8):
arg max q ( u , { ω m } m = 1 M | x , z ) Q E q ( u , { ω m } m = 1 M | x , z ) log p x , z | u , { ω m } m = 1 M D KL q u , { ω m } m = 1 M | x , z p u , { ω m } m = 1 M .
We denote the first and second terms in Equation (8) as J rec and J pri , respectively. Adhering to the mean-field principle, we assume the variational posterior as Equation (9):
q ( u , { ω m } m = 1 M | x , z ) = q ( u | x ) q ( { ω m } m = 1 M | x , z ) .
It should be noted that due to the absence of noisy label information during the prediction phase we assume that u is variationally conditioned on x to facilitate the prediction process. Because u adheres to a normal prior, we also employ a normal distribution to model the variational posterior of u :
q ( u | x ) = Norm ( u H ( x ; Θ mean ( q ) ) , SFP H ( x ; Θ std ( q ) ) )
where SFP ( · ) is the softplus function, i.e., SFP ( · ) = log ( 1 + exp ( · ) ) . To prevent the gradient information of the optimization objective from being removed by randomness, we can sample from q ( u | x ) by Equation (11):
u = H ( x ; Θ mean ( q ) ) + SFP H ( x ; Θ std ( q ) ) · ϵ : = R Norm ( ϵ ) , ϵ Norm ( ϵ 0 , diag ( 1 ) ) .
Similarly, we employ a Dirichlet distribution to model the variational posterior of { ω m } m = 1 M , as ω adheres to a Dirichlet prior:
q ( { ω m } m = 1 M | x , z ) = m = 1 M Diri ( ω m SFP H ( [ x , z ] ; Θ ω m ( q ) ) ) .
As suggested by [50], the gradient-preserved sampling for ω can be approximately formalized as follows:
ω m = ( δ t ξ m t Γ ( ξ m t ) ) 1 ξ m t i = 1 M ( δ t ξ m i Γ ( ξ m i ) ) 1 ξ m i t = 1 M : = R Diri ( δ ) , δ Unif ( δ 0 , 1 )
where ξ m = SFP H ( [ x , z ] ; Θ ω m ( q ) ) . According to the decomposition in Equation (9), the first term in Equation (8), i.e., J rec , can be estimated by Monte Carlo sampling:
J rec = E q ( u , { ω m } m = 1 M | x , z ) log p x , z | u , { ω m } m = 1 M 1 K k = 1 K log p x , z | R Norm ( ϵ ( k ) ) , { R Diri ( δ m ( k ) ) } m = 1 M
where ϵ ( k ) Norm ( ϵ 0 , diag ( 1 ) ) and δ m ( k ) Unif ( δ 0 , 1 ) . Intuitively, Equation (14) measures the quality of reconstructing the observations of features and labels from the variational posterior of latent variables, while maximizing Equation (14) encourages the latent variables to better explain the observations of features and noisy label distributions. The second term in Equation (8), i.e., J pri , can be transformed as follows:
J pri = D KL q u , { ω m } m = 1 M | x , z p u , { ω m } m = 1 M = D KL q u | x p ( u ) + m = 1 M D KL q ω m | x , z p ( ω m ) = D KL ( Norm ( u H ( x ; Θ mean ( q ) ) , SFP H ( x ; Θ std ( q ) ) ) Norm u m 0 , diag ( 1 ) ) + m = 1 M D KL ( Diri ( ω m SFP H ( [ x , z ] ; Θ ω m ( q ) ) ) Diri ( ω m [ α I ( t = m ) + 1 ] t = 1 M ) ) .
Intuitively, minimizing Equation (15) encourages the posteriors of the label distribution and confusion vectors to approach their priors. Computationally, Equation (15) entails the KL divergence between normal distributions as well as the KL divergence between Dirichlet distributions, both of which can be derived analytically. Therefore, by combining Equations (14) and (15), the final optimization objective can be summarized as follows:
maximize 1 N 1 K k = 1 K n = 1 N log Norm x ( n ) H ( R Norm ( ϵ ( k ) ) ; Θ mean ( p ) ) , diag ( λ 1 1 ) + log Diri z ( n ) R Diri ( δ m ( k ) ) , exp ( R Norm ( ϵ ( k ) ) ) m = 1 M D KL ( Norm ( u H ( x ( n ) ; Θ mean ( q ) ) , SFP H ( x ( n ) ; Θ std ( q ) ) ) Norm u 0 , diag ( 1 ) ) m = 1 M D KL ( Diri ( ω m SFP H ( [ x ( n ) , z ( n ) ] ; Θ ω m ( q ) ) ) Diri ( ω m [ α I ( t = m ) + 1 ] t = 1 M ) ) .

5. Experiments

5.1. Datasets and Evaluation Metrics

The datasets used in this paper are shown in Table 2. The SJAFFE [51] dataset comprises 213 facial images, each annotated by 60 experts who rated six basic emotions (happiness, sadness, surprise, fear, anger, and disgust) on a five-point intensity scale. The emotional intensity for each image is represented by the average score across all raters, which is subsequently normalized to construct the emotion label distribution. Similarly, the SBU-3DFE [52] dataset contains 2500 facial expression images, with each image evaluated by 23 experts following the same rating protocol as SJAFFE to generate corresponding label distributions. The Yeast datasets [1] (i.e. the datasets with IDs ranging from 3 to 8) encompass empirical data gathered from ten distinct biological experiments conducted on budding yeast. Each dataset encompasses 2465 yeast genes, with each gene represented by a 24-dimensional vector. The labels within each dataset denote the discrete time points during the biological experiment. The label distribution represents the expression levels of yeast genes at each time point.
We use six common LDL evaluation metrics: Cheb (Chebyshev distance), Clark (Clark distance), Canber (Canberra metric), KL (Kullback–Leibler divergence), Cosine (cosine coefficient), and Intersec (intersection similarity).
Cheb { z n } n = 1 N , { z ^ n } n = 1 N = 1 N n = 1 N max { | z n m z ^ n m | } m = 1 M Clark { z n } n = 1 N , { z ^ n } n = 1 N = 1 N n = 1 N m = 1 M ( z n m z ^ n m ) 2 ( z n m + z ^ n m ) 2 Canber { z n } n = 1 N , { z ^ n } n = 1 N = 1 N n = 1 N m = 1 M | z n m z ^ n m | z n m + z ^ n m KL { z n } n = 1 N , { z ^ n } n = 1 N = 1 N n = 1 N m = 1 M z n m log z n m z ^ n m Cosine { z n } n = 1 N , { z ^ n } n = 1 N = 1 N n = 1 N z n , z ^ n z n 2 · z ^ n 2 Intersec { z n } n = 1 N , { z ^ n } n = 1 N = 1 N n = 1 N m = 1 M min { z ^ n m , z n m }
Lower values of distance-based measures (i.e., Cheb, Clark, Canber, KL) and higher value of similarity-based measures (i.e., Cosine and Intersec) represent better performance, which are denoted by ↓ and ↑, respectively.

5.2. Experimental Procedure

In this section, we illustrate the methodology employed to assess the efficacy of our approach. Overall, we introduce confusion-based noise into the label distribution of the noise-free training set, then utilize the noise-free test set to evaluate the performance of comparison algorithms trained on the training set with noisy label distributions. Specifically, given a noise-free LDL dataset, we randomly split the dataset into two chunks (70% for training and 30% for testing). We then obtain the noisy label distributions z ~ by adding noise to the label distributions of the training instances according to Equation (18):
z ~ = [ π m , z ] m = 1 M , { π m } m = 1 M m = 1 M Diri ( π [ α * I ( t = m ) + 1 ] t = 1 M )
where the α ~ parameter controls the extent of noiselessness of the dataset. We then train LDL models on the noisy training set and evaluate the prediction performance on the noise-free test set. We repeat the above process ten times and record the mean and standard deviation.

5.3. Comparison Algorithms

We compare our proposed algorithms with seven existing LDL algorithms: AAkNN [1], SABFGS [1], BD-LDL [53], Duo-LDL [54], LDL-LRR [18], LDL-LDM [17], and LDL-DPA [19]. The hyperparameter settings for each algorithm adhere to the recommendations provided in their respective papers. The hyperparameter k in AAkNN is set to 5. The hyperparameters λ 1 and λ 2 in BD-LDL are selected among { 10 3 , 10 2 , , 10 3 } . The hyperparameters λ and β in LDL-LRR are selected among { 10 6 , 10 5 , , 10 1 } and { 10 3 , 10 2 , , 10 2 } , respectively. The hyperparameters λ 1 , λ 2 , and λ 3 in LDL-LDM are selected among { 10 3 , 10 2 , , 10 3 } , while the hyperparameter g in LDL-LDM is selected among { 1 , 2 , , 14 } . The hyperparameters λ 1 , λ 2 , and λ 3 in LDL-DPA are selected among { 10 9 , 10 8 , , 10 0 } , { 10 7 , 10 6 , , 10 2 } , and { 10 4 , 10 3 , , 10 1 } , respectively. Finally, the hyperparameters α and λ of our proposed algorithm are set as 6 and 10 1 , respectively.

5.4. Results and Discussions

The main experimental results are presented in Table 3, Table 4, Table 5 and Table 6, where Table 3 and Table 4 display the algorithm performance conducted under α * = 4 (low noise), and Table 5 and Table 6 display the algorithm performance conducted under α * = 8 (high noise). Each experimental result within these tables is formalized as “ mean ± std r / ”, where “mean” represents the mean value of the algorithm’s performance, “std” denotes the standard deviation of the algorithm performance, “r” denotes the rank of the corresponding algorithm among all comparison algorithms, and “†” denotes the statistical significance under the pairwise two-tailed t-test at 0.05 significance level; in addition, “ = ” denotes that the corresponding algorithm is significantly inferior to our algorithm, “ = ” denotes that the corresponding algorithm is significantly superior to our algorithm, and “ = ” denotes that there is no significant difference between the corresponding algorithm and our proposed algorithm. The best performance is highlighted by boldface. In these experiments, we systematically evaluate the performance of different weakly supervised learning methods when the data quality ( α * ) varies. When the data are noisy ( α * = 4 ), our method demonstrates a significant advantage on the SJAFFE and SBU-3DFE datasets; in particular, it is about 10–15% ahead of the second-best algorithm in terms of the distance metrics (Cheb, Clark, and Canber), which suggests that our method is more robust to the noisy data. While the performance of all methods improves as the data quality is improved to α * = 8 , our method still maintains the leading position, especially in terms of the KL and Cosine metrics on the Yeast datasets, for which the standard deviation is significantly smaller than that of the other methods (e.g., the standard deviation of the KL metrics for Yeast-cold is only 0.0008 ). These results indicate that the proposed method has stable prediction ability under different data distributions. Notably, LDL-LDM slightly outperforms our method for the Cheb metric ( 0.1425 vs. 0.1428 ) on the SBU-3DFE dataset with high-quality data ( α * = 8 ), which may stem from this algorithm’s advantage in learning clean labels. On the other hand, Duo-LDL and BD-LDL perform relatively poorly in all types of settings. This is particularly the case in noisy data environments, where the performance fluctuates drastically (e.g., the standard deviation of the Clark metric for Duo-LDL on SJAFFE at α * = 4 reaches 0.3578 ), which reflects the sensitivity of these methods to data quality. Overall, the experimental results validate the superiority of our method in both robustness and accuracy.

5.5. Further Analysis

In this section, we provide an empirical demonstration and discussion on the characteristics of our algorithm, which is beneficial for its practical application.

5.5.1. Hyperparameter Analysis

Here, we investigate the impact of the hyperparameters λ and α on the performance of our algorithm. Specifically, we define the value range of λ and α as ( λ , α ) A , where A = { 10 4 , 10 3 , , 10 } × { 2 , 4 , 6 , , 12 } . As described in Section 5.2, we conduct the experimental procedure with α * = 8 for each pair of hyperparameter values within A , obtaining the performance of the algorithm under each pair of hyperparameter values within A . The experimental results are visualized in Figure 2. In the heatmap, the x-axis and y-axis represent the values of λ and α , respectively, while the color indicates the performance quantified by the cosine coefficient or Kullback–Leibler divergence. The box plot shows the marginal impact of the hyperparameters λ or α on the algorithm’s performance. It can be observed that the algorithm generally performs better when λ is small and that the impact of α on performance is relatively minor. Therefore, in the actual implementation of the algorithm, λ can be typically set to a small value.

5.5.2. Convergence

Here, we demonstrate the convergence of the inference process of our proposed algorithm. As illustrated in Figure 3, the proposed algorithm demonstrates steady convergence behavior across multiple datasets. The curves are differentiated by color for each dataset and transparency for varying α * values. The results indicate that our algorithm rapidly approaches the optimal objective value within the first 100 iterations. This initial phase exhibits a steep descent, suggesting efficient gradient updates or search direction alignment. Subsequently, the optimization stabilizes, with marginal improvements in later iterations, indicating convergence to a near-optimal solution. The consistency of this trend across most datasets highlights the robustness of our algorithm to different data distributions.

6. Conclusions

This paper identifies a critical limitation of noise modeling in existing label distribution learning methods, i.e., the oversimplified assumption that noisy label distributions can be directly modeled as a mixture of the true label distribution and a random noise. To address this gap, we rigorously investigate the underlying generation mechanisms of noisy label distribution and propose an assumption that noisy label distributions primarily stem from semantic confusion among labels (i.e., inter-label semantic ambiguity). Grounded in this assumption, we develop a generative label distribution learning framework that explicitly accounts for label-wise semantic correlations and confusion patterns. To comprehensively validate our proposal, we conduct extensive experiments under varying noise levels across eight real-world benchmark datasets. The results demonstrate that our algorithm effectively models noisy label distributions originating from label semantic confusion, consistently achieving state-of-the-art performance in nearly all scenarios.

Author Contributions

Conceptualization, X.L., C.M., H.Z. and Y.L.; methodology, B.X., T.Y. and Y.L.; software, B.X. and T.Y.; validation, B.X. and T.Y.; formal analysis, B.X. and T.Y.; investigation, B.X. and T.Y.; resources, X.L., C.M. and H.Z.; data curation, B.X. and T.Y.; writing—original draft preparation, B.X. and T.Y.; writing—review and editing, X.L., C.M., H.Z. and Y.G.; visualization, B.X. and T.Y.; supervision, X.L.; project administration, X.L.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Project of China Southern Power Grid Company Limited, grant number 032000KC23120050/GDKJXM20231537.

Data Availability Statement

Restrictions apply to the availability of these data. Data were obtained from PALM lab and are available at https://palm.seu.edu.cn/xgeng/LDL/download.htm (accessed on 1 July 2025) with the permission of PALM lab.

Conflicts of Interest

Authors Xinhai Li, Chenxu Meng, and Heng Zhou were employed by the Zhongshan Power Supply Bureau, China Southern Power Grid Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The authors declare that this study received funding from the Science and Technology Project of China Southern Power Grid Company Limited. The funder was not involved in the study design, in the collection, analysis, and interpretation of data, in the writing of this article, or in the decision to submit it for publication.

References

  1. Geng, X. Label Distribution Learning. IEEE Trans. Knowl. Data Eng. 2016, 28, 1734–1748. [Google Scholar] [CrossRef]
  2. Geng, X.; Yin, C.; Zhou, Z.H. Facial Age Estimation by Learning from Label Distributions. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2401–2412. [Google Scholar] [CrossRef] [PubMed]
  3. Shen, W.; Guo, Y.; Wang, Y.; Zhao, K.; Wang, B.; Yuille, A. Deep Differentiable Random Forests for Age Estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 404–419. [Google Scholar] [CrossRef]
  4. Shen, W.; Zhao, K.; Guo, Y.; Yuille, A. Label Distribution Learning Forests. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  5. Hou, P.; Geng, X.; Huo, Z.W.; Lv, J. Semi-Supervised Adaptive Label Distribution Learning for Facial Age Estimation. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 2015–2021. [Google Scholar]
  6. Gao, B.B.; Zhou, H.Y.; Wu, J.; Geng, X. Age Estimation Using Expectation of Label Distribution Learning. In Proceedings of the International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 712–718. [Google Scholar]
  7. He, T.; Jin, X. Image Emotion Distribution Learning with Graph Convolutional Networks. In Proceedings of the International Conference on Multimedia Retrieval, Ottawa, ON, Canada, 10–13 June 2019; pp. 382–390. [Google Scholar]
  8. Yang, J.; Sun, M.; Sun, X. Learning Visual Sentiment Distributions via Augmented Conditional Probability Neural Network. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 224–230. [Google Scholar]
  9. Jia, X.; Zheng, X.; Li, W.; Zhang, C.; Li, Z. Facial Emotion Distribution Learning by Exploiting Low-Rank Label Correlations Locally. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 9841–9850. [Google Scholar]
  10. Peng, K.C.; Chen, T.; Sadovnik, A.; Gallagher, A. A Mixed Bag of Emotions: Model, Predict, and Transfer Emotion Distributions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 860–868. [Google Scholar]
  11. Machajdik, J.; Hanbury, A. Affective Image Classification Using Features Inspired by Psychology and Art Theory. In Proceedings of the ACM International Conference on Multimedia, Florence, Italy, 25–29 October 2010; pp. 83–92. [Google Scholar]
  12. Ren, Y.; Geng, X. Sense Beauty by Label Distribution Learning. In Proceedings of the International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 2648–2654. [Google Scholar]
  13. Geng, X.; Hou, P. Pre-Release Prediction of Crowd Opinion on Movies by Label Distribution Learning. In Proceedings of the International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015; pp. 3511–3517. [Google Scholar]
  14. Liu, S.; Huang, E.; Zhou, Z.; Xu, Y.; Kui, X.; Lei, T.; Meng, H. Lightweight Facial Attractiveness Prediction Using Dual Label Distribution. IEEE Trans. Cogn. Dev. Syst. 2025. early access. [Google Scholar] [CrossRef]
  15. Tang, Y.; Ni, Z.; Zhou, J.; Zhang, D.; Lu, J.; Wu, Y.; Zhou, J. Uncertainty-Aware Score Distribution Learning for Action Quality Assessment. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9836–9845. [Google Scholar]
  16. Xing, C.; Geng, X.; Xue, H. Logistic Boosting Regression for Label Distribution Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4489–4497. [Google Scholar]
  17. Wang, J.; Geng, X. Label Distribution Learning by Exploiting Label Distribution Manifold. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 839–852. [Google Scholar] [CrossRef] [PubMed]
  18. Jia, X.; Shen, X.; Li, W.; Lu, Y.; Zhu, J. Label Distribution Learning by Maintaining Label Ranking Relation. IEEE Trans. Knowl. Data Eng. 2023, 35, 1695–1707. [Google Scholar] [CrossRef]
  19. Jia, X.; Qin, T.; Lu, Y.; Li, W. Adaptive Weighted Ranking-Oriented Label Distribution Learning. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 11302–11316. [Google Scholar] [CrossRef]
  20. Kou, Z.; Wang, J.; Jia, Y.; Liu, B.; Geng, X. Instance-Dependent Inaccurate Label Distribution Learning. IEEE Trans. Neural Netw. Learn. Syst. 2023, 36, 1425–1437. [Google Scholar] [CrossRef]
  21. Guiasu, S.; Shenitzer, A. The Principle of Maximum Entropy. Math. Intell. 1985, 7, 42–48. [Google Scholar] [CrossRef]
  22. Zhai, Y.; Dai, J. Geometric Mean Metric Learning for Label Distribution Learning. In Proceedings of the International Conference on Neural Information Processing, Sydney, NSW, Australia, 12–15 December 2019; pp. 260–272. [Google Scholar]
  23. Xu, S.; Ju, H.; Shang, L.; Pedrycz, W.; Yang, X.; Li, C. Label Distribution Learning: A Local Collaborative Mechanism. Int. J. Approx. Reason. 2020, 121, 59–84. [Google Scholar] [CrossRef]
  24. Friedman, J.; Hastie, T.; Tibshirani, R. Additive Logistic Regression: A Statistical View of Boosting. Ann. Stat. 2000, 28, 337–407. [Google Scholar] [CrossRef]
  25. Zhai, Y.; Dai, J.; Shi, H. Label Distribution Learning Based on Ensemble Neural Networks. In Proceedings of the International Conference on Neural Information Processing, Siem Reap, Cambodia, 13–16 December 2018; pp. 593–602. [Google Scholar]
  26. Chen, M.; Wang, X.; Feng, B.; Liu, W. Structured Random Forest for Label Distribution Learning. Neurocomputing 2018, 320, 171–182. [Google Scholar] [CrossRef]
  27. Jia, X.; Li, W.; Liu, J.; Zhang, Y. Label Distribution Learning by Exploiting Label Correlations. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 3310–3317. [Google Scholar]
  28. Zhao, P.; Zhou, Z.H. Label Distribution Learning by Optimal Transport. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 4506–4513. [Google Scholar]
  29. Zheng, X.; Jia, X.; Li, W. Label Distribution Learning by Exploiting Sample Correlations Locally. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 4556–4563. [Google Scholar]
  30. Ren, T.; Jia, X.; Li, W.; Zhao, S. Label Distribution Learning with Label Correlations via Low-Rank Approximation. In Proceedings of the International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 3325–3331. [Google Scholar]
  31. Ren, T.; Jia, X.; Li, W.; Chen, L.; Li, Z. Label Distribution Learning with Label-Specific Features. In Proceedings of the International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 3318–3324. [Google Scholar]
  32. Peng, C.L.; Tao, A.; Geng, X. Label Embedding Based on Multi-Scale Locality Preservation. In Proceedings of the International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 2623–2629. [Google Scholar]
  33. Wang, K.; Geng, X. Binary Coding Based Label Distribution Learning. In Proceedings of the International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 2783–2789. [Google Scholar]
  34. Wang, K.; Geng, X. Discrete Binary Coding Based Label Distribution Learning. In Proceedings of the International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 3733–3739. [Google Scholar]
  35. Xu, M.; Zhou, Z.H. Incomplete Label Distribution Learning. In Proceedings of the International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 3175–3181. [Google Scholar]
  36. Zeng, X.Q.; Chen, S.F.; Xiang, R.; Wu, S.X.; Wan, Z.Y. Filling Missing Values by Local Reconstruction for Incomplete Label Distribution Learning. Int. J. Wirel. Mob. Comput. 2019, 16, 314–321. [Google Scholar] [CrossRef]
  37. Xu, C.; Gu, S.; Tao, H.; Hou, C. Fragmentary Label Distribution Learning via Graph Regularized Maximum Entropy Criteria. Pattern Recognit. Lett. 2021, 145, 147–156. [Google Scholar] [CrossRef]
  38. Xu, S.; Shang, L.; Shen, F.; Yang, X.; Pedrycz, W. Incomplete Label Distribution Learning via Label Correlation Decomposition. Inf. Fusion 2025, 113, 102600. [Google Scholar] [CrossRef]
  39. Jin, Y.; Gao, R.; He, Y.; Zhu, X. GLDL: Graph Label Distribution Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; pp. 12965–12974. [Google Scholar]
  40. Lu, Y.; Li, W.; Liu, D.; Li, H.; Jia, X. Adaptive-Grained Label Distribution Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; pp. 19161–19169. [Google Scholar]
  41. He, L.; Lu, Y.; Li, W.; Jia, X. Generative Calibration of Inaccurate Annotation for Label Distribution Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; pp. 12394–12401. [Google Scholar]
  42. Xu, N.; Tao, A.; Geng, X. Label Enhancement for Label Distribution Learning. In Proceedings of the International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 1632–1643. [Google Scholar]
  43. Xu, N.; Liu, Y.P.; Geng, X. Label Enhancement for Label Distribution Learning. IEEE Trans. Knowl. Data Eng. 2021, 33, 1632–1643. [Google Scholar] [CrossRef]
  44. Gao, Y.; Zhang, Y.; Geng, X. Label Enhancement for Label Distribution Learning via Prior Knowledge. In Proceedings of the International Joint Conference on Artificial Intelligence, Yokohama, Japan, 11–17 July 2020; pp. 3223–3229. [Google Scholar]
  45. Lu, Y.; Jia, X. Predicting Label Distribution from Ternary Labels. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 10–15 December 2024; pp. 70431–70452. [Google Scholar]
  46. Lu, Y.; Jia, X. Predicting Label Distribution from Multi-Label Ranking. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022; pp. 36931–36943. [Google Scholar]
  47. Lu, Y.; Li, W.; Li, H.; Jia, X. Predicting Label Distribution from Tie-Allowed Multi-Label Ranking. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 15364–15379. [Google Scholar] [CrossRef] [PubMed]
  48. Lu, Y.; Li, W.; Li, H.; Jia, X. Ranking-Preserved Generative Label Enhancement. Mach. Learn. 2023, 112, 4693–4721. [Google Scholar] [CrossRef]
  49. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. In Proceedings of the International Conference on Learning Representations, Scottsdale, AZ, USA, 2–4 May 2013. [Google Scholar]
  50. Joo, W.; Lee, W.; Park, S.; Moon, I.C. Dirichlet Variational Autoencoder. Pattern Recognit. 2020, 107, 107514. [Google Scholar] [CrossRef]
  51. Lyons, M.; Akamatsu, S.; Kamachi, M.; Gyoba, J. Coding Facial Expressions with Gabor Wavelets. In Proceedings of the IEEE International Conference on Automatic Face and Gesture Recognition, Nara, Japan, 14–16 April 1998; pp. 200–205. [Google Scholar]
  52. Yin, L.; Wei, X.; Sun, Y.; Wang, J.; Rosato, M.J. A 3D Facial Expression Database For Facial Behavior Research. In Proceedings of the International Conference on Automatic Face and Gesture Recognition, Southampton, UK, 10–12 April 2006; pp. 211–216. [Google Scholar]
  53. Liu, X.; Zhu, J.; Zheng, Q.; Li, Z.; Liu, R.; Wang, J. Bidirectional Loss Function for Label Enhancement and Distribution Learning. Knowl.-Based Syst. 2021, 213, 106690. [Google Scholar] [CrossRef]
  54. Zychowski, A.; Mandziuk, J. Duo-LDL Method for Label Distribution Learning Based on Pairwise Class Dependencies. Appl. Soft Comput. 2021, 110, 107585. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of our proposed model. The latent variables u and ω correspond to the true label distribution and label confusion vector, respectively. The observed variables x and z correspond to the feature vector and noisy label distribution provided in the dataset, respectively.
Figure 1. Schematic diagram of our proposed model. The latent variables u and ω correspond to the true label distribution and label confusion vector, respectively. The observed variables x and z correspond to the feature vector and noisy label distribution provided in the dataset, respectively.
Electronics 14 02736 g001
Figure 2. Performance with different values of hyperparameters on the SJAFFE dataset with α * = 8 .
Figure 2. Performance with different values of hyperparameters on the SJAFFE dataset with α * = 8 .
Electronics 14 02736 g002
Figure 3. Convergence process of our proposed algorithm. The color of the curve corresponds to the dataset, while the transparency of the curve corresponds to the value of α * .
Figure 3. Convergence process of our proposed algorithm. The color of the curve corresponds to the dataset, while the transparency of the curve corresponds to the value of α * .
Electronics 14 02736 g003
Table 1. Commonly used notation in this paper.
Table 1. Commonly used notation in this paper.
NotationDescription
v , u Inner product of vectors v and u
X D D-dimensional feature space, i.e., R D
Δ M M-dimensional distribution space, i.e., { v R + M v , 1 = 1 }
x X D Random vector of feature variables
z Δ M Label distribution
H ( v ; Θ ) Multi-layer perceptron with v as input and Θ as parameters
v 2 L 2 norm of a vector v
f h Composition of functions h and f, i.e., f h ( · ) : = f ( h ( · ) )
Norm Normal distribution
Diri Dirichlet distribution
Table 2. Dataset statistics.
Table 2. Dataset statistics.
IDDataset# Instances# Features# Labels
1SJAFFE [51]2132436
2SBU-3DFE [52]25002436
3Yeast-alpha [1]24652418
4Yeast-cdc [1]24652415
5Yeast-cold [1]24652414
6Yeast-diau [1]2465247
7Yeast-dtt [1]2465246
8Yeast-elu [1]2465246
Table 3. Prediction performance on the first four datasets with α * = 4 .
Table 3. Prediction performance on the first four datasets with α * = 4 .
Cheb (↓)Clark (↓)Canber (↓)KL (↓)Cosine (↑)Intersec (↑)
SJAFFE
Ours 0.1352 ± 0.0151 1 /∗  0.4605 ± 0.0462 1 /∗  0.9507 ± 0.0895 1 /∗  0.1016 ± 0.0249 1 /∗  0.9051 ± 0.0223 1 /∗  0.8302 ± 0.0200 1 /∗ 
LDL-DPA 0.1568 ± 0.0093 4 /•  0.5436 ± 0.0268 5 /•  1.1121 ± 0.0555 6 /•  0.1420 ± 0.0160 3 /•  0.8747 ± 0.0127 4 /•  0.8005 ± 0.0105 6 /• 
LDL-LDM 0.1592 ± 0.0095 6 /•  0.5544 ± 0.0287 7 /•  1.1346 ± 0.0611 7 /•  0.1480 ± 0.0175 7 /•  0.8707 ± 0.0133 6 /•  0.7965 ± 0.0111 7 /• 
LDL-LRR 0.1568 ± 0.0096 4 /•  0.5427 ± 0.0270 3 /•  1.1092 ± 0.0571 3 /•  0.1421 ± 0.0160 4 /•  0.8745 ± 0.0130 5 /•  0.8006 ± 0.0109 3 /• 
Duo-LDL 0.2003 ± 0.0432 8 /•  0.9613 ± 0.3578 8 /•  1.8097 ± 0.6623 8 /•  1.7878 ± 1.5211 8 /•  0.8046 ± 0.0675 8 /•  0.7261 ± 0.0765 8 /• 
BD-LDL 0.1539 ± 0.0096 2 /•  0.5480 ± 0.0286 6 /•  1.1101 ± 0.0611 4 /•  0.1444 ± 0.0156 5 /•  0.8753 ± 0.0121 2 /•  0.8018 ± 0.0114 2 /• 
SABFGS 0.1565 ± 0.0095 3 /•  0.5432 ± 0.0275 4 /•  1.1113 ± 0.0570 5 /•  0.1419 ± 0.0163 2 /•  0.8750 ± 0.0131 3 /•  0.8006 ± 0.0111 3 /• 
AAkNN 0.1594 ± 0.0109 7 /•  0.5384 ± 0.0303 2 /•  1.0891 ± 0.0626 2 /•  0.1470 ± 0.0180 6 /•  0.8701 ± 0.0144 7 /•  0.8006 ± 0.0126 3 /• 
SBU-3DFE
Ours 0.1550 ± 0.0083 1 /∗  0.4858 ± 0.0288 1 /∗  0.9951 ± 0.0509 1 /∗  0.1160 ± 0.0139 1 /∗  0.8902 ± 0.0122 1 /∗  0.8169 ± 0.0099 1 /∗ 
LDL-DPA 0.1668 ± 0.0018 3 /•  0.5802 ± 0.0031 3 /•  1.1726 ± 0.0063 3 /•  0.1532 ± 0.0026 3 /•  0.8647 ± 0.0021 3 /•  0.7889 ± 0.0015 4 /• 
LDL-LDM 0.1641 ± 0.0021 2 /•  0.5661 ± 0.0043 2 /•  1.1468 ± 0.0080 2 /•  0.1472 ± 0.0029 2 /•  0.8690 ± 0.0024 2 /•  0.7930 ± 0.0019 2 /• 
LDL-LRR 0.1670 ± 0.0023 4 /•  0.5805 ± 0.0033 4 /•  1.1731 ± 0.0070 4 /•  0.1535 ± 0.0028 5 /•  0.8646 ± 0.0024 4 /•  0.7889 ± 0.0019 4 /• 
Duo-LDL 0.1894 ± 0.0185 8 /•  0.7377 ± 0.2364 8 /•  1.4694 ± 0.3941 8 /•  0.7575 ± 1.2036 8 /∗  0.8269 ± 0.0315 8 /•  0.7500 ± 0.0378 8 /• 
BD-LDL 0.1723 ± 0.0020 7 /•  0.6014 ± 0.0043 7 /•  1.2334 ± 0.0083 7 /•  0.1655 ± 0.0038 7 /•  0.8562 ± 0.0024 7 /•  0.7802 ± 0.0018 7 /• 
SABFGS 0.1670 ± 0.0022 4 /•  0.5805 ± 0.0034 4 /•  1.1731 ± 0.0068 4 /•  0.1534 ± 0.0026 4 /•  0.8646 ± 0.0023 4 /•  0.7890 ± 0.0018 3 /• 
AAkNN 0.1675 ± 0.0023 6 /•  0.5819 ± 0.0038 6 /•  1.1823 ± 0.0092 6 /•  0.1560 ± 0.0034 6 /•  0.8638 ± 0.0029 6 /•  0.7878 ± 0.0020 6 /• 
Yeast-alpha
Ours 0.0171 ± 0.0006 1 /∗  0.2915 ± 0.0108 1 /∗  0.9806 ± 0.0395 1 /∗  0.0098 ± 0.0006 1 /∗  0.9904 ± 0.0007 1 /∗  0.9456 ± 0.0022 1 /∗ 
LDL-DPA 0.0230 ± 0.0000 2 /•  0.4181 ± 0.0021 4 /•  1.4358 ± 0.0084 4 /•  0.0197 ± 0.0005 2 /•  0.9810 ± 0.0000 2 /•  0.9206 ± 0.0005 3 /• 
LDL-LDM 0.0230 ± 0.0000 2 /•  0.4176 ± 0.0025 2 /•  1.4343 ± 0.0085 2 /•  0.0197 ± 0.0005 2 /•  0.9810 ± 0.0000 2 /•  0.9207 ± 0.0005 2 /• 
LDL-LRR 0.0230 ± 0.0000 2 /•  0.4179 ± 0.0024 3 /•  1.4357 ± 0.0084 3 /•  0.0197 ± 0.0005 2 /•  0.9810 ± 0.0000 2 /•  0.9206 ± 0.0005 3 /• 
Duo-LDL 0.0230 ± 0.0000 2 /•  0.4221 ± 0.0026 7 /•  1.4478 ± 0.0088 7 /•  0.0200 ± 0.0000 7 /•  0.9807 ± 0.0005 7 /•  0.9201 ± 0.0006 7 /• 
BD-LDL 0.0258 ± 0.0004 8 /•  0.4614 ± 0.0039 8 /•  1.5807 ± 0.0127 8 /•  0.0243 ± 0.0005 8 /•  0.9766 ± 0.0005 8 /•  0.9125 ± 0.0007 8 /• 
SABFGS 0.0230 ± 0.0000 2 /•  0.4181 ± 0.0021 4 /•  1.4360 ± 0.0084 5 /•  0.0197 ± 0.0005 2 /•  0.9810 ± 0.0000 2 /•  0.9206 ± 0.0005 3 /• 
AAkNN 0.0230 ± 0.0000 2 /•  0.4182 ± 0.0023 6 /•  1.4366 ± 0.0086 6 /•  0.0197 ± 0.0005 2 /•  0.9810 ± 0.0000 2 /•  0.9206 ± 0.0005 3 /• 
Yeast-cdc
Ours 0.0217 ± 0.0007 1 /∗  0.2891 ± 0.0088 1 /∗  0.8857 ± 0.0294 1 /∗  0.0115 ± 0.0007 1 /∗  0.9887 ± 0.0008 1 /∗  0.9417 ± 0.0018 1 /∗ 
LDL-DPA 0.0285 ± 0.0005 4 /•  0.4309 ± 0.0019 6 /•  1.3419 ± 0.0046 4 /•  0.0245 ± 0.0005 2 /•  0.9772 ± 0.0004 2 /•  0.9129 ± 0.0003 2 /• 
LDL-LDM 0.0284 ± 0.0005 2 /•  0.4307 ± 0.0018 2 /•  1.3408 ± 0.0043 2 /•  0.0245 ± 0.0005 2 /•  0.9772 ± 0.0004 2 /•  0.9129 ± 0.0003 2 /• 
LDL-LRR 0.0286 ± 0.0005 6 /•  0.4307 ± 0.0018 2 /•  1.3416 ± 0.0045 3 /•  0.0245 ± 0.0005 2 /•  0.9772 ± 0.0004 2 /•  0.9129 ± 0.0003 2 /• 
Duo-LDL 0.0290 ± 0.0005 7 /•  0.4330 ± 0.0090 7 /•  1.3457 ± 0.0232 7 /•  0.0248 ± 0.0008 7 /•  0.9771 ± 0.0007 7 /•  0.9125 ± 0.0014 7 /• 
BD-LDL 0.0311 ± 0.0003 8 /•  0.4681 ± 0.0035 8 /•  1.4550 ± 0.0108 8 /•  0.0290 ± 0.0005 8 /•  0.9735 ± 0.0005 8 /•  0.9054 ± 0.0008 8 /• 
SABFGS 0.0285 ± 0.0005 4 /•  0.4307 ± 0.0018 2 /•  1.3419 ± 0.0041 4 /•  0.0245 ± 0.0005 2 /•  0.9772 ± 0.0004 2 /•  0.9129 ± 0.0003 2 /• 
AAkNN 0.0284 ± 0.0005 2 /•  0.4308 ± 0.0018 5 /•  1.3420 ± 0.0051 6 /•  0.0246 ± 0.0005 6 /•  0.9772 ± 0.0004 2 /•  0.9129 ± 0.0003 2 /• 
* Best performance in each comparison group is boldfaced, with rank and statistical significance indicators shown in the top-left corner of each result.
Table 4. Prediction performance on the last four datasets with α * = 4 .
Table 4. Prediction performance on the last four datasets with α * = 4 .
Cheb (↓)Clark (↓)Canber (↓)KL (↓)Cosine (↑)Intersec (↑)
Yeast-cold
Ours 0.0569 ± 0.0021 1 /∗  0.1560 ± 0.0060 1 /∗  0.2701 ± 0.0117 1 /∗  0.0145 ± 0.0011 1 /∗  0.9862 ± 0.0009 1 /∗  0.9334 ± 0.0031 1 /∗ 
LDL-DPA 0.0654 ± 0.0008 4 /•  0.1829 ± 0.0021 4 /•  0.3243 ± 0.0035 5 /•  0.0195 ± 0.0005 3 /•  0.9814 ± 0.0005 4 /•  0.9198 ± 0.0009 4 /• 
LDL-LDM 0.0611 ± 0.0007 2 /•  0.1691 ± 0.0023 2 /•  0.2971 ± 0.0045 2 /•  0.0167 ± 0.0005 2 /•  0.9840 ± 0.0005 2 /•  0.9266 ± 0.0012 2 /• 
LDL-LRR 0.0654 ± 0.0008 4 /•  0.1827 ± 0.0021 3 /•  0.3242 ± 0.0038 4 /•  0.0195 ± 0.0005 3 /•  0.9815 ± 0.0005 3 /•  0.9197 ± 0.0011 5 /• 
Duo-LDL 0.0656 ± 0.0010 6 /•  0.1835 ± 0.0024 6 /•  0.3251 ± 0.0042 6 /•  0.0196 ± 0.0005 6 /•  0.9814 ± 0.0005 4 /•  0.9195 ± 0.0010 6 /• 
BD-LDL 0.0760 ± 0.0018 8 /•  0.2126 ± 0.0049 8 /•  0.3607 ± 0.0073 8 /•  0.0272 ± 0.0013 8 /•  0.9765 ± 0.0011 8 /•  0.9127 ± 0.0018 8 /• 
SABFGS 0.0653 ± 0.0007 3 /•  0.1829 ± 0.0021 4 /•  0.3240 ± 0.0032 3 /•  0.0195 ± 0.0005 3 /•  0.9814 ± 0.0005 4 /•  0.9199 ± 0.0009 3 /• 
AAkNN 0.0661 ± 0.0007 7 /•  0.1839 ± 0.0019 7 /•  0.3253 ± 0.0037 7 /•  0.0197 ± 0.0005 7 /•  0.9812 ± 0.0004 7 /•  0.9195 ± 0.0008 6 /• 
Yeast-diau
Ours 0.0450 ± 0.0017 1 /∗  0.2378 ± 0.0061 1 /∗  0.5083 ± 0.0119 1 /∗  0.0175 ± 0.0007 1 /∗  0.9836 ± 0.0008 1 /∗  0.9291 ± 0.0015 1 /∗ 
LDL-DPA 0.0623 ± 0.0005 5 /•  0.3301 ± 0.0024 5 /•  0.7049 ± 0.0051 5 /•  0.0322 ± 0.0006 3 /•  0.9715 ± 0.0005 4 /•  0.9028 ± 0.0008 3 /• 
LDL-LDM 0.0616 ± 0.0005 2 /•  0.3242 ± 0.0025 2 /•  0.6936 ± 0.0051 2 /•  0.0311 ± 0.0006 2 /•  0.9723 ± 0.0005 2 /•  0.9042 ± 0.0009 2 /• 
LDL-LRR 0.0624 ± 0.0005 6 /•  0.3301 ± 0.0024 5 /•  0.7049 ± 0.0054 5 /•  0.0322 ± 0.0006 3 /•  0.9715 ± 0.0005 4 /•  0.9028 ± 0.0008 3 /• 
Duo-LDL 0.0622 ± 0.0010 4 /•  0.3299 ± 0.0055 3 /•  0.7039 ± 0.0164 3 /•  0.0323 ± 0.0009 7 /•  0.9716 ± 0.0011 3 /•  0.9028 ± 0.0025 3 /• 
BD-LDL 0.0684 ± 0.0013 8 /•  0.3903 ± 0.0083 8 /•  0.8333 ± 0.0158 8 /•  0.0447 ± 0.0022 8 /•  0.9641 ± 0.0014 8 /•  0.8886 ± 0.0020 8 /• 
SABFGS 0.0624 ± 0.0005 6 /•  0.3299 ± 0.0024 3 /•  0.7048 ± 0.0054 4 /•  0.0322 ± 0.0006 3 /•  0.9715 ± 0.0005 4 /•  0.9028 ± 0.0008 3 /• 
AAkNN 0.0620 ± 0.0007 3 /•  0.3303 ± 0.0023 7 /•  0.7067 ± 0.0050 7 /•  0.0322 ± 0.0006 3 /•  0.9715 ± 0.0005 4 /•  0.9025 ± 0.0008 7 /• 
Yeast-dtt
Ours 0.0464 ± 0.0025 1 /∗  0.1274 ± 0.0069 1 /∗  0.2213 ± 0.0113 1 /∗  0.0099 ± 0.0009 1 /∗  0.9903 ± 0.0011 1 /∗  0.9453 ± 0.0027 1 /∗ 
LDL-DPA 0.0571 ± 0.0006 3 /•  0.1612 ± 0.0016 4 /•  0.2892 ± 0.0029 3 /•  0.0151 ± 0.0003 3 /•  0.9851 ± 0.0003 5 /•  0.9286 ± 0.0008 3 /• 
LDL-LDM 0.0523 ± 0.0007 2 /•  0.1466 ± 0.0018 2 /•  0.2614 ± 0.0037 2 /•  0.0129 ± 0.0003 2 /•  0.9876 ± 0.0005 2 /•  0.9352 ± 0.0008 2 /• 
LDL-LRR 0.0571 ± 0.0006 3 /•  0.1613 ± 0.0016 5 /•  0.2894 ± 0.0030 5 /•  0.0151 ± 0.0003 3 /•  0.9852 ± 0.0004 4 /•  0.9284 ± 0.0008 5 /• 
Duo-LDL 0.0575 ± 0.0013 6 /•  0.1627 ± 0.0031 7 /•  0.2918 ± 0.0055 7 /•  0.0154 ± 0.0007 6 /•  0.9851 ± 0.0006 5 /•  0.9278 ± 0.0012 7 /• 
BD-LDL 0.0786 ± 0.0023 8 /•  0.2331 ± 0.0098 8 /•  0.3852 ± 0.0146 8 /•  0.0599 ± 0.0336 8 /•  0.9738 ± 0.0019 8 /•  0.9096 ± 0.0028 8 /• 
SABFGS 0.0571 ± 0.0006 3 /•  0.1611 ± 0.0017 3 /•  0.2893 ± 0.0030 4 /•  0.0151 ± 0.0003 3 /•  0.9853 ± 0.0005 3 /•  0.9286 ± 0.0008 3 /• 
AAkNN 0.0575 ± 0.0008 6 /•  0.1622 ± 0.0018 6 /•  0.2909 ± 0.0034 6 /•  0.0155 ± 0.0005 7 /•  0.9850 ± 0.0005 7 /•  0.9281 ± 0.0011 6 /• 
Yeast-elu
Ours 0.0239 ± 0.0013 1 /∗  0.2909 ± 0.0054 1 /∗  0.8788 ± 0.0148 1 /∗  0.0129 ± 0.0006 1 /∗  0.9872 ± 0.0006 1 /∗  0.9370 ± 0.0009 1 /∗ 
LDL-DPA 0.0376 ± 0.0005 3 /•  0.4347 ± 0.0017 4 /•  1.3238 ± 0.0074 4 /•  0.0281 ± 0.0003 2 /•  0.9718 ± 0.0004 5 /•  0.9041 ± 0.0006 2 /• 
LDL-LDM 0.0374 ± 0.0005 2 /•  0.4341 ± 0.0019 2 /•  1.3230 ± 0.0075 2 /•  0.0281 ± 0.0003 2 /•  0.9719 ± 0.0003 2 /•  0.9041 ± 0.0006 2 /• 
LDL-LRR 0.0376 ± 0.0005 3 /•  0.4347 ± 0.0018 4 /•  1.3238 ± 0.0075 4 /•  0.0281 ± 0.0003 2 /•  0.9719 ± 0.0003 2 /•  0.9041 ± 0.0006 2 /• 
Duo-LDL 0.0376 ± 0.0005 3 /•  0.4379 ± 0.0037 7 /•  1.3333 ± 0.0128 7 /•  0.0286 ± 0.0005 7 /•  0.9714 ± 0.0005 7 /•  0.9035 ± 0.0010 7 /• 
BD-LDL 0.0407 ± 0.0005 8 /•  0.4924 ± 0.0043 8 /•  1.5042 ± 0.0140 8 /•  0.0361 ± 0.0006 8 /•  0.9653 ± 0.0005 8 /•  0.8925 ± 0.0008 8 /• 
SABFGS 0.0376 ± 0.0005 3 /•  0.4345 ± 0.0018 3 /•  1.3236 ± 0.0076 3 /•  0.0281 ± 0.0003 2 /•  0.9719 ± 0.0003 2 /•  0.9041 ± 0.0006 2 /• 
AAkNN 0.0376 ± 0.0005 3 /•  0.4351 ± 0.0021 6 /•  1.3259 ± 0.0078 6 /•  0.0281 ± 0.0003 2 /•  0.9717 ± 0.0005 6 /•  0.9041 ± 0.0006 2 /• 
* Best performance in each comparison group is boldfaced, with rank and statistical significance indicators shown in the top-left corner of each result.
Table 5. Prediction performance on the first four datasets with α * = 8 .
Table 5. Prediction performance on the first four datasets with α * = 8 .
Cheb (↓)Clark (↓)Canber (↓)KL (↓)Cosine (↑)Intersec (↑)
SJAFFE
Ours 0.1287 ± 0.0105 1 /∗  0.4447 ± 0.0346 1 /∗  0.9253 ± 0.0740 1 /∗  0.0900 ± 0.0172 1 /∗  0.9142 ± 0.0154 1 /∗  0.8357 ± 0.0144 1 /∗ 
LDL-DPA 0.1331 ± 0.0092 4 /∗  0.4789 ± 0.0275 6 /•  0.9885 ± 0.0609 6 /•  0.1059 ± 0.0148 4 /•  0.9046 ± 0.0118 4 /•  0.8263 ± 0.0113 5 /• 
LDL-LDM 0.1356 ± 0.0091 7 /•  0.4932 ± 0.0296 7 /•  1.0201 ± 0.0634 7 /•  0.1121 ± 0.0153 7 /•  0.9002 ± 0.0121 7 /•  0.8213 ± 0.0117 7 /• 
LDL-LRR 0.1328 ± 0.0089 3 /∗  0.4756 ± 0.0286 5 /•  0.9811 ± 0.0626 5 /•  0.1044 ± 0.0152 3 /•  0.9054 ± 0.0122 3 /•  0.8272 ± 0.0115 4 /• 
Duo-LDL 0.1705 ± 0.0300 8 /•  0.7557 ± 0.3338 8 /•  1.4350 ± 0.5434 8 /•  1.0511 ± 1.2603 8 /•  0.8534 ± 0.0427 8 /•  0.7748 ± 0.0535 8 /• 
BD-LDL 0.1352 ± 0.0093 5 /∗  0.4730 ± 0.0244 3 /•  0.9778 ± 0.0541 3 /∗  0.1077 ± 0.0126 6 /•  0.9012 ± 0.0106 6 /•  0.8253 ± 0.0103 6 /∗ 
SABFGS 0.1323 ± 0.0086 2 /∗  0.4755 ± 0.0225 4 /•  0.9803 ± 0.0497 4 /•  0.1042 ± 0.0130 2 /•  0.9056 ± 0.0107 2 /∗  0.8273 ± 0.0098 3 /∗ 
AAkNN 0.1361 ± 0.0100 6 /∗  0.4588 ± 0.0275 2 /∗  0.9384 ± 0.0589 2 /∗  0.1063 ± 0.0146 5 /•  0.9014 ± 0.0125 5 /∗  0.8289 ± 0.0117 2 /∗ 
SBU-3DFE
Ours 0.1428 ± 0.0037 2 /∗  0.4466 ± 0.0182 1 /∗  0.9295 ± 0.0316 1 /∗  0.0962 ± 0.0072 1 /∗  0.9078 ± 0.0061 1 /∗  0.8308 ± 0.0062 1 /∗ 
LDL-DPA 0.1451 ± 0.0021 5 /∗  0.4914 ± 0.0037 4 /•  0.9952 ± 0.0063 5 /•  0.1089 ± 0.0020 5 /•  0.8983 ± 0.0019 5 /•  0.8206 ± 0.0015 4 /• 
LDL-LDM 0.1425 ± 0.0023 1 /∗  0.4658 ± 0.0044 2 /•  0.9508 ± 0.0094 2 /∗  0.1013 ± 0.0024 2 /•  0.9045 ± 0.0020 2 /∗  0.8274 ± 0.0020 2 /∗ 
LDL-LRR 0.1449 ± 0.0020 4 /•  0.4916 ± 0.0035 6 /•  0.9950 ± 0.0067 4 /•  0.1088 ± 0.0020 4 /•  0.8985 ± 0.0019 3 /•  0.8207 ± 0.0016 3 /• 
Duo-LDL 0.1724 ± 0.0192 8 /•  0.7074 ± 0.2986 8 /•  1.3653 ± 0.4799 8 /•  0.7589 ± 1.1922 8 /∗  0.8566 ± 0.0354 8 /•  0.7764 ± 0.0451 8 /• 
BD-LDL 0.1561 ± 0.0021 7 /•  0.5078 ± 0.0044 7 /•  1.0467 ± 0.0096 7 /•  0.1239 ± 0.0027 7 /•  0.8850 ± 0.0023 7 /•  0.8095 ± 0.0020 7 /• 
SABFGS 0.1448 ± 0.0022 3 /∗  0.4914 ± 0.0041 4 /•  0.9948 ± 0.0079 3 /•  0.1087 ± 0.0020 3 /•  0.8985 ± 0.0019 3 /•  0.8206 ± 0.0014 4 /• 
AAkNN 0.1461 ± 0.0024 6 /•  0.4906 ± 0.0044 3 /•  0.9987 ± 0.0097 6 /•  0.1124 ± 0.0028 6 /•  0.8964 ± 0.0025 6 /•  0.8201 ± 0.0021 6 /• 
Yeast-alpha
Ours 0.0166 ± 0.0005 1 /∗  0.2737 ± 0.0074 1 /∗  0.9127 ± 0.0288 1 /∗  0.0088 ± 0.0006 1 /∗  0.9914 ± 0.0005 1 /∗  0.9496 ± 0.0016 1 /∗ 
LDL-DPA 0.0220 ± 0.0000 2 /•  0.3873 ± 0.0024 4 /•  1.3713 ± 0.0083 3 /•  0.0170 ± 0.0000 2 /•  0.9836 ± 0.0005 2 /•  0.9271 ± 0.0006 4 /• 
LDL-LDM 0.0220 ± 0.0000 2 /•  0.3864 ± 0.0025 2 /•  1.3145 ± 0.0085 2 /•  0.0170 ± 0.0000 2 /•  0.9836 ± 0.0005 2 /•  0.9275 ± 0.0007 2 /• 
LDL-LRR 0.0220 ± 0.0000 2 /•  0.3872 ± 0.0026 2 /•  1.3176 ± 0.0082 5 /•  0.0170 ± 0.0000 2 /•  0.9836 ± 0.0005 2 /•  0.9271 ± 0.0006 4 /• 
Duo-LDL 0.0220 ± 0.0000 2 /•  0.3910 ± 0.0032 7 /•  1.3304 ± 0.0112 7 /•  0.0173 ± 0.0005 7 /•  0.9832 ± 0.0004 7 /•  0.9266 ± 0.0007 7 /• 
BD-LDL 0.0252 ± 0.0006 8 /•  0.4375 ± 0.0047 8 /•  1.4867 ± 0.0155 8 /•  0.0217 ± 0.0005 8 /•  0.9789 ± 0.0006 8 /•  0.9179 ± 0.0010 8 /• 
SABFGS 0.0220 ± 0.0000 2 /•  0.3873 ± 0.0023 4 /•  1.3175 ± 0.0078 4 /•  0.0170 ± 0.0000 2 /•  0.9836 ± 0.0005 2 /•  0.9272 ± 0.0006 3 /• 
AAkNN 0.0220 ± 0.0000 2 /•  0.3873 ± 0.0025 4 /•  1.3186 ± 0.0089 6 /•  0.0170 ± 0.0000 2 /•  0.9836 ± 0.0005 2 /•  0.9271 ± 0.0006 4 /• 
Yeast-cdc
Ours 0.0207 ± 0.0007 1 /∗  0.2821 ± 0.0093 1 /∗  0.8627 ± 0.0312 1 /∗  0.0110 ± 0.0008 1 /∗  0.9894 ± 0.0005 1 /∗  0.9432 ± 0.0019 1 /∗ 
LDL-DPA 0.0269 ± 0.0003 2 /•  0.3979 ± 0.0021 5 /•  1.2301 ± 0.0051 4 /•  0.0210 ± 0.0000 2 /•  0.9806 ± 0.0005 3 /•  0.9201 ± 0.0003 2 /• 
LDL-LDM 0.0269 ± 0.0003 2 /•  0.3972 ± 0.0022 2 /•  1.2285 ± 0.0048 2 /•  0.0210 ± 0.0000 2 /•  0.9808 ± 0.0004 2 /•  0.9201 ± 0.0003 2 /• 
LDL-LRR 0.0269 ± 0.0003 2 /•  0.3977 ± 0.0022 4 /•  1.2300 ± 0.0052 3 /•  0.0210 ± 0.0000 2 /•  0.9805 ± 0.0005 5 /•  0.9201 ± 0.0003 2 /• 
Duo-LDL 0.0271 ± 0.0007 7 /•  0.4053 ± 0.0133 7 /•  1.2512 ± 0.0507 7 /•  0.0217 ± 0.0016 7 /•  0.9799 ± 0.0014 7 /•  0.9184 ± 0.0038 7 /• 
BD-LDL 0.0303 ± 0.0005 8 /•  0.4416 ± 0.0048 8 /•  1.3629 ± 0.0131 8 /•  0.0261 ± 0.0007 8 /•  0.9761 ± 0.0006 8 /•  0.9113 ± 0.0008 8 /• 
SABFGS 0.0269 ± 0.0003 2 /•  0.3976 ± 0.0021 3 /•  1.2303 ± 0.0051 5 /•  0.0210 ± 0.0000 2 /•  0.9806 ± 0.0005 3 /•  0.9201 ± 0.0003 2 /• 
AAkNN 0.0269 ± 0.0003 2 /•  0.3982 ± 0.0021 6 /•  1.2315 ± 0.0057 6 /•  0.0210 ± 0.0000 2 /•  0.9805 ± 0.0005 5 /•  0.9200 ± 0.0005 6 /• 
* Best performance in each comparison group is boldfaced, with rank and statistical significance indicators shown in the top-left corner of each result.
Table 6. Prediction performance on the last four datasets with α * = 8 .
Table 6. Prediction performance on the last four datasets with α * = 8 .
Cheb (↓)Clark (↓)Canber (↓)KL (↓)Cosine (↑)Intersec (↑)
Yeast-cold
Ours 0.0568 ± 0.0019 2 /∗  0.1542 ± 0.0048 1 /∗  0.2660 ± 0.0082 1 /∗  0.0145 ± 0.0008 1 /∗  0.9864 ± 0.0007 1 /∗  0.9344 ± 0.0020 1 /∗ 
LDL-DPA 0.0632 ± 0.0006 3 /•  0.1746 ± 0.0016 5 /•  0.3030 ± 0.0031 5 /•  0.0177 ± 0.0005 3 /•  0.9832 ± 0.0004 3 /•  0.9256 ± 0.0007 3 /• 
LDL-LDM 0.0567 ± 0.0008 1 /∗  0.1552 ± 0.0023 2 /∗  0.2684 ± 0.0043 2 /∗  0.0145 ± 0.0005 1 /∗  0.9864 ± 0.0005 1 /∗  0.9337 ± 0.0009 2 /∗ 
LDL-LRR 0.0632 ± 0.0006 3 /•  0.1744 ± 0.0020 3 /•  0.3029 ± 0.0034 3 /•  0.0177 ± 0.0005 3 /•  0.9832 ± 0.0004 3 /•  0.9256 ± 0.0007 3 /• 
Duo-LDL 0.0643 ± 0.0038 7 /•  0.1780 ± 0.0103 7 /•  0.3086 ± 0.0138 7 /•  0.0185 ± 0.0020 7 /•  0.9826 ± 0.0013 7 /•  0.9245 ± 0.0028 7 /• 
BD-LDL 0.0673 ± 0.0018 8 /•  0.1859 ± 0.0042 8 /•  0.3176 ± 0.0064 8 /•  0.0212 ± 0.0010 8 /•  0.9809 ± 0.0009 8 /•  0.9225 ± 0.0016 8 /• 
SABFGS 0.0632 ± 0.0006 3 /•  0.1744 ± 0.0020 3 /•  0.3029 ± 0.0033 3 /•  0.0177 ± 0.0005 3 /•  0.9832 ± 0.0004 3 /•  0.9256 ± 0.0007 3 /• 
AAkNN 0.0639 ± 0.0007 6 /•  0.1760 ± 0.0021 6 /•  0.3062 ± 0.0037 6 /•  0.0183 ± 0.0005 6 /•  0.9829 ± 0.0003 6 /•  0.9248 ± 0.0009 6 /• 
Yeast-diau
Ours 0.0443 ± 0.0014 1 /∗  0.2326 ± 0.0057 1 /∗  0.4979 ± 0.0122 1 /∗  0.0169 ± 0.0009 1 /∗  0.9844 ± 0.0008 1 /∗  0.9306 ± 0.0017 1 /∗ 
LDL-DPA 0.0584 ± 0.0005 5 /•  0.3011 ± 0.0025 5 /•  0.6395 ± 0.0053 5 /•  0.0272 ± 0.0004 3 /•  0.9755 ± 0.0005 3 /•  0.9112 ± 0.0009 4 /• 
LDL-LDM 0.0567 ± 0.0005 2 /•  0.2902 ± 0.0023 2 /•  0.6169 ± 0.0045 2 /•  0.0253 ± 0.0005 2 /•  0.9767 ± 0.0005 2 /•  0.9141 ± 0.0009 2 /• 
LDL-LRR 0.0584 ± 0.0005 5 /•  0.3008 ± 0.0027 3 /•  0.6394 ± 0.0053 4 /•  0.0272 ± 0.0004 3 /•  0.9755 ± 0.0005 3 /•  0.9112 ± 0.0009 4 /• 
Duo-LDL 0.0584 ± 0.0005 5 /•  0.3014 ± 0.0023 7 /•  0.6398 ± 0.0064 6 /•  0.0273 ± 0.0005 6 /•  0.9754 ± 0.0005 6 /•  0.9113 ± 0.0011 3 /• 
BD-LDL 0.0671 ± 0.0018 8 /•  0.4045 ± 0.0113 8 /•  0.7997 ± 0.0162 8 /•  0.0980 ± 0.0258 8 /•  0.9676 ± 0.0011 8 /•  0.8974 ± 0.0017 8 /• 
SABFGS 0.0583 ± 0.0005 4 /•  0.3009 ± 0.0026 4 /•  0.6393 ± 0.0051 3 /•  0.0272 ± 0.0004 3 /•  0.9755 ± 0.0005 3 /•  0.9112 ± 0.0009 4 /• 
AAkNN 0.0581 ± 0.0006 3 /•  0.3013 ± 0.0025 6 /•  0.6414 ± 0.0051 7 /•  0.0273 ± 0.0005 6 /•  0.9754 ± 0.0005 6 /•  0.9108 ± 0.0008 7 /• 
Yeast-dtt
Ours 0.0446 ± 0.0017 1 /∗  0.1224 ± 0.0061 1 /∗  0.2116 ± 0.0120 1 /∗  0.0093 ± 0.0009 1 /∗  0.9911 ± 0.0007 1 /∗  0.9479 ± 0.0029 1 /∗ 
LDL-DPA 0.0526 ± 0.0007 3 /•  0.1485 ± 0.0013 4 /•  0.2575 ± 0.0028 4 /•  0.0130 ± 0.0000 3 /•  0.9879 ± 0.0003 3 /•  0.9371 ± 0.0007 3 /• 
LDL-LDM 0.0458 ± 0.0004 2 /∗  0.1271 ± 0.0006 2 /•  0.2215 ± 0.0020 2 /•  0.0100 ± 0.0000 2 /•  0.9907 ± 0.0005 2 /∗  0.9455 ± 0.0005 2 /• 
LDL-LRR 0.0527 ± 0.0005 5 /•  0.1485 ± 0.0013 4 /•  0.2576 ± 0.0026 5 /•  0.0130 ± 0.0000 3 /•  0.9879 ± 0.0003 3 /•  0.9371 ± 0.0007 3 /• 
Duo-LDL 0.0546 ± 0.0051 7 /•  0.1540 ± 0.0143 7 /•  0.2668 ± 0.0211 7 /•  0.0139 ± 0.0022 7 /•  0.9872 ± 0.0016 7 /•  0.9350 ± 0.0044 7 /• 
BD-LDL 0.0675 ± 0.0022 8 /•  0.1996 ± 0.0080 8 /•  0.3353 ± 0.0143 8 /•  0.0278 ± 0.0033 8 /•  0.9791 ± 0.0017 8 /•  0.9204 ± 0.0031 8 /• 
SABFGS 0.0526 ± 0.0007 3 /•  0.1484 ± 0.0013 3 /•  0.2574 ± 0.0028 3 /•  0.0130 ± 0.0000 3 /•  0.9879 ± 0.0003 3 /•  0.9371 ± 0.0007 3 /• 
AAkNN 0.0534 ± 0.0008 6 /•  0.1505 ± 0.0018 6 /•  0.2613 ± 0.0033 6 /•  0.0133 ± 0.0005 6 /•  0.9875 ± 0.0005 6 /•  0.9364 ± 0.0008 6 /• 
Yeast-elu
Ours 0.0214 ± 0.0007 1 /∗  0.2700 ± 0.0050 1 /∗  0.8153 ± 0.0138 1 /∗  0.0110 ± 0.0005 1 /∗  0.9893 ± 0.0007 1 /∗  0.9420 ± 0.0011 1 /∗ 
LDL-DPA 0.0320 ± 0.0000 2 /•  0.3824 ± 0.0022 5 /•  1.1562 ± 0.0078 5 /•  0.0218 ± 0.0004 2 /•  0.9780 ± 0.0000 3 /•  0.9164 ± 0.0007 4 /• 
LDL-LDM 0.0320 ± 0.0000 2 /•  0.3817 ± 0.0021 2 /•  1.1545 ± 0.0079 2 /•  0.0218 ± 0.0004 2 /•  0.9782 ± 0.0004 2 /•  0.9168 ± 0.0008 2 /• 
LDL-LRR 0.0320 ± 0.0000 2 /•  0.3823 ± 0.0022 3 /•  1.1560 ± 0.0077 3 /•  0.0218 ± 0.0004 2 /•  0.9780 ± 0.0000 3 /•  0.9164 ± 0.0007 4 /• 
Duo-LDL 0.0322 ± 0.0004 7 /•  0.3845 ± 0.0023 7 /•  1.1621 ± 0.0091 7 /•  0.0220 ± 0.0000 6 /•  0.9780 ± 0.0000 3 /•  0.9162 ± 0.0008 6 /• 
BD-LDL 0.0361 ± 0.0007 8 /•  0.4474 ± 0.0052 8 /•  1.3617 ± 0.0166 8 /•  0.0298 ± 0.0009 8 /•  0.9718 ± 0.0006 8 /•  0.9032 ± 0.0010 8 /• 
SABFGS 0.0320 ± 0.0000 2 /•  0.3823 ± 0.0022 3 /•  1.1561 ± 0.0080 4 /•  0.0218 ± 0.0004 2 /•  0.9780 ± 0.0000 3 /•  0.9166 ± 0.0007 3 /• 
AAkNN 0.0321 ± 0.0003 6 /•  0.3835 ± 0.0021 6 /•  1.1603 ± 0.0083 6 /•  0.0220 ± 0.0000 6 /•  0.9780 ± 0.0000 3 /•  0.9161 ± 0.0006 7 /• 
* Best performance in each comparison group is boldfaced, with rank and statistical significance indicators shown in the top-left corner of each result.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, X.; Meng, C.; Zhou, H.; Guo, Y.; Xue, B.; Yu, T.; Lu, Y. Generative Learning from Semantically Confused Label Distribution via Auto-Encoding Variational Bayes. Electronics 2025, 14, 2736. https://doi.org/10.3390/electronics14132736

AMA Style

Li X, Meng C, Zhou H, Guo Y, Xue B, Yu T, Lu Y. Generative Learning from Semantically Confused Label Distribution via Auto-Encoding Variational Bayes. Electronics. 2025; 14(13):2736. https://doi.org/10.3390/electronics14132736

Chicago/Turabian Style

Li, Xinhai, Chenxu Meng, Heng Zhou, Yi Guo, Bowen Xue, Tianzuo Yu, and Yunan Lu. 2025. "Generative Learning from Semantically Confused Label Distribution via Auto-Encoding Variational Bayes" Electronics 14, no. 13: 2736. https://doi.org/10.3390/electronics14132736

APA Style

Li, X., Meng, C., Zhou, H., Guo, Y., Xue, B., Yu, T., & Lu, Y. (2025). Generative Learning from Semantically Confused Label Distribution via Auto-Encoding Variational Bayes. Electronics, 14(13), 2736. https://doi.org/10.3390/electronics14132736

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop