Next Article in Journal
Super-Resolution of Sentinel-2 Satellite Images: A Comparison of Different Interpolation Methods for Spatial Knowledge Extraction
Previous Article in Journal
Multivariate CO2 Emissions Forecasting Using Deep Neural Network Architectures
error_outline You can access the new MDPI.com website here. Explore and share your feedback with us.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classifying with the Fine Structure of Distributions: Leveraging Distributional Information for Robust and Plausible Naïve Bayes

1
Department of Mathematics and Computer Science, University of Marburg, Hans-Meerweinstraße 6, 35039 Marburg, Germany
2
IAP-GmbH Intelligent Analytics Projects, In Den Birken 10a, 29352 Adelheidsdorf, Germany
3
Department of Hematology, Oncology and Immunology, Philipps University Marburg, University Hospital Giessen and Marburg, 35043 Marburg, Germany
*
Author to whom correspondence should be addressed.
Mach. Learn. Knowl. Extr. 2026, 8(1), 13; https://doi.org/10.3390/make8010013
Submission received: 25 September 2025 / Revised: 11 December 2025 / Accepted: 22 December 2025 / Published: 5 January 2026
(This article belongs to the Section Learning)

Abstract

In machine learning, the Bayes classifier represents the theoretical optimum for minimizing classification errors. Since estimating high-dimensional probability densities is impractical, simplified approximations such as naïve Bayes and k-nearest neighbor are widely used as baseline classifiers. Despite their simplicity, these methods require design choices—such as the distance measures in kNN, or the feature independence in naïve Bayes. In particular, naïve Bayes relies on implicit assumptions by using Gaussian mixtures or univariate kernel density estimators. Such design choices, however, often fail to capture heterogeneous distributional structures across features. We propose a flexible naïve Bayes classifier that leverages Pareto Density Estimation (PDE), a parameter-free, non-parametric approach shown to outperform standard kernel methods in exploratory statistics. PDE avoids prior distributional assumptions and supports interpretability through visualization of class-conditional likelihoods. In addition, we address a recently described pitfall of Bayes’ theorem: the misclassification of observations with low evidence. Building on the concept of plausible Bayes, we introduce a safeguard to handle uncertain cases more reliably. While not aiming to surpass state-of-the-art classifiers, our results show that PDE-flexible naïve Bayes with uncertainty handling provides a robust, scalable, and interpretable baseline that can be applied across diverse data scenarios.

1. Introduction

In supervised classification the primary objective is to learn a decision rule that minimizes classification error. The Bayes classifier is the theoretical optimum for this criterion when the true class-conditional distributions are known [1,2,3]. In practice, however, estimating a nonparametric, potentially high-dimensional joint density is often infeasible [4,5] (e.g., due to lack of samples since the amount of data required grows exponentially) and building a fully parameterized model to represent the joint distribution can be impractical, because real distributions are too complex to model with simple parametric forms, and high-dimensional feature spaces require the estimation of too many parameters from too little data, leading to unstable or inaccurate models [6].
Consequently, a variety of classifiers have been developed to address classification challenges in practical scenarios, and a wide range of pragmatic approximations has been developed. One class of methods is nonparametric instance-based learning: k-nearest neighbors (kNN) is conceptually simple and, with a suitable choice of k, is a consistent approximation to the Bayes rule as the sample size grows [7]. Another popular approach is the naïve Bayes classifier, which drastically reduces estimation complexity by assuming feature independence and often performs well despite this strong simplification [6,8]. Both kNN and naïve Bayes offer fast and straightforward alternatives to the more intricate Bayes classifier, though they may not always achieve the same level of performance.
Therefore, the naïve Bayes and k-nearest neighbor (kNN) classifiers are commonly used as baseline algorithms for classification. However, in the case of naïve Bayes, the user must determine the most suitable strategy for model fitting based on the specific use case. Available strategies include fitting a mixture of Gaussian distributions or employing a non-parametric approach using kernel density estimation. However, these methods usually do not account for varying distributions across different features (e.g., see Appendix C for varying feature distributions). Moreover, analyzing the distributions of all features to select the optimal strategy for a naïve Bayes classifier is a complex task. Similarly, for kNN, the user must choose an appropriate distance measure, which poses a similar complex decision [9] as well as the number k of nearest neighbors.
Although this feature independence is quite constraining, naïve Bayes classifiers have been reported to achieve high performance in many cases [10,11,12]. Interestingly, it has been observed that a naïve Bayes classifier might perform well even when dependencies exist among features [12], though correlation does not necessarily imply dependence [13]. The impact of feature correlation on performance depends on both the degree of correlation and the specific measure used to evaluate it. Such a performance-dependent relation can be effectively demonstrated by applying multiple correlation measures (see Table 1). Likewise, the choice of class assignments for cases within a dataset also influences performance, which can be illustrated using a correlated dataset evaluated on two different classifications.
A pitfall in Bayesian theory occurs if observations with low evidence are assigned to a class with a higher likelihood than another, without considering that the probabilities in a probability density with higher variance decay more slowly than in one with lower variance, creating a situation wherein a class assignment chooses a distant class with high variance over a closer class with smaller variance [14]. An example of this would be the size of humans according to gender, where the female (Gaussian) distribution has higher variance than the male (Gaussian) distribution, and the female mean is left of that of males. In that case, the classical Bayes theorem would assign a giant’s size to the class of females, since the likelihood of the males’ distribution would decay faster than that of the females. Instead, the closest mode could be chosen if facing observations with low evidence (the giant size). As a consequence, the non-interpretable choice is to classify all giants as female. Figure 1 outlines this situation for univariate on the left side and sketches the influence of the flexible approach of smoothed Pareto density estimation (PDE) in combination with plausibility correction.
We are proposing a new methodology for the naïve Bayes classifier, overcoming these challenges and improving performance. Our approach does not make any prior assumption about density distribution, creating an algorithm free of assumptions about the data distribution. Furthermore, we dispose of any parameters to be optimized [15]. We use a model of density estimation based on information theory defined by a data-driven kernel radius without intrinsic assumptions about the data, which empirically outperforms typical density estimation approaches [16]. The main contributions of our work are as follows:
  • Solution to the above-mentioned pitfall of the Bayes theorem within a Naïve Bayes classifier framework.
  • Empirical benchmark showing a robust classification performance of the plausible naïve Bayes classifier using the Pareto Density Estimation (PDE).
  • Visualization of class-conditional likelihoods and posteriors to support model interpretability.
The aim of our work is to provide a classifier that can serve as a baseline. Hence, we will show that the performance of naïve Bayes is hard to associate with dependency measures, and that the design decisions are fluid across the degree of correlation and the choice of the correlation measure. We developed an open-access R package (0.2.8) on CRAN https://CRAN.R-project.org/package=PDEnaiveBayes, accessed on 17 November 2025).

2. Materials and Methods

Let a classification G = C 1 C k be a partition of a Dataset D consisting of N = |D| data points into k∈ℕ non-empty, disjoint subsets (classes) [17]. Each class C i D contains a subset of datapoints x 1 , . . , x l , and each datapoint is assigned is a class label c i ∈{1,…,k} via the hypothesis function ℎ: D→{1,…,k}. The task of a classifier is to learn the mapping function h given training data and labels c i .

2.1. Bayes Classification

Assume a set of continuous input variables X = { x 1 , , x n } , with vector x R d . In Bayesian classification, the prior is the initial belief (“knowledge”) about class membership.
The Bayesian classifier picks the class whose posterior p ( C i   | x ) has the highest probability
p ( C i   | x ) > p ( C j   | x ) j   i
The posterior probability captures what is known about C i , now that we have incorporated x . The posterior probability is obtained with the Bayes Theorem
p C i x = p x C i p ( C i ) i = 1 k p x C i p C i
Here p x C i is the conditional probability of x given class C i called class likelihood and the denominator is the evidence. Let H be the hypothesis space and h i being a MAP hypothesis, D R d the (not necessarily i.i.d.) data set. Let H = { h 1 , , h M } be the hypothesis space with h i being a MAP hypothesis, D R d the (not necessarily i.i.d.) data set and let f :   D > G be a classifier that is a function. a Bayes optimal classifier [18] is defined by
c j = f o p t ( x j ) = argmax C i h i H p C i | h i ( x j ) × p h i ( x j ) D
Equation (3) yields the optimal posterior probability to classify a data point in the sense that it minimizes the average probability error [6].
The general approach to Bayesian modeling is to set up the full probability model, i.e., the joint distribution of all entities in accordance with all that is known about the problem [19]. In practice, given sufficient knowledge about the problem, distributions can be assumed for the x and priors with hyperparameters θ that are estimated.
For the goal of this work, of using a Bayesian classifier to measure a baseline of performance, we make the assumption that the classification of a data point can be computed by the marginal class likelihoods p l with
c j = argmax C i   p ( C i ) l = 1 d p l x l C i
which is called the naive assumption because it assumes i.d.d. given the classification set G. In some domains, the performance of a naive Bayes classifier has been shown to be comparable to neural networks and decision trees [18]. We will show in the results that for typical datasets, even if this assumption does not hold true, an acceptable baseline performance can be computed.
Our objective is to exploit the empirical information in the training set to derive, in a fully data-driven manner and under as few assumptions as possible, an estimate of each feature’s distribution within every class C i , and then to evaluate classifier performance on unseen test samples. To this end, we adopt a frequentist strategy: we estimate the class-conditional densities directly from the observed data, and we compute the class priors as the relative frequencies of each C i in the training sample—assuming that the sample faithfully represents the underlying population. Because our decision rule is given by Equation (4), which compares unnormalized scores across classes, the global evidence term (the denominator in Bayes’ theorem) may be treated as a constant and hence omitted.
The challenge is to estimate the likelihood p ( x | C i ) , given the samples of the training data. However, by using the naive assumption, the challenge of estimating the d -dimensional density is simplified to estimating d one-dimensional densities. In prior works, the first naive Bayes classifier using density estimation was introduced as flexible Bayes [8]. This work improves significantly on the task of parameter-free univariate density estimation without making implicit assumptions about the data (c.f. Ref. [16]).

2.2. Density Estimation

The task of density estimation can be achieved in two general ways. First, through parametric statistics, meaning the fitting of parametrized functions as a superposition, where the fitting can be performed according to some quality measure or some sort of statistical testing. The drawback here is the rich possibility of available assumptions. A default approach could be the use of Gaussian distributions fitted to the data [20]. Second, nonparametric statistics, meaning local parametrized approximations using neighboring data around each available data point. It varies in its use of fixed or variable kernels (global or local estimated radius). The drawback of this approach is the complexity of tuning bandwidths for optimal kernel estimation, which is a computationally hard task [21]. A default approach to solve this problem can be a Gaussian kernel, where the bandwidth is selected according to Silverman’s rule of thumb [5].
In this work, we want to solve the density estimation with the data-driven approach called Pareto Density Estimation (PDE), taken from [15]. In this way, any prior model assumption is dropped. The PDE is a nonparametric estimation method with a variable kernel radius. For each available datapoint in the dataset, the PDE estimates a radius to estimate the density at this point. It uses a rule from information theory: The PDE maximizes the information content (reward) while minimizing the hypersphere radius (effort). Since this rule is historically known as the Pareto rule (or 80-20-rule, reward/effort-trade-off), the method is called the Pareto Density Estimation.

2.3. Pareto Density Estimation

Let a subset S D of data points have a relative size of p ~   = | S | | D | . If there is an equal probability that an arbitrary point x is observed, then its content is I ( p ~ ) = p ~   l n ( p ~   ) . The optimal set size is the Euclidean distance of S from the ideal point, which is the empty set with 100% information [15]. The unrealized potential is U R P ( S ) =   p ~   2 + ( 1 + p ~   l n ( p ~   ) ) 2 . Minimizing the U R P ( S ) yields the optimal set size p u ~ = 20.13 % . This set retrieves 88% of the maximum information [15]. For the purpose of univariate density estimation, computation of the univariate Euclidean distance is required. Under the MMI assumption, the squared Euclidean distances are χ-quadrat-distributed [22] leading to
R = 1 2 c d χ d 2 p u ~  
where c d ( χ d 2 ) ( p u ~ ) is the Chi-square cumulative distribution function for d degrees of freedom [15]. The pareto Radius is approximated by R p ~ 18 % for d = 1 .
The PDE is an adaptive technique for estimating the density at a datapoint following the Pareto rule (80-20-rule). The PDE maximizes the information content using a minimum neighborhood size. It is well-suited to estimating univariate feature distributions from sufficiently large samples and to highlighting non-normal characteristics (multimodality, skew, clipped ranges) that standard, default-parameter density estimators often miss. An empirical study, Ref. [16] demonstrated that PDE visualizations (mirrored-density plots, short MD plots) more faithfully revealed fine-scale distributional features than standard visualization defaults (histograms, violin plots, and bean plots) when evaluated across multiple density estimation approaches. Accordingly, PDE frequently yields superior empirical performance relative to commonly used, default-parameter density estimators [16].
Given the Pareto radius R, the raw PDE can be estimated as a discrete function solely at given kernel points x ^ with
f x ^ = 1 A x i   1 x i x ^ R  
where 1 is an indicator function that is 1 if the condition holds and 0 otherwise, A normalizes f x ^ Δ x ^ = 1 , and f x ^ a quantized function that is proportional to the number of samples falling within the interval [ x ^ −R, x ^ +R].
For the mirrored-density (MD) plots used in [16], we previously applied piecewise-linear interpolation of the discrete PDE f x ^ to fill gaps between kernel grid points and obtain a continuous visual approximation of a feature’s pdf f x . While linear interpolation produces a visually faithful representation and is adequate for exploratory plots, as it preserves continuity and is trivial to compute, it inherits the small-scale irregularities of the underlying discrete PDE estimate. If linearly interpolated densities are used directly as class-conditional likelihoods in Bayes’ theorem, their high-frequency noise propagates into the posteriors and can produce unstable or incorrect classifications.
Accordingly, in this work, we proceed as follows. After obtaining the raw and discrete conditional PDE f x ^ | C i l , we replace linear interpolation with several smoothing steps in the next section that produce a class likelihood p l x l C i , which (i) preserves the genuine distributional structure (modes, skewness, tails) and (ii) suppresses spurious high-frequency fluctuations that would destabilize posterior estimates.

2.4. Smoothed Pareto Density Estimation

Although the PDE of class likelihoods captures the overall shape of the distribution, it can be somewhat rough or piecewise constant due to its uniform kernel and finite sampling. This is disadvantageous, as irregular or noisy estimates yield fluctuating posteriors and unstable decisions. Therefore, smoothing the class likelihoods acts as a form of regularization: it suppresses sample-level noise and prevents high-frequency fluctuations like erratic spikes or dips in the likelihoods that lead to brittle posterior assignments. Balancing the fidelity to the PDE’s “true” features with the removal of artificial high-frequency components results in more accurate approximations of the true underlying distributions, leading to more reliable posterior estimates because they are less influenced by random sample noise.
For smoothing, we exploit the insight that the kernel estimate is a convolution of the data with the kernel by using fast Fourier transforms [5] p. 61. To produce a smooth, continuous density estimate, we convolve the PDE output with a Gaussian kernel using the Pareto radius as the bandwidth. Hence, the Gaussian smoothing kernel is defined as
K x = 1 R 2 π e x p ( x 2 2 R 2 )
where R is the Pareto radius. We use the Fast Fourier Transform (FFT), leveraging the convolution theorem, in order to implement this convolution efficiently, as follows.
First, we evaluate the Gaussian kernel on the same grid x ^ j j = 1 m as the PDE in which m is the number of grid points. Let Δ x ^ be the grid spacing, then the Gaussian kernel vector
k j = Δ x ^ × K x j
yields a normalized kernel vector aligned with the PDE grid, and we use the mean of adjacent differences to avoid numerical instabilities in spacing.
To perform a linear convolution without wrap-around artifacts, we zero-pad both the density f x ^ and kernel vectors k j before FFT [23] (part1, p. 260ff). We choose the padding length L 2 log 2 ( 2 m 1 ) as the next power of two. This length ensures that circular convolution via FFT corresponds to the linear convolution of the original sequences, and using a power of two leverages FFT efficiency [24]. We create padded vectors f p a d (the f x ^ with zero-pad) and k p a d ( k j with zeros zero-pad), each of length L. This padding avoids overlap of the signal with itself during convolution.
Next, we compute the FFT of both padded vectors, multiply them element-wise in the frequency domain, and then apply the inverse FFT. By the convolution theorem, the inverse FFT of the product
f p a d   k p a d = I F T ( F T ( f p a d ) F T ( k p a d ) ) / L
yields the linear convolution on the padded length. We divide by L in Equation (9) when taking the inverse FFT, as per the normalization convention.
The central m elements correspond to the convolved density over the original grid. This middle segment is the smoothed density vector f ~ x ^ , aligned with the original kernel grid. The approach is motivated by the idea from [25].
Finally, the montone Hermite spline approximation of f ~ x ^ [26] yields the likelihood function p ( x l | C i ) . and allows for a functional, fast computation of new points.
In sum, the empirical class PDE f x ^ l in dimension l can be rough due to the data being noisy. If used for a visualization task, the roughness would be inconsequential [16]. In order not to influence the posteriors through data noise, we propose as a solution a combination of filtering by convolution (c.f. Ref. [27]) and monotonous spline approximation (c.f. Ref. [28]) yielding p l ( x l | C i ) .

2.5. Plausible Naïve Bayes Classification

Ref. [14] showed that misclassification can occur when only low evidence is used in the Bayes’ theorem, i.e., the cases lie below a certain threshold ε. They define cases below ε as uncertain and provide two solutions [14]: reasonable Bayes (i.e., suspending a decision) and plausible Bayes (a correction of Equation (4)). To derive ε, they propose the use of the computed ABC analysis [29]. The algorithm allows users to compute precise thresholds that partition a dataset into interpretable subsets.
Closely related to the Lorenz curve, the ABC curve graphically represents the cumulative distribution function. Using this curve, the algorithm determines optimal cutoffs by leveraging the distributional properties of the data. Positive-valued data are divided into three disjoint subsets: A (the most profitable or largest values, representing the ‘important few’), B (values where yield matches effort), and C (the least profitable or smallest values, representing the ‘trivial many’).
Let { x 1 ,   x n } be a set of n observations, which, for the purpose of defining the plausible Naïve Bayes likelihoods in dimension l , are indexed in non-decreasing order in their respective dimensions, i.e., x 1 l , , x n l . Let s i = k = 1 i x k l ; the L ( P ) is defined by [30] as
L p i = 0 f o r   P i = 0 s i s n f o r   p i = i n
For all other p in [0, 1] with p ≠ pi, L(p) is calculated using a linear, spline, or other suitable interpolation on L ( p i ) [31].
Let L(p) be the Lorenz curve in Equation (11), then the ABC curve is formally defined as
A B C p = 1 L 1 p
Then the break-even point satisfies d ( A B C ( p ) ) d p = B x = 1 and the submarginal point B C p , B C A B C is located by minimizing the distance from the ABC curve to the maximal-yield point at (1,1) after passing the break-even point with B C p = argmin p > B x 1 A B C ( p ) . The break-even point yields the BC limit, and, hence, the threshold ε with
ε = B C A B C = A B C ( B C p )
Equation (12) defines the BC Limit.
Inspired by this idea, we reformulate the computation of epsilon from posterior to joint likelihood, as follows. An observation x   is considered uncertain in feature l whenever the joint likelihood of every class falls below the confidence threshold ε, i.e.,
Γ l ( x l ) = l = 1 k p l x l C i < ɛ
where p l denotes the marginal of the distribution in dimension l . We will apply the threshold ε to identify low-evidence regions where the plausibility correction (Equation (15)) may be considered. Such uncertain cases might be classified against human intuition to a class with a probability density center quite far away, despite closer available class centers [14]. Then, a “reasonable” assignment might be to assign the case to the class whose probability centroid is closest, which can be calculated using Voronoi cells for d > 1. For a one-dimensional case, they allow the closest mode to be determined for classification.
We estimate the univariate location of each class’s likelihood mode on a per-feature basis using the half-sample mode [32]. For small sample sizes (n < 100), we use the L 0 estimator recommended by [33]. As a safeguard mechanism, estimated modes are only considered for resolving uncertain cases if they are well-separated, i.e., have a distance from each other of at least the 10th percentile within the training data
Δ m i , j = m C i , x l m C j , x l > p ~ 10 % x l ,   i j , l = 1 d
This mechanism is motivated by the potential presence of inaccurately estimated modes or overlapping (non-separable) classes.
When the class likelihood p l x l C i for an observation x is uncertain in feature l (Equation (13) holds true), and there is a class i whose mode is well-separated from the highest-likelihood class (Equation (14)), we perform a conservative, local two-class correction of that feature class likelihoods as follows:
Let i m a x = argmax i   p l x l C i be the index of the uncorrected class likelihood with the largest value and i the index of the class likelihood with the closest mode to x for which Equation (14) holds true for i = i m a x and j = i .
Then we update the two involved class likelihoods by replacing the values of the class likelihood i with the (former) top class likelihood, i m a x . In addition, the operation subtracts δ from the prior top likelihood i m a x and adds δ to the runner-up, i , and the relative advantage of the runner-up versus the former top increases by 2δ, which is sufficient to resolve many marginal posterior ties or implausibilities (as shown in the example in Figure 1) while remaining conservative. All other class likelihoods for this feature remain unchanged in Equation (15):
p l , c o r r x l C i = p l x l C i δ   , i = i m a x , i f   Γ l ( x l ) < ɛ a n d   Δ m i m a x , i > p ~ 10 % p l x l C i m a x + δ , i = i   , i f   Γ l ( x l ) < ɛ a n d   Δ m i m a x , i > p ~ 10 % p l x l C i o t h e r w i s e  
Equation (4) is then used with the locally corrected class likelihoods p l , c o r r x l C i . Note that as δ 0 , the correction vanishes and the method reduces to the reasonable-Bayes rule; the transfer introduces a conservative “plausible-Bayes” adjustment, with δ controlling the strength of the plausibility correction.

2.6. Practical Considerations

In order to avoid numerical overflow, Equation (4), either uncorrected or the with the locally corrected class likelihoods p l , c o r r x l C i can be computed in log scale
c j = argmax C i log p C i + l o g ( i = 1 d p l x l C i )
to select the label of the class C j that with the highest probability.
In practice, before this function can be computed, it must be determined if there are enough samples to yield a proper PDE. Based on empirical benchmarks [16], if there are more than 50 samples and at least 12 uniquely defined samples, then f x ^ l can be estimated, otherwise the estimations might deviate. In case there are too few samples, the density estimation defaults to simple histogram binning with bin width defined by Scott‘s rule [34].
Let τ be a small constant, then, to ensure numerical stability in Equation 16, the likelihoods p l C i are clipped to the range of [ τ , 1 τ ] . The reason is that density after smoothing may result in values slightly below zero due to the convolution (c.f. [5]). In addition, density estimation can have spikes above 1. Moreover, we ensure numerical safety if we clip the corrected likelihoods to be non-negative after applying Equation (15).
There is also a possibility that Equation (4) may not allow a decision as two or more posteriors equal each other after the priors are considered. In such a case, the class assignment is randomly decided.
Equation (15)’s foundation is the assumption that modes can be estimated correctly in the data for each class, which could fail in practice. As a safeguard, we provide the following option for the user. We compute the classification assignments C j , I G ( I ) as defined in Equation (4), transformed according to Equation (16), and likewise without correction C j , I I G ( I I ) , using the training data.
Assuming the priors are not excessively imbalanced, we evaluate the Shannon entropy of each classification result C j and choose the configuration that yields the highest entropy. The Shannon entropy H of G with priors p ( C j ) for i = 1 , , k is defined as
H G = 1 Q i = 1 k p ( C j )   l o g ( p ( C j ) )
with the normalization factor Q = 1 k l o g ( 1 k ) .
A higher entropy indicates a potentially more informative classification.
Finally, due to the assumption leading to Equation (4) and subsequent equations, we provide a scalable multicore implementation of the plausible Bayes classifier by computing every feature separately, proceeding as follows: For each feature dimension   l we estimate a single Pareto radius R l independent of class, as defined in Equation (5), rather than separately for each feature–class combination. Empirical evaluations indicate that this approximation is sufficiently accurate for practical applications. For each class C i in each dimension l we compute the class-wise PDE on an evenly spaced kernel grid x i ^ i m to compute the raw conditional PDE f x ^ | C i l covering the range of the data. Then, we compute the smoothed likelihood functions p l ( x l | C i ) per feature dimension l from the discrete conditional PDE f x ^ | C i l . We call this approach the Plausible Pareto Density Estimation-based flexible Naïve Bayes classifier (PDENB).

2.7. Interpretability of PDENB

A one-dimensional density estimation is required for the Naïve Bayes Classifier to compute the class-conditional likelihood of a feature as one of three parts of the Bayes theorem, yielding the final Posterior. This class-conditional likelihood allows a two-dimensional visualization as a line plot for a single feature. The plot gives insight into the class-wise distribution of the feature. Scaling the likelihoods with the weight of the prior obtained from the frequentist approach [6] yields correct probabilistic proportions between the class-conditional likelihoods, which can be represented by different colors. Rotating the plots by 90 degrees and mirroring them in a similar way to that used for violin and mirrored density plots [16] allows a lineup of the likelihoods to be created for multiple features at once. Such a visualization allows for interpretation based on the class-wise distribution of features. Most often, the colored class conditional likelihoods are overlapping and non-separable by solely one feature. However, in case of a high performing naïve Bayes classifier, overlaps do not indicate non-separability, but rather, first of all, a class tendency for each feature, and second, the existence of certain combinations and the disqualification of other combinations in question. These visual implications can be the starting point for a domain expert to find relations between features and classes, resulting in explanations.
Additionally, we provide a visualization of one class versus all decision boundaries in 2D, as follows. Given a two-dimensional slice S D R d , the Voronoi cell associated with a point g S is the region of the plane consisting of all points that are closer to g than to any other point v in the slice, defined by
V g = y R 2   y g   y v   g   v }
That is, V g contains all points such that the Euclidean distance from any y to g is less than or equal to the distance to any other v S . Each Voronoi cell V g is binned according to the binned posterior probability P C i x , thereby mapping regions of the plane to their inferred class likelihoods by colors. The binning can be either performed in equal sizes, using Scotts rule as bin width [34], or less efficiently, by the DDCAL clustering algorithm [35]. The user can visualize the set of slices of interest. This approach is motivated by human pattern recognition and the subsequent classification of diseases from identified patterns in two-dimensional slices of data [36], which is apparently sufficient for a large variety of multivariate data distributions.

2.8. Benchmark Datasets and Conventional Naïve Bayes Algorithms

For the benchmark we selected 14 datasets: 13 from the UCI repository and a 14th dataset (“Cell populations”) containing manually identified cell populations [37]. Full dataset descriptions of the UCI datasets, attribute definitions, and links to original sources are available on the repository pages for each dataset [38], for example, Iris: https://archive.ics.uci.edu/dataset/53/iris, accessed on the 15 December 2025). The Cell populations dataset is an extended version of the data used by [37]; the set of populations provided here is larger than in that publication because the authors labeled populations at finer granularity. A detailed description of the cell populations dataset is given in [37].
The datasets were preprocessed prior to analysis using the methods of rotation [39,40] taking into consideration Refs. [41,42], as implemented in “ProjectionBasedClustering”, available on CRAN [43]. In it should be noted that although the signed log transformation for better interpretability was applied to the cell populations, no Euclidean optimized variance scaling was used [44]. Thereafter, correlations of features were computed. Important properties, meta information, and correlations of the processed datasets are presented in Table 1. The measure of correlation depends on the choice of algorithm. Here, the Pearson, Spearman’s rank, Kendall’s rank, and the Xi correlation coefficient [45] are summarized using the minimum and maximum values to characterize the correlations of the datasets. The values of the Pearson and Spearman correlation coefficients tend to be higher than the Xi correlation coefficient by around 0.3. Similarly, Kendall’s Tau values tend to be higher than the Xi correlation coefficients, but not as high as the Pearson’s or Spearman’s correlation coefficients. Some important attributes, such as the number of classes, the distribution of cases per class to judge class imbalance, and various dependency measures related to the independent feature assumption of the naïve Bayes classifier, are presented in Table 1.
Performance is evaluated for the Plausible Pareto Density Estimation-based flexible Naïve Bayes classifier (PDENB) in comparison to a Gaussian naïve Bayes classifier (GNB) and a nonparametric naïve Bayes classifier (NPNB) from the R package “naivebayes”, available on CRAN [20]); a fast implementation of k-nearest neighbor classifier (kNN as 7NN) from the R package “FNN”, available on CRAN [46]); a Gaussian naïve Bayes from the python package “sklearn” [47] (PyGNB); a Gaussian (klaRGNB) and non-parametric naïve Bayes (klaRNPNB) approach from the R package “klaR”, available on CRAN [48]; and last, but not least, a Gaussian naïve Bayes from the R package “e1071”, available on CRAN [49] (e1071GNB). The algorithms for the naïve Bayes methods were applied in their default settings differencing between “Gaussian” and “nonparametric” versions, while the parameter k for the kNN classifier is set to 7.

3. Results

The first subsection presents the classification performance in detail. Section 2 presents visualizations that support model interpretability, and Section 3 provides an application.

3.1. Classification Performance

Table 1 presents the benchmark across 14 datasets. Following the work of [50,51], small datasets with up to 10.000 samples are evaluated with 100-times repeated hold-out set using the 80–20 rule. Datasets with higher sample sizes are evaluated once with a hold-out set (80–20 rule) and 100 mean samples are obtained with a resampling technique. The performance in each trial is evaluated with the Matthews correlation coefficient (MCC) [52,53]. The results are summarized in tables by means of the 100 trials evaluated in Table 2. The distributions are shown using Mirrored Density plots (MD plots) in Appendix A. “NA” values indicate a failure of the classification algorithm. For the case of MiceProtein, several classifiers failed due to missing values in the dataset. The remaining cases of “NA” values can be observed for klaRGNB: in the datasets Dermatology, Spam, and Covertype, errors due to the estimation of variance are reported. Furthermore, the results show that high correlation values are not necessarily an indicator of low performance for naïve Bayes classifiers (compare Table 1 to Table 2).
We find that no single algorithm consistently outperforms all others, in line with the no-free-lunch theorem [54]. However, it is still possible to rank algorithms based on their mean performance, considering their best-performing variants across trials, while allowing for ties. To this end, we apply the permutation test [55] pairwise to each evaluation (see Table A1 in the Appendix). This statistical test determines whether performance differences are significant or not, enabling rankings with possible ties. From these results, we obtain a ranking of classifiers for all datasets (see Table A2). Finally, we compute the mean of these ranks to derive an overall score for each algorithm, presented in Table 3.

3.2. Interpretable Naïve Bayes Classifier

One advantage of the PDE-based, flexible naïve Bayes classifier is that it relies on one-dimensional density estimates, which we use to visualize the class-dependent fine structure of distributions for each feature. Inspecting all class-conditional likelihoods for a feature at once can reveal interesting relations or patterns when the number of dimensions allows such an overview; otherwise, a targeted feature selection is required. Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8 show class-conditional likelihoods for four representative datasets: Satellite, Iris, Penguin, and the Cell populations dataset on the training data for one arbitrary cross-validation trial.
Figure 2A–D present the features of the annotated biological populations in the Flowcytometry dataset called cell populations. Normally, cell populations are manually distinguished in sequential two-dimensional dot plots. In Figure 2A, CD45 is a pan-leukocyte antigen whose fluorescence intensity separates broad white blood cell compartments (lymphocytes, monocytes, granulocytes) along with side scatter, and helps distinguish hematopoietic from non-hematopoietic events. Within the lymphocytes, CD19 is a B-lineage surface marker used to identify and quantify B cells. In Figure 2B, CD3 separates T-cells, and NK-cells can be separated because they are CD3-negative and positive for CD16 and/or CD56. FS_INT (forward-scatter integral) is a proxy for cell size, and SS_INT (side scatter integral) for granularity in Figure 2D.
Figure 2B,C depict two features, labeled FL4_INT and FL8_INT. These channels correspond to detectors that do not match any used fluorochrome and therefore primarily record detector noise and light spillover from other detectors, see [56] for details. It is clearly visible that no class is distinguishable in FL4_INT in comparison to “CD16_FITC” and “CD4_PC7”. In FL8_INT, we see a spillover effect from another detector (possibly “CD14_APC700”) to FL8_INT, allowing us to distinguish the red class. Without medical knowledge, FL4_INT would be disregarded based on Figure 2C, and FL8_INT would be questioned because it only distinguishes the green class (Monocytes), with negative log light measures based on Figure 2B, which could be a light spillover of CD14, presented in Figure 2C. Based on Figure 2D, for the classification task, it seems that FS_Int and FS_PEAK contain the same information with respect to class likelihoods, which aligns with domain knowledge.
Figure 3 presents class-conditional likelihoods for the Satellite dataset and motivates the need for an assumption-free distribution estimator. The plot illustrates a variety of distributional shapes across classes and features: for example, the “mixture” class in feature 26 is skewed, the “grey soil” class exhibits increased kurtosis, and the “cotton crop” class shows long tails in several features.
Figure 4 shows the well-known Iris dataset, where clear per-class tendencies are evident; the strong performance of naïve Bayes in this example illustrates how informative one-dimensional patterns can be for classification.
Figure 5 displays the Penguin dataset: although class tendencies are visible, substantial class overlap is present. The good performance of naïve Bayes here implies that separability is achieved by combining the information from feature C2 with C4 after the ICA rotation is applied, and that Features C1 and C3 can be disregarded
Across the visualized datasets, PDENB achieves excellent performance (MCC ≥ 0.95). The per-feature PDE visualizations frequently reveal regions of class overlap at the single-feature level—not as a shortcoming, but as an exploratory strength—these plots make the limits of one-dimensional separation explicit and point to which features are complementary. By inspecting class-conditional shapes (modes, tails, skewness) across features we can identify feature combinations that produce clear separation in higher dimensions, guide feature selection, and may even generate meaningful explanations for the classifier’s decisions.
Figure 2. Figures (AD) show the class-conditional PDE likelihoods of dataset “Cell populations”. The colors depict different classes and are mapped to the cell populations. The range of values per data feature of the y-axis is presented after the transformation to the signed log scale as provided in the package “DataVisualizations” on CRAN [57]. The x-axis presents the range of class-conditional PDE likelihoods. The legend in Figure (A) presents the names of the cell population classes and also applies to the following subfigures (BD). Figure (A) shows the class-conditional PDE likelihoods for the first two features. Figure (B) shows the class-conditional PDE likelihoods for the next four features. Figure (C) shows the class-conditional PDE likelihoods for the next four features of the dataset “Cell populations”. Figure (D) shows the class-conditional PDE likelihoods for the last three features. For comparison, FS_TOF in a different range is shown again.
Figure 2. Figures (AD) show the class-conditional PDE likelihoods of dataset “Cell populations”. The colors depict different classes and are mapped to the cell populations. The range of values per data feature of the y-axis is presented after the transformation to the signed log scale as provided in the package “DataVisualizations” on CRAN [57]. The x-axis presents the range of class-conditional PDE likelihoods. The legend in Figure (A) presents the names of the cell population classes and also applies to the following subfigures (BD). Figure (A) shows the class-conditional PDE likelihoods for the first two features. Figure (B) shows the class-conditional PDE likelihoods for the next four features. Figure (C) shows the class-conditional PDE likelihoods for the next four features of the dataset “Cell populations”. Figure (D) shows the class-conditional PDE likelihoods for the last three features. For comparison, FS_TOF in a different range is shown again.
Make 08 00013 g002aMake 08 00013 g002bMake 08 00013 g002c
Figure 3. The figure visualizes the fine distribution structure of the class-conditional PDE likelihoods for a selected subset of features from the Satellite dataset, highlighting differences in skewness, tails, and multimodality across classes. The colors depict different classes as presented in the legend. Each feature contains varying non-classic distributions with different characteristics, such as long and fat tails, and multimodal and skewed distributions, which can be assessed visually.
Figure 3. The figure visualizes the fine distribution structure of the class-conditional PDE likelihoods for a selected subset of features from the Satellite dataset, highlighting differences in skewness, tails, and multimodality across classes. The colors depict different classes as presented in the legend. Each feature contains varying non-classic distributions with different characteristics, such as long and fat tails, and multimodal and skewed distributions, which can be assessed visually.
Make 08 00013 g003
Figure 4. The figure shows the class-conditional PDE likelihoods for the 4 features of the dataset “Iris”. The colors depict different classes as presented in the legend on the right.
Figure 4. The figure shows the class-conditional PDE likelihoods for the 4 features of the dataset “Iris”. The colors depict different classes as presented in the legend on the right.
Make 08 00013 g004
Figure 5. The figure shows the class-conditional PDE likelihoods for the 4 features of the dataset “Penguins”, obtained by an ICA transformation. The colors depict different classes as presented in the legend on the right. Features C1–C4 were derived by applying independent component analysis (ICA) to the original variables [58]. Hence, the resulting components are used as rotated features for downstream analysis [59].
Figure 5. The figure shows the class-conditional PDE likelihoods for the 4 features of the dataset “Penguins”, obtained by an ICA transformation. The colors depict different classes as presented in the legend on the right. Features C1–C4 were derived by applying independent component analysis (ICA) to the original variables [58]. Hence, the resulting components are used as rotated features for downstream analysis [59].
Make 08 00013 g005
Figure 6. The figure shows a customized 2D Voronoi tessellation based on the two ICA components C2 and C4 from the dataset Penguins. The posteriors for the three classes, Adelie, Gentoo, and Chinstrap, are highlighted from left to right. The color palette spans a continuous gradient from dark red through orange and yellow to white. Dark red corresponds to posterior values approaching one, with intermediate hues representing progressively lower posterior values, and white indicating posterior values equal to zero. A compact area consisting mainly of a dark red color can be detected, indicating a specific location of very high posterior values for each class in the relationship between the two features. See Figure 5 for comparison to class likelihoods.
Figure 6. The figure shows a customized 2D Voronoi tessellation based on the two ICA components C2 and C4 from the dataset Penguins. The posteriors for the three classes, Adelie, Gentoo, and Chinstrap, are highlighted from left to right. The color palette spans a continuous gradient from dark red through orange and yellow to white. Dark red corresponds to posterior values approaching one, with intermediate hues representing progressively lower posterior values, and white indicating posterior values equal to zero. A compact area consisting mainly of a dark red color can be detected, indicating a specific location of very high posterior values for each class in the relationship between the two features. See Figure 5 for comparison to class likelihoods.
Make 08 00013 g006
Figure 7. The figure shows a customized 2D Voronoi tessellation based on the two features CD16_FITC and CD14_APC700 from the Cell populations dataset. The color palette is identical to Figure 6, with dark red indicating high posterior values and white indicating zero. The posterior for the class of atypical monocytes 1 is highlighted. A compact area consisting mainly of a dark red color can be detected, indicating a specific location of high posterior values for this class in the relationship between the two features. Both features are typically used to detect atypical monocytes in Flow Cytometry.
Figure 7. The figure shows a customized 2D Voronoi tessellation based on the two features CD16_FITC and CD14_APC700 from the Cell populations dataset. The color palette is identical to Figure 6, with dark red indicating high posterior values and white indicating zero. The posterior for the class of atypical monocytes 1 is highlighted. A compact area consisting mainly of a dark red color can be detected, indicating a specific location of high posterior values for this class in the relationship between the two features. Both features are typically used to detect atypical monocytes in Flow Cytometry.
Make 08 00013 g007
Figure 8. The figure shows a customized 2D Voronoi tessellation based on the two features CD16_FITC and CD14_APC700 from the Cell populations dataset. The color palette is identical to Figure 6, with dark red indicating high posterior values and white indicating zero. The posterior for the class of classical monocytes is highlighted. The color palette is identical to Figure 6, with dark red indicating high posterior values and white indicating zero. A compact area consisting mainly of a dark red color can be detected, indicating a specific location of high posterior values for class 10 in the relationship between the two features. Both features are typically used to detect classical monocyte cells in Flow Cytometry.
Figure 8. The figure shows a customized 2D Voronoi tessellation based on the two features CD16_FITC and CD14_APC700 from the Cell populations dataset. The color palette is identical to Figure 6, with dark red indicating high posterior values and white indicating zero. The posterior for the class of classical monocytes is highlighted. The color palette is identical to Figure 6, with dark red indicating high posterior values and white indicating zero. A compact area consisting mainly of a dark red color can be detected, indicating a specific location of high posterior values for class 10 in the relationship between the two features. Both features are typically used to detect classical monocyte cells in Flow Cytometry.
Make 08 00013 g008
See Figure 6 for comparison to posteriors. Another take on visualizing patterns detected by the naïve Bayes classifier would be to visualize the posterior computed in high dimensions in a two-dimensional plot. An informative plot in two dimensions can be derived from a 2D scatter plot. Given the two-dimensional coordinates, a Voronoi tessellation can be used to partition the area defined by the coordinates, as is performed in Figure 6, Figure 7, Figure 8 and Figure 9 for the training data. This representation allows a coloration of decision areas based on the posterior. High posterior values are colored as dark red, zero values as white, and intermediate values as a gradient in between the range of these two colors, with yellow as the color for values in the middle. Within the 2D Voronoi visualization of the posterior decision boundaries, the recognition of a single compact area with high posterior values for a certain class suggests a decision pattern. For example, the hypothesis derived from Figure 8 is clearly visible in the high posterior values depending on the class in Figure 9.
Figure 7 and Figure 8 present the posteriors computed for the classes of atypical monocytes, classical monocytes, and B-cells on feature CD 14 vs. CD16. CD14 vs. CD16 is a standard plot for innate-cell phenotyping: it cleanly separates atypical monocyte subsets (classical CD14+ CD16- CD4+ and atypical CD14+/- CD16+ CD4 and CD14+/- CD16- CD4+. However, CD16 is not specific to classical monocytes (it is also found on neutrophils and some monocytes), so classical monocytes should be confirmed with CD56 and the exclusion of CD3/CD14. B cells are not identifiable on CD14 vs. CD16 in Figure 9, because they are typically CD14CD16 and overlap with many other CD14CD16 populations; a positive B-cell marker (e.g., CD19 or CD20) is required for reliable detection.
In sum, the proposed visualization approaches is meaningful because it reveals decision boundaries in the two-dimensional feature space that are not only predictive (i.e., yield good classification performance) but also explanatory, making the underlying decision patterns interpretable.

3.3. A Baseline for the Distinction of Blood vs. Bone Marrow in Biological Population Frequencies

Distinguishing bone marrow (BM) from peripheral blood (pB) is a routine but clinically important task in diagnostic hematology. BM and pB differ in their cellular composition and in the relative frequencies of their hematopoietic progenitors, immature myeloid and lymphoid populations, and other subpopulations; these differences are routinely exploited by clinicians in flow-cytometric two-dimensional scatter plots to identify diagnostically relevant populations. Accurate separation of BM from pB is also important for downstream tasks such as assessing Minimal Residual Disease (MRD), because inadvertent dilution of BM aspirates with peripheral blood can bias clinical interpretation. For background on aspiration and dilution effects, see [60].
We used the Dresden cohort from the public Flow Cytometry collection [61]. The Dresden data comprise N = 44 sample files measured on a BD FACSCanto II instrument: 22 bone marrow and 22 peripheral blood samples. Each sample consists of a high-event flow cytometry file comprising approximately 130,000–880,000 single-cell events. For each event, 10 parameters were recorded, including forward scatter and side scatter for cell size and granularity, as well as eight fluorescence channels corresponding to the antigens CD34, CD13, CD7, CD56, CD33, CD117, HLA-DR, and CD45. All files are anonymized, instrument-compensated, and log-scaled into the range [0, 6] for analysis. We follow the patient-level evaluation used in the dataset’s original benchmarking by identifying cell population frequencies through ALPODS [62]. This yields one label per sample file (BM vs. pB) and permits direct comparison with previously reported results [62]. Contrary to Ref. [62], we do not identify meaningful cell populations with ALPODS but use all generated cell population frequencies as a baseline.
Using the PDENB and 80/20% cross validation of 100 trials, PDENB achieves a classification accuracy of 99.3 ± 0.03% accuracy (98.8+-0.5 MCC) on the Dresden dataset (sample-level decision). This outperforms the previously reported accuracy of 96.8 ± 0.09% for the ALPODS explainable-AI pipeline on the same dataset. Figure 10 presents selected posterior decision boundaries in 2D. It is visible that low cell population frequencies of population C0024 (defined by CD45 < 2.0735 and CD13 ≥ 3.0485 and CD34 ≥ 3.4125), C0013 (CD45 < 2.0735 and CD13 ≥ 2.4105 and CD13 < 2.8655 and CD7 < 3.377 and FS < 5.464 and CD34 < 5.871 and CD33 ≥ 3.37451.91 and CD117 1.91 and CD117 < 3.4675 and CD117 ≥ 3.3245 & HLA_DR ≥ 2.2005) and C0014 (CD45 < 2.0735 and CD13 ≥ 2.4105 and CD13 < 2.8655 and CD7 < 3.377 and FS < 5.464 and CD34 < 5.871 & CD33 ≥ 3.3745 and CD56 ≥ 1.91 & CD117 ≥ 3.4675) depict peripheral blood and high cell population frequencies of bone marrow.
The improvement demonstrates that robust, nonparametric, PDE-based likelihood estimation combined with conservative smoothing and plausibility correction can yield a stronger baseline for this task, and that highly accurate classification is achievable for high-event flow-cytometry samples without bespoke gating rules.

4. Discussion

We introduced PDENB, a Pareto Density-based Plausible Naïve Bayes classifier that combines assumption-free, neighborhood-based density estimation with smoothing and visualization tools to produce robust, interpretable classification. Our empirical benchmark across 14 datasets and its dedicated application to multicolor flow cytometry demonstrate several consistent advantages of this approach.
First, PDENB is competitive with—and frequently superior to—established Naïve Bayes implementations and non-parametric variants. Using repeated 80/20 hold-out evaluations (or resampling for very large datasets) and Matthews Correlation Coefficient (MCC) as the performance measure, PDENB attains top average ranks (Table 3) and achieves very high per-dataset performance on several problems (e.g., MCC ≥ 0.95 for Iris, Penguins, Wine, Dermatology, and the Cell populations dataset). The permutation tests (with multiple-comparison correction) aggregated in Table 3 indicate that these improvements are not merely random fluctuations: they translate into statistically detectable differences for many dataset–classifier pairs (see also Appendix Table A1 and Table A2). Note that we did not apply variance-optimized feature scaling for the benchmark; because k-nearest neighbors’ decisions are distance-based, kNN is not expected to attain its best possible performance under our preprocessing. The choice of scaling and distance is often empirical and context-dependent, and there is no single universally “correct” recipe.
Second, PDENB’s core strength is its flexibility in modeling complex, non-Gaussian feature distributions without parametric assumptions. The Satellite dataset illustrates this point: feature distributions (see Appendix C) and class-conditional distributions for this set display long tails, multimodality, and skewness that violate Gaussian assumptions. In that setting, PDENB captures the fine structure that classical Gaussian Naïve Bayes misses, yielding substantially better discriminative performance. This example underscores the value of highly adaptive density estimation methods when confronted with complex, non-Gaussian data structures. This pattern—non-parametric methods outperforming Gaussian approximations when data depart from normality—is borne out across the benchmark: non-parametric Naïve Bayes variants tend to outrank their Gaussian counterparts (see Table 2).
Third, PDENB directly supports interpretability through visualization. The class-conditional mirrored density (MD) plots and the customized 2D Voronoi posterior maps provide intuitive, feature-level, and case-level explanations as outlined using Figure 2, Figure 3, Figure 4 and Figure 5: users can inspect class-conditional likelihood shapes (modes, skewness, overlaps) and identify the feature combinations that produce compact, high-posterior decision areas. These visual diagnostics do not replace formal model evaluation, but they materially aid exploratory analysis and hypothesis generation, and they help explain why a particular prediction was made in cases where two or more features jointly determine a compact posterior region (see Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10).
The FlowCytometry application of distinguishing blood vs. bone marrow illustrates a practical use case where PDENB’s combination of sensitivity to distributional fine structure, feature selection, and interpretability is valuable. In such features of cell population frequencies, absolute counts, and percentages, classical Gaussian assumptions are violated (see Appendix C Figure A16 and Figure A17). PDENB achieved 98.8 MCC under cross-validation. The improved baseline performance is practically meaningful because higher sample-level accuracy reduces the risk of mislabeling the origin of aspirates of bone marrow vs. peripheral blood, which in turn can reduce downstream diagnostic errors. This result is a first indication that PDENB could potentially support clinically relevant classification tasks even with modest sample sizes, provided that careful feature selection and validation are applied.
The influence of dependency was tracked with four different measures in Table 1 after preprocessing. High correlation values (>0.8) using all four measures are depicted in Table 1 for all datasets except for the following three: CoverType, Swiss, and Wine (Crabs and Penguin had low correlations due to rotation by ICA or PCA). Although naïve Bayes theoretically assumes feature independence, in practice, correlated features did not necessarily imply a low performance (<0.8). For example, the Cell populations dataset retained high performance despite correlated features above 0.9. The removal of correlated features did not necessarily improve performance. Still, correlations can affect interpretability and sometimes classifier reliability; feature decorrelation, conditional modeling, or methods that explicitly capture dependencies may further improve the model’s performance in specific domains. In our benchmark, we observed that high feature correlation does not uniformly impair PDENB.
Our benchmark covers a diverse but limited set of datasets; broader evaluations—especially in high-dimensional, noisy, or highly imbalanced settings—would strengthen generality claims, although benchmarking against all families of classifiers is a controversial topic [63,64]. We emphasize several methodological points and caveats. PDENB’s robust performance hinges on three design choices: (i) estimating a single, class-independent Pareto radius per feature; (ii) applying smoothing to the raw class-conditional PDE output prior to using it as a likelihood; and (iii) applying a plausibility correction to the class likelihoods. While these design choices increase estimator stability and interpretability, they come at the cost of greater computational demand. To mitigate the computational cost, we provide a multicore, shared-memory [65] implementation for large-scale applications. Finally, embedding PDENB visualizations within formal explainable-AI workflows (e.g., counterfactual analyses, local-explanation wrappers) would enhance their utility for decision-makers.
Several design alternatives (e.g., kernels, global or local Pareto radii) were explored during preliminary experimentation. For example, we approximate the Pareto radius through a single, class-independent parameter rather than estimating separate radii for each feature–class combination. While this could reduce the model’s flexibility, preliminary internal experiments and benchmarking indicate that the resulting approximation error is small and that the simplification is sufficiently accurate for practical applications. The final architecture and parameter configuration were fixed after extensive testing, as this setup consistently provided the most stable and reliable performance across validation settings. Therefore, we fixed the parameters, especially regarding the radius estimation, which could be estimated through class-dependent or global methods, the choice of kernel, and several competing approaches to assess low evidence areas in the naïve Bayes theorem. The curious reader will find even more possibilities for changing parameters in the presented setting.
As an example, we present results from our preliminary experimentation to support the process of deciding between a global versus a local (class-dependent) Pareto radius in Appendix D and showcase our reasoning. As Appendix D shows, the strategy using a global Pareto radius achieves better results in the high-performing examples, while the local radius achieves better results in the low-performing and moderately performing examples. Therefore, it is justifiable that the global strategy is the default setting, although the user is able to change the radius estimation to the local mode.
The computation of PDENB’s plausibility threshold ε is a data-driven optimal selection of the group of cases with smallest joint likelihood across all classes (see ABC-Analysis [29]). Therefore, the relevant cases for which a plausible class assignment could be considered are determined automatically.
The δ-based adjustment provides a controlled way to nudge likelihood mass from the current top class towards a nearby mode without distorting the global posterior landscape. The parameter δ governs the trade-off between conservatism and decisiveness: a small δ produces minor, local adjustments that yield stable and interpretable posterior changes, whereas a larger δ would raise the likelihood of MAP flips in low-evidence regions. Because Equation (15) transfers a bounded amount of mass from one class to another, the overall likelihood geometry is preserved and posterior trajectories remain smooth rather than exhibiting abrupt thresholding, particularly near class boundaries. This produces a gradual change in the likelihood profile (illustrated in Figure 1 and resembling a fading-variance effect) and enhances the MAP estimate for the closest-mode class while retaining Bayesian coherence and interpretability.
In Appendix E we show that random forest (RF) yields a small but consistent edge in raw MCC on four imbalanced datasets, while PDENB provides a transparent, well-calibrated, and parameter-free baseline the errors of which are interpretable from its class-conditional likelihood plots. The benchmark on four imbalanced datasets hints that PDENB cannot necessarily achieve high performance for minority classes if there is a lack of samples, since the minority might be underrepresented. When two classes overlap strongly in feature space, one class can suffer disproportionately poor performance—especially if it is less prevalent or more heterogeneous. Differences in class prevalence (priors) or in within-class variability shift posterior probabilities toward the dominant (or more concentrated) class, which increases misclassification of the disadvantaged class. An estimation of a risk leading to a reweighting of the posteriori might be helpful if the false negative rate of a specific minority class must be minimized.
The PDENB is applicable to numeric tabular data as shown in the benchmark study. Theoretically, it is limited to independent features, however in practice it shows high performance despite significant dependency measures within the data. The PDENB can process big and high-dimensional data, as was shown in the benchmark study. In the case of Covertype (over 450.000 training samples and over 100.000 test samples with 17 features), the model training time is under 35 s and the classification requires around 3 s. Larger datasets with millions of cases could be processed within a few minutes. A more explicit runtime test was executed and documented in Appendix F. The PDENB can compute large datasets of a million cases with up to 100 feature dimensions within a few minutes (<7 min). The computational speed of the training drastically decreases for hundreds of features and multiple millions of observations; however, it is still feasible, with a runtime of under an hour. Its prediction remains comparably fast, with a runtime of under 13 min for 4 million cases and 100 features.
PDENB assumes continuous feature domains. For features with only a few distinct numeric values or very small class-specific sample sizes, we therefore fall back to a simple histogram-based likelihood with an automatically chosen number of bins [16], as described in Appendix G. Truly categorical variables, i.e., unordered labels without a meaningful numeric scale, are mathematically different: they do not admit a probability density in the sense defined in the Section 2.3 ff., and we do not implement a dedicated categorical likelihood model for them in the present work. In principle, PDENB could be extended to such features by replacing the continuous density with a multinomial/Dirichlet class-conditional model over categories, or by embedding categories into a continuous representation and applying PDE in that space. We regard this as a natural direction for future work and, in this study, deliberately restrict ourselves to continuous and quasi-continuous features, for which PDE and the histogram fallback are well defined and interpretable.
The plausible correction in PDENB relies on effective mode separation, which is determined using quantile-based criteria. If class separation is not feasible, instead of plausible correction, the classical Bayes theorem will be used. Conversely, when class separation exists, skewed distributions do not affect the computation of the plausible correction, provided that mode recognition remains accurate. However, incorrect mode recognition may deteriorate the correction outcome. The presented framework offers a foundation for the development of autogating approaches. In future work, the two-dimensional representation based on Voronoi cells, together with the a posteriori probability derived from PDENB, could be utilized to identify regions of interest within the feature space. Convex hulls may then be employed to characterize specific properties associated with particular feature–class combinations. By combining these properties, it would be possible to define automated and data-driven class assignment strategies, thereby enabling fully autonomous gating within this framework.
The advantage of the PDENB is clearly its assumption-free modeling of the data and robust performance. PDENB can be applied to data to leverage fine details of the distributions. Its assumption-free smoothed density estimation, combined with a plausibility correction of the Bayes theorem, yields reliable and robust posterior estimates.
Patterns arising from the multiplicative combination of feature-wise likelihoods can be recognized by the user and serve as intuitive explanations of how different features jointly contribute to the final class assignment.

5. Conclusions

This work presents a novel way of solving the naïve Bayes classification considering critical decision details. An parameter-free density estimation allows for the adaptive modeling of continuous one-dimensional features. The result is a robust classifier achieving results in line with the state of the art. The advantage of the proposed classifier is its robust and adaptive modeling. Furthermore, its resulting class-conditional likelihoods can be visualized with Mirrored-Density plots, enabling an interpretable and explorative approach to machine learning. Interpretable visualizations allow swift hypothesis exploration. Future work will tackle the challenge of identifying the most important visualizations of likelihoods and posteriors. These robust results allow for its benchmarks to be referenced against other methods. The algorithm is accessible via https://CRAN.R-project.org/package=PDEnaiveBayes, accessed on the 17 November 2025.

Author Contributions

Conceptualization: M.C.T.; Methodology: M.C.T. and Q.S.; Formal analysis and investigation: Q.S. and M.C.T.; Writing—original draft preparation: Q.S. and M.C.T.; Writing—review and editing: Q.S., J.H. and M.C.T.; Application Data: J.H.; Resources: M.C.T. and J.H.; Supervision: M.C.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The FCS dataset was manually curated by Prof. Stefan W. Krause, Medizinische Klinik 5—Hämatologie/Onkologie Uniklinikum Erlangen. UCI [38] is an open-access platform.

Acknowledgments

We thank Stefan W. Krause, Uniklinikum Erlangen, for providing the FCS dataset [37] used in this work.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Appendix A presents 14 Mirrored Density plots (MD plots) as one of the main results of the classification benchmark study. MD plots are used to visualize the distribution of the Matthews Correlation Coefficient, representing a performance measure for multi-class problems. Each plot lines up the distributions for all classifiers applied on one particular dataset for direct visual comparison. For smaller datasets (less than 10 k samples), the mean from 100 trials of a repeated hold-out cross-validation is used, while for big datasets (>10 k samples), one hold-out set is used and 100 means are obtained by resampling methods from randomly drawn subsets.
Figure A1. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Cell population.
Figure A1. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Cell population.
Make 08 00013 g0a1
Figure A2. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Covertype.
Figure A2. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Covertype.
Make 08 00013 g0a2
Figure A3. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Crabs (Sex).
Figure A3. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Crabs (Sex).
Make 08 00013 g0a3
Figure A4. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Crabs (Sp).
Figure A4. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Crabs (Sp).
Make 08 00013 g0a4
Figure A5. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Dermatology.
Figure A5. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Dermatology.
Make 08 00013 g0a5
Figure A6. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Iris.
Figure A6. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Iris.
Make 08 00013 g0a6
Figure A7. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset LetterRecognition.
Figure A7. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset LetterRecognition.
Make 08 00013 g0a7
Figure A8. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset MiceProtein.
Figure A8. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset MiceProtein.
Make 08 00013 g0a8
Figure A9. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Penguin.
Figure A9. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Penguin.
Make 08 00013 g0a9
Figure A10. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Spam.
Figure A10. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Spam.
Make 08 00013 g0a10
Figure A11. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Satellite.
Figure A11. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Satellite.
Make 08 00013 g0a11
Figure A12. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Swiss.
Figure A12. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Swiss.
Make 08 00013 g0a12
Figure A13. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset WCBCD.
Figure A13. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset WCBCD.
Make 08 00013 g0a13
Figure A14. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Wine.
Figure A14. MD plot presenting the distributions of the Matthews Correlation Coefficient (MCC) for all classifiers on the dataset Wine.
Make 08 00013 g0a14

Appendix B

Table A1. The table presents the results of the permutation test for each combination of classifiers in the columns and for all datasets in the rows.
Table A1. The table presents the results of the permutation test for each combination of classifiers in the columns and for all datasets in the rows.
PDENB.GNBPDENB.NPNBPDENB.kNN7PDENB.PyGNBPDENB.klaRGNBPDENB.klaRNPNBPDENB.e1071GNBGNB.NPNBGNB.kNN7GNB.PyGNBGNB.klaRGNBGNB.klaRNPNBGNB.e1071GNBNPNB.kNN7
CellPopulations00000000000000
CoverType00000000001010
Crabs (Sex)0.4370.0100.4770.4430.010.4130.2210110.13310
Crabs (SP)0.022000.0240.01700.0220011010
Dermatology000000000.03800010
Iris000.93800000.2060.001110.2110.002
LetterRecognition00.1890000.1630000.2711010
MiceProtein00.237000000000010
Penguins0.0050.010.12200.0050.010.0010011010.635
Spam0000.994000000000.0080
Satellite00000000001010
Swiss0.0010.0030.887000.00101011110.006
WCBCD0.231000.2330.23100.220011010
Wine0.006000.0050.00800.0050.3020.077110.35310.269
NPNB.PyGNBNPNB.klaRGNBNPNB.klaRNPNBNPNB.e1071GNBkNN7.PyGNBkNN7.klaRGNBkNN7.klaRNPNBkNN7.e1071GNBPyGNB.klaRGNBPyGNB.klaRNPNBPyGNB.e1071GNBklaRGNB.klaRNPNBklaRGNB.e1071GNBklaRNPNB.e1071GNB
CellPopulations000.0700000000010
CoverType000.19700000000010
Crabs (Sex)0.1960.2010.4620.209000010.10410.12310.158
Crabs (SP)000.47100000101010
Dermatology00100000.044000000
Iris0.2080.1810.2040.0020.00300.00110.18910.20910.19
LetterRecognition000.511000000.23200.24010
MiceProtein00001110110100
Penguins001000.0010.6650101010
Spam00000000000000
Satellite000.2600000000010
Swiss1111000.0050111111
WCBCD000.03600000101010
Wine0.3080.30610.3080.0640.10.1940.08310.36310.34910.352
Table A2. The table presents the ranks of the permutation test for each combination of classifiers in the columns and for all datasets in the rows.
Table A2. The table presents the ranks of the permutation test for each combination of classifiers in the columns and for all datasets in the rows.
PDENBGNBNPNBkNN7PyGNBklaRGNBklaRNPe1071G
CellPopulations172.5845.52.55.5
CoverType273.51573.57
Crabs (Sex)44484444
Crabs (SP)336.58336.53
Dermatology162.56482.56
Iris1.55.55.51.55.55.55.55.5
LetterRecognition25.5285.55.525.5
MiceProtein1.53.51.56.56.56.56.53.5
Penguins6.52.56.56.52.52.56.52.5
Spam1.54.5631.5874.5
Satellite174.53274.57
Swiss7.53.53.57.53.53.53.53.5
WCBCD447.51447.54
Wine4.54.54.54.54.54.54.54.5

Appendix C

Figure A15. MD plot presenting the one-dimensional distributions of all features of the dataset Satellite.
Figure A15. MD plot presenting the one-dimensional distributions of all features of the dataset Satellite.
Make 08 00013 g0a15
Figure A16. MD plot presenting the one-dimensional distributions of the first 53 features of the ALPODS-identified population frequencies based on the Dresden dataset. Magenta overlay indicates that statistical testing resulted in the hypothesis of the distribution beeing Gaussian.
Figure A16. MD plot presenting the one-dimensional distributions of the first 53 features of the ALPODS-identified population frequencies based on the Dresden dataset. Magenta overlay indicates that statistical testing resulted in the hypothesis of the distribution beeing Gaussian.
Make 08 00013 g0a16
Figure A17. MD plot presenting the one-dimensional distributions of the last 53 features of the ALPODS-identified population frequencies based on the Dresden dataset.
Figure A17. MD plot presenting the one-dimensional distributions of the last 53 features of the ALPODS-identified population frequencies based on the Dresden dataset.
Make 08 00013 g0a17

Appendix D

In order to showcase our decision-making for deciding on the crucial parameters in our robust framework, we present results from our preliminary experiments, which give insights into the estimation of the Pareto radius used for density estimation. We differentiate between a global Pareto radius using all feature information independent of class, and a local one, which estimates the radius based on class information. We executed a benchmark in the same style as for the main results, with 100 trials evaluated with the MCC (Matthews Correlation Coefficient). The Table A3, Table A4 and Table A5 present the mean of the 100 trials between the global and the local method. When comparing the results between the two methods, we group the results into three distinct classes of different quality depending on their overall classification performance. We can distinguish between the high performers, with 0.96–0.99 MCC, the moderate performers, with 0.88–0.92 MCC, and the low performers, with 0.44–0.68 MCC.
Table A3. The table compares the performance results between a global and a local class-dependent estimated Pareto radius. The table presents the high-performing datasets according to the ABC analysis.
Table A3. The table compares the performance results between a global and a local class-dependent estimated Pareto radius. The table presents the high-performing datasets according to the ABC analysis.
GlobalLocal
Crabs (SP)0.990.97
Dermatology0.950.95
CellPopulations0.980.98
Iris0.950.95
Penguins0.980.97
Swiss0.980.98
Wine0.970.96
Table A4. The table compares the performance results between a global and a local class-dependent estimated Pareto radius. The table presents the moderately performing datasets according to the ABC analysis.
Table A4. The table compares the performance results between a global and a local class-dependent estimated Pareto radius. The table presents the moderately performing datasets according to the ABC analysis.
GlobalLocal
Crabs (Sex)0.910.92
MiceProtein0.840.85
Satellite0.770.78
WCBCD0.890.88
Table A5. The table compares the performance results between a global and a local class-dependent estimated Pareto radius. The table presents the low-performing datasets according to the ABC analysis.
Table A5. The table compares the performance results between a global and a local class-dependent estimated Pareto radius. The table presents the low-performing datasets according to the ABC analysis.
GlobalLocal
Spam0.610.68
CoverType0.440.46
LetterRecognition0.660.63
Figure A18. MD plot presenting the one-dimensional distributions of the performance differences on the benchmark datasets between two versions of the PDENB, one computed with a global estimated Pareto radius and one computed with a local class-dependent estimated Pareto radius.
Figure A18. MD plot presenting the one-dimensional distributions of the performance differences on the benchmark datasets between two versions of the PDENB, one computed with a global estimated Pareto radius and one computed with a local class-dependent estimated Pareto radius.
Make 08 00013 g0a18

Appendix E

To evaluate the influence of class imbalance, we selected two datasets from our benchmark, Dermatology and CellPopulations. In addition, we used the CellPopulations dataset described in Section 2 and the Ecoli dataset from the UCI Machine Learning Repository. The Dermatology and Ecoli datasets were evaluated without additional preprocessing.
Furthermore, we measured an additional healthy-donor sample at the University Hospital of Marburg, Germany, using a Navios flow cytometer (Beckman Coulter). Besides forward and side scatter (FS_PEAK, FS_INT, SS_INT), this sample included the following markers: CD34 FITC, CD13 PE, CD7 PerCP-Cy5.5, CD33 PE-Cy7, CD56 APC, CD117 AF750, HLA-DR PB, and CD45 Krome Orange. The file was compensated on the instrument but not further preprocessed. Within this sample, we manually annotated major leukocyte subsets (B cells, granulocytes, monocytes, NK cells, and T/NK cells), which both classifiers were then tasked to predict under the cross-validation procedure described in Section 2. All datasets used in this analysis are summarized in Table A6.
The results show that Random Forests (RF) often, but not always, attain slightly higher average MCCs than PDENB on four imbalanced datasets (Dermatology: 0.97 vs. 0.95; Ecoli: 0.84 vs. 0.81; CellPopulations: 0.99 vs. 0.98; Leucocytes: 0.92 vs. 0.94), although the differences are small, so their practical significance should be confirmed on additional imbalanced datasets and then by an appropriate statistical test. The detailed contingency tables reveal where those small aggregate differences arise.
For Dermatology, RF improves most notably in class 2 and class 6, reducing false discoveries relative to PDENB, whereas PDENB performs comparably or slightly worse across several classes but retains high per-class recall for the dominant classes. In Ecoli, the poorest performance of both methods occurs in extremely small classes (two-sample classes C3 and C4), which are essentially impossible to learn stably; both classifiers assign those cases to the large class 8, indicating that class scarcity—not model idiosyncrasy—is the primary problem. RF has a better performance on class 7, which also has a small sample size. For the large, multi-class Cell Populations problem, both methods perform near-perfectly for the major classes, but PDENB shows more confusion in very small classes (e.g., class 12), again reflecting sample size limitations. For the Leucocytes, the Random Forest is able to learn the small classes BCells, Monocytes, and TNKCells slightly better than the PDENB; however, RF fails to learn NKCells almost completely and confuses NKCells with TNKCells, while PDENB achieves a better distinction of NKCells, especially towards the heavily overlapping class TNKCells. The example highlights the potential of naïve Bayes to learn challenging relationships between various class-dependent features given enough training data, even in situations with high class overlap.
Despite the small number of datasets, which limits generalizability, the observed patterns are consistent with the known strengths and weaknesses of the two methods: Random Forests can capture complex, multivariate interactions and handle heterogeneous features flexibly, which often gives them an advantage on structured, imbalanced multi-class problems.. PDENB, by contrast, builds on one-dimensional class-conditional likelihoods (with PDE smoothing and a plausibility correction); this design makes it robust and highly interpretable and often competitive in overall performance, but inherently less able to exploit the higher-order feature interactions that RF leverages. Class imbalance amplifies these differences: very small classes produce noisy density estimates for PDENB and few training examples for RF, so both fail, but RF’s tree ensemble can sometimes use correlated features to recover tiny classes more effectively.
Table A6. The table presents a data summary of the imbalanced datasets used in this section.
Table A6. The table presents a data summary of the imbalanced datasets used in this section.
NDIMClass No.Cases per ClassPearsonSpearmanKendallXICOR
Dermatology358346111, 60, 71, 48, 48, 200–0.940–0.980–0.940–0.85
Ecoli33678143, 77, 2, 2, 35, 20, 5, 520.01–0.810–0.720–0.610–0.57
CellPopulations512114131128, 312, 829, 525, 768, 122, 76, 135, 283, 630, 229, 31, 530.01–0.990–0.980–0.90–0.85
Leucocytes482,30412514,488, 346,801, 43,556, 6092, 71,3670–0.960–0.950–0.80–0.72
Table A7. The table presents the overview of the overall performance (mean MCC ± AMAD) of the PDENB versus a Random Forest on four imbalanced datasets.
Table A7. The table presents the overview of the overall performance (mean MCC ± AMAD) of the PDENB versus a Random Forest on four imbalanced datasets.
DatasetPDENBRandom Forest
Dermatology0.95 ± 0.0340.97 ± 0.032
Ecoli0.81 ± 0.0710.84 ± 0.066
CellPopulations0.98 ± 0.00060.99 ± 0.002
Leucocytes0.94 ± 0.00.92 ± 0.0
Table A8. The table presents a contingency table of the dataset Dermatology for the PDENB classifier to allow an insight into the results of classification of an imbalanced dataset.
Table A8. The table presents a contingency table of the dataset Dermatology for the PDENB classifier to allow an insight into the results of classification of an imbalanced dataset.
TotalClass RatioClassPrediction of Class 1Prediction of Class 2Prediction of Class 3Prediction of Class 4Prediction of Class 5Prediction of Class 6False Discovery Rate
11131.01C198.681.2700.05001.32
6016.76C23.9289.3306.750010.67
7119.83C30099.790.070.1400.21
4813.41C42.43.7093.9006.1
4813.41C523.30094.705.3
205.59C66.5100092.57.5
Table A9. The table presents a contingency table of the dataset Dermatology for a Random Forest to allow an insight into the results of classification of an imbalanced dataset.
Table A9. The table presents a contingency table of the dataset Dermatology for a Random Forest to allow an insight into the results of classification of an imbalanced dataset.
TotalClass RatioClassPrediction of Class 1Prediction of Class 2Prediction of Class 3Prediction of Class 4Prediction of Class 5Prediction of Class 6False Discovery Rate
11131.01C199.910.0900000.09
6016.76C21.594.3303.6700.55.67
7119.83C30099.7100.2900.29
4813.41C409.6090.4009.6
4813.41C5000010000
205.59C610000991
Table A10. The table presents a contingency table of the dataset Ecoli for the PDENB classifier to allow an insight into the results of classification of an imbalanced dataset.
Table A10. The table presents a contingency table of the dataset Ecoli for the PDENB classifier to allow an insight into the results of classification of an imbalanced dataset.
TotalClass RatioClassPrediction of Class 1Prediction of Class 2Prediction of Class 3Prediction of Class 4Prediction of Class 5Prediction of Class 6Prediction of Class 7Prediction of Class 8False Discovery Rate
14342.56C198.550.03000001.421.45
7722.92C24.675.270.070.5319.47000.0724.74
20.6C30000000100100
20.6C40000000100100
3510.42C5237.140.29060000.5740
205.95C61.52.500083.75012.2516.25
51.49C7401127022211579
5215.48C87.33.10.100.10.5088.911.1
Table A11. The table presents a contingency table of the dataset Ecoli for a Random Forest to allow an insight into the results of classification of an imbalanced dataset.
Table A11. The table presents a contingency table of the dataset Ecoli for a Random Forest to allow an insight into the results of classification of an imbalanced dataset.
TotalClass RatioClassPrediction of Class 1Prediction of Class 2Prediction of Class 3Prediction of Class 4Prediction of Class 5Prediction of Class 6Prediction of Class 7Prediction of Class 8False Discovery Rate
14342.56C198.760000001.241.24
7722.92C22.685.470011.9300014.53
20.6C30000000100100
20.6C40000001000100
3510.42C52.5742.570054.57000.2945.43
205.95C61.25000086.7501213.25
51.49C7000000742626
5215.48C88.61.90000.10.189.310.7
Table A12. The table presents a contingency table of the dataset CellPopulations for the PDENB classifier to allow an insight into the results of classification of an imbalanced dataset.
Table A12. The table presents a contingency table of the dataset CellPopulations for the PDENB classifier to allow an insight into the results of classification of an imbalanced dataset.
TotalClass RatioClassClassPrediction of NeutrophilsPrediction of EosinophilsPrediction of B CellsPrediction of T Helper CellsPrediction of T Cells CD8+CD56-Prediction of T Cells CD8+CD56+Prediction of T Cells Double Neg.Prediction of NK Cells 1Prediction of NK Cells 2Prediction of Classical Mono.Prediction of Atypical Mono. 1Prediction of Atypical Mono. 2Prediction of BasophilesFalse Discovery Rate
112822.03NeutrophilsC199.920.08000000000000.08
3126.09EosinophilsC20.3199.230000000.060000.40.77
82916.19B CellsC30099.980.020000000000.02
52510.25T helper cellsC400099.990.01000000000.01
76815T cells CD8+CD56-C50000.0499.560.3600.04000000.44
1222.38T cells CD8+CD56+C600.960015.3883.67000000016.34
761.48T cells double neg.C70001.131.33097.530000002.46
1352.64NK cells 1C8000000097.112.8900002.89
2835.53NK cells 2C900000.3002.8496.8400.02003.16
63012.3Classical Mono.C1000000000099.950.05000.05
2294.47Atypical Mono. 1C11000000000.025.5294.1100.355.89
310.61Atypical Mono. 2C120004.170000011.1734.550.17049.84
531.03BasophilesC1301.450000000.8200097.732.27
Table A13. The table presents a contingency table of the dataset CellPopulations for a Random Forest to allow an insight into the results of classification of an imbalanced dataset.
Table A13. The table presents a contingency table of the dataset CellPopulations for a Random Forest to allow an insight into the results of classification of an imbalanced dataset.
TotalClass RatioClassPrediction of NeutrophilsPrediction of EosinophilsPrediction of B CellsPrediction of T Helper CellsPrediction of T Cells CD8+CD56-Prediction of T Cells CD8+CD56+Prediction of T Cells Double Neg.Prediction of NK Cells 1Prediction of NK Cells 2Prediction of Classical Mono.Prediction of Atypical Mono. 1Prediction of Atypical Mono. 2Prediction of BasophilesFalse Discovery Rate
112822.03Neutrophils1000000000000000
3126.09Eosinophils0.8298.69000000.03000.2900.161.3
82916.19B Cells0099.98000000.010000.010.02
52510.25T helper cells000.1299.880000000000.12
76815T cells CD8+CD56-0000.1199.030.73000.1400000.98
1222.38T cells CD8+CD56+00.46002.0497.500000002.5
761.48T cells double neg.000000.2799.730000000.27
1352.64NK cells 1000000098.261.7400001.74
2835.53NK cells 2000.1800.32001.6397.6800.1400.052.32
63012.3Classical Mono.00000000099.640.250.100.35
2294.47Atypical Mono. 1000000000.330.9198.570.201.44
310.61Atypical Mono. 20001.50000030.1795.3304.67
531.03Basophiles000000000.0900.73099.180.82
Table A14. The table presents a contingency table of the dataset Leucocytes for the PDENB classifier to allow an insight into the results of classification of an imbalanced dataset.
Table A14. The table presents a contingency table of the dataset Leucocytes for the PDENB classifier to allow an insight into the results of classification of an imbalanced dataset.
TotalClass RatioClassPrediction of BCellsPrediction of GranulocytesPrediction of MonocytesPrediction of NKCellsPrediction of TNKCellsFalse Discovery Rate
14,4883BCells99.590.24000.170.41
346,80171.91Granulocytes0.0199.70.120.020.140.29
43,5569.03Monocytes02.397.30.330.072.7
60921.26NKCells01.646.5755.0936.744.91
71,36714.8TNKCells0.253.670.596.1989.310.7
Table A15. The table presents a contingency table of the dataset Leucocytes for a Random Forest to allow an insight into the results of classification of an imbalanced dataset.
Table A15. The table presents a contingency table of the dataset Leucocytes for a Random Forest to allow an insight into the results of classification of an imbalanced dataset.
TotalClass RatioClassPrediction of BCellsPrediction of GranulocytesPrediction of MonocytesPrediction of NKCellsPrediction of TNKCellsFalse Discovery Rate
14,4883BCells99.90.03000.070.1
346,80171.91Granulocytes099.07000.930.93
43,5569.03Monocytes0097.641.071.32.37
60921.26NKCells00.97.642.9688.5197.05
71,36714.8TNKCells03.690.486.0689.7610.23

Appendix F

We computed PDENB for a runtime inspection on the CellPopulations dataset enhanced to 5 million cases, resulting in exemplary datasets of 1 and 4 million cases after being split into training and test datasets. Similar, the dataset can be artificially enlarged to 100 dimensions by reusing columns without a loss of computational load. We then executed the PDENB on the so-obtained dataset 30 times, using examples of both 14 and 100 dimensions, and 1 and 4 million cases. The final runtimes for training and prediction are documented with the median and AMAD (adjusted mean absolute deviation) in Table A16.
Table A16. The table presents the runtimes across 30 trials of the PDENB applied on a split of an enhanced version of the CellPopulations dataset. The CellPopulation dataset was enhanced to 5 million cases, resulting in a split into 1 and 4 million for training and testing purposes. The runtime was evaluated using the median and the AMAD (adjusted mean absolute deviation).
Table A16. The table presents the runtimes across 30 trials of the PDENB applied on a split of an enhanced version of the CellPopulations dataset. The CellPopulation dataset was enhanced to 5 million cases, resulting in a split into 1 and 4 million for training and testing purposes. The runtime was evaluated using the median and the AMAD (adjusted mean absolute deviation).
DIMNTrainingPrediction
MedianAMADMedianAMAD
141M0.653.740.372.96
4M2.376.411.284.72
1001M2.73.251.231.79
4M55.3675.725.537.2

Appendix G

The histogram-based fallback for the low-cardinality features in PDENB is designed primarily for continuous features and uses Pareto Density Estimation (PDE) when sufficient data are available per class. However, for features with few distinct values or very small class-specific sample sizes, PDE can become unstable. In these cases, we fall back to a simple histogram-based estimator with an automatically chosen number of bins, following the practical rules for the threshold of the minimal amount of values in data N u t and the quantity threshold (i.e., the threshold of the minimal amount of unique values in data) Q t , derived in [16].
Concretely, for a given feature and class, let N unique denote the number of distinct observed values and C i the number of samples in class C i . If the feature does not satisfy
N unique > N u t   and C i > Q t
we do not apply PDE (similar to the MD plot) but instead estimate a one-dimensional density via an equal-width histogram. First, we determine the number of bins n o p t using [34]. On the observed range x m i n , x m a x of the feature we construct equal-width bins h with edges in the range of data with
h = | x m a x x m i n | n o p t ,
e d g e s k = x m i n + k | x m a x x m i n | n o p t , k = 1 , ,   n o p t
The kernel grid points are defined as the bin midpoints between e d g e s k 1 and e d g e k
x ^ k = edges k 1 + edges k 2 , k = 2 , , n opt .
The class-conditional histogram density is obtained by normalizing bin counts by the total number of observations in the class and the bin width,
f ^ x ^ k C i = 1 M C i × h n = 1 N i 1 ( x n x ^ k   h 2 ) .
where 1 is 1 if the condition holds and 0 otherwise and M C i is the number of samples in class C i .
This yields a simple, well-normalized density estimate that is numerically stable for low-cardinality or small-sample situations and can be used directly as the class-conditional likelihood in PDENB. For categorical features we therefore currently fall back to a simple, practical histogram-style estimator based on the Q t threshold.

References

  1. Devroye, L.; Györfi, L.; Lugosi, G. A Probabilistic Theory of Pattern Recognition; Springer Science & Business Media: New York, NY, USA, 2013; Volume 31. [Google Scholar]
  2. Loizou, G.; Maybank, S.J. The nearest neighbor and the bayes error rates. IEEE Trans. Pattern Anal. Mach. Intell. 1987, PAMI-9, 254–262. [Google Scholar] [CrossRef] [PubMed]
  3. Fukunaga, K.; Kessell, D. Nonparametric Bayes error estimation using unclassified samples. IEEE Trans. Inf. Theory 1973, 19, 434–440. [Google Scholar] [CrossRef]
  4. Bellman, R.E. Adaptive Control Processes: A Guided Tour; Princeton University Press: Princeton, NJ, USA, 1961. [Google Scholar]
  5. Silverman, B.W. Density Estimation for Statistics and Data Analysis; Chapman and Hall: London, UK, 1998. [Google Scholar]
  6. Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification, 2nd ed.; A Wiley-Interscience Publication; John Wiley & Sons: New York, NY, USA, 2001. [Google Scholar]
  7. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  8. John, G.H.; Langley, P. Estimating continuous distributions in Bayesian classifiers. In Proceedings of the Eleventh Conference on Uncertainty in Artificial Intelligence, Montreal, QC, Canada, 18–20 August 1995; Morgan and Kaufman: San Mateo, CA, USA, 1995. [Google Scholar]
  9. Abu Alfeilat, H.A.; Hassanat, A.B.; Lasassmeh, O.; Tarawneh, A.S.; Alhasanat, M.B.; Salman, H.S.E.; Prasath, V.S. Effects of distance measure choice on k-nearest neighbor classifier performance: A review. Big Data 2019, 7, 221–248. [Google Scholar] [CrossRef]
  10. Domingos, P.; Pazzani, M. On the optimality of the simple Bayesian classifier under zero-one loss. Mach. Learn. 1997, 29, 103–130. [Google Scholar] [CrossRef]
  11. Zaidi, N.A.; Cerquides, J.; Carman, M.J.; Webb, G.I. Alleviating naive Bayes attribute independence assumption by attribute weighting. J. Mach. Learn. Res. 2013, 14, 1947–1988. [Google Scholar]
  12. Rish, I. An empirical study of the naive Bayes classifier. In Proceedings of the IJCAI 2001 Workshop on Empirical Methods in Artificial Intelligence, Seattle, WA, USA, 4 August 2001. [Google Scholar]
  13. van den Heuvel, E.; Zhan, Z. Myths about linear and monotonic associations: Pearson’s r, Spearman’s ρ, and Kendall’s τ. Am. Stat. 2022, 76, 44–52. [Google Scholar] [CrossRef]
  14. Ultsch, A.; Lötsch, J. Robust classification using posterior probability threshold computation followed by Voronoi cell based class assignment circumventing pitfalls of Bayesian analysis of biomedical data. Int. J. Mol. Sci. 2022, 23, 14081. [Google Scholar] [CrossRef]
  15. Ultsch, A. Pareto density estimation: A density estimation for knowledge discovery. In Innovations in Classification, Data Science, and Information Systems; Baier, D., Werrnecke, K.D., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; pp. 91–100. [Google Scholar]
  16. Thrun, M.C.; Gehlert, T.; Ultsch, A. Analyzing the Fine Structure of Distributions. PLoS ONE 2020, 15, e0238835. [Google Scholar] [CrossRef]
  17. Bock, H.H. Automatische Klassifikation: Theoret. u. prakt. Methoden z. Gruppierung u. Strukturierung von Daten (Cluster-Analyse). In Studia Mathematica; Grotemeyer, K.P., Morgenstern, D., Tietz, H., Eds.; Vandenhoeck & Ruprecht: Göttingen, Germany, 1974; Volume XXIV. [Google Scholar]
  18. Mitchell, T.M. Machine Learning, 24th ed.; McGraw-Hill Education: Noida, India, 1997; p. 414. [Google Scholar]
  19. Fukunaga, K.; Kessell, D.L. Estimation of classification error. IEEE Trans. Comput. 1971, 100, 1521–1527. [Google Scholar] [CrossRef]
  20. Majka, M. Naivebayes: High Performance Implementation of the Naive Bayes Algorithm in R, R package version 1.0.0; 2024. Available online: https://CRAN.R-project.org/package=naivebayes (accessed on 17 November 2025).
  21. Devroye, L.; Lugosi, G. Variable kernel estimates: On the impossibility of tuning the parameters. In High Dimensional Probability II; Springer: Boston, MA, USA, 2000; pp. 405–424. [Google Scholar]
  22. Ultsch, A. Optimal Density Estimation in Data Containing Clusters of Unknown Structure; University of Marburg, Department of Mathematics and Computer Science: Marburg, Germany, 2003. [Google Scholar]
  23. Blackman, R.B.; Tukey, J.W. The measurement of power spectra from the point of view of communications engineering—Part I. Bell Syst. Tech. J. 1958, 37, 185–282. [Google Scholar] [CrossRef]
  24. Jones, M.C.; Lotwick, H.W. Remark AS R50: A remark on algorithm AS 176. Kernel density estimation using the fast Fourier transform. J. R. Stat. Soc. Ser. C (Applied Stat.) 1984, 33, 120–122. [Google Scholar] [CrossRef]
  25. Silverman, B.W. Algorithm AS 176: Kernel density estimation using the fast Fourier transform. J. R. Stat. Soc. Ser. C (Applied Stat.) 1982, 31, 93–99. [Google Scholar] [CrossRef]
  26. Fritsch, F.N.; Carlson, R.E. Monotone piecewise cubic interpolation. SIAM J. Numer. Anal. 1980, 17, 238–246. [Google Scholar] [CrossRef]
  27. Scott, D.W. Multivariate Density Estimation: Theory, Practice, and Visualization; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  28. Silverman, B.W. Spline smoothing: The equivalent variable kernel method. Ann. Stat. 1984, 12, 898–916. [Google Scholar] [CrossRef]
  29. Ultsch, A.; Lötsch, J. Computed ABC Analysis for Rational Selection of Most Informative Variables in Multivariate Data. PLoS ONE 2015, 10, e0129767. [Google Scholar] [CrossRef]
  30. Gastwirth, J.L. A general definition of the Lorenz curve. Econom. J. Econom. Soc. 1971, 39, 1037. [Google Scholar] [CrossRef]
  31. Gastwirth, J.L.; Glauberman, M. The interpolation of the Lorenz curve and Gini index from grouped data. Econom. J. Econom. Soc. 1976, 44, 479–483. [Google Scholar] [CrossRef]
  32. Bickel, D.R.; Frühwirth, R. On a fast, robust estimator of the mode: Comparisons to other robust estimators with applications. Comput. Stat. Data Anal. 2006, 50, 3500–3530. [Google Scholar] [CrossRef]
  33. Ekblom, H. A Monte Carlo investigation of mode estimators in small samples. J. R. Stat. Soc. Ser. C (Applied Stat.) 1972, 21, 177. [Google Scholar] [CrossRef]
  34. Keating, J.P.; Scott, D.W. A primer on density estimation for the great homerun race of 1998. STATS 1999, 25, 16–22. [Google Scholar]
  35. Lux, M.; Rinderle-Ma, S. DDCAL: Evenly Distributing Data into Low Variance Clusters Based on Iterative Feature Scaling. J. Classif. 2023, 40, 106–144. [Google Scholar] [CrossRef] [PubMed]
  36. Shapiro, H.M. Practical Flow Cytometry; John Wiley & Sons: Hoboken, NJ, USA, 2005. [Google Scholar]
  37. Plank, K.; Dorn, C.; Krause, S.W. The effect of erythrocyte lysing reagents on enumeration of leukocyte subpopulations compared with a no-lyse-no-wash protocol. Int. J. Lab. Hematol. 2021, 43, 939–947. [Google Scholar] [CrossRef] [PubMed]
  38. Dua, D.; Graff, C. UCI Machine Learning Repository; University of California, Irvine, School of Information and Computer Science: Irvine, CA, USA, 2019. [Google Scholar]
  39. Pearson, K. On lines and planes of closest fit to systems of points in space. Lond. Edinb. Dublin Philos. Mag. J. Sci. 1901, 2, 559–572. [Google Scholar] [CrossRef]
  40. Hotelling, H. Analysis of a complex of statistical variables into principal components. J. Educ. Psychol. 1933, 24, 417. [Google Scholar] [CrossRef]
  41. Harmeling, S.; Meinecke, F.; Müller, K.-R. Injecting noise for analysing the stability of ICA components. Signal Process. 2004, 84, 255–266. [Google Scholar] [CrossRef]
  42. Karlis, D.; Saporta, G.; Spinakis, A. A simple rule for the selection of principal components. Commun. Stat.-Theory Methods 2003, 32, 643–666. [Google Scholar] [CrossRef]
  43. Thrun, M.C.; Ultsch, A. Using Projection-based Clustering to Find Distance- and Density-based Clusters in High-Dimensional Data. J. Classif. 2020, 38, 280–312. [Google Scholar] [CrossRef]
  44. Ultsch, A.; Lötsch, J. Euclidean distance-optimized data transformation for cluster analysis in biomedical data (EDOtrans). BMC Bioinform. 2022, 23, 233. [Google Scholar] [CrossRef]
  45. Chatterjee, S. A new coefficient of correlation. J. Am. Stat. Assoc. 2021, 116, 2009–2022. [Google Scholar] [CrossRef]
  46. Beygelzimer, A.; Kakadet, S.; Langford, J.; Arya, S.; Mount, D.; Li, S. FNN: Fast Nearest Neighbor Search Algorithms and Applications. 2024. Available online: https://CRAN.R-project.org/package=FNN (accessed on 17 November 2025).
  47. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V. Scikit-learn: Machine learning in python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  48. Roever, C.; Raabe, N.; Luebke, K.; Ligges, U.; Szepannek, G.; Zentgraf, M.; Meyer, D. klaR: Classification and Visualization. 2023. Available online: https://CRAN.R-project.org/package=klaR (accessed on 17 November 2025).
  49. Meyer, D.; Dimitriadou, E.; Hornik, K.; Weingessel, A.; Leisch, F. e1071: Misc Functions of the Department of Statistics, Probability Theory Group (Formerly: E1071), TU Wien. 2024. Available online: https://cran.r-project.org/package=e1071 (accessed on 17 November 2025).
  50. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  51. Hall, M.A.; Frank, E. Combining naive bayes and decision tables. In Proceedings of the FLAIRS Conference, Coconut Grove, FL, USA, 15–17 May 2008; Volume 2118, pp. 318–319. [Google Scholar]
  52. Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef] [PubMed]
  53. Chicco, D.; Tötsch, N.; Jurman, G. The Matthews correlation coefficient (MCC) is more reliable than balanced accuracy, bookmaker informedness, and markedness in two-class confusion matrix evaluation. BioData Min. 2021, 14, 13. [Google Scholar] [CrossRef] [PubMed]
  54. Wolpert, D.H.; Macready, W.G. No Free Lunch Theorems for Search; Technical Report SFI-TR-95-02-010; Santa Fe Institute: Santa Fe, NW, USA, 1995. [Google Scholar]
  55. Good, P. Permutation Tests: A Practical Guide to Resampling Methods for Testing Hypotheses; Springer Science & Business Media: New York, NY, USA, 2013. [Google Scholar]
  56. Novo, D. A comparison of spectral unmixing to conventional compensation for the calculation of fluorochrome abundances from flow cytometric data. Cytom. Part A 2022, 101, 885–891. [Google Scholar] [CrossRef]
  57. Thrun, M.C.; Ultsch, A. Effects of the payout system of income taxes to municipalities in Germany. In Proceedings of the 12th Professor Aleksander Zelias International Conference on Modelling and Forecasting of Socio-Economic Phenomena, Zakopane, Poland, 8–11 May 2018; Papież, M., Śmiech, S., Eds.; Foundation of the Cracow University of Economics: Cracow, Poland, 2018; pp. 533–542. [Google Scholar]
  58. Comon, P. Independent Component Analysis. In Higher-Order Statistics; Lacoume, J.L., Ed.; Elsevier: Issy-les-Moulineaux, France, 1992; pp. 29–38. [Google Scholar]
  59. Thrun, M.C. Projection-Based Clustering Through Self-Organization and Swarm Intelligence; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  60. Hoffmann, J.; Thrun, M.C.; Röhnert, M.A.; von Bonin, M.; Oelschlägel, U.; Neubauer, A.; Ultsch, A.; Brendel, C. Identification of critical hemodilution by artificial intelligence in bone marrow assessed for minimal residual disease analysis in acute myeloid leukemia: The Cinderella method. Cytom. Part A 2022, 103, 304–312. [Google Scholar] [CrossRef]
  61. Thrun, M.C.; Hoffmann, J.; Röhnert, M.; von Bonin, M.; Oelschlägel, U.; Brendel, C.; Ultsch, A. Flow Cytometry datasets consisting of peripheral blood and bone marrow samples for the evaluation of explainable artificial intelligence methods. Data Brief 2022, 43, 108382. [Google Scholar] [CrossRef]
  62. Ultsch, A.; Hoffmann, J.; Röhnert, M.A.; von Bonin, M.; Oelschlägel, U.; Brendel, C.; Thrun, M.C. An Explainable AI System for the Diagnosis of High-Dimensional Biomedical Data. BioMedInformatics 2024, 4, 197–218. [Google Scholar] [CrossRef]
  63. Fernández-Delgado, M.; Cernadas, E.; Barro, S.; Amorim, D. Do we need hundreds of classifiers to solve real world classification problems? J. Mach. Learn. Res. 2014, 15, 3133–3181. [Google Scholar]
  64. Wainberg, M.; Alipanahi, B.; Frey, B.J. Are random forests truly the best classifiers? J. Mach. Learn. Res. 2016, 17, 3837–3841. [Google Scholar]
  65. Thrun, M.C.; Märte, J. Memshare: Memory Sharing for Multicore Computation in R with an Application to Feature Selection by Mutual Information using PDE. arXiv 2025, arXiv:2509.08632. [Google Scholar] [CrossRef]
Figure 1. Left (AD) and right (EH) panels show results on artificial data for the classical and plausible PDE-based naïve Bayes classifier, respectively. Each panel contains four rows: N = 500 sampled points with predicted labels (A,H), class-conditional densities estimated from the training data (B,F), the posterior probability P(C1∣x) computed from the fitted model (C,G), and the test set of N = 5000 points with its predictions (D,H). Because class 1 (dark green) has a smaller variance, its posterior decays in both tails (C), and the MAP rule assigns extreme observations to class 2 in (D); we argue in favor of using the smoothed PDE to estimate the class likelihoods and the concept by [14] to correct assignments in regions of very low likelihood (F) that are not plausible in (G). In addition, the right panel shows that the fine structure of distributions should be accounted for in the class likelihoods (F). Without prior knowledge, applying the left model (C) to the test data produces misclassifications relative to the true boundary (magenta predictions to the left of the green predications in (D)) and is less interpretable in comparison to (H).
Figure 1. Left (AD) and right (EH) panels show results on artificial data for the classical and plausible PDE-based naïve Bayes classifier, respectively. Each panel contains four rows: N = 500 sampled points with predicted labels (A,H), class-conditional densities estimated from the training data (B,F), the posterior probability P(C1∣x) computed from the fitted model (C,G), and the test set of N = 5000 points with its predictions (D,H). Because class 1 (dark green) has a smaller variance, its posterior decays in both tails (C), and the MAP rule assigns extreme observations to class 2 in (D); we argue in favor of using the smoothed PDE to estimate the class likelihoods and the concept by [14] to correct assignments in regions of very low likelihood (F) that are not plausible in (G). In addition, the right panel shows that the fine structure of distributions should be accounted for in the class likelihoods (F). Without prior knowledge, applying the left model (C) to the test data produces misclassifications relative to the true boundary (magenta predictions to the left of the green predications in (D)) and is less interpretable in comparison to (H).
Make 08 00013 g001
Figure 9. The figure shows a customized 2D Voronoi tessellation based on the two features CD16_FITC and CD14_APC700 from the Cell populations dataset. The color palette is identical to Figure 6, with dark red indicating high posterior values and white indicating zero. It serves as a negative example, because B-cells cannot be reliably detected in CD16 and CD14. Very high posterior values approaching one for the B-cell class are highlighted in dark red. These dark red areas can be detected at various non-connected locations.
Figure 9. The figure shows a customized 2D Voronoi tessellation based on the two features CD16_FITC and CD14_APC700 from the Cell populations dataset. The color palette is identical to Figure 6, with dark red indicating high posterior values and white indicating zero. It serves as a negative example, because B-cells cannot be reliably detected in CD16 and CD14. Very high posterior values approaching one for the B-cell class are highlighted in dark red. These dark red areas can be detected at various non-connected locations.
Make 08 00013 g009
Figure 10. The figure shows three customized 2D Voronoi per tessellation per class based on three population frequencies from the Dresden dataset. (A) depicts the bone marrow class, and (B) depicts the blood class. Low posterior values are in white. Dark red areas present high posterior values.
Figure 10. The figure shows three customized 2D Voronoi per tessellation per class based on three population frequencies from the Dresden dataset. (A) depicts the bone marrow class, and (B) depicts the blood class. Low posterior values are in white. Dark red areas present high posterior values.
Make 08 00013 g010
Table 1. The table presents relevant meta information of the datasets used in the classification benchmark in the subsequent section. “N” states the number of cases, “DIM” the number of features, Class No. the number of classes, the number of cases per class, and the last four columns denote four different measures of dependency, namely Pearson’s correlation coefficient, Spearman’s Rank Correlation Coefficient, Kendall’s Tau, and Xi correlation coefficient.
Table 1. The table presents relevant meta information of the datasets used in the classification benchmark in the subsequent section. “N” states the number of cases, “DIM” the number of features, Class No. the number of classes, the number of cases per class, and the last four columns denote four different measures of dependency, namely Pearson’s correlation coefficient, Spearman’s Rank Correlation Coefficient, Kendall’s Tau, and Xi correlation coefficient.
NDIMClass No.Cases per ClassPearsonSpearmanKendallXICOR
Cell populations512114131128, 312, 829, 525, 768, 122, 76, 135, 283, 630, 229, 31, 530–0.990–0.980–0.90–0.85
CoverType581,012557211,840, 283,301, 35,754, 2747, 9493, 17,367, 20,5100–0.790–0.820–0.590–0.5
Crabs (Sex)20052100, 10000–0.070–0.030.01–0.24
Crabs (SP)20052100, 10000–0.070–0.030.01–0.24
Dermatology358346111, 60, 71, 48, 48, 200–0.940–0.980–0.940–0.85
Iris1504350, 50, 500.12–0.960.17–0.940.08–0.810.08–0.72
LetterRecognition20,0001626796, 755, 805, 783, 773, 748, 766, 789, 747, 792, 787, 753, 758, 775, 736, 734, 752, 761, 803, 768, 764, 786, 783, 813, 739, 7340–0.850–0.870–0.790–0.61
MiceProtein1080778150, 150, 135, 135, 135, 135, 105, 1350–10–10–10–0.99
Penguins34443152, 68, 12400–0.190–0.120–0.25
Satellite64353661533, 703, 1358, 626, 707, 15080–0.960–0.960.02–0.850.02–0.76
Spam46015722788, 18130–10–0.940–0.940–0.93
Swiss20062100, 1000.06–0.740.05–0.750.03–0.590.01–0.43
Wine17813359, 71, 480–0.860.01–0.880.01–0.70–0.6
WCBCD569302212, 3570–10–10–0.990–0.97
Table 2. The table presents the final results of the classifier performance evaluation. A total of 14 datasets were evaluated (rows) and 8 classification algorithms were applied. Performance was evaluated with the Matthews Correlation Coefficient (MCC) and the mean (denoted as μ ) + AMAD (Adjust Mean Absolute Deviation) (denoted as σ ) are used to determine the overall performance. NA values represent the impossibility of computation in specific cases.
Table 2. The table presents the final results of the classifier performance evaluation. A total of 14 datasets were evaluated (rows) and 8 classification algorithms were applied. Performance was evaluated with the Matthews Correlation Coefficient (MCC) and the mean (denoted as μ ) + AMAD (Adjust Mean Absolute Deviation) (denoted as σ ) are used to determine the overall performance. NA values represent the impossibility of computation in specific cases.
PDENBGNBNPNB7NNPyGNBklaRGNBklaRNPNBe1071GNB
μ σ μ σ μ σ μ σ μ σ μ σ μ σ μ σ
Cell populations0.9800.9700.9800.1600.9700.9700.9800.970
CoverType0.4400.2200.4100.8600.2200.2200.4100.220
Crabs (Sex)0.910.10.920.10.920.10.810.10.920.10.920.10.9200.920.1
Crabs (SP)0.9900.9900.9700.880.10.9900.9900.9700.990
Dermatology0.9500.8200.90.10.810.10.850NANA0.90.10.820
Iris0.950.10.940.10.940.10.9500.940.10.940.10.940.10.940.1
LetterRecognition0.720.10.660.10.720.10.180.10.660.10.660.10.720.10.660.1
MiceProtein0.8400.7500.840NANANANANANANANA0.750
Penguins0.9800.9800.9700.9700.9800.9800.9700.980
Spam0.680.10.5200.3800.5800.680NANA0.3700.520
Satellite0.7700.600.6200.6900.7500.600.6200.60
Swiss0.9800.9900.9900.9800.9900.9900.9900.990
WCBCD0.8900.890.10.8700.9300.890.10.890.10.8800.890.1
Wine0.970.10.960.10.960.10.950.10.960.10.960.10.960.10.960.1
Table 3. The table presents the grades computed as the mean of the ranks. Ranks were obtained in Appendix B based on the mean values of the study, and equal ranks are allowed and determined with a Permutation Test.
Table 3. The table presents the grades computed as the mean of the ranks. Ranks were obtained in Appendix B based on the mean values of the study, and equal ranks are allowed and determined with a Permutation Test.
PDENBPyGNBNPNBklaRNPNBe1071GNBGNBkNN7klaRGNB
Grade2.933.964.324.714.754.825.185.32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Stier, Q.; Hoffmann, J.; Thrun, M.C. Classifying with the Fine Structure of Distributions: Leveraging Distributional Information for Robust and Plausible Naïve Bayes. Mach. Learn. Knowl. Extr. 2026, 8, 13. https://doi.org/10.3390/make8010013

AMA Style

Stier Q, Hoffmann J, Thrun MC. Classifying with the Fine Structure of Distributions: Leveraging Distributional Information for Robust and Plausible Naïve Bayes. Machine Learning and Knowledge Extraction. 2026; 8(1):13. https://doi.org/10.3390/make8010013

Chicago/Turabian Style

Stier, Quirin, Jörg Hoffmann, and Michael C. Thrun. 2026. "Classifying with the Fine Structure of Distributions: Leveraging Distributional Information for Robust and Plausible Naïve Bayes" Machine Learning and Knowledge Extraction 8, no. 1: 13. https://doi.org/10.3390/make8010013

APA Style

Stier, Q., Hoffmann, J., & Thrun, M. C. (2026). Classifying with the Fine Structure of Distributions: Leveraging Distributional Information for Robust and Plausible Naïve Bayes. Machine Learning and Knowledge Extraction, 8(1), 13. https://doi.org/10.3390/make8010013

Article Metrics

Back to TopTop