Previous Article in Journal
Model-Based and Physics-Informed Deep Learning Neural Network Structures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Nonparametric Full Bayesian Significance Testing for Bayesian Histograms †

by
Fernando Corrêa
*,
Julio Michael Stern
and
Rafael Bassi Stern
*
Author to whom correspondence should be addressed.
Presented at the 43rd International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Ghent, Belgium, 1–5 July 2024.
These authors contributed equally to this work.
Phys. Sci. Forum 2025, 12(1), 11; https://doi.org/10.3390/psf2025012011
Published: 20 October 2025

Abstract

In this article, we present an extension of the Full Bayesian Significance Test (FBST) for nonparametric settings, termed NP-FBST, which is constructed using the limit of finite dimension histograms. The test statistics for NP-FBST are based on a plug-in estimate of the cross-entropy between the null hypothesis and a histogram. This method shares similarities with Kullback–Leibler and entropy-based goodness-of-fit tests, but it can be applied to a broader range of hypotheses and is generally less computationally intensive. We demonstrate that when the number of histogram bins increases slowly with the sample size, the NP-FBST is consistent for Lipschitz continuous data-generating densities. Additionally, we propose an algorithm to optimize the NP-FBST. Through simulations, we compare the performance of the NP-FBST to traditional methods for testing uniformity. Our results indicate that the NP-FBST is competitive in terms of power, even surpassing the most powerful likelihood-ratio-based procedures for very small sample sizes.

1. Introduction and General Setting

Full Bayesian Significance Testing (FBST) [1] is a Bayesian method for testing if a parameter θ belongs to some set Θ 0 . In traditional statistical setting, researchers analyze a collection of n observations X n = ( X 1 , , X n ) that are presumed to conform to a specified distribution f θ characterized by an unobserved parameter θ . A Bayesian statistician makes inferences about θ by updating a prior density π ( θ ) , supported by the set of all possibilities Θ . After observing X n , one obtains a posterior density f θ | X n . Often, one needs to determine whether f θ | X n supports a scientific hypothesis framed with respect to θ belonging to some subset Θ 0 Θ , written H 0 : θ Θ 0 . FBST tests H 0 by comparing the posterior density f θ | X n of points inside and outside Θ 0 . This comparison is represented by the posterior probability of the tangential set:
T ( Θ 0 ) = { θ : f θ | X ( θ ) sup t Θ 0 f θ | X ( t ) } .
T ( Θ 0 ) encompasses all points in the parameter space that exhibit lower posterior density compared to those in Θ 0 . The FBST methodology posits that if the posterior probability of T ( Θ 0 ) is low, the hypothesis H 0 should be rejected, as it is located in a region characterized by low posterior density.
Definition 1.
In a standard Bayesian statistical model, let Θ be a finite dimensional parametric space, X n an observed sample, L the likelihood function, π be the prior distribution, and f θ | X n the posterior density proportional to π ( θ ) L ( θ ) . Also, let Π θ | X n be the measure on Θ induced by f θ | X n . The Full Bayesian Significance Test (FBST) for testing H 0 : θ Θ 0 consists on rejecting the null hypothesis based on the e-value statistic
e v ( Θ 0 ; X n ) = Π θ | X n ( T ( Θ 0 ) )
where the tangential set T ( Θ 0 ) is given by Equation (1). H 0 is rejected if e v ( Θ 0 ; X n ) < α for some fixed significance level α [ 0 , 1 ] .
In other words, the e-value quantifies the credibility of a hypothesis H 0 using the maximum probability argument, whereby a system is optimally represented by its most probable realization. This probability is defined as the posterior density f θ | X n , which quantifies the continuous probability associated with a specific point θ Θ 0 . The e-value directly addresses the question “What is the posterior probability of observing a θ with a posterior density exceeding that of any point in Θ 0 ?”. A higher e-value signifies that H 0 : θ Θ 0 is deemed more credible, whereas a lower e-value suggests that H 0 is considered less credible.
In this paper, we extend this concept to a nonparametric framework for density estimation using histograms. Bayesian nonparametric approaches for density estimation can be divided into two main categories. The first type focuses on defining priors Π 0 in the infinite-dimensional space of probability densities. Upon observing the data X n , these priors are updated into infinite-dimensional posteriors, facilitating an adapted approach to Bayesian inference. Well-established examples of such priors include the Dirichlet Process Mixtures (DPM) and its extensions [2]. In contrast, the second type of Bayesian nonparametric approach employs regular finite-dimensional Bayesian modeling in parameter spaces Θ k ( n ) that maintain a fixed dimension k ( n ) , while allowing k ( n ) for gradual expansion as the sample sizes increase. This includes truncated versions of infinite-dimensional priors and histograms with a fixed number of bins that increases with the sample size. This paper specifically examines a variant of the FBST applied in the context of increasing dimensionality.
In this paper, we propose an FBST for the problem of Bayesian density estimation using Dirichlet-Multinomial models, interpreted as histograms where the number of bins increases with the sample size. This methodology is in alignment with the Bayesian frameworks outlined by [3,4,5]. Therefore, we will use consistent notation to leverage results from the existing literature. The primary advantages of leveraging the Dirichlet-Multinomial model include (1) the feasibility of deriving an explicit formula for the FBST test statistic in a nonparametric context, (2) the implicit relation between the formula for the FBST test statistic and the differential entropy estimation, and (3) the potential to extend frequentist consistency results from the literature to this method. These attributes collectively establish a robust framework for nonparametric hypothesis testing that is mathematically rigorous, interpretable through the lens of information geometry, and consistent from a frequentist standpoint.
This paper is structured as follows. Section 2 outlines the essential definitions and properties of our proposed methodology. Section 3 provides simulations demonstrating the statistical power of our test. Finally, Section 4 offers a discussion of our findings and potential avenues for future research. The proof of our results are presented in Appendix A.

2. FBST for Random Histograms

We start with a formal definition of our model. To maintain clarity, we will restrict our analysis to densities on [ 0 , 1 ] .
Definition 2.
For k N , consider the set of densities with support on [ 0 , 1 ] , defined as
H k = f L 1 ( [ 0 , 1 ] ) : f ( t ) = k i = 1 k 1 I i ( t ) w i and i = 1 k w i = 1 , w i 0
where I j = [ ( j 1 ) / k , j / k ) for j = 1 , 2 , , k . A random histogram θ is a random variable that selects a random element of H k ( n ) .
The distribution of θ is fully characterized by the distribution of the vector of random weights W = ( W 1 , , W k ) . Bayesian posterior inference on θ may also be conducted with respect to W if the likelihood is given by
L ( θ ) = i = 1 n θ ( X i ) w 1 i = 1 n 1 I 1 ( X i ) w k i = 1 n 1 I k ( X i )
which corresponds to the assumption that the sample values X 1 , , X n | θ are conditionally independent and share an identical density θ . In this paper, we shall consider random histograms sampled implicitly by Dirichlet priors of the weights W. This approach guarantees that the posterior inference on θ is conjugate and computationally tractable, as it is equivalent to inference on a Dirichlet-Multinomial Bayesian model.
Proposition 1.
Consider θ a random histogram with weights W Dirichlet ( α 1 , , α k ) . If X i X j | θ , i j and X i | θ θ , then the posterior θ | X n remains a random histogram with weights
W | X n Dirichlet ( α 1 + N 1 , , α k + N k )
where N j = i = 1 n 1 I j ( X i ) , 1 j k .
A usual approach for Bayesian nonparametric inference on a histogram is letting k ( n ) grow slowly with the sample size n. This may be interpreted as a data-dependent prior; the full parameter space being considered is the set of all densities and, contingent on n, random histograms puts mass only on specific subsets of this set. One could define priors that do not depend on n, but this would come at a heavy computational cost. Moreover, meaningful and computationally sound inference might be conducted both in frequentist and Bayesian perspective if we also require the priors of w to depend on n [2,4].
Fixing n and k ( n ) , the original definition of e v ( Θ 0 ; X n ) may be adapted to conduct tests regarding θ | X n . Given that there exists a bijection between an element of H k ( n ) and its corresponding weights ( W 1 , , W k ( n ) ) , the FBST test statistic may be defined in terms of the Dirichlet distribution defined in Equation (3). However, this approach comes at the price of being able to only test hypothesis of the form Θ 0 H k ( n ) . Therefore, if a researcher is interested in testing a hypothesis framed in terms of a general Θ 0 , our proposed procedure specifies a test statistic based on its finite-dimensional counterpart. As k ( n ) is permitted to increase with the sample size, this translation process becomes increasingly negligible.
Definition 3
(FBST for random histograms). Let θ be a random histogram defined by Dirichlet weights, and X n | θ represent an i.i.d. sample drawn from θ. The FBST test statistic for testing a hypothesis H 0 : θ Θ 0 , where Θ 0 is an arbitrary set of densities on [ 0 , 1 ] , is given by
e v ( Θ 0 ; X n ) = Π θ | X n j = 1 k ( n ) ( N j + α j ) log W j sup w S ( Θ 0 ) i = 1 k ( n ) ( N j + α j ) log w j
where S ( Θ 0 ) denotes probabilities attributed to the sets I 1 , , I k ( n ) by each element of Θ 0 :
S ( Θ 0 ) = I 1 f ( t ) d t , I 2 f ( t ) d t , , I k ( n ) f ( t ) d t : f Θ 0
The FBST for random histograms may be interpreted in the context of information theory. Let p and q be two m dimensional probability vectors. We recall that the cross-entropy divergence H ( p , q ) between those vectors is given by j = 1 m p i log q i and the Kullback–Leibler divergence D K L ( p | | q ) is given by p i log p i q i . By introducing w ^ n = N 1 + α j n + j = 1 k ( n ) α j , , N k ( n ) + α j n + j = 1 k ( n ) α j , Equation (3) can be articulated as
e v ( Θ 0 ; X n ) = Π θ | X n H ( w ^ n , W ) sup w S ( Θ 0 ) H ( w ^ n , w ) ,
e v ( Θ 0 ; X n ) = Π θ | X n D K L ( w ^ n | | W ) inf w S ( Θ 0 ) D K L ( w ^ n | | w ) .
These equations demonstrate that the application of the FBST definition leads to statistical tests grounded in an information-theoretic measure of divergence between a distribution of the sample into k ( n ) bins and the expected value of counts on those same bins under the assumption that θ is some hypothesized density f. Indeed, in the context of this particular test, a related concept has emerged in the literature on goodness-of-fit testing, notably in G-tests [6] and other methodologies rooted in frequentist nonparametric estimates of the continuous variant of the Kullback–Leibler divergence for probability densities [7]. Both tests utilize a χ 2 asymptotic distribution under the null hypothesis. For the FBST, there are specific rates of increase of k ( n ) that ensure the presence of analogous results.
Theorem 1.
If (1) X n is an independent and identically distributed sample of X 1 with density f * Lipchitz continuous on [ 0 , 1 ] ; (2) θ is a random histogram satisfying M > α i > m for all i and fixed quantities m , M and (3) k ( n ) = n 1 / 6 ( log n ) 1 / 6 + ϵ , for any ϵ > 0 , then the FBST for random histograms with H 0 : θ = f 0 satisfies
1. 
e v ( Θ 0 ; X n ) L Unif ( 0 , 1 ) if f 0 = f * and
2. 
e v ( Θ 0 ; X n ) P 0 if f 0 f * , where P denotes convergence in probability with respect to X n .
One particular virtue of Equation (4) is the simplicity of the optimization step in the FBST. This fact is due the convexity of the cross-entropy functional. Also, this optimization will be able to reject false null hypotheses as the sample size grows larger, as we exemplify for the case of fixed parametric families.
Theorem 2.
Let P α be a parametric family of differentiable distribution functions f α , such that min | | f α f * | | 2 2 = ϵ > 0 and Θ 0 k = { ( F α ( I 1 ) , . . . , F α ( I k ( n ) ) ) : F α P α } is the corresponding subset of the k ( n ) dimensional simplex. Then, the FBST on histograms for goodness-of-fit of this parametric family satisfies e v ( Θ 0 k ) 0 in f * probability.
This procedure is similar to other nonparametric methods that do not rely on maximum likelihood estimates for testing, but instead optimize specific statistics. This idea dates back to Berkson’s suggestion to minimize chi-squared rather than maximize likelihood [8], although there have been few attempts to directly optimize test statistics, such as the Kolmogorov–Smirnov statistic [9]. This may be because optimizing usual test statistics for goodness-of-fit, such as Kolmogorov–Smirnov, Anderson–Darling, and Cramer–von Mises [10], requires specialized optimization procedures, like the one developed in [9].
Alternatively, the most common approach for testing adherence to a parametric family of distributions involves estimating parameters by maximum likelihood and then deriving the null distribution of an existing test through resampling [11]. Our new test, as we will demonstrate in simulations, could also require corrections when the optimization suggested by Theorem 2 is used.

3. Simulations

In this section, we will compare the statistical power of our test with that of other available techniques through simulations. Both simple and composite null hypotheses will be considered. Simulations will be conducted using the R programming language [12] and its public repository of packages. For simple hypotheses, the following tests are compared:
  • The e-value for histograms, as defined in Definition 3, adopting with α i , j = 1 , with the number of bins defined as the hypothesis of Theorem 1;
  • Classic Kolmogorov–Smirnov (KS) test, as described in [6];
  • Alternative versions of KS, AD, and CV, constructed by [10], implemented by the R Package [13].
Following [10], we shall compare the NP-FBST of Definition 3 with
k ( n ) = max { 2 , n 1 / 6 log ( n ) 1 / 6 1 }
using sample sizes n = 10 , 20 , 30 , 50 , 70 , 100 , 150 , 200 , 300 . The null hypothesis tested shall be H 0 : f * = U n i f ( 0 , 1 ) and H 1 will be simulated in 4 scenarios: B e t a ( 1.6 , 1.6 ) , B e t a ( 1.3 , 1.3 ) , B e t a ( 0.8 , 0.8 ) , and B e t a ( 0.6 , 0.6 ) . We calculate the statistical power as the % of correct rejection of H 0 , with rejection at the 5 % level, on 500 Monte Carlo sample. The results are summarized in Figure 1.
Analyzing Figure 1, we observe the following:
  • For α > 1 and β > 1 , the NP-FBST power may be much more powerful for small sample sizes, but it is still competitive for large sample sizes.
  • For α < 1 and β < 1 , which are non-Lipchitz alternative hypotheses, the test performs worse than Zhang’s alternatives [10]. However, it still shows comparable or superior power compared to the usual Kolmogorov–Smirnov statistic.
  • For sample sizes below 1000, k ( n ) will usually be very small, such as 2 or 3, so the test is just a regular multinomial test.
To showcase the approximation properties of the NP-FBST based on Lemma A1, we shall simulate one last example, but this time we will adopt k ( n ) = log 2 ( n ) , usually referred as Sturge’s Law, for histogram binning [14]. This is, of course, appropriate asymptotically, as log 2 ( n ) o ( n 1 / 6 / log ( n ) 1 / 6 ϵ ) for all ϵ > 0 . However, this produces very competitive statistical power for testing if a B e t a ( 2 , 2 ) is a Irwin-Hall ( 2 ) , as highlighted in Figure 2.

4. Conclusions

In this paper, we presented a new nonparametric Bayesian procedure extending the usual FBST for Bayesian histograms. We summarize our results as practical and theoretical.
On the practical and applied front, we draw the following conclusions:
  • For small sample sizes, our method is competitive in terms of statistical power, even compared to sophisticated alternatives such as Zhang’s tests [10].
  • For larger sample sizes, the very slow sample size growth required by Theorem 1 harms the statistical power. Therefore, other binning rules could be considered. Further research shall look for an adaptable number of bins. Desirable binning rules should be larger than n 1 / 6 log ( n ) 1 / 6 + ϵ for small sample sizes, but smaller for large sample sizes. Our simulations suggest that the usual k ( n ) = O ( log 2 ( n ) ) , known as Sturge’s Law, is a competitive alternative for moderate sample sizes lower than 1000.
  • Unlike previous attempts, our method is computationally inexpensive with competitive statistical power.
From a theoretical perspective, we derive the following conclusions:
  • The natural Dirichlet-Multinomial formulation of Bayesian histograms induces statistical tests based on estimates of Kullback–Leibler divergences. This formulation logically follows from the definition of the FBST, and the same logic could be applied to other Bayesian density estimation methods. The frequentist properties of other versions of this NP-FBST remain to be studied in future works, but our results highlight the Kullback–Leibler divergence as a possible “canonical” statistic for nonparametric versions of the FBST for density estimation.
  • Our results show that taking the limit of a slowly increasing finite-dimensional parameter space is a viable strategy for building nonparametric versions of the FBST. The frequentist properties of the FBST are intimately related to the Bernstein–von Mises theorem. Therefore, if these types of Gaussian approximations are available, our arguments should also hold. In fact, all the main references of this paper build specific growth rates of the dimension of the parameter space and could be used to find other versions of nonparametric FBSTs [3,4,5].
  • For composite hypotheses, our method can be used both for testing based on the maximum likelihood estimate of nuisance parameters and for directly optimizing the test statistics, which may be interpreted as a weighted likelihood function. Usual numerical methods for optimizing the likelihood will work for our test statistic, which is not the case for other usual statistics such as Anderson–Darling, Cramer–von Mises, or Kolmogorov–Smirnov.
For future research, we highlight that adaptively choosing the number of bins k ( n ) is crucial, as the statistical power is heavily influenced by this quantity. Additionally, Theorem 1 requires a Lipschitz continuous data-generating density, a usual assumption for histograms, but excludes unbounded densities, which are important from a practical and theoretical point of view. Extending our results to Hölder continuous densities is particularly important but requires the derivation of other versions of Bernstein–von Mises theorems.

Author Contributions

Conceptualization, J.M.S. and R.B.S.; methodology, R.B.S.; software, F.C.; validation, R.B.S. and J.M.S.; formal analysis, R.B.S.; investigation, F.C.; writing—original draft preparation, F.C.; writing—review and editing, J.M.S. and R.B.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by CAPES grant number 88887.613569/2021-00.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All code and simulated datasets are available on https://github.com/azeloc/histograms.maxent.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
FBSTFull Bayesian Significance Testing
NPNonparametric

Appendix A. Proofs

The following lemma is the main ingredient of the proof. As k ( n ) grows slowly, we are able to employ a normal approximation to n J ^ ( θ h ^ ) , where J is the square root of the Fisher information matrix, evaluated at h ^ .
Lemma A1
(Adapted from Theorem 2.4 and the discussion of Section 3 of [3]). Consider
  • f * a Lipchitz p.d.f. on [0, 1];
  • f * the posterior obtained using the model defined in Definition 2, with k ( n ) = n 1 / 6 log ( n ) 1 / 6 ϵ with a fixed ϵ > 0 ;
  • J ( θ ) as the square root of the multinomial Fisher information matrix evaluated at θ S k ( n ) , and J ^ = J ( h ^ ) the sample estimate of J ( θ ) ;
  • D ^ as the diagonal matrix with h ^ in its entries.
Then:
1. 
f θ | X n { θ : | | θ θ 0 | | ϵ } 0 in f * probability;
2. 
n J ^ ( h ^ θ 0 ) N ( 0 , 1 ) in distribution;
3. 
the largest eigenvalue of J ^ 2 is O ( k ( n ) 2 ) ;
4. 
f θ | X n ( θ ) Φ ( v ; h ^ , J ^ 2 / n ) d θ 0 in f * probability
Proof of Theorem 1.
Let N n and Φ n be the measure and density induced by the Gaussian approximation of Equation (A1). It follows from Equation (A1) that | Π n ( A ) N n ( A ) | 0 in probability for all measurable A. Equation (A1) also implies that the following expectation vanishes in probability:
f θ | X n ( θ ) f θ | X n ( θ ) Φ n ( θ ) f θ | X n ( θ ) d θ = E θ | X n f θ | X n ( θ ) Φ n ( θ ) f θ | X n ( θ ) 0 .
Now let B ϵ = { θ : | Φ n ( θ ) f θ | X n ( θ ) 1 | ϵ } . By Markov’s inequality it follows that
Π θ | X n B ϵ c E θ | X n f θ | X n ( θ ) Φ n ( θ ) f θ | X n ( θ ) δ 0
in f * probability for all ϵ .
Now, consider
T ( θ 0 ) = { θ : Φ n ( θ ) Φ n ( θ 0 ) }
and for a fixed γ > 0
C γ = { | | θ θ 0 | | < γ } .
One may verify that ( T ( θ 0 ) B ϵ 1 C δ ) T ( θ 0 ) and ( T ( θ 0 ) B ϵ 2 C δ ) T ( θ 0 ) for any choice of ϵ 1 , ϵ 2 and δ . It follows that
Π θ | X n ( T ( θ 0 ) ) = Π θ | X n ( T ( θ 0 ) ( B ϵ C δ ) ) + Π θ | X n ( T ( θ 0 ) ( B ϵ C δ ) c )
Π θ | X n ( T ( θ 0 ) ) + Π θ | X n ( B ϵ c C δ c )
and analogously
Π θ | X n ( T ( θ 0 ) ) Π θ | X n ( T ( θ 0 ) ) + Π θ | X n ( B ϵ c C δ c ) .
Therefore
| Π θ | X n ( T ( θ 0 ) ) Π θ | X n ( T ( θ 0 ) ) | Π θ | X n ( B ϵ c ) + Π θ | X n ( C δ c ) 0
in probability. Finally we conclude that the last convergence implies that
| Π θ | X n ( T ( θ 0 ) ) N n ( T ( θ 0 ) ) |
| Π θ | X n ( T ( θ 0 ) ) Π θ | X n ( T ( θ 0 ) ) | + | Π θ | X n ( T ( θ 0 ) ) N n ( T ( θ 0 ) ) | 0
in probability. This conclusion ensures that the e-value statistic may be approximated by N n ( T ( θ 0 ) ) . Now, note that as N n ( T ( θ 0 ) ) is based on a quadratic form, this probability shall might expressed in terms of the χ 2 distribution:
N n ( T ( θ 0 ) ) = N n n ( θ h ^ ) T J ^ 2 ( θ h ^ ) ( θ 0 h ^ ) T J ^ 2 ( θ 0 h ^ ) = 1 χ k ( n ) 2 ( θ 0 h ^ ) T J ^ 2 ( θ 0 h ^ ) .
Also, note that χ 2 may be approximated in distribution by a Gaussian with same mean and variance. Therefore we may rewrite
N n ( T ( θ 0 ) ) 1 Φ ( θ 0 h ^ ) T J ^ 2 ( θ 0 h ^ ) k ( n ) 2 k ( n )
and also, as h ^ is approximate Gaussian, ( θ 0 h ^ ) T J ^ 2 ( θ 0 h ^ ) is approximate χ k ( n ) 2 and therefore, ( θ 0 h ^ ) T J ^ 2 ( θ 0 h ^ ) k ( n ) 2 k ( n ) D N ( 0 , 1 ) . Therefore, by the continuous mapping theorem 1 Φ ( θ 0 h ^ ) T J ^ 2 ( θ 0 h ^ ) k ( n ) 2 k ( n ) converges to a uniform distribution.
For part 2, note that under H 1 ( θ 0 h ^ ) T J ^ 2 ( θ 0 h ^ ) might be approximated by a non-central chi-square with mean k ( n ) + ( θ i * θ 0 i ) 2 k ( n ) + n ( f 0 ( t ) f * ( t ) ) d t , which is asymptotically larger than k ( n ) . Hence,
( θ 0 h ^ ) T J ^ 2 ( θ 0 h ^ ) k ( n ) 2 k ( n ) ,
and then N n ( T ( θ 0 ) ) must converge to 0. □
Proof of Theorem 2.
By Lemma A1 we have that the largest eigenvalue of J 2 is of order k ( n ) 2 , and therefore the normal approximation obtained at Equation (A3) becomes
e v ( Θ 0 k ) 1 Φ ( θ α * h ^ ) T J ^ 2 ( θ α * h ^ ) k ( n ) 2 k ( n )
Now we note that the condition of the theorem implies that ( θ α * h ^ ) T J ^ 2 ( θ α * h ^ ) k ( n ) 2 k ( n ) is asymptotically larger than k ( n ) . Then it follows that the RHS of the above quantity converges to 0 in f * probability. □

References

  1. de B. Pereira, C.A.; Stern, J.M.; Wechsler, S. Can a significance test be genuinely Bayesian? Bayesian Anal. 2008, 3, 79–100. [Google Scholar] [CrossRef]
  2. Ghosal, S.; van der Vaart, A.W. Fundamentals of Nonparametric Bayesian Inference; Number 44 in Cambridge Series in Statistical and Probabilistic Mathematics; Cambridge University Press: Cambridge, UK; New York, NY, USA, 2017; pp. xxiv+646. [Google Scholar] [CrossRef]
  3. Ghosal, S. Asymptotic Normality of Posterior Distributions for Exponential Families when the Number of Parameters Tends to Infinity. J. Multivar. Anal. 2000, 74, 49–68. [Google Scholar] [CrossRef]
  4. Castillo, I.; Nickl, R. On the Bernstein–von Mises phenomenon for nonparametric Bayes procedures. Ann. Stat. 2014, 42, 1941–1969. [Google Scholar] [CrossRef]
  5. Castillo, I.; Rousseau, J. A Bernstein–von Mises theorem for smooth functionals in semiparametric models. Ann. Stat. 2015, 43, 2353–2383. [Google Scholar] [CrossRef]
  6. D’Agostino, R.B.; Stephens, M.A. Goodness-of-Fit Techniques; Marcel Dekker, Inc.: New York, NY, USA, 1986. [Google Scholar]
  7. Song, K.S. Goodness-of-fit tests based on Kullback-Leibler discrimination information. IEEE Trans. Inf. Theory 2002, 48, 1103–1117. [Google Scholar] [CrossRef]
  8. Berkson, J. Minimum Chi-Square, not Maximum Likelihood! Ann. Stat. 1980, 8, 457–487. [Google Scholar] [CrossRef]
  9. Zvi Drezner, O.T.; Zerom, D. A Modified Kolmogorov—Smirnov Test for Normality. Commun. Stat.-Simul. Comput. 2010, 39, 693–704. [Google Scholar] [CrossRef]
  10. Zhang, J. Powerful Goodness-of-Fit Tests Based on the Likelihood Ratio. J. R. Stat. Soc. Ser. (Stat. Methodol.) 2002, 64, 281–294. [Google Scholar] [CrossRef]
  11. Babu, G.J.; Rao, C.R. Goodness-of-Fit Tests When Parameters Are Estimated. Sankhyā Indian J. Stat. (2003–2007) 2004, 66, 63–74. [Google Scholar]
  12. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2022. [Google Scholar]
  13. Cui, N.; Zhou, M. DistributionTest: Powerful Goodness-of-Fit Tests Based on the Likelihood Ratio. R package version 1.1. Comprehensive R Archive Network (CRAN), 2020. Available online: https://CRAN.R-project.org/package=DistributionTest (accessed on 31 July 2024).
  14. Sturges, H.A. The Choice of a Class Interval. J. Am. Stat. Assoc. 1926, 21, 65–66. [Google Scholar] [CrossRef]
Figure 1. Simulated power for uniformity tests of several test statistics under specific simulated H 1 . All power percentages were calculated considering 500 Monte Carlo samples. (a) H 1 : B e t a ( 1.6 , 1.6 ) , (b) H 1 : B e t a ( 1.3 , 1.3 ) , (c) H 1 : B e t a ( 0.8 , 0.8 ) , (d) H 1 : B e t a ( 0.6 , 0.6 ) .
Figure 1. Simulated power for uniformity tests of several test statistics under specific simulated H 1 . All power percentages were calculated considering 500 Monte Carlo samples. (a) H 1 : B e t a ( 1.6 , 1.6 ) , (b) H 1 : B e t a ( 1.3 , 1.3 ) , (c) H 1 : B e t a ( 0.8 , 0.8 ) , (d) H 1 : B e t a ( 0.6 , 0.6 ) .
Psf 12 00011 g001
Figure 2. Simulated power for several tests with H 0 : f * = B e t a ( 2 , 2 ) under H 1 : f * = Irwin-Hall ( 2 ) .
Figure 2. Simulated power for several tests with H 0 : f * = B e t a ( 2 , 2 ) under H 1 : f * = Irwin-Hall ( 2 ) .
Psf 12 00011 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Corrêa, F.; Stern, J.M.; Stern, R.B. Nonparametric Full Bayesian Significance Testing for Bayesian Histograms. Phys. Sci. Forum 2025, 12, 11. https://doi.org/10.3390/psf2025012011

AMA Style

Corrêa F, Stern JM, Stern RB. Nonparametric Full Bayesian Significance Testing for Bayesian Histograms. Physical Sciences Forum. 2025; 12(1):11. https://doi.org/10.3390/psf2025012011

Chicago/Turabian Style

Corrêa, Fernando, Julio Michael Stern, and Rafael Bassi Stern. 2025. "Nonparametric Full Bayesian Significance Testing for Bayesian Histograms" Physical Sciences Forum 12, no. 1: 11. https://doi.org/10.3390/psf2025012011

APA Style

Corrêa, F., Stern, J. M., & Stern, R. B. (2025). Nonparametric Full Bayesian Significance Testing for Bayesian Histograms. Physical Sciences Forum, 12(1), 11. https://doi.org/10.3390/psf2025012011

Article Metrics

Back to TopTop