A Bayesian Decision-Theoretic Approach to Logically-Consistent Hypothesis Testing

This work addresses an important issue regarding the performance of simultaneous test procedures: the construction of multiple tests that at the same time are optimal from a statistical perspective and that also yield logically-consistent results that are easy to communicate to practitioners of statistical methods. For instance, if hypothesis A implies hypothesis B, is it possible to create optimal testing procedures that reject A whenever they reject B? Unfortunately, several standard testing procedures fail in having such logical consistency. Although this has been deeply investigated under a frequentist perspective, the literature lacks analyses under a Bayesian paradigm. In this work, we contribute to the discussion by investigating three rational relationships under a Bayesian decision-theoretic standpoint: coherence, invertibility and union consonance. We characterize and illustrate through simple examples optimal Bayes tests that fulfill each of these requisites separately. We also explore how far one can go by putting these requirements together. We show that although fairly intuitive tests satisfy both coherence and invertibility, no Bayesian testing scheme meets the desiderata as a whole, strengthening the understanding that logical consistency cannot be combined with statistical optimality in general. Finally, we associate Bayesian hypothesis testing with Bayes point estimation procedures. We prove the performance of logically-consistent hypothesis testing by means of a Bayes point estimator to be optimal only under very restrictive conditions. Entropy 2015, 17 6535


Introduction
One could (...) argue that 'power is not everything'.In particular for multiple test procedures one can formulate additional requirements, such as, for example, that the decision patterns should be logical, conceivable to other persons, and, as far as possible, simple to communicate to non-statisticians.
-G. Hommel and F. Bretz [1] Multiple hypothesis testing, a formal quantitative method that consists of testing several hypotheses simultaneously [2], has gained considerable ground in the last few decades with the aim of drawing conclusions from data in scientific experiments regarding unknown quantities of interest.Most of the development of multiple hypothesis testing has been focused on the construction of test procedures satisfying statistical optimality criteria, such as the minimization of posterior expected loss functions or the control of various error rates.These advances are detailed, for instance, in [2], [3] (p. 7), [4] and the references therein.However, another important issue concerning multiple hypothesis testing, namely the construction of simultaneous tests that yield coherent results easier to communicate to practitioners of statistical methods, has not been so deeply investigated yet, especially under the Bayesian paradigm.As a matter of fact, most traditional multiple hypothesis testing schemes do not combine statistical optimality with logical consistency.For example, [5] (p.250) presents a situation regarding the parameter, θ, of a single exponential random variable, X, in which uniformly most powerful (UMP) tests of level 0.05 for the one-sided hypothesis H p1q 0 : θ ¤ 1 and the two-sided hypothesis H p2q 0 : θ ¤ 1 θ ¥ 2, say ϕ 1 and ϕ 2 , respectively, lead to puzzling decisions.In fact, for the sample outcome X 0.7, the test ϕ 2 rejects H p2q 0 , and because H p1q 0 implies H p2q 0 , one may decide to reject H p1q 0 , as well.On the other hand, the test ϕ 1 does not reject H p1q 0 , a fact that makes a practitioner confused given these conflicting results.In this example, an inconsistency related to nested hypotheses named coherence [6] takes place.Frequently, other logical relationships one may expect from the conclusions drawn from multiple hypothesis testing, such as consonance [6] and compatibility [7], are not met either.
Although several of these properties have been deeply investigated under a frequentist hypothesis-testing framework, Bayesian literature lacks such analyses.In this work, we contribute to this discussion by examining three rational requirements in simultaneous tests under a Bayesian decision-theoretic perspective.In short, we characterize the families of loss functions that induce multiple Bayesian tests that satisfy partially such desiderata.In Section 2, we review and illustrate the concept of a testing scheme (TS), a mathematical object that assigns to each statistical hypothesis of interest a test function.In Section 3, we formalize three consistency relations one may find important to hold in simultaneous tests: coherence, union consonance and invertibility.In Section 4, we provide necessary and sufficient conditions on loss functions to ensure Bayesian tests to meet each desideratum separately, whatsoever the prior distribution for the relevant parameters is.In Section 5, we prove, under quite general conditions, the impossibility of creating multiple tests under a Bayesian decision-theoretic framework that fulfill the triplet of requisites simultaneously with respect to all prior distributions.We also explore the connection between logically-consistent Bayes tests and Bayes point estimation procedures.Final remarks and suggestions for future inquiries are presented in Section 6.All theorems are proven in the Appendix.

Testing Schemes
We start by formulating the mathematical setup for multiple Bayesian tests.For the remainder of the manuscript, the parameter space is denoted by Θ and the sample space by X .Furthermore, σpΘq and σpX q represent σ-fields of subsets of Θ and X , respectively.We consider the Bayesian statistical model pX ¢ Θ, σpX ¢ Θq, I Pq.The I P-marginal distribution of θ, namely the prior distribution for θ, is denoted by π, while π x p.q represents the posterior distribution for θ given X x, x X .Moreover, P p.|θq stands for the conditional distribution of the observable X given θ, and L x pθq represents the likelihood function at the point θ Θ generated by the sample observation x X .Finally, let Ψ be the set of all test functions, that is the set of all t0, 1u-valued measurable functions defined on X .As usual, "1" denotes the decision of rejecting the null hypothesis and "0" the decision of not rejecting or accepting it.
Next, we review the definition of a TS, a mathematical device that formally describes the idea that to each hypothesis of interest it is assigned a test function.Although the specification of the hypotheses of interest most of the times depends on the scientific problem under consideration, here, we assume that a decision-maker has to assign a test to each element of σpΘq.This assumption not only enables us to precisely define the relevant consistency properties, but it also allows multiple Bayesian testing based on posterior probabilities of the hypotheses (a deeper discussion on this issue may be found in [3] (p. 5) and [8]).Definition 1. (Testing scheme (TS)) Let the σ-field of subsets of the parameter space σpΘq be the set of hypotheses to be tested.Moreover, let Ψ be the set of all test functions defined on X .A TS is a function ϕ : σpΘq Ñ Ψ that assigns to each hypothesis A σpΘq the test ϕ A Ψ for testing A.
Thus, for A σpΘq and x X , ϕ A pxq 1 represents the decision of rejecting the hypothesis A when the datum x is observed.Similarly, ϕ A pxq 0 represents the decision of not rejecting A. We now present examples of testing schemes.
Example 1. (Tests based on posterior probabilities) Assume Θ R d and σpΘq BpR d q, the Borelians of R d .Let π be the prior probability distribution for θ.For each A σpΘq, let ϕ A : X Ñ t0, 1u be defined by: where π x p.q is the posterior distribution of θ, given x.This is the TS that assigns to each hypothesis A BpR d q the test that rejects it when its posterior probability is smaller than 1{2.
Recall that, under a Bayesian decision-theoretic perspective, a hypothesis testing for the hypothesis ) is a decision problem in which the action space is t0, 1u and the loss function L : t0, 1u ¢ Θ Ñ R satisfies: that is, L is such that the wrong decision ought to be assigned a loss at least as large as that assigned to a correct decision (many authors consider strict inequalities in Equation ( 1)).We call such a loss function a (strict) hypothesis testing loss function.
A solution of this decision problem, named a Bayes test, is a test function ϕ ¦ Ψ derived, for each sample point x X , by minimizing the expectation of the loss function L over t0, 1u with respect to the posterior distribution.That is, for each x X , ϕ ¦ pxq 1 ðñ ErLp1, θq|X xs ErLp0, θq|X xs, where ErLpd, θq|X xs ³ Θ Lpd, θqdπ x pθq, d t0, 1u.In the case of the equality of the posterior expectations, both zero and one are optimal decisions, and either of them can be chosen as ϕ ¦ pxq.
When dealing with multiple tests, one can use the above procedure for each hypothesis of interest.
Hence, one can derive a Bayes test for each null hypothesis A σpΘq considering a specified loss function L A : t0, 1u ¢ Θ Ñ R satisfying Equation (1).This is formally described in the following definition.Definition 2. (TS generated by a family of loss functions) Let pX ¢ Θ, σpX ¢ Θq, I Pq be a Bayesian statistical model.Let pL A q AσpΘq be a family of hypothesis testing loss functions, where L A : t0, 1u ¢ Θ Ñ R is the loss function for testing A σpΘq.A TS generated by the family of loss functions pL A q AσpΘq is any TS ϕ defined over σpΘq, such that, dA σpΘq, ϕ A is a Bayes test for hypothesis A with respect to π considering the loss L A .
The following example illustrates this concept.
Example 2. (Tests based on posterior probabilities) Assume the same scenario as Example 1 and that pL A q AσpΘq is a family of loss functions, such that dA σpΘq and dθ Θ, L A p0, θq Ipθ Aq and L A p1, θq Ipθ Aq, that is, L A is the 0-1 loss for A ( [5] (p.215)).The testing scheme introduced in Example 1 is a TS generated by the family of 0-1 loss functions.
The next example shows a TS of Bayesian tests motivated by different epistemological considerations (see [9,10] for details), the full Bayesian significance tests (FBST).
Example 3. (FBST testing scheme) Let Θ R d , σpΘq BpR d q and f p.q be the prior probability density function (pdf) for θ.Suppose that, for each x X , there exists f p.|xq, the pdf of the posterior distribution of θ, given x.For each hypothesis A σpΘq, let: be the set tangent to the null hypothesis, and let ev x pAq 1 ¡ π x pT A x q be the Pereira-Stern evidence value for A (see [11] for a geometric motivation).One can define a TS ϕ by: ϕ A pxq I pev x pAq ¤ cq , dA σpΘq and dx X , in which c r0, 1s is fixed.In other words, one does not reject the null hypothesis when its evidence is larger than c.
We end this section by defining a TS generated by a point estimation procedure, an intuitive concept that plays an important role in characterizing logically-consistent simultaneous tests.Definition 3. (TS generated by a point estimation procedure) Let δ : X ÝÑ Θ be a point estimator for θ ( [5] (p.296)).The TS generated by δ is defined by: Hence, the TS generated by the point estimator δ rejects hypothesis A after observing x if, and only if, the point estimate for θ, δpxq, is not in A.
Example 4. (TS generated by a point estimation procedure) Let Θ R, σpΘq PpΘq and X 1 . . . ,X n |θ i.i.d.N pθ, 1q.The TS generated by the sample mean, s X, rejects A σpΘq when x is observed if s x A.

The Desiderata
In this section, we review three properties one may expect from simultaneous test procedures: coherence, invertibility and union consonance.

Coherence
When a hypothesis is tested by a significance test and is not rejected, it is generally agreed that all hypotheses implied by that hypothesis (its "components") must also be considered as non-rejected.
-K. R. Gabriel [6] The first property concerns nested hypotheses and was originally defined by [6] 1.These numbers were taken from a study presented by [12] that had the aim of verifying the hypothesis that subunits of the gene GABA A contribute to a condition known as methamphetamine use disorder.Here, the set of all possible genotypes is G tAA, AB, BBu.Let γ pγ AA , γ AB , γ BB q, where γ i is the probability that an individual from the case group has genotype i.Similarly, let π pπ AA , π AB , π BB q, where π i is the probability that an individual of control group has genotype i.
In this context, two hypotheses are of interest: the hypothesis that the genotypic proportions are the same in both groups, H G 0 : γ π, and the hypothesis that the allelic proportions are the same in both groups The p-values obtained using chi-square tests for these hypotheses are, respectively, 0.152 and 0.069.Hence, at the level of significance α 10%, the TS given by chi-square tests rejects H A 0 , but does not reject H G 0 .That is, the TS leads a practitioner to believe that the allelic proportions are different in both groups, but it does not suggest any difference between the genotypic proportions.This is absurd!If the allelic proportions are not the same in both groups, the genotypic proportions cannot be the same either.Indeed, if the latter were the same, then γ i π i , di G, and hence, θ H A 0 .This example is further discussed in [8,13].
Coherence is by far the most emphasized logical requisite for simultaneous test procedures in the literature.It is often regarded as a sensible property by both theorists and practitioners of statistical methods who perceive a hypothesis test as a two-fold (accept/reject) decision problem.On the other hand, adherents to evidence-based approaches to hypothesis testing [15] do not see the need for coherence.Under the frequentist approach to hypothesis testing, the construction of coherent procedures is closely associated with the so-called closure methods [16,17].Many results on coherent classical tests are shown in [6,17], among others.On the other hand, coherence has not been deeply investigated from a Bayesian standpoint yet, except for [18], who relate coherence with admissibility and Bayesian optimality in certain situations of finitely many hypotheses of interest.In Section 4, we provide a characterization of coherent testing schemes under a decision-theoretic framework.

Invertibility
There is a duality between hypotheses and alternatives which is not respected in most of the classical hypothesis-testing literature.(...) suppose that we decide to switch the names of alternative and hypothesis, so that Ω H becomes Ω A , and vice versa.Then we can switch tests from φ to ψ 1 ¡ φ and the "actions" accept and reject become switched.
-M.J. Schervish [5] (p.216) The duality mentioned in the quotation above is formally described in the next definition.Definition 5. (Invertibility) A testing scheme ϕ satisfies invertibility if: In other words, it is irrelevant to decision-making which hypothesis is labeled as null and which is labeled as alternative.
Unlike coherence, there is no consensus among statisticians on how reasonable invertibility is.While it is supported by many decision-theorists, invertibility is usually discredited by advocates of the frequentist theory owing to the difference between the interpretations of "not reject a hypothesis" and "accept a hypothesis" under various epistemological viewpoints (the reader is referred to [7] for a discussion on this distinction).As a matter of fact, invertibility can also be seen, from a logic perspective, as a version of the law of the excluded middle, which itself represents a gap between schools of logic ( [19] (p.32)).In spite of the controversies on invertibility, it seems to be beyond any argument the fact that the absence of invertibility in multiple tests may lead a decision-maker to be puzzled by senseless conclusions, such as the simultaneous rejections of both a hypothesis and its alternative.The following example illustrates this point.Example 6. Suppose that X|θ Normalpθ, 1q, and consider that the parameter space is Θ t¡3, 3u.
Assume one wants to test the following null hypotheses: The Neyman-Pearson tests for these hypotheses have the following critical regions, at the level 5%, respectively: tx R : x 1.35u and tx R : x ¡ ¡1.35u.
Hence, if we observe x ¡0.5, we reject both H A 0 and H B 0 , even though The testing schemes of Examples 2 and 4 satisfy invertibility.In Example 4, it is straightforward to verify this.In Example 2, it follows essentially from the equivalence π x pAq 1{2 ô π x pA c q ¡ 1{2.
If π x pAq $ 1{2 for each sample x and for all A σpΘq, the unique TS generated by the 0-1 loss functions satisfies invertibility.Otherwise, there is a testing scheme generated by such losses that is still in line with this property.Indeed, for any A σpΘq and x 0 X , such that π x 0 pAq 1{2, the decision of rejecting A (not rejecting A c ) after observing x 0 has the same expected loss as the decision of not rejecting (rejecting) it.Thus, among all testing schemes generated by the 0-1 loss functions, which are all equivalent from a decision-theoretic point of view ( [20] (p.123)), a decision-maker can always choose a TS ϕ I , such that ϕ I A c px 0 q 1 ¡ϕ I A px 0 q for all A σpΘq, and x 0 X , such that π x 0 pAq 1{2.
Such a TS ϕ I meets invertibility.

Consonance
... a test for p iI H i q c versus p iI H i q may result in rejection which then indicates that at least one of the hypotheses H i , i I, may be true.
-H. Finner and K. Strassburger [21] The third property concerns two hypotheses, say A and B, and their union, A B. It is motivated by the fact that in many cases, it seems reasonable that a testing scheme that retains the union of these hypotheses should also retain at least one of them.This idea is generalized in Definition 6.
Definition 6. (Union Consonance) A TS ϕ satisfies the finite (countable) union consonance if for all finite (countable) set of indices I, In other words, if we retain the union of the hypotheses iI A i , we should not reject at least one of the A i 's.
There are several testing schemes that meet union consonance.For instance, TSs generated by point estimation procedures, TSs of Aitchison's confidence-region tests [22] and FBST TSs (under quite general conditions; see [8]) satisfy both finite and countable union consonance.
Although union consonance may not be considered as appealing as coherence for simultaneous test procedures, it was hinted at in a few relevant works.For instance, the interpretation given by [21] on the final joint decisions derived from partial decisions implicitly suggests that union consonance is reasonable: they suggest one should consider B : A:ϕ A pxq1 A to be the set of all parameter values rejected by the simultaneous procedure at hand when x is observed.Under this reading, it seems natural to expect that ϕ B pxq 1, which is exactly what the union consonance principle states.As a matter of fact, the general partitioning principle proposed by these authors satisfies union consonance.It should also be mentioned that union consonance, together with coherence, plays a key role in the possibilistic abstract belief calculus [23].In addition, an evidence-based approach detailed in [24] satisfies both consonance and invertibility.
We end this section by stating a result derived from putting these logical requirements together.
Theorem 1.Let Θ be a countable parameter space and σpΘq PpΘq.Let ϕ be a testing scheme defined on σpΘq.The TS ϕ satisfies coherence, invertibility and countable union consonance if, and only if, there is a point estimator δ : X Ñ Θ, such that ϕ is generated by δ.
Theorem 1 is also valid for finite union consonance with the obvious adaptation.

A Bayesian Look at Each Desideratum
In the previous section, we provided several examples of testing schemes satisfying some of the logical properties reviewed therein.In particular, a testing scheme generated by the family of 0-1 loss functions (Example 2) was shown to fulfill both coherence and invertibility.However, not all families of loss functions generate a TS meeting any of these requisites, as is shown in the examples below.
Example 7. Suppose that X|θ Bernoullipθq and that one is interested in testing the null hypotheses:

and H B
0 : θ ¤ 0.5.Furthermore, assume θ U nif ormp0, 1q a priori and that he uses the loss functions from Table 2

State of Nature
Intuitively, incoherence takes place because the loss of falsely rejecting H A 0 is three-times as large as the loss of falsely rejecting H B 0 , while the corresponding errors of Type II are of the same magnitude.Hence, these loss functions reveal that the decision-maker is more reluctant to reject H A 0 than to reject H B 0 in such a way that he only needs little evidence to accept H A 0 (posterior probability greater than 1/7) when compared to the amount of evidence needed to accept H B 0 (posterior probability greater than 1/3).Thus, it is not surprising at all that in this case, the tests do not cohere for some priors.

State of Nature
The absence of invertibility is somewhat expected here, because the degree to which the decision-maker believes an incorrect decision of choosing H B 0 to be more serious than an incorrect decision of choosing H B c 0 is not the same whether H B 0 is regarded as the "null" or the "alternative" hypothesis.More precisely, while the decision-maker assigns a loss to the error of Type I that is the double of the one assigned to the error of Type II when testing the null hypothesis H B 0 , he evaluates the loss of falsely accepting H B c 0 to be four-times (not twice!) as large as that of falsely rejecting it when H B c 0 is the null hypothesis.
The examples we have examined so far give rise to the question: from a decision-theoretic perspective, what conditions must be imposed on a family of loss functions so that the resultant Bayesian testing scheme meets coherence (invertibility)?Next, we offer a solution to this question.We first give a definition in order to simplify the statement of the main results of this section.Definition 7. (Relative loss) Let L A be a loss function for testing the hypothesis A σpΘq.The function ∆ A : Θ Ñ R defined by: In short, the relative loss measures the difference between losses of taking the wrong and the correct decisions.Thus, the relative loss of any hypothesis testing loss function is always non-negative.
A careful examination of Example 7 hints that in order to obtain coherent tests, the "larger" (the "smaller") the null hypothesis of interest is, the more cautious about falsely rejecting (accepting) it the decision-maker ought to be.This can be quantified as follows: for hypotheses A and B, such that A B and with corresponding hypothesis testing loss functions L A and L B , if θ 1 A, then ∆ B pθ 1 q should be at least as large as ∆ A pθ 1 q.Similarly, if θ 2 B c , then ∆ B pθ 2 q should be at most ∆ A pθ 2 q.Such conditions are also appealing, since it seems reasonable that greater relative losses should be assigned to greater "distances" between the parameter and the wrong decision.For instance, if θ A (and consequently, θ B), the rougher error of rejecting B should be penalized more heavily than the error of rejecting A;  These conditions, namely: are sufficient for coherence.As a matter of fact, Theorem 2 states that the weaker condition: is necessary and sufficient for a family of hypothesis testing loss functions to induce a coherent testing scheme with respect to each prior distribution for θ.Henceforward, we assume that EpL A pd, θq|xq V, for all A σpΘq, d t0, 1u and x X .Theorem 2. Let pL A q AσpΘq be a family of hypothesis testing loss functions.Suppose that for all θ 1 , θ 2 Θ, there is x X , such that L x pθ 1 q, L x pθ 2 q ¡ 0.Then, for all prior distributions π for θ, there exists a testing scheme generated by pL A q AσpΘq with respect to π that is coherent if, and only if, pL A q AσpΘq is such that for all A, B σpΘq with A B: Notice that the "if" part of Theorem 2 still holds for families of hypothesis testing loss functions that depend also on the sample.Theorem 2 characterizes, under certain conditions, all families of loss functions that induce coherent tests, no matter what the decision-maker's opinion (prior) on the unknown parameter is.Although the result of Theorem 2 is not properly normative, any Bayesian decision-maker can make use of it to prevent himself from drawing incoherent conclusions from multiple hypothesis testing by checking whether his personal losses satisfy the condition in Equation (2).
Many simple families of loss functions generate coherent tests, as we illustrate in Examples 9 and 10.
Example 9. Consider, for each A σpΘq, the loss function L A in Table 4 to test the null hypothesis A, in which λ : σpΘq Ñ R is any finite measure, such that λpΘq ¡ 0. This family of loss functions satisfies the condition in Equation (2) for coherence as for all A, B σpΘq, such that A B, and for all θ 1 A and θ 2 B c , ∆ A pθ 1 q λpAq, ∆ B pθ 2 q λpB c q, ∆ A pθ 2 q λpA c q and ∆ B pθ 1 q λpBq.Table 4. Loss function L A for testing A.

State of Nature
then the family pL A q AσpΘq will induce a coherent TS for each prior for θ.Example 10.Assume Θ is equipped with a distance, say d.Define, for each A σpΘq the loss function L A for testing A by: L A p0, θq d ¦ pθ, Aq and L A p1, θq d ¦ pθ, A c q , where d ¦ pθ, Aq inf aA dpθ, aq is the distance between θ Θ and A. For A, B σpΘq, such that A B, and for θ 1 A and θ 2 B c , ∆ A pθ 1 q d ¦ pθ 1 , A c q, ∆ B pθ 2 q d ¦ pθ 2 , Bq, ∆ A pθ 2 q d ¦ pθ 2 , Aq and ∆ B pθ 1 q d ¦ pθ 1 , B c q.These values satisfy Equation (2) from Theorem 2. Hence, families of loss functions based on distances as the above generate Bayesian coherent tests.
Next, we characterize Bayesian tests with respect to invertibility.In order to obtain TSs that meet invertibility, it seems reasonable that when the null and alternative hypotheses are switched, the relative losses ought to remain the same.That is to say, when testing the null hypothesis A, the relative loss at each point θ Θ, ∆ A pθq, should be equal to the relative loss ∆ A c pθq when A c is the null hypothesis instead.This condition is sufficient, but not necessary for a family of loss functions to induce tests fulfilling this logical requisite with respect to all prior distributions.In Theorem 3, however, we provide necessary and sufficient conditions for invertibility.Theorem 3. Let pL A q AσpΘq be a family of hypothesis testing loss functions.Suppose that for all θ 1 , θ 2 Θ, there is x X , such that L x pθ 1 q, L x pθ 2 q ¡ 0.Then, for all prior distributions π for θ, there exists a testing scheme generated by pL A q AσpΘq with respect to π that satisfies invertibility if, and only if, pL A q AσpΘq is such that for all A σpΘq: Condition Equation ( 3) is equivalent (for strict hypothesis testing loss functions) to impose, for each A σpΘq, that the function ∆ A p.q ∆ A c p.q to be constant over Θ.We should mention that the "if" part of Theorem 3 still holds for hypothesis testing loss functions satisfying (Equation ( 3)) that also depend on the sample x.
The families of loss functions introduced in Examples 9 and 10 satisfy (Equation ( 3)).Thus, such families of losses ensure the construction of simultaneous Bayes tests that are in conformity with both coherence and invertibility for all prior distributions on σpΘq.Thus, if one believes these (two) logical requirements to be of primary importance in multiple hypothesis testing, he can make use of any of these families of loss functions to perform tests satisfactorily.Other simple loss functions also lead to TSs that meet invertibility: for instance, any family of 0-1-c loss functions for which c A c 1{c A for all A σpΘq leads to invertible TSs.
We end this section by examining union consonance under a decision-theoretic point of view.From Definition 6, it appears that a necessary condition for the derivation of consonant tests is that "smaller" ("larger") null hypotheses ought to be assigned greater losses for false rejection (acceptance).
More precisely, for A, B σpΘq, if θ 1 A B, then it seems that either ∆ AB pθ 1 q ¤ ∆ A pθ 1 q or ∆ AB pθ 1 q ¤ ∆ B pθ 1 q should hold.If θ 2 pA Bq c , then it is reasonable that either ∆ AB pθ 2 q ¥ ∆ A pθ 2 q or ∆ AB pθ 2 q ¥ ∆ B pθ 2 q.The next theorem shows that this is nearly the case.However, it is still unknown whether sufficient conditions for union consonance are determinable.Theorem 4. Let pL A q AσpΘq be a family of hypothesis testing loss functions.Suppose that for all θ 1 , θ 2 Θ, there is x X , such that L x pθ 1 q, L x pθ 2 q ¡ 0. If for all prior distribution π for θ, there exists a testing scheme generated by pL A q AσpΘq with respect to π that satisfies finite union consonance, then pL A q AσpΘq is such that for all A, B σpΘq and for all θ 1 A B, θ 2 pA Bq c , either ∆ AB pθ 1 q∆ A pθ 2 q ¤ ∆ AB pθ 2 q∆ A pθ 1 q or ∆ AB pθ 1 q∆ B pθ 2 q ¤ ∆ AB pθ 2 q∆ B pθ 1 q.

Putting the Desiderata Together
In Section 4, we showed that there are infinitely many families of loss functions that induce, for each prior distribution for θ, a TS that satisfies both coherence and invertibility (Examples 9 and 10).However, requiring the three logical consistency properties we presented to hold simultaneously with respect to all priors is too restrictive: under mild conditions, no TS constructed under a Bayesian decision-theoretic approach to hypothesis testing fulfills this, as stated in the next theorem.
Theorem 5. Assume that Θ and σpΘq are such that |Θ| ¥ 3 and that there is a partition of Θ composed of three nonempty measurable sets.Assume also that for all triplet θ 1 , θ 2 , θ 3 Θ, there is x X , such that L x pθ i q ¡ 0 for i 1, 2, 3.Then, there is no family of strict hypothesis testing loss functions that induces, for each prior distribution for θ, a testing scheme satisfying coherence, invertibility and finite union consonance.
Theorem 5 states that Bayesian optimality (based on standard loss functions that do not depend on the sample) cannot be combined with complete logical consistency.This fact can lead one to wonder whether such properties are indeed sensible in multiple hypothesis testing.The following result shows us that the desiderata are in fact reasonable in the sense that a TS meeting these requirements does correspond to the optimal tests of some Bayesian decision-makers.We return to this point in the concluding remarks.Theorem 6.Let Θ be a countable (finite) parameter space, σpΘq PpΘq, and X be a countable sample space.Let ϕ be a testing scheme that satisfies coherence, invertibility and countable (finite) union consonance.Then, there exist a probability measure µ over PpΘ ¢ X q and a family of strict hypothesis testing loss functions pL A q AσpΘq , such that ϕ is generated by pL A q AσpΘq with respect to the µ-marginal distribution of θ.
We end this section by associating logically-consistent Bayesian hypothesis testing with Bayes point estimation procedures in case both Θ and X are finite.This relationship is characterized in Theorem 7.
Theorem 7. Let Θ and X be finite sets and σpΘq PpΘq.Let ϕ be the testing scheme generated by the point estimator δ : X Ñ Θ. Suppose that for all x X , L x pδpxqq ¡ 0.
(a) If there exist a probability measure π : σpΘq Ñ r0, 1s for θ, with πpδpxqq ¡ 0 for all x X , and a loss function L : Θ ¢ Θ Ñ R , satisfying Lpθ, θq 0 and Lpd, θq ¡ 0 for d $ θ, such that δ is a Bayes estimator for θ generated by L with respect to π, then there is a family of hypothesis testing loss functions pL A q AσpΘq , L A : t0, 1u ¢ pΘ ¢ X q Ñ R for each A σpΘq, such that ϕ is generated by pL A q AσpΘq with respect to π.(b) If there exist a probability measure π : σpΘq Ñ r0, 1s for θ, with πpδpxqq ¡ 0 for all x X , and a family of strict hypothesis testing loss functions pL A q AσpΘq , L A : t0, 1u ¢ Θ Ñ R for each A σpΘq, such that ϕ is generated by pL A q AσpΘq with respect to π, then there is a loss function L : Θ ¢ Θ Ñ R , with Lpθ, θq 0 and Lpd, θq ¡ 0 for d $ θ, such that δ is a Bayes estimator for θ generated by L with respect to π.
Theorem 7 ensures that multiple Bayesian tests that fulfill the desiderata cannot be separated from Bayes point estimation procedures.One may find in Theorem 7, Part (a), a decision-theoretic justification for performing simultaneous tests by means of a Bayes point estimator.However, the optimality of such tests is derived under very restrictive conditions, as the underlying loss functions depend both on the sample and on a point estimator.This fact reinforces that one can reconcile statistical optimality and logical consistency in multiple tests only in very particular cases.We should also emphasize that, under the conditions of Part (a), if, in addition, πpθq ¡ 0 for all θ Θ, then, for all A σpΘq, ϕ A is an admissible test for A with regard to L A (the standard proof of this result developed for losses that do not depend on the sample also works here).The second part of Theorem 7 states that if a Bayesian testing scheme meets coherence, invertibility and finite union consonance, then the point estimator that generates it cannot be devoid of optimality: it must be a Bayes estimator for specific loss functions.Example 11 illustrates the first part of this theorem.
Example 11.Assume that Θ tθ 1 , θ 2 , . . ., θ k u and X is finite.Assume also that there is a maximum likelihood estimator (MLE) for θ, δ M L : X Ñ Θ, such that L x pδ M L pxqq ¡ 0, for all x X .Then, the testing scheme generated by δ M L is a TS of Bayes tests.Indeed, when Θ is finite, an MLE for θ is a Bayes estimator generated by the loss function Lpd, θq Ipd $ θq, d, θ Θ, with respect to the uniform prior over Θ (that is, δ M L pxq corresponds to a mode of the posterior distribution π x , for each x X ).Consequently (recall that |Θ| k), π x pδ M L pxqq ¥ 1{k and ErLpδ M L pxq, θq|xs 1 ¡ π x pδ M L pxqq, for each x X .Thus, as g : p0, 1s Ñ R given by gptq p1 ¡ tq{t is strictly decreasing.By Theorem 7, it follows that the TS generated by the MLE δ M L is a Bayesian TS generated by (for instance) the family of loss functions pL A q AσpΘq given, for each A σpΘq, by L A p1, pθ, xqq 0 and L A p0, pθ, xqq 1, for θ A c , and L A p0, pθ, xqq 0 and L A p1, pθ, xqq kI A pδ M L pxqq p1{kqI A c pδ M L pxqq, for θ A.
It is worth mentioning that the development of Theorem 7(a) and Example 11 is in a sense related to the optimality of least relative surprise estimators under prior-based loss functions [24] (Section 2).

Conclusions
While several studies on frequentist multiple tests deal with the question of seeking for a balance between statistical optimality and logical consistency, this issue has not been addressed yet under a decision-theoretic standpoint.For this reason, in this work, we examine simultaneous Bayesian hypothesis testing with respect to three rational properties: coherence, invertibility and union consonance.Briefly, we characterize the families of loss functions that yield Bayes tests meeting each of these requisites separately, whatever the prior distribution for the relevant parameter is.These results not only shed some light on when each of these relationships may be considered to be sensible for a given scientific problem, but they also serve as a guide for a Bayesian decision-maker aiming at performing tests in line with the requirement he finds more important.In particular, this can be done through the usage of the loss functions described in the paper.
We also explore how far one can go by putting these properties together.We provide examples of fairly intuitive loss functions that induce testing schemes satisfying both coherence and invertibility, no matter what one's prior opinion on the parameter is.On the other hand, we prove that no family of reasonable loss functions generates Bayes tests that respect the logical properties as a whole with respect to all priors, although any testing scheme meeting the desiderata corresponds to the optimal tests of several Bayesian decision-makers.
Finally, we discuss the relationship between logically-consistent Bayesian hypothesis testing and Bayes point estimations procedures when both the parameter space and the sample space are finite.We conclude that the point estimator generating a testing scheme fulfilling the rational properties is inevitably and unavoidably a Bayes estimator for certain loss functions.Furthermore, performing logically-consistent procedures by means of a Bayes estimator is one's best approach towards multiple hypothesis testing only under very restrictive conditions in which the underlying loss functions depend not only on the decision to be made and the parameter as usual, but also on the observed sample.See [24][25][26] for some examples of such loss functions.That is, a more complex framework is needed to combine Bayesian optimality with logical consistency.This fact and the impossibility result of Theorem 5 corroborate the thesis that full rationality and statistical optimality rarely can be combined in simultaneous tests.In practice, this suggests that when testing hypotheses at once, a practitioner may abandon in part the desiderata so as to preserve statistical optimality.This is further discussed in [8].
Several issues remain open, among which we mention three.First, the extent to which the results derived in this work can be generalized to infinite (continuous) parameter spaces is an important problem from both theoretical and practical aspects.Furthermore, the consideration of different decision-theoretic approaches to hypothesis testing, such as the "agnostic" tests with three-fold action spaces proposed by [27], may bring new insight into which logical properties may be expected, not only in the current, but also in alternative frameworks.In epistemological terms, one may be concerned with the question of whether multiple hypothesis testing is the most adequate way to draw inferences about a parameter of interest from data given the incompatibility between full logical consistency and the achievement of statistical optimality.As a matter of fact, many Bayesians regard the whole posterior distribution as the most complete inference one can make about the unknown parameter.These analyses may contribute to better decision-making.the corresponding posterior distribution given x I is π ¦ x I pθ 1 q α 0 and π ¦ x I pθ 2 q 1 ¡ α 0 .Let ϕ ¦ be any TS generated by pL A q AσpΘq with respect to π ¦ .Thus, From Equation (4), we have ϕ ¦ A px I q 0 and ϕ ¦ B px I q 1.Therefore, there is a prior distribution π ¦ for θ with respect to which any TS generated by pL A q AσpΘq is not coherent.
We now prove the "if" part.We suppose that the family pL A q AσpΘq satisfies the condition that for all A, B σpΘq with A B, ∆ A pθ 1 q∆ B pθ 2 q ¤ ∆ A pθ 2 q∆ B pθ 1 q, dθ 1 A, dθ 2 B c .Integrating (with respect to θ 1 ) over A with respect to any probability measure P , we obtain: Similarly, integration (with respect to θ 2 ) over B c with respect to the same measure P yields: Now, let ϕ be a testing scheme generated by the family pL A q AσpΘq .For A, B σpΘq with A B and x X , where π x p.q denotes the posterior distribution of θ given X x.Thus, Multiplying the last inequality by ³ B c ∆ B pθqdπ x pθq ¥ 0, we get: From inequality Equation ( 5), it follows that: we have that: and, consequently, If this integral is equal to zero, then both zero and one are optimal solutions, and we can choose the decision zero as ϕ B pxq in order to ensure that ϕ B pxq ¤ ϕ A pxq. Hence, with respect to each prior π, there is a TS generated by pL A q AσpΘq that is coherent.

C. Proof of Theorem 3
The proof is analogous to that of Theorem 2. First, we prove the necessary condition by the contrapositive.Suppose that there are A σpΘq and θ 1 A and θ 2 A c , such that: Assume ∆ A pθ 1 q∆ A c pθ 2 q ∆ A c pθ 1 q∆ A pθ 2 q (the other case is developed in the same way), which implies that ∆ A c pθ 1 q ¡ 0 and ∆ A pθ 2 q ¡ 0. Adding ∆ A c pθ 2 q∆ A pθ 2 q to both sides of the inequality, we easily obtain that: Thus, there is α 0 p0, 1q, such that: In addition, there is x I X , such that L x I pθ 1 q, L x I pθ 2 q ¡ 0. For the prior distribution π ¦ for θ given by: π ¦ pθ 1 q α 0 L x I pθ 2 q α 0 L x I pθ 2 q p1 ¡ α 0 qL x I pθ 1 q and π ¦ pθ 2 q 1 ¡ π ¦ pθ 1 q , the posterior distribution given x I is π ¦ x I pθ 1 q α 0 and π ¦ x I pθ 2 q 1 ¡ α 0 .Let ϕ ¦ be any TS generated by pL A q AσpΘq with respect to π ¦ .Thus, ϕ ¦ A px I q 0 , if α 0 ¡ ∆ A pθ 2 q ∆ A pθ 1 q ∆ A pθ 2 q and ϕ ¦ A c px I q 0 , if α 0 From Equation (6), we have ϕ ¦ A px I q 1 and ϕ ¦ A c px I q 1.Therefore, there is a prior distribution π ¦ for θ with respect to which any TS generated by pL A q AσpΘq does not meet invertibility.Now, we prove the sufficiency.Suppose that for all A σpΘq: Integrating (with respect to θ 2 ) over the set A c with respect to any probability measure P defined on σpΘq, we have: Similarly, integrating (with respect to θ 1 ) over A, we get: Let ϕ be a TS generated by pL A q AσpΘq .If ϕ A pxq 0, then: Multiplying both sides by From Equation (7), it follows that: Thus, If the integral is zero, then we can choose ϕ A c pxq 1, so as to obtain ϕ A c pxq 1 ¡ ϕ A pxq.Similarly, we prove that if ϕ A pxq 1, then there is a Bayes test for A c , ϕ A c , generated by L A c , such that ϕ A c pxq 0. Consequently, there is a TS generated by pL A q AσpΘq that satisfies invertibility.

D. Proof of Theorem 4
Suppose that there are A, B σpΘq, θ 1 A B and θ 2 pA Bq c such that both: ∆ AB pθ 1 q∆ A pθ 2 q ¡ ∆ AB pθ 2 q∆ A pθ 1 q and ∆ AB pθ 1 q∆ B pθ 2 q ¡ ∆ AB pθ 2 q∆ B pθ 1 q hold, from which it follows that ∆ AB pθ 1 q ¡ 0, ∆ A pθ 2 q ¡ 0 and ∆ B pθ 2 q ¡ 0. Proceeding as in the previous proofs, we obtain that: Thus, there is α 0 p0, 1q such that: In addition, there is x I X such that L x I pθ 1 q, L x I pθ 2 q ¡ 0. For the prior distribution π ¦ for θ given by: π ¦ pθ 1 q α 0 L x I pθ 2 q α 0 L x I pθ 2 q p1 ¡ α 0 qL x I pθ 1 q and π ¦ pθ 2 q 1 ¡ π ¦ pθ 1 q , the posterior distribution is π ¦ x I pθ 1 q α 0 and π ¦ x I pθ 2 q 1 ¡ α 0 .Let ϕ ¦ be any TS generated by pL A q AσpΘq with respect to π ¦ .Next, we consider three cases: for any C tA, B, A Bu. Thus, we have ϕ ¦ A px I q 1, ϕ ¦ B px I q 1 and ϕ ¦ AB px I q 0; (ii) if θ A, then: pθ 2 q and ϕ ¦ AB px I q 0 , if α 0 ¡ ∆ AB pθ 2 q ∆ AB pθ 1 q ∆ AB pθ 2 q , and: Thus, ϕ ¦ A px I q 1, ϕ ¦ B px I q 1 and ϕ ¦ AB px I q 0; (iii) if θ B, a development similar to that of Case (ii) yields the same results: ϕ ¦ A px I q 1, ϕ ¦ B px I q 1 and ϕ ¦ AB px I q 0.
Therefore, in any case, there is a prior distribution π ¦ for θ with respect to which no TS generated by pL A q AσpΘq meets finite union consonance, concluding the proof.

E. Proof of Theorem 5
The proof of Theorem 5 consists of verifying the inexistence of such a family of loss functions that generates Bayes tests satisfying the desiderata with respect to all priors concentrated on three points in Θ (of course, there will not be such a family satisfying these requisites with respect to all priors over σpΘq).
Let tA 1 , A 2 , A 3 u be a measurable partition of Θ and θ 1 , θ 2 , θ 3 Θ, such that θ i A i , i 1, 2, 3. First, notice that for all x X , such that L x pθ i q ¡ 0 for i 1, 2, 3, there is a one-to-one correspondence between prior and posterior distributions concentrated on tθ 1 , θ 2 , θ 3 u.Indeed, for all pα 1 , α 2 , α 3 q A tpa, b, cq R 3 : a b c 1u and x X , such that L x pθ i q ¡ 0 for i 1, 2, 3, there is a unique prior distribution for θ, π, such that the corresponding posterior distribution given x, π x , satisfies π x pθ i q α i , i 1, 2, 3, namely: Henceforth, we will refer to the above posterior by pα 1 , α 2 , α 3 q for short.Let pL A q AσpΘq be any family of strict hypothesis testing loss functions.For each pα 1 , α 2 , α 3 q A, the difference between the posterior risk of accepting H piq 0 : θ A i and that of rejecting it is given by: » where ∆ A i pθ j q L A i p0, θ j q ¡ L A i p1, θ j q (note that ∆ A i pθ j q ¡ 0, if i $ j, while ∆ A i pθ i q 0).In order to evaluate the tests for the hypotheses H p1q 0 , H p2q 0 and H p3q 0 with respect to all posterior distributions concentrated on tθ 1 , θ 2 , θ 3 u, we consider the transformation T : A Ñ R 3 defined by: Thus, T assigns to each posterior pα 1 , α 2 , α 3 q A the differences between the risks of accepting H piq 0 and of rejecting it, i 1, 2, 3.It is easy to verify that B T pAq tTpα 1 , α 2 , α 3 q : pα 1 , α 2 , α 3 q Au is a convex set.Indeed, B is a triangle (see Figure 2) with vertices P 1 T p1, 0, 0q p∆ A 1 pθ 1 q, ∆ A 2 pθ 1 q, ∆ A 3 pθ 1 qq, P 2 T p0, 1, 0q p∆ A 1 pθ 2 q, ∆ A 2 pθ 2 q, ∆ A 3 pθ 2 qq and P 3 T p0, 0, 1q p∆ A 1 pθ 3 q, ∆ A 2 pθ 3 q, ∆ A 3 pθ 3 qq (these points are not aligned owing to the restrictions on the quantities ∆ A i pθ j q, [14]).Now, we turn to the main argument of the proof.By Theorem 4.3 from [8], it is necessary for a Bayesian testing scheme to satisfy the logical requirements with respect to all priors over σpΘq that exactly one of the A I i s is accepted for each vector of probabilities pα 1 , α 2 , α 3 q.Geometrically, such a necessary condition is equivalent to the triangle B to be contained in the union of the octants that comprise the triplets with only one negative coordinate, namely R ¡ ¢ R ¢ R , R ¢ R ¡ ¢ R and R ¢ R ¢ R ¡ .However, this is impossible.To verify this fact, we consider three cases (Figure 3 illustrates the projection of B over the plane w tpu, v, 0q : u, v Ru in each of these cases): ( , then the projection of the line segment joining P 1 and P 2 over the plane w intersects the (third) quadrant R ¡ ¢ R ¡ ¢ t0u (see the first graphic in Figure 3).Thus, there is γ p0, 1q, such that γ∆ A i pθ 1 q p1¡γq∆ A i pθ 2 q 0, i 1, 2. As γP 1 p1¡γqP 2 B, there is a posterior pα 1 , α 2 , α 3 q concentrated on tθ 1 , θ 2 , θ 3 u with respect to which any TS generated by pL A q AσpΘq does not reject both A 1 and A 2 and, therefore, does not respect coherence, invertibility and finite union consonance; (ii) if ∆ A 1 pθ 1 q∆ A 2 pθ 2 q ∆ A 1 pθ 2 q∆ A 2 pθ 1 q, then the projection of the line segment joining P 1 and P 2 over w intersects the origin p0, 0, 0q (see the second graphic in Figure 3).Thus, there is t 0 ¡ 0, such that the point P 0 p0, 0, t 0 q B. Considering now the line segment joining P 0 and P 3 , it is easily seen that for any γ p ¡∆ A 3 pθ 3 q t 0 ¡∆ A 3 pθ 3 q , 1q, γ0 p1 ¡ γq∆ A 1 pθ 3 q ¡ 0,γ0 p1 ¡ γq∆ A 2 pθ 3 q ¡ 0 and γt 0 p1 ¡ γq∆ A 3 pθ 3 q ¡ 0. As γP 0 p1 ¡ γqP 3 B, there is a posterior distribution with respect to which any TS generated by pL A q AσpΘq rejects A 1 , A 2 and A 3 and, therefore, does not satisfy the logical consistency properties all together; (iii) if ∆ A 1 pθ 1 q∆ A 2 pθ 2 q ∆ A 1 pθ 2 q∆ A 2 pθ 1 q, then the projection of the above-mentioned segment over w intersects the (first) quadrant R ¢R ¢t0u (third graphic in Figure 3).Thus, there is γ p0, 1q such that γ∆ A i pθ 1 q p1 ¡ γq∆ A i pθ 2 q ¡ 0, i 1, 2. As γP 1 p1 ¡ γqP 2 B, there is a posterior pα 1 , α 2 , α 3 q concentrated on tθ 1 , θ 2 , θ 3 u with respect to which any TS generated by pL A q AσpΘq rejects A 1 , A 2 and A 3 and, consequently, does not meet the desiderata.

F. Proof of Theorem 6
Let ϕ be a TS satisfying coherence, invertibility and countable union consonance.From Theorem 1, there is a unique point estimator δ : X Ñ Θ, such that for all A σpΘq and x X , ϕ A pxq Ipδpxq Aq.For each x X , define µ x : σpΘq Ñ R by: that is, µ x is the probability measure degenerate at the point δpxq [14].Furthermore, let µ 0 be any probability measure defined on PpX q.Defining µ : PpΘ ¢ X q Ñ R by: µpBq pθ,xqB µ 0 ptxuqµ x ptθuq , B PpΘ ¢ χq , it is immediate that µ is a probability measure and that µ x is the conditional distribution of θ given X x, for each x X .Next, let pL A q AσpΘq be any family of strict hypothesis testing loss functions.Let ϕ ¦ be a testing scheme generated by pL A q AσpΘq with respect to the µ-marginal distribution of θ.Let us verify that ϕ ¦ coincides with ϕ.Indeed, for x X and A σpΘq, we have: PpΘ ¢ X q and families of strict hypothesis testing loss functions that yield the result of Theorem 6.For instance, considering for each x X , a conditional probability measure µ I x , such that µ I x pδpxqq ¡ 1{2 and µ

I
x pθq ¡ 0, for all θ Θ, together with the family of 0-1 loss functions, one will obtain a Bayesian TS that coincides with ϕ, as well (see [14] for the details).

G. Proof of Theorem 7
To prove Part (a), we define a family of loss functions that generates Bayesian testing schemes satisfying both coherence and invertibility with respect to all prior distributions for θ, which implies, by Theorem 3.1 from [28], that, for each sample point, at most one hypothesis of each partition of Θ is not rejected.Next, we prove that, for each x X , there is a singleton that is not rejected with respect to the prior π.Combining these facts, we prove that, for each sample point, exactly one hypothesis of each partition of Θ is accepted, which is equivalent (Theorem 4.3 from [8]) to asserting that the TS generated by that family of losses with respect to π meets the desiderata.Thus, for A σpΘq, let L A : t0, 1u ¢ pΘ ¢ X q Ñ R be given, for θ A c and x X , by L A p1, pθ, xqq 0 and: A .These hypothesis testing loss functions do not penalize correct decisions.They also reflect the decision-maker's tendency to not reject the hypotheses that comprise the best estimate for θ, δpxq.
Analogously, we prove that the condition in Equation (3) of Theorem 3 is fulfilled if δpxq A. Thus, there are testing schemes generated by pL A q AσpΘq that respect invertibility with respect to all priors.Finally, let us prove that a TS ϕ generated by pL A q AσpΘq is such that ϕ tδpxqu pxq 0, for all x X .Indeed,

Example 8 .
In the setup of Example 7, suppose one also needs to test the null hypothesis H B c 0 : θ ¡ 0.5 by taking into account the loss function in

Figure 1
Figure 1 enlightens this idea.

Figure 1 .
Figure 1.Interpretation of sensible relative losses: rougher errors of decisions should be assigned larger relative losses.

Figure 3 .
Figure 3. Projection of B in u ¢ v.

LI
A p0, pθ, xqq min " A c pδpxqq : d A * , and, for θ A and x X , by L A p0, pθ, xqq 0 and: L A p1, pθ, xqq min " 1 C min !Lpd, θq; 1 Lpd, θq ) I A c pδpxqq C max !Lpd, θq; 1 Lpd, θq ) I A pδpxqq : d A c * , where C ¡ 1 is any constant greater than max 3 ErLpδpxq,θq|xs πxpδpxqq : x X B σpΘq, A B ñ ϕ A ¥ ϕ B , i.e., dx X , ϕ A pxq ¥ ϕ B pxq.The testing schemes introduced in Examples 1, 3 and 4 are coherent.Indeed, in Example 1, coherence is a consequence of the monotonicity of probability measures, while in Example 3, it follows from the fact that if A B, then T Bx T A x and, therefore, ev x pAq ¤ ev x pBq.

Table 3 .
The Bayes test for H B c

Table 3 .
Loss function for Example 8.