Abstract
ROC (Receiver Operating Characteristic) analyses are considered under a variety of assumptions concerning the distributions of a measurement X in two populations. These include the binormal model as well as nonparametric models where little is assumed about the form of distributions. The methodology is based on a characterization of statistical evidence which is dependent on the specification of prior distributions for the unknown population distributions as well as for the relevant prevalence w of the disease in a given population. In all cases, elicitation algorithms are provided to guide the selection of the priors. Inferences are derived for the AUC (Area Under the Curve), the cutoff c used for classification as well as the error characteristics used to assess the quality of the classification.
    Keywords:
                                                                    ROC;                    AUC;                    optimal cutoff;                    statistical evidence;                    relative belief;                    binormal;                    mixture Dirichlet process        1. Introduction
An ROC (Receiver Operating Characteristic) analysis is used in medical science to determine whether or not a real-valued diagnostic variable X for a disease or condition is useful. If the diagnostic indicates that an individual has the condition, then this will typically mean that a more expensive or invasive medical procedure is undertaken. So it is important to assess the accuracy of the diagnostic variable  These methods have a wider class of applications but our terminology will focus on the medical context.
An approach to such analyses is presented here that is based on a characterization of statistical evidence and which incorporates all available information as expressed via prior probability distributions. For example, while p-values are often used in such analyses, there are questions concerning the validity of these quantities as characterizations of statistical evidence. As will be discussed, there are many advantages to the framework adopted here.
A common approach to the assessment of the diagnostic variable X is to estimate its AUC (Area Under the Curve), namely, the probability that an individual sampled from the diseased population will have a higher value of diagnostic variable X than an individual independently sampled from the nondiseased population. A good diagnostic should give a value of the AUC near 1 while a value near 1/2 indicates a poor diagnostic test (if the AUC is near 0, then the classification is reversed). It is possible, however, that a diagnostic with AUC  may not be suitable (see Examples 1 and 6). In particular, a cutoff value c needs to be selected so that if  then an individual is classified as requiring the more invasive procedure. Inferences about the error characteristics for the combination  such as the false positive rate, etc., are also required.
This paper is concerned with inferences about the AUC, the cutoff c and the error characteristics of the classification based on a valid measure of evidence. A key aspect of the analysis is the relevant prevalence w. The phrase “relevant prevalence” means that X will be applied to a certain population, such as those patients who exhibit certain symptoms, and w represents the proportion of this subpopulation who are diseased. The value of w may vary by geography, medical unit, time, etc. To make a valid assessment of X in an application, it is necessary that the information available concerning w be incorporated. This information is expressed here via an elicited prior probability distribution for w, which may be degenerate at a single value if w is assumed known, or be quite diffuse when little is known about w. In fact, all unknown population quantities are given elicited priors. There are many contexts where data are available relevant to the value of w and this leads to a full posterior analysis for w as well as for the other quantities of interest. Even when such data are not available, however, it is still possible to take the prior for w into account so the uncertainties concerning w always play a role in the analysis and this is a unique aspect of the approach taken here.
While there are some methods available for the choice of c, these often do not depend on the prevalence w which is a key factor in determining the true error characteristics of  in an application, see [,,,,]. So it is preferable to take w into account when considering the value of a diagnostic in a particular context. One approach to choosing c is to minimize some error criterion that depends on w to obtain  As will be demonstrated in the examples, however, sometimes  results in a classification that is useless. In such a situation a suboptimal choice of c is required but the error characteristics can still be based on what is known about w so that these are directly relevant to the application.
Others have pointed out deficiencies in the AUC statistic and proposed alternatives. For example, it can be argued that taking into account the costs associated with various misclassification errors is necessary and that using the AUC is implicitly making unrealistic assumptions concerning these costs, see []. While costs are relevant, costs are not incorporated here as these are often difficult to quantify. Our goal is to express clearly what the evidence is saying about how good  is via an assessment of its error characteristics. With the error characteristics in hand, a user can decide whether or not the costs of misclassifications are such that the diagnostic is usable. This may be a qualitative assessment although, if numerical costs are available, these could be subsequently incorporated. The principle here is that economic or social factors be considered separately from what the evidence in the data says, as it is a goal of statistics to clearly state the latter.
The framework for the analysis is Bayesian as proper priors are placed on the unknown distribution  (the distribution of X in the nondiseased population), on  (the distribution of X in the diseased population) and the prevalence  In all the problems considered, elicitation algorithms are presented for how to choose these priors. Moreover, all inferences are based on the relative belief characterization of statistical evidence where, for a given quantity, evidence in favor (against) is obtained when posterior beliefs are greater (less) than prior beliefs, see Section 2.2 for discussion and []. So evidence is determined by how the data change beliefs. Section 2 discusses the general framework, defines relevant quantities and provides an outline for how specific relative belief inferences are determined. Section 3 develops the inferences for the quantities of interest for three contexts (1) X is an ordered discrete variable with and without constraints on  (2) X is a continuous variable and  are normal distributions (the binormal model) (3) X is a continuous variable and no constraints are placed on 
There is previous work on using Bayesian methods in ROC analyses. For example, a Bayesian analysis for the binormal model when there are covariates present is developed in []. An estimate of the ROC using the Bayesian bootstrap is discussed in []. A Bayesian semiparametric analysis using a Dirichlet mixture process prior is developed in [,]. The sampling regime where the data can be used for inference about the relevant prevalence and where a gold standard classifier is not assumed to exist is presented in []. Considerable discussion concerning the case where the diagnostic test is binary, covering the cases where there is and is not a gold standard test, as well as the situation where the goal is to compare diagnostic tests and to make inference about the prevalence distribution can be found in [] and also see []. Application of an ROC analysis to a comparison of linear and nonlinear approaches to a problem in medical physics is in []. Further discussion of nonlinear methodology can be found in [,].
The contributions of this paper, that have not been covered by previous published work in this area, are as follows:
- (i)
 - The primary contribution is to base all the inferences associated with an ROC analysis on a clear and unambiguous characterization of statistical evidence via the principle of evidence and the relative belief ratio. While Bayes factors are also used to measure statistical evidence, there are serious limitations on their usage with continuous parameters as priors are restricted to be of a particular form. The approach via relative belief removes such restrictions on priors and provides a unified treatment of estimation and hypothesis assessment problems. In particular, this leads directly to estimates of all the quantities of interest, together with assessments of the accuracy of the estimates, and a characterization of the evidence, whether in favor of or against a hypothesis, together with a measure of the strength of the evidence. Moreover, no loss functions are required to develop these inferences. The merits of the relative belief approach over others are more fully discussed in Section 2.2.
 - (ii)
 - A prior on the relevant prevalence is always used to determine inferences even when the posterior distribution of this quantity is not available. As such the prevalence always plays a role in the inferences derived here.
 - (iii)
 - The error in the estimate of the cut-off is always quantified as well as the errors in the estimates of the characteristics evaluated at the chosen cut-off. It is these characteristics, such as the sensitivity and specificity, that ultimately determine the value of the diagnostic test.
 - (iv)
 - The hypothesis AUC is first assessed and if evidence is found in favor of this, the prior is then conditioned on this event being true for inferences about the remaining quantities. Note that this is equivalent to conditioning the posterior on the event AUC > when inferences are determined by the posterior but with relative belief inferences both the conditioned prior and conditioned posterior are needed to determine the inferences.
 - (v)
 - Precise conditions are developed for the existence of an optimal cutoff with the binormal model.
 - (vi)
 - In the discrete context (1), it is shown how to develop a prior and the analysis under the assumption that the probabilities describing the outcomes from the diagnostic variable X are monotone.
 
The relative belief ratio, as a measure of evidence, is seen to have a connection to relative entropy. For example, it is equivalent, in the sense that the inferences are the same, to use the logarithm of the relative belief ratio as the measure of evidence. The relative entropy is then the posterior expectation of this quantity and so can be considered as a measure of the overall evidence provided by the model, prior and data concerning a quantity of interest.
The methods used for all the computations in the paper are simulation based and represent fairly standard Bayesian computational methods. In each context considered, sufficient detail is provided so that these can be implemented by a user.
2. The Problem
Consider the formulation of the problem as presented in [,] but with somewhat different notation. There is a measurement  defined on a population  with  where  is comprised of those with a particular disease, and  represents those without the disease. So  is the conditional cdf of X in the nondiseased population, and  is the conditional cdf of X in the diseased population. It is assumed that there is a gold standard classifier, typically much more difficult to use than  such that for any  it can be determined definitively if  or  There are two ways in which one can sample from  namely,
- (i)
 - take samples from each of and separately or
 - (ii)
 - take a sample from
 
The sampling method used affects the inferences that can be drawn. For many studies (i) is the relevant sampling mode, as in case-control studies, while (ii) is relevant in cross-sectional studies.
It supposed that the greater the value  is for individual  the more likely it is that  For the classification, a cutoff value c is required such that, if , then  is classified as being in  and otherwise is classified as being in  However, X is an imperfect classifier for any c and it is necessary to assess the performance of . It seems natural that a value of c be used that is optimal in some sense related to the error characteristics of this classification. Table 1 gives the relevant probabilities for classification into  and , together with some common terminology, in a confusion matrix.
       
    
    Table 1.
    Error probabilities when  indicates a positive.
  
Another key ingredient is the prevalence  of the disease in  In practical situations, it is necessary to also take w into account in assessing the error in  The following error characteristics depend on 
      
        
      
      
      
      
    Under sampling regime (ii) and cutoff c, Error(c) is the probability of making an error, FDR(c) is the conditional probability of a subject being misclassified as positive given that it has been classified as positive and FNDR(c) is the conditional probability of a subject being misclassified as negative given that it has been classified as negative. In other words, FDR(c) is the proportion of those individuals in the population consisting of those who have been classified by the diagnostic test as having the disease, but in fact do not have it. It is often observed that when w is very small and FNR(c) and FPR(c) are small, then FDR(c) can be big. This is sometimes referred to as the base rate fallacy as, even though the test appears to be a good one, there is a high probability that an individual classified as having the disease will be misclassified. For example, if  FNR(c) = FPR then Error, FDR FNDR and when  then Error, FDR FNDR In these cases the false nondiscovery rate is quite small while the false discovery rate is large. If the disease is highly contagious, then these probabilities may be considered acceptable but indeed they need to be estimated. Similarly, FNDR may be small when FNR is large and w is very small.
It is naturally desirable to make inference about an optimal cutoff  and its associated error quantities. For a given value of  the optimal cutoff will be defined here as  Error, the value which minimizes the probability of making an error. Other choices for determining a  can be made, and the analysis and computations will be similar, but our thesis is that, when possible, any such criterion should involve the prior distribution of the relevant prevalence  As demonstrated in Example 6 this can sometimes lead to useless values of  even when the AUC is large. While this situation calls into question the value of the diagnostic, a suboptimal choice of c can still be made according to some alternative methodology. For example, sometimes Youden’s index, which maximizes Error over c with , is recommended, or the closest-to-(0,1) criterion which minimizes , see [] for discussion. Youden’s index and the closest-to-(0,1) criterion do not depend on the prevalence and have geometrical interpretations in terms of the ROC curve, but as we will see, the ROC curve does not exist in full generality and this is particularly relevant in the discrete case. The methodology developed here provides an estimate of the c to be used, together with an exact assessment of the error in this estimate, as well as providing estimates of the associated error characteristics of the classification.
Letting  denote the estimate of , the values of Error and  are also estimated and the recorded values used to assess the value of the diagnostic test. There are also other characteristics that may prove useful in this regard such as the positive predictive value (PPV)
      
      
        
      
      
      
      
    
      namely, the conditional probability a subject is positive given that they have tested positive, which plays a role similar to FDR. See [] for discussion of the PPV and the similarly defined negative predictive value (NPV). The value of  can be estimated in the same way as the other quantities as is subsequently discussed.
2.1. The AUC and ROC
Consider two situations where  are either both absolutely continuous or both discrete. In the discrete case, suppose that these distributions are concentrated on a set of points  When  are selected using sampling scheme (i), then the probability that a higher score is received on diagnostic X by a diseased individual than a nondiseased individual is
        
      
        
      
      
      
      
    Under the assumption that  is constant on  for every  there is a function ROC (receiver operating characteristic) such that  ROC so AUCROC Putting  then ROC In the absolutely continuous case, AUCROC which is the area under the curve given by the ROC function. The area under the curve interpretation is geometrically evocative but is not necessary for (1) to be meaningful.
It is commonly suggested that a good diagnostic variable X will have an AUC close to 1 while a value close to 1/2 suggests a poor diagnostic test. It is surely the case, however, that the utility of X in practice will depend on the cutoff c chosen and the various error characteristics associated with this choice. So while the AUC can be used to screen diagnostics, it is only part of the analysis and inferences about the error characteristics are required to truly assess the performance of a diagnostic. Consider an example.
Example 1.  
Suppose that  for some  where  is continuous, strictly increasing with associated density  Then using (1), AUC  which is approximately 1 when q is large. The optimal c minimizes Error which implies c satisfies  when  and the optimal c is otherwise . If  then AUC  and with  so FNR FPR Error FDR and FNDR. So X seems like a good diagnostic via the AUC and the error characteristics that depend on the prevalence although within the diseased population the probability is  of not detecting the disease. If instead  then the AUC is the same but  and the optimal classification always classifies an individual as non-diseased which is useless. So the AUC does not indicate enough about the characteristics of the diagnostic to determine if it is useful or not. It is necessary to look at the error characteristics of the classification at the cutoff value that will actually be used, to determine if a diagnostic is suitable and this implies that information about w is necessary in an application.
2.2. Relative Belief Inferences
Suppose there is a model  for data  together with a prior probability measure  with density  on  These ingredients lead, via the principle of conditional probability, to beliefs about the true value of  as initially expressed by  being replaced by the posterior probability measure  with density  Note that if interest is instead in a quantity  where  and we use the same notation for the function and its range, then the model is replaced by  where  is obtained by integrating out the nuisance parameters, and the prior is replaced by the marginal prior  This leads to the marginal posterior  with density 
For the moment suppose that all the distributions are discrete. The principle of evidence then says that there is evidence in favor of the value  if  evidence against the value  if  and no evidence either way if  So, for example, there is evidence in favor of  if the probability of  increases after seeing the data. To order the possible values with respect to the evidence, we use the relative belief ratio
      
        
      
      
      
      
    Note that  indicates whether there is evidence in favor of (against) the value  If there is evidence in favor of both  and  then there is more evidence in favor of  than  whenever  and, if there is evidence against both  and  then there is more evidence against  than  whenever  For the continuous case consider a sequence of neighborhoods  as  and then
        
      
        
      
      
      
      
    
        under very weak conditions such as  and  being continuous at 
All the inferences about quantities considered in the paper are derived based upon the principle of evidence as expressed via the relative belief ratio. For example, it is immediate that the value  indicates whether or not there is evidence in favor of or against the hypothesis  Furthermore, the posterior probability  measures the strength of this evidence for, if  and this probability is large, then there is strong evidence in favor of  as there is a small belief that the true value has a larger relative belief ratio and if  and this probability is small, then there is strong evidence against  as there is high belief that the true value has a larger relative belief ratio. For estimation it is natural to estimate  by the relative belief estimate  as this value has the maximum evidence in its favor. Furthermore, the accuracy of this estimate can be assessed by looking at the plausible region  consisting of all those values for which there is evidence in favor, together with its size and posterior content which measures how strongly it is believed the true value lies in this set. Rather than using the plausible region to assess the accuracy of , one could quote a relative belief credible region
      
        
      
      
      
      
    
        where the constant  is the largest value such that . It is necessary, however, that  as otherwise  will contain values for which there is evidence against, and this is only known after the data have been seen.
It is established in [], and in papers referenced there, that these inferences possess a number of good properties such as consistency, satisfy various optimality criteria and clearly they are based on a direct measure of the evidence. Perhaps most significant is the fact that all the inferences are invariant under reparameterizations. For if , where  is a smooth bijection, then
        
      
        
      
      
      
      
    
        and so, for example,  This invariance property is not possessed by the most common inference methods employed such as MAP estimation or using posterior means and this invariance holds no matter what the dimension of  is. Moreover, it is proved in [] that relative belief inferences are optimally robust among all Bayesian inferences for , to linear contaminations of the prior on .
An analysis, using relative belief, of the data obtained in several physics experiments that were all concerned with examining whether there was evidence in favor of or against the quantum model versus hidden variables is available in []. Furthermore, an approach to checking models used for quantum mechanics via relative belief is discussed in []. Other applications of relative belief inferences to common problems of statistical practice can be found in [].
The Bayes factor is an alternative measure of evidence and is commonly used for hypothesis assessment in Bayesian inference. To see why the relative belief ratio has advantages over the Bayes factor for evidence-based inferences consider first assessing the hypothesis  When the prior probability of  satisfies  then the Bayes factor is defined as the ratio of the posterior odds in favor of  to the prior odds in favor of  namely,
        
      
        
      
      
      
      
    It is easily shown that the Bayes factor satisfies the principle of evidence and  is evidence in favor (against)  so in this context it is a valid measure of evidence.
One might wonder why it is necessary to consider a ratio of odds as opposed to the simpler ratio of probabilities, as specified by the relative belief ratio, for the purpose of measuring evidence but in fact there is a more serious issue with the Bayes factor. For suppose, as commonly arises in applications, that  is a continuous probability measure so that  as then the Bayes factor for  is not defined. The common recommendation in this context is to require the specification of the following ingredients: a prior probability  a prior distribution  concentrated on  which provides the prior predictive density , a prior distribution  concentrated on  which provides the prior predictive density  and then the full prior is taken to be the mixture  With this prior the Bayes factor for  is defined, as now the prior probability of  equals  and an easy calculation shows that  Typically the prior  is taken to be the prior that we might place on  when interest is in estimating .
Now consider the problem of estimating  and the prior is such that  for every value of  as with a continuous prior. The Bayes factor is then not defined for any value of  and, if we wished to use the Bayes factor for estimation purposes, it would be necessary to modify the prior to be a different mixture for each value of  so that there would be in effect multiple different priors. This does not correspond to the logic underlying Bayesian inference. When using the relative belief ratio for inference only one prior is required and the same measure of evidence is used for both hypothesis assessment and estimation purposes.
Another approach to dealing with the problem that arises with the Bayes factor and continuous priors is to take a limit as in (2) and, when this is done, we obtain the result
        
      
        
      
      
      
      
    
        as  whenever the prior density of  is continuous and positive at  In other words the relative belief ratio can be also considered as a natural definition of the Bayes factor in continuous contexts.
3. Inferences for an ROC Analysis
Suppose we have a sample of  from , namely,  and a sample of  from , namely,  and the goal is to make inference about the AUC, the cutoff c and the error characteristics FNR FPR Error FDR and FNDR. For the AUC it makes sense to first assess the hypothesis  AUC  via stating whether there is evidence for or against  together with an assessment of the strength of this evidence. Estimates are required for all of these quantities, together with an assessment of the accuracy of the estimate.
3.1. The Prevalence
Consider first inferences for the relevant prevalence  If w is known, or at least assumed known, then nothing further needs to be done but otherwise this quantity needs to be estimated when assessing the value of the diagnostic and so uncertainty about w needs to be addressed.
If the full data set is based on sampling scheme (ii), then  binomial A natural prior  to place on w is a beta distribution. The hyperparameters are chosen based on the elicitation algorithm discussed in [] where interval  is chosen such that it is believed that  with prior probability  Here  is chosen so that we are virtually certain that  and  then seems like a reasonable choice. Note that choosing  corresponds to w being known and so  in that case. Next pick a point  for the mode of the prior and a reasonable choice might be  Then putting  leads to the parameterization beta beta where  locates the mode and  controls the spread of the distribution about  Here  gives the uniform distribution and  gives the distribution degenerate at  With  specified,  is the smallest value of  such that the probability content of  is  and this is found iteratively. For example, if  and  so w is known reasonably well, then  and  so the prior is beta and the posterior is beta
The estimate of w is then
        
      
        
      
      
      
      
    In this case the estimate is the MLE, namely,  The accuracy of this estimate is measured by the size of the plausible region . For example, if  and  then  and  which has posterior content  So the data suggest that the upper bound of  is too strong although the posterior belief in this interval is not very high.
The prior and posterior distributions of w play a role in inferences about all the quantities that depend on the prevalence. In the case where the cutoff is determined by minimizing the probability of a misclassification, then  FNR FPR Error FDR and FNDR all depend on the prevalence. Under sampling scheme (i), however, only the prior on w has any influence when considering the effectiveness of  Inference for these quantities is now discussed in both cases.
3.2. Ordered Discrete Diagnostic
Suppose X takes values on the finite ordered scale  and let  so  and . These imply that FPR FNR
      
        
      
      
      
      
    
        with the remaining quantities defined similarly. Ref. [] can be used to obtain independent elicited Dirichlet priors
        
      
        
      
      
      
      
    
        on these probabilities by placing either upper or lower bounds on each cell probability that hold with virtual certainty  as discussed for the beta prior on the prevalence. If little information is available, it is reasonable to use uniform (Dirichlet) priors on  and  This together with the independent prior on w leads to prior distributions for the AUC,  and all the quantities associated with error assessment such as FNR etc.
Data  lead to counts  and  which in turn lead to the independent posteriors
        
      
        
      
      
      
      
    Under sampling regime (ii) this, together with the independent posterior on  leads to posterior distributions for all the quantities of interest. Under sampling regime (i), however, the logical thing to do, so the inferences reflect the uncertainty about  is to only use the prior on w when deriving inferences about any quantities that depend on this such as  and the various error assessments.
Consider inferences for the AUC. The first inference should be to assess the hypothesis  AUC  for, if  is false, then X would seem to have no value as a diagnostic (the possibility that the directionality is wrong is ignored here). The relative belief ratio  is computed and compared to 1. If it is concluded that  is true, then perhaps the next inference of interest is to estimate the AUC via the relative belief estimate. The prior and posterior densities of the AUC are not available in closed form so estimates are required and density histograms are employed here for this. The set  is discretized into L subintervals  and putting  the value of the prior density  is estimated by  proportion of prior simulated values of AUC in  and similarly for the posterior density  Then  is maximized to obtain the relative belief estimate AUC together with the plausible region and its posterior content.
These quantities are also obtained for  in a similar fashion, although  has prior and posterior distribution concentrated on  so there is no need to discretize. Estimates of the quantities FNR FPR Error FDR and FNDR are also obtained as these indicate the performance of the diagnostic in practice. The relative belief estimates of these quantities are easily obtained in a second simulation where  is fixed.
Consider now an example.
Example 2.  
  
    
      
      
    
  
  
  
    
      
      
    
  
  
  
    
        
       
    
  
  
  
    
      
      
    
  
  
Simulated example.
For  and , data were generated as
      
        
      
      
      
      
    With these choices for  the true values are AUC, and with ,  FNR FPR Error FDR and FNDR. So X is not an outstanding diagnostic but with these error characteristics it may prove suitable for a given application. Uniform, namely, Dirichlet priors were placed on  and  reflecting little knowledge about these quantities.
Simulations based on Monte Carlo sample sizes of  from the prior and posterior distributions of  and  were conducted and the prior and posterior distributions of the quantities of interest obtained. The hypothesis  AUC  is assessed by  So there is evidence in favor of  and the strength of this evidence is measured by the posterior probability content of  which equals  to machine accuracy and so this is categorical evidence in favor of 
For the continuous quantities a grid based on  equispaced points  was used and all the mass in the interval  assigned to the midpoint  Figure 1 contains plots of the prior and posterior densities and relative belief ratio of the AUC. The relative belief estimate of the AUC is AUC with  having posterior content  Certainly a finer partition of  than just 24 intervals is possible, but even in this relatively coarse case the results are quite accurate.
      
    
    Figure 1.
      In Example 2, plots of the prior (- - -), the posterior (—) and the RB ratio of the AUC.
  
Supposing that the relevant prevalence is known to be  Figure 2 contains plots of the prior and posterior densities and relative belief ratio of  The relative belief estimate is  with  with posterior probability content  so the correct optimal cut-off has been identified but there is a degree of uncertainty concerning this. The error characteristics that tell us about the utility of X as a diagnostic are given by the relative belief estimates (column (a)) in Table 2. It is interesting to note that the estimate of Error is determined by the prior and posterior distributions of a convex combination of FPR and FNR and the estimate is not the same convex combination of the estimates of FPR and FNR. So, in this case Error seems like a much better assessment of the performance of the diagnostic.
      
    
    Figure 2.
      In Example 2, plots of the the prior (+), the posterior (×) and the RB ratio of .
  
       
    
    Table 2.
    The estimates of the error characteristcs of X at  in Example 2 where (a) w is assumed known, (b) only the prior for w is available, (c) the posterior for w is also available.
  
Suppose now that the prevalence is not known but there is a beta prior specified for w and consider the choice discussed in Section 3.1 where  and  When the data are produced according to sampling regime (i), then there is no posterior for w but this prior can still be used in determining the prior and posterior distributions of  and the associated error characteristics. When this simulation was carried out  with  with posterior probability content  and column (b) of Table 2 gives the estimates of the error characteristics. So other than the estimate of the FPR, the results are similar. Finally, assuming that the data arose under sampling scheme (ii), then w has a posterior distribution and using this gives  with  with posterior probability content  and error characteristics as in column (c) of Table 2. These results are the same as if the prevalence is known which is sensible as the posterior concentrates about the true value more than the prior.
Another somewhat anomalous feature of this example is the fact that uniform priors on  and  do not lead to a prior on the AUC that is even close to uniform. In fact one could say that this prior has a built-in bias against a diagnostic with AUC  and indeed most choices of  and  will not satisfy this. Another possibility is to require  and  namely, require monotonicity of the probabilities. A result in [] implies that  satisfies this iff  where  the standard -dimensional simplex, and  with i-ith row equal to  and  satisfies this iff  where  and  where  contains all 0’s except for 1’s on the crossdiagonal. If  and  are independent and uniform on  then  and  are independent and uniform on the sets of probabilities satisfying the corresponding monotonicities and Figure 3 has a plot of the prior of the AUC when this is the case. It is seen that this prior is biased in favor of AUC  Figure 3 also has a plot of the prior of the AUC when  is uniform on the set of all nondecreasing probabilities and  is uniform on  This reflects a much more modest belief that X will satisfy AUC  and indeed this may be a more appropriate prior than using uniform distributions on  Ref. [] also provides elicitation algorithms for choosing alternative Dirichlet distributions for  and 
      
    
    Figure 3.
      Prior density of the AUC when  is uniform on the set of nondecreasing probabilities independent of  uniform on the set of nonincreasing probabilities (–) as well as when  is uniformly distributed on the set of nondecreasing probabilities independent of  uniform on  (- -).
  
When  AUC  is accepted, it makes sense to use the conditional prior, given that this event is true, in the inferences. As such it is necessary to condition the prior on the event  In general, it is not clear how to generate from this conditional prior but depending on the size of m and the prior, a brute force approach is to simply generate from the unconditional prior and select those samples for which the condition is satisfied and the same approach works with the posterior.
Here  and using uniform priors for  and , the prior probability of AUC  is  while the posterior probability is  so the posterior sampling is much more efficient. Choosing priors that are more favorable to AUC  will improve the efficiency of the prior sampling. Using the conditional priors led to AUC with  with posterior content . This is similar to the results obtained using the unconditional prior but the conditional prior puts more mass on larger values of the AUC hence the wider plausible region with lower posterior content. Moreover,  with  with posterior probability content approximately  (actually ) which reflects virtual certainty that the true optimal value is in 
3.3. Binormal Diagnostic
Suppose now that X is a continuous diagnostic variable and it is assumed that the distributions  and  are normal distributions. The assumption of normality should be checked by an appropriate test and it will be assumed here that this has been carried out and normality was not rejected. While the normality assumption may seem somewhat unrealistic, many aspects of the analysis can be expressed in closed form and this allows for a deeper understanding of ROC analyses more generally.
With  denoting the  cdf, then FNRFPR so  and
        
      
        
      
      
      
      
    For given  and  all these values can be computed using  except the AUC and for that quadrature or simulation via generating  is required.
The following results hold for the AUC with the proofs in the Appendix A.
        
Lemma 1.  
From Lemma 1 it is clear that it makes sense to restrict the parameterization so that  but we need to test the hypothesis  first. Clearly ErrorFNRFPR as  and Error as  so, if Error does not achieve a minimum at a finite value of  then the optimal cut-off is infinite and the optimal error is  It is possible to give conditions under which a finite cutoff exists and express  in closed form when the parameters and the relevant prevalence w are all known.AUC  iff  and when  the AUC is a strictly increasing function of 
Lemma 2.  
(i) When  then a finite optimal cut-off minimizing Error exists iff  and in that case
      
        
      
      
      
      
    (ii) When  then a finite optimal cut-off exists iff
      
        
      
      
      
      
    and in that case
Corollary 1.  
The restrictions  and (6) hold iff
Consider now examples with equal and unequal variances.
Example 3.  
  
    
      
      
    
  
  
  
    
      
      
    
  
  
Binormal with 
There may be reasons why the assumption of equal variance is believed to hold but this needs to be assessed and evidence in favor found. If evidence against the assumption is found, then the approach of Example 4 can be used. A possible prior is given by  where
      
        
      
      
      
      
    which is a conjugate prior. The hyperparameters to be elicited are  Consider first eliciting the prior for  For this an interval  is specified such that is it believed that  with virtual certainty (say with probability  Then putting  implies
      
        
      
      
      
      
    which implies  where  The interval  will contain an observation from  with virtual certainty and let  be lower and upper bounds on the half-length of this interval so  with virtual certainty. This implies  This leaves specifying the hyperparameters  and letting  denote the cdf of the gamma distribution, then  satisfying
      
        
      
      
      
      
    will give the specified γ coverage. Noting that  first specify  and solve the first equation in (9) for  and then solve the second equation in (9) for  and continue this iteration until the probability content of  is sufficiently close to γ. Using  the posterior is then
      
        
      
      
      
      
    where
      
        
      
      
      
      
    Suppose the following values of the mss were obtained based on samples of  from  and  from 
      
        
      
      
      
      
    So the true values of the parameters are  In this case AUC  Supposing that the relevant prevalence is  FNR, FPR, Error, FDR, FNDR,
For the prior elicitation, suppose it is known with virtual certainty that both means lie in  and  so we take  and the iterative process leads to  For inference about  it is necessary to specify a prior distribution for the prevalence  This can range from w being completely known to being completely unknown whence a uniform(0,1) (beta) would be appropriate. Following the developments of Section 3.1, suppose it is known that  with prior probability  so in this case  and  and the prior is  beta
The first inference step is to assess the hypothesis  AUC  which is equivalent to  by computing the prior and posterior probabilities of this event to obtain the relative belief ratio. The prior probability of  given  is
      
        
      
      
      
      
    and averaging this quantity over the prior for  we get  The posterior probability of this event can be easily obtained via simulating from the joint posterior. When this is done in the specific numerical example, the relative belief ratio of this event is  with posterior content  so there is strong evidence that  AUC  is true.
If evidence is found against  then this would indicate a poor diagnostic. If evidence is found in favor, then we can proceed conditionally given that  holds and so condition the joint prior and joint posterior on this event being true when making inferences about AUC,  etc. So for the prior it is necessary to generate  gamma and then generate  from the joint conditional prior given  and that  Denoting the conditional priors given  by  and  we see that this joint conditional prior is proportional to
      
        
      
      
      
      
    While generally it is not possible to generate efficiently from this distribution we can use importance sampling to calculate any expectations by generating  with  serving as the importance sampling weight and where  denotes the  distribution conditioned to  with density
      
        
      
      
      
      
    for  and 0 otherwise. Generating from this distribution via inversion is easy since the cdf is  Note that, if we take the posterior from the unconditioned prior and condition that, we will get the same conditioned posterior as when we use the conditioned prior to obtain the posterior. This implies that in the joint posterior for  it is only necessary to adjust the posterior for  as was done with the prior and this is also easy to generate from. Note that Lemma 2 (i) implies that it is necessary to use the conditional prior and posterior to guarantee that  exists finitely.
Since  was accepted, the conditional sampling was implemented and the estimate of the AUC is  with plausible region  which has posterior content  So the estimate is close to the true value but there is substantial uncertainty. Figure 4 is a plot of the conditioned prior, the conditioned posterior and relative belief ratio for this data.
      
    
    Figure 4.
      The conditioned prior (- -) and posterior (–) densities (left panel) and the relative belief ratio (right panel) of the AUC in Example 3.
  
With the specified prior for  the posterior is beta  which leads to estimate  for w with plausible interval  having posterior probability content  Using this prior and posterior for w and the conditioned prior and posterior for  we proceed to an inference about  and the error characteristics associated with this classification. A computational problem arises when obtaining the prior and posterior distributions of  as it is clear from (5) that these distributions can be extremely long-tailed. As such, we transform to  (the Cauchy cdf), obtain the estimate  where  and its plausible region and then, applying the inverse transform, obtain  and its plausible region. It is notable that relative belief inferences are invariant under 1-1 smooth transformations, so it does not matter which parameterization is used, but it is much easier computationally to work with a bounded quantity. Furthermore, if a shorter tailed cdf is used rather than a Cauchy, e.g., a  cdf, then errors can arise due to extreme negative values being always transformed to 0 and very extreme positive values always transformed to 1. Figure 5 is a plot of the prior density, posterior density and relative belief ratio of  For these data  with plausible interval  having posterior content  Large Monte Carlo samples were used to get smooth estimates of the densities and relative belief ratio but these only required a few minutes of computer time on a desktop. The estimated error characteristics at this value of  are as follows: FNR, FPR, Error, FDR, FNDR which are close to the true values.
      
    
    Figure 5.
      Plots of the prior (- -), posterior (left panel) and relative belief ratio (right panel) of  in Example 3.
  
Example 4.  
  
    
      
      
    
  
  
Binormal with 
In this case the prior is given by  where
      
        
      
      
      
      
    Although this specifies the same prior for the two populations, this is easily modified to use different priors and, in any case, the posteriors are different. Again it is necessary to check that the AUC  but also to check that  exists using the full posterior based on this prior and for this we have the hypothesis  given by Corollary 1. If evidence in favor of  is found, the prior is replaced by the conditional prior given this event for inference about  This can be implemented via importance sampling as was done in Example 3 and similarly for the posterior.
Using the same data and hyperparameters as in Example 3 the relative belief ratio of  is  with posterior content  so there is reasonably strong evidence in favor of  Estimating the value of the AUC is then based on conditioning on  being true. Using the conditional prior given that  is true, the relative belief estimate of the AUC is  with plausible interval  with posterior content  The optimal cutoff is estimated as  with plausible interval  having posterior content  Figure 6 is a plot of the prior density, posterior density and relative belief ratio of  The estimates of the error characteristics at  are as follows: FNR, FPR, Error, FDR, FNDR
      
    
    Figure 6.
      Plots of the prior (- -), posterior (left panel) and relative belief ratio (right panel) of  in Example 4.
  
It is notable that these inferences are very similar to those in Example 3. It is also noted that the sample sizes are not big and so the only situation where it might be expected that the inferences will be quite different between the two analyses is when the variances are substantially different.
3.4. Nonparametric Bayes Model
Suppose that X is a continuous variable, of course still measured to some finite accuracy, and available information is such that no particular finite dimensional family of distributions is considered feasible. The situation is considered where a normal distribution  perhaps after transforming the data, is considered as a possible base distribution for X but we want to allow for deviation from this form. Alternative choices can also be made for the base distribution. The statistical model is then to assume that the  and  are generated as samples from  and  where these are independent values from a DP (Dirichlet) process with base  for some  and concentration parameter  Actually, since it is difficult to argue for some particular choice of  it is supposed that  also has a prior  The prior on  is then specified hierarchically as a mixture Dirichlet process,
        
      
        
      
      
      
      
    To complete the prior it is necessary to specify  and the concentration parameters  and  For  the prior is taken to be a normal distribution elicited as discussed in Section 3.3 although other choices are possible. For eliciting the concentration parameters, consider how strongly it is believed that normality holds and for convenience suppose  If  DP with H a probability measure, then  and  When F a random measure from  then  which, when  DP equals
        
      
        
      
      
      
      
    
        where  denotes the beta measure. This upper bound on the probability that the random F differs from H by at least  on an event can be made as small as desirable by choosing a large enough. For example, if  and it is required that this upper bound be less than  then this satisfied when  and if instead , then  is necessary. Note that, since this bound holds for every continuous probability measure  it also holds when H is random, as considered here. So a is controlling how close it is believed that the true distribution is to H. Alternative methods for eliciting a can be found in [,].
Generating  from the prior for given  can only be done approximately and the approach of [] is adopted. For this, integer  is specified and measure  is generated where  independent of  since  DP as  So to carry out a priori calculations proceed as follows. Generate
        
      
        
      
      
      
      
    
        and similarly for  and  Then  is the random cdf at  and similarly for , so AUC  is a value from the prior distribution of the AUC. This is done repeatedly to get the prior distribution of the AUC as in our previous discussions and we proceed similarly for the other quantities of interest.
Now  independent of  with  and  is the empirical cdf (ecdf) based on  and similarly for  The posteriors of  and  are obtained via results in [,]. The posterior density of  given  is proportional to
        
      
        
      
      
      
      
    
        where  is the number of unique values in  and  is the set of unique values with mean  and sum of squared deviations . From this it is immediate that
        
      
        
      
      
      
      
    
        where  A similar result holds for the posterior of 
To approximately generate from the full posterior specify some  put  and generate
        
      
        
      
      
      
      
    
        and similarly for  and  If the data does not comprise a sample from the full population, then the posterior for w is replaced by its prior.
There is an issue that arises when making inference about  namely, the distributions for  that arises from this approach can be very irregular and particularly the posterior distribution. In part this is due to the discreteness of the posterior distributions of  and . This does not affect the prior distribution because the points on which the generated distributions are concentrated vary quite continuously among the realizations and this leads to a relatively smooth prior density for  For the posterior, however, the sampling from the ecdf leads to a very irregular, multimodal density for  So some smoothing is necessary in this case.
Consider now applying such an analysis to the dataset of Example 3, where we know the true values of the quantities of interest and then to a dataset concerned with the COVID-19 epidemic.
Example 5.  
Binormal data (Examples 3 and 4)
The data used in Example 3 are now analyzed but using the methods of this section. The prior on  and w is taken to be the same as that used in Example 4 so the variances are not assumed to be the same. The value  is used and requiring (11) to be less than  leads to  So the true distributions are allowed to differ quite substantially from a normal distribution. Testing the hypothesis  AUC  led to the relative belief ratio  (maximum possible value is 2) and the strength of the evidence is  so there is strong evidence that  is true. The AUC, based on the prior conditioned on  being true, is estimated to be equal to  with plausible interval  having posterior content  For these data  with plausible interval  having posterior content . The true value of the AUC is  and the true value of  is  so these inferences are certainly reasonable although, as one might expect, when the length of the plausible intervals are taken into account, they are not as accurate as those when binormality is assumed as this is correct for this data. So the DP approach worked here although the posterior density for  was quite multimodal and required some smoothing (averaging 3 consecutive values).
Example 6.  
  
    
        
       
    
  
  
  
    
        
       
    
  
  
  
    
        
       
    
  
  
COVID-19 data.
A dataset was downloaded from https://github.com/YasinKhc/Covid-19 containing data on 3397 individuals diagnosed with COVID-19 and includes whether or not the patient survived the disease, their gender and their age. There are 1136 complete cases on these variables of which 646 are male, with 52 having died, and 490 are female, with 25 having died. Our interest is in the use of a patient’s age X to predict whether or not they will survive. More detail on this dataset can be found in []. The goal is to determine a cutoff age so that extra medical attention can be paid to patients beyond that age. Furthermore, it is desirable to see whether or not gender leads to differences so separate analyses can be carried out by gender. So, for example, in the male group ND refers to those males with COVID-19 that will not die and D refers to the population that will. Looking at histograms of the data, it is quite clear that binormality is not a suitable assumption and no transformation of the age variable seems to be available to make a normality assumption more suitable. Table 3 gives summary statistics for the subgroups. Of some note is that condition (8), when using standard estimates for population quantities such as  for Males and  for females, is not satisfied which suggests that in a binormal analysis no finite optimal cutoff exists.
       
    
    Table 3.
    Summary statistics for the data in Example 6.
  
For the prior, it is assumed that  and  are independent values from the same prior distribution as in (10). For the prior elicitation, as discussed in Example 3, suppose it is known with virtual certainty that both means lie in  and  so we take  and the iterative process leads to  which implies a prior on the σ’s with mode at  and the interval  containing  of the prior probability. Here the relevant prevalence refers to the proportion of COVID-19 patients that will die and it is supposed that  with virtual certainty which implies  beta So the prior probability that someone with COVID-19 will die is assumed to be less than 15% with virtual certainty. Since normality is not an appropriate assumption for the distribution of  the choice  with the upper bound (11) equal to  seems reasonable and so  This specifies the prior that is used for the analysis with both genders and it is to be noted that it is not highly informative.
For males the hypothesis AUC  is assessed and  (maximum value 2) with strength effectively equal to  was obtained, so there is extremely strong evidence that this is true. The unconditional estimate of the AUC is  with plausible region  having posterior content  so there is a fair bit of uncertainty concerning the true value. For the conditional analysis, given that AUC  the estimate of the AUC is  with plausible region  having posterior content  So the conditional analysis gives a similar estimate for the AUC with a small increase in accuracy. In either case it seems that the AUC is indicating that age should be a reasonable diagnostic. Note that the standard nonparametric estimate of the AUC is  so the two approaches agree here. For females the hypothesis AUC  is assessed and  with strength effectively equal to 1 was obtained, so there is extremely strong evidence that this is true. The unconditional estimate of the AUC is  with plausible region  having posterior content . For the conditional analysis, given that AUC  the estimate of the AUC is  with plausible region  having posterior content  The traditional estimate of the AUC is  so the two approaches are again in close agreement.
Inferences for  are more problematical in both genders. Consider the male data. The data set is very discrete as there are many repeats and the approach samples from the ecdf about 84% of the time for the males that died and 98% of the time for the males that did not die. The result is a plausible region that is not contiguous even with smoothing. Without smoothing the estimate is  for males, which is a very dominant peak for the relative belief ratio. The plausible region contains  of the posterior probability and, although it is not a contiguous interval, the subinterval  is a -credible interval for  that is in agreement with the evidence. If we make the data continuous by adding a uniform(0,1) random error to each age in the data set, then  and plausible interval  with posterior content  is obtained. These cutoffs are both greater than the maximum value in the ND data, so there is ample protection against false positives but it is undoubtedly false negatives that are of most concern in this context. If instead the FNDR is used as the error criterion to minimize, then  and plausible interval  with posterior content  is obtained and so in this case there will be too many false positives. So a useful optimal cutoff incorporating the relevant prevalence does not seem to exist with these data.
If the relevant prevalence is ignored and FNRFPR is used for some fixed weight  to determine , then more reasonable values are obtained. Table 4 gives the estimates for various  values. With  (corresponding to using Youden’s index)  while if  then  When  is too small or too large then the value of  is not useful. While these estimates do not depend on the relevant prevalence, the error characteristics that do depend on this prevalence (as expressed via its prior and posterior distributions) can still be quoted and a decision made as to whether or not to use the diagnostic. Table 5 contains the estimates of the error characteristics at  for various values of  where these are determined using the prior and posterior on the relevant prevalence  Note that these estimates are determined as the values that maximize the corresponding relative belief ratios and take into account the posterior of  So, for example, the estimate of the Error is not the convex combination of the estimates of FNR and FPR based on the  weight. Another approach is to simply set the cutoff Age at a value at a value  and then investigate the error characteristics at that value. For example, with  then the estimated values are given by FNR FPR Error FDR  and FNDR
       
    
    Table 4.
    Weighted error FNR+(FPR determining  for Males in Example 6.
  
       
    
    Table 5.
    Error characteristics for Males in Example 6 at various weights.
  
Similar results are obtained for the cutoff with female data although with different values. Overall, Age by itself does not seem to be useful classifier although that is a decision for medical practitioners. Perhaps it is more important to treat those who stand a significant chance of dying more extensively and not worry too much that some treatments are not necessary. The clear message from this data, however, is that a relatively high AUC does not immediately imply that a diagnostic is useful and the relevant prevalence is a key aspect of this determination.
4. Conclusions
ROC analyses represent a significant practical application of statistical methodology. While previous work has considered such analyses within a Bayesian framework, this has typically required the specification of loss functions. Losses are not required in the approach taken here which is entirely based on a natural characterization of statistical evidence via the principle of evidence and the relative belief ratio. As discussed in Section 2.2 this results in a number of good properties for the inferences that are not possessed by inferences derived by other approaches. While the Bayes factor is also a valid measure of evidence, its usage is far more restricted than the relative belief ratio which can be applied with any prior, without the need for any modifications, for both hypothesis assessment and estimation problems. This paper has demonstrated the application of relative belief to ROC analyses under a number of model assumptions. In addition, as documented in points (ii)–(vi) of the Introduction, a number of new results have been developed for ROC analyses more generally.
Author Contributions
Methodology, L.A.-L. and M.E.; Investigation, Q.L.; Writing—original draft, M.E.; Supervision, M.E. All authors have read and agreed to the published version of the manuscript.
Funding
Evans was supported by grant 10671 from the Natural Sciences and Engineering Research Council of Canada.
Data Availability Statement
The data and R code used for the examples in Section 3.2, Section 3.3 and Section 3.4 can be obtained at https://utstat.utoronto.ca/mikevans/software/ROCcodeforexamples.zip (accessed on 15 November 2022).
Conflicts of Interest
The authors declare no conflict of interest.
Appendix A
Proof of Lemma 1.  
Consider  as a function of  so
          
      
        
      
      
      
      
    When  then  is increasing in b for , decreasing in b for  equals 0 when  and when  it is decreasing in b for , increasing in b for  Therefore, when  then  and when  then     □
Proof of Lemma 2.  
Note that  will satisfy
          
      
        
      
      
      
      
    
          which implies
          
      
        
      
      
      
      
    So  is a root of the quadratic . A single real root exists when  and is given by (5).
If , then there are two real roots when the discriminant
          
      
        
      
      
      
      
    
          establishing (6). To be a minimum the root c has to satisfy
          
      
        
      
      
      
      
    
          and by (A1), this holds iff
          
      
        
      
      
      
      
    
          which is true iff  When  this is true iff  which completes the proof of (i). When  this, together with the formula for the roots of a quadratic establishes (7).    □
Proof of Corollary 1.  
Suppose  and (6) hold. Then putting
          
      
        
      
      
      
      
    
          we have that, for fixed  and  then  is a quadratic in  This quadratic has discriminant  and so has no real roots whenever  and, noting a does not depend on  the only restriction on  is  When  the roots of the quadratic are given by  and so, since the quadratic is negative between the roots and  the two restrictions imply  Combining the two cases gives (8).
References
- Metz, C.; Pan, X. “Proper” binormal ROC curves: Theory and maximum-likelihood estimation. Math. Psychol. 1999, 43, 1–33. [Google Scholar] [CrossRef] [PubMed]
 - Perkins, N.J.; Schisterman, E.F. The inconsistency of “optimal” cutpoints obtained using two criteria based on the Receiver Operating Characteristic Curve. Am. J. Epidemiol. 2006, 163, 670–675. [Google Scholar] [CrossRef] [PubMed]
 - López-Ratón, M.; Rodríguez-Álvarez, M.X.; Cadarso-Suárez, C.; Gude-Sampedro, F. OptimalCutpoints: An R package for selecting optimal cutpoints in diagnostic tests. J. Stat. 2014, 61, 8. [Google Scholar] [CrossRef]
 - Unal, I. Defining an optimal cut-point value in ROC analysis: An alternative approach. Comput. Math. Model. Med. 2017, 2017, 3762651. [Google Scholar] [CrossRef]
 - Verbakel, J.Y.; Steyerberg, E.W.; Uno, H.; De Cock, B.; Wynants, L.; Collins, G.S.; Van Calster, B. ROC plots showed no added value above the AUC when evaluating the performance of clinical prediction models. Clin. Epidemiol. 2020, in press. [Google Scholar]
 - Hand, D. Measuring classifier performance: A coherent alternative to the area under the ROC curve. Mach. Learn. 2009, 99, 103–123. [Google Scholar] [CrossRef]
 - Evans, M. Measuring Statistical Evidence Using Relative Belief. In Monographs on Statistics and Applied Probability; CRC Press: Boca Raton, FL, USA; Taylor & Francis: Abingdon, UK, 2015; Volume 144. [Google Scholar]
 - O’Malley, A.J.; Zou, K.H.; Fielding, J.R.; Tempany, C.M.C. Bayesian regression methodology for estimating a receiver operating characteristic curve with two radiologic applications: Prostate biopsy and spiral CT of ureteral stiones. Acad. Radiol. 2001, 8, 5407–5420. [Google Scholar]
 - Gu, J.; Ghosal, S.; Roy, A. Bayesian bootstrap estimation of ROC curve. Stat. Med. 2008, 27, 5407–5420. [Google Scholar] [CrossRef]
 - Erkanli, A.; Sung, M.; Costello, E.J.; Angold, A. Bayesian semi-parametric ROC analysis. Stat. Med. 2006, 25, 3905–3928. [Google Scholar] [CrossRef] [PubMed]
 - De Carvalho, V.; Jara, A.; Hanson, E.; de Carvalho, M. Bayesian nonparametric ROC regression modeling. Bayesian Anal. 2013, 3, 623–646. [Google Scholar] [CrossRef]
 - Ladouceur, M.; Rahme, E.; Belisle, P.; Scott, A.; Schwartzman, K.; Joseph, L. Modeling continuous diagnostic test data using approximate Dirichlet process distributions. Stat. Med. 2011, 30, 2648–2662. [Google Scholar] [CrossRef] [PubMed]
 - Christensen, R.; Johnson, W.; Branscum, A.; Hanson, T.E. Bayesian Ideas and Data Analysis; Chapman and Hall/CRC: Boca Raton, FL, USA, 2011. [Google Scholar]
 - Rosner, G.L.; Laud, P.W.; Johnson, W.O. Bayesian Thinking in Biostatistics; Chapman and Hall/CRC: Boca Raton, FL, USA, 2021. [Google Scholar]
 - Diab, A.; Hassan, M.; Marquea, C.; Karlsson, B. Performance analysis of four nonlinearity analysis methods using a model with variable complexity and application to uterine EMG signals. Med. Eng. Phys. 2014, 36, 761–767. [Google Scholar] [CrossRef] [PubMed]
 - Gao, X.-Y.; Guo, Y.-J.; Shan, W.-R. Regarding the shallow water in an ocean via a Whitham-Broer-Kaup-like system: Hetero-Bäcklund transformations, bilinear forms and M solitons. Chaos Solitons Fractals 2022, 162, 112486. [Google Scholar] [CrossRef]
 - Gao, X.-T.; Tian, B. Water-wave studies on a (2+1)-dimensional generalized variable-coefficient Boiti–Leon–Pempinelli system. Appl. Math. Lett. 2022, 128, 107858. [Google Scholar] [CrossRef]
 - Obuchowski, N.; Bullen, J. Receiver operating characteristic (ROC) curves: Review of methods with applications in diagnostic medicine. Phys. Med. Biol. 2018, 63, 07TR01. [Google Scholar] [CrossRef]
 - Zhou, X.; Obuchowski, N.; McClish, D. Statistical Methods in Diagnostic Medicine, 2nd ed.; Wiley: Hoboken, NJ, USA, 2011. [Google Scholar]
 - Al-Labadi, L.; Evans, M. Optimal robustness results for some Bayesian procedures and the relationship to prior-data conflict. Bayesian Anal. 2017, 12, 702–728. [Google Scholar] [CrossRef]
 - Gu, Y.; Li, W.; Evans, M.; Englert, B.-G. Very strong evidence in favor of quantum mechanics and against local hidden variables from a Bayesian analysis. Phys. Rev. A 2019, 99, 022112. [Google Scholar] [CrossRef]
 - Englert, B.-G.; Evans, M.; Jang, G.-H.; Ng, H.-K.; Nott, D.; Seah, Y.-L. Checking the model and the prior for the constrained multinomial. arXiv 2018, arXiv:1804.06906. [Google Scholar]
 - Evans, M.; Guttman, I.; Li, P. Prior elicitation, assessment and inference with a Dirichlet prior. Entropy 2017, 19, 564. [Google Scholar] [CrossRef]
 - Swartz, T. Subjective priors for the Dirichlet process. Commun. Stat. Theory Methods 1993, 28, 2821–2841. [Google Scholar] [CrossRef]
 - Swartz, T. Nonparametric goodness-of-fit. Commun. Stat. Theory Methods 1999, 22, 2999–3011. [Google Scholar] [CrossRef]
 - Ishwaran, H.; Zarepour, M. Exact and approximate sum representations for the Dirichlet process. Can. J. Stat. 2002, 30, 269–283. [Google Scholar] [CrossRef]
 - Antoniak, C.E. Mixtures of Dirichlet processes with applications to Bayesian nonparametric problems. Ann. Stat. 1974, 2, 1152–1174. [Google Scholar] [CrossRef]
 - Doss, H. Bayesian Nonparametric Estimation for Incomplete Data Via Successive Substitution Sampling. Ann. Stat. 1994, 22, 1763–1786. [Google Scholar] [CrossRef]
 - Charvadeh, Y.K.; Yi, G.Y. Data visualization and descriptive analysis for understanding epidemiological characteristics of COVID-19: A case study of a dataset from January 22, 2020 to March 29, 2020. J. Data Sci. 2020, 18, 526–535. [Google Scholar] [CrossRef]
 
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.  | 
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).