Next Article in Journal
Thermodynamics for a Rotating Chiral Fermion System in the Uniform Magnetic Field
Next Article in Special Issue
Introducing Structural Symmetry and Asymmetry Implications in Development of Recent Pharmacy and Medicine
Previous Article in Journal
Visualising Departures from Symmetry and Bowker’s X2 Statistic
Previous Article in Special Issue
Microsolvation of Histidine—A Theoretical Study of Intermolecular Interactions Based on AIM and SAPT Approaches
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Binomial Distributed Data Confidence Interval Calculation: Formulas, Algorithms and Examples

by
Lorentz Jäntschi
1,2
1
Department of Physics and Chemistry, Technical University of Cluj-Napoca, 400641 Cluj, Romania
2
Chemical Doctoral School, Babes-Bolyai University, 400028 Cluj-Napoca, Romania
Symmetry 2022, 14(6), 1104; https://doi.org/10.3390/sym14061104
Submission received: 25 March 2022 / Revised: 11 May 2022 / Accepted: 24 May 2022 / Published: 27 May 2022

Abstract

:
When collecting experimental data, the observable may be dichotomous. Sampling (eventually with replacement) thus emulates a Bernoulli trial leading to a binomial proportion. Because the binomial distribution is discrete, the analytical evaluation of the exact confidence interval of the sampled outcome is a mathematical challenge. This paper proposes three alternative confidence interval calculation methods that are characterized and exemplified.

1. Introduction

While many continuous distributions are known, the list of discrete ones (usually derived from counting) is relatively short. This list becomes even shorter when dealing with dichotomous observables: binomial, hypergeometric, negative binomial, and uniform.
The sampling strategy is foundational when dealing with finite populations. After extracting an individual (case) from a population, it can be selected for insertion in the population, having a non-zero probability of being extracted again in the future; “sampling with replacement” or not back in the population (“sampling without replacement”). Sampling with or without replacement from an infinite population should produce the same effect. For a more complete discussion involving quantiles, please see [1].
While uniform distribution is of essential importance in order statistics [2,3,4], approximating the maximum of a discrete uniform distribution from sampling without replacement for a population of n numbered (from 1 to n) items is the (German tank) problem of estimating n from a random sample [5].
Typical examples of negative binomial distribution include the hospital stay [6], while negative binomial distribution has significance in capturing diversity [7], being introduced by Fisher [8] in the context of sampling without replacement experiment (also known as Fisher’s logseries, [9]). Other related distributions include the Pòlya urn (and the associated Chinese restaurant process) and Indian buffet process [10].
Sampling a dichotomous outcome emulates a Bernoulli trial, and the ratio between the desired (positive) outcome (x) and the number of trials (m) leads to a proportion ( 0 x / m 1 ).
From a population of size m containing x objects of interest, sampling (following a Bernoulli trial, counting successes, x vs. failures, m x ) with replacement leads to a binomial distribution ( f B , Equation (1)), while the alternative—sampling without replacement—leads to the hypergeometric distribution ( f H , Equation (2)).
f B ( y ; x , n , m ) = n y x m y 1 x m n y
f H ( y ; x , n , m ) = x y m x n y / m n
For infinite populations, the number of successes in the population is expressed as a proportion ( p = x / m ) when Gamma function replaces the factorial ( Γ ( w + 1 ) = w ! ).
Dümbgen et al. [11] appropriately present the differences between Hypergeometric and Binomial distributions [11].
The confidence interval is computed from the sample statistics to give a range of values for an unknown population parameter, usually calculated for a confidence level, i.e., the probability with which a random resampling will contain the parameter’s value or the true population value.

2. Background

Considering x as the variable counting a part of a sampled dichotomy (the other part being m x ; for instance, counting the number of successes), confidence intervals will be provided for it.
The probability to sample u successes out of m trials when we have sampled x successes from m trials could be obtained by substituting n = m and y = u in Equations (1) and (2) ( f RB and f RH in Equations (3) and (4)).
f RB ( u ; x , m ) = m u x m u 1 x m m u
f RH ( u ; x , m ) = x u m x m u
Equation (4) imposes simultaneously x u and u x , meaning that u = x and f RH ( u ; x , m ) is a Dirac function of x u (probability mass function f RH ( u ; x , m ) has only one event possible and that is u = x ). This is the expected result, since for the sampling without replacement, sampling the whole population (of size m) always leads to the same number of successes (x) and failures.
The confidence in drawing x successes from m trials should be related to the number of trials (m). Any repetition of the sampling experiment is related to the data from the experiment through Equation (3). All information from the sampling has been used and no more information is available, it could be said that Equation (3) defines a sufficient statistic.
The statistical experiment is as follows: let us have a bag with black and white balls. We draw from the bag a sample of m balls, out of which x are white (a series of m Bernoulli trials). Without a counting function providing the total number of the balls in the bag (finite or not), how many white balls out the m will be extracted if the experiment is repeated, with the result expressed as a 95% rate of success confidence interval.
We will never know what is to come from a draw or a sampling in the case of sampling with replacement Equation (3). A resampling of size m will have a variable number of successes u. However, Equation (3) defines a probability mass function (it is easy to check that Equation (3) is the ( u + 1 ) -th term from ( ( x / m ) + ( 1 x / m ) ) m binomial expansion). A confidence interval supporting drawing x successes from m trials could be constructed around x value, with Equation (3).

3. Literature Review

From the beginning, it should be mentioned that there are two approaches to the same problem. One is to say and to provide “asymptotic” (working better—more precisely—with the increasing of the sample size) intervals with or without “continuity corrections” (patches meant to improve their behavior for small sample sizes) while the other is concerned with “exact” ones.
Each alternative has certain advantages and certain disadvantages. Asymptotic with continuity correction (confidence) intervals will never work well for small samples (a disadvantage) but will, most of the time, have a convenient formula and a relatively simple and fast way of being calculated (an advantage). For a more complete discussion, please see [12].
An exact confidence interval will rarely have a convenient formula (disadvantage) but will always provide some kind of an interval of what it is expected to do (advantage).
Since the previous formulation for the advantage of an exact confidence interval is quite confusing, an explanation is due. It is now clear how “perfect” it should be (for example, having a confidence interval with 5% risk of being in error would be perfect; perfect agreement would be if α = α ^ everywhere), but one has no way of obtaining it (if there are two alternatives, and a person needs to choose between one providing a actual probability of 4.5% and one providing 5.2%), this choice may then only have different opinions about different alternatives bringing the person closer to “perfection” (some may argue that 4.5% is closer to perfection than 5.2% while others may say the opposite), and each of those alternatives can be as exact (what was asked for from the method) as any other.
A lot of studies on improving (and further comparing) asymptotic (with or without continuity correction) methods are reported in the scientific literature.
Comparisons of different methods are given in [13]—seven methods, Ref. [12]—sixteen methods, and in [14]—twenty methods—but the central problem of them is that they are not exact methods.
The formulation of “exact” confidence intervals for proportion started with the Clopper–Pearson method [15], and continued with [16,17,18,19,20]. Recently, a strategy for computing confidence intervals for expressions of two proportions has been proposed [21].
Validity, optimality, and invariance are the desirable properties of constructing confidence intervals [22], but the specifics of their implementation depends on the problem in question. By going further with the strategy implemented in [21], this manuscript presents three alternatives (to being “exact” in the calculation of the confidence intervals for one proportion).
Assessing confidence interval methods may raise problems of different natures, and may be evaluated using various measures of the departure from the imposed level (such as the ones defined in [20]). Some authors propose a set of rules, such as monotony of the boundaries functions in both binomial variable (on half of the interval) and sample size (see [13,23]), or to be the smallest [24].

4. Methodology

In any study, but especially in medical studies [25], it is more important to ensure that the risk of being in error does not surpass the imposed value. If it surpasses it, we must ensure that it is the smallest value possible. The true coverage probability that the interval contains x is plotted against nominal coverage probability (expected or imposed probability that the interval contains x) to show the performances of the confidence interval method. Figure 1 (see also Figure 3 in [26] and Figure 5 in [12]) illustrate the non-coverage probabilities for Wald confidence interval (CI) based on Equation (5).
C I N = [ x κ x ( m x ) / m , x + κ x ( m x ) / m ] , κ = I n v C D F ( Normal , 1 α / 2 )
For m = 100 the “standard” (or “Wald”; “asymptotic interval”, “no continuity correction”). If ϵ = α ^ α is the error, then its standard error, SE( ϵ ) is 0.119.
Figure 2, Figure 3 and Figure 4 illustrates “exact method” with different algorithms at the same specific significance level ( α = 5%) and number of trials (m = 100).
Each of the proposed alternatives (see Figure 3 and Figure 4) is, in a way, an improvement of the normal approximation interval (Equation (5); Figure 3 in [26]; Figure 1 here). Thus, for the actual non-coverage probability ( α ^ , in %) when α = 5% is the imposed level, m = 100 and x = 0, 1, …, m:
  • Figure 1 (asymptotic interval) illustrates α ^ for CI from Equation (5).
  • In Figure 2, using CI from Algorithm 1, always α ^ α (x = 0, 1, …, m) and SE( ϵ ) = 0.127.
  • Figure 1 and Figure 3 (the solutions given by Equation (5) and Algorithm 2) are similar in shape but the amplitude of the oscillation is reduced (SE( ϵ ) from 0.119 to 0.099).
  • In Figure 4 (from Algorithm 3) again α ^ α , and SE( ϵ ) = 0.117 is lower than from Algorithm 1 (0.127).
All proposed methods, as well as the classical Wald method provide symmetry (relative to the middle, p = 0.5, x = 50 on Figure 1, Figure 2, Figure 3 and Figure 4 depicting the true error for a sample of size m = 100). Figure 3 shows a visible improvment in the variation of the error around its imposed level (of 5% in Figure 1, Figure 2, Figure 3 and Figure 4). The solution depicted in Figure 4 is an improvement of the solution depicted in Figure 2 as it has a lower variation of error arround its imposed level.
Althrough providing an accurately calculated confidence interval may closely show what is happening in the population, most researchers are still reticent to use the “exact methods” when caculating confidence intervals. Reputed statisticians suggest that ’approximate is better than exact’ [27]. A better argument for not using the “exact methods” is the complexity of the calculation [21]. The people developing such a method and the researchers supposed to use it are not the same and do not have the same background. In this article, we propose three exact alternatives for calculating the confidence interval to slightly alleviate this inconvenience, examples and illustrations accompany these alternatives.
Following the pattern described by Jäntschi [21] (see Section 5.2 in [21]; the methodology does not change when the distribution is changed), the applied methodology is:
  • Collect all possible drawings and their associated probabilities ( ( u , f RB ( u ; x , m ) ) from Equation (3);
  • The values (u) are already sorted (and grouped) when u ranges increasingly from 0 to m and the probability mass function (PMF) is defined by the { f RB ( 0 ; x , m ) , f RB ( 1 ; x , m ) , …, f RB ( u ; x , m ) , …, f RB ( m ; x , m ) } set;
  • Construct cumulative distribution function (CDF; the result is formally defined by Equation (6)) and (from which) CI in increasing coverage for which the result is formally defined by Equation (7)).
( u , F RB ( u ; x , m ) ) , where F RB ( u ; x , m ) i = 0 u f RB ( i ; x , m )
C I ( i 1 , i 2 ) = [ x i 1 , x + i 2 ] ; p C I ( i 1 , i 2 ) = i = x i 1 x + i 2 f RB ( i ; x , m )
One should notice that the methodology is straightforward and easy to follow, even for a non-specialist. The three proposed alternatives are given when a certain interval needs to be chosen.

Proposed CI Calculation Algorithms

Additionally, one can inspect, for instance, the plots from Figure 2 (solution obtained with Algorithm 1), Figure 3 (solution obtained with Algorithm 2), and Figure 4 (solution obtained with Algorithm 3), or the entries in Table 1 (where [ 1 , 7 ] is obtained with Algorithm 1, [ 2 , 7 ] is obtained with Algorithm 2 and [ 2 , 8 ] with Algorithm 3).
Algorithm 1: Foundational confidence interval, CI_v0
Input: α , x, m    //imposed level, number of successes, number of trials
  •   procedureFound( β , x , m , & r , & i , & j , & q )
  •       i x ; j x ; q r [ x ]
  •      For( ; ; )
  •         if( q β ) Break; if( ( i = 0 ) AND ( j = m ) ) Break
  •         If( i = 0 ) j j + 1 ; q q + r [ j ] ; Continue EndIf
  •         If( j = m ) i i 1 ; q q + r [ i ] ; Continue EndIf
  •         If( r [ i 1 ] = r [ j + 1 ] ) i i 1 ; j j + 1 ; q q + r [ i ] + r [ j ] ; Continue EndIf
  •         If( r [ i 1 ] > r [ j + 1 ] ) i i 1 ; q q + r [ i ] ; Continue EndIf
  •         If( r [ i 1 ] < r [ j + 1 ] ) j j + 1 ; q q + r [ j ] ; Continue EndIf
  •      EndFor
  •   end procedure
  •   procedureCI_v0( α , x , m , & i 1 , & i 2 , & q )
  •     PMF_B( x , m , r ); β 1 α ; Found( β , x , m , r , i 1 , i 2 )
  •   end procedure
Output: i 1 , i 2 , q    // [ i 1 , i 2 ] is the CI; q is the actual CI’s coverage
Algorithm 2: First improved confidence interval, CI_v1
Input: α , x, m    imposed level, number of successes, number of trials
  •   procedureImp_1( β , x , m , & r , & i , & j , & q )
  •      For( ; ; )
  •         if( j i ) Break
  •         If( r [ i ] = r [ j ] )
  •            If( Abs( β q ) > Abs( β q + r [ i ] + r [ j ] ) )
  •                q q r [ i ] r [ j ] ; i i + 1 ; j j 1
  •            EndIf
  •            Break
  •         EndIf
  •         If( r [ i ] < r [ j ] )
  •            If( Abs( β q ) > Abs( β q + r [ i ] ) ) q q r [ i ] ; i i + 1 ; EndIf
  •            Break
  •         EndIf
  •         If( r [ i ] > r [ j ] )
  •            If( Abs( β q ) > Abs( β q + r [ j ] ) ) q q r [ j ] ; j j 1 EndIf
  •            Break
  •         EndIf
  •      EndFor
  •   end procedure
  •   procedureCI_v1( α , x, m, i1, i2, er)
  •      PMF_B( x , m , r ); β 1 α ; Found( β , x , m , r , i 1 , i 2 ); Imp_1( β , x , m , r , i 1 , i 2 , q )
  •   end procedure
Output: i 1 , i 2 , q    // [ i 1 , i 2 ] is the CI; q is the actual CI’s coverage
Algorithm 3: Second improved confidence interval, CI_v2
Input: α , x, m    imposed level, number of successes, number of trials
  •   procedureImp_2( β , x , m , & r , & i , & j , & q )
  •      For( ; ; )
  •         if( q β ) Break; if( ( i = 0 ) AND ( j = m ) ) Break
  •         If( i = 0 ) j j + 1 ; q q + r [ j ] ; Continue EndIf
  •         If( j = m ) i i 1 ; q q + r [ i ] ; Continue EndIf
  •         If( r [ i 1 ] = r [ j + 1 ] ) i i 1 ; j j + 1 ; q q + r [ i ] + r [ j ] ; Continue EndIf
  •         If( r [ i 1 ] < r [ j + 1 ] ) i i 1 ; q q + r [ i ] ; Continue EndIf
  •         If( r [ i 1 ] > r [ j + 1 ] ) j j + 1 ; q q + r [ j ] ; Continue EndIf
  •      EndFor
  •   end procedure
  •   procedureCI_v2( α , x, m, i1, i2, er)
  •      PMF_B( x , m , r ); β 1 α ; Found( β , x , m , r , i 1 , i 2 )
  •      Imp_1( β , x , m , r , i 1 , i 2 , q ); Imp_2( β , x , m , r , i 1 , i 2 , q )
  •   end procedure
Output: i 1 , i 2 , q    // [ i 1 , i 2 ] is the CI; q is the actual CI’s coverage

5. Results and Discussion

Consider one example ( m = 10 and x = 4 ) and follow the methodology described above until Equation (7). Table 2 contains all possible outcomes (sampling with replacement) from the population when m = 10 drawn had x = 4 successes.
Table 3 contains the cumulative distribution function (CDF; Equation (6)) for the data from Table 2.
Table 4 lists true coverage probabilities for the same primary data in Table 2, as the final step (from the CDF given in Table 3) in the common strategy providing the CI (Equation (7)).
As Table 4 reveals, the construction of the confidence interval follows the same strategy as the one used for CDF (from Table 3; F RB ( u ; x , m ) = P ( u x ) , p C I ( i 1 , i 2 ) = P ( i 1 u i 2 ) ) but in the calculation of CI the CDF is not particularly effective ( p C I ( i 1 , i 2 ) = P ( u i 2 ) P ( u < i 1 ) ); a more direct approach is from PMF. The construction of the CI from the draws sorted in descending order of likelihood is the choice making more sense for me (Bayesian reasoning; see [28] for the foundations of Bayesian reasoning and [29] for more details on Bayesian reasoning).
The solution presented in Table 4 is the simplest way to construct CI (Crow was probably the first that proposed the construction of the CI in this manner [23]) but it is merely one alternative. To exemplify, the proposed example should be reconsidered (with x = 4 and m = 10 ) for an imposed level (of non-coverage) α = 0.05. Table (Table 1) presents three alternatives.
The 2nd strategy exemplified in Table 1 should operate on the C I provided by the foundational (1st) strategy and try to improve it by narrowing it, i.e., by trying to increase the lower boundary and trying to decrease the upper boundary such that | α p C I | is minimized. Similarly, 3rd strategy exemplified in Table 1 try to improve the 2nd strategy by broader it (operate to decreases the lower bound and to increase the upper bound such that | α p C I | is minimized).
PMF for a binomial distribution is calculated with Algorithm 1 from [21]. By calling PMF_B ( x , m , r ) the output r is an array indexed from 0 to m and containing the series of probabilities from Equation (3). There are differences between the solutions proposed by the Algorithm 1–3 algorithms (see Figure 5).

5.1. Properties of the Proposed Solutions

Each proposed solution is (anti)symmetrical, not relatively to the observed successes but relatively to its complement compared to half. Specifically, if [ i 1 , i 2 ] is the CI of x successes from m draws, then [ m i 1 , m i 2 ] is the CI of m x successes from m draws. This property is available for the normal approximation interval as well (Equation (5)). It can be easily checked that if C I N ( x ) = [ C I L N ( x ) , C I U N ( x ) ] = [ x κ x ( m x ) / m , x + κ x ( m x ) / m ] (see Equation (5)), then substituting x with m x , C I N ( m x ) = [ C I L N ( m x ) , C I U N ( m x ) ] = [ m x κ x ( m x ) / m , m x + κ x ( m x ) / m ] and C I L N ( m x ) + C I U N ( x ) = m = C I U N ( m x ) + C I L N ( x ) and the actual coverage probability is the same (see Figure 1, Figure 2, Figure 3 and Figure 4; all are symmetrical). When considering the proposed solutions (Algorithms 1–3) the (anti)symmetry is kept due to the presence of the If(r[i-1]=r[j+1])…EndIf (Algorithms 1 and 3), and If(r[i]=r[j])…EndIf (Algorithm 2) instructions blocks. Additionally, an important remark could be made about the Normal approximation interval (Equation (5)). This interval is actually (as defined by Equation (5)) even more symmetrical in relation to the observed number of successes (x), but if anyone considers the confidence intervals proposed by Equation (5) to be a very small (such as is x = 1, 2, …) or very large (such as is x = m 1 , m 2 , …) number of successes (see Table 5), it will be seen that a bound greater than m or smaller than 0 is not logical; therefore, it must be immediately patched (Equation 8).
C I N P ( x , m , α ) = [ max ( 0 , C I L N ( x , m , α ) ) , min ( m , C I U N ( x , m , α ) ) ]
All integer boundaries formulas (column C I N N ( x ) in Table 5) can be easily obtained (Equation (9), where · is the ceil function and · is the floor function) from Equation (8).
C I N N ( x , m , α ) = [ C I L N P ( x , m , α ) , C I U N P ( x , m , α ) ]
C I N P ( x , m , α ) and C I N N ( x , m , α ) possess exactly the same (anti)symmetrical properties like the proposed ones (Algorithms 1–3).
A consequence derives from this property: it is sufficient to know one bound for all range of successes in order to know the other bound as well. For convenience, three series are listed in the Appendix A (Table A1, Table A2 and Table A3) of lower boundaries for the proposed CI alternatives. In each case, if ( N i ) 0 i m is the series given, then the confidence interval of x = i is [ N i , m N m i ] . Each table presents the run of one algorithm, so the solution proposed by individual algorithms (Table A1 for Algorithm 1, Table A2 for Algorithm 2, and Table A3 for Algorithm 3). The algorithms were implemented in PHP language and the outputs are given. There is no serious constraint in the computation, other than memory limits of the computer; see Section 5.3 for details. An example is suitable here. It would be useful to know the smallest sample size for which the proposed algorithms provide different solutions. That is m = 8 . The CI calculated from the N i numbers are given in Table 6.
Please note that when m is very small the whole probability space is small (Equation (3) defines a ( m + 1 ) × ( m + 1 ) matrix) so any method (optimized or not) have to choose from a very small set of possibble choices. This peculiar behavior is visible in the Appendix A Table A1, Table A2, Table A3 and Table A4 as well: the entries for m = 2 and m = 3 are identical and until m = 8 it is in one way or another an overlap between the proposed solutions.

5.2. Smoothing of the Proposed Solutions

Smoothing a data set is typically done to create an approximating function that captures important patterns in the data, leaving out noise or other fine-scale variations [30]. In this instance, smoothing is proposed to be used to introduce noise, or a fine-scale variation, to make the confidence interval more similar to the normal approximation (Equation (5); to all intents and purposes like its patch Equation (8). Ergo, someone may say that C I N N is the anti-smoothed version of the patched normal approximation confidence interval ( C I N P ).
Considering that two adjacent intervals that natively do not intersect will still not intersect after smoothing. It means that, by smoothing, they should not be enlarged with more than a half of 1/m. Also, smoothing should balance the transition from one number of successes (x) to adjacent numbers of successes ( x 1 and x + 1 ). An interval is made with common sense by picking its limits between two integers such that the ratio of the probabilities of their extraction is also the ratio in which the limit divides the allotted distance between them (0.5, Equation (10).
If [ i 1 , i 2 ] non-smoothed interval , then [ x 1 , x 2 ] smoothed interval , with : x 1 = 0 , if i 1 = 0 i 1 0.5 f RB ( i 1 , x , m ) f RB ( i 1 , x , m ) + f RB ( i 1 1 , x , m ) , otherwise x 2 = m , if i 2 = m i 2 + 0.5 f RB ( i 2 , x , m ) f RB ( i 2 , x , m ) + f RB ( i 2 + 1 , x , m ) , otherwise
Smoothing with Equation (10) does not change the coverage probability (the coverage probabilities of [ i 1 , i 2 ] and [ x 1 , x 2 ] are equal) because the change is too small, smaller than the increment between the success events.

5.3. General Discussion

What if one follows the (numerous) sets of rules, such as having a confidence interval of which boundaries are monotonic functions? It may be checked that the solutions proposed by Algorithms 1 and 2 are both monotonic (all the data given in Table A1 and Table A2. A screening conducted for m from 21 to 100 also supported the monotony. So, in actuality, two proposed solutions follow some rules (another rule implemented in the algorithm is that the confidence interval is constructed from redrawn successes taken in descending order of the probabilities—most likely first).
The series of numbers giving one boundary may be inspected in more details. Consider the series for m = 25 and α = 0.05 provided by Algorithm 1: N = {0, 0, 0, 0, 1, 2, 2, 3, 4, 5, 6, 7, 8, 8, 9, 10, 11, 12, 14, 15, 16, 18, 19, 21, 22, 25}. Independently of the algorithm, sample size, and imposed significance level, the first ( N 0 ) and last ( N m ) are known ( N 0 = 0, N m = m) as well as actual non-coverage probabilities associated with a u = 0 and u = m draw (0%, [ 0 , 0 ] and [ m , m ] intervals cover all possible cases, see Equation (3). Since CI is [ N i , m N m i ] (see Section 5.1) bigger numbers in the series means narrower confidence intervals (the bounds are then closer) while smaller numbers in the series mean larger confidence intervals (the bounds are then farther). This is an important property which allows us to simplify the reasoning.
Nevertheless, someone may argue that this is not enough (see [17] patched in [18]), so what is happening when arbitrary (but with a sort of common sense) rules are imposed (see Equation 1.3 and 1.4 in [17]) should be analyzed. Moving from the CI for m = 25 to the CI for m = 26 (whenever; applicable for any succession), one may argue that the lower bound of the CIs for x successes from m + 1 draws should not be greater than the bound of the CI for x successes from m draws (the general trend in the columns in Table A1 and Table A2 should be considered). As result, once you have constructed the series of N for m + 1 from the series of N for m, all that is needed is to add the new sample size ( m + 1 ) at the end, thus getting a series of bounds which are (the first ones) following the defined rule and may contain numbers generally higher than the desired ones. A such series should be labeled N S ; N S series for m = 26 and α = 0.05 is N S = {0, 0, 0, 0, 1, 2, 2, 3, 4, 5, 6, 7, 8, 8, 9, 10, 11, 12, 14, 15, 16, 18, 19, 21, 22, 25, 26}). By following an exactly mirrored reasoning, by adding a 0 at the beginning, the result is a series of too large but still rule complying series. A such series may contain numbers generally smaller than the desired ones and this series should be labeled N I ( N I series for m = 26 and α = 0.05 is N I = {0, 0, 0, 0, 0, 1, 2, 2, 3, 4, 5, 6, 7, 8, 8, 9, 10, 11, 12, 14, 15, 16, 18, 19, 21, 22, 25}).
The N I and N S series defines a trap for any confidence interval following an extended monotony rule (if N N is the best option then N S x N N x N I x for 0 x m ). Why not find the best one? It is a matter of extensive search and the algorithm is of no interest, since its result is a failure. Figure 6 depicts the “optimal” solution for m = 31, α = 0.05 providing the smallest departure between the imposed and the actual non-coverage probability, when the intervals were constructed iteratively starting from the series of N for m = 1 ( N = { 0 , 1 } ); raw data in Table A4. For comparison, the Figure 7, Figure 8 and Figure 9 show the confidence intervals obtained from the proposed strategies followed by the illustration of actual non-coverage probabilities (Figure 10, Figure 11, Figure 12 and Figure 13).
As can be observed from the above figures, the confidence interval boundaries of Algorithms 1 and 2 are monotonic, while those of Algorithm 3 are not (see Figure 9 left). Nevertheless, the confidence interval itself is monotonic and symmetric from the middle (see Figure 9 right). In contrast, the (supposed) optimum confidence interval built on monotonies (see Figure 6 right) is symmetric but not monotonic relative to the center.
Comparing Figure 6 with Figure 8 and Figure 10 with Figure 12 it can be observed that the monotonic increase of the boundaries leads to erroneous confidence intervals in the middle (Figure 6 vs. Figure 8), and is departing from the optimal (see SE( ϵ ) for Figure 10 vs. Figure 12).
The standard error of the estimate (SE) as function of the sample size for samples from 10 to 100 shows the expected behavior of increasing precision with sample size (Figure 14).
In retrospect, the improvement proposed by Algorithm 3 vs. Algorithms 1 and 2 seems insignificant (see SE in Figure 12 vs. in Figure 13) in relation to the effects—the intervals, even if they are of monotonous width relative to the middle (Figure 8 right and Figure 9 right), and are still in a zig-zag form (see Figure 8 left and Figure 9 left).
Once we have the confidence interval for a binomially distributed variable (such as the ones proposed by the Algorithm 1 to Algorithm 3) we can derive the confidence interval for the proportion (Equation 11).
If [ x 1 , x 2 ] confidence interval for x successes from m trials Then x 1 m , x 2 m confidence interval for x m
Roughly the same calculations are involved when the binomial distribution (Equation (3)) is replaced with multinomial distribution (Equation (12)) thus permitting the usual factorial analysis [31] to be enriched with statistical significance.
f M ( u 1 , , u k ; x 1 , , x k ) i = 1 k x i u i i = 1 k u i ! m ! m m ; m i = 1 k u i = i = 1 k x i
The multinomial distribution proportion is a generalization from the binomial distribution proportion (Equation (12); when k 2 Equation (12) becomes Equation (3). Further research is required to adapt the procedures for binomial distributed samples confidence intervals calculation for multinomial distributed variables and proportions.
Since Poisson distribution can be applied to systems with a large number of possible events, each of which is rare, a future work is to adapt the approach for it.

6. Conclusions

A strategy to provide exact confidence intervals for binomial distributed variables is elaborated. Three algorithms for calculation of three alternate confidence intervals are given with examples. All proposed alternatives provide antisymmetrical confidence intervals with symmetrical true errors improving the classical asymptotic method.

Funding

This research received no external funding.

Data Availability Statement

Implementation of the algorithms in PHP is available upon request and a future work of the author intends to make it freely available.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A

Table A1. Numbers providing the confidence intervals from Algorithm 1 at α = 0.05: C I ( α , x , m ) = [ N x , m N m x ] .
Table A1. Numbers providing the confidence intervals from Algorithm 1 at α = 0.05: C I ( α , x , m ) = [ N x , m N m x ] .
m+++20191817161514131211109876543210x
101+201716141311109876543221000020
2002+1916151312109876543321000019
30013+18151412119876543321000018
400014+1714131110876543321000017
5000125+16131210976544321000016
60001236+151211986544321000015
700012347+14111087643321000014
8000112458+131097654321100013
90001223569+1298654321100012
10000112346710+118754321100011
x012345678910+++++++++++++m
Table A2. Numbers providing the confidence intervals from Algorithm 2 at α = 0.05: C I ( α , x , m ) = [ N x , m N m x ] .
Table A2. Numbers providing the confidence intervals from Algorithm 2 at α = 0.05: C I ( α , x , m ) = [ N x , m N m x ] .
m+++20191817161514131211109876543210x
101+201816141311109876544321100020
2002+1917151312119876544321100019
30013+181614121110876554321100018
400024+1715131110976554321100017
5000235+16141210986554321100016
60001246+151311987644321100015
700012357+14121087654321100014
8000123468+131197654321100013
90001223579+12108654322100012
10000122346810+119754322100011
x012345678910+++++++++++++m
Table A3. Numbers providing the confidence intervals from Algorithm 3 at α = 0.05: C I ( α , x , m ) = [ N x , m N m x ] .
Table A3. Numbers providing the confidence intervals from Algorithm 3 at α = 0.05: C I ( α , x , m ) = [ N x , m N m x ] .
m+++20191817161514131211109876543210x
101+20171613131197876544321100020
2002+1916151212119776544311100019
30013+181514111110865554311100018
400014+1714131010975454311100017
5000125+1613129986544311100016
60001236+151211887644211100015
700012347+14111077653221100014
8000013458+131097654321100013
90001223569+1298604321100012
10000122246710+118750321100011
x012345678910+++++++++++++m
Table A4. Numbers providing the confidence intervals with monotonic boundaries constrain at α = 0.05: C I ( α , x , m ) = [ N x , m N m x ] .
Table A4. Numbers providing the confidence intervals with monotonic boundaries constrain at α = 0.05: C I ( α , x , m ) = [ N x , m N m x ] .
m∖x012345678910111213141516171819202122232425262728293031
2002
30013
400024
5000235
60001246
700012357
8000123468
90000224579
10000022256810
110000222267911
12000022222781012
130000222222891113
14000022222229101214
1500002222222210111315
16000012222222911121416
1700001222222291012131517
180000112222229101113141618
19000011222222910111214151719
2000001112222291011121315161820
210000111122229101112131416171921
22000011111222910111213141517182022
230000111111228911121314151618192123
24000011111122891011131415161719202224
2500001111112279101112141516171820212325
260000111111227810111213141617181921222426
27000011111122671011121314151718192022232527
280000111111226791112131415161819202123242628
29000011111122667111213141516171820212224252729
3000001111112256610121314151617181921222325262830
31000011111122256612131415161718192022232426272931

References

  1. Ialongo, C. Confidence interval for quantiles and percentiles. Biochem. Med. 2019, 29, 010101. [Google Scholar] [CrossRef]
  2. Jäntschi, L.; Bolboacă, S.-D. Performances of Shannon’s Entropy Statistic in Assessment of Distribution of Data. Ovidius Univ. Ann. Chem. 2017, 28, 30. [Google Scholar] [CrossRef] [Green Version]
  3. Jäntschi, L. A test detecting the outliers for continuous distributions based on the cumulative distribution function of the data being tested. Symmetry 2019, 11, 835. [Google Scholar] [CrossRef] [Green Version]
  4. Jäntschi, L. Detecting extreme values with order statistics in samples from continuous distributions. Mathematics 2020, 8, 216. [Google Scholar] [CrossRef] [Green Version]
  5. Ruggles, R.; Brodie, H. An Empirical Approach to Economic Intelligence in World War II. J. Am. Stat. Assoc. 1947, 42, 72. [Google Scholar] [CrossRef]
  6. Carter, E.-M.; Potts, H.-W.-W. Predicting length of stay from an electronic patient record system: A primary total knee replacement example. BMC Med. Inform. Decis. Mak. 2014, 14, 26. [Google Scholar] [CrossRef] [Green Version]
  7. Jäntschi, L. Distribution fitting 16. How many colors are actually in the field? Buasvmcn. Hortic. 2012, 69, 184. [Google Scholar]
  8. Fisher, R.A. The relation between the number of species and the number of individuals in a random sample of an animal population. Part 3. A Theoretical distribution for the apparent abundance of different species. J. Anim. Ecol. 1943, 12, 54. [Google Scholar] [CrossRef]
  9. Devaurs, D.A.; Gras, R. Species abundance patterns in an ecosystem simulation studied through Fisher’s logseries. Simul. Model. Pract. Theor. 2010, 18, 100–123. [Google Scholar] [CrossRef] [Green Version]
  10. Frigyik, B.A.; Kapila, A.; Gupta, M.R. Introduction to the Dirichlet Distribution and Related Processes: UWEE Technical Report UWEETR-2010-0006; University of Washington: Seattle, WA, USA, 2010; pp. 18–25. [Google Scholar]
  11. Dümbgen, L.; Samworth, R.J.; Wellner, J.A. Bounding distributional errors via density ratios. Bernoulli 2021, 27, 818–852. [Google Scholar] [CrossRef]
  12. Bolboacă, S.-D.; Achimaş-Cadariu, B.-A. Binomial Distribution Sample Confidence Intervals Estimation 2. Proportion-like Medical Key Parameters. Leonardo Electron. J. Pract. Technol. 2003, 3, 75. [Google Scholar]
  13. Newcombe, R.-G. Two–sided confidence intervals for the single proportion: Comparison of seven methods. Stat. Med. 1998, 17, 857. [Google Scholar] [CrossRef]
  14. Pires, A.-M.; Amado, C. Interval estimators for a binomial proportion: Comparison of twenty methods. REVSTAT Stat. J. 2008, 6, 165. [Google Scholar]
  15. Clopper, C.; Pearson, E.-S. The use of confidence or fiducial limits illustrated in the case of the binomial. Biometrika 1934, 26, 404. [Google Scholar] [CrossRef]
  16. Sterne, T.E. Some Remarks on Confidence or Fiducial Limits. Biometrika 1954, 41, 275. [Google Scholar]
  17. Blyth, C.R.; Still, H.A. Binomial confidence intervals. J. Am. Stat. Assoc. 1983, 78, 108. [Google Scholar] [CrossRef]
  18. Casella, G. Refining Binomial Confidence Intervals. Can. J. Stat. 1986, 14, 113. [Google Scholar] [CrossRef] [Green Version]
  19. Blaker, H. Confidence curves and improved exact confidence intervals for discrete distributions. Can. J. Stat. 2000, 28, 783. [Google Scholar] [CrossRef] [Green Version]
  20. Bolboacă, S.-D.; Jäntschi, L. Optimized confidence intervals for binomial distributed samples. Int. J. Pure Appl. Math. 2008, 47, 1. [Google Scholar]
  21. Jäntschi, L. Formulas, Algorithms and Examples for Binomial Distributed Data Confidence Interval Calculation: Excess Risk, Relative Risk and Odds Ratio. Mathematics 2021, 9, 2506. [Google Scholar] [CrossRef]
  22. Sprott, D.A. What Is Optimality in Scientific Inference? Lect. Notes Monog. Ser. 2004, 44, 133–152. [Google Scholar]
  23. Crow, E.L. Confidence Intervals for a Proportion. Biometrika 1956, 43, 423. [Google Scholar] [CrossRef]
  24. Wang, W. Smallest confidence intervals for one binomial proportion. J. Stat. Plan. Inference 2006, 136, 4293. [Google Scholar] [CrossRef]
  25. Eisenstein, J.; Chang, D.I. Case study: Improving population and individual health through health system transformation in Washington state. NAM Perspectives. Nat. Acad. Med. 2017, 4, 201704e. [Google Scholar] [CrossRef]
  26. Brown, L.-D.; Cai, T.-T.; DasGupta, A. Interval Estimation for a Binomial Proportion. Stat. Sci. 2001, 16, 101. [Google Scholar] [CrossRef]
  27. Agresti, A.; Coull, B.-A. Approximate is Better than “Exact” for Interval Estimation of Binomial Proportions. Am. Stat. 1998, 52, 119–126. [Google Scholar]
  28. Bayes, T. An Essay Towards Solving a Problem in the Doctrine of Chances. Philos. Trans. R. Soc. Lond. 1763, 50, 370. [Google Scholar] [CrossRef]
  29. Fienberg, S.-E. When Did Bayesian Inference Become “Bayesian”? Bayesian Anal. 2006, 1, 1. [Google Scholar] [CrossRef]
  30. Vasil’ev, V.; Vasilyeva, M. An Accurate Approximation of the Two-Phase Stefan Problem with Coefficient Smoothing. Mathematics 2020, 8, 1924. [Google Scholar] [CrossRef]
  31. Jäntschi, L.; Bálint, D.; Pruteanu, L.-L.; Bolboacă, S.-D. Elemental factorial study on one–cage pentagonal faces nanostructure congeners. Mater. Discov. 2016, 5, 14. [Google Scholar] [CrossRef]
Figure 1. Actual non-coverage probability ( α ^ , in %) as function of the binomial random variable (x) for the asymptotic interval.
Figure 1. Actual non-coverage probability ( α ^ , in %) as function of the binomial random variable (x) for the asymptotic interval.
Symmetry 14 01104 g001
Figure 2. Actual non-coverage probability ( α ^ , in %) as function of the binomial random variable (x) for the first “exact method” proposed (see below).
Figure 2. Actual non-coverage probability ( α ^ , in %) as function of the binomial random variable (x) for the first “exact method” proposed (see below).
Symmetry 14 01104 g002
Figure 3. Actual non-coverage probability ( α ^ , in %) as function of the binomial random variable (x) for the second “exact method” proposed (see below).
Figure 3. Actual non-coverage probability ( α ^ , in %) as function of the binomial random variable (x) for the second “exact method” proposed (see below).
Symmetry 14 01104 g003
Figure 4. Actual non-coverage probability ( α ^ , in %) as function of the binomial random variable (x) for the third “exact method” proposed (see below).
Figure 4. Actual non-coverage probability ( α ^ , in %) as function of the binomial random variable (x) for the third “exact method” proposed (see below).
Symmetry 14 01104 g004
Figure 5. Lower and upper bound of four alternatives to construct the CI: Normal (Equation (5)) and the three proposed algorithms (Algorithms 1–3) when m = 10 , x = 0 , 1 , , 10 and α = 0.05 .
Figure 5. Lower and upper bound of four alternatives to construct the CI: Normal (Equation (5)) and the three proposed algorithms (Algorithms 1–3) when m = 10 , x = 0 , 1 , , 10 and α = 0.05 .
Symmetry 14 01104 g005
Figure 6. Optimum CIs when monotony rule is fully enacted for m = 31 and α = 0.05 (bounds in the left hand image and width in the right hand image).
Figure 6. Optimum CIs when monotony rule is fully enacted for m = 31 and α = 0.05 (bounds in the left hand image and width in the right hand image).
Symmetry 14 01104 g006
Figure 7. Algorithm 1 CIs for m = 31 and α = 0.05 (bounds in the left hand image and width in the right hand image).
Figure 7. Algorithm 1 CIs for m = 31 and α = 0.05 (bounds in the left hand image and width in the right hand image).
Symmetry 14 01104 g007
Figure 8. Algorithm 2 CIs for m = 31 and α = 0.05 (bounds in the left hand image and width in the right hand image).
Figure 8. Algorithm 2 CIs for m = 31 and α = 0.05 (bounds in the left hand image and width in the right hand image).
Symmetry 14 01104 g008
Figure 9. Algorithm 3 CIs for m = 31 and α = 0.05 (bounds in the left hand image and width in the right hand image).
Figure 9. Algorithm 3 CIs for m = 31 and α = 0.05 (bounds in the left hand image and width in the right hand image).
Symmetry 14 01104 g009
Figure 10. Optimum CIs when monotony rule is fully enacted: α ^ (in %) when α = 5% is the imposed level (m = 31; x = 0, 1, …, m; CI in Figure 6; α ^ is probably the smoothest one, SE( ϵ ) = 0.301).
Figure 10. Optimum CIs when monotony rule is fully enacted: α ^ (in %) when α = 5% is the imposed level (m = 31; x = 0, 1, …, m; CI in Figure 6; α ^ is probably the smoothest one, SE( ϵ ) = 0.301).
Symmetry 14 01104 g010
Figure 11. Algorithm 1 CIs: α ^ (in %) when α = 5% is the imposed level (m = 31; x = 0, 1, …, m; α ^ α for all x = 0, 1, …, m; SE( ϵ ) = 0.331).
Figure 11. Algorithm 1 CIs: α ^ (in %) when α = 5% is the imposed level (m = 31; x = 0, 1, …, m; α ^ α for all x = 0, 1, …, m; SE( ϵ ) = 0.331).
Symmetry 14 01104 g011
Figure 12. Algorithm 2 CIs: α ^ (in %) when α = 5% is the imposed level (m = 31; x = 0, 1, …, m; α ^ better than the one from Figure 10, SE( ϵ ) = 0.288).
Figure 12. Algorithm 2 CIs: α ^ (in %) when α = 5% is the imposed level (m = 31; x = 0, 1, …, m; α ^ better than the one from Figure 10, SE( ϵ ) = 0.288).
Symmetry 14 01104 g012
Figure 13. Algorithm 3 CIs: α ^ (in %) when α = 5% is the imposed level (m = 31; x = 0, 1, …, m; α ^ α for all x = 0, 1, …, m; α ^ is better than the one from Figure 10, SE( ϵ ) = 0.287).
Figure 13. Algorithm 3 CIs: α ^ (in %) when α = 5% is the imposed level (m = 31; x = 0, 1, …, m; α ^ α for all x = 0, 1, …, m; α ^ is better than the one from Figure 10, SE( ϵ ) = 0.287).
Symmetry 14 01104 g013
Figure 14. Standard error of α ^ (in %) for CIs from Algorithms 1–3, when α = 5% is the imposed level (m ranges from 10 to 100).
Figure 14. Standard error of α ^ (in %) for CIs from Algorithms 1–3, when α = 5% is the imposed level (m ranges from 10 to 100).
Symmetry 14 01104 g014
Table 1. Alternate CIs for replica ( u , m ) of ( x , m ) draw.
Table 1. Alternate CIs for replica ( u , m ) of ( x , m ) draw.
C I ( i 1 , i 2 ) p C I ( i 1 , i 2 ) Strategy
[ 1 , 7 ] 0.9817Descending probabilities (foundational strategy)
[ 2 , 7 ] 0.94141st Improvement—decrease such that: | α p C I | m i n .
[ 2 , 8 ] 0.95202nd Improvement—increase such that: | α p C I | m i n .
The first CI of (x = 4, m = 10) (is constructed by inclusion of the likelihood values (fRB(u; 4, 10)) in a decreasing order. The 1st improvement proposes a shorter interval with a non-coverage probability (1 − 0.9414 = 0.0586) greater than the imposed level (0.0500), but with a much better proximity to it (1 − 0.9817 = 0.0183). The 2nd improvement expands the 1st improvement and provides an even better proximity (1 − 0.9520 = 0.0480).
Table 2. PMF example for a replica ( u , m ) of a ( x , m ) draw ( f RB from Equation (3)).
Table 2. PMF example for a replica ( u , m ) of a ( x , m ) draw ( f RB from Equation (3)).
u f RB ( u ; 4 , 10 ) u f RB ( u ; 4 , 10 )
00.006060.1115
10.040370.0425
20.120980.0106
30.215090.0016
40.2508100.0001
50.2007Any1.0000
PMF for (u, m = 10) resampling (with replacement) of (x = 4, m = 10) is given {(u, fRB(u; 4, 10)), 0 ≤ u ≤ 10}. It can be observed that u = x is the most probable event. The numeric values of the probabilities are given with four significant decimals. However, the calculations and the actual values are recommended to be done with machine-like precision (see Table 2 in [4] and Section 5.3 in [21] for an extended explanation).
Table 3. CDF example for a replica ( u , m ) of a ( x , m ) draw ( F RB from Equation (6)).
Table 3. CDF example for a replica ( u , m ) of a ( x , m ) draw ( F RB from Equation (6)).
u F RB ( u ; 4 , 10 ) u F RB ( u ; 4 , 10 )
00.006060.9452
10.046370.9877
20.167280.9983
30.382290.9999
40.6330101.0000
50.8337
CDF for (u, m = 10) resampling (with replacement) of (x = 4, m = 10) is given {(u, fRB(u; 4, 10)), 0 ≤ u ≤ 10}.
Table 4. CIs for any replica ( u , m ) of a ( x , m ) draw ( C I i 1 , i 2 from Equation (7)).
Table 4. CIs for any replica ( u , m ) of a ( x , m ) draw ( C I i 1 , i 2 from Equation (7)).
i 1 i 2 C I ( i 1 , i 2 ) p C I ( i 1 , i 2 ) i 1 i 2 C I ( i 1 , i 2 ) p C I ( i 1 , i 2 )
44[4, 4]0.250817[1, 7]0.9817
34[3, 4]0.465818[1, 8]0.9923
35[3, 5]0.666508[0, 8]0.9983
25[2, 5]0.787409[0, 9]0.9999
26[2, 6]0.8989010[0, 10]1
27[2, 7]0.9414
CI for any (u, m = 10) resampling (with replacement) of (x = 4, m = 10) is constructed by inclusion in decreasing value of the likelihood (fRB(u; 4, 10)). The values of the probabilities are given with four significant decimals.
Table 5. Patching the normal CIs for binomial data.
Table 5. Patching the normal CIs for binomial data.
x 1 pCI ( x ) C I N ( x ) C I N P ( x ) C I N N ( x )
10.0702[−0.9 , + 2.9 ] [ 0.0 , 2.9 ] [ 0 , 2 ]
20.0328[−0.5 , + 4.5 ] [ 0.0 , 4.5 ] [ 0 , 4 ]
30.0756 [ + 0.2 , + 5.8 ] [ 0.2 , 5.8 ] [ 1 , 5 ]
70.0756 [ + 4.2 , + 9.8 ] [ 4.2 , 9.8 ] [ 4 , 9 ]
80.0328 [ + 5.5 , 10.5 ] [ 5.5 , 10.0 ] [ 5 , 10 ]
90.0702 [ + 7.1 , 10.9 ] [ 7.1 , 10.0 ] [ 7 , 10 ]
In this case, patching is exemplified for α = 0.05, m = 10 and x = 0, 1, 2. Coverage probability (pCI(x)) is not changed due to patching. CIN(x) is the CI calculated with Equation (5); one decimal digit is given. Since only an integer between 0 and m = 10 can be recorded as number of successes, CINP(x) is the adjustment of the CI calculated with Equation (5) and adjusted to make sense at limits (0 and m = 10); one decimal digit is given. Furthermore, since only a natural number of successes can be recorded, a decimal number as boundary may be implausible, so CINP(x) is the adjustment of the CINP(x) to have integer boundaries.
Table 6. Example of distinct solutions proposed by the algorithms.
Table 6. Example of distinct solutions proposed by the algorithms.
Bin N i CI α ^ , in %
0000[0, 0][0, 0][0, 0]0.000.000.00
1000[0, 3][0, 2][0, 3]1.126.731.12
2000[0, 4][0, 4][0, 4]2.732.732.73
3110[1, 6][1, 5][0, 5]2.895.933.60
4121[1, 7][2, 6][1, 7]0.787.030.78
5233[2, 7][3, 7][3, 8]2.895.933.60
6444[4, 8][4, 8][4, 8]2.732.732.73
7565[5, 8][6, 8][5, 8]1.126.731.12
8888[8, 8][8, 8][8, 8]0.000.000.00
xA1A2A3A1A2A3A1A2A3
It should be noted that the proposed solutions corresponds to m = 8 and α = 5%, so the expected value for each entry in α ^ column is 5.00%.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jäntschi, L. Binomial Distributed Data Confidence Interval Calculation: Formulas, Algorithms and Examples. Symmetry 2022, 14, 1104. https://doi.org/10.3390/sym14061104

AMA Style

Jäntschi L. Binomial Distributed Data Confidence Interval Calculation: Formulas, Algorithms and Examples. Symmetry. 2022; 14(6):1104. https://doi.org/10.3390/sym14061104

Chicago/Turabian Style

Jäntschi, Lorentz. 2022. "Binomial Distributed Data Confidence Interval Calculation: Formulas, Algorithms and Examples" Symmetry 14, no. 6: 1104. https://doi.org/10.3390/sym14061104

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop