Next Article in Journal
Bootstrap Assessment of Crop Area Estimates Using Satellite Pixels Counting
Next Article in Special Issue
A Log-Det Heuristics for Covariance Matrix Estimation: The Analytic Setup
Previous Article in Journal
Has the Market Started to Collapse or Will It Resist?
Previous Article in Special Issue
Properties and Limiting Forms of the Multivariate Extended Skew-Normal and Skew-Student Distributions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Omnibus Tests for Multiple Binomial Proportions via Doubly Sampled Framework with Under-Reported Data

Department of Defense, Fort Meade, MD 20755, USA
Disclaimer Statement: This research represents the author’s own work and opinion. It does not reflect any policy nor represent the official position of the U.S. Department of Defense nor any other U.S. Federal Agency.
Stats 2022, 5(2), 408-421; https://doi.org/10.3390/stats5020024
Submission received: 27 March 2022 / Revised: 18 April 2022 / Accepted: 20 April 2022 / Published: 23 April 2022
(This article belongs to the Special Issue Multivariate Statistics and Applications)

Abstract

:
Previously, Rahardja (2020) paper (in the first reference list) developed a (pairwise) multiple comparison procedure (MCP) to determine which (proportions) pairs of Multiple Binomial Proportions (with under-reported data), the significant differences came from. Generally, such an MCP test (developed by Rahardja, 2020) is the second part of a two-stage sequential test. In this paper, we derived two omnibus tests (i.e., the overall equality of multiple proportions test) as the first part of the above two-stage sequential test (with under-reported data), in general. Using two likelihood-based approaches, we acquire two Wald-type (Omnibus) tests to compare Multiple Binomial Proportions (in the presence of under-reported data). Our closed-form algorithm is easy to implement and not computationally burdensome. We applied our algorithm to a vehicle-accident data example.

1. Introduction

As stated in the Abstract, this paper is the continuation of Rahardja (2020) paper [1]. In regard to counts-and-proportions (categorical) type of data, the topic of interest here is about miscategorization of categorical data, particularly, binomial outcome data. For example, counts of disease/no-disease cancer screening in a hospital, proportions of male/female vehicle-driver accidents in an insurance company, counts of correct/incorrect audited payments in an accounting firm, etc. In an ideal (error-free) world, there is no miscategorization. Then, the simple and straightforward estimators (counts-and-proportions) taught in an introductory statistics (Stat 101) course can just be applied. However, the fact that in this imperfect world, errors or flaws (miscategorization) exist and cannot be ignored because such presence of under-reported or over-reported errors may create severe bias [2,3]. The fact that such imperfections exist in this world (due to the presence of under-reported or over-reported errors) is the motivation behind this research: to remedy the bias by incorporating a doubly sampled framework and to achieve model identifiability. Consequently (as a brief overview), our objective here is to provide a (bias remedy) method, by using a doubly sampled framework, in the presence of either under-reported or over-reported data. In the next paragraphs, we will go over the bibliography and novelty of this research.
In reality, we frequently run into data structure of a binomial output variable and a nominal input variable with multiple groups or levels, within that nominal input variable. In the scenario that both the binomial output variable and the nominal input variable are perfectly counted with no mistake, one can just use the Pearson chi-square statistics (taught in a Stat 101 course) to test the relationship between the input and output variables. Nevertheless, there are data cases where the nominal input variable (for example, a patient-category nominal input variable with g = 3 levels: male/female/pediatric) is categorized flawlessly with no error, while the binomial/binary output variable (e.g., disease/no-disease) is categorized using an error-prone classifier. Such error-prone classifier can misclassify a disease-free (male/female/pediatric) patient into disease output category and vice versa. Therefore, the objective is to examine whether there is any relationship between the nominal input variable (patient-category, with three levels: male/female/pediatric) and the binomial output variable. In other words, to test if the patient-category matters (i.e., testing whether the proportions of the three patient-category are the same, hence, H0: pmale = pfemale = ppediatric) in determining the output (disease-free/not). Among many past research, Bross in 1954 [2] and Goldberg in 1975 [3] examined that with error-prone classifier, the bias can be severe, in their case of determining the odds ratio for misclassified binomial data with two levels (g = 2) of input variables. Subsequently, to adjust for the bias, a sampling framework is needed to obtain valid estimators.
In the case that an error-free classifier is also accessible for acquiring the binomial output variable, one can use a doubly sampled framework, pioneered by Tenenbein in 1970 [4], to reach model identifiability and to remedy bias. Here, is how the doubly sampled framework operates. First, one chooses either an arbitrary subset from the main data or another new sample. In the former scenario where one chooses an arbitrary subset from the main data, the error-free classifier is utilized to attain the correct measurement of the binomial output variable for such subset. In the second scenario, one chooses another new sample, and utilize both the error-prone and the error-free classifier to attain the binomial output variable for such new sample. Additionally, one also measures the nominal input variable for this new sample. To put it another way, the subset or new sample can contain three categorical variables: a nominal input variable measured with no error, a miscategorized output variable by the error-prone classifier, and an output variable measured with no error by the error-free classifier. In both scenarios, we designate the sample categorized only by the error-prone classifier for the binomial output variable as the original study and the sample categorized by both the error-prone and error-free classifiers for the binomial output variable as the validation substudy. Then, we combine the original study and the validation substudy to be the overall data, where the inferential statistics are based on.
There are a lot of bibliography for g equals 1 or 2 (but not for g ≥ 3). When g = 1 and only over-reported or under-reported data exists, Moors et al. in 2000 [5] proposed a one-sided exact confidence interval and Boese et al. in 2006 [6] devised several likelihood-based confidence intervals (CIs) for the proportion parameter. Moreover, Raats and Moors in 2003 [7] and Lee and Byun in 2008 [8] considered Bayesian credible intervals for the proportion parameter. When g = 1 and both over-reported and under-reported data exist, Raats and Moors in 2003 [7] obtained both an exact confidence interval and a Bayesian credible set for the proportion parameter while Lee in 2013 [9] offered a Bayesian approach. When g = 2 with only over-reported or under-reported data, Rahardja et al. derived both Bayesian in 2019 [10] and Likelihood-Based inference in 2019 [11] for the difference of two proportion parameters. For g = 2 with both over-reported and under-reported data, Prescott and Garthwaite in 2002 [12] offered Bayesian credible interval; while Morrissey and Spiegelman in 1999 [13] presented likelihood-based confidence intervals for the odds ratio.
Usually, we encounter both over-reported and under-reported data at the same time. However, occasionally data also exist for only over-reported or only under-reported. From the existing bibliography, the statistical methods derivation and the corresponding computing-algorithm developments are different and cannot be used automatically for either type of misclassifications. Hence, it is necessary to specifically derive different method for each type of misclassification. To date, the bibliography only presented methods for g = 1 with only one error type (over-/under-reported only); g = 1 for both error types; g = 2 with only one error type (over-/under-reported only); g = 2 for both error types; moreover, all their algorithms are non-closed form (except, in all the aforementioned Rahardja et al. [10,11] papers, where algorithms were closed-form) and hence, their algorithms encountered a lot of computational difficulties, hard to reproduce, and computationally very burdensome.
Therefore, in light of novelty, in this article, we propose two likelihood-based (two Wald-type Omnibus Testing: naïve and modified) methods to analyze data obtained via a doubly sampled framework for g ≥ 3. Without loss of generality (WLOG), here we limit the scope of our omnibus testing methods to only one error type (under-reported or over-reported only) data. Again, the objective is to test the relationship between the nominal input variable (with multiple levels or groups) and the binomial output variable. Equivalently, this objective translates into H0: p1 = p2 = … = pg, or testing the equality of the multiple binomial proportions among g groups (i.e., the equality of the levels of this nominal input variable). If the proportions are not significantly different, i.e., H0 is not rejected, then there is no relationship between the nominal input variable and the binomial output variable. Vice versa, if H0 is rejected, then there is relationship. The rest of the article is managed (described) this way. In Section 2 we lay down the data structure. In Section 3 we formulate two omnibus tests (Wald-type) statistics to test the relationship between the two variables (input and output). In Section 4, as an example, we demonstrate our two omnibus tests by employing real data. In Section 5 we assess the Type I error rates and power of the tests via simulations. Finally, in Section 6, we present the conclusion, limitations, and future research.

2. Data Structure

In this section, we consider multiple-sample binomial data with only under-reported data obtained using a doubly sampled framework. Since there has been various discussion and methods proposed in the literature, for g = 1 and g = 2, here we provide a description of the kind of data structure that has been discussed (see Table 1 for such data structure under consideration) and generalized it to g ≥ 3 scenario.
As previously mentioned in Section 1, for each of the g groups, the overall data structure is composed of two independent datasets: the original study and the validation substudy. Due to economic viability, one only utilizes the perfect classifier in the validation substudy; while using the imperfect classifier in both the original and validation studies. We exhibit the summary counts in both the original and the validation studies for the data structure in Table 1. We let the size of observations mi and ni in both the original and the validation studies, for each group i (i = 1, …, g), respectively. We then define Ni = mi + ni be the total number of units for group i.
In the original study, we define xi and yi be the numbers of positive and negative classifications in group i using the error-prone classifier. In the validation study (for i = 1, …, g, j = 0, 1, and k = 0, 1), we denote nijk as the sample size in group i classified as j and k using the error-free and the error-prone classifiers, respectively.
Here, we have more notations to define. For the jth unit in the ith group, while i = 1, …, g and j = 1, …, Ni, we designate Fij and Tij as the binomial output variables for the error-prone and error-free classifiers, respectively. Note that both Fij and Tij are observable for all units in both the original and the validation studies, while Tij is only observable units in the validation substudy. If the result is positive by the error-prone classifier, we mark Fij = 1; else, Fij = 0. Likewise, we mark Tij = 1 if the result is truly positive using the error-free classifier; else, Tij = 0. Clearly, misclassification occurs when FijTij.
Next, we mark the parameters for group i as follows. Again, in this article, WLOG we consider under-reported data only, since the same formulation generality will apply (to over-reported only), as well. We let the true proportion parameters of interest be p i = Pr ( T i j = 1 ) , the proportion parameters of the error-prone classifier be π i = Pr ( F i j = 1 ) , and the false positive rates of the error-prone classifier be φ i = Pr ( F i j = 1 | T i j = 0 ) for i = 1, 2, …, g. Note that πi is not an additional unique parameter because it is obtainable by using all other parameters. In particular, using the law of total probability, we have
π i = Pr ( T i j = 1 ) Pr ( F i j = 1 | T i j = 1 ) + Pr ( T i j = 0 ) Pr ( F i j = 1 | T i j = 0 ) = p i + q i φ i ,
where qi = 1 − pi. In relation to the summary counts arranged in Table 1, we exhibit the associated cell probabilities in Table 2.
Our goal is to test the relationship between the categorical (nominal) input variable and the (binomial) output variable. In other words, whether the levels within the (nominal) input variable matters (testing equal proportions), in predicting the binomial output variable. In particular, the statistical hypotheses of interest can be written as
H 0 : p 1 = p 2 = = p g   vs.   H 1 :   at   least   one   pair   is   not   equal .
Again, this goal translates into testing the equality of the multiple binomial proportions among g groups (i.e., if the proportions are not significantly different, or H0 is not rejected, then there is no relationship between the categorical (nominal) input variable and the binomial output; and vice versa, if H0 is rejected, then there is association).

3. Methods

In this section we aim to derive two (omnibus) statistical tests for testing the hypotheses in Expression (2). We begin with estimating the Maximum Likelihood Estimator (MLE) along with its asymptotic variance-covariance matrix for every single one of the parameters based on the full-likelihood function. Next, we acquire two chi-square tests for testing the Expression (2) where the second test is essentially an improved version of the first test.

3.1. The Maximum Likelihood Estimator (MLE)

We display the data structure under consideration in Table 1. For group i, the observed counts ( n i 00 , n i 01 , n i 11 ) from the validation substudy have a Trinomial distribution with total size ni and associated probabilities displayed in an upper right 2 × 2 submatrix in Table 2, i.e.,
( n i 00 , n i 01 , n i 11 ) | p i , φ i ~ Trinomial [ n i , ( q i ( 1 φ i ) , q i φ i , p i ) ] .
Note that Multinomial Distribution [14] is the generalization of Binomial [15] distribution. Trinomial Distribution [16] is another special case of Multinomial Distribution [14]. Moreover, the observed counts (xi, yi) in the original study have the following binomial distribution:
( x i , y i ) | p i , φ i ~ Bin [ m i , ( π i , 1 π i ) ] .
Since ( n i 00 , n i 01 , n i 11 ) and ( x i , y i ) are independent for group i and these cell counts are independent across groups, up to a constant, the full-likelihood function is
L ( η ) = i = 1 g { [ q i ( 1 φ i ) ] n i 00 [ q i φ i ] n i 01 p i n i 11 π i x i ( 1 π i ) y i } ,
where η = ( p 1 , φ 1 , , p g , φ g ) is the parameter vector.
Directly maximizing (3) with respect to η is not straightforward. Instead of using numerical methods or non-closed form algorithm (which have been the historical challenges in the bibliography), we use a reparameterization of η and derive a closed-form solution. Specifically, we newly introduce
λ i = p i π i .
Then, we introduce the new set of parameters to be a vector γ = ( λ 1 , π 1 , , λ g , π g ) . Via Expression (3), the full log likelihood function in γ is
l ( γ ) = i = 1 g [ n i 11 log λ i + n i 01 log ( 1 λ i ) + ( x i + n i 1 ) log π i + ( y i + n i 00 ) log ( 1 π i ) ] .
The corresponding score vector has the following form:
( n 111 λ 1 n 101 1 λ 1 , x 1 + n 1 1 π 1 y 1 + n 100 1 π 1 , , n g 11 λ g n g 01 1 λ g , x g + n g 1 π g y g + n g 00 1 π g ) .
By setting the above score vector to 0, we obtain the MLE for γ as a vector λ ^ i = n i 11 / n i 1 and π ^ i = ( x i + n i 1 ) / N i . By solving Expressions (1) and (4) and applying the invariance property of MLE, the MLE for the vector η is p ^ i = π ^ i λ ^ i and φ ^ i = ( 1 λ ^ i ) π ^ i / q ^ i .
Using Expression (5), the expected Fisher information matrix I ( γ ) is computed as a diagonal matrix with the following diagonal elements:
( n 1 π 1 λ 1 ( 1 λ 1 ) , N 1 π 1 ( 1 π 1 ) , , n g π g λ g ( 1 λ g ) , N g π g ( 1 π g ) ) .
We can inspect that the regularity conditions are fulfilled for this model. Hence, the MLE γ ^ = ( λ ^ 1 , π ^ 1 , , λ ^ g , π ^ g ) has an asymptotic multivariate normal distribution with mean γ and covariance matrix I 1 ( γ ) , which is a diagonal matrix with the following diagonal elements:
( λ 1 ( 1 λ 1 ) n 1 π 1 , π 1 ( 1 π 1 ) N 1 , , λ g ( 1 λ g ) n g π g , π g ( 1 π g ) N g ) .
Hence, asymptotically, we have
V ( λ ^ i ) = λ i ( 1 λ i ) / ( n i π i )   and   V ( π ^ i ) = π i ( 1 π i ) / N i .
Moreover, λ ^ 1 , π ^ 1 , , λ ^ g , π ^ g are asymptotically independent.
Since p ^ i = π ^ i λ ^ i and λ ^ i ,   π ^ i are asymptotically independent, using the Delta method, the variance of p ^ i can be written as follows:
σ i 2 = π i λ i ( 1 λ i ) n i + λ i 2 π i ( 1 π i ) N i .
A consistent estimator of σ i 2 is
σ ^ i 2 = π ^ i λ ^ i ( 1 λ ^ i ) n i + λ ^ i 2 π ^ i ( 1 π ^ i ) N i .

3.2. The Two Chi-Square Tests

Here, we present two chi-square tests: the naïve Wald test and the modified Wald test, where the second test is essentially an improved version of the first test. The hypotheses in Expression (2) are equivalent to
H 0 ( 1 ) : C p = 0 vs. H 1 ( 1 ) :   Not   H 0 ( 1 ) ,
where C = [ 1 1 0 0 1 0 0 1 ] and p = ( p 1 , , p g ) .

3.2.1. Naïve Wald (nWald) Test

Under Expressions (2) or (6), asymptotically, we have
T 1 = ( C p ^ ) T ( C   Σ ^   C T ) 1 ( C p ^ ) ~ χ g 1 2 ,
where p ^ = ( p ^ 1 , , p ^ g ) and Σ ^ = diag ( σ ^ 1 2 , , σ ^ g 2 ) . Hence, we reject the null hypothesis in Expressions (2) or (6) at Type I error level α if T 1 > χ g 1 , 1 α / 2 2 , where χ g 1 , 1 α / 2 2 is the ( 1 α / 2 ) quantile of the chi-square distribution with g 1 degrees of freedom. Note that we derive T1 by applying a naïve Wald test procedure and therefore we name it as a naïve Wald (nWald) test. Such tests are known to be liberal and tend to have Type I error rates above the nominal level.

3.2.2. Modified Wald (mWald) Test

To better refine the naïve Wald test, we utilize s i = logit ( p i ) . The MLE for s i is s ^ i = logit ( p ^ i ) . Since the range of s i is the real line, the asymptotic distribution of s ^ i is better approximated by a normal distribution than that of p ^ i . Subsequently, in what follows we will construct a test based on s ^ i . By the Delta method, we have τ i 2 V ( s ^ i ) = σ i 2 / [ p i ( 1 p i ) ] 2 . A consistent estimator for τ i 2 is τ ^ i 2 = σ ^ i 2 / [ p ^ i ( 1 p ^ i ) ] 2 .
We note that the hypotheses in Expression (2) are also equivalent to
H 0 ( 2 ) : C s = 0 vs. H 1 ( 2 ) :   Not   H 0 ( 2 ) ,
where s = ( s 1 , , s g ) . Now, because under Expressions (2) or (7), asymptotically, we have
T 2 = ( C s ^ ) T ( C   Ω ^   C T ) 1 ( C s ^ ) ~ χ g 1 2 ,
where s ^ = ( s ^ 1 , , s ^ g ) and Ω ^ = diag ( τ ^ 1 2 , , τ ^ g 2 ) .
We reject the null hypothesis in Expressions (2) or (7) at Type I error level α if T 2 > χ g 1 , 1 α / 2 2 . Such testing method is called the modified Wald (mWald) test.

4. Example

To demonstrate the two Omnibus tests (nWald and mWald) derived in Section 3 of this manuscript (as the First Part/Stage of our Two-Stage Sequential Test), here (for example-continuity purpose), we apply the same vehicle-accident data published by Hochberg in 1977 [17] and demonstrated by Rahardja in 2020 [1] as the Second Part/Stage of our Two-Stage Sequential Testing development.
To recap, the narrative of such vehicle-accident data [1,17] is as follows. There are two classifiers for categorizing accidents into two [output] categories: with injuries (=1) or without injuries (=0). Our error-prone classifier is the police classifier which is prone to error. The other error-free classifier (gold standard) is the non-police classifier which does not miscategorize (e.g., hospital records of injury examination and vehicle-insurance company examination). The original study is based on police reports on vehicle accidents in North Carolina in 1974. The validation substudy is based on accidents that occurred in North Carolina during an unspecified short period in early 1975. The vehicle-accident data are displayed in Table 3.
In such vehicle-accident example [1,17], the objective is to contrast the probability of injury among four (4) groups: Male and High vehicle damage (A), Female and High vehicle damage (B), Male and Low vehicle damage (C), and Female and Low vehicle damage (D). Again, for example and continuity with Rahardja’s manuscript in 2020 [1], we adopt the same Hochberg vehicle-accident data published in 1977 [17] as displayed in Table 3.
Recall that from Expression (2), the null hypothesis in this context is to test if the chance of injuries among the four (4) accident groups are equal. Using the two (omnibus) tests developed in Section 3, the probabilities of injury for the four groups are obtained as p1 = 0.175, p2 = 0.204, p3 = 0.028, and p4 = 0.046. The chi-square statistics and p-values for testing the equality of the four proportions are nWald (T1 = 220.67, p-value < 0.0001) and mWald (T2 = 143.98, p-value < 0.0001). Since the p-value are less than 0.05 (hence, at the 5% level of significance), we conclude that the vehicle-accident study provided enough statistical evidence in support of an association/relationship between the output (injure risks) and the (nominal) input variable (with four-level of category: A = Male and High vehicle damage, B = Female and High vehicle damage, C = Male and Low vehicle damage, and D = Female and Low vehicle damage) because those risks of injuries (p1 = 0.175, p2 = 0.204, p3 = 0.028, and p4 = 0.046) are significantly different among those four groups (A, B, C, D). In other words, the four category as levels of input variable matters in predicting the injury risks (4-category). Since there were statistically significant differences among the four-level of input variable, then it is necessary to proceed to the Second Part/Stage (of the Two-Stage Sequential Testing Procedures) to determine which pair(s) contributes the proportions difference. In addition to the numeric version of results described above (in terms of test statistics, p-value, and probabilities of injuries), it is also evidenced from the graphical version of results graphed in Figure 1—visually, the height (probabilities) of the histogram showed significant different among the four accident groups (A, B, C, D). As anticipated, both the numeric and graphical versions of results are consistent in showing that there are significant different among the four accident groups. Such results concluded the First Part/Stage testing (of the Two-Stage Sequential Testing Procedure). Next, one can proceed to the Second Part/Stage testing.
Note again that such Second Part/Stage testing (of the Two-Stage Sequential Testing Procedures) was already completed/demonstrated by Rahardja in 2020 manuscript [1]; and as a summary recap, those significant differences came from the following pairs: AC, AD, BC, and BD.

5. Simulation

In this Simulation Section, we would like to mention that there is no software package and/or library package readily available for such methodology developed above (in Section 3) because generally, the methodology was developed first, before the software package. Then, typically, after a decade or more, when the method developed has become popular then commercial/non-commercial software companies or individuals start building software package and/or library package; not the reverse. Hence, we resorted to writing our own manual coding via R software [18] version 4.1.2. Since the methodologies or algorithm developed above (in Section 3) are closed-form (unlike those non-closed form algorithms formerly developed in all the previous literature reviews), it does not require much/intensive computing resource. Consequently, we can merely utilize a regular laptop/desktop computer. In other words, parallel super computers are not necessary. Here, we use a regular computer with the following specifications: Dual Core Processor of 4GB RAM CPU and a hard drive storage of 1TB; and each case of the simulations we describe below only takes about 2 min to run.
Simulations were performed to evaluate and contrast the performance of nWald and mWald tests under varying cases using empirical Type I error rate. A two-sided nominal-level Type I error rate of α = 0.1 was used for a selection of g = 3 levels. Even though not required by the two tests, for the sake of simplifying the simulation and display of results, equal sample sizes N1 = N2 = N3 = N, substudy sample sizes n1 = n2 = n3 = n, and false-positive (over-reported) rates φ1 = φ2 = φ3 = φ were utilized in simulations. For each simulation configuration, we generated K = 10,000 datasets.

5.1. Type I Error Rate

First, the performance of these two tests was investigated in terms of Type-I error rate by differing the total sample sizes. In these simulations, we assumed:
  • False positive rate φ = 0.1;
  • Ratio of validation sample size versus the total sample size r = n/N = 0.2;
  • Total sample sizes N = 100, 200, 300, 400, 500, 600;
  • True proportion parameters of interest p = 0.1 and p = 0.5.
The upper two panels in Figure 2 display the graphs of Type I error rate of both tests versus N for p = 0.1 and p = 0.5, respectively. On the upper-left panel (p = 0.1), binomial distributions are skewed; hence are not sufficiently well-behaved (because the distribution is not symmetric but skewed). Hence, in this scenario, the two tests are not anticipated to behave/carry-out well for (relatively) small samples. Nevertheless, the mWald test is conservative (as the graph showed close-to-nominal, in this case, type I error rate of 10% horizontal line) for all sample sizes studied. Both tests have close-to-nominal Type I error rates when the sample size N is greater than 300 (i.e., as the sample size larger than 300, the graph approaches to the nominal/horizontal line, asymptotically). On the contrary, it is well-known that on the upper-right panel (p = 0.5), binomial distributions are quite symmetric around their means and therefore we anticipate both tests to behave/carry-out well in this scenario. It is not surprising that the upper right panel of Figure 2 shows that both tests have close-to-nominal Type I error rates in any case of the sample sizes (i.e., both graphs approach to the nominal level or the horizontal line, in all sample sizes on the graph). The Type I error rate of the mWald test is consistently better than that of the nWald test (because the graphs consistently showed that the mWald have smaller Type I error rates, across all sample size studied).
Second, we examine the performance of the two tests in terms of Type-I error by varying p. In these simulations, the following are assumed:
  • False positive rate φ = 0.1;
  • Ratio of validation sample size versus the total sample size r = n/N = 0.2;
  • Total sample sizes N = 200 and 500;
  • True proportion parameters of interest p = 0.1, 0.2, 0.3, 0.4, 0.5.
The lower two panels of Figure 2 graph the Type-I error rate of both tests versus p. On the lower-left panel (N = 200), the graph showed that the mWald test (dashed-line curve) reaches nominal level of the Type-I error rate (which is the horizontal line) when p > 0.2 (as shown on the x-axis), while the nWald test (solid-line curve) attains nominal level of the Type-I error rate when p > 0.4. On the lower-right panel (N = 500), the mWald test reaches nominal level of the Type-I error rate for all p, while the nWald test attains nominal level of the Type-I error rate when p > 0.2. We note that via manually coding [18] using R version 4.1.2 software with the above specified computer machine, it took 2.069883 min to produce the simulations displayed in Figure 2.
Third, we also evaluate the Type-I error rate of the two tests by varying r. In the simulations, the following are the specifications.
  • False positive rate φ = 0.1.
  • Ratio of substudy sample size versus the total sample size r = n/N = 0.1, 0.2,…, 0.5.
  • Total sample sizes N = 200 and 500.
  • True proportion parameters of interest p = 0.2.
In the upper two panels of Figure 3, we graph Type I error rate of both omnibus tests versus r for N = 200 and N = 500, subsequently. When N = 200, the upper left panel of Figure 3 demonstrates that the nWald test (solid-line curve) has inflated Type-I error rates for all r (since the solid-line curve are way above the nominal/horizontal line), especially when r is small. The nWald test has better coverage as r increases. On the contrary, the mWald test has close-to-nominal levels for all the r values studied here (since the dashed-line curve are very close to the horizonal line). Similar observations can be made for N = 500 with both tests having better Type-I error rates.
Forth and finally, we assess the Type-I error rate of the two tests by varying φ. In the simulations, below are the specifications:
  • False positive rate φ = 0.1, 0.2, …, 0.5.
  • Ratio of substudy sample size versus the total sample size r = n/N = 0.2.
  • Total sample sizes N = 200 and 500.
  • True proportion parameters of interest p = 0.2.
The lower two panels in Figure 3 graph Type-I error rate of both tests versus φ for N = 200 and 500, respectively. For both N, the lower two panels of Figure 3 demonstrate that the nWald test has inflated Type-I error rates (because the solid-line curve are inflated above the horizontal line), while the mWald test has close-to-nominal levels for all the φ values studied here (because the dashed-line curve almost overlapped the horizontal line). The values of φ do not affect the Type-I error rates for both tests and both N (because both curves are quite flat across all studied values of φ, as shown in the graph). Note that the above computer machine took 1.548624 min using R version 4.1.2 software [18] to run the simulations displayed in Figure 3.

5.2. Power

We also examine the power of the two tests. In these simulations, we choose the following.
  • False-positive rate φ = 0.1.
  • Ratio of substudy sample size versus the total sample size r = n/N = 0.2, 0.4.
  • Total sample sizes N = 100, 150, …, 400.
  • True proportion parameters of interest p = (0.1, 0.2, 0.3) and p = (0.4, 0.5, 0.6).
The upper two panels of Figure 4 graph the power of both tests versus N for p = (0.1, 0.2, 0.3); the lower two panels graph the power of both tests versus N for p = (0.4, 0.5, 0.6). Note that theoretically, binomial distribution sufficiently well behaved when p close to 0.5 (i.e., “fair-coin” behavior) and not sufficiently well behaved when p is further away from 0.5—this is why we selected such contrasted selections of p = (0.1, 0.2, 0.3) and p = (0.4, 0.5, 0.6). For both sets of p and both r values, the nWald test is uniformly more powerful than the mWald test (as shown on the graph that the solid-line curve uniformly lay above the dashed-line curve), especially when the sample size is small. For larger sample sizes, both tests have similar powers. However, note that when sample size is small, the nWald test inflates the Type-I error rate and, therefore, should be not be used; therefore, the power comparison is not meaningful in these cases. Additionally, note that with the above computer machine specifications, the processing time in R software [18] is 1.904431 min for the simulation results displayed in Figure 4.

6. Summary Discussion, Limitations and Future Research

The summary is as follows. In this manuscript, we considered testing the relationship between a nominal input variable (with multiple levels or groups) and a binomial output variable in doubly sampled framework with under-reported data. We begin with deriving closed-form formulas for the MLE of all the parameters via reparameterization. Subsequently, we constructed two omnibus tests (nWald test and mWald test) by applying the Wald test procedures.
As example to demonstrate the procedures, both tests were applied to the same vehicle-accident dataset used in the Rahardja (2020) paper [1], first introduced by Hochberg in 1977 [17]. Note that to be deemed valid, in general, most statistical methodologies developed and published historically in the bibliography are derived based on the Asymptotic (Large-Sample) Theory. This is why in the standard university-level introductory statistics (Stat 101) course, as the general rule-of-thumb and simplified case, students are taught that large sample was defined as big as 30 samples. Therefore, as anticipated, by Asymptotic (Large-Sample) Theory, the two omnibus tests yielded similar results due to their large-sample size nature of Hochberg (1977) vehicle-accident dataset [17].
Note also that in the case of small sample situations, for examples, in the Biology field, or in a Pilot Study of typical drug developments, where practitioners often encounter small sample, then the results should be noted with precautions. Meaning, if the results from the omnibus test and/or MCP test are not significant, it does not necessarily mean that it is truly not significant; it could be due to insufficient (small) sample size. Such precaution should be included/noted in the labeling or study report. Then, more information needs to be consulted with the subject matter experts (SME) in the field of study, for the final direction of decision making for such small-sample (biology or pilot) study. For example, whether making the decision of “Go/No-Go”; in other words, whether proceeding/not, to next phase (large sample) study, by spending more funds to recruit more patients/sample (with such noted risk or precaution); or terminate the study since it is not promising, etc.
Simulation studies were performed to assess the empirical Type I error rate of both tests under various simulation configurations. Since the two tests were developed based on large-sample theories, both methods were anticipated to implement well under large samples. Such anticipation was affirmed since the resulting Type I error rates were close to the nominal level for large samples. When sample size is relatively small and p is close to 0 or 1, the nWald test resulted in a lot more inflated Type I error rate while the mWald test preserved the Type I error rate. Hence, the mWald test is recommended over nWald test.
The limitations are as follows. When sample size is afforded at relatively small size, p is close to 0 or 1, even in larger samples, both methods still suffer from inflation of type I error rate due to: not only asymptotic approximation, but also both tests can be conservative, i.e., slow to reject the null hypothesis (Figure 2). In the standard/built-in statistical software packages such as R, SAS, etc., there is the use of continuity adjustment in chi-square test which is most useful for small sample sizes, to prevent overestimation of statistical significance for small data. The formula is chiefly used when at least one cell of the table has an expected count smaller than five (5). Unfortunately, such correction may tend to overcorrect. This use of the continuity adjustment is somewhat controversial and can result in an overly conservative result that cause the chi-square test fails to reject the null hypothesis when it should (i.e., too conservative), for small sample size.
Finally, in this manuscript, as the first part of a two-stage sequential test (with over-reported data), we have contributed the development of two omnibus tests (first-stage), to test the statistical hypothesis in expression (2), with an easy-to-implement, closed-form algorithm. Indeed, due to our closed-form algorithm via reparameterization, we were able to derive such omnibus tests for g ≥ 3 scenario with the under-report (or WLOG, over-report), which was the reason why previously other researchers were not able to derive such scenario. If such test in expression (2) is not rejected then we declare there is no proportion difference among the g groups. Otherwise, if the test is rejected, then we may want to proceed to the second stage (of the two-stage sequential test) to find-out which pair(s) contributes the difference. In such case, Rahardja pairwise comparison method developed in 2020 [1], which also known as Pairwise Multiple Comparison Procedures (MCP) with over-reported adjustment incorporated via doubly sampled framework, can further be applied. Note also that without incorporating any misclassification (under-reported or over-reported) into the statistical model, such MCP tests are just the standard/classical procedures in the literature review, as prescribed by Hsu in 1996 [19], Edwards and Hsu in 1983 [20], Hsu and Edwards in 1983 [21], and Hsu in 1982 [22]. Moreover, for the future research, one can add multiple covariates (with multiple levels, too) as the input variables, into considerations—in the presence of under-reported or over-reported (output) data. Examples of such multiple-level (input) covariates may be: age category (below 25, 25–40, 40–60, above 60), height category, body-mass-index category, drinking category (never, sometimes, often), etc.

7. Conclusions

In conclusion, we have completed a Two-Stage Sequential Testing development (for multiple binomial proportions using doubly sampled framework with only under-reported or over-reported data). In this manuscript, we derived two Omnibus Tests with the under-reported or over-reported only. Such Omnibus Tests function as the First-Stage of the Two-Stage Sequential Test. Additionally, the Pairwise-MCP Test was already developed by Rahardja in 2020 [1] and function as the Second-Stage of the Two-Stage Sequential Test (with the over-reported or under-reported).

Funding

This research received no external funding.

Acknowledgments

The author thanks both the anonymous referees and the Associate Editor for their insightful and constructive suggestions which have improved the final presentation of this manuscript.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Rahardja, D. Multiple Comparison Procedures for the Differences of Proportion Parameters in Over-Reported Multiple-Sample Binomial Data. Stats 2020, 3, 56–67. [Google Scholar] [CrossRef] [Green Version]
  2. Bross, I. Misclassification in 2 × 2 tables. Biometrics 1954, 10, 478–486. [Google Scholar] [CrossRef]
  3. Goldberg, J.D. The effects of misclassification on the bias in the difference between two proportions and the relative odds in the fourfold table. J. Am. Stat. Assoc. 1975, 70, 561–567. [Google Scholar]
  4. Tenenbein, A. A double sampling scheme for estimating from binomial data with misclassifications. J. Am. Stat. Assoc. 1970, 65, 1350–1361. [Google Scholar] [CrossRef]
  5. Moors, J.J.A.; van der Genugten, B.B.; Strijbosch, L.W.G. Repeated audit controls. Stat. Neerl. 2000, 54, 3–13. [Google Scholar] [CrossRef]
  6. Boese, D.H.; Young, D.M.; Stamey, J.D. Confidence intervals for a binomial parameter based on binary data subject to false-positive misclassification. Comput. Stat. Data Anal. 2006, 50, 3369–3385. [Google Scholar] [CrossRef]
  7. Raats, V.M.; Moors, J.J.A. Double-checking auditors: A Bayesian approach. Stats 2003, 52, 351–365. [Google Scholar] [CrossRef]
  8. Lee, S.C.; Byun, J.S. A Bayesian approach to obtain confidence intervals for binomial proportion in a double sampling scheme subject to false-positive misclassification. J. Korean Stat. Soc. 2008, 37, 393–403. [Google Scholar] [CrossRef]
  9. Lee, S.C. Bayesian confidence intervals of proportion with misclassified binary data. J. Korean Stat. Soc. 2013, 42, 291–299. [Google Scholar] [CrossRef]
  10. Rahardja, D. Bayesian Inference for the Difference of Two Proportion Parameters in Over-Reported Two-Sample Binomial Data Using the Doubly Sample. Stats 2019, 2, 111–120. [Google Scholar] [CrossRef] [Green Version]
  11. Rahardja, D.; Wu, H.; Zhang, Z.; Tiedt, A.D. Maximum likelihood estimation for the proportion difference of two-sample binomial data subject to one type of misclassification. J. Stat. Manag. Syst. 2019, 22, 1365–1379. [Google Scholar] [CrossRef]
  12. Prescott, G.J.; Garthwaite, P.H. A simple bayesian analysis of misclassified binary data with a validation substudy. Biometrics 2002, 58, 454–458. [Google Scholar] [CrossRef] [PubMed]
  13. Morrissey, M.J.; Spiegelman, D. Matrix methods for estimating odds ratios with misclassified exposure data: Extensions and comparisons. Biometrics 1999, 55, 338–344. [Google Scholar] [CrossRef] [PubMed]
  14. Wikipedia. Multinomial Distribution. Available online: https://en.wikipedia.org/wiki/Multinomial_distribution (accessed on 14 March 2022).
  15. Wikipedia. Binomial Distribution. Available online: https://en.wikipedia.org/wiki/Binomial_distribution (accessed on 14 March 2022).
  16. Wikipedia. Trinomial Distribution. Available online: https://online.stat.psu.edu/stat414/lesson/17/17.3 (accessed on 14 March 2022).
  17. Hochberg, Y. On the use of double sampling schemes in analyzing categorical data with misclassification errors. J. Am. Stat. Assoc. 1977, 72, 914–921. [Google Scholar]
  18. Lim, A.; Tjhi, W. R High Performance Programming; Packt Publishing Ltd.: Birmingham, UK, 2015; ISBN 978-1-78398-926-3. [Google Scholar]
  19. Hsu, J.C. Multiple Comparisons: Theory and Methods; Chapman & Hall/CRC: Boca Raton, FL, USA; London, UK; New York, NY, USA; Washington, DC, USA, 1996; ISBN1 0-41298-281-1, ISBN2 978-0412982811, ISBN3 0412982811; Available online: https://www.asc.ohio-state.edu/hsu.1//mc.html (accessed on 3 May 2019).
  20. Edward, D.G.; Hsu, J.C. Multiple comparisons with the best treatment. J. Am. Stat. Assoc. 1983, 78, 965–971. [Google Scholar] [CrossRef]
  21. Hsu, J.C.; Edwards, D.G. Sequential multiple comparisons with the best. J. Am. Stat. Assoc. 1983, 78, 958–964. [Google Scholar] [CrossRef]
  22. Hsu, J.C. Simultaneous inference with respect to the best treatment in block designs. J. Am. Stat. Assoc. 1982, 77, 461–467. [Google Scholar] [CrossRef]
Figure 1. Probability (risk or chance of injury) versus the 4 Accident Groups (A, B, C, D).
Figure 1. Probability (risk or chance of injury) versus the 4 Accident Groups (A, B, C, D).
Stats 05 00024 g001
Figure 2. Type I error rate versus total sample sizes N for the upper two panels and versus the true proportion parameter p for the lower two panels.
Figure 2. Type I error rate versus total sample sizes N for the upper two panels and versus the true proportion parameter p for the lower two panels.
Stats 05 00024 g002
Figure 3. Type-I error rate versus r for the upper two panels and versus φ for the lower two panels.
Figure 3. Type-I error rate versus r for the upper two panels and versus φ for the lower two panels.
Stats 05 00024 g003
Figure 4. Power versus N with p = (0.1, 0.2, 0.3) for the upper two panels and with p = (0.4, 0.5, 0.6) for the lower two panels.
Figure 4. Power versus N with p = (0.1, 0.2, 0.3) for the upper two panels and with p = (0.4, 0.5, 0.6) for the lower two panels.
Stats 05 00024 g004
Table 1. Data for group i.
Table 1. Data for group i.
StudyError-Free ClassifierError-Prone Classifier
01Total
Validation0ni00ni01ni0∙
1NAni11ni11
Totalni00ni∙1ni
OriginalNAyiximi
NA: Not Available.
Table 2. Cell probabilities for group i.
Table 2. Cell probabilities for group i.
StudyError-Free ClassifierError-Prone Classifier
01Total
Validation0qi (1 − φi)qi φiqi
1NApipi
OriginalNA1 − πiπi1
NA: Not Available.
Table 3. Vehicle accident data.
Table 3. Vehicle accident data.
GroupStudyError-Free
Classifier
Error-Prone
Classifier
01
A (Male and High Vehicle Damage)Validation0369NA
175132
OriginalNA19,6317329
B (Female and High Vehicle Damage)Validation0123NA
16187
OriginalNA76924004
C (Male and Low Vehicle Damage)Validation0529NA
15939
OriginalNA25,5421886
D (Female and Low Vehicle Damage)Validation0249NA
14330
OriginalNA12,4611539
NA: Not Available.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rahardja, D. Omnibus Tests for Multiple Binomial Proportions via Doubly Sampled Framework with Under-Reported Data. Stats 2022, 5, 408-421. https://doi.org/10.3390/stats5020024

AMA Style

Rahardja D. Omnibus Tests for Multiple Binomial Proportions via Doubly Sampled Framework with Under-Reported Data. Stats. 2022; 5(2):408-421. https://doi.org/10.3390/stats5020024

Chicago/Turabian Style

Rahardja, Dewi. 2022. "Omnibus Tests for Multiple Binomial Proportions via Doubly Sampled Framework with Under-Reported Data" Stats 5, no. 2: 408-421. https://doi.org/10.3390/stats5020024

Article Metrics

Back to TopTop