Next Article in Journal
A Note on New Near-Extremal Type I Z4-Codes of Length 48
Previous Article in Journal
Weakly Ricci-Symmetric Space-Times and f (R,G) Gravity
Previous Article in Special Issue
Bayesian Estimation of the Stress–Strength Parameter for Bivariate Normal Distribution Under an Updated Type-II Hybrid Censoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Inference with Pólya-Gamma Augmentation for US Election Law

US Census Bureau, 4600 Silver Hill Rd., Washington, DC 20233, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2025, 13(6), 945; https://doi.org/10.3390/math13060945
Submission received: 7 February 2025 / Revised: 6 March 2025 / Accepted: 11 March 2025 / Published: 13 March 2025
(This article belongs to the Special Issue Statistical Simulation and Computation: 3rd Edition)

Abstract

:
Pólya-gamma (PG) augmentation has proven to be highly effective for Bayesian MCMC simulation, particularly for models with binomial likelihoods. This data augmentation strategy offers two key advantages. First, the method circumvents the need for analytic approximations or Metropolis–Hastings algorithms, which leads to simpler and more computationally efficient posterior inference. Second, the approach can be successfully applied to several types of models, including nonlinear mixed-effects models for count data. The effectiveness of PG augmentation has led to its widespread adoption and implementation in statistical software packages, such as version 2.1 of the R package BayesLogit. This success has inspired us to apply this method to the implementation of Section 203 of the Voting Rights Act (VRA), a US law that requires certain jurisdictions to provide non-English voting materials for specific language minority groups (LMGs). In this paper, we show how PG augmentation can be used to fit a Bayesian model that estimates the prevalence of each LMG in each US voting jurisdiction, and that uses a variable selection technique called stochastic search variable selection. We demonstrate that this new model outperforms the previous model used for 2021 VRA data with respect to model diagnostic measures.

1. Introduction

1.1. A Motivating Example

Section 203(b) of the Voting Rights Act (VRA) is an election law that requires some US states or their political subdivisions to make voting materials available in languages other than English. The US Census Bureau determines whether a jurisdiction, e.g., county, is subject to such requirements based on the rates of limited English language proficiency and illiteracy within designated population subgroups, called language minority groups (LMGs), in that jurisdiction. For the purposes of VRA coverage determinations, limited English proficiency (LEP) is defined as speaking a language other than English at home, and speaking English less than “Very Well”. Illiteracy (ILL) is defined as having less than a 5th-grade education. A more detailed discussion of the legal and statistical criteria used to make VRA coverage determinations is available in Slud et al. [1].
To make these determinations, the Census Bureau must estimate the number of voting-age (18 or over) people (VOT), voting-age citizens (CIT), voting-age citizens who have limited English proficiency (LEP), and voting-age citizens with limited English proficiency who are illiterate (ILL) in each of these geographies. The Census Bureau initially used “direct” estimates based on data from the long-form of the Decennial Census (later replaced by the American Community Survey, or ACS) to approximate these quantities. The term “direct” indicates that an estimate does not use any statistical model or auxiliary information beyond the survey weights and observations within that area. However, since 2011, Section 203 determinations have been based on estimates from small-area models. These statistical models were updated in 2014 by Joyce et al. [2], who developed a Bayesian hierarchical model, and Slud et al. [3], who adopted a Dirichlet-multinomial model.
In the VRA 2021 determinations, model-based estimates were obtained using both frequentist and Bayesian methods [1]. The frequentist method used adaptive Gaussian quadrature [4] to fit random effects models for LMGs with a sample in less than 700 jurisdictions (counties and MCDs), or less than 3000 jurisdictions where most citizens did not have severe illiteracy. The Bayesian model was fitted using STAN [5], which only worked well for large LMGs. Models fit with STAN had relatively slow and unstable convergence in smaller LMGs even when using priors recommended for computational efficiency, such as the LKJ prior for the random-effects covariance distribution. In addition, both methods used covariates that were selected based on their performance in a few large LMGs to model all LMGs, rather than performing independent variable selection procedures in the models for smaller groups. More details on the VRA 2021 methods are described in Slud et al. [1].
Table 1 shows a dataset that contains hypothetical examples of jurisdiction-level estimates that VRA models might produce for an LMG. For this LMG, let V O T j , C I T j , L E P j , and I L L j represent VOT, CIT, LEP, and ILL, respectively, in jurisdiction j. Then, in Table 1, CITprop, LEPprop, and ILLprop represent the direct estimates of the ratios C I T j / V O T j , L E P j / V O T j , and I L L j / V O T j , respectively. The model-based predictions of these ratios are represented by π C I T , π L E P , and π I L L , respectively.
Note that the model-based estimates are similar to the direct estimates for jurisdiction 1, which has a large number of voting-age residents. However, these estimates differ in jurisdictions with fewer voting-age persons. This pattern occurs because direct estimates perform well in large samples, but poorly when sample sizes are small. Model-based estimates use vectors of auxiliary covariates, denoted in Table 1 by X, to help improve their low sample performance.
The new model introduced in this article uses Pólya-gamma (PG) augmentation [6] to estimate Bayesian models for all LMGs and jurisdictions. This enables the new model to more easily implement stochastic search variable selection (SSVS) within each LMG to select covariates, which allows our new model to outperform the older models in terms of diagnostic metrics.

1.2. Utility of Pólya-Gamma Augmentation in Bayesian Methods

Pólya-gamma (PG) augmentation, as introduced by Polson et al. [6], is an efficient method for Bayesian inference in models with binomial likelihoods. The application of PG augmentation yields closed-form expressions for full conditionals of posterior distributions in sampling algorithms like Markov chain Monte Carlo (MCMC), which eliminates the need for analytic approximations, numerical integration, or Metropolis–Hastings algorithms. This allows the construction of simple and effective Gibbs samplers for models with binomial likelihoods, particularly logistic regression. The method has been successfully applied to various models, including logistic regression and nonlinear mixed-effects models [6]. In practice, the PG Gibbs sampler for Bayesian logistic regression involves two main steps in each iteration: sampling PG random variables and drawing from a multivariate normal distribution. This approach has been implemented in various statistical software packages, such as Version 2.1 of the R package BayesLogit [7].
The remainder of this paper is organized into the following sections. Section 2 reviews the 2021 VRA model, introduces PG augmentation, and explains a basic application in the context of the VRA data. Section 3 builds on this foundation to introduce the new Bayesian multinomial logit model, and describes a Gibbs sampling algorithm that makes use of PG augmentation to fit it. Section 4 applies the new model to the data used for the 2021 VRA determinations and shows its superior diagnostic results compared with the 2021 model results. Finally, Section 5 discusses the results.

2. The Pólya-Gamma Distribution for Binomial VRA Datasets

2.1. The 2021 VRA Model

The 2021 VRA model was constructed by modeling the counts mentioned in Section 1 using a multinomial distribution. Specifically, for each jurisdiction j where j { 1 , , 7859 } , and n j is the number of voting-age people in j, we model
M j , 1 , M j , 2 , M j , 3 , M j , 4 M u l t i n o m n j , p j , 1 , p j , 2 , p j , 3 , p j , 4 ,
where M j , 1 , M j , 2 , M j , 3 , and M j , 4 are V O T j - C I T j , C I T j - L E P j , L E P j - I L L j , and I L L j , respectively.
An equivalent formulation of this model treats the M variables in Equation (1) as the composition of three conditionally binomial ( B i n o m ) models:
M j , 1 B i n o m n j , p j , 1 M j , 2 B i n o m n j M j , 1 , p j , 2 1 p j , 1 given M j , 1 M j , 3 B i n o m n j M j , 1 M j , 2 , p j , 3 1 p j , 1 p j , 2 given M j , 1 , M j , 2 ,
The intuition behind the success probabilities in these distributions is as follows. In the multinomial model, p j , 1 represents the marginal probability that a voting-age person is not a citizen. Thus, if we ignore the other categories, M j , 1 can be modeled as a binomial distribution counting the number of voting-age non-citizens against the number of voting-age citizens. Similarly, the marginal probability that a person is a voting-age citizen and is not LEP is p j , 2 in the multinomial model. It can be inferred that 1 p 1 is the marginal probability that a voting-age person is a citizen. Thus, the probability of a voting-age person not having limited English proficiency conditional on being a citizen is p j , 2 / ( 1 p j , 1 ) . By the same logic, the marginal probability that a voting-age person is a citizen and LEP but not ILL is p 3 in the multinomial model, and the marginal probability that a voting-age person is a citizen and is LEP is 1 p j , 1 p j , 2 . Thus, the conditional probability of a voting-age person not being ILL given that they are a citizen and have LEP is p j , 3 / ( 1 p j , 1 p j , 2 ) . M j , 4 is known given M j , 1 , M j , 2 , M j , 3 , and n j , and thus does not need to be modeled.
The complements of the “success” probabilities of the binomial distributions for M j , 1 , M j , 2 , and M j , 3 can be interpreted as estimates of the ratios v j , 1 = C I T j / V O T j , v j , 2 = L E P j / C I T j , and v j , 3 = I L L j / L E P j in jurisdiction j. With this notation, the probabilities in Equation (1) can be written for each jurisdiction j as
p j , 1 , p j , 2 , p j , 3 , p j , 4 1 v j , 1 , v j , 1 1 v j , 2 , v j , 1 v j , 2 1 v j , 3 , v j , 1 v j , 2 v j , 3
This means that we can model the multinomial success probabilities in (1) by modeling the binomial success probabilities in (2), and we can model the binomial success probabilities in (2) by modeling the v j , k ratios. To model the v j , k ratios, we assume that
v j , k = exp η j , k 1 + exp η j , k ,
where
η j , k X j T β k + u j , k
and the u are multivariate normal random variables with a zero mean and variance parameter
Σ = σ 1 2 σ 12 σ 13 σ 12 σ 2 2 σ 23 σ 13 σ 23 σ 3 2 .
This specification is called the “multinomial logit normal” (MLN) model due to the multinomial outcome described in Equation (1), the logistic link function in Equation (3), and the multivariate normal random effects. When these parameters are estimated, we can use them to derive model-based estimates of π C I T , π L E P , and π I L L similar to the hypothetical examples in Table 1. Specifically, for jurisdiction j, we estimate π C I T as v j , 1 , π L E P as v j , 1 × v j , 2 , and π I L L as v j , 1 × v j , 2 × v j , 3 . The difficulty of estimating these parameters stems from the nonlinearity of the logistic link function in Equation (3). The Pólya-gamma augmentation makes the estimation process linear, resulting in closed-form solutions for the posterior distributions of our parameters.

2.2. The Pólya-Gamma (PG) Distribution

A PG random variable is defined as an infinite weighted sum of gamma random variables. Specifically, if ω P G ( b , c ) , where b > 0 and c R , then:
ω = d 1 2 π 2 k = 1 g k ( k 0.5 ) 2 + c 2 4 π 2
where = d denotes equality in distribution, and g k G a m m a ( b , 1 ) are independent gamma random variables. Polson et al. [6] proved the following useful property of the PG ω variable. Let ψ denote a constant, e.g., a linear combination of covariates X and their coefficients β . Then, the following equation holds for an exponential functional form:
( e ψ ) a ( 1 + e ψ ) b = 2 b e κ ψ 0 e ω ψ 2 / 2 P ( ω ) d ω
where κ = a b / 2 and p ω = P G ω | b , 0 . The utility of the above equation is that for j = 1 , , N , we can express the contribution of jurisdiction j to the likelihood as
L j ( β ) = ( e ψ j ) y j ( 1 + e ψ j ) n j = 2 n j e κ j ψ j 0 e ω j ψ j 2 / 2 p ( ω j ) d ω j ,
where κ j = y j n j / 2 . Note that if we condition on Ω = d i a g ( ω 1 , , ω N ) , the kernel of the posterior distribution for β looks similar to a Gaussian density function:
P β | Ω , y , n = P ( β ) j = 1 N L j ( β | ω j )
P ( β ) j = 1 N e κ j ψ j e ω j ψ j 2 / 2
P ( β ) j = 1 N exp ω j 2 ψ j κ j ω j 2
= P ( β ) exp 1 2 Z X β T Ω Z X β
where Z = κ 1 / ω 1 , , κ N / ω N . Step (8) is justified because conditional on ω j , the expectation in (6) is constant. Step (9) is derived by completing the square. From this, Step (10) can be derived using algebra.
In summary, if our prior on β is Gaussian (quadratic in β ), then (7) has a closed-form solution. This enables us to construct a Gibbs sampler for β , by repeatedly sampling Ω | β and then sampling β | Ω .

2.3. A Simplified Application of Pólya-Gamma Augmentation

To illustrate the application of PG augmentation to the 2021 VRA model, consider jurisdiction j, for j { 1 , , N } , where N denotes the total number of jurisdictions. Let n j denote the number of voting-age citizens, v j denote the probability that a voting-age citizen is LEP, and y j denote the number of voting-age citizens with LEP. For simplicity, we do not include random effects in the example for this section. Then, PG augmentation as introduced by Polson et al. [6] can be applied to write the probability of observing y j out of n j as
1 v j n j y j v j y j = exp ψ j y j 1 + exp ψ j n j N κ j / ω j | X j β , ω j 1 ,
where κ j = y j n j / 2 , ω j P G ( n j , ψ j ) , and ψ j = X j β . Following Polson et al. [6], we let Z = ( κ 1 / ω 1 , , κ N / ω N ) , which allows us to write the β posterior distribution as
P β | y , ω P β j = 1 N L j β | ω j P β exp 1 2 Z X β T Ω Z X β = N β | b , B N Z | X β , Ω 1 = N β | m ω , V ω ,
where X = X 1 T , , X N T T , κ = κ 1 , , κ N T , Ω = d i a g ω 1 , , ω N , V ω = X T Ω X + B 1 1 , and m ω = V ω X T κ + B 1 b and a standard normal distribution N b , B for the β prior. Thus, a simple Gibbs sampling algorithm to fit this model iterates:
ω j | β P G n j , X j T β β | y j , n j , ω j N m ω , V ω .
This Gibbs sampling algorithm was implemented using the R package BayesLogit [7] and provides closed-form equations for the β posterior. This allows users to adopt methods developed for linear regressions, including the Bayesian model selection procedure [8]. Appendix A explains the derivation of the closed-form β posterior distribution.

3. New MLN Model with PG Augmentation

3.1. Basic Setup

In this section, we introduce the model specification for our new Bayesian VRA model. This new model is distinguished by our adoption of stochastic search variable selection [8], which is the most important difference between it and previous VRA models. However, the new model retains the multinomial logit normal structure from the 2021 VRA model [1] described in Section 2.1. In particular, we can write this model as a composition of three conditional binomial distributions with success probabilities v 1 , v 2 , and v 3 . Table 2 summarizes the relationships between the multinomial category proportions p 1 , p 2 , p 3 , and p 4 and the modeled ratios v 1 , v 2 , and v 3 in this model. It also shows the contribution that each jurisdiction, American Indian Area (AIA), and/or Alaskan Native Regional Corporation (ANRC) makes to the log-likelihood when fitting this statistical model.
Note that the log-likelihood contribution for v k compares M k with c = k + 1 4 M c for k = 1, 2, 3.

3.2. Model Likelihood and Priors

Regression coefficients β , the random effects u, and the covariance matrix Σ are specified with the following likelihood contribution for the jth jurisdiction:
L H j = 1 v j , 1 V O T j C I T j v j , 1 C I T j 1 v j , 2 C I T j L E P j v j , 2 L E P j 1 v j , 3 L E P j I L L j v j , 3 I L L j = 1 v j , 1 M j , 1 v j , 1 c = 2 4 M j , c 1 v j , 2 M j , 2 v j , 2 c = 3 4 M j , c 1 v j , 3 M j , 3 v j , 3 M j , 4 = c = 1 3 exp η j , c k = c + 1 4 M j , k 1 + exp η j , c k = c 4 M j , k ,
Using this notation, the likelihood for all N jurisdictions multiplied by the prior distributions for the parameters can be expressed as
j = 1 N L H j × k = 1 3 P β k P Σ P U j , k = u j , k .
The Bayesian model in Equation (12) treats the random effects variables u as parameters to be estimated. In contrast, frequentist inference averages out random effects as “missing data” in the optimization process. However, the widely used frequentist adaptive Gaussian quadrature method [4] for fitting random effects models treats the random effects as parameters in the process of averaging them out. Thus, both Bayesian and frequentist inferences estimate random effects parameters from this technical perspective.

3.3. Conditional Posterior Distributions for Regression Coefficients and PG Variables

For each outcome k { 1 , 2 , 3 } and jurisdiction j { 1 , , 7859 } , the regression coefficients β k and PG random variables ω j , k are sampled similarly to what was done in Polson et al. [6]. Recall the key result from Section 2 that, conditionally given ω , exp ψ y / 1 + exp ψ n N ω 1 y n / 2 | ψ , ω 1 , which indicates a Gaussian normal distribution for ω 1 y n / 2 with its mean ψ and variance ω 1 . Using this result it can be shown that, conditional on ω j , k , the posterior distributions for the β k coefficients have the following closed-form representations:
  • V O T j C I T j vs. C I T j ( k = 1 4 M j , k - k = 2 4 M j , k vs. k = 2 4 M j , k )
    β 1 N V β 1 1 m β 1 , V β 1 1
    conditional on ω j , 1 being drawn from ω j , 1 P G k = 1 4 M j , k , η j , 1 = d P G V O T j , η j , 1 for each j.
  • C I T j L E P j vs. L E P j ( k = 2 4 M j , k k = 3 4 M j , k vs. k = 3 4 M j , k )
    β 2 N V β 2 1 m β 2 , V β 2 1
    conditional on ω j , 2 being drawn from ω j , 2 P G k = 2 4 M j , k , η j , 2 = d P G C I T j , η j , 2 for each j.
  • L E P j I L L j vs. I L L j ( k = 3 4 M j , k k = 4 4 M j , k vs. k = 4 4 M j , k )
    β 3 N V β 3 1 m β 3 , V β 3 1
    conditional on ω j , 3 being drawn from ω j , 3 P G k = 3 4 M j , k , η j , 3 = d P G L E P j , η j , 3 for each j.
Here, for each outcome category k { 1 , 2 , 3 } , we define: η j , k = X j T β k + u j , k , X = X 1 T , , X N T T , κ j , k = c = k + 1 4 M j , c 1 2 c = k 4 M j , c , κ k = κ 1 , k , , κ N , k T , Ω k = d i a g ( ω 1 , k , , ω N , k ) , U k = u 1 , k T , , u N , k T T , V β k = X T Ω k X + B 1 , m β k = X T κ k X T Ω k U k + B 1 b , and b and B are from the normal β k prior N ( b , B ) . In general, the conditional β k posterior for outcome category k is P ( β k | X , κ k , U k , Ω k ) N V β k 1 m β k , V β k 1 . This posterior distribution is derived in Appendix B.

3.4. Conditional Posterior Distribution for the Variable Selection Parameters

George and McCulloch [8] developed a Bayesian variable selection procedure known as stochastic search variable selection (SSVS) in the context of linear regression. Recall that the regression in Equation (11) assumes the β prior distribution N b , B . To perform SSVS, we specify the prior distribution for β as a mixture of two normal distributions: one with a smaller variance to signify a coefficient close to zero, and the other with a larger variance to signify a coefficient farther away from zero. Specifically, George and McCulloch [8] introduced the latent inclusion variable γ i (1 if included, 0 otherwise) for the normal mixture prior distribution for the ith coefficient β i ( i { 1 , , q } ):
1 γ i N 0 , τ i 2 + γ i N 0 , c i 2 τ i 2 .
The interpretation of this equation is as follows. When γ i is 0, β i is assumed to have N(0, τ i 2 ), and when γ i is 1, β i is assumed to have N(0, c i 2 τ i 2 ). τ i is set to be small so that if γ i is 1, then β i will be close to 0. c i is set to be large ( c i > 1 always) so that if γ i is 1, then a non-zero estimate of β i will be included in the model. For specific choices of γ i and c i , George and McCulloch [8] recommended that users should consider varying the settings to extract more information rather than treat any particular settings as rules that guarantee good results.
To implement the normal mixture prior in Equation (16), the prior covariance matrix B from the prior distribution N b , B is specified as D γ R D γ , where γ = ( γ 1 , , γ q ) , R is the prior correlation matrix (an identity matrix), and
D γ d i a g ( a 1 τ 1 , , a q τ q ) ,
with a i = 1 if γ i = 0 and a i = c i if γ i = 1. That is, a i τ i = { ( 1 γ i ) + γ i c i } × τ i and a i 2 τ i 2 = { ( 1 γ i ) + γ i c i 2 } × τ i 2 . Thus, D γ determines the scaling of the prior covariance matrix in such a way that Equation (16) is satisfied.
The conditional posterior distribution of γ i ( m ) for the ith covariate at the mth MCMC sampling step is
P γ i ( m ) = 1 | β ( m 1 ) , γ ( i ) ( m ) ,
where γ ( i ) ( m ) = ( γ 1 ( m ) ,…, γ i 1 ( m ) , γ i + 1 ( m 1 ) ,…, γ p ( m 1 ) ), and β ( m 1 ) is the coefficients vector for the ( m 1 ) th MCMC sample with m { 1 , , 10 5 } . Notice that the posterior for γ i ( m ) does not depend on the rest of the parameters, including the outcomes. This yields a substantial simplification that reduces computational requirements and allows for faster convergence of the MCMC subsequence for all γ i . Finally,
P γ i ( m ) = 1 | β ( m 1 ) , γ ( i ) ( m ) = a a + b ,
where
a = P β ( m ) | γ ( i ) ( m ) , γ i ( m ) = 1 × p i
and
b = P β ( m ) | γ ( i ) ( m ) , γ i ( m ) = 0 × ( 1 p i ) ,
where p i was chosen to be 0.5, implying equal weights for a and b.

3.5. Conditional Posterior Distribution for Random Effects Covariance Σ

Recall U = ( u 1 , u 2 , , u N ) and u j = ( u j , 1 , u j , 2 u j , 3 ). Σ was defined in Equation (4). Assuming U has a multivariate normal distribution M V N 0 , Σ , the conjugate Σ posterior is
I W j = 1 N u j u j T + V 0 , N + s 0 ,
where V 0 can be set to be I 3 × 3 , and s 0 can be set to 3, which is the number of rows of Σ . These are typical choices for the inverse Wishart priors. s 0 can be chosen to be bigger than 3, but larger values of s 0 tend to attenuate the random effect covariance elements because the elements of the posterior mean j = 1 N u j u j T + V 0 / N + s 0 3 1 will get closer to zero as s 0 increases. We keep s 0 at 3, and set the off-diagonial elements to zero. Slud et al. [1] called the multinomial logit model with this random effects covariance structure “MLN-D”. When the off-diagonal elements were not set to zero, Slud et al. [1] called the model “MLN-F”. Fitting MLN-D is generally more stable than fitting MLN-F, but this does not necessarily mean that one is better than the other in terms of diagnostic results.

3.6. Conditional Posterior Distribution for Random Effects Variables

Consider a Gaussian prior distrubtion N 0 , Σ on u j = u j , 1 , u j , 2 , u j , 3 . Then, the conditional posterior P ( u j | Σ ) can be shown to be N V j , u 1 m j , u , V j , u 1 , where V j , u = Ω j + Σ 1 , m j , u = Ω j M j u , Ω j = d i a g ω j , 1 , ω j , 2 , ω j , 3 , M j , k u = M j , k - X j β k , M j , k = ω j , k 1 c = k + 1 4 M j , c 1 2 k = c 4 M j , c , and η j , k X j T β k + u j , k .
Note that m j , u = Ω j M j u can be expressed as:
ω j , 1 0 0 0 ω j , 2 0 0 0 ω j , 3 × M j , 1 X j β 1 M j , 2 X j β 2 M j , 2 X j β k = k = 2 4 M j , k 1 2 k = 1 4 M j , k ω j , 1 × X j β 1 k = 3 4 M j , k 1 2 k = 2 4 M j , k ω j , 2 × X j β 2 k = 4 4 M j , k 1 2 k = 3 4 M j , k ω j , 3 × X j β 3 .
This is computationally helpful because M j , k is equal to ω k 1 c = k + 1 4 M j , c 1 2 k = c 4 M j , c . Calculating this expression requires taking the inverse of ω j , k , which may be close to zero. For numerical stability, we offset ω j , k 1 by multiplying it by ω j , k and set the mininum value of ω j , k to 10 16 .

3.7. Gibbs Sampling Algorithm

The following Gibbs sampling steps are for respective posterior distributions for the aforementioned parameters— β , ω , γ , Σ , and u—with each step being conditioned on all other sampled parameters from previous steps. The initial values were drawn from either a standard normal distribution or a uniform distribution.
  • ω j , k (see Section 3.3): Draw PG random values ω j , k for each jurisdiction j { 1 , , N } and outcome category k { 1 , 2 , 3 } as follows:
    ω j , k P G c = k 4 M j , c , η j , k
    where η j , k is as defined in Section 3.3.
  • β k (see Section 3.3): Draw coefficients β k for k { 1 , 2 , 3 } as follows:
    β k N V β k 1 m β k , V β k 1
    where V β k and m β k are defined as in Section 3.3, and the prior covariance matrix B for β k is defined as in Section 3.4.
  • γ i (see Section 3.4): For each covariate i { 1 , , p } , draw
    γ i P γ i = 1 | Y , β , γ i = N β | 0 , D γ R D γ ; γ i , γ i = 1 × p i k { 0 , 1 } N β | 0 , D γ R D γ ; γ i , γ i = k p i k 1 p i 1 k .
    where the matrices D γ and R, as well as the proportions p i for each i, are as defined in Section 3.4.
  • Σ (see Section 3.5): Draw
    Σ I W j = 1 N u j u j T + V 0 , N + s 0
    where u j is defined as in Section 3.5, V 0 is set to be the 3 × 3 identity matrix I 3 × 3 , and s 0 is set to 3, due to the three disjoint outcome categories.
  • u j (see Section 3.6): For each j { 1 , , N } , draw
    u j = u j , 1 , u j , 2 , u j , 3 N V j , u 1 m j , u , V j , u 1 ,
    where V j , u and m j , u are as defined in Section 3.6.

4. 2021 VRA Data Analysis Results

4.1. Data Description

This paper uses the ACS 2015–2019 5-year data, which was used for the most recent 2021 VRA determination. The ACS releases 1-year and 5-year data products. The 5-year products aggregate and re-weight data collected over a 5-year period, allowing increased precision of population estimates. The 5-year data are particularly useful for estimating features of small geographic areas or small domains in which 1-year estimates are too imprecise for release under Census Bureau statistical quality guidelines.
Data for the four disjoint categorical outcomes described in Equation (1) were obtained from the ACS 2015–2019 data, as were the following covariates:
  • Logit-transformed fraction of voting-age persons who are citizens;
  • Logit-transformed fraction of citizens that are limited English-proficient;
  • Proportion of voting-age persons who are non-Hispanic White in each geography;
  • Proportion of voting-age persons with no college education in each geography;
  • Average number of voting-age people per housing unit in each geography;
  • Average age among voting-age persons in any AIAN LMG in each geography;
  • Proportion of voting-age persons in poverty in each geography;
  • Proportion of voting-age persons speaking a language other than English at home in each jurisdiction;
  • Proportion of foreign-born voting-age persons in each jurisdiction;
  • Average years in US (as of 2019) of voting-age foreign-born persons in each jurisdiction.

4.2. Model Comparison

This section highlights the improvement in model diagnostic results made by the Bayesian model described in this paper over the diagnostic results for the model used to produce the 2021 VRA determinations. We estimate the new model for 73 Asian, Hispanic, and Native American LMGs in the 7859 jurisdictions with respondents in the 2015–2019 ACS.
Note that direct population estimates from the ACS are reliable when they are made within domains with large sample sizes. Slud et al. [1] evaluate the performance of models by examining the discrepancy between their estimates and direct ACS estimates when aggregated over larger domains. The model is judged to be performing well when the direct and model-based estimates agree in these large sample domains.
We follow Slud et al. [1] in using three metrics for model evaluation: Delta , PctRel Δ , and Stdiz Δ . Delta is the difference between model predictions N ˜ and the ACS direct estimates N ^ . PctRel Δ is the percent relative difference 100 · ( N ˜ N ^ ) / N ^ . Stdiz Δ is the standardized relative difference ( N ˜ N ^ ) / S E ( N ^ ) , where S E ( N ^ ) is the square root of the “successive difference replication” (SDR) estimate of the variance of the direct estimate N ^ described in the Appendix of Slud et al. [1]. This is the standard error estimate generally used by the ACS, and for large domains, the standardized difference ( N ^ N t r u e ) / S E ( N ^ ) is distributed approximately as a standard normal. Because of this, Stdiz Δ roughly measures the departure from the null hypothesis that the N ˜ prediction and N ^ estimate are the same. However, since model-based predictions are conditioned on the direct estimate, N ˜ and N ^ will be positively correlated, and Stdiz Δ will be smaller in absolute value than a N ( 0 , 1 ) deviate. We also follow Slud et al. [1] by evaluating these estimates in domains created by aggregating jurisdictions with voting-age sample sizes in the following pre-specified ranges: 1–4, 5–12, 13–25, 26–50, 51–200, and 201– 10 6 (the release of the statistics in this paper has been approved by the Census Bureau’s Disclosure Review Board (DRB) with the approval number CBDRB-FY25-0130). Priors were chosen such that model-diagnostic results were optimized.

4.2.1. Bangladeshi LMG Results

Table 3 compares the 2021 VRA model results for the Bangladeshi LMG published by Slud et al. [1] to the model results for the same LMG from our new Bayesian MLN model. In Table 3, the first three lines are results from the 2021 VRA model and the next three lines are results from the new model.
For the jurisdictions with the number of voting-age persons larger than 4, i.e., [5, 12], [13, 25],…, [201, 1 × 10 6 ], both the new Bayesian MLN model and the 2021 VRA model provided comparable Stdiz Δ values that were less than 1.96, the 97.5th percentile point of the standard normal distribution. The new Bayesian MLN model outperformed the 2021 VRA model for the jurisdictions with smaller number of voting-age persons ranging from 1 to 4; the old model resulted in a StdzDelt of 3.2, while the new model produced 1.2. The enhanced prediction accuracy in small areas can likely be attributed to the impact of the SSVS variable selection procedure in the new model.

4.2.2. Sri Lankan LMG Results

Table 4 shows similar results for the Sri Lankan LMG, which was also used as an example in Slud et al. [1]. Like the Bangladeshi LMG, the new Bayes MLN model resulted in more accurate prediction for jurisdictions with a voting-age sample size ranging from 1 to 4.

4.2.3. Overall Results

Table 5 compares the results for the new Bayesian MLN and old 2021 VRA models across all LMGs. The sum of absolute Stdiz Δ values across all 73 LMGs was produced for each sample size bin. The 73 LMGs include 21 Asian LMGs, 51 American Indian and Alaska Native (AIAN) LMGs, and a single Hispanic LMG that includes all racial groups. Overall, the sum of the absolute StdzDelt values across all of these LMGs is smaller for the new model than it was for the 2021 VRA model.

4.2.4. Improvement Rates

For each sample size bin, the first row of Table 6 shows the number of LMGs with absolute Stdiz Δ statistics larger than 1.96 in the 2021 VRA models, and the second shows the number of LMGs with absolute Stdiz Δ statistics larger than 1.96 in the new Bayes model. The third row shows the number of LMGs with absolute Stdiz Δ values greater than 1.96 in at least one of the two models, and the fourth row shows the number of LMGs where both models produce absolute Stdiz Δ values greater than 1.96. The fifth row of Table 6 shows the number of LMGs where |Stdiz Δ | > 1.96 for both the old and new models, but the new model produces better absolute Stdiz Δ values than the old model. Finally, the sixth row shows the same data as the fifth row as a percentage of the number of LMGs where at least one model has |Stdiz Δ | > 1.96.
It should be noted that in most LMGs, the (0, 4] sample size bin contains more jurisdictions than the other bins, since many LMGs are sparsely sampled. In the first two rows of this table, we see that the number of LMGs with model estimates that are more than 1.96 SDR SDs away from the direct estimates in the (0, 4] sample size bin decreases by almost half in the new model compared to the old model. These numbers remain the same as the old model in the other sample size bins, except in the (4, 12] bin, for which there is one additional LMG that is “far” from the direct estimates. We also see that the new model performs better than the old model even when the model estimates are far from the direct estimates. In particular, the final row of Table 6 shows that in 85% of LMGs where the model estimates were more than 1.96 SDs away from the direct estimates, the new model performed better than the old model.

4.2.5. Residual Plots

Figure 1 compares the performance of the old and new models. Each point represents a jurisdiction across all 73 LMGs. The residual value here is the difference between “Direct” and "Model" estimates. LEPprop denotes the proportion of voting-age citizens who have limited English proficiency (or LEP / CIT), while ILLrat represents the illiteracy rate among citizens with limited English proficiency (or ILL / LEP). If all the points in this residual plot were on the 45-degree line, that would indicate that there was no difference between the old and new models. The closer the points are to a horizontal line in the center of the plot, the better the new model is performing for those jurisdictions. The fact that both plots show several points below the 45-degree lines and above the horizontal line at the center indicates that the old model has larger residuals than the new model for those jurisdictions. This pattern is consistent through all sample size bins.

5. Discussion

In this paper, a new statistical methodology was proposed for the Bayesian multinomial logit model for US Voting Rights Act Section 203 determinations. Our methodological contribution involved adopting PG augmentation to linearize the nonlinear optimization problem, resulting in closed-form solutions for posterior distributions. This enhanced computational stability in the Gibbs sampling steps, which enabled us to fit one model for all jurisdictions. The former 2021 VRA determinations used different models for different jurisdictions—that is, if one model did not fit, then other models were tried for the same jurisdiction data.
Our method was developed based on the works of George and McCulloch [8] and Polson et al. [6]. George and McCulloch [8] did not suggest formal convergence criteria or diagnostics. Instead, they relied on empirical observations of convergence. Polson et al. [6] implemented the PG sampling algorithm in the R package BayesLogit. Note that while convergence diagnostics such as trace plots can indicate non-convergence, they cannot definitively prove convergence, especially given the substantial number of parameters in our analysis. The ultimate evidence of convergence was the improvement in the new model’s diagnostic measures over the 2021 VRA model.
Using SSVS requires the specification of the prior parameters c i and τ i . Setting τ i to 0.5 and c i to 10 produced desirable diagnostic results for all LMGs, as discussed in Section 4. In exchange for the additional effort required to identify suitable SSVS priors, the new PG-based Bayesian model demonstrates improved efficiency compared to the previous STAN-based Bayesian model developed for the 2021 VRA project. For example, the new model produced parameter estimates for the Bangladeshi LMG’s 517 jurisdictions in approximately 8 min. In contrast, the 2021 Bayesian VRA model for this LMG could not be estimated at all using STAN. The use of other Bayesian variable selection procedures, such as Bayesian LASSO [9], is a topic of interest for future research.
The estimates from the new model can be viewed as Bayesian model averages, because the posterior parameters were sampled and averaged over various models that had different sets of selected covariates. The overall results for the 2021 VRA data showed that the new model, which used stochastic search variable selection(SSVS) [8] separately in each LMG, outperformed the previous model, which used the same covariates in every LMG. The diagnostic data presented for the Bangladeshi and Sri Lankan LMGs are examples of how the new model yielded better results than the 2021 VRA models, especially in jurisdictions with small sample sizes of between one and four voting-age persons.
The nonlinearity of the logistic link function of the MLN model and the unseen latent random effects variables were computationally challenging issues. Both the frequentist adaptive Gaussian quadrature procedure and Bayesian inference can handle random effects, as was demonstrated in previous 2021 VRA determinations [1]. The frequentist model was fit by using adaptive Gaussian quadrature to optimize the likelihood function, which required the random effects variable to be averaged out. This approach makes it technically difficult to adopt variable selection procedures, since the entire score function would have to be modified for such an additional procedure. In contrast, Bayesian methods treat the random effects as just another parameter, which can be drawn in Gibbs sampling. STAN was used to sample the random effects and their covariance matrix for the 2021 VRA Bayesian model, but it did not converge for smaller LMGs. Our new Bayesian model used PG augmentation to obtain closed-form solutions for posterior distributions, which made implementing Bayesian stochastic search variable selection computationally feasible in more LMGs. This lead to noticeable improvements in our overall diagnostic results.

Author Contributions

Conceptualization, J.K. and A.C.H.; data curation, A.C.H.; formal analysis, J.K. and A.C.H.; methodology, J.K.; project administration, J.K.; software, J.K. and A.C.H.; supervision, J.K.; validation, J.K.; visualization, J.K.; writing—original draft preparation, J.K.; writing—review and editing, A.C.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The datasets presented in this article are not readily available because they are protected by Title 13, U.S.C., in the interests of respondent privacy.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Derivation of the β Posterior [6]

Without loss of generality, we omit the outcome category indicator k from variable notations in this section. This means that β k will be written as β , U k will be written as U, Ω k will be written as Ω , etc. Recall that Z= z 1 , . , z N T , z j denote y j n j / 2 / ω j , and suppose ω j is sampled from P G n j , X j β . Then, P Z | β , X P β | b , B is proportional to
exp 1 2 Z X β T Ω Z X β + β b T B 1 β b = exp 1 2 Z X β T Ω 1 2 T Ω 1 2 Z X β + β b T B 1 β b = exp 1 2 Ω 1 2 Z X β T Ω 1 2 Z X β + β b T B 1 β b = exp 1 2 Z X β T Z X β + β b T B 1 β b = exp 1 2 Z X β ^ + X β ^ X β T Z X β ^ + X β ^ X β + β b T B 1 β b = exp 1 2 Z X β ^ T Z X β ^ + X β X β ^ T X β X β ^ + β b T B 1 β b = exp 1 2 Z X β ^ T Z X β ^ + β β ^ T X T X β β ^ + β b T B 1 β b = exp 1 2 Z X β ^ T Ω Z X β ^ + β β ^ T X T Ω X β β ^ + β b T B 1 β b exp 1 2 β β ^ T X T Ω X β β ^ + β b T B 1 β b = exp 1 2 β V β 1 m β T V β β V β 1 m β
where V β is X T Ω X + B 1 and m β is X p × N T Ω N × N X N × p β ^ p × 1 + B 1 b . Note that X T Ω X β ^ is X T Ω Z because β ^ is such that X T Z X β ^ = 0, where A = Ω 1 2 A and because X T Z X β ^ is the same as X T Ω Z X T Ω X β ^ .
Thus, the posterior distribution of β |Z,X is
N V β 1 m β , V β 1 ,
where V β is X T Ω X + B 1 and m β is X T Ω z + B 1 b , which is X T κ + B 1 b , where κ i = y i n i / 2 .

Appendix B. Derivation of the β Posterior for the VRA Model

Without loss of generality, we omit the outcome category indicator k from variable notations in this section. This means that β k will be written as β , U k will be written as U, Ω k will be written as Ω , etc. Recall X is the matrix of covariates. Let U denote the N × 1 matrix of random effects variables, N ( b , B ) denote a standard normal prior for β , and the likelihood function P ( Z | β , U , X ) denote N ( U + X β , Ω 1 ) . Then, P Z | β , U , X P β | b , B is proportional to
exp 1 2 Z U X β T Ω Z U X β + β b T B 1 β b = exp 1 2 Z ˜ X β T Ω 1 2 T Ω 1 2 Z ˜ X β + β b T B 1 β b = exp 1 2 Ω 1 2 Z ˜ X β T Ω 1 2 Z ˜ X β + β b T B 1 β b = exp 1 2 Z X β T Z X β + β b T B 1 β b = exp 1 2 Z X β ^ + X β ^ X β T Z X β ^ + X β ^ X β + β b T B 1 β b = exp 1 2 Z X β ^ T Z X β ^ + x β X β ^ T X β X β ^ + β b T B 1 β b = exp 1 2 Z X β ^ T Z X β ^ + β β ^ T X T X β β ^ + β b T B 1 β b = exp 1 2 Z ˜ X β ^ T Ω Z ˜ X β ^ + β β ^ T X T Ω X β β ^ + β b T B 1 β b exp 1 2 β β ^ T X T Ω X β β ^ + β b T B 1 β b = exp 1 2 β V β 1 m β T V β β V β 1 m β
where V β is X T Ω X + B 1 and m β is X T Ω X β ^ + B 1 b , A denotes Ω 1 2 A for a variable A, and y and n denote y 1 ,…, y N and n 1 ,…, n N , respectively, as described in Section 2. Note that X T Ω X β ^ is X T y n / 2 X T Ω U because β ^ is such that X T Z X β ^ = 0, and hence X T Ω Z ˜ = X T Ω X β ^ . Now, X T Ω Z ˜ is X T Ω Z U and X T y n / 2 X T Ω U . Thus, X T y n / 2 X T Ω U = X T Ω X β ^ .

Appendix C. Derivation of P U | M , X , β , ω

Let us assume a Gaussian prior distribution N 0 , Σ for the random effects variables u j (= u j , 1 , u j , 2 , u j , 3 ) for jurisdiction j. Recall that M j denoted the counts of persons with language conditions. The conjugate Gaussian posterior for u j is proportinal to P M j | u j × P u j | Σ , as shown as follows.
exp η j , 1 k = 2 4 M j , k 1 + exp η j , 1 k = 1 4 M j , k exp η j , 2 k = 3 4 M j , k 1 + exp η j , 2 k = 2 4 M j , k exp η j , 3 k = 4 4 M j , k 1 + exp η j , 3 k = 3 4 M j , k × exp 1 2 u j T Σ 1 u j = N M j , 1 | η j , 1 , ω j , 1 1 × N M j , 2 | η j , 2 , ω 2 1 × N M j , 3 | η j , 3 , ω 3 1 × exp 1 2 u j T Σ 1 u j exp 1 2 ω j , 1 M j , 1 η j , 1 2 + ω j , 2 M j , 2 η j , 2 2 + ω j , 3 M j , 3 η j , 3 2 + u j T Σ 1 u j = exp 1 2 ω j , 1 u j , 1 M j , 1 u 2 + ω j , 2 u j , 2 M j , 2 u 2 + ω j , 3 u j , 3 M j , 3 u 2 + u j T Σ 1 u j = exp 1 2 u j M j u 1 × 3 T Ω j 3 × 3 u j M j u 3 × 1 + u j 0 T Σ 1 u j 0 N V j , u 1 m j , u , V j , u 1
where V j , u = Ω j + Σ 1 , m j , u = Ω j M j u , Ω j = d i a g ω j , 1 , ω j , 2 , ω j , 3 , M j , k u = M j , k X j β k , M j , 1 = ω 1 1 k = 2 4 M j , k 1 2 k = 1 4 M j , k , M j , 2 = ω 2 1 k = 3 4 M j , k 1 2 k = 2 4 M j , k , and M j , 3 = ω 3 1 k = 4 4 M j , k 1 2 k = 3 4 M j , k , and η j , k X j T β k + u j , k . The last two lines of the equations result by the algebra in Appendix E.

Appendix D. Derivation of P ( Σ | U )

Assume a multivariate normal distribution M V N 0 , Σ on U, which is
P U | Σ = j = 1 N 2 π 1 2 | Σ | 1 2 exp 1 2 u j T Σ 1 u j | Σ | N 2 exp 1 2 j = 1 N u j T Σ 1 u j = | Σ | N 2 exp 1 2 t r j = 1 N u j u j T Σ 1 .
An inverse Wishart I W V 0 , s 0 for the prior distribution with known scale matrix V 0 and the degrees of freedom s 0 for a p × p covariance matrix Σ is
P Σ = | V 0 | 1 2 s 0 2 s 0 p / 2 Γ p s 0 / 2 | Σ | 1 2 s 0 + p + 1 exp 1 2 t r V 0 Σ 1
The conditional posterior P Σ | U is proportional to
P U | Σ × P Σ | Σ | 1 2 N exp 1 2 t r j = 1 N u j u j T Σ 1 × | Σ | 1 2 s 0 + p + 1 exp 1 2 t r V 0 Σ 1 = | Σ | 1 2 N + s 0 + p + 1 exp 1 2 t r j = 1 N u j u j T Σ 1 + V 0 Σ 1 = | Σ | 1 2 N + s 0 + p + 1 exp 1 2 t r j = 1 N u j u j T + V 0 Σ 1 .

Appendix E. Algebraic Basics: Sum of Quadratic Matrices

Appendix E.1. Sum of Two Quadratic Matrices

x μ A T Σ A 1 x μ A + x μ B T Σ B 1 x μ B = x T Σ A 1 x 2 μ A T Σ A 1 x + μ A T Σ A 1 μ A + x T Σ B 1 x 2 μ B T Σ B 1 + μ B T Σ B 1 μ B = x T Σ A 1 + Σ B 1 x 2 Σ A 1 μ A + Σ B 1 μ B T x + μ A T Σ A 1 μ A + μ B T Σ B 1 μ B = x T V x 2 m T x + R = x V 1 m T V x V 1 m m T V 1 m + R N V 1 m , V 1
where V= Σ A 1 + Σ B 1 and m = Σ A 1 μ A + Σ B 1 μ B . Thus, μ A | x N μ A , Σ A and x N μ B , Σ B , so x | μ A N V 1 m , V 1 . V 1 = Σ A 1 + Σ B 1 1 and m = Σ A 1 μ A + Σ B 1 μ B .

Appendix E.2. Sum of N Quadratic Matrices

Consider x to be a vector of the VRA’s three disjoint outcomes. Then,
j = 1 N x μ j T Σ j 1 x μ j = x μ 1 T Σ 1 1 x μ 1 + x μ 2 T Σ 2 1 x μ 2 + + x μ N T Σ N 1 x μ N = x T Σ 1 1 x 2 μ 1 T Σ 1 1 x + x T Σ 2 1 x 2 μ 2 T Σ 2 1 + + x T Σ N 1 x 2 μ N T Σ N 1 + R = x T Σ 1 1 + Σ 2 1 + + Σ N 1 x 2 Σ 1 1 μ 1 + Σ 2 1 μ 2 + + Σ N 1 μ N T x + R = x T V N x 2 m N T x + R = x V N 1 m N T V N x V N 1 m N + R N V 1 m , V 1
V N = j = 1 N Σ j 1 and m N = j = 1 N Σ j 1 μ j . Thus, x | μ 1 , , μ N N V N 1 m N , V N 1 .

References

  1. Slud, E.V.; Franco, C.; Hall, A.; Kang, J. Statistical Methodology (2021) for Voting Rights Act, Section 203 Determinations. 2022. Available online: https://www.census.gov/library/working-papers/2022/adrm/RRS2022-06.html (accessed on 9 April 2024).
  2. Joyce, P.M.; Malec, D.; Little, R.J.; Gilary, A.; Navarro, A.; Asiala, M.E. Statistical Modeling Methodology for the Voting Rights Act Section 203 Language Assistance Determinations. J. Am. Stat. Assoc. 2014, 109, 36–47. [Google Scholar] [CrossRef]
  3. Slud, E.V.; Ashmead, R.; Joyce, P.; Wright, T. Statistical Methodology (2016) for Voting Rights Act, Section 203 Determinations. 2018. Available online: https://www.census.gov/library/working-papers/2018/adrm/RRS2018-12.html (accessed on 9 April 2024).
  4. Pinheiro, J.C.; Bates, D.M. Approximations to the Log-Likelihood Function in the Nonlinear Mixed-Effects Model. J. Comput. Graph. Stat. 1995, 4, 12–35. [Google Scholar] [CrossRef]
  5. Carpenter, B.; Gelman, A.; Hoffman, M.D.; Lee, D.; Goodrich, B.; Betancourt, M.; Brubaker, M.; Guo, J.; Li, P.; Riddell, A. Stan: A probabilistic programming language. J. Stat. Softw. 2017, 76, 1–32. [Google Scholar] [CrossRef] [PubMed]
  6. Polson, N.G.; Scott, J.G.; Windle, J. Bayesian Inference for Logistic Models Using Polya Gamma Latent Variables. J. Am. Stat. Assoc. 2013, 108, 1339–1349. [Google Scholar] [CrossRef]
  7. Polson, N.G.; Scott, J.G.; Windle, J. Package: BayesLogit 2.1. 2019. Available online: https://cran.r-project.org/web/packages/BayesLogit/index.html (accessed on 9 April 2024).
  8. George, E.I.; McCulloch, R.E. Variable Selection via Gibbs Sampling. J. Am. Stat. Assoc. 1993, 88, 881–889. [Google Scholar] [CrossRef]
  9. Park, T.; Casella, G. The Bayesian Lasso. J. Am. Stat. Assoc. 2008, 103, 681–686. [Google Scholar] [CrossRef]
Figure 1. Residual plots.
Figure 1. Residual plots.
Mathematics 13 00945 g001
Table 1. Hypothetical VRA proportions for an LMG.
Table 1. Hypothetical VRA proportions for an LMG.
Jurisdiction# of Voting-Aged PersonsCITpropLEPpropILLprop π CIT π LEP π ILL X j
Jurisdiction 110000.800.100.010.800.100.01 X 1
Jurisdiction 2100.950.250.020.920.190.07 X 2
............... ... ... ... ...
Jurisdiction N11000.930.100.01 X N
Table 2. Likelihood contributions.
Table 2. Likelihood contributions.
Binomial Prob.RatioEquivalent Prob.Log-Likelihood Contribution for a Jurisdiction
v 1 CIT/VOT 1 p 1 M 1 log p 1 + n M 1 log 1 p 1
v 2 LEP/CIT 1 p 2 1 p 1 M 2 log p 2 1 p 1 + n M 1 M 2 log 1 p 2 1 p 1
v 3 ILL/LEP 1 p 3 1 p 1 p 2 M 3 log p 3 1 p 1 p 2 + n M 1 M 2 M 3 log 1 p 3 1 p 1 p 2
Table 3. Comparing new Bayes model and older results for Bangladeshi LMG.
Table 3. Comparing new Bayes model and older results for Bangladeshi LMG.
[1, 4][5, 12][13, 25][26, 50][51, 200][201, 1 × 10 6 ]
Delta (Old)611.99.7−143.0−47.5−319.9−64.3
PctRel Δ (Old)47.20.4−6.2−1.9−3.1−0.4
Stdiz Δ (Old)3.20.0−0.5−0.2−0.6−0.1
Delta (New)223.6−117.3−126.1−62.442.48.5
PctRel Δ (New)17.3−5.3−5.5−2.50.40.0
Stdiz Δ (New)1.2−0.4−0.5−0.20.10.0
Table 4. Comparing new Bayes model and older results for Sri Lankan LMG.
Table 4. Comparing new Bayes model and older results for Sri Lankan LMG.
[1, 4][5, 12][13, 25][26, 50][51, 200][201, 1 × 10 6 ]
Delta (Old)228.5−143.3−55.9−8.3−49.6−11.4
PctRel Δ (Old)54.3−18.4−6.7−1.9−3.6−1.4
Stdiz Δ (Old)2.5−1.0−0.3−0.1−0.2−0.1
Delta (New)87.2−174.311.562.9−2.5−9.7
PctRel Δ (New)20.7−22.41.414.4−0.2−1.2
Stdiz Δ (New)0.9−1.20.10.60.0−0.1
Table 5. Comparing new Bayes model and older results across all LMGs.
Table 5. Comparing new Bayes model and older results across all LMGs.
[1, 4][5, 12][13, 25][26, 50][51, 200][201, 1 × 10 6 ]
Old104.950.14328.721.712.9
New79.744.03922.017.08.5
Table 6. Comparing new Bayes (denoted by “New”) and older results across all LMGs.
Table 6. Comparing new Bayes (denoted by “New”) and older results across all LMGs.
(0, 4](4, 12](12, 25](25, 50](50, 200](200, 1 × 10 6 ]All
# LMGs with |Stdiz Δ |>1.96 (Old)155211024
# LMGs with |Stdiz Δ |>1.96 (New)76210016
# LMGs with |Stdiz Δ |>1.96 (Old and New)74210014
# LMGs with |Stdiz Δ |>1.96 (Old or New)157211026
# LMGs with |Stdiz Δ |>1.96 (Improved)144211022
% LMGs with |Stidz Δ |>1.96 (Improved)93%57%100%100%100%NA85%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hall, A.C.; Kang, J. Inference with Pólya-Gamma Augmentation for US Election Law. Mathematics 2025, 13, 945. https://doi.org/10.3390/math13060945

AMA Style

Hall AC, Kang J. Inference with Pólya-Gamma Augmentation for US Election Law. Mathematics. 2025; 13(6):945. https://doi.org/10.3390/math13060945

Chicago/Turabian Style

Hall, Adam C., and Joseph Kang. 2025. "Inference with Pólya-Gamma Augmentation for US Election Law" Mathematics 13, no. 6: 945. https://doi.org/10.3390/math13060945

APA Style

Hall, A. C., & Kang, J. (2025). Inference with Pólya-Gamma Augmentation for US Election Law. Mathematics, 13(6), 945. https://doi.org/10.3390/math13060945

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop