Next Article in Journal
Unique Method for Prognosis of Risk of Depressive Episodes Using Novel Measures to Model Uncertainty Under Data Privacy
Previous Article in Journal
Performance Improvement for Discretely Modulated Continuous-Variable Measurement-Device-Independent Quantum Key Distribution with Imbalanced Modulation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fiducial Inference in Linear Mixed-Effects Models

School of Mathematics and Statistics, Qingdao University, Qingdao 266071, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(2), 161; https://doi.org/10.3390/e27020161
Submission received: 30 September 2024 / Revised: 20 January 2025 / Accepted: 28 January 2025 / Published: 3 February 2025
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
We develop a novel framework for fiducial inference in linear mixed-effects (LME) models, with the standard deviation of random effects reformulated as coefficients. The exact fiducial density is derived as the equilibrium measure of a reversible Markov chain over the parameter space. The density is equivalent in form to a Bayesian LME with noninformative prior, while the underlying fiducial structure adds new benefits to unify the inference of random effects and all other parameters in a neat and simultaneous way. Our fiducial LME needs no additional tests or statistics for zero variance and is more suitable for small sample sizes. In simulation and empirical analysis, our confidence intervals (CIs) are comparable to those based on Bayesian and likelihood profiling methods. And our inference for the variance of random effects has competitive power with the likelihood ratio test.

1. Introduction

The linear mixed-effects (LME) model is very popular in many fields like physics, biology and social sciences [1], where both within-subject and between-subject variations are studied. There have been a wealth of LME-based methodologies developed in the past decades, such as model selection [2], model averaging [3], semiparametric [4] and nonparametric [5] estimations, etc. Large-sample asymptotics of likelihood-based LME models are well studied [6]. Bayesian LME models also perform well under medium-to-large samples. For small samples, the inference based on the maximum likelihood needs finite-sample corrections or approximations [7], and Bayesian models rely much more on the prior.
The analysis of variance (ANOVA) and t-test summaries are two common ways of statistical inference in LME models. For fixed effects, F statistics in ANOVA and t statistics are available in most software. However, testing zero-variance components in random effects is challenging for small samples and has not been well resolved in the statistical literature. When some of the within-subject variances are zero or close-to-zero, Bayesian estimations may encounter issues [8]. And the inference of zero variance involves additionally constructed statistics like the profiling restricted maximum likelihood (REML) ratio test [9] or posterior evidence ratio with properly built prior [10]. Comparisons among multiple nested models are often unskippable.
From a frequentist viewpoint, aside from asymptotic approximations, the inference can be facilitated by nonparametric approaches [11,12], which are also known to be performant in LME models for small sample sizes. According to a comparative study [13] on various bootstrap methods in LME models, a nonparametric bootstrap with resampling on subjects is recommended [11].
Fiducial inference, initiated by Fisher [14] and revitalized in the 21st century [15], has unique advantages in small-sample inference in parametric statistical models, despite its limits in nonparametric statistics and computing cost [16]. Unlike Bayesian posterior built upon prior, fiducial distribution derived from data generation equations thus relieves the need for prior. For LME, Li et al. [17] proposed a fiducial test for within-subject variances, but they did not consider other parameters. Hari et al. [18] studied a two-component LME with partially known structures on covariances. Hannig et al. [19] developed a generalized fiducial inference (GFIlmm) for interval data. Although GFIlmm can be adapted to non-interval cases by artificially adding the intervals, the estimations vary with the interval widths. And its numerical algorithm in practice is inefficient on models with more than nine parameters.
In this study, we propose a new framework for non-asymptotic inference in LME models via a fiducial approach, which reduces computational burden and does not require duplicates within each group. Moreover, our approach naturally incorporates the zero-variance inference, thus cutting the labor of extra tests and calculation.
The rest of this article is organized as follows: Section 2 and Section 3 are about the methodology and algorithms. Section 4 is the simulation and Section 5 is a real data illustration. Section 6 is the conclusion and discussion of our study.

2. Fiducial Distribution in LME

We assume the data is generated as
y i = x i β + z i γ i + e i , i = 1 , , m ,
where β R p × 1 is a vector of fixed coefficients; γ i N q ( 0 , D ) is a random coefficient vector; y i is the n i × 1 response of the ith subject; x i and z i are the n i × p and n i × q covariate matrix of fixed and random effects, respectively; and e i N ( 0 , σ 2 I n i ) is independent of γ i , n i 1 , i = 1 , , m , i = 1 m n i = n . In this study, we assume a diagonal D = d i a g ( σ 1 2 , , σ q 2 ) .

2.1. Conditional Fiducial Distribution

Let δ = ( σ 1 , , σ q ) , U i i . i . d N ( 0 , I q ) and z U , i = z i d i a g ( U i ) , then model (1) can be rewritten as
y i = x i β + z U , i δ + e i , i = 1 , , m .
It can be further specified in matrix form as
Y = X β + Z U δ + ϵ ,
with Y = ( y 1 , , y m ) , X = ( x 1 , , x m ) , Z U = ( z U , 1 , , z U , m ) , U = ( U 1 , , U m ) and ϵ = ( e 1 , , e m ) . The σ j in random effects are formulated as a coefficient vector δ in Equation (2). This brings two benefits: 1. The inference of random effects can be realized by a general fiducial recipe [16] like all other parameters, avoiding extra tests and statistics like the LR tests. 2. We do not need n i q . The model can still be estimated even if n i = 1 . So, our method can handle inadequate within-subject measures.
Let θ = ( β , δ ) , X U = ( X , Z U ) n × ( p + q ) and P U = X U ( X U X U ) 1 X U , then model (2) given U = u can be reformed as
Y = X β + Z u δ + ϵ = X u θ + ϵ ,
For observed Y = y , the conditional fiducial density of θ and σ 2 given u, denoted as r y ( θ | σ 2 , u ) and r y ( σ 2 | u ) , respectively, could be derived from the conditional fiducial quantities [17] below:
Q y ( σ 2 | u ) = y ( I P u ) y / ξ ,
Q y ( θ | u ) = ( X u X u ) 1 X u y + Q y 1 2 ( σ 2 | u ) ( X u X u ) 1 2 ζ ,
where ξ χ n p q 2 , ζ N ( 0 , I p + q ) independently, irrespective of any parameters in an LME model. Let η = ( θ , σ 2 ) , then the conditional fiducial density of η given u is
r y ( η | u ) = r y ( σ 2 | u ) r y ( θ | σ 2 , u ) = e x p y ( I P u ) y 2 σ 2 ( y ( I P u ) y 2 σ 2 ) n p q 2 1 σ 4 Γ ( n p q 2 ) e x p 1 2 σ 2 ( θ ( X u X u ) 1 X u y ) ( X u X u ) ( θ ( X u X u ) 1 X u y ) ( 2 π ) p + q 2 σ p + q | X u X u | 1 2 = ( y ( I P u ) y ) n p q 2 1 σ n p q + 2 Γ ( n p q 2 ) 2 n p q 2 1 e x p ( y X u θ 2 2 σ 2 ) ( 2 π ) p + q 2 σ p + q | X u X u | 1 2 = 2 | X u X u | 1 2 ( y ( I P u ) y ) n p q 2 1 Γ ( n p q 2 ) π p + q n 2 e x p ( y X u θ 2 2 σ 2 ) ( 2 π ) n 2 σ n + 2 = C ( X u , y ) p ( y | u , η ) σ 2 .
Here, C ( X u , y ) = 2 | X u X u | 1 2 ( y ( I P u ) y ) n p q 2 1 Γ ( n p q 2 ) π ( p + q n ) / 2 and p ( y | u , η ) is the likelihood function of (2) given U = u .
U is invisible in reality. But, it is easy to see that U i | η , y N q ( E ( U i | η , y ) , V a r ( U i | η , y ) ) with
E ( U i | η , y ) = D 1 2 z i ( σ 2 I n i + z i D z i ) 1 ( y i x i β ) , V a r ( U i | η , y ) = I q D 1 2 z i ( σ 2 I n i + z i D z i ) 1 z i D 1 2 ,
i = 1 , , m . We denote the conditional density of U as p ( u | η , y ) hereinafter.

2.2. Gibbs Sampler and the Final Fiducial Distribution

With both r y ( η | u ) and p ( u | η , y ) available, the fiducial distribution could be realized by a Gibbs sampler. Let K ( η , η ˜ ) = r y ( η | u ) p ( u | η ˜ , y ) d u . By Equations (5) and (6), we have
K ( η , η ˜ ) = C ( X u , y ) p ( y | u , η ) σ 2 p ( u | y , η ˜ ) d u = C ( X u , y ) p ( y , u | η ) σ 2 p ( u ) p ( u | y , η ˜ ) d u = p ( y | η ) σ 2 C ( X u , y ) p ( u ) p ( u | y , η ) p ( u | y , η ˜ ) d u = p ( y | η ) σ 2 h ( η , η ˜ ) ,
where h ( η , η ˜ ) = C ( X u , y ) p ( u ) p ( u | y , η ) p ( u | y , η ˜ ) d u and p ( y | η ) is the likelihood function with U integrated out. Obviously, h ( · , · ) is symmetric and p ( y | η ) σ 2 satisfies the detailed balance below:
p ( y | η ˜ ) σ ˜ 2 K ( η , η ˜ ) = p ( y | η ˜ ) σ ˜ 2 p ( y | η ) σ 2 h ( η , η ˜ ) = p ( y | η ) σ 2 p ( y | η ˜ ) σ ˜ 2 h ( η ˜ , η ) = p ( y | η ) σ 2 K ( η ˜ , η ) .
The final fiducial density, as the stationary distribution of η , can be derived by the reversibility [20] of the Markov chain on η :
r y ( η ) p ( y | η ) σ 2 ,
which is equivalent to a Bayesian LME with uniform prior on β and δ and prior 1 σ on σ .
In practice, once a u sampled by Equation (6) makes X u ( X u X u ) 1 X u y close to y, the Q y ( σ 2 | u ) in Equation (3) also becomes close to 0, making Q y ( θ | u ) in Equation (4) nearly constant and the Monte Carlo Markov Chain (MCMC) degenerated. So, we restrict σ ’s parameter space away from zero by a small c 0 > 0 so that σ 2 [ c 0 , + ) and use the constrained fiducial quantity by a trimmed χ n p q 2 as follows:
Q y * ( σ 2 | u ) = y ( I X u ( X u X u ) 1 X u ) y F χ n p q 2 1 ( F χ n p q 2 ( y ( I X u ( X u X u ) 1 X u ) y c 0 ) U ( 0 , 1 ) ) ,
where F ( · ) is the cumulative distribution function, F 1 ( · ) is its inverse, and U ( 0 , 1 ) is a uniform random variable on ( 0 , 1 ) . In this paper, we set c 0 = 0.01 . The Gibbs sampling algorithm is given in Algorithm 1.
Algorithm 1: Gibbs sampling for r y ( η )
  • Initialize U ( 0 ) by U i ( 0 ) i . i . d N ( 0 , I q ) , i = 1 , , m .
  • Given current U ( l ) , sample β ( l ) , δ ( l ) from Equation (4) and σ ( l ) from Equation (10), respectively.
  • Update U ( l + 1 ) by Equation (6).
  • For l = 0 , 1 , 2 , iterate between step 2 to 3 until the MCMC chain gets well mixed.
Remark 1. 
We can also start with Equation (9) by setting corresponding prior in ready-made Bayesian packages that employ algorithms with better quality like HMC or NUTS [21]. However, HMC is time consuming and loses the advantages of direct inference for σ j = 0 . The histogram of a Gibbs sample on δ can directly reflect how much the distribution is concentrated around zero. So, we mainly use Gibbs in this study and leave HMC for future research.
Remark 2. 
When D is non-diagonal, U i i . i . d N ( 0 , I q ) becomes U i i . i . d N q ( 0 , R ) , where R is the correlation matrix. And p ( u | η , y ) becomes p ( u | η , R , y ) . Given U = u , the fiducial density of R, say r u ( R ) , is also available [22]. The Gibbs sampler turns into iterations among three parts: r y ( η | u ) , p ( u | η , R , y ) and r u ( R ) . The fiducial density here is no longer equivalent to Equation (9) and cannot be readily implemented by Bayesian packages. We also leave this for future research.

3. Fiducial Inference for LME

Given the observed y, the fiducial distribution of a parameter of interest g = g ( η ) can be given as
F y ( g ) = { g ( η ) g } r y ( η ) d η .

3.1. Interval Estimation

A 1 α fiducial interval for parameter g ( η ) is formed by the α 2 and 1 α 2 quantiles of F y ( g ) , denoted as g α 2 and g 1 α 2 , respectively, which are the solutions of
F y ( g α 2 ) = α 2 , F y ( g 1 α 2 ) = 1 α 2 .
Then, the 1 α equal-tailed fiducial interval is [ g α 2 , g 1 α 2 ] .
Once we find a numerical sample of η as Section 2.2 described, the 1 α fiducial interval of g = g ( η ) can be obtained from the empirical distribution of F y ( g ) , which is given in Algorithm 2.
Algorithm 2: Interval estimation for g ( η )
  • Generate η 1 , , η N according to Algorithm 1.
  • Compute g i = g ( η i ) , i = 1 , , N , and sort g i in ascending order to get ( g ( 1 ) , , g ( N ) ) .
  • Find the α 2 sample quantile g ( [ N · α 2 ] ) and 1 α 2 sample quantile g ( [ N · ( 1 α 2 ) ] ) . Then [ g ( [ N · α 2 ] ) , g ( [ N · ( 1 α 2 ) ] ) ] is the 1 α fiducial interval estimation.
Noting that σ j is conventionally nonnegative, we use its zero-truncated distribution to construct the CI such that the left end is max { 0 , σ ( N α 2 ) , j } . According to remark 2.1 and theorem 2.2 in [17], a truncated fiducial quantity uniquely solved from the fiducial structural equations has the same theoretical property as the untruncated one for inference in a restricted parameter space.

3.2. Fiducial p-Value

For a hypothesis test H 0 : g ( η ) = g 0 with a two-sided alternative, the fiducial p-value is
p = 2 m i n { F y ( g 0 ) , 1 F y ( g 0 ) } .
Given a significance level α , H 0 is rejected when p < α . The computing procedure is given in Algorithm 3.
Algorithm 3: Fiducial p-value for H 0 : g ( η ) = g 0
  • Generate η 1 , , η N according to Algorithm 1.
  • Compute g i = g ( η i ) , i = 1 , , N .
  • Calculate p = 2 min i = 1 N 1 [ g i > g 0 ] N , i = 1 N 1 [ g i g 0 ] N .
  • Reject H 0 if p < α .
Remark 3. 
Algorithms 2 and 3 are routine constructions of the CI and p value in unimodal distributions. All the histograms of δ that we have plotted so far are unimodal in a Gibbs sampler, probably because the sign of δ and that of U are interdependent in model (2) while Gibbs updates δ largely based on U. But, this does not mean the bimodal shape will not occur for δ in other algorithms bypassing the sampling of r ( θ | u , y ) since δ in Equation (9) is symmetric at about 0. For a bimodal distribution, routine p value and CI are not applicable but a histogram or density curve can directly tell.

4. Simulation

4.1. Confidence Intervals

We compare the coverages and lengths of CIs constructed by our method (Fiducial), the profiling method (Profiling) in R package lme4 [9], the Bayesian LME (Bayeisan) with default prior in R package brms (version 2.22.0) [10], and the nonparametric bootstrap (Bootstrap) in [13]. We generate a chain over 6000 in length, with the first 300 as a warmup, and select every 3 steps to gather a sample size N = 2000 for the Gibbs sampler in Algorithm 1. We also try the HMC in Remark 1 by setting a uniform prior for β and σ j , and 1 σ is approximated by an inverse gamma prior IG(0.001, 0.001) in brms (The density of IG(a, b) is b a e b / σ Γ ( a ) σ a + 1 , σ > 0 . 1 σ = lim b 0 lim a 0 b a e b / σ Γ ( a ) σ a + 1 thus can be approximated by IG(a,b) with a and b close to 0.) Bootstrap CIs are constructed by percentiles of 2000 replicates resampled among subjects, and each replicate is estimated by REML (lmer in package lme4). All MCMCs are set as a single chain with a sample size of 2000. Coverages (nominal level 95 % ) are annotated in figures and tables, averaged over 1000 runs.
Example 1. 
y i j = β 0 + z i j γ i 1 + e i j with z i j N ( 0 , 1 ) , γ i 1 N ( 0 , σ 1 2 ) , e i j N ( 0 , σ 2 ) , 1 j 2 , 1 i 10 , β 0 = 0.8 , σ 1 = 0.75 , σ = 0.7 . In general, the Profiling CIs are narrower but susceptible to low coverages (Figure 1). Fiducial CIs are comparable to the Bayesian except those on σ 1 are often wider. This can be improved by HMC yet it is much slower. We use Gibbs for our method afterward. Bayesian LME is time-consuming yet performs well, partly due to its advantages in small samples and partly due to the suitability of the default prior for this parameter setting. Bootstrap has the lowest coverages; perhaps the sample size here is too small for it to take effect. The relative CPU time for model fit plus CI construction in fiducial (Gibbs), profiling, Bayesian, and bootstrap is about 2:1:30:40. Given the underperformance of bootstrap in coverage and computation efficiency, we do not discuss it in later examples. The variations in CI length are similar across different methods. So, we summarize the median length in the table hereinafter.
Example 2. 
y i j = β 0 + γ i 0 + z i j γ i 1 + e i j with γ i 0 N ( 0 , σ 0 2 ) , σ 0 = 0 , and the others the same as example 1. That is, we add a false random effect γ i 0 0 . Bayesian LME is unable to directly detect zero variance (coverage = 0 for σ 0 , Table 1). Profiling CIs can have zero end points to keep reasonable coverages, but other parameters are affected, especially σ. Also, 5 % 10 % of l m e r ’s output contains NA or ∞ when n i = q = 2 . This is not a big issue when n i = 4 > q , but σ’s coverage is still lower. In contrast, fiducial CIs capture both zero and nonzero variances in random effects and have stable coverages (Table 1).
Example 3. 
y i = β 0 + γ i 0 + i · β 1 + i · γ i 1 + e i with γ i 0 , γ i 1 , e i all from N ( 0 , 1 ) , i = 1 , . . . , 10 , i . e . , a longitudinal setting with n i = 1 , q = 2 so n i < q . l m e r does not work in this scenario, which makes profiling and bootstrap inapplicable. Bayesian CIs, overall, are wider with lower coverages than fiducial CIs (Table 2).

4.2. Zero-Variance Test for Random Effects

We plot the histograms of fiducial samples on zero and nonzero δ . They are indeed different (Figure 2). Negative δ generated by Equation (4) helps in measuring the significance of the random effect. We further compare p value distributions under null (=0) and alternative (≠0) hypotheses by treating example 2 as y i j = β 0 + γ i 0 + z i j β 1 + z i j γ i 1 + e i j with β 1 = 0 and others unchanged. Frequentist p values are calculated by the package lmerTest [23] with fixed effects based on Satterthwaite-approximated t tests and random effects based on likelihood ratio (LR) tests. Fiducial p values are calculated simultaneously as described in Section 3. The type I error and power at a 5 % significance level are annotated in Figure 3. The Satterthwaite p value has a higher type I error than the fiducial test (Figure 3). The type I errors of the fiducial and LR tests do not differ much, but the power of the fiducial test is higher. Moreover, the LR p value has a weird peak near 1 under alternative hypothesis (Figure 3), which is an obvious drawback in small data analysis. From this perspective, we think our fiducial approach can contribute to the inference in LME for small samples.

4.3. Comparison with GFIlmm

For the GFIlmm in [19] (simplified as GFI hereinafter), we set a simpler scenario that their R package gfilmm (version 2.0.5) can run for comparsion: y i j = β 0 + γ i + β 1 x i j + e i . Here, β 0 = 0.8 , β 1 = 0.5 , γ i N ( 0 , δ 2 ) , x i j N ( 0 , 1 ) and e i N ( 0 , 1 ) , j = 1, 2; i = 1 , , 10 . Intervals are added as y i j 0.01 < y i j < y i j + 0.01 . Performances on both zero and nonzero δ are compared (Table 3). The CIs in the nonzero case are comparable, and GFI is better in the zero case. We also plot histograms of the two fiducial p values. At a 5 % significance level, our method has lower type I error and higher power than GFI (Figure 4).

5. Empirical Analysis

We use the sleepstudy data in R package lme4 as an illustration, which contains a sleep-deprived group of 18 subjects for 10 days of the study in [24]. Day0 and Day1 were adaptation and training, so altogether 18 × 8 = 144 observations. The reaction time of the ith subject on D a y j after sleep deprivation is modeled as
r e a c t i o n i j = β 0 + γ i 1 + β 1 D a y j + γ i 2 D a y j + e i j ,
with γ i 1 N ( 0 , σ 1 2 ) , γ i 2 N ( 0 , σ 2 2 ) , and e i N ( 0 , σ 2 ) . Both the fixed and random effects are significant [9]. We resample 25 % of the data for a subset with n = 9 × 4 = 36 to see how much the inference based on a smaller sample agrees with that on the whole dataset. We conduct resampling 50 times and plot the CI lengths of β 0 , β 1 , σ 1 , and σ 2 in Figure 5. Different from the simulation, the Bayesian intervals are much wider this time since the default prior does not well suit the real data. We annotate the percentage of CIs excluding 0 in Figure 5 as an estimated power, on which the fiducial LME are generally higher. We further try a combination of m = 18 and n i = 2 , retaining n = 36, but neither lmer nor brms can fit all the cases successfully this time. Our fiducial LME still works. The percentages of significance are 100 % on β 0 , 80 % on β 1 , 21 % on σ 1 , and 74 % on σ 2 , respectively.

6. Conclusions

We study the fiducial inference in LME with the group-level variation innovatively modeled as coefficients. We derive the fiducial density by the stationary theory of Markov chains and demonstrate its advantages in small sample sizes by using simulation and real data. Our study facilitates the inference of zero variance in random effects and reveals a deep relationship between fiducial and Bayesian inference. On one hand, it confirms the rationality of the commonly used noninformative prior in Bayesian LME. On the other, the equivalence to a Bayesian LME is only for independent random effects. It is difficult to make a direct inference of zero variance in random effects if we simply treat the final density as a posterior distribution with a flat prior. But, from the perspective of fiducial structural equations, it turns out to be feasible.
We compare four inferential methods on n i > q , n i = q , and n i < q respectively. Our approach outperforms others on n i q . This is helpful in prescreening random effects since q can easily exceed n i when we test if all the predictors have random effects. The minimal sample size on which our fiducial LME can work stably is much smaller than other methods. We also tried the parametric bootstrap in lme4. But, its improvement over nonparametric bootstrap is quite limited regarding coverage under the scenarios in Section 4.1.
There is also much space for improvement. The Gibbs algorithm is not fine enough. More advanced numerical algorithms can also be considered and will be studied in the future. The theoretical property of the truncated δ in the fiducial sample and the inference for correlated random effects (non-diagonal D) also need further investigation.

Author Contributions

C.Z. designed and conducted the research. X.L. provided theoretical support and edited the article. J.Y. participated in the simulation. H.G. provided resources, study materials and financial support. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Grants 11901328, 11871294 and 72171126 from the National Natural Science Foundation of China and Systems Science Plus Joint Research Program of Qingdao University (XT2024301).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jiang, J. Linear and Generalized Linear Mixed Models and Their Applications; Springer: New York, NY, USA, 2007. [Google Scholar]
  2. Buscemi, S.; Plaia, A. Model selection in linear mixed-effect models. Asta-Adv. Stat. Anal. 2019, 104, 1–47. [Google Scholar] [CrossRef]
  3. Zhang, X.; Zou, G.; Liang, H. Model averaging and weight choice in linear mixed-effects models. Biometrika 2014, 101, 205–218. [Google Scholar] [CrossRef]
  4. Waterman, M.J.; Birch, J.B.; Abdel-Salam, A.-S.G. Several nonparametric and semiparametric approaches to linear mixed model regression. J. Stat. Comput. Simul. 2015, 85, 956–977. [Google Scholar] [CrossRef]
  5. Dion, C. New adaptive strategies for nonparametric estimation in linear mixed models. J. Stat. Plan. Inference 2014, 150, 30–48. [Google Scholar] [CrossRef]
  6. Jiang, J. Asymptotic Analysis of Mixed Effects Models: Theory, Applications, and Open Problems; Chapman and Hall/CRC: Boca Raton, FL, USA, 2017. [Google Scholar]
  7. Breslow, N.E.; Clayton, D.G. Approximate inference in generalized linear mixed models. J. Am. Stat. Assoc. 1993, 88, 9–25. [Google Scholar] [CrossRef]
  8. Chung, Y.; Rabe-Hesketh, S.; Dorie, V.; Gelman, A.; Liu, J. A non-degenerate estimator for variance parameters in multilevel models via penalized likelihood estimation. Psychometrika 2013, 78, 685–709. [Google Scholar] [CrossRef] [PubMed]
  9. Bates, D. Fitting linear mixed-effects models using lme4. arXiv 2014, arXiv:1406.5823. [Google Scholar]
  10. Burkner, P.-C. brms: An r package for bayesian multilevel models using stan. J. Stat. Softw. 2017, 80, 1–28. [Google Scholar] [CrossRef]
  11. Das, S.; Krishen, A. Some bootstrap methods in nonlinear mixed-effect models. J. Stat. Plan. Inference 1999, 75, 237–245. [Google Scholar] [CrossRef]
  12. Flores-Agreda, D.; Cantoni, E. Bootstrap estimation of uncertainty in prediction for generalized linear mixed models. Comput. Stat. Data Anal. 2019, 130, 1–17. [Google Scholar] [CrossRef]
  13. Thai, H.-T.; Mentre, F.; Holford, N.; Veyrat-Follet, C.; Comets, E. A comparison of bootstrap approaches for estimating uncertainty of parameters in linear mixed-effects models. Pharm. Stat. 2013, 12, 129–140. [Google Scholar] [CrossRef]
  14. Fisher, R.A. On a point raised by m. s. bartlett on fiducial probability. Ann. Hum. Genet. 1937, 7, 370–375. [Google Scholar] [CrossRef]
  15. Hannig, J.; Iyer, H.; Lai, R.C.S.; Lee, T.C.M. Generalized fiducial inference: A review and new results. J. Am. Stat. Assoc. 2016, 111, 1346–1361. [Google Scholar] [CrossRef]
  16. Martin, R.; Liu, C. Inferential models: A framework for prior-free posterior probabilistic inference. J. Am. Stat. Assoc. 2013, 108, 301–313. [Google Scholar] [CrossRef]
  17. Li, X.; Su, H.; Liang, H. Fiducial generalized p-values for testing zero-variance components in linear mixed-effects models. Sci. China Math. 2018, 61, 1–16. [Google Scholar] [CrossRef]
  18. Lidong, E.; Hannig, J.; Iyer, H.K. Fiducial intervals for variance components in an unbalanced two-component normal mixed linear model. J. Am. Stat. Assoc. 2008, 103, 854–865. [Google Scholar]
  19. Cisewski, J.; Hannig, J. Generalized fiducial inference for normal linear mixed models. Ann. Stat. 2012, 40, 2102–2127. [Google Scholar] [CrossRef]
  20. Grimmett, G.R.; Stirzaker, D.R. Probability and Random Processes; Oxford University Press: Oxford, UK, 2001. [Google Scholar]
  21. Betancourt, M. A Conceptual Introduction to Hamiltonian Monte Carlo. arXiv 2017, arXiv:1701.02434. [Google Scholar] [CrossRef]
  22. Shi, W.J.; Hannig, J.; Lai, R.C.S.; Lee, T.C.M. Covariance estimation via fiducial inference. Stat. Theory Relat. Fields 2021, 5, 316–331. [Google Scholar] [PubMed]
  23. Kuznetsova, A.; Brockhoff, P.B.; Christensen, R.H.B. lmerTest package: Tests in linear mixed effects models. J. Stat. Softw. 2017, 82, 1–26. [Google Scholar] [CrossRef]
  24. Belenky, G.; Wesensten, N.; Thorne, D.; Thomas, M.; Sing, H.; Redmond, D.; Russo, M.; Balkin, T. Patterns of performance degradation and restoration during sleep restriction and subsequent recovery: A sleep dose-response study. J. Sleep Res. 2003, 12, 1–12. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Confidence intervals in example 1.
Figure 1. Confidence intervals in example 1.
Entropy 27 00161 g001
Figure 2. Fiducial density for δ = 0 and δ = 1 .
Figure 2. Fiducial density for δ = 0 and δ = 1 .
Entropy 27 00161 g002
Figure 3. p value comparison for fixed and random effects, n = 20.
Figure 3. p value comparison for fixed and random effects, n = 20.
Entropy 27 00161 g003
Figure 4. p value distributions on zero and nonzero random effects ( n i = 2 , m = 10).
Figure 4. p value distributions on zero and nonzero random effects ( n i = 2 , m = 10).
Entropy 27 00161 g004
Figure 5. CI length in resampled sleep data.
Figure 5. CI length in resampled sleep data.
Entropy 27 00161 g005
Table 1. Coverage (length) of CIs in example 2.
Table 1. Coverage (length) of CIs in example 2.
n i = 2 , m = 10 β 0 = 0.8 σ 0 = 0 σ 1 = 0.75 σ = 0.7
Fiducial0.99 (0.99)0.95 (1.01)0.97 (1.63)0.95 (0.71)
Profiling0.93 (0.80)0.97 (0.70)0.89 (1.38)0.87 (0.64)
Bayesian0.97 (1.03)0.00 (0.98)0.97 (1.49)0.96 (0.77)
n i = 4 , m = 5 β 0 = 0.8 σ 0 = 0 σ 1 = 0.75 σ = 0.7
Fiducial1.00 (1.30)0.98 (1.50)0.96 (2.21)0.95 (0.60)
Profiling0.95 (0.86)0.99 (0.65)0.93 (1.78)0.92 (0.54)
Bayesian1.00 (1.15)0.00 (1.38)0.97 (1.99)0.95 (0.67)
Table 2. Coverage (length) of CIs in example 3.
Table 2. Coverage (length) of CIs in example 3.
n i = 1 , m = 10 β 0 = 10 β 1 = 0 σ 0 = 1 σ 1 = 1 σ = 1
Fiducial0.99 (16.31)0.98 (3.36)1.00 (9.89)0.98 (2.06)1.00 (6.26)
ProfilingNANANANANA
Baysian0.97 (20.87)0.94 (4.36)0.97 (9.01)0.96 (2.79)0.77 (8.58)
Table 3. Coverages (length) of CIs in comparison with GFI.
Table 3. Coverages (length) of CIs in comparison with GFI.
Parameter β 0 = 0.8 β 1 = 0.5 δ = 0.75 σ = 1
Fiducial0.97 (1.56)0.94 (0.83)0.98 (1.55)0.96 (0.82)
Profling0.94 (1.43)0.92 (0.82)0.94 (1.25)0.96 (0.79)
GFI0.96 (1.60)96 (0.93)1.0 (1.45)0.95 (0.93)
Parameter β 0 = 0.8 β 1 = 0.5 δ = 0 σ = 1
Fiducial0.98 (1.23)0.95 (1.00)0.98 (1.21)0.95 (0.78)
Profling0.93 (0.96)0.93 (0.75)0.96 (0.88)0.91 (0.66)
GFI0.97 (1.15)0.94 (0.82)0.98 (1.09)0.96 (0.77)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, J.; Li, X.; Gao, H.; Zou, C. Fiducial Inference in Linear Mixed-Effects Models. Entropy 2025, 27, 161. https://doi.org/10.3390/e27020161

AMA Style

Yang J, Li X, Gao H, Zou C. Fiducial Inference in Linear Mixed-Effects Models. Entropy. 2025; 27(2):161. https://doi.org/10.3390/e27020161

Chicago/Turabian Style

Yang, Jie, Xinmin Li, Hongwei Gao, and Chenchen Zou. 2025. "Fiducial Inference in Linear Mixed-Effects Models" Entropy 27, no. 2: 161. https://doi.org/10.3390/e27020161

APA Style

Yang, J., Li, X., Gao, H., & Zou, C. (2025). Fiducial Inference in Linear Mixed-Effects Models. Entropy, 27(2), 161. https://doi.org/10.3390/e27020161

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop