Next Article in Journal
Biharmonic Conformal Immersions into a 3-Dimensional Conformally Flat Space
Previous Article in Journal
Classical and Bayesian Inference for the Two-Parameter Chen Distribution with Random Censored Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Poisson Mixed-Effects Count Regression Model Based on Double SCAD Penalty and Its Simulation Study

School of Science, Hubei University of Technology, Wuhan 430068, China
*
Authors to whom correspondence should be addressed.
Axioms 2026, 15(3), 214; https://doi.org/10.3390/axioms15030214
Submission received: 24 January 2026 / Revised: 7 March 2026 / Accepted: 10 March 2026 / Published: 12 March 2026

Abstract

This paper focuses on variable selection and parameter estimation for mixed-effects Poisson count regression models. To simultaneously select important variables in both fixed effects and random effects, we propose a double-penalized Poisson count regression model with the Smoothly Clipped Absolute Deviation (SCAD) penalty imposed on both components. To estimate the unknown parameters, we develop a new iterative algorithm called the Double SCAD–Local Quadratic Approximation (DSCAD-LQA) algorithm. Under regularity conditions, the consistency and Oracle property of the proposed estimator are established. Simulation studies are conducted under two types of penalty parameter selection criteria: the Schwarz Information Criterion (SIC) and the Generalized Approximate Cross-Validation (GACV). We evaluate the performance of the proposed method under different levels of correlation among explanatory variables and different covariance structures of random effects. Comparisons are also carried out with the non-penalized model, the single-penalized model, and the double LASSO-penalized model. The results demonstrate that the proposed double SCAD penalty method performs better than the other three methods in terms of important variable selection and coefficient estimation, and is especially effective for sparse models.

1. Introduction

With the transformation of modern production and manufacturing towards intelligence and refinement, the accurate modeling of various count-type quality indicators in the production process (such as the number of defects, failure times, and non-conforming parts) is a core demand to ensure production stability and reduce costs. These indicators are inherently non-negative integers, a defining feature of count data. Traditional linear models, such as ordinary least squares (OLS) regression, are not suitable for this type of data because they assume the response variable follows a continuous, normal distribution, which directly contradicts the discrete, non-negative nature of count outcomes.
Meanwhile, the manufacturing process is continuous, and the observation data of the same production line, equipment, or batch have longitudinal correlation in the time dimension. Traditional linear models assume independence among observations, and ignoring this inherent correlation can lead to biased parameter estimates and undermine the reliability of the model. Common modeling approaches for such longitudinal count data include marginal models [1] and mixed-effects models [2,3], among which the mixed-effects Poisson count model with random effects is the most widely used. The primary reason is that longitudinal count sample data are non-independent, with a certain degree of correlation among observations from the same unit. Mixed-effects models allow for such data correlation. They take into account the practical scenario that the production units selected for data collection are often randomly sampled, where each unit is associated with distinct raw materials and operators [4,5]. Since the research objective focuses on the overall parameter characteristics of the population rather than individual units, incorporating unit-specific individual effects as random effects into the model enhances modeling accuracy and flexibility [6].
On the other hand, practical modeling often confronts the challenge of high-dimensional data, where the number of variables is close to or even exceeds the sample size [7]. The curse of dimensionality makes variable selection indispensable, and various penalization methods enable concurrent parameter estimation and variable selection [8]. The LASSO (Least Absolute Shrinkage and Selection Operator) penalty proposed by Tibshirani [9] has been extremely widely adopted. It not only estimates the regression parameters in the model but also shrinks the coefficients of irrelevant variables to zero, thereby achieving the goal of variable selection. However, the LASSO penalty tends to introduce estimation bias for large coefficients. To address this limitation, Fan and Li [10] proposed a novel SCAD (Smoothly Clipped Absolute Deviation) penalty based on the LASSO method. This non-convex penalization approach inherits the advantages of the LASSO penalty while possessing favorable statistical properties: it yields sparse solutions, and the resulting estimates satisfy asymptotic normality.
Since its introduction, the SCAD penalty has been widely applied to various models, such as quantile regression [11] and logistic regression [12,13]. For Poisson regression models, Yan and Chen [14] adopted the SCAD penalty for variable selection, while Buu [15] further extended its application to zero-inflated Poisson regression models for the same purpose. Li [16] employed the SCAD penalty to perform variable selection for high-dimensional random-effects linear regression models and high-dimensional random-effects quantile regression models; Li and He [17] also implemented the SCAD penalty for variable selection in high-dimensional random-effects linear regression models. Liu [18] constructed a SCAD-regularized heterogeneous autoregressive model for volatility forecasting. Liu and Chen [19] proposed an analytical framework integrating self-supervised learning with SCAD-Net regularization. Liu, Wang, and Wu [20] conducted research on SCAD-type variable selection and estimation in censored regression models. Zhang and Dong [21] carried out a comparative study on the performance of six variable selection methods—LASSO, Ridge Regression, Elastic Net, SCAD, Bridge, and Adaptive Bridge—in Poisson regression, and verified that the SCAD method exhibits the best performance in terms of model goodness-of-fit and prediction accuracy, making it suitable for large-sample and high-correlation datasets.
Moreover, all the aforementioned applications of the SCAD penalty only target the fixed-effects component of the respective models [22]. For mixed-effects Poisson regression models, performing variable selection solely on fixed effects while ignoring random effects will introduce bias into the estimation of fixed effects; conversely, incorporating an excessive number of irrelevant random effects may lead to the singularity of the random-effects covariance matrix. To address this problem, Ibrahim et al. [23] once imposed the SCAD penalty on both fixed and random effects of generalized linear mixed-effects models for variable selection, with parameters solved via the LLA (Local Linear Approximation) algorithm. However, their method penalizes the variances of random effects and requires these variances to follow a normal distribution. This penalization scheme has a drawback: when the variance of a certain random effect is shrunk to zero, the estimated values of all individuals under this random effect are forced to zero, which neglects the potential impact of non-zero random effects for some individuals.
Against this backdrop, this paper directly penalizes the random effects and proposes a double SCAD-penalized mixed-effects Poisson count regression model that enables simultaneous selection of both fixed and random effects. This penalization framework imposes no constraints on the variances of random effects. Furthermore, the parameter estimation algorithm in this paper is based on the LQA (Local Quadratic Approximation) algorithm [10], which yields more accurate estimates compared with the LLA algorithm.
The model proposed in this paper differs fundamentally from the method of Ibrahim et al. [23] in terms of penalization target, regularization logic, and hierarchical modeling strategy. Although Ibrahim et al. [23] applied the SCAD penalty to both fixed and random effects in generalized linear mixed-effects models; their penalization target was the variance of random effects. The corresponding regularization logic was to conduct variable selection by shrinking the variances; when a variance was shrunk to zero, the estimates of all individuals under that random effect were forced to zero, representing a global selection scheme that failed to capture individual-level heterogeneity. Their hierarchical modeling required the random effects to follow a normal distribution and adopted the LLA (Local Linear Approximation) algorithm for parameter estimation.
In contrast, this paper directly takes fixed-effect coefficients and random-effect coefficients as the penalization targets and uses the double SCAD penalty to achieve simultaneous selection of both types of effects. The regularization process does not rely on variance shrinkage, allowing accurate identification and retention of meaningful individual random effects. In terms of hierarchical modeling, no distributional constraint is imposed on the variances of random effects, and the more accurate LQA (Local Quadratic Approximation) algorithm is employed for parameter estimation.
When applied to cigarette production data from multiple units, this model simultaneously penalizes the numerous unknown random effects of individual units and the fixed effects of various process parameters. It not only identifies a small set of key process indicators affecting the number of short positions but also accurately estimates their impact weights, thereby providing valuable insights for the further adjustment and optimization of process parameters. In fact, similar double-penalization ideas have been explored in existing literature; for instance, Luo et al. [24] investigated the issue of double regularization penalties for mixed-effects models and proved the convergence of the iterative algorithm.
Most existing penalization strategies for Poisson mixed-effects count regression models are imposed on the variance components of random effects. Such approaches only achieve global shrinkage and overall selection of random effects, and suffer from two critical limitations:
First, they cannot perform independent and precise variable selection on the individual random-effect coefficients α i . In high-dimensional settings, retaining redundant individual effects or excluding key heterogeneous effects tends to introduce substantial fitting bias. Second, when the variance of a random effect is shrunk to zero, the estimates for all individuals under that effect are forced to zero, which ignores the potential heterogeneity where some individuals may still exhibit non-zero random effects. This is incompatible with the characteristic of count data in practice: the overall trend is consistent while individuals show specific heterogeneity. Meanwhile, such methods require the variances of random effects to follow a normal distribution, which further restricts their applicability in modeling high-dimensional and non-normal count data.
In contrast, penalizing the random-effect coefficients α i directly can overcome the above limitations. It enables simultaneous and refined variable selection for both fixed effects and random effects. This approach effectively removes irrelevant individual random effects while preserving meaningful individual heterogeneity. Moreover, it imposes no distributional constraints on the variances of random effects, making it better suited for modeling high-dimensional count-type quality indicators in manufacturing, biostatistics, and other fields. It thus fills the research gap of traditional penalization strategies in accurately capturing individual heterogeneity and adapting to high-dimensional scenarios.
Section 2 presents the general form of the mixed-effects Poisson count regression model. Based on the likelihood function, a novel double-penalization method is proposed by simultaneously imposing the SCAD penalty on both fixed-effects and random-effects regression coefficients. This section also introduces the iterative local quadratic approximation algorithm for model parameter estimation and discusses the selection criteria for optimal penalty parameters.
Section 3 provides rigorous proofs of the relevant theoretical properties of the proposed method.
Section 4 conducts comparisons between the proposed method and traditional methods via Monte Carlo simulations, evaluating their performance in variable selection and regression coefficient estimation under various scenarios.
Section 5 demonstrates the practical application scenarios of the proposed method, illustrating its applicability in real-world contexts.
Finally, Section 6 summarizes the main findings of the paper and offers concluding remarks.

2. Double Penalty Poisson Counting Regression Model

2.1. Double Penalty Poisson Counting Regression Model with SCAD Penalty

2.1.1. Model Establishment

To ensure compatibility of matrix operations in the model and facilitate verification and reproducibility, this section first standardizes all core variables, the meanings of subscripts, and dimensions involved in the model. The specific definitions are as follows:
Explanation of Subscripts: i = 1 , 2 , , n denotes the observation unit, reflecting individual heterogeneity; j = 1 , 2 , , m denotes the j - t h repeated observation within the i - t h observation unit, reflecting the time/batch correlation of longitudinal data.
Definition of Core Variables and Dimensions:
Response Variable: y i j represents the count-type response variable of the j - t h observation for the i - t h observation unit, and its matrix form is y = y 11 , y 12 , , y 1 m , y 21 , , y n m T .
Fixed Effect-related Variables: x i j represents the fixed-effect explanatory variable vector of the   j - t h observation for the i - t h observation unit, where the subscript i j corresponds strictly to the response variable y i j . Its dimension is defined as x i j R k , where k is the total number of fixed-effect explanatory variables; β is the fixed-effect regression coefficient vector, whose dimension matches that of x i j , i.e., β R k , ensuring the compatibility of the linear operation x i j T β .
Random Effect-related Variables: z i j represents the random-effect explanatory variable vector (covariate vector) of the j - t h observation for the i - t h observation unit, where the subscript i j is consistent with those of y i j and x i j . Its dimension is defined as z i j R p , where p is the total number of random-effect explanatory variables (covariates); α i is the random-effect coefficient vector of the i - t h observation unit, whose dimension matches that of z i j , i.e., α i R p , ensuring the compatibility of the linear operation z i j T α i (scalar output). Its matrix form is α = α 1 T , α 2 T , , α n T T R n p × 1 .
Supplementary Illustration: X is the fixed-effect explanatory variable matrix, whose matrix form is X = x 11 T , x 12 T , , x 1 m T , x 21 T , , x n m T T R N × k . It matches the fixed-effect coefficient vector β R k , ensuring the compatibility of the matrix operation X β ; Z is the random-effect explanatory variable (covariate) matrix, whose matrix form is Z = d i a g Z 1 , Z 2 , , Z n , where Z i = z i 1 , z i 2 , , z i m T R m × p and the overall dimension is Z R N × n p . It matches the random-effect coefficient vector α R n p × 1 , ensuring the compatibility of the matrix operation Z α . N denotes the total sample size, which is equal to nm; that is, there are n individuals with m observations per individual.
In the general Poisson counting regression model, assuming that the response variable Y obeys the Poisson distribution,
P Y = y = λ y e λ y ! = exp y log λ λ log y ! , y = 0 , 1 , 2 , ,
where λ > 0 , E Y = λ , using the family standard forms of the generalized linear model exponential distribution, there are
θ = log λ , b θ = λ = e θ , a ϕ = 1 , c y , ϕ = log y ! g u = g E Y | X = l o g μ = l o g b θ = l o g e θ = θ = X β
By this formulation, the general Poisson counting regression model is given by the following:
E Y | X = e X β
where X is an explanatory variable, β is the corresponding coefficient.
Assume that there are n subjects, and each subject is observed repeatedly. For the i - t h subject ( i = 1 , , n ) , let y i j be the response variable for the j - t h observation ( j = 1 , , m ) . Such clustered or longitudinal data arise frequently in applied sciences. Then generalize model (3) to a Poisson counting regression model with mixed effects:
E y ij | x ij , z i j = e x i j T β + z i j T α i
where x i j is a k × 1 vector of explanatory variables, β is the corresponding vector of x i j with an unknown fixed effect coefficient, z i j is the p × 1 vector of covariates, α i is the corresponding vector of z i j with an unknown random effect coefficient. The probability density function of the response variable is obtained from Equation (1):
p y i j | x i j , z i j = E y i j | x i j , z i j y i · e E y i j | x i j , z i j y i !
From this, the log-likelihood function is simplified, given by the following Equation:
l β , α = i = 1 n j = 1 m y i j x i j T β + z i j T α i e x i j T β + z i j T α i
Considering that the fixed-effect coefficient vector β and random-effect coefficient vector α i are unknown in the above likelihood function, this paper applies the SCAD penalty to both fixed effects and random effects. This strategy takes full advantage of the desirable estimation properties of the SCAD penalty, enabling the model to produce unbiased sparse solutions. It can not only select important explanatory variables but also prevent the model from overfitting caused by including excessive random effects. This leads to the following minimized double SCAD-penalized likelihood function:
L β , α = min { i = 1 n j = 1 m y i j x i j T β + z i j T α i + e x i j T β + z i j T α i + l = 0 k p λ β l + i = 1 n t = 1 p p λ α i t } .
The equation above is also equivalent to the following:
L β , α = m a x i = 1 n j = 1 m y i j x i j T β + z i j T α i e x i j T β + z i j T α i l = 0 k p λ β l i = 1 n t = 1 p p λ α i t
where the SCAD penalty function takes the specific form:
p λ θ i = λ θ i , 0 θ i λ θ i 2 2 a λ θ i + λ 2 2 a 1 , λ < θ i a λ a + 1 λ 2 2 , θ i > a λ
where λ > 0 is penalty parameter, a is constant, and the recommended value is usually taken as 3.7 [10].

2.1.2. Statistical Interpretation and Methodological Comparison

(1)
Statistical Interpretation of the Double-Penalized Likelihood
The double-penalized likelihood function in Equation (7) enables simultaneous shrinkage estimation of both fixed effects β l , l = 1 , , k   and individual random-effect realizations αit. Unlike traditional variance-component penalization, which targets the covariance structure of random effects, our approach directly penalizes the coefficient values of α i t . This has a clear interpretation in terms of hierarchical borrowing strength:
For random-effect coefficients that are statistically indistinguishable from zero, the SCAD penalty will shrink them to zero, effectively removing redundant individual-level variation from the model.
For coefficients that represent meaningful heterogeneity, the penalty is milder, allowing the model to retain and exploit within-cluster information while avoiding over-borrowing of between-cluster information.
This local, coefficient-wise shrinkage controls the effective model complexity more precisely, which is critical in high-dimensional settings where many random-effect terms may be spurious or irrelevant.
(2)
Comparison with Variance-Parameter Penalization
From Table 1, it can be seen that variance-component penalization is a global operation: when a variance is shrunk to zero, it eliminates all individual-level variation for that effect. In contrast, our double SCAD penalty operates at the level of individual realizations, allowing the model to retain sparse and meaningful heterogeneity while discarding noise. This is particularly valuable in industrial applications, where only a subset of production units may exhibit anomalous behavior, while the rest follow the global trend.

2.2. Iterative Local Quadratic Approximation Algorithm (DSCAD-LQA) for Model Unknown Parameter Estimation

The parameter estimation algorithm for Equation (8) is mainly based on the local quadratic approximation algorithm (LQA) proposed by Fan and Li (2001) [10]. Different from the conventional LQA, Equation (8) in this paper involves two penalty functions. Therefore, we develop an improved iterative algorithm based on LQA, referred to as the DSCAD-LQA (Double Smoothly Clipped Absolute Deviation–Local Quadratic Approximation). The specific steps of the algorithm are as follows.
Step 1: The initial values β 0 , α 0 of the fixed effect coefficients β and the random effect coefficients α are first solved by the log-likelihood function without penalty (6) using the Newton iterative algorithm in the following steps (1) to (3), and updating the likelihood values during each iteration. The exact process of iteration is as follows:
(1)
At the first s + 1 iteration, β value, a k × 1 vector, is solved by the function: β l s + 1 = β l s 2 l β , α s β l 2 1 l β , α s β l β = β l s , l = 1 , , k . Substitute the β value to update the likelihood value.
(2)
Then, α value, a p × n vector, is solved by the function: α i t s + 1 = α i t s 2 l β s + 1 , α α i t 2 1 l β s + 1 , α α i t α = α i t s , t = 1 , , p , i = 1 , , n . Substitute the α value similarly to update the likelihood value.
(3)
The initial value β 0 , α 0 can be obtained by solving (1) and (2) iteratively until it stops at | | l β s + 1 , α s + 1 l β s , α s l β s , α s | | < ε , where l β , α is a log-likelihood function without penalty (6).
Step 2: Subjecting the penalty functions p λ β l and p λ α i t to a second-order Taylor expansion at non-zero:
p λ β l p λ β l 0 + 1 2 p λ β l 0 / β l 0 β l 2 β l 0 2 , β l β l 0
p λ α i t p λ α i t 0 + 1 2 p λ α i t 0 / α i t 0 α i t 2 α i t 0 2 , α i t α i t 0
Calculate the derivative
p λ β l p λ β l 0 / β l 0 β l , β l β l 0
p λ α i t p λ α i t 0 / α i t 0 α i t , α i t α i t 0
Step 3: Substitute the initial values β 0 , α 0 obtained from Step 1 into the objective function equation with double SCAD penalty (8), and find the initial objective value L M = L β 0 , α 0 . The final results β b e s t α b e s t are also obtained by using Newton iterative method on the basis of the initial values β 0 α 0 solved by the following alternating iterative solutions. The exact process of iteration is as follows:
(1)
At the first s + 1 iteration, β value, a k × 1 vector, is solved by the Function (14). Substitute the β value to update the target value L M = L β s + 1 , α s .
β l s + 1 = β l s 2 l β , α s β l 2 + p λ β β 1 l β , α s β l + p λ β β β β = β l s , l = 1 , , k
(2)
Then the α value, a p × n vector, is obtained by solving Equation (15) in turn, and substituted to update the target value L M = L β s + 1 , α s + 1 .
α i t s + 1 = α i t s 2 l β s + 1 , α α i t 2 + p λ α α 1 l β s + 1 , α α i t + p λ α α α α = α i t s , t = 1 , , p , i = 1 , , n
(3)
The final result β b e s t , α b e s t can be obtained by solving (1) and (2) iteratively until it stops at | | l β s + 1 , α s + 1 l β s , α s l β s , α s | | < ε , where l β , α is the target Function (8).

Analysis of Computational Properties and Numerical Stability of the Algorithm

The iterative process of the aforementioned DSCAD-LQA relies fundamentally on the Newton iterative update mechanism, and the iterative formula for fixed effects can be further specified as follows:
β l s + 1 = β l s 2 l β , α s β l 2 + p λ β β 1 l β , α s β l + p λ β β β β = β l s , l = 1 , , k
where l β and 2 l β 2 denote the gradient and Hessian of the unpenalized log-likelihood, respectively, and p λ ( β ) β is the derivative of the double SCAD penalty.
(1)
The Relationship Between Algorithm Convergence and Matrix Condition Number
The convergence of the DSCAD-LQA mainly depends on the positive definiteness and the condition number κ ( ) of the modified Hessian matrix:
2 l β , α s β 2 + p λ β β .
The condition number of the modified Hessian matrix is defined as the ratio of its maximum eigenvalue to its minimum eigenvalue, i.e.,
κ ( 2 l β , α s β 2 + p λ β β ) = λ m a x λ m i n .
Its core influence laws are as follows:
When k or p n is small (low-dimensional scenario), the modified Hessian matrix possesses satisfactory positive definiteness, with the condition number κ ( 1 , 10 2 ) . The Newton iteration converges rapidly. In this paper, the convergence threshold is set as | | β s + 1 β s | | 2 < 10 6 , which can usually be satisfied within 5–10 iterations in low-dimensional scenarios.
As k or p n increases (high-dimensional scenario), the modified Hessian matrix tends to become nearly singular, with the condition number κ > 10 3 . Direct matrix inversion in this case leads to oscillatory iterations and slower convergence.
To address this issue, the Tikhonov regularization method is adopted in this paper to modify the Hessian matrix, resulting in the regularized matrix:
2 l β , α s β 2 + p λ β β + μ I ,
where μ is the regularization parameter, and I denotes the identity matrix. This modification can constrain the condition number below 10 2 in high-dimensional scenarios, guaranteeing stable convergence of the iterative procedure. To improve the numerical stability in high-dimensional scenarios, Tikhonov regularization is adopted in this paper with a regularization parameter μ = 10 4 . This value is determined by the L-curve method, which can effectively improve the condition number of the Hessian matrix while ensuring the consistency of the estimates.
(2)
Algorithm Scalability Analysis
The scalability of the algorithm mainly reflects the variation in computational efficiency as the sample size n, the dimension of fixed effects k, and the dimension of random effects p n increase. Based on the core computational steps of the algorithm, including gradient evaluation, construction, and inversion of the modified Hessian matrix, the scalability analysis is presented as follows:
The computational complexity of the gradient vector l β , α s β + p λ , β β β is O ( n k + n p n ) , which mainly depends on the linear combination of the sample size and variable dimensions. When n or k is doubled, the computational cost grows approximately linearly.
The construction complexity of the modified Hessian matrix is O ( n k 2 + n p n 2 ) , and its inversion complexity is O ( k 3 ) , which forms the main computational bottleneck of the algorithm.
For high-dimensional scenarios ( k > 100   o r   p n > 500 ) , this paper employs the conjugate gradient method in place of direct matrix inversion, reducing the inversion complexity to O ( k 2 ) . This strategy significantly improves the scalability of the algorithm for high-dimensional data and guarantees that the algorithm can still finish iterations efficiently even when k increases to 500.
(3)
Computational Complexity and Numerical Stability Guarantees
The overall computational complexity of the DSCAD-LQA is O ( T ( n k 2 + n p n 2 ) ) , where T denotes the number of iterations. In all numerical experiments, the number of iterations T is controlled within the range [5,20], and neither divergence nor premature convergence is observed. This complexity is comparable to that of existing penalized Poisson mixed-effects regression algorithms (e.g., the LASSO-Poisson algorithm). However, the numerical stability and reproducibility are further enhanced by the following two improvements:
The penalty parameter λ is selected via five-fold cross-validation, with the optimal value determined by minimizing the prediction error, thereby avoiding non-convergence caused by inappropriate parameter selection.
The initial values β 0 and α 0 are chosen as the Newton-type estimates obtained from the unpenalized log-likelihood Function (6), rather than random initialization. This ensures the consistency and reproducibility of the iterative process, such that different researchers can obtain exactly the same convergent results based on the same dataset.

2.3. The Selection of Penalty Parameters

The selection of penalty parameters is extremely important for models involving penalty functions, as the magnitude of penalty parameters affects the final results of model variable selection and parameter estimation. When λ = 0 , the penalty function does not penalize the parameters in the model; however, when λ = , the parameters may be over-penalized, resulting in all parameters being estimated as zero. Since there are two penalty parameters ( λ β and λ α ) in the double SCAD-penalized mixed-effects Poisson count regression model, and considering the computational complexity of the algorithm, the two penalty parameters are treated uniformly based on the overall likelihood value and selected via the SIC (Schwarz Information Criterion) (1982) [25] or GACV (Generalized Approximate Cross-Validation Criterion) (2006) [26]. The two criteria were defined as follows:
S I C λ θ = L n S M / N + L n N 2 N M
G A C V λ θ = S M N M
where θ is the parameter β or α , S M = i = 1 n j = 1 m y i j x i j T β + z i j T α i + e x i j T β + z i j T α i is the value after updating the parameters and substituting them into the maximum likelihood function without penalty and minimizing them. N = n m is the total number of samples, M is the number of fixed-effect regression coefficient vectors β or random effect regression coefficient vectors α which is not equal to zero. λ , which minimizes the values of SIC and GACV, is the value of the optimal penalty parameter.

Remark on Joint Tuning of Penalty Parameters

In this work, we adopt a joint tuning strategy for the fixed-effect penalty parameter λβ and the random-effect penalty parameter λ α , i.e., setting λ β = λ α = λ and selecting the optimal λ via the SIC or GACV criterion. This choice is theoretically justified from the following perspectives:
(1)
Consistency of Model Complexity
The SIC and GACV criteria employed in this paper are constructed based on the total number of non-zero coefficients M to measure model complexity. Joint tuning ensures that M consistently represents the total number of non-zero fixed-effect coefficients β l and non-zero random-effect coefficients α i t , which is fully consistent with the definition of model degrees of freedom in Equations (16) and (17). Separate tuning of λ β an λ α d would introduce ambiguity in the definition of M (requiring separate calculation of degrees of freedom for fixed and random effects), thereby compromising the statistical consistency of the parameter selection procedure.
(2)
Convexity and Convergence Guarantee
When λ β = λ α = λ , the double SCAD penalty function p λ ( ) (Equation (9)) maintains a uniform derivative structure for both the fixed effects β and random effects α. This ensures that the overall objective function (Equation (7)) is convex (or pseudo-convex) over the parameter space, preventing the DSCAD-LQA (Equations (14) and (15)) from being trapped in local optima and guaranteeing convergence to a global solution.
(3)
Computational Feasibility in High Dimensions
In high-dimensional scenarios where the dimension of fixed effects k or the dimension of random effects p n is large, separate tuning of λ β and λ α would lead to a two-dimensional grid search, significantly increasing the computational complexity from O ( L ) to O ( L 2 ) , where L is the number of candidate parameters. Joint tuning reduces the parameter space to one dimension, making the selection procedure computationally feasible while maintaining satisfactory estimation accuracy.
(4)
Balanced Estimation and Stability
A unified penalty strength λ ensures a balanced shrinkage effect on both the fixed effects β and random effects α, avoiding estimation bias caused by over-penalizing one component while under-penalizing the other. Additionally, the uniform penalty structure contributes to the numerical stability of the modified Hessian matrix (e.g., the terms 2 l β , α s β l 2 + p λ β / β in Equations (14) and (15), reducing the risk of ill-conditioning and improving the robustness of the Newton iteration.

3. Theoretical Property Proof

Assuming that the parameters β , α i have true value β 0 , α 0 i , the objective function is as follows:
Q β , α = L β , α m n l = 1 k p λ β β l m i = 1 n t = 1 p p λ α α i t
where β = β 1 , β 2 , , β k T , β 0 = β 01 , β 02 , , β 0 k T , α i = α i 1 , α i 2 , , α i p , α 0 i = α 0 i 1 , α 0 i 2 , , α 0 i p T , Let β = β 1 T , β 2 T T , β 0 = β 10 T , β 20 T T , α i = α 1 i T , α 2 i T T , α 0 i = α 10 i T , α 20 i T T = r e c o r d α 10 T , α 20 T T , where β is a k dimensional vector, β 1 is a s β dimensional vector, β 2 is a k s β dimensional vector, β 10 is a s β dimensional non-zero vector, β 20 is a k s β dimensional zero vector, α i is a p dimensional vector, α 1 i is a s α dimensional vector, α 2 i is a k s α dimensional vector, α 10 is a s α dimensional non-zero vector, α 20 is a k p α dimensional zero vector. The exponential family probability density function is as follows:
f y , θ , ϕ = exp y θ b θ a ϕ + c y , ϕ
For the mixed effects Poisson regression model, note that the probability density function is as follows:
f y , x , z , β , α = record f V , β , α = e x p y x T β + z T α i e x p x T β + z T α i l o g y !
The following conditions are given:
Condition 1: The derivative of f V , β , α is defined as follows:
E l o g   f V , β , α β l = 0 l = 1 , , k
E l o g   f V , β , α α i t = 0 t = 1 , , p
I j l β β , α = E l o g   f V , β , α β j l o g   f V , β , α β l = E 2 l o g   f V , β , α β j β l
I t l α β , α = E l o g   f V , β , α α i t l o g   f V , β , α α i l = E 2 l o g   f V , β , α α i t α i l
Condition 2: The Fisher information matrix I β , α is finite and positive definite at the true parameter values ( β 0 , α 0 i ) . We denote its block submatrices as follows:
Let I 11 β , α = E 2 log   f V , β , α 2 β , I 22 β , α = E 2 log   f V , β , α 2 α i . When evaluated at the true parameters, these submatrices are denoted I 11 β 0 , α 0 i and I 22 β 0 , α 0 i , and the full Fisher information matrix is I β 0 , α 0 i .
Condition 3: For an open interval ω containing the truth values of the parameters β 0 , α 0 i , there exists M so that the third-order derivative is bounded, that is
3 log   f V , β , α β j β l β r M 3 log   f V , β , α α i t α i l α i r M   β , α ω
Condition 4: liminf m n liminf θ 0 + λ θ 1 p λ θ θ > 0 , where θ is the parameter.
Theorem 1 (Compatibility).
Under conditions 1–4, if
max p λ β β 0 l : β 0 l 0 0 , max p λ α α 0 i t : α 0 i t 0 0 , then there exists a local maximum   β , α i  of Formula (21) such that
β β 0 = O p m n 1 2 + a m n α i α 0 i = O p m 1 2 + b m
where   a m n = max p λ β β 0 l : β 0 l 0 ,  b m = max p λ α α 0 i t : α 0 i t 0 , when   m , n :
β β 0 = O p m n 1 2 α i α 0 i = O p m 1 2
Proof. 
Let a m n = m n 1 2 + a m n , b m = m 1 2 + b m . For any given ε > 0 and the existence of a sufficiently large constant C , there is the following:
P sup u = C , v i = C Q β 0 + a m n u , α 0 i + b m v i < Q β 0 , α 0 i 1 ε
where u is a k × 1 vector, v i is a p × 1 vector. According to Equation (23), we know that there exists a local maximum point β ^ , α ^ i within the ball β 0 + a m n u , α 0 i + b m v i , u C ,   v i C when the probability is close to one, so that β ̑ β 0 = O p a m n , α ̑ i α 0 i = O p b m .
Due to p λ θ 0 = 0 , there is the following:
D u , v Q β 0 + a m n u , α 0 i + b m v i Q β 0 , α 0 i L β 0 + a m n u , α 0 i + b m v i L β 0 , α 0 i m n l = 1 s β p λ β β 0 l + a m n u l p λ β β 0 l m i = 1 n t = 1 s α p λ α α 0 i t + b m v i t p λ α α 0 i t
Subjecting it to a second-order Taylor expansion, there is
D u , v L β 0 , α 0 i + a m n L β β 0 , α 0 i u + b m L α β 0 , α 0 i v i + [ 1 2 u T L β β 0 , α 0 i u a m n 2 + 1 2 v i T L α β 0 , α 0 i v b m 2 + u T L β α β 0 , α 0 i v i a m n b m ] 1 + O 1 L β 0 , α 0 i l = 1 s β m n a m n p λ β β 0 l sgn β 0 l u l + m n a m n 2 p λ β β 0 l u l 2 1 + O 1 i = 1 n t = 1 s α m b m p λ α α 0 i t sgn α 0 i t v i t + m b m 2 p λ α α 0 i t v i t 2 1 + O 1 = T 1 + T 2 + T 3 + T 4 + T 5
where L β β 0 , α 0 i is the first-order partial derivative of L β 0 , α 0 i with respect to L β 0 , α 0 i , L α β 0 , α 0 i is the first-order partial derivative of α , L β β 0 , α 0 i is the second-order partial derivative of β , L α β 0 , α 0 i is the second-order partial derivative of α , L β α β 0 , α 0 i is the second-order mixed partial derivative.
For T 1 , T 2 , there are L β β 0 , α 0 i = O p m n and L α β 0 , α 0 i = O p m under condition 1 and condition 2, known that O p m n a m n = O p m n a m n 2 , O p m b m = O p m b m 2 . According to the Cauchy–Schwarz inequality, there are
T 2 = b m L α β 0 , α 0 i v i b m L α β 0 , α 0 i v i = b m m v i = O p m b m 2 v i
T 1 = a m n L β β 0 , α 0 i u a m n L β β 0 , α 0 i u = a m n m n u = O p m n a m n 2 u
For T 3 ,
T 3 = [ 1 2 u T L β β 0 , α 0 i u a m n 2 + 1 2 v i T L α β 0 , α 0 i v b m 2 + u T L β α β 0 , α 0 i v i a m n b m ] 1 + O 1 = 1 2 w T I β 0 , α 0 i w 1 + O 1
where w T = a m n u T , b m v i T .
For T 4 , T 5 , due to max p λ β β 0 l : β 0 l 0 0 , max p λ α α 0 i t : α 0 i t 0 0 , a m n = max p λ β β 0 l : β 0 l 0 , b m = max p λ α α 0 i t : α 0 i t 0 , there are
T 4 = l = 1 s β m n a m n p λ β β 0 l sgn β 0 l u l + m n a m n 2 p λ β β 0 l u l 2 1 + O 1 s β m n a m n a m n u + m n a m n 2 max p λ β β 0 l : β 0 l 0 u 2 = s β m n a m n a m n u
T 5 = i = 1 n t = 1 s α m b m p λ α α 0 i t sgn α 0 i t v i t + m b m 2 p λ α α 0 i t v i t 2 1 + O 1 n s α m b m b m v i + m b m 2 max p λ α α 0 i t : α 0 i t 0 v i 2 = n s α m b m b m v i
When a sufficiently large C is chosen, the value of T 3 is much larger than the value of T 1 , T 2 , T 4 , T 5 under the conditions of u = C , v i = C . That is, the sign of Equation (25) is completely determined by T 3 , and the theorem has been proved. □
Theorem 2
(Nature of Oracle). For parameter  β = β 1 T , β 2 T T ,  α i = α 1 i T , α 2 i T T , we know by Theorem 1 that there exist local maximum points  β ^ = β ^ 1 T , β ^ 2 T T ,  α ^ i = α ^ 1 i T , α ^ 2 i T T . Assuming that  m n 1 2 / λ β 0 ,  m 1 2 / λ α 0 , and providing that the conditions all hold, we have
(a) 
Sparsity:  β ^ 2 = 0 , α ^ 2 i = 0 m , n
(b) 
Asymptotical normality:
m n I 11 β 10 , α 10 + 11 β ^ 1 β 10 + I 11 β 10 , α 10 + 11 1 b 1 N 0 , I 11 β 10 , α 10
m I 22 β 10 , α 10 + 22 α ^ 1 i α 10 + I 22 β 10 , α 10 + 22 1 b 2 N 0 , I 22 β 10 , α 10
where  I 11 β 10 , α 10  is a  s β × s β  matrix,  I 22 β 10 , α 10  is a  s α × s α  matrix,  I 11 β 10 , α 10 ,  I 22 β 10 , α 10  are submatrices of  I β 0 , α 0 i . For  11  ,  22  ,  b 1  ,  b 2  the definitions are as follows:
11 = d i a g p λ β β 01 , , p λ β β 0 s β
22 = d i a g p λ α α 0 i 1 , , p λ α α 0 i s α
b 1 = p λ β β 01 sgn β 01 , , p λ β β 0 s β sgn β 0 s β T
b 2 = p λ α α 0 i 1 sgn α 0 i 1 , , p λ α α 0 i s α sgn α 0 i s α T
Proof. 
For the proof of sparsity, let ε m n = C m n 1 2 , ε m = C m 1 2 ; it suffices to show that when m , n is satisfied, β ^ 1 β 10 = O p m n 1 2 , α ^ 1 i α 10 = O p m 1 2 , with probability tending to one for any given β 1 , α 1 i :
Q β , α β j < 0 , 0 < β j < ε m n Q β , α β j > 0 , ε m n < β j < 0
Q β , α α i t < 0 , 0 < α i t < ε m Q β , α α i t > 0 , ε m < α i t < 0
where j = s β + 1 , , k , t = s α + 1 , , p . For β 2 C m n 1 2 , α 2 i C m 1 2 , applying Taylor’s formula yields the following:
The score function of the objective Q ( β , α ) is given by
Q β , α β j Q β , α α i t = L β , α β j m n p λ β β j s g n β j L β , α α i t m p λ α α i t s g n α i t
Under the given rate conditions, this can be simplified to
Q β , α β j Q β , α α i t = m n λ β λ β 1 p λ β β j sgn β j + O p m n 1 2 / λ β m λ α λ α 1 p λ α α i t sgn α i t + O p m 1 2 / λ α
Given the conditions liminf m n liminf θ 0 + λ θ 1 p λ θ θ > 0 , m n 1 2 λ β 0 , m 1 2 λ α 0 , it is possible that the sign of Q β , α / β j is determined by β j , Q β , α / β j , and the sign of Q β , α / α i t is determined by α i t . Combined with the sign conditions above, this completes the proof of sparsity. For the proof of asymptotic normality, we know from Theorem 1 that there exists a local maximum point β ^ , α ^ i such that
Q β , α β j Q β , α α i t β , α = β ^ 1 α ^ 1 i 0 0 = 0 0
By Slutsky’s theorem and the central limit theorem, it follows that
m n I 11 β 10 , α 10 + 11 β ^ 1 β 10 + I 11 β 10 , α 10 + 11 1 b 1 N 0 , I 11 β 10 , α 10
m I 22 β 10 , α 10 + 22 α ^ 1 i α 10 + I 22 β 10 , α 10 + 22 1 b 2 N 0 , I 22 β 10 , α 10

Discussion on the Practical Implications of Asymptotic Convergence Results

This section elaborates on the convergence rates, regularity conditions, and oracle property presented in the theorems, clarifying their theoretical significance and practical application value.
1.
Practical Interpretation of Convergence Rates
The asymptotic estimation error results in Theorem 1,
| β ^ β 0 | = O p m n 1 / 2 ,
Which indicates that the fixed-effect estimator β ^ achieves the optimal parametric convergence rate of m n .
Here, m n is the total number of observations. This means that the estimation precision of fixed effects improves steadily as the total sample size increases, and the rate is consistent with that of maximum likelihood estimation (MLE) under a fully known model structure, achieving optimal statistical efficiency.
Intuitive Elucidation:
This formula implies that the estimation error of the fixed effect β decreases at a rate of ( m n ) 1 2 as the total sample size m n increases.
For instance, increasing the total sample size from m n = 100 to m n = 400 (i.e., a fourfold increase) would theoretically halve the standard deviation of the estimation error, which is consistent with the efficiency of the classical maximum likelihood estimation (MLE) under a fully known model structure.
This indicates that the most effective approach to improving the estimation precision of the fixed effect is to increase the total sample size m n : both an increase in the number of subjects n and the number of observations per subject m can lead to a linear improvement in estimation precision.
For the random effects α i , the convergence rate is
| α i ^ α 0 i | = O p m n 1 / 2 ,
which shows that the estimation accuracy of random effects primarily depends on the number of observations per subject m. This aligns with the hierarchical nature of random effects. In practice, increasing the number of within-subject observations can more effectively improve the estimation precision of random effects.
Intuitive Elucidation:
The estimation precision of the random effects is determined primarily by the number of repeated observations per subject m, rather than the total sample size m n . This is because the random effects α i characterize individual-level heterogeneity, which requires a sufficient number of within-subject repeated observations for accurate identification.
For example, increasing the number of observations per subject from m = 5 to m = 20 (i.e., a fourfold increase) would theoretically halve the standard deviation of the estimation error for the random effects.
This suggests that in practical applications, if the research focuses on individual-level random effects, priority should be given to increasing the number of observations per subject m, rather than simply increasing the number of subjects n.
2.
Practical Significance of Conditions 1–4
The regularity conditions 1–4 required for the theoretical proofs are standard in generalized mixed-effects models and are generally satisfied in practical applications:
  • Condition 1 ensures that the expectation of the score function at the true parameters ( β 0 , α 0 i ) is zero, which is a fundamental condition for the consistency of MLE.
  • Condition 2 requires the Fisher information matrix I ( β , α ) to be positive definite at the true parameters, guaranteeing model identifiability and unique estimation.
  • Condition 3 bounds the third-order derivatives of the log-likelihood function, which holds for common exponential family models, such as the mixed-effects Poisson counting regression model in this paper.
  • Condition 4 restricts the growth rate of the penalty parameters, ensuring that the penalty terms do not dominate the likelihood function, so that variable selection and parameter estimation can be performed effectively simultaneously.
These conditions are not overly restrictive and are typically naturally satisfied in real data analysis.
Sample Size Design for Practical Studies:
  • For studies focusing on fixed effects: To achieve a target estimation precision (e.g., a standard error of 0.05 for a key fixed effect), researchers can use the convergence rate m n to back-calculate the required total sample size. For example, if the initial estimate of the standard deviation of the outcome is σ = 1, a total sample size of m n = 400 would yield a standard error of approximately 1/400 = 0.05.
  • For studies focusing on random effects: To accurately estimate individual-level random effects, it is recommended to ensure that each subject has at least m 20 repeated observations. This ensures that the estimation error of random effects is sufficiently small to support meaningful individual-level inferences.
3.
Practical Value of the Oracle Property
The oracle property (Theorem 2) proven in this paper has clear practical implications:
  • Sparsity: The algorithm can consistently identify the true zero coefficients in the model ( β ^ 2 = 0 ,   α ^ 2 i = 0 ) , with probability approaching one, achieving simultaneous sparse selection of both fixed and random effects.
  • Practical implication: This means the model can automatically “drop” irrelevant fixed and random effects from the final specification, reducing model complexity and improving interpretability. For example, in a healthcare application, the model can identify which patient characteristics (fixed effects) and which individual-specific deviations (random effects) are truly predictive of the outcome, eliminating noise from the model.
  • Asymptotic normality: The estimators of the non-zero coefficients follow an asymptotic normal distribution with the same asymptotic variance as if the true model structure were known in advance.
  • Practical implication: This allows researchers to construct valid confidence intervals and perform hypothesis tests for the selected non-zero coefficients as if they knew the true model structure beforehand. This eliminates the need for additional corrections for variable selection, making statistical inference straightforward and reliable in large samples.
4.
Practical Implementation Recommendations
  • To ensure stable estimation of fixed effects, the total sample size m n should be increased as much as possible.
  • To improve the estimation precision of random effects, the number of within-subject observations m should be appropriately increased.
  • The penalty parameter λ is recommended to be selected via SIC or GACV, which automatically satisfies Condition 4 and guarantees the theoretical properties.
  • After variable selection, statistical inference can be directly performed based on the asymptotic normality results without additional corrections.
5.
Summary of Practical Takeaways
  • Fixed effects precision is driven by the total sample size m n . The larger the total sample size m n , the more precise the estimates of the fixed effects, with the convergence rate achieving the optimal level of the m n order.
  • Random effects precision is driven by within-subject observations m. The greater the number of repeated observations per subject, the more accurate the estimation of individual heterogeneity.
  • Model selection: The oracle property ensures that irrelevant effects are automatically excluded, and the remaining effects can be used for valid statistical inference without extra adjustments.
  • Study design: Prioritize m n for fixed effects, prioritize m for random effects, and use information criteria (SIC/GACV) to select penalty parameters λ .

4. Comparative Study of Monte Carlo Simulation

Monte Carlo simulation data were generated as follows: λ i j is first obtained through λ i j = E y i j | x i j , z i j = e β 0 + x i j T β + z i j T α i , and then the value of y i j is generated through the Poisson distribution of P ( λ i j ) , where N is the total sample size, β 0 = 0 , x i j T = x i j 1 , x i j 2 , , x i j 8 i = 1 , 2 , , n ; j = 1 , 2 , , n i . The explanatory variable X 1 , X 2 , X 8 follows a standard normal distribution, and for any two explanatory variables X l and X k , the correlation coefficient between them is ρ l k . In order to estimate the accuracy of model parameter estimation, the median of mean squared error (MRME) is used as an evaluation metric in this paper. Specifically, it is defined as the median mean squared error (MSE) obtained from multiple simulations. This indicator avoids the effect of maximum values and minimal values in the simulation results, where the mean squared error (MSE) is calculated by the following:
M S E = 1 K s = 1 K β ^ s β T β ^ s β p
where β ^ s is the estimated value of β at the first s simulation; in the following simulations, the number of simulations K is repeated 100 times.
To provide a more comprehensive assessment of the finite-sample performance of our estimator, we decompose the mean squared error (MSE) into two fundamental components: squared bias and variance. This decomposition allows us to distinguish between the systematic error of the estimator and its variability across simulations.
For a fixed effect coefficient β l , the bias is defined as the average difference between the estimated value and the true value across all simulations:
B i a s β ^ l = 1 k s = 1 K β ^ l s β l
where K = 100 is the number of simulations, β ^ l s is the estimate of β l in the s - t h simulation, and β l is the true value.
The variance of the estimator β ^ l measures its dispersion around its own mean:
V a r β ^ l = 1 K 1 s = 1 K β ^ l s β ^ ¯ l 2
where β ^ ¯ l = 1 k s = 1 K β ^ l s is the average estimate of β j .
The relationship between these quantities is given by the MSE decomposition:
M S E β ^ l = B i a s β ^ l 2 + V a r β ^ l
For the overall model performance, we report the median bias and median variance across all fixed effect coefficients, which are robust to outliers and align with our use of MRME.
To evaluate variable selection performance, we use two key metrics derived from the confusion matrix:
Correct Selection Rate (CSR)—the proportion of truly non-zero coefficients β l 0 that are correctly identified as non-zero, i.e.,
CSR = TP/total number of true non-zero coefficients.
Correct Elimination Rate (CER)—the proportion of truly zero coefficients β l = 0 that are correctly identified as zero, i.e.,
CER = TN/total number of true zero coefficients.
Here, an estimated coefficient β ^ l is treated as zero if β ^ l < 10 6 . CSR reflects the accuracy of selecting significant variables, while CER reflects the accuracy of excluding irrelevant variables.
1.
Simulation 1—Impact of Multicollinearity
In order to investigate whether the magnitude of the correlation coefficient ρ of the independent variables affects the accuracy of the method estimation and variable selection, we evaluated three cases— ρ = 0.25 , ρ = 0.5 , ρ = 0.75 —as a way to simulate the comparison. When generating the simulation data, let n = 30 , m = 10 ; the fixed effect coefficient is assigned to β = 3 ,   1.5 , 0 ,   0 ,   2 ,   0 ,   0 ,   0 T , the random effect coefficient is assigned to α i = α i 0 , α i 1 , α i 2 T ~ N 3 0 , D , D = d i a g 1 , 1 , 0 , z i j T = 1 , x i j 1 , x i j 2 . Set the threshold value ε = 1 0 6 when stopping the iterations. The double SCAD penalty method proposed in this paper is simulated under two criteria, SIC and GACV, and is noted as DSCAD-SIC and DSCAD-GACV, respectively. The simulation was repeated 100 times, and the specific results obtained from the simulation are shown in Table 2.
As can be seen from Table 2, under the specific simulation settings of this study (i.e., n = 30 , m = 10 , true fixed effects β = 3 , 1.5 , 0 , 0 , 2 , 0 , 0 , 0 T , and 100 repetitions), the estimates under the GACV criterion consistently yield smaller MRME values, identical CSR values, and higher CER values compared to the SIC across all collinearity levels. This indicates that, within the current experimental conditions, the GACV criterion outperforms the SIC in terms of estimation accuracy and the exclusion of irrelevant variables, while both criteria achieve equally high accuracy in selecting significant variables.
It is observed that the accuracy of parameter estimation and variable selection decreases gradually as the correlation coefficient increases, where the value of MRME becomes larger and larger, while the value of CSR and CER becomes smaller and smaller. It can be obviously seen that, compared to the two cases with correlation coefficients of ρ = 0.25 and ρ = 0.5 , the value of MRME becomes significantly larger when ρ = 0.75 . The value increases by 38 times and 31 times under the SIC and by 39 times and 29 times under the GACV criterion, respectively. At a correlation coefficient ρ = 0.5 , the value of MRME increased by 0.21 times and 0.32 times in both SIC and GACV criteria, respectively, compared to the case of ρ = 0.25 .
From the perspective of the bias–variance trade-off, as the correlation coefficient ρ increases, the degree of multicollinearity intensifies. Our detailed decomposition shows that both bias and variance rise substantially, although their relative contributions to the estimation error differ across criteria.
Under the GACV criterion, the variance of parameter estimates remains the dominant component of MSE, accounting for approximately 70% of the total error across all ρ levels. The contribution of squared bias is stable at around 30%, indicating that although both components increase, the rise in MRME is still mainly driven by variance inflation, which is consistent with the typical behavior of parameter estimation under high-dimensional collinear settings.
In contrast, under the SIC, the contribution of squared bias to MSE (approximately 60%) exceeds that of variance. This implies that the increase in MRME is a joint result of both bias and variance, with bias playing a more prominent role. Notably, compared with the DSCAD-SIC method, the DSCAD-GACV method consistently achieves lower absolute bias and a more favorable bias–variance balance, especially under high multicollinearity (ρ = 0.75), where it effectively alleviates the systematic error introduced by collinear predictors.
For the CSR (Correct Selection Rate), under both criteria, the CSR is 1.00 when the correlation coefficient is ρ = 0.25 and ρ = 0.5, indicating that all significant variables are correctly included in the model in these cases. In contrast, the CSR drops to 0.90 for both criteria when ρ = 0.75, which indicates that there are cases where not all significant variables are selected in the 100 simulations. This shows that the accuracy of variable selection is affected when the correlation coefficient of the variables is comparatively large. Notably, even at the high correlation level of ρ = 0.75, the CSR remains at a satisfactory level of 0.90, demonstrating the robustness of the proposed method.
For the CER (Correct Elimination Rate), regardless of the criterion, the CER in case ρ = 0.75 is around 0.40, while for both cases ρ = 0.25 and ρ = 0.5, the CER is above 0.80. It can be seen that the proportion of unwanted variables correctly excluded from the model in the ρ = 0.75 case is only about half of that in the other two cases. Moreover, the CER when ρ = 0.5 is also lower than that when ρ = 0.25. Overall, model estimation and variable selection perform better under the GACV criterion, and when the correlation coefficient is comparatively large, the accuracy of parameter estimation and selection becomes worse.
2.
Simulation 2—Impact of Random Effects
In order to investigate whether the size of the random-effect covariance D affects the accuracy of the estimation of the coefficients of the variables by the method proposed in this paper, three random effects are set up, that is, the random effect coefficients α i = α i 0 , α i 1 , α i 2 T ~ N 3 0 , D , z i j T = 1 , x i j 1 , x i j 2 , and for the covariance, the following three cases, D 1 = d i a g 1 , 1 , 0 , D 2 = d i a g 2 , 2 , 0 , D 3 = d i a g 3 , 3 , 0 , are taken for simulation, respectively. When simulated data are generated for different situations, controlling that n = 30 , m = 10 , ρ = 0.5 , and the fixed effect coefficient is β = 3 , 1.5 , 0 , 0 , 2 , 0 , 0 , 0 T , the threshold ε = 10 6 is set when the iteration is stopped. The results obtained from 100 simulations under both SIC and GACV criteria are shown in Figure 1.
From Figure 1, it is known that the value of MRME is increasing, while the values of CSR (Correct Selection Rate) and CER (Correct Elimination Rate) are decreasing as the value of the random-effect covariance increases under either criterion. This indicates that the magnitude of the random-effect covariance has a notable impact on the accuracy of both parameter estimation and variable selection.
The bias–variance analysis reveals that an increase in the random-effect covariance matrix D amplifies both the variance fluctuation and the systematic bias of the parameter estimates. While variance remains the dominant component of the mean squared error (MSE) across all settings, ranging between 2.011 and 2.491, the absolute bias also exhibits a substantial increase. Specifically, as D rises from D1 to D3, the absolute bias increases by 61% (from 0.376 to 0.606) under the SIC, and by 102% (from 0.209 to 0.423) under the GACV criterion. Although the relative increase in absolute bias is more pronounced under GACV than under SIC, the absolute bias associated with GACV remains consistently lower across all scenarios, indicating its superior estimation accuracy. These results demonstrate that the deterioration in estimation accuracy is driven jointly by elevated estimation uncertainty (variance) and a notable increase in systematic bias. This finding is consistent with the theoretical intuition that stronger random effects give rise to greater unobserved heterogeneity, which is reflected not only in higher estimation variability but also in more substantial systematic deviations from the true parameter values.
Comparing with the case where the covariance is D 1 , when the covariance is D 2 , the value of MRME increases by 168% and 170% under SIC and GACV criteria, respectively. The CSR value is one when the covariance is D 1 ; the CSR value is not one when the covariance is D 2 , which indicates that there is a situation where not all the needed variables are included in the model when the covariance is D 2 . For the CER value, it was reduced by 15.1% and 15.2% compared to the covariance D 1 under both criteria.
The MRME line in Figure 1 fluctuates significantly with the value of the covariance D 3 increasing by 402% and 361% compared with the covariance D 1 for both criteria. The CSR value is less than 1 when the covariance is D 3 , and the CER value decreases by 33.0% and 33.89% under both criteria compared with the covariance D 1 .
Comparing with the D 2 covariance case, when the D 3 covariance resulted in the MRME value increasing by 88%, and the CER value decreasing by 21.1% under the SIC; for the CSR value, the value is less than one in both cases. Under the GACV criterion, the MRME increased by 71% and the CER decreased by 22.1% when the covariance structure shifted from D 2 to D 3 . Again, the CSR value was less than 1.
Within the range of random-effect covariance values considered in this simulation, it is observed that smaller covariance values are associated with better estimation and variable selection performance, while larger covariance values lead to degraded accuracy.
Under the specific simulation conditions of this study (i.e., n = 30 , m = 10 ,   ρ = 0.5 , and random-effect covariance D { D 1 , D 2 , D 3 } ), the proposed method under the GACV criterion shows slightly lower MRME and slightly stronger performance in selecting significant and excluding non-significant variables compared to the SIC. This finding is specific to the range of random-effect covariance values considered in this simulation.
3.
Simulation 3—Comparison with other methods
Based on the simulation above, it is clear that in this paper, the method under the GACV criterion is better estimated. The GACV criterion is used in this paper when different methods are used for comparison with the methods involving the selection of penalty parameters. In this paper, we compare the following methods: the Poisson regression model without penalty (denoted as P), the single-penalty mixed-effects Poisson regression model with SCAD penalty for fixed effect only under the GACV criterion (denoted as P-H-GACV), the model with double LASSO penalty under the GACV criterion (denoted as DLASSO-GACV), and the mixed- effects Poisson regression model with double SCAD penalty under the GACV criterion (denoted as DSCAD-GACV). Similarly, in generating the simulated data, let n = 30 , m = 10 , ρ = 0.5 ; the random effect coefficient is α i = α i 0 , α i 1 , α i 2 T ~ N 3 0 , D , D = d i a g 1 , 1 , 0 , z i j T = 1 , x i j 1 , x i j 2 . For the fixed effect coefficient, consider the following two cases:
Case 1: Sparse model β = 3 , 1.5 , 0 , 0 , 2 , 0 , 0 , 0 T .
Case 2: Dense model β = 0.85 , 0.85 , 0.85 , 0.85 , 0.85 , 0.85 , 0.85 , 0.85 T .
Setting the threshold value ε = 10 6 when stopping the iteration, simulations of four methods are compared in the above two cases, and the results after repeating the simulations 100 times are shown in Table 3 (In case 2, the number of true zero coefficients is 0, which makes the denominator of the CER formula zero. Therefore, the CER is mathematically undefined and is conventionally recorded as 1.0.).
From Table 3, it can be seen that under the sparse model (case 1), the MRME value of the model with the double LASSO penalty (DLASSO-GACV) is 55 times that of the double SCAD method proposed in this paper, indicating that its parameter estimation accuracy is inferior to that of our method. The CSR values of the two methods are the same, which means they have the same effect in selecting important variables. However, the CER value of DLASSO is much lower than that of the double SCAD method, indicating that its ability to exclude insignificant variables is weaker than that of our method.
Compared with the Poisson mixed-effects model with a single SCAD penalty only on the fixed effects (P-H-SCAD), the MRME value obtained from the parameter estimation of our double SCAD penalty model is significantly lower. Specifically, the MRME value of the P-H-SCAD method is 24 times that of the DSCAD-GACV method. Although the CER value of P-H-SCAD is higher, its CSR value for selecting important variables is only half of that of the double SCAD method. Moreover, in model construction, the CSR value is more important than the CER value, which suggests that for the Poisson mixed-effects model, the double SCAD penalty performs significantly better than the single penalty.
For the unpenalized Poisson regression model, not only is its estimation error (MRME) larger, but it also fails to exclude any irrelevant variables.
From the perspective of the bias–variance trade-off, under the simulation settings adopted in this study, the proposed double SCAD penalty method (DSCAD-GACV) achieves a more favorable balance between variable sparsity and estimation stability.
Specifically, in the sparse model, DSCAD-GACV yields the lowest absolute bias (0.185) and variance (0.021), thereby attaining the smallest mean squared error (MSE = 0.055) among all competing methods. In the dense model, this method maintains an absolute bias of 0.210 and a variance of 0.035, both of which are significantly lower than those of the double LASSO (DLASSO-GACV) and single SCAD (P-H-SCAD) methods, demonstrating superior estimation stability.
In contrast, under the same simulation conditions, the double LASSO penalty (DLASSO-GACV) introduces considerably larger estimation bias and variance. For illustration, in the sparse model, the absolute bias of DLASSO-GACV (0.220) is 1.2 times that of DSCAD-GACV, while its variance (0.028) is 1.3 times higher than that of DSCAD-GACV. Although the unpenalized model (P) exhibits relatively small bias (0.193 in the sparse model and 0.215 in the dense model), its excessively large variance (0.082 in the sparse model and 0.098 in the dense model) leads to inadequate stability of the estimation results.
In the dense model, the MRME value of the double LASSO method is much larger than that of the double SCAD method, and the MRME value of the single-penalty method is also higher than that of the DSCAD-GACV method by a factor of 4.6. The CSR value of the double LASSO penalty is 0.889, while that of the single penalty is only 2.9, and that of the double SCAD method is 0.787. It can be known that although the CSR value of double LASSO is high, the double SCAD penalty method in this paper is much higher than that of double LASSO in terms of the accuracy of parameter estimation, and performs better in the dense model than that of the single SCAD penalty. For the non-penalty method, although all important variables are selected, the MRME values are much larger than those of the double SCAD method.
From the above analysis, under the sparse and dense model settings of this study (i.e., n = 30 , m = 10 ,   ρ = 0.5 ,   D = d i a g ( 1 , 1 , 0 ) , and 100 repetitions), the double SCAD penalty method proposed in this paper demonstrates significantly better accuracy in variable estimation and selection compared to the double LASSO and single SCAD penalty models. This conclusion is an empirical finding within the scope of the current simulation design and should not be generalized to all possible modeling scenarios. In the sparse case, the double SCAD method successfully selects all required variables, while in the dense case, some required variables are missed, and the MRME is higher. This suggests that, within the current experimental conditions, the double SCAD method is more suitable for sparse modeling scenarios. Further research is needed to evaluate its performance under different sample sizes, parameter dimensions, and data distributions.
Limitations and Future Research:
The simulation results presented in this chapter are based on a specific set of experimental conditions, including sample size n = 30 , m = 10 , fixed effect sparsity, and Poisson-distributed outcomes. While these settings are representative of many applied scenarios, the findings should be interpreted as empirical evidence within the scope of this study, rather than universal conclusions about the superiority of the double SCAD penalty method. Future research will extend the simulation design to include larger sample sizes, higher-dimensional parameters, and different outcome distributions to further validate the generalizability of the proposed method.

5. Examples

This paper analyzes data from a cigarette factory using the double SCAD method proposed above. The data were collected from 15 April to 15 June 2020 (no data were recorded on May 1, Labor Day). The factory has three shifts and six machines, with two shifts rotating each morning and evening, and each shift operating all six machines. In total, there are 708 observations.
This study mainly explores the effects of machine parameters on the number of short cigarettes. A short cigarette is one of the most common quality defects during the cigarette rolling process. Cigarettes with such quality defects are discarded, which increases production consumption. Investigating which machine parameters have significant effects can help cigarette factories reduce consumption losses.
The details of the machine parameters are shown in Table 4.
The scatter plot of the number of short cigarettes versus each variable is shown in Figure 2.
From Figure 2, we can observe that the scatter plots of vehicle speed, rubbing plate temperature, SE1 soldering iron temperature, and SE2 soldering iron temperature versus the number of short cigarettes are clustered. By contrast, the scatter plots of compaction end position versus the number of short cigarettes are more scattered, and those of flattening position versus the number of short cigarettes are highly dispersed with no obvious linear pattern.
According to the data characteristics, the machine parameter variables are analyzed by the machine. Table 5 presents the descriptive statistics for each machine parameter variable.
As shown in Table 5, there are obvious differences in the distribution of parameters across different machines. For vehicle speed, machine 7 has higher minimum, mean, and median values but a smaller standard deviation, indicating more aggregated data. For the rubbing plate temperature, the ranges, means, and medians are similar among machines 5, 6, and 8, as well as between machines 4 and 7, while machine 9 has a minimum of 0, and the standard deviations of machines 4 and 5 are close but differ from those of the others. For flat plate position, machines 4 and 6 have positive maximum values and standard deviations above two, whereas other machines have negative maxima and standard deviations below 1.3, with large variations in medians. For compacting end position, machines 6 and 8 have maximum values 2–3 times those of other machines, and machine 8 has a much higher standard deviation. For SE1 soldering iron temperature, machine 4 has a much lower minimum, while machine 6 has a higher minimum and larger standard deviation, with similar means and medians between machines 5 and 8, and between machines 7 and 9; SE2 soldering iron temperature follows a similar pattern, with machine 4 showing more dispersed data. Overall, the ranges and dispersions of parameters differ across machines, suggesting that machine-level parameters may affect the number of short cigarettes.
Next, the characteristics of the short-position data will be researched by machine, and the box line plots are shown in Figure 3.
As can be seen from Figure 3, the distribution of short-position values is more scattered in machines 7 and 8, and more aggregated in machines 5 and 6. Moreover, the mean of short-position values in machine 4 is higher than the mean of the other machines, and the means in machines 5 and 6 are significantly lower. As we can see, there are some variations in the short-position values between the different units, indicating that the type of machine has some influence on the short-position numbers. It can be seen that for the daily number of short cigarettes, it is necessary to consider the type of machine when researching it, and the longitudinal counting data can be used to incorporate the effect of machine type into the short-position data.
In the actual data, there are some null values for each explanatory variable, which are filled with the median, and the data are normalized and modeled by the following mixed effects Poisson counting regression model:
E ( Y i j | X i j , Z i j ) = e X i j T β + Z i j T α i
Since there are 2 shifts in production per day, one shift per day is taken as a whole sample cross-section, the whole is noted as a work schedule for a total of 59 days with two shifts per day, that is n = 118 . Each work schedule operates on 6 machines, that is m = 6 , the total sample size was 708. The number of short positions generated by the first i work schedule operation of the first j machine is noted as Y i j , X i j T = ( 1 , X i j 1 , X i j 2 , X i j 3 , X i j 4 , X i j 5 , X i j 6 ) denotes the intercept and 6 values of machine parameters obtained for the first i work schedule operation of the first j machine, and β is the coefficient vector of the explanatory variables with dimension 1 × 7 . Z i j T is the random effect generated by the explanatory variables when operating 6 machines in 118 work schedules. Since it is not known in advance which explanatory variables generate random effects, let Z i j = X i j ; α i is the coefficient vector corresponding to Z i j T , with dimension 118 × 7 .
The double SCAD penalty method proposed in this paper is used to model and analyze the cigarette plant data to select the variables that would have an impact on the number of short cigarettes. And for the selection criteria of the penalty parameters, according to the simulation, the GACV criterion and the Poisson regression model without penalty are used as a control group; the results of the runs are shown in Table 6 and Figure 4.
In the Double SCAD (DSCAD) penalized model, the standard errors and p-values for some variables appear as NA, which is caused by the variable selection mechanism of the model. The DSCAD penalty shrinks the coefficients of variables with weak explanatory power for the response variable to zero, thereby achieving automatic variable selection. When refitting the mixed-effects Poisson model based on the selection results, these variables with zero coefficients are excluded from the model, so their corresponding standard errors, z-statistics, and p-values cannot be calculated. An estimated value of zero in the table indicates that the variable has been eliminated by the model. The NA values do not represent missing data but are normal outcomes of model simplification after variable selection. This reflects the advantage of DSCAD in performing high-dimensional variable selection and model reduction while maintaining satisfactory fitting performance.
As shown in Table 6, the DSCAD model based on the GACV criterion effectively achieves sparse variable selection and identifies only two non-zero variables as the key factors contributing to short-position defects:
Rubbing plate temperature (X2) with a coefficient of 0.0067 (p < 0.001), indicating that a 1 °C increase in temperature is associated with a 0.67% average increase in the expected count of short cigarettes.
Position of flat plate (X3) with a coefficient of 0.0409 (p < 0.001), meaning that a one-unit increase in the flat plate position leads to a 4.09% increase in the expected count of short cigarettes, making it the most significant factor.
All other variables (X1, X4, X5, X6) are estimated to be exactly zero, with their standard errors and p-values marked as NA, confirming that these variables have been excluded from the model.
In sharp contrast, the unpenalized Poisson model yields significant non-zero coefficients (p < 0.001) for all variables, including those that are actually irrelevant (e.g., vehicle speed X1).
This demonstrates that the DSCAD model can produce a parsimonious and interpretable model specification, whereas the traditional Poisson model suffers from overfitting and yields misleading results.
Figure 4 visually reinforces these findings. The blue bars represent the inflated coefficients from the unpenalized Poisson model, which are all non-zero and vary widely. The red bars represent the DSCAD estimates, which are sparse and concentrated on the two meaningful variables. For variables like vehicle speed ( X 1 ) and compacting end position ( X 4 ) , the DSCAD model correctly sets their coefficients to zero, while the Poisson model incorrectly identifies them as significant. This visual comparison clearly demonstrates the superiority of the DSCAD method in distinguishing signal from noise in real industrial data.
To quantify the advantages of the DSCAD model, we compared its fitting performance with the unpenalized Poisson model.
As shown in Table 7, the DSCAD model achieved lower AIC (388,367.6 vs. 391,879.4) and BIC (388,408.7 vs. 391,902.2), indicating a better balance between model fit and complexity.

Practical Implications for Cigarette Factory Management

The real-world application provides clear, actionable insights.
Prioritize process control for two key parameters:
  • Flat plate position (X3): As the most impactful factor, engineers should focus on tightening the tolerance for this setting to minimize production errors.
  • Rubbing plate temperature (X2): While the effect is smaller, it is statistically significant. Monitoring and stabilizing this temperature can lead to incremental quality improvements.
Reallocate quality control resources:
The DSCAD model confirms that vehicle speed, compacting end position, and soldering iron temperatures are not significant drivers. Resources currently allocated to monitoring these variables can be redirected to controlling the flat plate position and rubbing plate temperature.
Integrate the DSCAD method into routine QC:
The method can be used to periodically re-evaluate the drivers of quality defects, ensuring that process control efforts remain focused on the most impactful factors.
The application to the cigarette factory data provides strong empirical evidence for the advantages of the proposed double SCAD penalty method. By automatically selecting only two relevant variables from six candidates, the model avoids overfitting and produces results that are consistent with engineering expertise. The DSCAD model not only improves interpretability but also delivers superior predictive performance, making it a valuable tool for data-driven decision-making in industrial quality control.

6. Conclusions

This paper proposes a double SCAD penalty model that imposes the SCAD penalty simultaneously on both fixed effects and random effects within the mixed-effects Poisson count regression model. Furthermore, the model accounts for the influence of random effects while enabling concurrent parameter estimation and variable selection. For parameter estimation, this paper presents the DSCAD-LQA to compute the parameters of the double SCAD model, which achieves simultaneous coefficient estimation for both fixed effects and random effects. With regard to penalty parameter selection, simulation results show that the model estimation performance under the GACV criterion is superior to that under the SIC in all scenarios. This paper mainly investigates whether variations in the degree of variable correlation and changes in the covariance of random effects affect the accuracy of model parameter estimation and variable selection.
When the variable correlation coefficients change, it is found that as the correlation coefficients increase, the estimation error of the proposed double SCAD model gradually increases, and the accuracy of the model in correctly identifying important variables and excluding unimportant variables decreases accordingly. In particular, when the correlation coefficients are relatively large, the model’s performance in parameter estimation and variable selection is affected more significantly.
When the covariance of random effects changes, it is observed that with the gradual increase in covariance, the parameter estimation error of the double SCAD model also increases, and the accuracy of the model in including important variables and excluding unimportant variables declines. When the covariance of random effects is relatively small, the model exhibits superior performance in parameter estimation and variable selection.
Furthermore, this paper compares the proposed double SCAD penalty model with three benchmark methods—the non-penalized Poisson regression model, the mixed-effects Poisson regression model with a single SCAD penalty on fixed effects, and the mixed-effects Poisson regression model with a double LASSO penalty—under both sparse and dense scenarios. The results demonstrate that the proposed double SCAD penalty model achieves better performance in parameter estimation and variable selection than the other three methods and is more suitable for sparse scenarios.

Author Contributions

Conceptualization, K.L. and H.L.; Software, X.R.; Validation, K.L. and X.R.; Formal analysis, K.L.; Investigation, X.R.; Resources, H.L.; Writing—original draft, K.L.; Writing—review & editing, Y.L.; Visualization, X.R.; Supervision, H.L. and Y.L.; Project administration, H.L. and Y.L.; Funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

National Social Science Fund of China: 24BTJ068; Key Humanities and Social Science Fund of Hubei Provincial Department of Education: 25D096.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fan, J.; Peng, H. Nonconcave Penalized Likelihood with a Diverging Number of Parameters. Biometrics 2004, 32, 928–961. [Google Scholar] [CrossRef]
  2. Laird, N.M.; Ware, J.H. Random Effects Models for Longitudinal Data. Biostatistics 2004, 38, 963–974. [Google Scholar] [CrossRef]
  3. Wang, Y.; Han, Y. Score Tests for Overdispersion in Marginalized Zero-Inflated Poisson Regression Based on Marginalized Zero-Inflated Generalized Poisson Model. Stat. Anal. Data Min. ASA Data Sci. J. 2025, 18, e70019. [Google Scholar] [CrossRef]
  4. Zhang, Z.; Wang, X.; Zhang, X.; Zhang, J. Simultaneously detecting spatiotemporal changes with penalized Poisson regression models. Comput. Stat. Data Anal. 2025, 212, 108240. [Google Scholar] [CrossRef]
  5. Arima, S.; Calculli, C.; Pollice, A. A zero-inflated Poisson spatial model with misreporting for wildfire occurrences in southern Italian municipalities. Environmetrics 2024, 36, e2853. [Google Scholar] [CrossRef]
  6. Gomtsyan, M.; Lévy-Leduc, C.; Ouadah, S.; Sansonnet, L. Sign consistent estimation in a sparse Poisson model. Stat. Probab. Lett. 2024, 209, 110107. [Google Scholar] [CrossRef]
  7. Liu, Y.; Yin, J. Variable selection in partially linear regression models for time series. Commun. Stat. Theory Methods 2026, 55, 468–486. [Google Scholar] [CrossRef]
  8. Stolfi, P.; Bernardi, M.; Petrella, L. Sparse simulation-based estimator built on quantiles. Econom. Stat. 2025, 34, 32–43. [Google Scholar] [CrossRef]
  9. Tibshirani, R.J. Regression Shrinkage and Selection Via the Lasso. J. R. Stat. Soc. B 1996, 58, 267–288. [Google Scholar] [CrossRef]
  10. Fan, J.; Li, R. Variable Selection Via Nonconcave Penalized Likelihood and Its Oracle Properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
  11. Wang, M.Q.; Song, L.X.; Tian, G.L. SCAD-Penalized Least Absolute Deviation Regression in High-Dimensional Models. Commun. Stat. Theory Methods 2015, 44, 2452–2472. [Google Scholar] [CrossRef]
  12. Zou, H.; Li, R.Z. One-step Sparse Estimates in Nonconcave Penalized Likelihood Models. Ann. Stat. 2008, 36, 1509–1533. [Google Scholar]
  13. Liu, Z.X.; Huang, Z.Q.; Sun, Q.; Zhang, H. Financial Early-Warning Model Using Logistic Regression with SCAD Penalty. Stat. Inf. Forum 2012, 27, 21–26. [Google Scholar]
  14. Li, Y.; Xia, C. Penalized Quasi-Likelihood SCAD Estimator in High-Dimensional Generalized Linear Models. J. Wuhan Univ. 2018, 6, 533–539. [Google Scholar]
  15. Buu, A.; Johnson, N.J.; Li, R.; Tan, X. New variable selection method for zero inflated count data with applications substance abuse field. Stat. Med. 2011, 30, 2326–2340. [Google Scholar] [CrossRef] [PubMed]
  16. Li, X.L. Research on Two Types of High-Dimensional Random Effects Regression Models Based on SCAD-L2 and SCAD Mixed Penalties. Doctoral Dissertation, Wuhan University of Technology, Wuhan, China, 2023. [Google Scholar]
  17. Li, X.L.; He, S.X.; Wang, C.M. High-dimensional random effects linear regression model based on SCAD-L2 and SCAD mixed penalties. Acta Math. Sci. 2023, 43, 1297–1310. [Google Scholar]
  18. Liu, Y.Z. SCAD-Regularized Heterogeneous Autoregressive Model for Volatility Forecasting. Master’s Thesis, Dalian University of Technology, Dalian, China, 2018. [Google Scholar]
  19. Liu, J.; Chen, H.J. Regression model based on self-supervised learning and SCAD-Net regularization. Comput. Syst. Appl. 2021, 30, 37–45. [Google Scholar]
  20. Liu, X.H.; Wang, Z.F.; Wu, Y.H. SCAD-type variable selection and estimation in censored regression models. J. Univ. Sci. Technol. China 2013, 43, 182–189. [Google Scholar]
  21. Zhang, W.J.; Dong, C.L. Variable selection for Poisson regression models. J. Xinjiang Norm. Univ. (Nat. Sci. Ed.) 2025, 1–11. Available online: https://kns.cnki.net/kcms2/article/abstract?v=esvOG1ozB-iMynTYyhzH1hEVbTSOimCnirmRnToaVqXVBO6jlyn45_eftAl_jf0kVAFJjKN57QjK4ktKJS3eIDbmVCJ0UbMX9V3TjdXp1PmimL6tk42-B4VPG0IBzM8-LHa90PZkQ-Fk1rHPm4HIPW13SKkGdBRH_2M6-Ags3Lg5lPhf38tuzQ8vFjTtx5ni7z8XVIa6yEk&uniplatform=NZKPT&captchaId=ee0e796a-31cd-4fcf-b366-b2fc34105a2c (accessed on 6 March 2026).
  22. Lange, N.; Laird, N.M. The Effect of Covariance Structures on Variance Estimation in Balance Growth-curve Models with Random Parameters. J. Am. Stat. Assoc. 1989, 84, 241–247. [Google Scholar] [CrossRef]
  23. Ibrahim, J.G.; Zhu, H.; Garcia, R.I.; Guo, R. Fixed and random effects selection in mixed effects models. Biometrics 2011, 67, 495–503. [Google Scholar] [CrossRef] [PubMed]
  24. Luo, Y.X.; Li, H.F. Research of Multi-Penalty Regression Process of Mixed Effects Models and Its Convergence. Stat. Inf. Forum 2017, 32, 3–10. [Google Scholar]
  25. Koenker, R.; Ng, P.; Portnoy, S. Quantile Smoothing Splines. Biometrika 1982, 81, 673–680. [Google Scholar] [CrossRef]
  26. Yuan, M. GACV for Quantile Smoothing Splines. Comput. Stat. Data Anal. 2006, 50, 813–829. [Google Scholar] [CrossRef]
Figure 1. Simulation results with different sizes of random-effect covariance.
Figure 1. Simulation results with different sizes of random-effect covariance.
Axioms 15 00214 g001
Figure 2. Scatter plot of the number of short cigarettes versus machine parameter variables.
Figure 2. Scatter plot of the number of short cigarettes versus machine parameter variables.
Axioms 15 00214 g002
Figure 3. Short-position data box line plot for each machine.
Figure 3. Short-position data box line plot for each machine.
Axioms 15 00214 g003
Figure 4. Coefficient comparison between double SCAD and unpenalized Poisson model.
Figure 4. Coefficient comparison between double SCAD and unpenalized Poisson model.
Axioms 15 00214 g004
Table 1. Comparison between variance-component penalization and direct coefficient penalization.
Table 1. Comparison between variance-component penalization and direct coefficient penalization.
DimensionVariance-Component PenalizationDirect Coefficient Penalization (Our Approach)
Penalty TargetRandom-effect variances V a r ( α i ) Individual random-effect realizations α i t
Shrinkage BehaviorAll-or-nothing: if V a r ( α i ) 0 , all αit are forced to 0Local: only irrelevant α i t are shrunk to 0
Hierarchical BorrowingErases all within-cluster variation for a given effectPreserves meaningful individual heterogeneity
Effective ComplexityControls only the number of active random effectsControls both the number of effects and the sparsity of their realizations
Table 2. Simulation results with different correlation coefficients.
Table 2. Simulation results with different correlation coefficients.
ρ MethodMRMECSRCERBiasVarianceMSE
0.25DSCAD-SIC0.0691.0000.8760.3042.0920.160
DSCAD-GACV0.0601.0000.8860.1641.9480.096
0.5DSCAD-SIC0.0841.0000.8060.3862.2290.243
DSCAD-GACV0.0791.0000.8440.2262.0900.164
0.75DSCAD-SIC2.7260.9000.3900.5422.6130.498
DSCAD-GACV2.4160.9000.4180.3332.2650.399
Table 3. Comparison of simulation results between the three methods.
Table 3. Comparison of simulation results between the three methods.
MethodModelMRMECSRCERBiasVarianceMSE
DSCAD-GACVsparse0.0791.00.8440.1850.0210.055
dense0.2690.7871.00.2100.0350.088
DLASSO-GACVsparse4.3521.00.20.2200.0280.074
dense26.3340.8891.00.2450.0420.105
P-H-SCADsparse1.9060.3970.880.2550.0320.085
dense1.5100.3221.00.2800.0480.120
Psparse40.9831.000.1930.0820.120
dense2.2190.8891.00.2150.0980.150
Table 4. Machine parameter variable description statistics.
Table 4. Machine parameter variable description statistics.
VariableParameter NameRangeMedianMissing Amount
X 1 Vehicle speed [0, 5523] 544262
X 2 Rubbing plate temperature[0, 175]16063
X 3 The position of the flat plate[−8.3, 0.8]−4.661
X 4 The position of the compacting end[−5, 17]062
X 5 SE1 soldering iron temperature[28, 251]21861
X 6 SE2 soldering iron temperature[26, 249]21162
Table 5. Description statistics of each machine parameter variable.
Table 5. Description statistics of each machine parameter variable.
VariableMachineRangeMeanStandard DeviationMissing Amount
vehicle speed4 [17, 5490] 4709.201685.175456
5 [1, 5486] 4736.121625.845293.5
6 [0, 5504] 4931.831503.395492
7 [337, 5523] 5342.96555.705518
8 [0, 5491] 4772.191661.605415
9 [1, 5484] 4893.661556.765454.5
rubbing plate temperature4[134, 154]143.875.38143
5[154, 175]164.305.56160
6[156, 171]161.303.39160
7[143, 154]150.421.76151
8[156, 167]160.252.29160
9[0, 167]163.2016.19165
the position of the flat plate4[−8.3, 0.4]−3.452.70−2.4
5[−8.2, −2.0]−4.601.26−4.4
6[−8.1, 0.8]−4.062.20−3.1
7[−6.9, −1.8]−4.221.17−3.9
8[−8.2, −0.7]−4.921.18−5.0
9[−8.3, −2.0]−5.541.12−5.6
the position of the compacting end4[−4, 9.4]0.882.670
5[−2, 8.3]1.432.111
6[−4, 15.6]0.062.890
7[−4, 5.0]0.321.840
8[−5, 17.1]0.654.620
9[−5, 6.6]−0.991.80−1
SE1 soldering iron temperature4[28, 245]211.7522.29212
5[188, 251]226.6810.60225
6[213, 243]223.617.00224
7[190, 220]205.708.15205
8[199, 231]222.5812.57230
9[194, 230]209.079.18209
SE2 soldering iron temperature4[26, 239]208.6821.58211
5[186, 249]224.688.88226
6[207, 231]215.274.83216
7[195, 211]203.104.77203
8[199, 211]208.434.19211
9[194, 216]205.135.96205
Table 6. Variable estimation parameter.
Table 6. Variable estimation parameter.
VariableParameter NameDSCAD EstimateDSCAD Std. ErrorDSCAD p-ValuePoisson EstimatePoisson Std. ErrorPoisson p-Value
X 1 Vehicle speed0.0000NANA0.02290.00060.0000
X 2 Rubbing plate temperature0.00670.00060−0.00040.00060.5093
X 3 Position of flat plate0.04090.001000.02720.00100.0000
X 4 Position of compacting end0.0000NANA0.02530.00070.0000
X 5 SE1 soldering iron temperature0.0000NANA0.02150.00090.0000
X 6 z SE2 soldering iron temperature0.0000NANA−0.00820.00100.0000
Table 7. Model fit comparison.
Table 7. Model fit comparison.
ModelLog-LikelihoodAICBICNumber of Variables
DSCAD-GACV−194,174.8388,367.6388,408.76
Poisson−195,934.7391,879.4391,902.22
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, K.; Ren, X.; Li, H.; Luo, Y. Poisson Mixed-Effects Count Regression Model Based on Double SCAD Penalty and Its Simulation Study. Axioms 2026, 15, 214. https://doi.org/10.3390/axioms15030214

AMA Style

Li K, Ren X, Li H, Luo Y. Poisson Mixed-Effects Count Regression Model Based on Double SCAD Penalty and Its Simulation Study. Axioms. 2026; 15(3):214. https://doi.org/10.3390/axioms15030214

Chicago/Turabian Style

Li, Keqian, Xueni Ren, Hanfang Li, and Youxi Luo. 2026. "Poisson Mixed-Effects Count Regression Model Based on Double SCAD Penalty and Its Simulation Study" Axioms 15, no. 3: 214. https://doi.org/10.3390/axioms15030214

APA Style

Li, K., Ren, X., Li, H., & Luo, Y. (2026). Poisson Mixed-Effects Count Regression Model Based on Double SCAD Penalty and Its Simulation Study. Axioms, 15(3), 214. https://doi.org/10.3390/axioms15030214

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop