Next Article in Journal
Additive Derivations of Incidence Modules
Previous Article in Journal
Nonlinear Mixed Jordan-Type Derivations on ∗-Algebra
Previous Article in Special Issue
A Novel Family of CDF Estimators Under PPS Sampling: Computational, Theoretical, and Applied Perspectives
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Estimation for Longitudinal Varying Coefficient Partially Nonlinear Models Based on QR Decomposition

1
School of Mathematics, North University of China, Taiyuan 030051, China
2
School of Mathematics and Big Data, Dezhou University, Dezhou 253023, China
3
School of Mathematics and Statistics, Shandong University of Technology, Zibo 255022, China
*
Author to whom correspondence should be addressed.
Axioms 2025, 14(12), 875; https://doi.org/10.3390/axioms14120875
Submission received: 28 October 2025 / Revised: 21 November 2025 / Accepted: 26 November 2025 / Published: 28 November 2025
(This article belongs to the Special Issue Computational Statistics and Its Applications, 2nd Edition)

Abstract

To address the estimation efficiency issues arising from multicollinearity and longitudinal data correlation in the varying coefficient partially nonlinear models (VCPNLM), a method based on QR decomposition and quadratic inference function (QIF) is proposed to obtain the orthogonality estimation of parameter components and varying coefficient functions. QR decomposition eliminates the pathology of the design matrix, and combines the adaptive weighting of the relevant structures within the group by QIF to effectively capture the complex correlation structure of longitudinal data. The theoretical analysis proves the asymptotic nature of the estimator, and the efficiency of the estimation method proposed in this paper is verified by simulation experiments.

1. Introduction

In recent years, longitudinal data have attracted much attention due to their wide applications in domains like biomedicine, economics and social sciences. This kind of data results from repeatedly observing the same subject at various time intervals. As a result, multiple measurement results of the same subject exhibit temporal or spatial dependence, which makes the traditional independent and identically distributed assumption no longer applicable. To address this issue, the VCPNLM has become one of the research hotspots, as it combines the interpretability of parametric models with the flexibility of nonparametric models. Li and Mei [1] put forward a profile nonlinear least squares estimation approach to estimate the parameter vector and coefficient function vector of the VCPNLM, and further derived the asymptotic characteristics of the obtained estimators.
Important progress has been made in model estimation in existing studies. Yu et al. [2] constructed robust estimators for the partially linear additive model under functional data. Yan et al. [3] explored the empirical likelihood inference targeting the partially linear errors-in-variables model under longitudinal data. VCPNLM, based on mode regression, a robust two-stage method for estimation and variable selection, was proposed by Xiao and Liang [4]. Zhao et al. [5] put forward a new orthogonality-based empirical likelihood inference method through orthogonal estimation techniques and empirical likelihood inference methods, which is used to estimate the parametric and nonparametric components in a class of varying coefficient partially linear instrumental variable models for longitudinal data. Liang and Zeger [6] applied the generalized estimating equation to longitudinal data analysis for effective handling of its correlation issue. For the varying coefficient partially nonlinear quantile regression model under randomly left-censored data, Xu et al. [7] introduced a three-stage estimation weighting method. A comprehensive overview of longitudinal data analysis methods can be found in Diggle et al. [8].
The VCPNLM has received widespread attention in statistical research due to its flexibility in capturing dynamic relationships between variables and adaptability to complex data structures [4,6,7]. It integrates the advantages of varying coefficient models and partially nonlinear models, making it a powerful tool for practical applications such as biomedical research, economic forecasting, and environmental monitoring. However, the complexity of the model structure and the characteristics of real-world data pose great challenges to accurate parameter estimation, which motivates the need for further improvements in existing methods.
Despite the notable advancements mentioned above, significant challenges still exist in the existing methods for estimating the VCPNLM. For instance, the estimation of nonlinear parameters is easily disturbed by nonparametric components and longitudinal correlation structures, leading to reduced estimation efficiency; when there is multicollinearity among explanatory variables, the ill-posed nature of the design matrix will further decrease the stability of estimation. Such issues severely limit the performance and practicality of the model in practical applications.
To address the above issues, existing studies have proposed a variety of improved methods. For example, Qu et al. [9] used the QIF to advance the estimating equation, and this method maintains optimal performance even as the working correlation structure is misspecified. Bai et al. [10] applied the quadratic inference function to handle longitudinal data, and the results demonstrate that the proposed estimation method exhibits excellent asymptotic properties. For semiparametric varying coefficient partially linear models, Tian et al. [11] proposed penalized QIF. Schumaker [12] utilized B-spline basis functions to approximate the varying coefficient part, which improves the computational efficiency. For linear models with randomly missing responses, Wei et al. [13] introduced a model averaging approach. Jiang et al. [14] proposed an estimation method for the VCPNLM based on the exponential squared loss function. Xiao and Chen [15] advanced a procedure for local bias-corrected cross-sectional nonlinear least squares estimation. By adopting an orthogonality-projection method, Yang and Yang [16] developed smooth-threshold estimating equations for VCPNLM. Additional studies focusing on this model include Refs. [7,17,18,19]. These investigations address various data scenarios and model settings, offering targeted estimation methods and inference strategies to further supplement and optimize the research framework within this field.
This paper proposes an orthogonal estimation framework that integrates QR decomposition and the QIF. QR decomposition can effectively eliminate the ill-posed nature of the design matrix and improve numerical stability. The QIF method avoids the limitations of traditional generalized estimating equations by adaptively weighting the intra-group correlation structure, significantly enhancing estimation efficiency. This study not only provides a new theoretical tool for longitudinal data analysis but also offers a feasible solution to complex modeling problems in practical applications.
The structure of this paper is arranged as follows: Section 2 introduces the model specification and estimation method, including the specific implementation of QR decomposition and QIF; Section 3 addresses the asymptotic properties of the estimators; Section 4 verifies the superiority of the proposed method using simulation experiments; Appendix A contains the proof process of the key conclusions.

2. Models and Methods

Consider the VCPNLM introduced by Li and Mei [1]:
Y = X T α U + g Z , β + ε ,
where Y R is the response variable, X R p , U R , and Z R r are covariates, α U = α 1 U , α 2 U , , α p U T and α k U , k = 1 , 2 , , p are unknown smooth functions, g ( · , · ) is a nonlinear function with a known form, β = β 1 , β 2 , , β q T denotes a q-dimensional unknown parameter vector, and ε is the model error with mean zero and variance σ 2 .

2.1. Estimation of Parameter Vector

Considering the model (1) under longitudinal data, suppose the j-th observation of the i-th individual satisfies
Y i j = X i j T α U i j + g Z i j , β + ε i j , i = 1 , 2 , , n , j = 1 , 2 , , n i ,
among them, ε i j has a mean of zero, and X i j R p , Z i j R r , U i j are covariates.
Based on the idea of Schumaker [12], the unknown function α k · , k = 1 , 2 , , p is approximated through basis functions. The B-spline basis functions are denoted by the vector B u = B 1 u , B 2 u , , B L u T . The dimension L is defined as L = K + M , where K and M represent the number of interior knots and the order of the spline, respectively. Then α k u is approximately expressed as
α k u B u T γ k , k = 1 , 2 , , p ;
here, γ k = γ k 1 , γ k 2 , , γ k L T are the coefficient vector for the B-spline basis functions. model (2) is expressed as
Y i j W i j T γ + g Z i j , β + ε i j , i = 1 , 2 , , n , j = 1 , 2 , , n i ,
define W i j = I p B U i j · X i j , where ⊗ is the Kronecker product, and let γ = γ 1 T , γ 2 T , , γ p T T , Y i = Y i 1 , Y i 2 , , Y i n i T , W i = W i 1 , W i 2 , , W i n i T , g ( Z i , β ) = g ( Z i 1 , β ) , g ( Z i 2 , β ) , ,   g ( Z i n i , β ) T , ε i = ε i 1 , ε i 2 , , ε i n i T . Then model (4) is expressed as
Y i W i γ + g Z i , β + ε i , i = 1 , 2 , , n .
We first state a fundamental result on the QR decomposition of full-column-rank matrices, which is essential for the subsequent steps.
Lemma 1
(QR Decomposition for Full-Column-Rank Matrices). Let W R m × k be a matrix with full column rank k. Then, an orthogonal matrix Q R m × m and an upper triangular matrix R R k × k exist, satisfying
W = Q R 0 ,
where 0 is a ( m k ) × k zero matrix. Moreover, Q can be partitioned as Q = ( Q 1 , Q 2 ) , with Q 1 R m × k and Q 2 R m × ( m k ) , satisfying Q 2 Q 1 = 0 .
We now proceed with the derivation from Equation (5). Suppose that for all i = 1 , 2 , , n , the matrices W i have full column rank, their QR decomposition can be expressed as
W i = Q i R i 0 , i = 1 , 2 , , n ,
where the definitions of Q i , R, and 0 are the same as those in Lemma 1 above. The matrix Q i can be divided into two parts as Q i = Q i 1 , Q i 2 , where Q i 1 is a n i × p L matrix and Q i 2 is a n i × n i p L matrix. Substitute the decomposition of Q i into (6) to obtain W i = Q i 1 R i , from the properties of orthogonal matrices, Q i 2 T Q i 1 = 0 can be derived, and then Q i 2 T W i = Q i 2 T Q i 1 R i = 0 is obtained. Multiplying both sides of Equation (5) by Q i 2 T , we get
Q i 2 T Y i Q i 2 T g Z i , β + Q i 2 T ε i , i = 1 , 2 , , n .
Obviously, model (7) is a regression model containing only unknown parameters. Following Liang [6], the generalized estimating equation for β can be formulated as
i = 1 n g ( 1 ) ( Z i , β ) T Q i 2 V i 1 Q i 2 T ( Y i g ( Z i , β ) ) = 0 ,
among them, g ( 1 ) ( Z i , β ) = ( g ( 1 ) ( Z i 1 , β ) , , g ( 1 ) ( Z i n i , β ) ) T , V i = Q i 2 T Σ i Q i 2 , Σ i is the covariance matrices of Y i , and the structure of V i can also be expressed as V i = A i 1 2 R ρ A i 1 2 according to the method of Liang [6], where A i is a diagonal matrix, R ρ is a working correlation matrix, and ρ is a correlation parameter. Since a consistent estimator of ρ is not always available in practice, we adopt the QIF method to approximate the working correlation matrix R 1 ( ρ ) through several basis matrices, thereby avoiding directly specifying the correlation structure. Drawing on the classic specifications proposed by Qu et al. [9] and combined with the correlation structure characteristics of the data in this study, we select a set of basis matrices M 1 , M 2 , , M s with corresponding coefficients a 1 , a 2 , , a s that satisfy
R 1 ( ρ ) a 1 M 1 + a 2 M 2 + + a s M s .
Substituting (9) into (8), we obtain
i = 1 n g ( 1 ) ( Z i , β ) T Q i 2 A i 1 2 ( a 1 M 1 + + a s M s ) A i 1 2 Q i 2 T ( Y i g ( Z i , β ) ) = 0 .
Define the extended score function as follows:
G n ( β ) = 1 n i = 1 n η i ( β ) 1 n i = 1 n g ( 1 ) ( Z i , β ) T Q i 2 A i 1 2 M 1 A i 1 2 Q i 2 T ( Y i g ( Z i , β ) ) g ( 1 ) ( Z i , β ) T Q i 2 A i 1 2 M 2 A i 1 2 Q i 2 T ( Y i g ( Z i , β ) ) g ( 1 ) ( Z i , β ) T Q i 2 A i 1 2 M s A i 1 2 Q i 2 T ( Y i g ( Z i , β ) ) .
Thus, the QIF for β can be defined as
M n ( β ) = G n T ( β ) Φ n 1 ( β ) G n ( β ) ,
where Φ n ( β ) = 1 n i = 1 n η i ( β ) η i T ( β ) , in this case, we obtain the estimator β ^ of β by minimizing the objective function M n ( β ) :
β ^ = arg min β M n ( β ) .

2.2. Estimation of the Coefficient Functions

The QIF effectively handles the correlation in longitudinal data and improves estimation efficiency by constructing a set of estimating equations. Therefore, after obtaining the initial estimates of the parameters β , we also use the QIF method to estimate the coefficient functions α k u , k = 1 , 2 , , p .
Substitute the estimator of into Equation (5), resulting in
Y ˜ i W i γ + ε i , i = 1 , 2 , , n ,
where Y ˜ i = Y i g i Z i , β ^ , assuming that the covariance matrix of Y ˜ i is Ψ i , and its structure is expressed following Liang [6] as Ψ i = A i 1 2 R ρ A i 1 2 , where A i and R ρ are defined as in the previous subsection. Assuming V i is known, construct the estimating equation
n i = 1 W i T Ψ i 1 Y ˜ i W i γ = 0 .
In practical applications, V i is usually unknown. Based on this, we still adopt the QIF method and use M 1 , M 2 , , M s to approximate R i 1 , then we have
n i = 1 W i T A i 1 2 a 1 M 1 + a 2 M 2 + + a s M s A i 1 2 Y i W i γ = 0 .
Define the extended score function as follows:
G n ( γ ) = 1 n i = 1 n φ i ( γ ) 1 n i = 1 n W i T A i 1 2 M 1 A i 1 2 Y i W i γ W i T A i 1 2 M 2 A i 1 2 Y i W i γ W i T A i 1 2 M s A i 1 2 Y i W i γ .
Thus, the QIF for γ is defined as follows:
Q n γ = G n T γ C n 1 ( γ ) G n γ ,
where C n ( γ ) = 1 n i = 1 n φ i γ φ i T γ . In this case, we obtain the estimator γ ^ of γ by minimizing the objective function Q n ( γ ) ,
γ ^ = arg min Q n γ .
Thus, the estimate of the coefficient function α k u can be expressed as
α ^ k u B u T γ ^ k , k = 1 , 2 , , p ,
where γ ^ k is the component corresponding to the k-th coefficient function in γ ^ .

3. Main Conclusions

This section studies the asymptotic properties of the estimators β ^ and α ^ k u , k = 1 , 2 , , p , assuming that β 0 and α 0 ( · ) are the true values of β and α ( · ) , respectively, γ 0 is the true value of γ , and α k 0 ( · ) and γ k 0 correspond to the k-th elements of α 0 ( · ) and γ , respectively. First, we present some common regularity conditions in longitudinal data analysis as follows:
(C1)
The support Ω of the random variable U is bounded, and its probability density function f u · has continuous second-order derivatives.
(C2)
The varying coefficient functions α 1 u , , α p u are continuously differentiable of order r on 0 , 1 , where r > 2 .
(C3)
For arbitrary Z, g Z , β exhibits continuity with respect to β , and g Z , β has continuous partial derivatives of order r.
(C4)
sup i E ( ε i ε i T ) < holds, and there exists some δ > 0 satisfying E ( ε i 2 + δ ) < .
(C5)
The covariates X i j and Z i j are assumed to satisfy the following conditions: sup i j E X i j 4 < , sup i j E Z i j 4 < , i = 1 , 2 , , n , j = 1 , 2 , , n i .
(C6)
Let t 1 , , t k be interior nodes on 0 , 1 . Furthermore, let t 0 = 0 , t k + 1 = 1 , ξ i = t i t i 1 , then a constant C 0 exists such that: max ξ i min ξ i C 0 , max ξ i + 1 ξ i = o K 1
(C7)
Define P i k = Q i 2 A i 1 2 M k A i 1 2 Q i 2 T , then we have:
1 n i = 1 n E g ( 1 ) ( Z i , β ) T P i k g ( 1 ) ( Z i , β ) P D k , k = 1 , 2 , , s .
1 n i = 1 n E g ( 1 ) ( Z i , β ) T P i l T V i P i l g ( 1 ) ( Z i , β ) P Ω l k , l , k = 1 , 2 , , s .
Among them, D k and Ω l k are constant matrices, and P denotes convergence in probability. Define Ω 0 = ( Ω l k ) p s × p s and Γ 0 = ( D 1 T , , D s T ) T , additionally, assume that Ω 0 and Γ 0 T Ω 0 1 Γ 0 are both invertible.
It is noted that conditions (C1) to (C5) are common conditions in VCPNLM components, condition (C6) indicates that t 0 , , t k + 1 is a uniform partition sequence over the interval 0 , 1 , and condition (C7) is used for subsequent proofs.
Theorem 1.
Under conditions (C1) to (C7), and when n , thus
n ( β ^ β 0 ) L N ( 0 , Σ ) ,
where the matrix Σ = ( Γ 0 T Ω 0 1 Γ 0 ) 1 , L denotes convergence in distribution.
Theorem 2.
Under conditions (C1) to (C7), the number of nodes K = O n 1 2 r + 1 , and when n , it follows that
α ^ k · α k 0 · = O p n r 2 r + 1 , k = 1 , 2 , , p ,
in which · denotes the function’s L 2 norm.

4. Simulation Study

This section assesses the finite-sample performance of the proposed orthogonality estimation method based on QR decomposition and QIF in VCPNLM through a Monte Carlo simulation study. We define the following model:
Y = X T α U + g Z , β + ε ,
among them, the covariates X , Z = ( Z 1 , Z 2 ) T both follow normal distributions, g ( Z , β ) is defined as g ( Z , β ) = β 1 Z 1 + β 2 exp ( Z 2 ) , with the parameter vector β = ( β 1 , β 2 ) T = ( 0.8 , 0.5 ) T . Additionally, U U ( 0 , 1 ) , the coefficient function α ( U ) = sin ( 2 π U ) , the error term ε follows an AR(1) process, and its structure is: ε i , j = 0.5 ε i , j 1 + v i , j , where v i , j N ( 0 , 0.5 1 0.5 2 ) .
The sample size is set to n = 100 , 150 , 200 ; for the i-th subject, the number of repeated measurements is m i 8 , and 1000 simulation runs are conducted for each case. The method combining QR decomposition and QIF proposed in this paper (OQIF) is compared with the profile nonlinear least squares method (PNLS) introduced by Li and Mei [1]. After sorting out, Table 1 and Table 2 are obtained.
As illustrated in Table 1, increasing the sample size leads to a reduction in both bias and standard deviation for both methods. Notably, the OQIF method demonstrates smaller bias and standard deviation, indicating superior estimation accuracy. Although larger sample sizes generally enhance the precision of both methods, the OQIF method outperforms in terms of bias control.
The results in Table 2 show that the mean length for the confidence interval of the OQIF method is significantly shorter than that of the PNLS method, demonstrating higher estimation efficiency; the coverage rate of the OQIF method is closer to the ideal 95%, while although the coverage rate of the PNLS method has improved, it is still below 95%.
In conclusion, as the sample size grows from 100 to 200, the OQIF method demonstrates higher estimation accuracy and stronger robustness across all evaluation indicators. These trend analyses further confirm the theoretical advantages and practical application value of the OQIF method in handling nonlinear longitudinal data.
We further conducted 1000 simulations and plotted boxplots of the 1000 RMSE values for parameters β ^ 1 and β ^ 2 , as shown in Figure 1 and Figure 2. From the figures, we can observe the following: The RMSE values of β ^ 1 and β ^ 2 for both methods decrease as the sample size increases; however, the OQIF method already exhibits good performance with a small sample size (n = 100), while the PNLS method requires a larger sample size to achieve similar accuracy. The boxplots of the OQIF method are more symmetric and compact, indicating that the distribution of its estimators is closer to the normal distribution and has better statistical properties. Additionally, the OQIF method has significantly fewer outliers than the PNLS method, demonstrating stronger robustness against abnormal data. This suggests that the overall performance of the OQIF method in parameter estimation is superior to that of the PNLS method, with more pronounced advantages especially in cases of finite samples.
Next, we estimate the coefficient function α ^ U , simulate 1000 times, and draw the box plot of the RMSE of α ^ U under different samples, resulting in Figure 3 below.
As can be seen from the above boxplots, with the increase in sample size, the error distributions of both the OQIF method and the PNLS method become more concentrated, but the OQIF method has much smaller errors than the PNLS method. This further verifies the superiority of the OQIF method in the VCPNLM.

Author Contributions

Conceptualization, J.G., X.Z. and C.W.; methodology, J.G., X.Z. and C.W.; software, J.G., and C.W.; validation, J.G., X.Z. and C.W.; data curation, J.G. and X.Z.; writing—original draft preparation, J.G.; writing—review and editing, X.Z.; funding acquisition, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Shandong Province (Grant No. ZR2022MA065).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
VCPNLMVarying coefficient partially nonlinear model
QIFQuadratic inference function
OQIFThe combination of QR decomposition and QIF
PNLSProfile nonlinear least squares

Appendix A

To prove Theorems 1 and 2, the following lemmas are presented as follows:
Lemma A1.
Let c R + , and let ξ i , i = 1 , 2 , n . be a sequence of independent and identically distributed random variables. If it satisfies conditions E ( ξ i ) = 0 and E ( ξ i 2 ) < c < , we get max 1 k n i = 1 k ξ i = O p ( n log n ) .
Proof. 
The result has been proven in Lemma A2 of Zhao and Xue [20]. □
Lemma A2.
Assume that the conditions (C1) to (C7) hold, when n , we get
n G n ( β 0 ) L N ( 0 , Ω 0 ) ,
where Ω 0 is given in condition (C7).
Proof. 
Let R ( U i ) = ( X i 1 T r ( U i 1 ) , , X i n i T r ( U i n i ) ) T , r ( u ) = ( r 1 ( u ) , , r p ( u ) ) T , r k ( u ) = α k ( u ) B ( u ) T γ k , combining with Q i 2 T W i = 0 , we can obtain
Q i 2 T Y i = Q i 2 T g ( Z i , β 0 ) + Q i 2 T W i γ 0 + Q i 2 T R ( U i ) + Q i 2 T ε i = Q i 2 T g ( Z i , β 0 ) + Q i 2 T R ( U i ) + Q i 2 T ε i .
Let G n , k ( β ) represent the k-th component of G n ( β ) , combining with Equation (11), we can obtain
n G n , k ( β 0 ) = 1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T ( Y i g ( Z i , β 0 ) ) = 1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T R ( U i ) + 1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T ε i .
Conclusion R ( U i ) = O ( K r ) can be derived from conditions (C2), (C5), and Corollary 6.21 of Schumaker [12], further combining with Lemma A1, we can obtain
1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T R ( U i ) = O p ( n 1 2 n 1 2 log n K r ) = o p ( 1 ) .
Let ζ i k = g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T ( Y i g ( Z i , β 0 ) ) , combining the conditions that the expectation of ε i is 0 and the covariance is V i given X i and Z i , we can conclude E ( ζ i k X i , Z i ) = 0 and cov ( ζ i k X i , Z i ) = g ( 1 ) ( Z i , β 0 ) T P i k V i P i k T g ( 1 ) ( Z i , β 0 ) , in which P i k is defined in condition (C7). We further let ζ i = ( ζ i 1 T , ζ i 2 T , , ζ i s T ) T , we can obtain E ( ζ i X i , Z i ) = 0 and
cov ( ζ i X i , Z i ) = N i 11 N i 1 s N i s 1 N i s s ,
where N i l k = g ( 1 ) ( Z i , β 0 ) T P i l V i P i k T g ( 1 ) ( Z i , β 0 ) , l , k = 1 , 2 , , s . combining with condition (C7) and the Law of Large Numbers, we conclude that
1 n i = 1 n N i l k P Ω l k .
Furthermore, for any constant vector a R s p that satisfying condition a T a = 1 , we have the expectation of a T ζ i is 0 and sup i E a T ζ i 2 + δ a 2 + δ sup i E ζ i 2 + δ C sup i E ε i 2 + δ < , where C is a positive constant. Therefore, a T ζ i satisfies the Lyapunov condition. Thus, we obtain: i = 1 n a T ζ i ( a T i = 1 n cov ( ζ i ) a ) 1 2 L N ( 0 , 1 ) . Further combining with Equations (A4) and (A5), we can obtain
1 n i = 1 n cov ( ζ i ) P Ω 0 .
Further combining with Equations (A2) and (A3), we get
n G n ( β 0 ) = 1 n i = 1 n ζ i + o p ( 1 ) L N ( 0 , Ω 0 ) .
Lemma A3.
Under the conditions (C1) to (C7), we have G ˙ n ( β 0 ) P Γ 0 and Ω n ( β 0 ) P Ω 0 , and Γ 0 and Ω 0 are specified in condition (C7).
Proof. 
After simple calculations, we can obtain
G ˙ n ( β 0 ) = 1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M 1 A i 1 2 Q i 2 T g ( 1 ) ( Z i , β 0 ) g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M s A i 1 2 Q i 2 T g ( 1 ) ( Z i , β 0 ) .
Combining with condition (C7) and the Law of Large Numbers, it follows that
1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T g ( 1 ) ( Z i , β 0 ) P D k , k = 1 , 2 , s .
Further combining with Equation (A8), we can obtain G ˙ n ( β 0 ) P Γ 0 . As demonstrated in the proof of Lemma 2:
Ω n ( β 0 ) = 1 n i = 1 n ζ i ζ i T + o p ( 1 ) = 1 n i = 1 n ζ i 1 ζ i 1 T ζ i 1 ζ i s T ζ i s ζ i 1 T ζ i s ζ i s T + o p ( 1 ) .
Similar to the proof of Equation (A5), we get
1 n i = 1 n ζ i l ζ i k T P Ω l k .
Therefore, Ω n ( β 0 ) P Ω 0 . □
Lemma A4.
Under the conditions (C1) to (C7), we can obtain
M ˙ n ( β 0 ) 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) = O p ( n 1 ) ,
and
M ¨ n ( β 0 ) 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) G ˙ n ( β 0 ) = o p ( 1 ) .
Proof. 
According to the definition of M n ( β ) , we can obtain the following result through calculation:
M ˙ n ( β 0 ) = 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) + G n T ( β 0 ) Ω n 1 ( β 0 ) Ω ˙ n ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) .
It follows from Lemma A2 that: G n ( β 0 ) = O p ( n 1 2 ) . Further, from condition (C7) and Lemma A3, we can obtain
G n T ( β 0 ) Ω n 1 ( β 0 ) Ω ˙ n ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) = O p ( n 1 ) .
Combining the Equations (A12) and (A13), we can obtain
M ˙ n ( β 0 ) = 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) + O p ( n 1 ) .
Similarly, we can obtain the following through calculation:
M ¨ n ( β 0 ) = 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) G ˙ n ( β 0 ) + R n ,
where,
R n = 2 G ¨ n T ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) 4 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) Ω ˙ n ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) + 2 G n T ( β 0 ) Ω n 1 ( β 0 ) Ω ˙ n ( β 0 ) Ω n 1 ( β 0 ) Ω ˙ n ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) G n T ( β 0 ) Ω n 1 ( β 0 ) Ω ¨ n ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) = T 1 + T 2 + T 3 + T 4 .
Combining Lemmas A2 and A3, we can obtain the following: T 1 and T 2 are both O p ( n 1 2 ) , while T 3 and T 4 are both O p ( n 1 ) , thus, R n = o p ( 1 ) . □
Proof of Theorem 1.
Noting that β ^ = arg min β M n ( β ) , thus M ˙ n ( β ^ ) = 0 , therefore,
M ˙ n ( β 0 ) + M ¨ n ( β * ) ( β ^ β 0 ) = 0 ,
specifically β * lies between β ^ and β 0 , thus, we have
n ( β ^ β 0 ) = n M ¨ n 1 ( β * ) M ˙ n ( β 0 ) .
Combining Lemma A3 and Lemma A4, we can obtain
M ¨ n 1 ( β * ) = 2 G ˙ n T ( β * ) Ω n 1 ( β * ) G ˙ n ( β * ) + o p ( 1 ) 1 = 1 2 Γ 0 T Ω 0 1 Γ 0 1 + o p ( 1 ) ,
and
n M ˙ n ( β 0 ) = 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) n G n ( β 0 ) + o p ( 1 ) = 2 Γ 0 Ω 0 1 n G n ( β 0 ) + o p ( 1 ) .
Combining Equations (A18) to (A20) and Slutsky’s Theorem, we get
n ( β ^ β 0 ) = 1 2 Γ 0 T Ω 0 1 Γ 0 1 + o p ( 1 ) 2 Γ 0 Ω 0 1 n G n ( β 0 ) + o p ( 1 ) = Γ 0 T Ω 0 1 Γ 0 1 Γ 0 Ω 0 1 n G n ( β 0 ) + o p ( 1 ) L N ( 0 , Σ ) ,
where, Σ = Γ 0 T Ω 0 1 Γ 0 1 . □
Proof of Theorem 2.
Let τ = n r 2 r + 1 , γ = γ 0 + τ s . That is, we need to prove that for arbitrarily given ε > 0 , we can find a constant C ensuring the following holds:
P { sup s = C Q n ( γ ) > Q n ( γ 0 ) } < 1 ε .
Let Δ ( γ ) = n K 1 [ Q n ( γ ) Q n ( γ 0 ) ] . By Taylor’s Formula, we can obtain
Δ ( γ ) = n K 1 [ Q n ( γ 0 + τ s ) Q n ( γ 0 ) ] = n K 1 τ s T Q n ( γ 0 ) + 1 2 n K 1 τ 2 s T Q n ( γ 0 ) s + 1 6 n K 1 τ 3 ( γ { s T Q n ( γ * ) s } ) T s = J 1 + J 2 + J 3 ,
where γ * lies between γ and γ 0 . Noting that K = O n 1 2 r + 1 , we can obtain the following through the assumption conditions and some calculations: J 1 = O p ( n τ K 1 ) s = o p ( 1 ) s , J 2 = O p ( n τ 2 K 1 ) s 2 = O p ( 1 ) s 2 , J 3 = O p ( n τ 3 K 1 ) s 3 = o p ( 1 ) s 3 . Thus, there is a sufficiently large constant C ensuring that when s = C , J 2 can control J 1 and J 3 ; therefore, Equation (A8) holds, and there further exists a maximum point γ ^ satisfying
γ ^ γ 0 = O p ( τ ) = O p ( n r 2 r + 1 ) ,
therefore, we get
α ^ k ( · ) α ^ k 0 ( · ) = O p ( n r 2 r + 1 ) , k = 1 , 2 , , p .

References

  1. Li, T.; Mei, C. Estimation and inference for varying coefficient partially nonlinear models. J. Stat. Plan. Inference 2013, 143, 2023–2037. [Google Scholar] [CrossRef]
  2. Yu, P.; Zhu, Z.; Shi, J.; Ai, X. Robust estimation for partial functional linear regression model based on modal regression. J. Syst. Sci. Complex. 2020, 33, 527–544. [Google Scholar] [CrossRef]
  3. Yan, L.; Tan, X.y.; Chen, X. Empirical likelihood for partially linear errors-in-variables models with longitudinal data. Acta Math. Appl. Sin. Engl. Ser. 2022, 38, 664–683. [Google Scholar] [CrossRef]
  4. Xiao, Y.; Liang, L. Robust estimation and variable selection for varying-coefficient partially nonlinear models based on modal regression. J. Korean Stat. Soc. 2022, 51, 692–715. [Google Scholar] [CrossRef]
  5. Zhao, P.; Zhou, X.; Wang, X.; Huang, X. A new orthogonality empirical likelihood for varying coefficient partially linear instrumental variable models with longitudinal data. Commun. Stat.-Simul. Comput. 2020, 49, 3328–3344. [Google Scholar] [CrossRef]
  6. Liang, K.Y.; Zeger, S.L. Longitudinal data analysis using generalized linear models. Biometrika 1986, 73, 13–22. [Google Scholar] [CrossRef]
  7. Xu, H.X.; Fan, G.L.; Liang, H.Y. Quantile regression for varying-coefficient partially nonlinear models with randomly truncated data. Stat. Pap. 2024, 65, 2567–2604. [Google Scholar] [CrossRef]
  8. Diggle, P.J. Analysis of Longitudinal Data; Oxford University Press: Oxford, UK, 2002. [Google Scholar]
  9. Qu, A.; Lindsay, B.G.; Li, B. Improving generalised estimating equations using quadratic inference functions. Biometrika 2000, 87, 823–836. [Google Scholar] [CrossRef]
  10. Bai, Y.; Fung, W.K.; Zhu, Z.Y. Penalized quadratic inference functions for single-index models with longitudinal data. J. Multivar. Anal. 2009, 100, 152–161. [Google Scholar] [CrossRef]
  11. Tian, R.; Xue, L.; Liu, C. Penalized quadratic inference functions for semiparametric varying coefficient partially linear models with longitudinal data. J. Multivar. Anal. 2014, 132, 94–110. [Google Scholar] [CrossRef]
  12. Schumaker, L. Spline Functions: Basic Theory; Wiley: Hoboken, NJ, USA, 1981. [Google Scholar]
  13. Wei, Y.; Wang, Q.; Liu, W. Model averaging for linear models with responses missing at random. Ann. Inst. Stat. Math. 2021, 73, 535–553. [Google Scholar] [CrossRef]
  14. Jiang, Y.; Ji, Q.; Xie, B. Robust estimation for the varying coefficient partially nonlinear models. J. Comput. Appl. Math. 2017, 326, 31–43. [Google Scholar] [CrossRef]
  15. Xiao, Y.T.; Chen, Z.S. Bias-corrected estimations in varying-coefficient partially nonlinear models with measurement error in the nonparametric part. J. Appl. Stat. 2018, 45, 586–603. [Google Scholar] [CrossRef]
  16. Yang, J.; Yang, H. Smooth-threshold estimating equations for varying coefficient partially nonlinear models based on orthogonality-projection method. J. Comput. Appl. Math. 2016, 302, 24–37. [Google Scholar] [CrossRef]
  17. Qian, Y.; Huang, Z. Statistical inference for a varying-coefficient partially nonlinear model with measurement errors. Stat. Methodol. 2016, 32, 122–130. [Google Scholar] [CrossRef]
  18. Wang, X.; Zhao, P.; Du, H. Statistical inferences for varying coefficient partially non linear model with missing covariates. Commun. Stat.-Theory Methods 2021, 50, 2599–2618. [Google Scholar] [CrossRef]
  19. Zhou, Y.; Mei, R.; Zhao, Y.; Hu, Z.; Zhao, M. Orthogonality-based bias-corrected empirical likelihood inference for partial linear varying coefficient EV models with longitudinal data. J. Comput. Appl. Math. 2024, 443, 115751. [Google Scholar] [CrossRef]
  20. Zhao, P.; Xue, L. Empirical likelihood inferences for semiparametric varying-coefficient partially linear errors-in-variables models with longitudinal data. J. Nonparametric Stat. 2009, 21, 907–923. [Google Scholar] [CrossRef]
Figure 1. The boxplots of 1000 RMSE values for β ^ 1 under the OQIF (A) and PNLS (B) methods.
Figure 1. The boxplots of 1000 RMSE values for β ^ 1 under the OQIF (A) and PNLS (B) methods.
Axioms 14 00875 g001
Figure 2. The boxplots of 1000 RMSE values for β ^ 2 under the OQIF (A) and PNLS (B) methods.
Figure 2. The boxplots of 1000 RMSE values for β ^ 2 under the OQIF (A) and PNLS (B) methods.
Axioms 14 00875 g002
Figure 3. The boxplots of 1000 RMSE for α ^ U .
Figure 3. The boxplots of 1000 RMSE for α ^ U .
Axioms 14 00875 g003
Table 1. The bias and standard deviation of β ^ 1 and β ^ 2 measured using distinct methods.
Table 1. The bias and standard deviation of β ^ 1 and β ^ 2 measured using distinct methods.
MethodParameter n = 100 n = 150 n = 200
BiasSDBiasSDBiasSD
OQIF β ^ 1 0.002570.020120.001230.018900.000790.01568
β ^ 2 −0.003900.03890−0.002350.03568−0.001570.02946
PNLS β ^ 1 0.014570.030120.012350.027890.009880.02346
β ^ 2 −0.027890.05890−0.024570.05235−0.020120.04457
Table 2. Confidence interval length and coverage probability comparison.
Table 2. Confidence interval length and coverage probability comparison.
MethodParameter n = 100 n = 150 n = 200
Length Coverage Length Coverage Length Coverage
OQIF β ^ 1 0.076710.9430.071900.9470.059710.951
β ^ 2 0.148530.9370.135510.9410.113290.948
PNLS β ^ 1 0.113290.9110.100650.9180.085410.925
β ^ 2 0.209950.8990.187730.9050.159460.912
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ge, J.; Zhou, X.; Wang, C. Estimation for Longitudinal Varying Coefficient Partially Nonlinear Models Based on QR Decomposition. Axioms 2025, 14, 875. https://doi.org/10.3390/axioms14120875

AMA Style

Ge J, Zhou X, Wang C. Estimation for Longitudinal Varying Coefficient Partially Nonlinear Models Based on QR Decomposition. Axioms. 2025; 14(12):875. https://doi.org/10.3390/axioms14120875

Chicago/Turabian Style

Ge, Jiangcui, Xiaoshuang Zhou, and Cuiping Wang. 2025. "Estimation for Longitudinal Varying Coefficient Partially Nonlinear Models Based on QR Decomposition" Axioms 14, no. 12: 875. https://doi.org/10.3390/axioms14120875

APA Style

Ge, J., Zhou, X., & Wang, C. (2025). Estimation for Longitudinal Varying Coefficient Partially Nonlinear Models Based on QR Decomposition. Axioms, 14(12), 875. https://doi.org/10.3390/axioms14120875

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop