Next Article in Journal
A Forgotten Effects Approach to the Analysis of Complex Economic Systems: Identifying Indirect Effects on Trade Networks
Previous Article in Journal
Scale Mixture of Maxwell-Boltzmann Distribution
Previous Article in Special Issue
Equivalence Analysis of Statistical Inference Results under True and Misspecified Multivariate Linear Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multicollinearity and Linear Predictor Link Function Problems in Regression Modelling of Longitudinal Data

1
Department of Statistics, Faculty of Mathematical Sciences, Ferdowsi University of Mashhad, Mashhad 9177948974, Iran
2
Department of Statistics, Faculty of Natural and Agricultural Sciences, University of Pretoria, Pretoria 0028, South Africa
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(3), 530; https://doi.org/10.3390/math11030530
Submission received: 1 October 2022 / Revised: 13 December 2022 / Accepted: 26 December 2022 / Published: 18 January 2023
(This article belongs to the Special Issue Contemporary Contributions to Statistical Modelling and Data Science)

Abstract

:
In the longitudinal data analysis we integrate flexible linear predictor link function and high-correlated predictor variables. Our approach uses B-splines for non-parametric part in the linear predictor component. A generalized estimation equation is used to estimate the parameters of the proposed model. We assess the performance of our proposed model using simulations and an application to an analysis of acquired immunodeficiency syndrome data set.

1. Introduction

Generalized linear models (GLMs) [1], which have a link function connecting the predictors linearly, are now part of regression models toolbox. Assuming a linear predictor link function could be very restrictive as the true relationship could be non-linear. Generalized partial linear models (GPLMs) accommodates both parametric and nonparametric connections. For example, in modelling longitudinal data, the GPLMs have proved useful as the model account for possible dependencies in the data [2,3,4,5,6].
One of the basic assumptions in the regression analysis is that all the explanatory variables are linearly independent. However two or more of the explanatory variables could be correlated with one another, resulting into a multicollinearity problem. Thus, it becomes harder to distinguish between effects of the independent variables on the outcome variable. It also results in the inflation of the variance of the regression parameter estimates. Ridge regression is widely used in regression model analyses with a large number of highly correlated independent variables [7,8,9,10,11,12,13,14,15,16,17]. While ridge regression techniques have widely been used in modelling cross-section data, there have been very few studies and applications in longitdinal data [18,19,20,21,22].
In this paper, we consider two typical problems name: of multicollinearity among predictor variables and linear predictor link function in the analysis of longitudinal non-normal data. For the former, we employ ridge regression and for the later we adopt the use of B-splines for nonparametric component of the linear predictor in an integrated approach. We concentrate on the estimation of population averaged model parameters. For this, we have the marginal mean model specification and account for the possible dependencies in the longitudinal data in a nonparametric manner through a convenient working within-subject covariance and employ the generalised estimation equations (GEE) for estimating the parameters.
In Section 2, we specify the underlying longitudinal model with the nonparametric part using splines. The estimation and asymptotic properties of the model parameters are also presented in Section 2. Simulation studies and an application to acquired immunodeficiency syndrome data set are in Section 3. We conclude the paper in Section 4.

2. Model and Estimation Procedure

2.1. GPLMs for Longitudinal Data

Suppose we have n subjects and subject i has n i observations denoted by y i j ( i = 1 , , n ,   j = 1 , , n i ) for a total of N = i = 1 n n i observations. Also, X i j be a vector of time-varying covariate. Thus, the total observed data set for the analysis is { ( X i j , y i j , t i j ) , i = 1 , , n ,   j = 1 , , n i } . Further let E ( y i j ) = μ i j and Var ( y i j ) = ϕ ν ( μ i j ) , where ϕ is a scale parameter and ν ( . ) is a known variance function. We model the longitudinal data with a GPLM, and specify a marginal model on the first two moments of y i j . Especially, the marginal mean μ i j is modeled as
η i j = g ( μ i j ) = X i j β + f ( t i j ) ; μ i j = g 1 ( η i j ) ,
where g(.) is a link function for the GLM, β is the regression coefficient vector with dimension p, and f ( . ) is an unknown smooth function. We also assume independency between observations from different subjects. Finally, we assume t i j are all scaled into the interval [ 0 , 1 ] .
Similar to [3,23,24], we approximate the unspecified smooth function by the following polynomial spline
f ( t i j ) = α 0 + α 1 t i j + + α d t i j d + l = 1 L n α ( d + 1 ) + l ( t i j t i ( l ) ) + d = B ( t i j ) α ,
where d is the degree of the polynomial component, L n is the number of interior knots (rate of L n will be specified in Remark 1), t i ( l ) are knots of the ith subject, B ( t i j ) = 1 , t i j , , t i j d , t i j t i ( 1 ) + d , , t i j t i ( L n ) + d is a h n × 1 vector of basis functions, h n is the number of basis functions used to approximate f ( t i j ) , h n = d + 1 + L n , ( a ) + = max ( 0 , a ) , and α n = ( α 0 , , α d , α d + 1 , , α d + 1 + L n ) is the spline coefficients vector of dimension h n . The nonparametric part in the linear prediction part is set as the basis functions with pseudo-design variables. In this way, the regression model problem in (1) could be linearised
η i j ( θ ) = g ( μ i j ( θ ) ) = X i j β + B ( t i j ) α = D i j θ .
D i j = X i j , B ( t i j ) is a ( p + h n ) × 1 design matrix, which combines both the fixed and spline effects for the jth outcome of the ith subject. The combined regression coeffieicnts θ = ( β , α ) has dimension ( p + h n ) × 1 . Let μ i = ( μ i 1 , , μ i n i ) , Y i = ( y i 1 , , y i n i ) , where μ i j = g 1 ( D i j θ ) , and D i = ( X i , B ( t i ) ) . By the linear form of the GPLM in (1), using the spline approach, any computational algorithm developed for the GLM can be used for the GPLM.
Remark 1.
In spline smoothing, it is important to select the knots efficiently. Concentrating on the estimation of β, Ref. [3] noticed that knot selection is more important for estimating f ( · ) rather than β. Because in most of the studies, the focus is on β and providing sufficient statistical inference, and one only needs some basic information about f ( . ) . Therefore they particularly used the sample quantiles of { t i j , i = 1 , , n ; k = 1 , , n i } as knots. For instance, with three internal knots, we take three quartiles of the observed t i j . Considering splines of order 4, they applied cubic splines with the integer part of M 1 / 5 , the number of internal knots, where M is the number of distinct values in t i j . Another study, Ref. [25], proposed that the number of distinct knots should increase with sample size to achieve asymptotic consistency. One must note that having too many knots increases the variance of estimators. Thus, the number of knots must be appropriately selected. When n goes to ∞, the number of knots should increase at n 1 / ( 2 m + 1 ) . Thus, here, we use L n n 1 / ( 2 m + 1 ) with integer value m as the number of internal knots. We fix m = 2 , and choose L n n 1 / 5 , for asymptotic consistency. However, it is mainly based on practical experience and a desire for simplicity and is not an optimal choice. We considered a similar procedure in simulation and real data.

2.2. Ridge Generalized Estimating Equation (RGEE)

In most applications of GPLMs, the primary research interest is to make statistical inferences on the regression coefficient θ , along with and understanding of basic features of f ( t ) . For the ridge GEE procedure for β , we briefly review the GEE method. For the estimating equation of θ , we have
n 1 i = 1 n μ i j ( θ ) θ V i 1 ( Y i μ i ) = 0 ,
where V i is a covariance matrix of Y i . For most applications, the actual intracluster covariance is regularly unknown. We take the working correlation matrix as R ( τ ) : V i = A i 1 / 2 ( θ ) R ( τ ) A i 1 / 2 ( θ ) , with the finite-dimensional parameter τ . Some commonly used working correlation structures include independence, autocorrelation (AR)-1, equally correlated (also called compound symmetry), or unstructured correlation. For a given working correlation structure, τ can be estimated using the residual-based moment method. Here, similar to [26], the marginal density of Y i j follows a canonical exponential family. Consequently, μ i j ( θ ) = a ( θ i j ) and σ i j 2 ( θ ) = ϕ a ( θ i j ) , where θ i j = D i j θ , for a differentiable function a ( · ) and a scaling constant ϕ . Assume R ^ is the estimated working correlation matrix. Then, (3) simplifies to
1 n i = 1 n D i A i 1 / 2 ( β ) R ^ 1 A i 1 / 2 ( θ ) Y i μ i ( θ ) = 0 .
We formally define the GEE estimator as the solution β ^ of the above-estimating equations. For ease of exposition, we assume ϕ = 1 in the rest of the article.
To account for multicollinearity in longitudinal data, we use the ridge GEE in the GPLM in (1) for parameter estimation. We do this by adding a shrinkage term λ β β to the objective function for handling correlated predictors. The ridge GEE has form
U ( θ ) = S ( θ ) λ β ,
where
S ( θ ) = 1 n i = 1 n D i A i 1 / 2 ( θ ) R ^ 1 A i 1 / 2 ( θ ) Y i μ i ( θ ) ,
are the estimating functions defining the GEE. Here, λ is the tuning parameter that determines the shrinkage amount. The RGEE estimator β ^ R is the solution to U ( θ ) = 0 . We use the Newton–Raphson algorithm along with (6) to get the following iterative algorithm
θ ^ k + 1 = θ ^ k + H ( θ ^ k ) + n λ E ( θ ^ k ) 1 × S ( θ ^ k ) n λ E ( θ ^ ( k ) θ ^ k ) .
Here, H ( θ ^ k ) = n 1 i = 1 n D i A i 1 / 2 ( θ ) R ^ 1 A i 1 / 2 ( θ ) D i , E ( θ ^ k ) = diag 1 p , 0 N k . Further, 1 p and 0 N k represent a vector of 1 with dimension p, and a zero vector of dimension N k , respectively. The suggested estimation approach can be implemented step by step, and the detailed computation procedure can be summarized in Algorithm 1, describing the combination of ridge regression into the Newton–Raphson iterative algorithm of GEE. With prespecified λ and initial value β , the above algorithm is repeated to update β ^ k + 1 untill convergence.
Algorithm 1: Monte Carlo Newton–Raphson (MCNR) algorithm
  • Step 1. Approximate each predictor trajectory f ( t i j ) , by regression splines technique where described in Section 2.1. The smoothed predictor trajectories are then denoted as f ( t i j ) = π ( t i j ) α .
  • Step 2. Set k = 0 . Choose initial values for parameter space θ 0 = ( β 0 , α 0 ) and correlation parameter τ 0 where determine covariance matrix V i 0 = A i 1 / 2 ( θ 0 ) R ( τ 0 ) A i 1 / 2 ( θ 0 ) . Use these values to find the RGEE estimates of θ . Specially,
    (a)
    Compute θ ( k + 1 ) from the expression
    θ k + 1 = θ k + H ( θ k ) + n λ E ( θ k ) 1 × S ( θ k ) n λ E ( θ ( k ) θ k ) ,
    where
    H n θ k = n 1 i = 1 n D i A i 1 2 ( θ k ) R 1 ( τ k ) A i 1 2 ( θ k ) D i , S n θ k = i = 1 n D i A i 1 2 ( θ k ) R ^ 1 ( τ k ) A i 1 2 ( θ k ) ) y i μ i ( θ k ) .
    (b)
    Compute V i k + 1 by A i 1 / 2 ( θ k + 1 ) R ( τ k + 1 ) A i 1 / 2 ( θ k + 1 ) where τ k + 1 For a given working correlation structure, τ can be estimated using the residual-based moment method according to the prespecified working correlation structure. For more details refer to
    (c)
    Set k = k + 1 .
  • Step 3. Go to step 2 until convergence is achieved. Choose θ k + 1 and V i k + 1 to be the RGEE estimates of θ and V i .
It is critical to choose a suitable value of tuning parameter λ to achieve satisfactory performance of the selection procedure. Many authors have introduced many methods for choosing an optimal tuning parameter within a given set of candidates. Traditional model selection criteria, such as AIC and BIC, have several limitations. The generalized cross validation (GCV) suggested by [27], later [28] proposed a Bayesian information criterion (BIC). How to choose λ for high-dimensional data discussed by [29]. They proposed a modified BIC. Further [30] extended the BIC information criterion. For the selection of the tuning parameter λ , here, we apply the suggested GCV of [27], given by
GCV λ = 1 n RSS ( λ n ) 1 1 n d ( λ n ) 2 ,
where
RSS ( λ ) = 1 n i = 1 n y i μ i ( θ ^ ) V ^ i 1 y i μ i ( θ ^ ) ,
is the residual sum of squares, and effective number is equal to
d ( λ ) = tr [ { H ( θ ^ ) + n E ( θ ^ ) } 1 × H n θ ^ ] .
The optimal parameter denoted by λ o p t is the minimizer of the GCV λ . In practical implementation, one can use PGEE package of R software, where function CVfit computes cross-validated tuning parameter value for longitudinal data. In numerical studies of the current paper, we used R codes similarly to compute λ o p t .

2.3. Asymptotics

We now discuss the asymptotic properties of the estimators β and f ( . ) for the ridge GEE. Assuming (A.1)–(A.4) in Appendix A, the following theorems state the large sample property for f ( . ) and β , respectively. For the proofs, refer to Appendix A.
Theorem 1.
For k n = n 1 / ( 2 m + 1 ) we have
N 1 i , j f ^ ( t i j ) f ( t i j ) 2 = O p ( n ν ) ,
where i , j = i = 1 n j = 1 n i and ν = 2 m / ( 2 m + 1 ) .
Theorem 2.
Under the named regularity Conditions (A.1)–(A.4), n β ^ β + b ( β ) has the asymptotic p-variate normal distribution with the zero-vector mean and covariate matrix V , where
b ( β ) = λ H 1 ( β ) β , V = R ( λ ) H 1 ( β ) R ( λ ) , H ( β ) = n 1 i = 1 n X i A i 1 / 2 ( β ) R ^ 1 A i 1 / 2 ( β ) X i , R ( λ ) = 1 p + λ H 1 ( β ) 1 .

3. Numerical Analyses

3.1. Simulations

Here, we assess the performance of the GEE compared to its counterpart, the ridge GEE, for multiple correlated predictor variables. We generate the explanatory variables using
x i k = ( 1 γ 2 ) 1 2 ω i k + γ 2 ω i p , i = 1 , , n = 100 , k = 1 , , p 1 , p = 5 ,
where ω i k are assumed to be independent and generated from a normal distribution with zero mean and unit variance. The parameter γ reflects the correlation such that any two explanatory variables correlate equally to γ 2 . We generality, we consider γ { 0.70 , 0.80 , 0.90 , 0.99 } . The nonparametric part of (1) has form f ( t i j ) = 2 s i n ( 2 π t i j ) . In the entire process, n i = 4 for each subject i and use the following GPLM for simulation
y i j = x i , 1 , j β 1 + x i , 2 , j β 2 + x i , 3 , j β 3 + x i , 4 , j β 4 + x i , 5 , j β 5 + f ( t i j ) + ε i j .
We set the true β as β T = ( 0.5 , 1 , 1.5 , 2 , 0.1 , 0.2 ) . We generate t i j from the uniform distribution over ( 0 , 1 ) . The ε i j is generated from a normal distribution with zero mean, a common marginal variance σ 2 = 1 . Moreover, the correlation structure is AR ( 1 ) , i.e., corr ( ε i s , ε i t ) = ρ | t s | for s t , ρ = 0.9 . Each simulated data set is fitted separately by the GEE approach of [26] and our ridge GEE using Algorithm 1. Then, 200 replications are run for each combination of ρ and γ .
To assess the behavior of both estimators encountering misspecified correlation structures, we conduct a comparison between β ^ G E E and β ^ R G E E . We test the exchangeable working correlation structure (GEE-I) or (RGEE-I) when the true correlation structure is AR(1), (GEE-C), or (RGEE-C). For each of the estimators, we measure the accuracy in estimation using the mean squared error (MSE) given by MSE = ( β ^ β ) ( β ^ β ) . We recall the tuning parameter λ was obtained using the GCV, where λ o p t was the minimizer of the G C V λ . Alternatively, one can use the ridge trace to find λ o p t . Figure 1 illustrates the ridge trace for the first generated random sample. As can be seen, the minimizer occurs at K = 0.05 , which is the same value as the minimizer of GCV. The simulation results for MSE are presented in the latest column of Table 1. Table 1 reports the empirical biases and standard deviations (SDs) of the estimated β from the GEE and RGEE methods. We can take the following observations: Our proposed RGEE has a superior performance in the MSE criterion. For misspecified correlation structures, the RGEE outperforms. However, it has more bias and offers a smaller SD in most cases compared to the GEE. By increasing the correlation among predictors, the increase in RGEE MSE is lesser than the GEE for all considered criteria. The conclusion is evident at the extreme level of correlation γ = 0.99 .

3.2. AIDS Data Analysis

For illustration, in this section, the proposed model is used to analyze the CD4 cell data. From the number of 369 patients, 2376 CD4 measurements are recorded. The population’s average time course of CD4 decay is regressed on the following covariates: packs per day for an indication of smoking; binary variable recreational drug use; SEXP as an indication of the number of sexual partners; and depression symptoms as measured by the CESD scale (larger values indicate increased depressive symptoms). Similar to most literature, we take the square root of CD4 numbers. For the reason of the latter transformation, the reader is referred to [31,32]. For the correlation structure, we follow the approach of [31] and fit the compound symmetry covariance, where ρ = 0.509 . We then used our proposed RGEE compared with the GEE for parameter estimation. We computed the standard errors (SDs) calculated using the bootstrap method. Table 2 provides the parameter estimates. From the result of this table, our proposed RGEE gives effectively smaller SD values compared to the GEE.

4. Concluding Remarks and Discussion

We considered a generalized partially linear model (GPLM) and ridge regression to tackle the problems of multicollinearity and non-linearity in the relationship between the mean response and covariates in the longitudinal data analysis. The generalized estimation Equation (GEE) methods were used to estimate parameters in our proposed model. Using simulation studies, our methods resulted in smaller biases for estimating parameters than could have been obtained in a standard GEE. The performance of our proposed method decreased with increased dependencies between the model predictors. We also applied our model to a specific data set on AIDS data analysis.

Author Contributions

Conceptualization, M.T., M.A. and S.M.; Funding acquisition, M.A. and S.M.; Methodology, M.T. and M.A.; Software, M.T.; Supervision, M.A. and S.M.; Visualization, M.T., M.A. and S.M.; Formal analysis, M.T. and M.A.; Writing—original draft preparation, M.T.; Writing—review and editing, M.T., M.A. and S.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was based upon research supported in part by the National Research Foundation (NRF) of South Africa, SARChI Research Chair UID: 71199, the South African DST-NRF-MRC SARChI Research Chair in Biostatistics (Grant No. 114613) and STATOMET at the Department of Statistics at the University of Pretoria, South Africa. The opinions expressed and conclusions arrived at are those of the authors and are not necessarily to be attributed to the NRF.

Data Availability Statement

The data is publicly available.

Acknowledgments

We sincerely thank two anonymous reviewers for their constructive comments that significantly improved the presentation and led to putting many details in the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

To study the asymptotic properties of RGEE estimators, the following regularity conditions are required.
(A.1)
Number of observations over time ( n i ) is a bounded sequence of positive integers, and the distinct values of t i j form a quasi-uniform sequence that grows dense on [ 0 , 1 ] , and the kth derivative of f ( t i j ) is bounded for some k 2 ;
(A.2)
The covariates X i j , 1 i n , 1 j m are uniformly bounded;
(A.3)
The unknown parameter β belongs to a compact subset B R p , the true parameter value lies in the interior of B ;
(A.4)
There exist two positive constants, b 1 and b 2 , such that
b 1 λ min n 1 i = 1 n X i X i λ max n 1 i = 1 n X i X i b 2 ,
where λ min (resp. λ max ) denotes the minimum (resp. maximum) eigenvalue of a matrix.
To verify Theorem 1, we need the following lemma.
Lemma A1.
Under Condition (A.1), there exists a constant C depending only on k n such that
sup t 0 , 1 | f ( t ) π ( t ) α | C k n m .
The proof of this lemma follows readily from Theorem 12.7 of Schumaker [33].
Proof of Theorem 1.
By Lemma A1, we approximate f ( t i j ) by π ( t ) α , then by choosing k n n 1 / ( 2 m + 1 ) we have
f ^ ( t ) f ( t ) 2 = | f ^ ( t ) f ( t ) | | f ^ ( t ) f ( t ) | sup t 0 , 1 | f ^ ( t ) f ( t ) | sup t 0 , 1 | f ^ ( t ) f ( t ) | C k n m C k n m = C 2 n 2 m / ( 2 m + 1 ) = O p ( n 2 m / ( 2 m + 1 ) ) ,
which proves Theorem 1. □
Proof of Theorem 2.
To proof Theorem 2 define the linear operator R ( λ ) = 1 p + λ H 1 ( β ) 1 . It is straightforward to calculate that the ridge estimator β ^ R G E E can be expressed as R ( λ ) β ^ G E E where β ^ G E E is ordinary GEE estimator. The expectation of the ridge estimator can be expressed as
E β ^ R G E E = E R ( λ ) β ^ G E E = β λ H 1 ( β ) β .
Clearly, E β ^ R G E E β = λ H 1 ( β ) β 0 for any λ > 0 . Hence, the ridge estimator is biased with λ H 1 ( β ) β = b ( β ) . The variance of the RGEE estimator is straightforwardly obtained when exploiting its linearly relation with the GEE estimator. Then,
Var β ^ R G E E = Var R ( λ ) β ^ G E E = R ( λ ) Var β ^ G E E R ( λ ) = R ( λ ) H 1 ( β ) R ( λ ) ,
where R ( λ ) H 1 ( β ) R ( λ ) = V . Combining the expectation and variance terms, the proof is complete. □

References

  1. McCullagh, P.; Nelder, J.A. Generalized Linear Models, 2nd ed.; Chapman and Hall: London, UK, 1989. [Google Scholar]
  2. He, X.; Zhu, Z.Y.; Fung, W.K. Estimation in a Semiparametric model for longitudinal data with unspecified dependence structure. Biometrika 2002, 89, 579–590. [Google Scholar] [CrossRef]
  3. He, X.M.; Fung, W.K.; Zhu, Z.Y. Robust estimation in a generalized partially linear model for cluster data. J. Am. Stat. Assoc. 2005, 100, 1176–1184. [Google Scholar] [CrossRef]
  4. Qin, G.; Bai, Y.; Zhu, Z. Robust empirical likelihood inference for generalized partial linear models with longitudinal data. J. Multivar. Anal. 2012, 105, 32–44. [Google Scholar] [CrossRef] [Green Version]
  5. Chen, B.; Zhou, X.H. Generalized partially linear models for incomplete longitudinal data in the presence of population-Level information. Biometrics 2013, 69, 386–395. [Google Scholar] [CrossRef] [Green Version]
  6. Zhang, J.; Xue, L. Empirical likelihood inference for generalized partially linear models with longitudinal data. Open J. Stat. 2020, 10, 188–202. [Google Scholar] [CrossRef] [Green Version]
  7. Hoerl, A.; Kennard, R. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  8. Hoerl, A.; Kennard, R. Ridge regression: Application to nonorthogonal problems. Technometrics 1970, 12, 69–82. [Google Scholar] [CrossRef]
  9. Theobald, C.M. Generalization of mean squer error applied to ridge regresion. J. R. Stat. Soc. 1974, 36, 103–106. [Google Scholar]
  10. Tikhonov, A. On the stability of inverse problems. Proc. USSR Acad. Sci. 1943, 39, 267–288. [Google Scholar]
  11. Saleh, A.K.M.; Kibria, B.M.G. Performances of some new preliminary test ridge regression estimators and their properties. Commun. Stat.—Theory Methods 1993, 22, 2747–2764. [Google Scholar] [CrossRef]
  12. Kibria, B.M.G.; Saleh, A.K.M.E. Effect of W,LR and LM tests on the performance of preliminary test ridge regression estimators. J. Jpn. Stat. Soc. 2003, 33, 119–136. [Google Scholar] [CrossRef] [Green Version]
  13. Kibria, B.M.G.; Saleh, A.K.M.E. Preliminary test ridge regression estimators with student’s /errors and conflicting test-statistics. Metrika 2004, 59, 105–124. [Google Scholar] [CrossRef]
  14. Arashi, M.; Tabatabaey, S.M.M.; Iranmanesh, A. Improved estimation in stochastic linear models under elliptical symmetry. J. Appl. Probab. Stat. 2010, 5, 145–160. [Google Scholar]
  15. Bashtian, H.M.; Arashi, M.; Tabatabaey, S.M.M. Using improved estimation strategies to combat multicollinearity. J. Stat. Comput. Simul. 2011, 81, 1773–1797. [Google Scholar] [CrossRef]
  16. Bashtian, H.M.; Arashi, M.; Tabatabaey, S.M.M. Ridge estimation under the stochastic restriction. Commun. Stat.—Theory Methods 2011, 40, 3711–3747. [Google Scholar] [CrossRef]
  17. Arashi, M.; Tabatabaey, S.M.M.; Soleimani, H. Simple regression in view of elliptical models. Linear Algebra Its Appl. 2012, 437, 1675–1691. [Google Scholar] [CrossRef] [Green Version]
  18. Zhang, B.; Horvath, S. Ridge regression based hybrid genetic algorithms for multi-locus quantitative trait mapping. Bioinform. Res. Appl. 2006, 1, 261–272. [Google Scholar] [CrossRef]
  19. Malo, N.; Libiger, O.; Schork, N. Accommodating linkage disequilibrium in genetic-association analyses via ridge regression. Am. J. Hum. Genet. 2008, 82, 375–385. [Google Scholar] [CrossRef] [Green Version]
  20. Eliot, M.; Ferguson, J.; Reilly, M.P.; Foulkes, A.S. Ridge regression for longitudinal biomarker data. Int. J. Biostat. 2011, 7, 37. [Google Scholar] [CrossRef]
  21. Rahmani, M.; Arashi, M.; Mamode Khan, N.; Sunecher, Y. Improved mixed model for longitudinal data analysis using shrinkage method. Math. Sci. 2018, 12, 305–312. [Google Scholar] [CrossRef] [Green Version]
  22. Taavoni, M.; Arashi, M. Semiparametric ridge regression for longitudinal data. In Proceedings of the 14th Iranian Statistics Conference, Shahrood University of Technology, Shahrood, Iran, 25–27 August 2018. [Google Scholar]
  23. Qin, G.Y.; Zhu, Z.Y. Robustified maximum likelihood estimation in generalized partial linear mixed model for longitudinal data. Biometrics 2009, 65, 52–59. [Google Scholar] [CrossRef] [PubMed]
  24. Taavoni, M.; Arashi, M. High-dimensional generalized semiparametric model for longitudinal data. Statistics 2021, 55, 831–850. [Google Scholar] [CrossRef]
  25. Qin, G.Y.; Zhu, Z.Y. Robust estimation in generalized semiparametric mixed models for longitudinal data. J. Multivar. Anal. 2007, 98, 1658–1683. [Google Scholar] [CrossRef]
  26. Liang, K.Y.; Zeger, S.L. Longitudinal data analysis using generalized linear models. Biometrika 1986, 73, 13–22. [Google Scholar] [CrossRef]
  27. Fan, J.Q.; Li, R. Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
  28. Wang, H.S.; Li, R.Z.; Tcai, C.L. Tuning parameter selectors for the smoothly clipped absolute deviation method. Biometrika 2007, 94, 553–568. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Wang, H.S.; Li, B.; Leng, C.L. Shrinkage tuning parameter selection with a diverging number of parameters. J. R. Stat. Soc. Ser. B 2009, 71, 671–683. [Google Scholar] [CrossRef]
  30. Li, G.R.; Peng, H.; Zhu, L.X. Nonconcave penalized M-estimation with a diverging number of parameters. Stat. Sin. 2011, 21, 391–419. [Google Scholar]
  31. Zeger, S.L.; Diggle, P.J. Semi-parametric models for longitudinal data with application to CD4 cell numbers in HIV seroconverters. Biometrics 1994, 50, 689–699. [Google Scholar] [CrossRef]
  32. Wang, N.; Carroll, R.; Lin, X.H. Efficient semiparametric marginal estimation for longitudinal/clustered data. J. Am. Stat. Assoc. 2005, 100, 147–157. [Google Scholar] [CrossRef]
  33. Schumaker, L.L. Spline Functions; Wiley: New York, NY, USA, 1981. [Google Scholar]
Figure 1. The ridge trace plot, for the first simulated data.
Figure 1. The ridge trace plot, for the first simulated data.
Mathematics 11 00530 g001
Table 1. Estimated regression coefficients for the important variables; bias (SD) based on 200 replications.
Table 1. Estimated regression coefficients for the important variables; bias (SD) based on 200 replications.
MethodsParameters γ = 0.70 γ = 0.80 γ = 0.90 γ = 0.99
RGEE-C β 1 0.028(0.032)0.033(0.038)0.046(0.052)0.141(0.142)
β 2 0.024(0.027)0.029(0.032)0.039(0.044)0.122(0.126)
β 3 0.056(0.027)0.067(0.032)0.092(0.044)0.284(0.127)
β 4 0.055(0.029)0.066(0.035)0.090(0.048)0.279(0.137)
β 5 0.080(0.035)0.097(0.045)0.137(0.068)0.446(0.209)
MSE0.2430.2910.4041.271
RGEE-I β 1 0.041(0.033)0.049(0.039)0.068(0.053)0.209(0.150)
β 2 0.057(0.029)0.068(0.034)0.093(0.047)0.289(0.137)
β 3 0.055(0.029)0.065(0.034)0.090(0.047)0.278(0.137)
β 4 0.061(0.031)0.073(0.037)0.101(0.051)0.311(0.146)
β 5 0.070(0.037)0.076(0.048)0.092(0.073)0.249(0.236)
MSE0.2840.3310.4441.335
GEE-C β 1 0.027(0.055)0.033(0.066)0.045(0.090)0.129(0.279)
β 2 0.024(0.044)0.028(0.052)0.039(0.072)0.113(0.223)
β 3 0.058(0.051)0.069(0.061)0.096(0.084)0.308(0.259)
β 4 0.055(0.050)0.065(0.060)0.089(0.082)0.244(0.254)
β 5 0.082(0.053)0.100(0.068)0.143(0.103)0.449(0.391)
MSE0.2460.2950.4111.243
GEE-I β 1 0.040(0.057)0.048(0.068)0.065(0.094)0.184(0.289)
β 2 0.055(0.048)0.065(0.057)0.088(0.078)0.237(0.242)
β 3 0.057(0.052)0.069(0.062)0.096(0.085)0.320(0.262)
β 4 0.061(0.056)0.072(0.067)0.099(0.092)0.271(0.284)
β 5 0.072(0.055)0.079(0.071)0.099(0.110)0.262(0.423)
MSE0.2860.3330.4471.275
Table 2. Regression coefficient estimates (SD) in the analysis of the CD4 data.
Table 2. Regression coefficient estimates (SD) in the analysis of the CD4 data.
CoefficientsMethodsCoefficientsMethods
RGEEGEERGEEGEE
AGE3.987 (0.006)4.298 (0.009)AGE*CESD−0.268 (0.008)−0.262 (0.001)
SMOKE32.780 (0.053)32.916 (0.062)SMOKE*DRUG−16.204 (0.046)−16.221 (0.055)
DRUG17.949 (0.066)18.254 (0.075)SMOKE*SEXP4.051 (0.002)4.057 (0.005)
SEXP2.801 (0.009)2.797 (0.013)SMOKE*CESD−0.268 (0.003)−0.251 (0.002)
CESD−3.077 (0.002)−3.077 (0.005)DRUG*SEXP−1.205 (0.005)−1.292 (0.013)
AGE*SMOKE0.039 (0.002)−0.007 (0.003)DRUG*CESD0.274 (0.003)0.273 (0.005)
AGE*DRUG−1.006 (0.003)−1.017 (0.009)SEXP*CESD0.033 (0.008)0.026 (0.001)
AGE*SEXP−0.565 (0.003)−0.596 (0.001)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Taavoni, M.; Arashi, M.; Manda, S. Multicollinearity and Linear Predictor Link Function Problems in Regression Modelling of Longitudinal Data. Mathematics 2023, 11, 530. https://doi.org/10.3390/math11030530

AMA Style

Taavoni M, Arashi M, Manda S. Multicollinearity and Linear Predictor Link Function Problems in Regression Modelling of Longitudinal Data. Mathematics. 2023; 11(3):530. https://doi.org/10.3390/math11030530

Chicago/Turabian Style

Taavoni, Mozhgan, Mohammad Arashi, and Samuel Manda. 2023. "Multicollinearity and Linear Predictor Link Function Problems in Regression Modelling of Longitudinal Data" Mathematics 11, no. 3: 530. https://doi.org/10.3390/math11030530

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop