Abstract
Variable selection is often needed in many fields and has been discussed by many authors in various situations. This is especially the case under linear models and when one observes complete data. Among others, one common situation where variable selection is required is to identify important risk factors from a large number of covariates. In this paper, we consider the problem when one observes interval-censored failure time data arising from generalized linear models, for which there does not seem to exist an established method. To address this, we propose a penalized least squares method with the use of an unbiased transformation and the oracle property of the method is established along with the asymptotic normality of the resulting estimators of regression parameters. Simulation studies were conducted and demonstrated that the proposed method performed well for practical situations. In addition, the method was applied to a motivating example about children’s mortality data of Nigeria.
Keywords:
interval-censored data; unbiased transformation; linear model; variable selection; large sample properties MSC:
62J12; 62N02; 62J07; 62G20
1. Introduction
Variable selection is an important field in statistics, and there is a lot of literature on variable selection, especially in the context of linear models for complete data, such as stepwise regression, ridge regression, Bayesian variable selection, least absolute shrinkage and selection operator (LASSO), model averaging, smoothly clipped absolute deviation (SCAD), elastic net, adaptive LASSO (ALASSO), minimax concave penalty (MCP), seamless- (SELO) and broken adaptive ridge (BAR) (Goldberger et al., 1961 [1]; Hoerl and Kennard, 1970 [2]; Mitchell and Beauchamp, 1988 [3]; Tibshirani, 1996 [4]; Raftery et al., 1997 [5]; Zou and Hastie, 2005 [6]; Zou, 2006 [7]; Fan and Li, 2001 [8]; Zhang, 2010 [9]; Dicker et al., 2013 [10]; Liu and Li, 2016 [11]; Dai et al., 2018 [12]; Zheng et al., 2021 [13]). Among others, one type of methods that have recently attracted a lot of attention is the penalized method, for which various penalty functions have been proposed. LASSO by Tibshirani (1996) [4], SCAD by Fan and Li (2001) [8], elastic net by Zou and Hastie (2005), ALASSO by Zou (2006) [7], MCP by Zhang (2010) [9], SELO by Dicker et al. (2013) [10], and BAR by Liu and Li (2016) [11] are included for penalty functions.
Variable selection has been investigated by many authors for incomplete data such as right-censored and interval-censored failure time data (Cai et al., 2005 [14]; Fan and Li, 2002 [15]; Tibshirani, 1997 [16]; Zhang and Lu, 2007 [17]; Nasrullah Khan et al., 2018 [18]; Zhao et al., 2020 [19]; Li et al., 2020 [20]; Du and Sun 2021 [21]; Ali et al., 2021 [22]). By interval censored data, we mean that the failure time of interest is known or observed only to belong to an interval instead of being observed exactly. It is easy to see that this is usually the case for periodic follow-up studies such as clinical trials and they include right-censored failure time data as a special case. Notably, the analysis of interval-censored data is much more challenging than the analysis of right-censored data. In terms of the well-known Cox model, the classical partial likelihood method for right-censored data is no longer available with interval censoring. This is because for the case of interval censoring, we have to estimate not only regression parameters of interest but also the nuisance parameters simultaneously. Recently, many authors have discussed the analysis of interval-censored data in various situations. For example, Zhao et al. (2015) [23], Wang et al. (2016) [24] and Li et al. (2018) [25] studied inference procedures for the Cox model, the additive hazard model, and the linear transformation model, respectively. Sun (2006) [26] provided a relatively complete review of the literature for the interval-censored failure time data analysis.
As mentioned above, several authors have discussed variable selection for interval-censored failure time data ([19,20]). However, existing methods cannot be applied directly to linear models or generalized linear models. In the variable selection procedure proposed below, we employ the unbiased transformation approach and the main idea behind it is to transfer two variables representing interval-censored data to a new, single variable that has the same conditional expectation. One of the early applications of the approach was given by Buckley and James (1979) [27] in dealing with right-censored data. Among others, one advantage of the proposed method over the existing methods is that it can be relatively easily implemented since one can make use of the existing variable selection programs for complete data with simple modifications.
Deng (2004) [28] and Deng et al. (2012) [29] discussed the use of the unbiased transformation approach for the analysis of interval-censored data. In particular, the latter considered the situation similar to that discussed here, but not on variable selection, under the assumption that the joint density function of the two variables representing the observed data is known. It is easy to see that this may not be true in many applications. To address this, we will adopt the kernel estimation approach [30,31] to estimate the needed density function and develop a unified approach to variable selection with different penalty functions for generalized linear models based on interval-censored data.
The rest of the article is organized as follows. In Section 2, we will first introduce some notations and assumptions to be used throughout the paper and present the proposed variable selection procedure. For the implementation of the presented method, a coordinate descent algorithm is developed in Section 3. In Section 4, the asymptotic properties of the proposed method under the BAR penalty are established, and Section 5 gives some numerical results obtained from a simulation study, which suggest that the method works well in practical situations. In Section 6, the method is applied to children’s mortality data of Nigeria that motivated this study, and Section 7 contains some discussion and concluding remarks.
2. Unbiased Transformation Variable Selection Procedure
Consider a failure time study that consists of n independent subjects and for subject i, let denote the failure time of interest and a p-dimensional vector of covariates. Suppose that for each , two observations are available at the observation times and such that they divide the axis into three parts and we know which part falls in. Thus, the observed data on subject i have the form , where and with being the indicator function,
. In the following, we will assume that is independent of and given .
To describe the covariate effects, we will assume that given , has the form
where is a known function, is a zero mean random error with distribution unknown, and and are unknown parameters. Note that for the estimation of the model above, if was exactly observed, a simple method would be to take as a new response variable and transform model (1) into a linear model. For the situation here where is interval-censored, however, the above method cannot be carried out directly.
To overcome the problem, we adopt the unbiased transformation approach to first convert into the variable
where , and are some continuous functions with finite continuous partial derivatives and also independent of the distribution of . Let denotes the joint density function of and and assume that , and satisfy the following conditions
Then according to Deng et al. (2012) [29], we have that
for .
Under Equation (2), one can see that , and not only depend on and but also depend on . More specifically, one can rewrite as
Thus, for estimating model (1) or , it is natural to consider the least squares method to minimize the mean square of the residual after the unbiased transformation
if was known. Of course, in practice, is unknown and for this, we propose to first estimate by the kernel density estimator and then consider
Note that given above involves the estimation of the two-dimensional function g and according to Section 5 of Deng (2004) [28], one could equivalently replace it by
where and denote the kernel estimators of the marginal density functions and In Lemma A1 given in Appendix A, we show that converge to in probability. Since and are positive variables, we propose to adopt the log-transformation technique and employ the log kernel density estimator
and
(Parzen, 1962 [30]). Then by following Deng (2004) [28], one can estimate by minimizing the mean squared residual after kernel estimation and unbiased transformation
Now we consider the variable selection or the selection of important covariates. For this, and motivated by (4), we propose to use the penalized least squares estimation method and to minimize the mean squared residual after kernel estimation and unbiased transformation with penalized criteria
where denotes a penalty function with a and (some comments are given below). In the following, several commonly used penalty functions are considered, including the LASSO penalty proposed by Tibshirani (1996) [4] and the SCAD penalty
with by Fan and Li (2001) [8]. For a in the SCAD penalty, we set suggested by Fan and Li (2001) [8]. Furthermore, we investigate the use of the MCP penalty
with given in Zhang (2010) [9] and the BAR penalty discussed in Zhao et al. (2020) [19], where denotes a nonzero “good” estimator of .
Note that for the application of the method described above, one needs to choose the functions and satisfying Equation (2). For this, many functions can be used and for simplicity, we suggest to employ (I) ; (II) ; or (III) , , The numerical study below indicates that they give satisfactory and robust results and more discussion on this can be found in Deng et al. (2012) [29].
3. Penalized Least Squares Coordinate Descent Algorithm
Let denote the estimator of given by minimizing the penalized criterion function in (5). In this section, we discuss the determination of and investigate a coordinate descent algorithm that takes turn to update each element of while making all other elements of fixed at their current estimates except .
Define
Then, at the k th iteration, the value of is derived by minimizing to determine . Note that by drawing the locally quadratic approximation idea discussed in Fan and Li (2001) [8], the locally quadratic approximation can be presented as
when . In other words, a quadratic function can locally approximate at as
Meanwhile, can be obtained by the second-order Taylor expansion,
where and represent the first and second derivatives of , respectively. Therefore, minimizing is equivalent to minimizing the function
with respect to . A closed form solution is given as
Note that it is easy to see that the resulting solution (6) and the approximation used above for the penalty function apply to any penalty function. However, the BAR penalty is not necessary to do a locally quadratic approximation, due to the fact that it is already a quadratic function of coefficients. For that situation, we can obtain the closed-form iterative solution proposed by Wu et al., 2020 [32] as
where and are the first and second derivatives of with respect to evaluated at , respectively. By combining the discussion above, the algorithm can be implemented as follows:
- Step 1:
- Set and choose the initial estimate
- Step 2:
- Step 3:
- Repeat Step 2 until the convergence or k exceeds a given large number.
There are many various criteria to check the convergence in Step 3 above. In the simulation studies below, the proposed algorithm is declared to achieve convergence if the maximum of the absolute differences of the estimates between two successive iterations is less than (Sun et al., 2019) [33]. To implement the algorithm above, one also needs to choose the tuning parameter . For the results given below, we use the Bayesian information criterion (BIC) proposed by Schwarz (1978) [34] which is data-dependent and defined as
In the above, denotes the final estimator of and represents the total number of nonzero estimates in , which serves as the degree of freedom. Alternatively, one could employ other methods such as a K-fold cross-validation (CV) (Verweij and Houwelingen, 1993) [35]. For given , we define BIC as above. Then, one can choose the value of that minimize . For variance estimation of the proposed estimators, we suggest using the nonparametric bootstrap method, which seems to work well as the numerical study indicates.
4. Asymptotic Properties
In this section, we establish the asymptotic properties of the variable selection procedure or the estimator proposed above with the use of the BAR penalty function. For this, let denote the true parameter value. Note that here, we replace p by to emphasize the dependence of p on n and assume that can diverge to infinity but . For simplicity, we write and assume and , where and . Let denote the BAR estimator of corresponding to the same participation. Set and , where is jth column of for . Let = , , = , for . For the asymptotic properties, we need the following regularity conditions.
. For every , exists, where is the derivative of and .
. and are positive i.i.d random vectors with uniformly continuous density functions and , respectively.
. and are independent given .
. .
. with probability 1 as n tends to ∞, where is a positive definite matrix ([19]).
. is uniformly continuous and of bound variation, and as . and . , and .
. There exists a constant such that for every integer n.
. and . As , and .
Note that Conditions – are necessary to obtain an unbiased transformation ([29]), where uniformly continuous density functions and are needed to make kernel estimation and converging to density functions and almost surely and Condition guarantees that converges to in probability ([31]). Condition assumes that is positive definite almost surely and its eigenvalues are bounded away from zero and infinity. Condition gives some sufficient, but not necessary, conditions needed to prove the convergence and asymptotic properties of the BAR estimator and nonzero coefficients are assumed to be uniformly bounded away from zero and infinity ([12]). Define , where and are and vectors, respectively. The following theorem gives the asymptotic properties.
Theorem 1.
Assume that Conditions – given above hold and , , and are continuous functions with finite continuous partial derivatives. Then, we have that:
(i) The fixed point of exists and is unique, where .
(ii) (Oracle property) the BAR estimator exists and is unique, where and is the unique fixed point of .
(iii) (Asymptotic normality) in distribution, where is defined in Appendix A.
5. Simulation Study
Now we present some results from a simulation study to assess the finite sample performance of the variable selection procedure presented in the previous sections. In the study, we generated the vector of covariates from and the covariance matrix with the element being . Given the ’s, the true failure times were generated based on model (1) with and the ’s following the standard normal distribution. For the generation of the observed interval-censored data or the two observation times for each subject, they were set to be times from the homogeneous Poisson process with the interexamination times being independently and identically distributed in the exponential distribution with a mean of 0.4 (Li et al., 2019 [20]).
For the application of the proposed variable selection procedure, we set
by following Deng et al. (2012) [29]. Meanwhile, we set
and
For the kernel estimators and , we considered the following biquadratic kernel function
and several different bandwidths. They include (a) , (b) , (c) with being the 0.75 quantile minus the 0.25 quantile, and (d) , where denotes the sample standard deviation and is selected by the CV method over 20 numbers from with equal distance. The simulation results of above four circumstances are listed in Supplemental Materials. We also considered the selection of the tuning parameter and the bandwidth choice (d) together, based on the BIC described above with over 50 numbers from 0.001 to 0.01 with equal distance and over 10 numbers from with equal distance. The results given below are based on the sample size or with 100 replications and , 30, 50, or 100.
Table 1, Table 2 and Table 3 are based on (8), (9), (10), respectively. Table 1, Table 2 and Table 3 present the results on the covariate selection with and , corresponding to relatively moderate and weak signals, respectively, as well as . They include the average number of nonzero estimates of the parameters whose true values are not zero (TP) and the average number of nonzero estimates of the parameters whose true values are zero (FP). In addition, we calculated and included in the table the median mean squared errors (MMSE) given by , measuring the prediction accuracy, and the standard deviation of MSE (SD), where denotes the population covariance matrix of the covariates. In the table, in addition to the BAR penalty function, we also considered the LASSO, MCP, and SCAD penalty functions, and the joint selection of the tuning parameter and the bandwidth based on the BIC was used. We also added backward stepwise variable selection based on BIC and for in Table 1.
Table 1.
Simulation results based on the joint selection of the tuning parameter and bandwidth for (I).
Table 2.
Simulation results based on the joint selection of the tuning parameter and bandwidth for (II).
Table 3.
Simulation results based on the joint selection of the tuning parameter and bandwidth for (III).
Furthermore, we added lower and greater sample sizes and with in Table 4 with . Furthermore, we added an extra simulation to demonstrate how this method works in the presence of noncontinuous covariates, i.e., the last covariates are generated from a Bernoulli distribution with a 0.5 success probability. The simulation results of this setup are presented in Table 5 with , , and . Finally, we considered a toy example in which the left endpoint imputation is considered (Sun, 2006). We illustrated the results in Table 6 to show the error that would be made if the data were considered as uncensored with , and .
Table 4.
Simulation results based on the joint selection of the tuning parameter and bandwidth for or with .
Table 5.
Simulation results based on the joint selection of the tuning parameter and bandwidth for existing noncontinuous covariate.
Table 6.
Comparison of the proposed method to the left endpoint method.
One can see from Table 1, Table 2 and Table 3 that the proposed approach seems to perform well with all penalty functions considered and in general, the method with the BAR penalty function gave smaller FP and thus parsimonious models. Meanwhile, results are similar for (I), (II), and (III). As expected, the method with the LASSO penalty function gave slightly larger FP and tended to select more noises than other penalty functions, and the method gave better results on both variable selection and prediction accuracy when the sample size increased. Furthermore, as expected, the important covariates with weak effects were more difficult to be identified than moderate effect covariates. The stepwise variable selection gave the largest MMSE and SD. One can see from Table 4 that the results with a larger sample have lower MMSE and SD. The proposed method only needed 17 s, 15 s, 20 s, and 5 s for LASSO, SCAD, MCP and BAR, respectively, on average in Table 4. One can see from Table 5 that the proposed procedure works in the presence of noncontinuous covariates. The unbiased transformation method can improve accuracy as shown when comparing the proposed method to the left endpoint imputation in Table 6.
6. An Application
In this section, we apply the methodology proposed in the previous sections to a set of children’s mortality data arising from the 2003 Nigeria Demographic and Health Survey (Kneib, 2006) [36]. The data set consists of 5730 children with their survival information and six covariates. They are AGE (the age of the children’s mother when giving birth), BMI (the mother’s body mass index at birth), HOSP (1 if the baby was delivered in a hospital and 0 otherwise), GENDER (1 for boys and 0 otherwise), EDU (1 if the mother received higher education and 0 otherwise), and URBAN (1 if the family lived in an urban area and 0 otherwise). Among others, one of the objectives of the study was to identify the covariates that had a significant effect on children’s mortality in Nigeria.
In the study, for each subject, if death occurred within the first two months after birth, the failure time could be observed exactly, and otherwise, only interval-censored observations were obtained based on the interview times of their mothers. Among the 5730 children, 233 gave exact failure times, 430 interval-censored observations, and the others provided right-censored observations. To apply the proposed approach, as in the simulation study, we set and used the same four penalty functions. For the three functions , and , we considered three choices and they are (I) ; (II) ; (III) , , , which will be referred to as choices (I), (II), and (III) below, respectively. Note that for exact failure times, we used as the response and made no unbiased transformation.
The results on the covariate selection and the estimated covariate effects are presented in Table 7, Table 8 and Table 9 with the use of choices (I)–(III), respectively. In addition, the estimated standard errors were obtained by using the nonparametric bootstrap method with 100 bootstrap samples and are included in the tables. One can see from these tables that the results seem to be robust with respect to the three choices, and all methods selected the factor AGE, suggesting that the age of the mother giving birth has a significant effect on the mortality risk of the children. The factor EDU was also selected by the LASSO, SCAD, and MCP penalty functions, and the results indicate that the level of the mother’s education seem to have a significant effect on the mortality risk. Xu et al. (2020) [37] have analyzed this application. They also found that EDU had a significant effect on the mortality risk of the children by the LASSO, ALASSO, and SCAD penalty functions. In contrast, the results suggest that the mother’s body mass index, the child’s gender, the baby’s birth location, and the family’s location did not appear to have significant effects on the children’s mortality or death rate.
Table 7.
Analysis results of children’s mortality data based on choice (I).
Table 8.
Analysis results of children’s mortality data based on choice (II).
Table 9.
Analysis results of children’s mortality data based on choice (III).
7. Discussion and Concluding Remarks
This paper discussed the variable selection for generalized linear models when only interval-censored failure time data are available, and for that problem, a new unbiased-transformation-based approach was proposed. One advantage for the use of the unbiased transformation is that it allows one to employ the simple penalized least squares approach for estimation. The proposed approach can accommodate any penalty function such as LASSO, SCAD, MCP, and BAR, and a coordinate descent algorithm was developed for the implementation of the proposed procedure. In addition, the asymptotic properties of the resulting estimators were established, and the simulation study indicated that the proposed methodology works well for practical situations.
There exist several directions for future research. One is that in the proposed method, it was assumed that and it is clear that this may not hold in practice. One such example is the accelerated failure time (AFT) model, in which case the proposed method would not be valid. In other words, one needs to modify the proposed unbiased transformation or develop some new methods that can be applied to the AFT model. Another direction is that in practical applications, one may encounter multivariate failure time data; in that scenario, Cai et al. (2005) [14] proposed a penalized marginal likelihood method for right-censored data with a mass of covariates. It would be helpful to develop some flexible and reliable methods to handle other types of multivariate failure time data including interval-censored data. A third direction is that it has been assumed in the previous sections that the dimension of covariates can diverge to infinity but is smaller than the sample size n. Obviously, there may be case where is greater than n such as in genetic or biomarker studies. In other words, some new methods that allow for need to be developed.
As one anonymous reviewer pointed out, neutrosophic statistics (Smarandache, 1998 [38]; Smarandache, 2013 [39]) is the extension of classical statistics and is applied when the data is coming from a complex process or from an uncertain environment. The current study can be extended using neutrosophic statistics as future research.
Supplementary Materials
The following are available online at https://www.mdpi.com/article/10.3390/math10050763/s1, Table S1: Simulation results based on the bandwidth selection method (a), Table S2: Simulation results based on the bandwidth selection method (b), Table S3: Simulation results based on the bandwidth selection method (c), Table S4: Simulation results based on the bandwidth selection method (d).
Author Contributions
Formal analysis, R.L.; Investigation, S.Z.; Methodology, T.H.; Software, R.L.; Supervision, T.H. and J.S. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Beijing Natural Science Foundation Z210003 and National Nature Science Foundation of China (Grant Nos: 12171328,11971064). Shishun Zhao’s study was supported by National Natural Science Foundation of China (NSFC) (12071176), Science and Technology Developing Plan of Jilin Province (No. 20200201258JC, 20190902012TC, 20190304129YY).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Publicly available datasets were analyzed in this study. Interested researchers can obtain the data in the application section of this paper on the official website of Demographic and Health Surveys (DHS) Program: https://dhsprogram.com/data/available-datasets.cfm (accessed on 19 February 2022). One can registered for a download account and apply for the data.
Conflicts of Interest
The authors declare no conflict of interest.
Appendix A
In this appendix, we sketch the proof of Theorem 1. For completeness, we introduce a few lemmas and notations, which are useful for the proof of Theorem 1. Let . When we consider BAR penalty, i.e.,
In Section 2 we can get
For simplicity, we write and as and . can be partitioned as
where the is a matrix. Multiplying to Equation (A2) gives
where , , and .
Lemma A1.
Assume that the conditions hold and , , and are continuous functions with finite continuous partial derivatives. We have
Proof.
Under the conditions and , we have that and according to Theorem A in Silverman (1978) [31]. Using Taylor’s expansion, , , and are continuous functions with finite continuous partial derivatives yielding
The proof of Lemma A1 is completed. □
Lemma A2.
Let denote the least squares estimator defined in (3), is a true value of and suppose that the conditions to hold. Then, we have that
Proof.
When we consider (A1), according to , we get . According to (3), we get . After some simple algebraic manipulations, we get
Under the conditions to , we have that . According to Lemma A1, . . We get that . □
Lemma A3.
Suppose that the conditions , , and hold. Then, we have that
Proof.
According to , we know . According to Lemma A1, converges to in probability for . Then, So Lemma A3 is proved. □
Proof of Theorem 1 (Oracle Property).
For a zero component, according to Lemma 1 in [12] conclusion (i), let be a sequence of positive real numbers satisfying and . Define , Then, with probability tending to 1, we have
For a nonzero component, according to Lemma 2 in [12], is a contraction mapping from to itself, where
Let is the unique fixed point of defined by .
Hence, to prove the consistency of nonzero parts in Theorem 1, it is sufficient to show that
Define if . It is easy to see from (A3) that for any ,
According to conclusion b in Lemma 1 of Dai et al. (2018) [12], g in (A2) is a mapping from to itself. This, together with (A3) and (A6), implies that, as ,
with probability tending to 1. Note that
where the last step is on the basis of . Let , for every integer . From (A7), for any , there exists a positive integer N such that for every integer
with probability 1 as n tends to ∞ and the right-hand term tends to 0 as . This proves (A4). Therefore, . □
Proof (Proof of Theorem 1) (Asymptotic Normality).
We next prove the asymptotic normality of the nonzero components in the BAR regression. Finally, based on (A6), we have
where
According to the first-order resolvent expansion formula, we obtain
This yields
With conditions and , we can get
Therefore, we can get
According to Lemma A3, we know that converge to in probability. Then, we get
in distribution. Denote as . This complete the proof.
□
References
- Goldberger, A.S.; Jochems, D.B. Note on stepwise least squares. J. Am. Stat. Assoc. 1961, 56, 105–110. [Google Scholar] [CrossRef]
- Hoerl, A.E.; Kennard, R.W. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
- Mitchell, T.J.; Beauchamp, J.J. Bayesian Variable Selection in Linear Regression. J. Am. Stat. Assoc. 1988, 83, 1023–1041. [Google Scholar] [CrossRef]
- Tibshirani, R. Regression Shrinkage and Selection via the Lasso. J. R. Stat. Soc. Ser. B 1996, 58, 267–288. [Google Scholar] [CrossRef]
- Raftery, A.E.; Madigan, D.; Hoeting, J.A. Bayesian Variable model averaing for Linear Regression Models. J. Am. Stat. Assoc. 1997, 92, 179–191. [Google Scholar] [CrossRef]
- Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B 2005, 67, 301–320. [Google Scholar] [CrossRef] [Green Version]
- Zou, H. The Adaptive Lasso and Its Oracle Properties. J. Am. Stat. Assoc. 2006, 101, 1418–1429. [Google Scholar] [CrossRef] [Green Version]
- Fan, J.; Li, R. Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
- Zhang, C.H. Nearly Unbiased Variable Selection Under Minimax Concave Penalty. Ann. Stat. 2010, 38, 894–942. [Google Scholar] [CrossRef] [Green Version]
- Dicker, L.; Huang, B.; Lin, X. Variable Selection and Estimation with the Seamless-L0 penalty. Stata Sin. 2013, 23, 929–962. [Google Scholar] [CrossRef]
- Liu, Z.; Li, G. Efficient Regularized Regression with L0 Penalty for Variable Selection and Network Construction. Comput. Math. Methods Med. 2016, 2016, 3456153. [Google Scholar] [CrossRef] [Green Version]
- Dai, L.; Chen, K.; Sun, Z.; Liu, Z.; Li, G. Broken Adaptive Ridge Regression and Its Asymptotic Properties. J. Multivar. Anal. 2018, 168, 334–351. [Google Scholar] [CrossRef] [PubMed]
- Zheng, X.; Rong, Y.; Liu, L.; Cheng, W. A More Accurate Estimation of Semiparametric Logistic Regression. Mathematics 2021, 9, 2376. [Google Scholar] [CrossRef]
- Cai, J.; Fan, J.; Li, R.; Zhou, H. Variable Selection for Multivariate Failure Time Data. Biometrika 2005, 92, 303–316. [Google Scholar] [CrossRef] [PubMed]
- Fan, J.; Li, R. Variable Selection for Cox’s Proportional Hazards Model and Frailty Model. Ann. Stat. 2002, 30, 74–99. [Google Scholar] [CrossRef]
- Tibshirani, R. The Lasso Method for Variable Selection in the Cox Model. Stat. Med. 1997, 16, 385–395. [Google Scholar] [CrossRef] [Green Version]
- Zhang, H.H.; Lu, W. Adaptive Lasso for Cox’s proportional hazards model. Biometrika 2007, 94, 691–703. [Google Scholar] [CrossRef]
- Khan, N.; Aslam, M.; Raza, S.M.; Jun, C. A New Variable Control Chart under Failure-censored Reliability Tests for Weibull distribution. Qual. Reliab. Eng. Int. 2018, 35, 572–581. [Google Scholar] [CrossRef]
- Zhao, H.; Wu, Q.; Li, G.; Sun, J. Simultaneous Estimation and Variable Selection for Interval-Censored Data With Broken Adaptive Ridge Regression. J. Am. Stat. Assoc. 2020, 115, 204–216. [Google Scholar] [CrossRef]
- Li, S.; Wei, Q.; Sun, J. Penalized estimation of semiparametric transformation models with interval-censore data and application to Alzheimer’s disease. Stat. Methods Med. Res. 2019, 29, 2151–2166. [Google Scholar] [CrossRef]
- Du, M.; Sun, J. Variable selection for interval-censored failure time data. Int. Stat. Rev. 2021. accepted. [Google Scholar] [CrossRef]
- Ali, S.; Raza, S.M.; Aslam, M.; Butt, M.M. CEV-Hybrid DEWMA charts for censored data Using Weibull distribution. Commun. Stat.—Simul. Comput. 2021, 50, 446–461. [Google Scholar] [CrossRef]
- Zhao, S.; Hu, T.; Ma, L.; Wang, P.; Sun, J. Regression analysis of informative current status data with the additive hazards model. Lifetime Data Anal. 2015, 21, 241–258. [Google Scholar] [CrossRef]
- Wang, P.; Zhao, H.; Sun, J. Regression Analysis of Case K Interval Censored Failure Time Data in the Presence of Informative Censoring. Biometrics 2016, 72, 1103–1112. [Google Scholar] [CrossRef] [PubMed]
- Li, S.; Hu, T.; Wang, P.; Sun, J. A Class of Semiparametric Transformation Models for Doubly Censored Failure Time Data. Scand. J. Stat. 2018, 45, 682–698. [Google Scholar] [CrossRef]
- Sun, J. The Statistical Analysis of Interval-Censored Failure Time Data; Springer: New York, NY, USA, 2006. [Google Scholar]
- Buckley, J.; James, I. Linear regression with censored data. Biometrika 1979, 66, 429–436. [Google Scholar] [CrossRef]
- Deng, W. Some Issues on Interval Censored Data. Ph.D. Dissertation, Fudan University, Shanghai, China, 2004. [Google Scholar]
- Deng, W.; Tian, Y.; Lv, Q. Parametric Estimator of Linear Model with Interval-Censored Data. Commun. Stat.—Simul. Comput. 2012, 41, 1794–1804. [Google Scholar] [CrossRef]
- Parzen, E. On Estimation of a Probability Density Function and Mode. Ann. Math. Stat. 1962, 33, 1065–1076. [Google Scholar] [CrossRef]
- Silverman, B.W. Weak and Strong Uniform Consistency of the Kernel Estimate of a Density and its Derivatives. Ann. Statist. 1978, 6, 177–184. [Google Scholar] [CrossRef]
- Wu, Q.; Zhao, H.; Zhu, L.; Sun, J. Variable Selection for High-dimensional Partly Linear Additive Cox model with application to Alzheimer’s Disease. Stat. Med. 2020, 39, 3120–3134. [Google Scholar] [CrossRef]
- Sun, L.; Li, S.; Wang, L.; Song, X. Variable Selection in semiparametric nonmixture cure model with interval-censored failure time data: An application to the prostate cancer screening study. Stat. Med. 2019, 38, 3026–3039. [Google Scholar] [CrossRef] [PubMed]
- Schwarz, G. Estimating the Dimension for a Model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
- Verweij, P.; Houwelingen, H. Cross-validation in survival analysis. Stat. Med. 2010, 12, 2305–2314. [Google Scholar] [CrossRef]
- Kneib, T. Mixed Model-Based Inference in Geoadditive Hazard Regression for Interval-Censored Survival Times. Comput. Stat. Data Anal. 2006, 51, 777–792. [Google Scholar] [CrossRef] [Green Version]
- Xu, Y.; Zhao, S.; Tao, H.; Sun, J. Variable Selection for generalized odds rate mixture cure models wth interval-censored failure time data. Comput. Stat. Data Anal. 2020, 156, 107115. [Google Scholar] [CrossRef]
- Smarandache, F. Neutrosophic Probability, Set, and Logic; American Research Press: Rehoboth, NM, USA, 1998. [Google Scholar]
- Smarandache, F. Introduction to Neutrosophic Measure, Neutrosophic Integral, and Neutrosophic Probability; Sitech-Eduacation Publisher: Craiova, Romania, 2013. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).