Abstract
Cure models and receiver operating characteristic (ROC) curve estimation are two important issues in survival analysis and have received attention for many years. In the development of biostatistics, these two topics have been well discussed separately. However, a rare development in the estimation of the ROC curve has been made available based on survival data with the cure fraction. On the other hand, while a large body of estimation methods have been proposed, they rely on an implicit assumption that the variables are precisely measured. In applications, measurement errors are generally ubiquitous and ignoring measurement errors can cause unexpected bias for the estimator and lead to the wrong conclusion. In this paper, we study the estimation of the ROC curve and the area under curve (AUC) when variables or biomarkers are subject to measurement error. We propose a valid procedure to handle measurement error effects and estimate the parameters in the cure model, as well as the AUC. We also make an effort to establish the theoretical properties with rigorous justification.
Keywords:
area under curve; bias correction; cure; censoring; EM algorithm; incomplete response; insertion method; mismeasurement MSC:
62Nxx; 62N02
1. Introduction
Survival analysis has been an important research topic in biostatistics. The main purpose of survival analysis is to understand the time at which a specific event or disease happens among patients and make a prediction. The key feature of survival data is that the time-to-event variable either is observed at a finite time (i.e., the event, such as death, happens during the observation period) or cannot be observed due to censoring. A more detailed introduction can be found in some monographs, such as [1]. However, in the context of survival analysis, there is a situation in which patients never experience the failure event (e.g., death) in the study period; this phenomenon is called cure, and the corresponding model and data structure is the well-known cure model. There are some estimation methods to analyze the cure model in the literature. For example, [2] considered the transformation model for survival data and implemented the logistic regression model to characterize the probability of cure. Ref. [3] proposed the conditional likelihood function given truncation variables. Ref. [4] explored the cure model with a mixture of single-index and Cox models. More comprehensive discussions are summarized in [5].
In applications, noisy data are usually inevitable due to the data collection or sampling mechanism. One characteristic of typical noisy data is the measurement error, which indicates that the observed variables cannot fully reflect what they should be. This phenomenon is usually caused by machines that are imprecise or incorrect records by researchers. In the framework of survival data with cure models, some methods have also been developed to deal with measurement error problems. To name a few, ref. [6] proposed the simulation and extrapolation (SIMEX) method to correct the measurement errors. Ref. [7] developed the conditional expectation approach to correct the error effects. Unlike [6,7], who focused on the Cox proportional hazards model, ref. [8] studied the transformation model and adopted the SIMEX method to handle measurement error.
The other important issue in survival analysis is the estimation of the receiver operating characteristic (ROC) curve, whose primary goal is to display the sensitivity and the specificity of a continuous marker for a given disease. From the existing literature, the estimation procedures lie in the censored data, including [9,10,11,12,13]. However, to our knowledge, rare methods have been available to estimate the ROC curve when survival data contain the cured group and variables or biomarkers suffer from measurement error, except for [14,15], who considered the Cox model and the mixture cure model with the decomposition of sensitivity and specificity, respectively. The other challenging feature in analysis is the measurement error in biomarkers, which frequently appears in real-world applications. For example, [16] studied the Mayo Clinic primary biliary cirrhosis dataset with measurement errors in serum bilirubin, serum albumin, and prothrombin time; ref. [17] pointed out the biomarker systolic blood pressure (SBP), which is one of the key prognostic factors for cardiovascular risk scores, is possibly subject to measurement error; ref. [18] implemented three error-prone biomarkers, ER, PR, and Ki67 proteins, to analyze health study data on breast cancer in nurses. All the literature reveals that measurement error in biomarkers possibly affects the performance of standard diagnostic measures, such as ROC curves and area under the ROC curve (AUC). While error-prone biomarkers were discussed in survival analysis or the estimation of ROC curves (e.g., [16,17]), it seems that estimation methods for the cure fraction and error-prone biomarkers are unavailable.
To fill this research gap, we aim to develop a new estimation procedure in this paper. Specifically, we first consider the transformation model with the cure fraction in the survival data and the measurement error in covariates or biomarkers. Our setting generalizes the conventional implementation of the Cox model. Unlike existing methods that adopted the SIMEX method to correct for measurement error (e.g., [6,8]), we propose the insertion method, whose key spirit is to construct the new function so that its conditional expectation can be recovered to the true function based on the unobserved covariates or biomarkers. The corrected estimation function ensures the consistency of the estimator. After that, we further explore the estimation of the ROC curve when the survival data contain cured samples and mismeasured variables. We develop the time-independent and time-dependent estimation for the ROC curves and the corresponding estimation of the area under curve (AUC). Theoretically, we also establish the consistency and the asymptotic normality of the AUC estimator. The contribution of this paper is a valid estimation strategy, and the main focus is its theoretical establishment with rigorous justification.
The remainder is organized as follows. In Section 2, we introduce notation and models. In Section 3, we first present the proposed method to correct the measurement error effect and derive the estimator. Moreover, the corresponding theoretical properties are presented. In Section 4, we present numerical experiments and their results. We conclude the paper with discussion in Section 5. Finally, the proofs of the theoretical properties are placed in Appendix A.
2. Notation and Models
2.1. Cure Model
Let T and C denote the (uncured) failure time and the censoring time, respectively. With subjects cured and uncured, the failure time is determined by , where indicates whether a subject is cured () or uncured (). To characterize A, we consider the logistic regression
where X is the p-dimensional vector of covariates or biomarkers and is the p-dimensional vector of the parameters associated with X.
On the other hand, regarding the uncured failure time T, we consider the survivor function that follows the transformation model (e.g., [19,20])
where is the p-dimensional vector of the parameters associated with X, is the unknown strictly increasing function, and is the known parameter. In particular, when , (2) follows the proportional odds (PO) model; when , then (2) becomes the proportional hazards (PH) model. Given (2), the survivor function of is given by
Finally, let C denote the censoring time, and we further define as the observed survival time and , with being the indicator function. Let denote the independent and identically distributed sample as .
To develop the method in the following sections, some typical assumptions are imposed in the context of survival analysis, including noninformative censoring times, and the independence of failure time is independent of the censoring time, given the covariates.
2.2. Measurement Error Model
In applications, the covariates or biomarkers are possibly subject to measurement error. Here, we consider the case where is error-contaminated, and let denote the observed or surrogate measurement of . Consistent with most work in the literature (e.g., [21]), we consider that the are continuous and linked with by the additive measurement error model:
where is independent of , with the covariance matrix , and represents the zero vector. Let denote the reliability ratio of the covariances of jth and kth covariates and their surrogates for , which reflects the magnitude of variation between the unobserved and observed covariates, and higher values indicate a minor measurement error. Note that , which implies that . In applications, we may assume that the reliability ratio is not a small value for all (e.g., [22]), which implies that the values in are not too large.
In situations where the parameters for the measurement error model (3) are unknown, we may utilize the information carried by additional data sources, such as repeated measurements or validation subsamples. As a result, to highlight key ideas, we focus our attention on the estimation of parameters associated with the survival model and assume that the parameter for measurement error model (3) is known. This assumption is reasonable, typically arising in two circumstances: (i) prior studies provide the information on the covariate mismeasurement and offer an estimate of , and (ii) in conducting sensitivity analyses, different values of are specified to understand how mismeasurement effects may affect inference results about the parameters associated with the survival model.
3. Methodology
Suppose that we have a sample of n subjects and that for , has the same distribution as . We first estimate the unknown parameters and the survivor function of based on formulation (2) with suitable error correction. After that, we develop the error-corrected method to estimate the time-independent/time-dependent AUC. Moreover, theoretical results are also established after presenting the proposed estimation procedures.
3.1. Construction of the Error-Corrected Likelihood Function
To develop our approach, we start the discussion by pretending that is available. After that, we replace with and propose the error-corrected method to adjust error effects.
For cured patients with and , we have . In addition, for those patients who encounter the failure event, i.e., and , we have , where represents the density function of uncured failure time. Moreover, for observations who are censored but not cured ( and ), we can obtain . Therefore, for subject , the complete likelihood function is given by
where is the derivative of , and the corresponding log-likelihood function of (4) is
However, note that we usually have in (5) instead of . It is crucial to make a suitable correction. Our strategy is to find the new likelihood function based on , denoted by , such that the conditional expectation is satisfied. In this way, we have , showing that is an error-corrected likelihood function and provides the same optimizer as those obtained from ℓ. By measurement error model (3) and the moment-generating function under the normal distribution, it is direct to have and . According to (5), however, the challenge is that the error-prone covariate appears in and , which are nonlinear terms.
To address this issue, we first consider the function . Since the expectation of a ratio random variable is sometimes difficult to compute, we may adopt the approximation. Specifically, let and . By the second-order Taylor series expansion (e.g., [23], pp. 69–72), we have the following approximation:
Taking the conditional expectation gives that
On the other hand, the conditional variance of and conditional covariance of and , given , are respectively derived as
and
Provided that the reliability ratio is not too small or, equivalently, that the variance of the noise term is not too large, the two values and are close to zero, yielding that the second and last terms in (6) are close to zero; thus, we have that
and performing integration on (7) with respect to yields
where the last approximation is due to that “plus one” being almost ignorable. Thus, (8) suggests that the suitable error correction of is
3.2. Estimation of Parameters and Functions
After establishing the corrected log-likelihood function (11), we aim to estimate unknown parameters and , as well as unknown function .
However, the other concern is that for those censored observations (), whether they are cured or not is unknown; thus, can be regarded as a missing value. To address this concern, we apply the expectation–maximization (EM) algorithm with error correction. Specifically, in the E-step, we start by considering to replace . Recall that we have only instead of ; then, it is crucial to make a suitable adjustment. Let . We aim to find , which is in terms of the observed covariate , such that the measurement error would be adjusted, i.e., .
When , we have , and it is easy to check that . When , then . Let , which satisfies . In particular, we apply the idea in [22] to express by the linear approximation
with , , , and .
Following the derivation of [22], we apply the second-order Taylor series expansion for and around with respect to and , respectively; then, we can obtain the following two approximations:
and
Combining (13) and (14) gives that
Then, taking the expectation on (15) yields that
where the expectation of the subtraction of two quadratic forms in (16) is given by
since . Consequently, (16) implies that
Therefore, by the given data and the fact from measurement error model (3), (12) can be estimated by
where and are the empirical estimates of and , respectively. Thus, replacing in (17) with yields
When the E-step is established, we examine the M-step. Specifically, replacing in (11) with gives
where contains information on only and includes and , which reflect the survival process. Thus, the maximization of for can be performed separately. More specifically, the estimator of , denoted by , is determined by . Due to the unavailability of the closed form, one can adopt the Newton–Raphson approach to derive .
Naturally, the estimator of can also be determined by
However, the challenge comes from nonparametric function . To address it, we follow the same line as [19] and “parametrize” function by , where represents the index of the latest jump point before and are the corresponding logs of jump sizes of at , which are the sorted times. Based on such representation, can be expressed as
Note that (21) is a concave function; then, taking the negative sign on (21) makes it become convex. To derive the estimators of and from (21), we employ the coordinate-descent algorithm, which is frequently implemented in the optimization problem (e.g., [24]). Detailed descriptions of the steps are given below:
- Step 1:
- Choose an initial value for and , and denote them by and .
- Step 2:
- For , given and , , update by finding
- Step 3:
- For , given and , update by finding
- Step 4:
- Repeat Steps 2 and 3 until convergence, and let and denote the limit of and as .
We comment that the coordinate-descent algorithm is a common strategy to obtain the minimizer (e.g., [24]). Its key idea is to iteratively update the scale parameter with other values given under the previous step. This approach enables us to obtain a stable numeric scheme and avoid estimating high-dimensional vectors of parameters. In the literature, the alternative strategy to address the optimization in (20) is called the minimization–maximization (MM) algorithm (e.g., [19,25]). While the MM algorithm is also an iterative approach based on the historical values that is similar to the coordinate-descent algorithm, the key step of this approach is to find and optimize the surrogate function that improves the value of the objective function or leave it unchanged, but in general, such suitable surrogate functions are not easy to find. Instead, the coordinate-descent algorithm enables us to directly examine (21).
Finally, to assess the validity of the proposed method, we present the following theoretical results, and the proof is postponed to Appendix A.2.
Theorem 1.
Let , , and denote the true values of parameters. Under regularity conditions, the proposed estimator is consistent. That is, as ,
where and are the - and supremum norms, respectively.
Let , and let us define as the collection of directions, where and are p-dimensional vectors, is a function defined in with being the end time of the study, and is the total variation, defined as
The next theorem shows the convergence in distribution of the proposed method with the rate .
Theorem 2.
Under regularity conditions, the proposed estimator has an asymptotic distribution. That is, as , converges in distribution to a mean-zero Gaussian process in the functional space on .
Let . According to (19), the parameters and appear in and . With fixed, by the theory of M-estimation, the marginal asymptotic distribution of is given by
where , with
and for a vector .
3.3. Estimation of ROC and AUC
When the estimators of , , and are obtained, we now propose the estimation procedure for the time-independent and time-dependent ROC curves, as well as the area under curve (AUC) with measurement error, where the former says that the event status and marker value for an individual are independent of the change in the time and the latter reflects that the curve varies at different time points and seems more flexible than the former. In real-world applications, it is common for the disease status to change over time, so the time-dependent ROC curve is more widely explored. However, in the presence of the cure fraction in survival time and measurement error in the covariates, both time-independent and time-dependent cases may not be fully discussed. To make the discussion on measurement error effects more comprehensive, we still study two different types of ROC curve estimation in the following two subsections.
3.3.1. Time-Independent AUC
We first discuss the time-independent case. Let be the binary and time-independent classifier with biomarker X, which classifies patients into the uncured group with 1. On the other hand, as discussed by [26], the linear composite biomarker is an optimal classifier among all functions of biomarkers. As a result, the event is equivalent to the inequality for some constant u. Then, the time-independent true-positive rate (TPR) and false-positive rate (FPR), denoted by and , are defined as
Moreover, (23) can be further expressed as
and
respectively, with and being the density function of X and the expectation evaluated with respect to X, respectively, where the last equality in (24) comes from
with and being the conditional density function of T given X; similar derivations show that
with , yielding the last equality of (25).
In practice, however, biomarker X may also be measured with error. Let denote the surrogate version of X. We adopt measurement error model (3) to build the relationship between X and .
In (24), appears in and . To correct for error effects, our strategy is to follow a similar approach to that in Section 3.2 and find the surrogate functions in terms of , say and , such that and . In this way, we have and ; thus, and can be replaced by and , respectively. A similar strategy is also adopted for the numerator and denominator terms in (25) and (26). To this end, we develop and .
According to the Mean Value Theorem, there exists between X and , such that
and taking the conditional expectation gives
This indicates that the corrected version is given by the original function with X being replaced by . In particular, this result is essentially similar to the result in [27], which pointed out the satisfactory approximation of implementing in the logit function.
On the other hand, note that based on the measurement error model, if , then we have . This indicates that implies . Therefore, takes the value when , and by a similar strategy to (27), taking the conditional expectation yields
thus, is taken as with X being replaced by .
Consequently, the corrected TPR, FPR, and AUC are given by (24), (25), and (26), respectively. Moreover, with the consistent estimator in Section 3.2, the corresponding estimators of the corrected TPR, FPR, and AUC are
and
where represents with being replaced by .
3.3.2. Time-Dependent AUC
We now explore the time-dependent case. Denote the time-dependent TPR and FPR, respectively, by
Similar to the discussion on (23), one can further express (31) as
and
Thus, the time-dependent AUC is defined as
Moreover, following a similar discussion to that in Section 3.3.1, the corrected time-dependent TPR, FPR, and AUC are given by (32), (33), and (34) with X being replaced by , respectively. With the consistent estimators , the corresponding estimators are determined by
and
Finally, at the end of this section, we establish the following theorems to justify the proposed estimators of time-independent/-dependent AUC, including consistency and asymptotic normality.
Theorem 3.
Suppose that the conditions in Theorem 1 hold. With a fixed value u, as , the following applies:
- (a)
- For the time-independent result in Section 3.3.1,
- (b)
Theorem 4.
Suppose that conditions in Theorems 1 and 2 hold. Then as , the following applies:
- (a)
- For the time-independent result in Section 3.3.1,where is the asymptotic variance whose exact form is given in Appendix A.5;
- (b)
- For the time-dependent result in Section 3.3.2,where is the asymptotic variance whose exact form is given in Appendix A.5.
4. Numerical Studies
Let or 500 denote the sample size. For a subject , let be generated by the standard normal distribution. Let , , and be the true values of the parameters in (1) and (2), respectively. We consider the function . We then generate and by (1) and (2), respectively. We further independently generate the censoring time from the uniform distribution in an interval , where is a pre-specified constant, such that the censoring rate is approximately 30%. In addition, we further apply measurement error model (3) to generate for , where is independently generated by the normal distribution with mean zero and variance , and 0.55, which reflect minor, moderate, or severe measurement error effects, respectively. Consequently, the collected dataset is denoted by . For each setting, we run the simulation 500 times repeatedly.
There are two goals in this study. First, we aim to estimate two parameters and in (1) and (2), respectively. Based on two estimates and , the next goal is to estimate the time-independent AUC (26) and the time-dependent AUC (34) at the time points and 10. In addition to implementing the proposed method with measurement error correction, to see the impact of the measurement error on estimation, we also examine the naive method, which is given by the same estimation procedure in Section 3 but with error-prone covariates equipped.
The simulation results are summarized in Table 1, where we report the bias, the standard error (S.E.) obtained by repeated simulations, and the mean squared error (MSE). In general, we find that biases are increasing when becomes large, which implies that the measurement error may affect the estimation even though the correction is taken into account. In particular, larger biases of the estimated AUC indicate that the shape of the ROC curves slightly moves to the 45-degree straight line. For the comparisons of two estimation methods, the proposed method generally outperforms the naive method due to the smaller biases of the estimators, regardless of the change of the values of and sample sizes n. This indicates that the proposed method is valid to produce precise estimators. Note that the S.E. of the naive method is relatively smaller than the proposed method, as discussed in [22]; it is due to removing biases in point estimators and is a typical phenomenon in the framework of measurement error analysis. By the trade-off between bias and variation, we observe from the MSE that the proposed method is still better than the naive method with smaller MSE. In summary, the numerical results verify the validity of the proposed method and the measurement error correction.
Table 1.
Simulation results. “Parameter” indicates the some parameters of interest; “Method” refers to the proposed and naive methods; “Bias” is the bias of the estimators; “S.E.” is the standard error; and “MSE” is the mean squared error.
5. Summary
In this paper, we study the transformation model to analyze the survival data with the cure fraction and its extension to the estimation of the ROC curve. In addition, we also explore the measurement error in covariates or biomarkers, which is one type of noisy data. To deal with measurement error effects and derive a reliable estimator, we propose a corrected likelihood function whose expectation recovers to the likelihood function under the true covariates. In addition, we also propose a measurement error correction strategy to estimate the time-independent/time-dependent AUC. To derive the optimization of (19), we introduce the EM algorithm with finite stepwise iterations to obtain the estimators, but one should keep in mind that the computational implementation is not unique, and some alternative strategies can be adopted, such as quasi-Newton methods or damped Anderson acceleration, when the dimension of variables is large or the settings are complex. The other contribution of this paper is the focus on the establishment of the asymptotic properties for the proposed estimator with rigorous derivations. While we have no real data application in the current stage, it is also expected that the proposed method is valid to handle the relevant data, since simulation studies show that the performance of the proposed method is satisfactory.
The current development focuses on the estimation of the parameters, as well as the AUC, and the handling of measurement error problems, but some further issues related to the AUC could be explored in depth. For example, the current setting involves the cure fraction, which may induce long follow-up periods for the survival time. An interesting but unsolved concern is to understand whether the dynamic AUC estimation is stabilized and the corresponding impact caused by measurement error. Exploring this issue in depth might make the diagnosis of the disease more precise. In the next stage of our research, we wish to explore this issue from a theoretical perspective and provide rigorous justification. In addition to the time-dependent AUC, as mentioned by a referee, the other challenging feature in applications is the time-varying covariates. While this issue was studied in the literature (e.g., [28]), the corresponding applications to AUC estimation seem to be rarely discussed. Moreover, in the presence of measurement error in covariates, the other challenge includes the characterization of time-dependent covariates and the corresponding measurement error correction. We expect that the current method can be extended to handle this complex structure in the near future.
Funding
This research was funded by National Science and Technology Council, Taiwan, with grant number 112-2118-M-004-005-MY2.
Data Availability Statement
All the data in this paper are simulated data generated by the procedure in Section 4.
Conflicts of Interest
The authors declare no conflicts of interest.
Appendix A. Theoretical Justification
Appendix A.1. Regularity Conditions
- (C1)
- is a compact set, and the true parameter value is an interior point of .
- (C2)
- Let be the finite maximum support of the failure time.
- (C3)
- are independent and identically distributed for .
- (C4)
- The covariates or biomarkers are bounded.
- (C5)
- Censoring time is noninformative. That is, the failure time and the censoring time are independent, given the covariate .
Condition (C1) is a basic condition that is used to derive the maximizer of the target function. (C2) to (C5) are standard conditions for survival analysis, which allow us to obtain the sum of i.i.d. random variables and hence to derive the asymptotic properties of the estimators.
Appendix A.2. Proof of Theorem 1
Proof.
Note that the function is parameterized as , and the corresponding estimator is given by . Following Helly’s Lemma (e.g., [29], Section 2.1) and a similar discussion in [30], we can show that converges to , provided that is replaced by the true value for . As a result, to show the consistency of and , it suffices to examine the relationship between and ; thus, we can explore the asymptotic behavior of .
In the following derivation, we aim to prove that
Specifically, if (A1) holds, then we conclude that , and as .
To prove (A1), we follow the strategy in [24] and start by showing that
for every and a given value , where , with , , and .
We write . By the Taylor series expansion around , we have
where , , and are the first-order derivatives of with respect to , , and , respectively; , , and are the second-order derivatives of with respect to , , and , respectively; and is the derivative of with respect to and .
We now separately examine each term in (A3). Note that is the maximizer of , and according to our development. This implies that is also the maximizer of or, equivalently, the solutions of , , and . Therefore, we have
For the second-order derivative , its explicit form is given by
By the Law of Large Numbers, we have that as . As a result, to find , it suffices to compute , where the last equality is based on the interchange of the derivative and integral.
Taking the derivative of (19) with respect to and then taking the conditional expectation give
which also implies that
As a result, by the Law of Large Numbers and Condition (C3), is determined by
By Conditions (C1) and (C4), the covariates and the parameter are bounded, and is positive definite, which indicates that is bounded componentwise and its negative definite ensures that the maximizer of exists.
Next, we explore . By taking the derivative of with respect to , we have
Note that the moment-generating function gives
and
then, the conditional expectation of given is
This further shows that
where
and
Consequently, by the Law of Large Numbers, we have that
and is the matrix of with stack in row for , where
and they are bounded elementally due to positive definite matrix and Conditions (C1) and (C4).
Finally, we examine . Taking the first derivative of with respect to yields
Additionally, by (A5), we have
This also implies that
and
for . Therefore, by the Law of Large Numbers, we have with diagonal entries
and off-diagonal elements
for .
Appendix A.3. Proof of Theorem 2
Proof.
In this proof, we adopt Theorem 3.3.1 in [31] to derive the asymptotic distribution. A similar discussion was also presented by [32].
Let and . Moreover, define and . According to Theorem 3.3.1 in [31], we need to check the following:
- (a)
- , where Z is a tight random element;
- (b)
- The map is Fréchet differentiable at with a continuously invertible derivative , where denotes an operator of the derivative of f with respect to x;
- (c)
- , and satisfies .
If Conditions (a)–(c) are verified, then we conclude that
Lastly, we examine Conditions (a)–(c) separately.
- Check Condition (a):
Since is the true value, we automatically have . Then, , where
and
By Conditions (C1) and (C4), is Donsker bounded because it belongs to the finite-dimensional class of a measurable score function. In addition, and are bounded real valued functions on the bounded support due to Conditions (C2) and (C4). This indicates that and are Donsker (e.g., [29], p. 270). Therefore, by Example 2.10.7 in [31], we conclude that is also Donsker. Consequently, by Section 2.8.2 in [31] and by Theorem 19.3 in [29], we obtain that as , converges in distribution to a tight random element, denoted by Z.
- Check Condition (b):
The first part is to show Fréchet differentiability. That is, it suffices to prove that (e.g., [29], p. 297)
Let , and let denote the true value of . We write , with , where and satisfies bound variation on . Define .
On one hand, by the Taylor series expansion on around , we have
On the other hand, by taking the derivative of based on the definition in p. 296 of [29], we have
Finally, by combining (A7) and (A8), we obtain (A6) as , because of Theorem 1 and because as .
The second part is to discuss the continuous invertibility of . Similar to the discussion in [32], the continuous invertibility of the Fréchet derivative can be justified by showing that there exists a constant such that
Since can be expressed as a linear combination of the three -operators, where three terms in are continuously differentiable functions that map to a finite-dimensional space. Therefore, we conclude that is the summation of invertible and compact operators; thus, (A9) holds by the similar arguments in [32]. Consequently, is verified to be continuously differentiable.
- Check Condition (c):
The first claim, , has been verified by the argument of Condition (a). In addition, is the maximizer of , and is the consistent estimator due to Theorem 1. Therefore, the second claim, , holds. □
Appendix A.4. Proof of Theorem 3
Proof.
In this appendix, we mainly prove results (a) and (b) separately.
- Proof of part (a)
In this proof, we only show that, as ,
and we omit the proofs of and because the derivations are similar to those for (A10).
We write , where
and
In the following, we examine and separately.
Claim 1:
Let . Show that as ,
Note that by (28), we have , yielding that . By Theorem 1, which shows that is the consistent estimator of , we have as . On the other hand, by Condition (C3), which shows that is independent, applying the Law of Large Number gives that as , and . Thus, Claim 1 holds.
Claim 2:
Let . Show that as ,
Similar to Claim 1, we can obtain that is equal to due to (29). By adding and subtracting an additional term, we have
where
can be expressed as
with
By a similar discussion to that for Claim 1, we have as . In addition, applying Theorem 1 gives . By and the continuous mapping theorem (e.g., [33], Theorem 3.2.4), we have that as ,
Moreover, similar derivations show that
By (A14) and the fact that for any event , we have
Combining (A13) and (A15) yields ; thus, by Chebyshev’s inequality, we have
On the other hand, since (A11) is formulated as the U-statistics with bounded kernel, then applying the similar derivation of [34] gives
- Proof of part (b)
The proof of part (b) is similar to that of part (a) except for the involvement of the estimates of the function and the parameter . According to Theorem 1, since and are consistent with the true values, then applying a similar argument in part (a) yields the desired result. Thus, the proof is completed. □
Appendix A.5. Proof of Theorem 4
Proof.
In this appendix, we mainly prove results (a) and (b) separately.
- Proof of part (a)
Define as (1) with being replaced by . Let denote (30) with and being replaced by and , respectively. By adding and subtracting an additional term , we have
Note that by applying the Law of Large Numbers and Theorem 1 to the denominator terms of and , we have that as ,
To complete this proof, we examine the numerator terms of and separately.
- Step 1: Examine
The numerator term of can be expressed as
since if . By the Taylor series expansion, we have
then, the first term in the right-hand side of (A20) gives that
where in the second step is defined as , and the last step is due to the Law of Large Numbers.
In addition, by the Taylor series expansion, we have
Then, the second term of (A20) can be expressed as
where the equality holds due to the Law of Large Numbers in the U-statistics. Therefore, combining (A21) and (A22) yields
where
and
- Step 2: Examine
Similarly to Step 1, we only examine the numerator term of because the denominator term comes from (A19).
For the numerator term of , we have
is the kernel and forms the U-statistics. By Conditions (C3) and (C4), , and by the construction in Section 3.3.1, we have . Then, applying Theorem 12.3 in [29] and Slutsky’s theorem yields that, as ,
with .
Finally, combining (A23) and (A24) gives that as ,
with and being the asymptotic variance of defined after Theorem 2.
- Proof of part (b)
Following a similar discussion to that in part (a), we consider the decomposition
By applying the Law of Large Numbers and Theorem 1 to the denominator terms of and , we have that as ,
On the other hand, the numerator term of can be expressed as
In addition, the numerator term of can be written as
where is the kernel of the U-statistic in the numerator term of .
References
- Lawless, J.F. Statistical Models and Methods for Lifetime Data; Wiley: New York, NY, USA, 2003. [Google Scholar]
- Lu, W.; Ying, Z. On semiparametric transformation cure models. Biometrika 2004, 91, 331–343. [Google Scholar] [CrossRef]
- Chen, C.-M.; Shen, P.-S.; Wei, J.C.-C.; Lin, L. A semiparametric mixture cure survival model for left-truncated and right-censored data. Biom. J. 2017, 59, 270–290. [Google Scholar] [CrossRef]
- Amico, M.; Keilegom, I.V.; Legrand, C. The single-index/Cox mixture cure model. Biometrics 2019, 75, 452–462. [Google Scholar] [CrossRef]
- Amico, M.; Keilegom, I.V. Cure models in survival analysis. Annu. Rev. Stat. Its Appl. 2018, 5, 311–342. [Google Scholar] [CrossRef]
- Bertrand, A.; Legrand, C.; Carroll, R.J.; Meester, C.D.; Keilegom, I.V. Inference in a survival cure model with mismeasured covariates using a simulation-extrapolation approach. Biometrika 2017, 104, 31–50. [Google Scholar] [CrossRef][Green Version]
- Ma, Y.; Yin, G. Cure rate model with mismeasured covariates under transformation. J. Am. Stat. Assoc. 2008, 103, 743–756. [Google Scholar] [CrossRef]
- Chen, L.-P. Semiparametric estimation for cure survival model with left-truncated and right-censored data and covariate measurement error. Stat. Probab. Lett. 2019, 154, 108547. [Google Scholar] [CrossRef]
- Beyene, K.M.; El Ghouch, A. Smoothed time-dependent receiver operating characteristic curve for right censored survival data. Stat. Med. 2020, 39, 3373–3396. [Google Scholar] [CrossRef]
- Heagerty, P.J.; Lumley, T.; Pepe, M.S. Time-dependent ROC curves for censored survival data and a diagnostic marker. Biometrics 2000, 56, 337–344. [Google Scholar] [CrossRef]
- Kamarudin, A.N.; Cox, T.; Kolamunnage-Dona, R. Time-dependent ROC curve analysis in medical research: Current methods and applications. BMC Med. Res. Methodol. 2017, 17, 53. [Google Scholar] [CrossRef]
- Li, L.; Greene, T.; Hu, B. A simple method to estimate the time-dependent receiver operating characteristic curve and the area under the curve with right censored data. Stat. Methods Med. Res. 2018, 27, 2264–2278. [Google Scholar] [CrossRef]
- Song, X.; Zhou, X.-H. A semiparametric approach for the covariate specific roc curve with survival outcome. Stat. Sin. 2008, 18, 947–965. [Google Scholar]
- Zhang, Y.; Han, X.; Shao, Y. The ROC of Cox proportional hazards cure models with application in cancer studies. Lifetime Data Anal. 2021, 27, 195–215. [Google Scholar] [CrossRef]
- Amico, M.; Keilegom, I.V.; Han, B. Assessing cure status prediction from survival data using receiver operating characteristic curves. Biometrika 2021, 108, 727–740. [Google Scholar] [CrossRef]
- Kolamunnage-Dona, R.; Kamarudin, A.N. Adjustment for the measurement error in evaluating biomarker performances at baseline for future survival outcomes: Time-dependent receiver operating characteristic curve within a joint modelling framework. Res. Methods Med. Health Sci. 2021, 2, 51–60. [Google Scholar] [CrossRef]
- Crowther, M.J.; Lambert, P.C.; Abrams, K.R. Adjusting for measurement error in baseline prognostic biomarkers included in a time-to-event analysis: A joint modelling approach. BMC Med. Res. Methodol. 2013, 13, 146. [Google Scholar] [CrossRef]
- Nevo, D.; Zucker, D.M.; Tamimic, R.M.; Wang, M. Accounting for measurement error in biomarker data and misclassification of subtypes. Stat. Med. 2016, 35, 5686–5700. [Google Scholar] [CrossRef]
- Mao, M.; Wang, J.-L. Semiparametric efficient estimation for a class of generalized proportional odds cure models. J. Am. Stat. 2010, 105, 302–311. [Google Scholar] [CrossRef]
- Chen, L.-P. Semiparametric estimation for the transformation model with length-biased data and covariate measurement error. J. Stat. Comput. Simul. 2020, 90, 420–442. [Google Scholar] [CrossRef]
- Carroll, R.J.; Ruppert, D.; Stefanski, L.A.; Crainiceanu, C.M. Measurement Error in Nonlinear Model; CRC Press: New York, NY, USA, 2006. [Google Scholar]
- Chen, L.-P.; Yi, G.Y. Semiparametric estimation methods for left-truncated and right-censored survival data with covariate measurement error. Ann. Inst. Stat. Math. 2021, 73, 481–517. [Google Scholar] [CrossRef]
- Elandt-Johnson, R.C.; Johnson, N.L. Survival Models and Data Analysis; John Wiley & Sons: New York, NY, USA, 1980. [Google Scholar]
- Chen, L.-P.; Yi, G.Y. Analysis of noisy survival data with graphical proportional hazards measurement error models. Biometrics 77, 956–969. [CrossRef]
- Hunter, D.R.; Lange, K. Computing estimates in the proportional odds model. Ann. Inst. Stat. Math. 2002, 54, 155–168. [Google Scholar] [CrossRef]
- Zheng, Y.; Cai, T.; Feng, Z. Application of the time-dependent ROC curves for prognostic accuracy with multiple biomarkers. Biometrics 2006, 62, 279–287. [Google Scholar] [CrossRef]
- Carroll, R.J.; Spiegelman, C.H.; Lan, K.K.G.; Bailey, K.T.; Abbot, R.D. On errors-in-variables for binary regression model. Biometrika 1984, 71, 19–25. [Google Scholar] [CrossRef]
- Webb, A.; Ma, J. Cox models with time-varying covariates and partly-interval censoring—A maximum penalised likelihood approach. Stat. Med. 2023, 42, 815–833. [Google Scholar] [CrossRef]
- van der Vaart, A.W. Asymptotic Statistics; Cambridge University Press: New York, NY, USA, 1998. [Google Scholar]
- Kim, J.P.; Lu, W.; Sit, T.; Ying, Z. A unified approach to semiparametric transformation models under general biased sampling schemes. J. Am. Stat. Assoc. 2013, 108, 217–227. [Google Scholar] [CrossRef][Green Version]
- van der Vaart, A.W.; Wellner, J.A. Weak Convergence and Empirical Processes; Springer: New York, NY, USA, 1996. [Google Scholar]
- Su, Y.-R.; Wang, J.-L. Modeling left-truncated and right-censored survival data with longitudinal covariates. Ann. Stat. 2012, 40, 1465–1488. [Google Scholar] [CrossRef]
- Durrett, R. Probability: Theory and Examples; Cambridge University Press: New York, NY, USA, 2010. [Google Scholar]
- Hoeffding, W. A class of statistics with asymptotically normal distribution. Ann. Math. Stat. 1948, 19, 293–325. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).