Next Article in Journal
Advances and Challenges in the Treatment of HPV-Associated Lower Genital Tract Cancers by Immune Checkpoint Blockers: Insights from Basic and Clinical Science
Next Article in Special Issue
Dynamic Prediction of Rectal Cancer Relapse and Mortality Using a Landmarking-Based Machine Learning Model: A Multicenter Retrospective Study from the Italian Society of Surgical Oncology—Colorectal Cancer Network Collaborative Group
Previous Article in Journal
Radiomic Profiling of Orthotopic Mouse Models of Glioblastoma Reveals Histopathological Correlations Associated with Tumour Response to Ionising Radiation
Previous Article in Special Issue
Identifying Rural Hotspots for Head and Neck Cancer Using the Bayesian Mapping Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving the Estimation of Prediction Increment Measures in Logistic and Survival Analysis

by
Danielle M. Enserro
* and
Austin Miller
Department of Biostatistics and Bioinformatics, Roswell Park Comprehensive Cancer Center, Buffalo, NY 14263, USA
*
Author to whom correspondence should be addressed.
Cancers 2025, 17(8), 1259; https://doi.org/10.3390/cancers17081259
Submission received: 11 March 2025 / Revised: 2 April 2025 / Accepted: 7 April 2025 / Published: 8 April 2025
(This article belongs to the Special Issue Application of Biostatistics in Cancer Research)

Simple Summary

Confidence interval estimation of discrimination improvement measures, including the area under the receiver operating characteristic curve, the net reclassification index, and the integrated discrimination improvement, is an area of ongoing research. The most common confidence interval estimation methods employ normal theory. Literature suggests that degeneration of the normal assumption under the null hypothesis exists, and normal theory confidence intervals estimated may be invalid. Bootstrapped confidence intervals do not rely on normal theory assumptions. We examine the performance of discrimination improvement measures in both the logistic and survival regression context through simulation. Normal theory intervals are only appropriate with a strong effect size of the added parameter, and the percentile bootstrap interval exhibits reasonable coverage while maintaining the shortest width in nearly all simulated scenarios, making this interval the most reliable choice. The intent is that these recommendations improve the accuracy in the estimation and the overall assessment of discrimination improvement.

Abstract

Background/Objectives: Proper confidence interval estimation of the area under the receiver operating characteristic curve (AUC), the net reclassification index (NRI), and the integrated discrimination improvement (IDI) is an area of ongoing research. The most common confidence interval estimation methods employ asymptotic theory. However, developments demonstrate that degeneration of the normal distribution assumption under the null hypothesis exists for measures such as the change in AUC (ΔAUC) and IDI, and confidence intervals estimated under the normal distribution assumption may be invalid. We aim to study the performance of confidence intervals derived assuming asymptotic theory and those derived with non-parametric bootstrapping methods. Methods: We examine the performance of ΔAUC, NRI, and IDI in both the logistic and survival regression context. We explore empirical distributions and compare coverage probabilities of asymptotic confidence intervals with those produced from bootstrapping methods through simulation. Results: The primary finding in both the logistic framework and the survival analysis framework is that the percentile CIs performed well regarding coverage, without compromise to their width; this finding was robust in most scenarios. Conclusions: Our results suggest that the asymptotic intervals are only appropriate when a strong effect size of the added parameter exists, and that the percentile bootstrap interval exhibits at least a reasonable coverage while maintaining the shortest width in nearly all simulated scenarios, making this interval the most reliable choice. The intent is that these recommendations improve the accuracy in the estimation and the overall assessment of discrimination improvement.

1. Introduction

In the pursuit of eradicating cancer, a key area of research is cancer prevention and control. Both observational and treatment trials may assess collected patient risk factors for their associations with cancer survival outcomes, such as objective tumor response, progression-free survival, and overall survival. Cancer risk prediction models can be developed and utilized by health care providers to assess a patient’s risk of developing cancer given specific patient characteristics. An example of a model used for breast cancer risk assessment is the National Cancer Institute’s Breast Cancer Risk Assessment Tool (BCRAT), which was originally developed by Mitchell Gail and known as the Gail Model [1].
It is important that established risk prediction models are accurate, reliable, and robust for their application in the population at-risk for cancer development or recurrence. The common practice is to develop the risk prediction model on a discovery data set, followed by assessing the model performance through internal or external validation, or some combination [2,3]. The BCRAT model and its performance were later assessed with external validation in a range of diverse populations [4,5,6,7]. Common statistical modeling techniques for model development and validation include logistic regression for binary outcomes and Cox proportional hazards (PH) regression for time-to-event outcomes [8].
Risk prediction model validation first requires establishing clinical and statistical significance of the marker of interest [9]. Additionally, model calibration and discrimination should be assessed: calibration quantifies how close the predictions are to the observed outcomes while discrimination quantifies the model’s ability to correctly distinguish between events and nonevents [8]. A common area of ongoing research is taking previously developed and validated models and updating them with the addition of new markers to improve both model calibration and discrimination. This paper focuses on the improvement in discrimination by adding a new marker to a standard risk model.

1.1. ΔAUC for Evaluating Discrimination Improvement

There are multiple measures for assessing discrimination improvement. The most common metric used in logistic regression is the difference in area under the receiver operating characteristic curve (AUC), where ΔAUC = AUC(expanded model) − AUC(standard model) [10]. The overall C is an extension of the AUC for survival analysis, and ΔC can be calculated as C(expanded model) − C(standard model) [11]. For comparing nested models, a strong added marker should theoretically increase the AUC or the overall C. Pepe et al. demonstrated that if the model is correct under normality, then the significance of the new marker is equivalent to significant improvement in AUC [9]. However, the magnitude of ΔAUC often fails to recognize some promising new markers, the paradox being that a statistically significant marker will fail to produce a large increase in AUC, especially if the standard model’s baseline AUC is large [12]. This phenomenon motivated a search for new methods for assessing a predictor’s incremental value [13,14].

1.2. NRI, NRI > 0, and IDI: Definitions and Controversaries

Pencina et al. proposed new measures for assessing discrimination in logistic regression including a risk-category-based measure known as the net reclassification index (NRI), the continuous NRI (NRI > 0), and the integrated discrimination improvement (IDI) [12,15]. Extensions of these measures were later derived for survival analysis [16]. The NRI attempts to quantify whether a new variable provides clinically relevant improvement in risk prediction, assuming that a valuable marker will tend to increase predicted risks for events and decrease predicted risks for nonevents [12]. Calculation is based on the categorization of model-based predicted probabilities from models being compared into ordinal categories of absolute risk by event status, which are cross-tabulated in a reclassification table. NRI typically assumes either two or three risk categories, but it can also assume many. In the case that risk categories are not clinically assignable, the definition was extended so that model-based predicted probabilities from the standard prediction model and the updated prediction model can be compared directly with NRI > 0 [15]. Finally, IDI accounts for the size of movements up and down in risk categories, jointly quantifying the overall improvement in sensitivity and specificity over all possible cutoffs on the (0, 1) interval [12].
NRI and NRI > 0 have been subject to criticism. Pepe argued that NRI as a single summary measure is less clinically relevant than the event NRI (eNRI) and nonevent NRI (neNRI) as separate components, and Kerr et al. demonstrate scenarios where the sum of eNRI and neNRI mask detriment by the new model within either the events or nonevents [17,18]. NRI’s interpretation is commonly mistaken as a proportion, and Pepe argues it is better to interpret eNRI and neNRI separately, since eNRI is the net proportion of events assigned a higher risk and neNRI is the net proportion of nonevents assigned a lower risk [17]. Kerr et al. show that NRI > 0 can be dramatically large in magnitude, even for models with an uninformative added marker, thus alluding that null new markers are predictive and falsely leading investigators to incorrectly assign them roles in risk prediction [18]. Hilden and Gerds argue that the estimation of NRI > 0 and IDI may not be valid with poorly calibrated risk models and that model re-calibration may not correct all issues [19,20]. Leening et al. counter-argue that mean calibration of risk prediction models is automatically achieved by any maximum likelihood estimation (MLE) based algorithm and therefore Hilden and Gerds’ example is not applicable [21]. They also reiterate the importance of ensuring properly calibrated models before evaluating the inherent value of an added marker [21]. Additionally, McKeague and Qian show that the empirical distribution of NRI > 0 can be bimodal in nature [22]. With the amount of conflicting literature, investigators should take great care in the estimation and interpretation of NRI, NRI > 0, and IDI, as these measures can be estimated in conjunction with other measures of discrimination improvement such as ΔAUC [17,21].

1.3. Challenges in Estimation of Discrimination Improvement Measures

Proper confidence interval (CI) estimation of these metrics is important to consider. The most common confidence interval estimation methods employ asymptotic theory. However, developments demonstrate that degeneration of the normal distribution assumption under the null hypothesis exists for discrimination improvement metrics, and confidence intervals estimated under the normal distribution assumption may be invalid. Demler et al. demonstrated misuse of the DeLong test comparing AUCs for two nested models, observing that the empirical distribution of ΔAUC was highly right-skewed with the median near zero, suggesting that the distribution of the ΔAUC estimator under the null is dramatically different from the normal distribution assumed by the DeLong test [23]. Additionally, Kerr et al. demonstrated via simulation that the standard error formula of IDI underestimates the standard error, and that the normal theory test testing the null hypothesis of IDI = 0 is not valid due to violation of the normal assumption [24]. We hypothesize that this phenomenon also exists for NRI and NRI > 0. The asymptotic CI formulas rely heavily on the assumption that the distribution of these measures is always normal in nature, which may not be the case, especially for weak added factors. They also rely on proper estimation of standard errors; the developed formulas for the standard errors of NRI and IDI may not be correctly estimating the true variance.
Alternative methods for CI estimation employ bootstrapping techniques and do not assume normality under the null. However, with the potential for large data, deriving CIs from bootstrapping can be computationally intensive, so we must ensure that what we gain from bootstrap CIs is worth the extra time and computer power. In this paper, we examine the performance of ΔAUC, NRI, NRI > 0, and IDI in the binary outcome framework and their extensions in survival analysis. We explore empirical distributions and compare the coverage probabilities for CIs derived from various methods. We consider two event rates, several effect sizes, two sample size options, and two types of censoring for survival analysis. Finally, we make recommendations for CI methods to utilize based on their performance.

2. Methods

2.1. General Framework: Logistic Regression

We let Y be a binary outcome, where Y = 1 for an event of interest and Y = 0 for a nonevent. We define a column vector X of p + q predictor variables conditional on Y that follows a multivariate normal distribution for nonevents X|Y = 0~N0, Σ) and for events X|Y = 1~N1, Σ). While we assume equal variance for events and non-events, we clarify that this is not a necessary assumption. We assume that Y and X are available for all N patients. We also assume that the p + q predictor variables of X are uncorrelated. We use X to predict Y; a projection based on the full set of p + q predictors compared with a projection based on a reduced set of predictors, p, where the reduced model is nested within the full model. In logistic regression, the respective models produce linear coefficient estimates a’ = (a1,…, ap+q), based on the full model, and b’ = (b1,…, bp), based on the reduced model, with risk scores functions of a’Xp+q and b’Xp. We assume here that higher values of risk scores correspond to a higher probability of the event of interest. We wish to test whether the last q coefficients are equal to 0 and to then determine whether the full model discriminates as well as or better between the two subgroups of Y as compared to the reduced model.

2.2. General Framework: Time-to-Event Regression

Logistic regression modeling can be employed in scenarios with full follow-up. While collecting complete follow-up on all participants is ideal, situations without complete follow-up time are more realistic. Survival analysis methods, such as Kaplan–Meier and Cox proportional hazards regressions, have become some of the most used tools in risk prediction modeling [3,8,25]. We assume there is a time-to-event variable T for all individuals, such that Ti represents either when an event has occurred or the time of censoring, where censoring can occur as either loss to follow-up or at the end of the study. All individuals also have an event indicator Y, where Y = 1 for those with the event observed or Y = 0 for those censored. We also have collected data on p + q fixed covariates at baseline in the vector X, made up of x1, …, xp+q. We assume that the censoring is non-informative, meaning that the censoring mechanism is unrelated to the outcome of interest or the covariates, and we also assume right-censoring for the event times. Thus, if the proportionality of hazards assumption is true, we can use the Cox proportional hazards (PH) regression model [26]. The fitted risk model can then be used to estimate the predicted probability of survival within the specified time frame (0, t). Measuring discrimination improvement in survival analysis is complicated by the fact that the event of interest involves a time-to-event component, and the model thus makes predictions about these survival times.

2.3. Measures of Discrimination Improvement

2.3.1. AUC and Overall C

The area under the receiver operating characteristic (ROC) curve (AUC) is defined as the probability that the risk of event for a randomly selected nonevent is less than the risk of event for a randomly selected event. Given a risk score threshold t, sensitivity (or true positive fraction, TPF) is the probability that an event is classified correctly as an event:
Sensitivity = TPF = P(risk score > t | Y = 1)
Specificity is defined as the probability of correctly classifying a nonevent:
Specificity = P(risk score < t | Y = 0)
Each risk score is compared to t for classification of the subject with that risk score to a predicted group, i.e., predicted event vs. predicted nonevent. Subtracting specificity from 1 yields the false positive fraction (FPF):
1 − Specificity = FPF = P(risk score > t) | Y = 0)
Depending on the choice of threshold t, some events and nonevents will be correctly classified while the remainder will be misclassified. Varying t yields different pairs of values of TPF and FPF. Plotting these pairs (FPF, TPF) for all possible choices of t results in the ROC curve. The AUC is the area under the curve, which ranges from 0.5 (no discriminatory ability) to 1.0 (perfect discrimination).
There is no explicit formula for the AUC, so typically it is estimated by the Mann–Whitney statistic [10]. This is a non-parametric, unbiased estimator, referred to as npAUC and known as the C-statistic [25]., e.g., for the full model discussed with risk scores a’Xp+q, the formula is
C     npAUC = 1 n 0 n 1 x i Y 0 x j Y 1 I a x i ,   a x j ,
where Y1 and Y0 are the sets of subjects with and without events, respectively, n1 and n0 are the sizes of the sets, and I is given as follows:
I a x i ,   a x j = 1 ,   i f   a x i < a x j   0.5 ,   i f   a x i = a x j 0 ,   o t h e r w i s e  
DeLong et al. developed an approximate test for testing if two AUCs from different models on the same group of subjects are equal and produced a CI for ∆AUC. The approximate standard error used in both calculations is
S E A U C = 1 , 1 S ( 1 , 1 ) 1 / 2
where S is the estimated variance-covariance matrix of (npAUCp+q, npAUCp) [10]. We use the C-statistic as the AUC estimator in this study. To assess improvement in discrimination, we estimate the difference in area under of the receiver operating characteristic curve (AUC) for two models: ∆AUC = AUC(new or updated model) − AUC(standard model) [10].
We now extend the AUC to the survival analysis context. Using the survival regression method described, we denote the predicted survival time for everyone at baseline, T1, T2, …, Tn. We have two categories of subjects at a given time point T ≤ Tfinal: those who developed the event during the study (Y = 1) and those who were censored (Y = 0). We denote these actual survival times by U1, U2, …, Un. Harrell states that predicted probabilities of survival until any fixed time point (which we denote V1, V2, …, Vn) can be used in place of the corresponding T1, T2, …, Tn if those estimates remain in one-to-one correspondence [25]. M. Pencina et al. show that this point holds true in the most common models in survival analysis, specifically accelerated failure time models and proportional hazards models [11].
Now we consider all pairs of subjects, (i, j), such that i < j, to ensure no repetitions. Thus, for a given pair, we have a concordant pair if Ui < Uj and Vi < Vj or Ui > Uj and Vi > Vj, and we have a discordant pair if Ui < Uj and Vi > Vj or Ui > Uj and Vi < Vj. Since not all pairs will be concordant or discordant due to ties, we can only use the usable pairs of subjects in which at least one had an event in the construction of the C index, resulting in either event vs. event or event vs. nonevent comparisons. The overall C is defined as the proportion of all usable concordant pairs:
C = π c π c + π d
where
π c = P U i < U j   a n d   V i < V j   o r   U i > U j   a n d   V i > V j = P U i < U j   a n d   V i < V j + P ( U i > U j   a n d   V i > V j )
π d = P U i < U j   a n d   V i > V j   o r   U i > U j   a n d   V i < V j = P U i < U j   a n d   V i > V j + P ( U i > U j   a n d   V i < V j )
M. Pencina et al. [11] demonstrate that since i and j are interchangeable in the definitions of πc and πd, and since the V’s have a continuous distribution (assuming at least one predictor is continuous), the overall C index reduces to
C = P ( U i < U j   a n d   V i < V j ) P ( U i < U j ) = P ( V i < V j / U i < U j ) .
A natural estimate for the overall C is
C ^ = 1 Q ( i , j ) O c i j
where O is the set of all usable pairs of subjects (i, j), Q is the total number of usable pairs, and cij takes on the value of 1 for concordant pairs and 0 otherwise [27]. To assess improvement in discrimination, we estimate the difference in overall Cs for two models: ∆C = C(expanded model) − C(standard model).

2.3.2. NRI, NRI > 0, and NRI(t)

Pencina et al. introduced the net reclassification improvement (NRI) to quantify whether a new variable provides clinically relevant improvement in risk prediction, assuming a valuable new marker will tend to increase predicted risks for events and decrease predicted risks for nonevents [12]. Calculation is based on the categorization of model-based predicted probabilities from the standard risk prediction model and the updated prediction model into clinically meaningful ordinal categories of absolute risk, which are cross-tabulated in a reclassification table. The table tabulates events and nonevents separately. Observations are categorized into the table cells based on assigned risk categories by the standard (old) risk model and the updated (new) risk model. Upward movement is defined as moving into a higher category based on the updated model, while downward movement is defined as moving into a lower category. For those where Y = 1, upward movement is preferable while downward movement is problematic. For those where Y = 0, the opposite preferences hold. Thus, the NRI is defined as:
NRI = [P(up|Y = 1) − P(down|Y = 1)] + [P(down|Y = 0) − P(up|Y = 0)].
The quantity in the first bracket is referred to as the event NRI (eNRI), or the net proportion of events assigned a higher risk category. The quantity in the second bracket is referred to as the nonevent NRI (neNRI), or the net proportion of non-events assigned a lower risk category. Each component contains a penalty for any wrong movement in categories. The probabilities can be estimated using sample data obtained from the reclassification table:
P ^ ( up | Y = 1 ) = p ^ u p , Y = 1 = #   e v e n t s   m o v i n g   u p #   e v e n t s
P ^ ( d o w n | Y = 1 ) = p ^ d o w n , Y = 1 = #   e v e n t s   m o v i n g   d o w n #   e v e n t s
P ^ ( up | Y = 0 ) = p ^ u p , Y = 0 = #   n o n e v e n t s   m o v i n g   u p #   n o n e v e n t s
P ^ ( d o w n | Y = 0 ) = p ^ d o w n , Y = 0 = #   n o n e v e n t s   m o v i n g   d o w n #   n o n e v e n t s .
Thus, NRI can be estimated as
N R I ^ = e N R I ^ + n e N R I ^
where
e N R I ^ = ( p ^ u p , Y = 1 p ^ d o w n , Y = 1 )
n e N R I ^ = ( p ^ d o w n , Y = 0 p ^ u p , Y = 0
The asymptotic standard errors for eNRI and neNRI can be estimated as
S E e N R I ^ = p ^ u p , Y = 1 + p ^ d o w n , Y = 1 n 1 ( p ^ u p , Y = 1 p ^ d o w n , Y = 1 ) 2 n 1
S E n e N R I ^ = p ^ d o w n , Y = 0 + p ^ u p , Y = 0 n 0 ( p ^ d o w n , Y = 0 p ^ u p , Y = 0 ) 2 n 0
and
S E N R I ^ = S E ( e N R I ^ ) 2 + S E ( n e N R I ^ ) 2
where n1 is the number of events and n0 is the number of nonevents [12].
NRI with categories typically assumes either two or three risk categories but can assume up to k risk categories. The choice for risk category cutoffs is of much debate. Pencina et al. recommended that clinically useful category cutoffs should be used, but the choice of category cutoff may not be easy to make [12]. Pencina et al. later recommended that under circumstances that the choice in category cutoff is not obvious, investigators should consider using the event rate as the category cutoff, which yields an NRI measure with relationships with metrics in decision analytics and possesses meaningful interpretations [28].
In the scenario where clinically meaningful categories are not assigned, Pencina et al. extended the definition of NRI so that model-based predicted probabilities from the standard prediction model and the updated prediction model can be compared directly [15]. Instead of tabulating upward and downward movement in categories, increases and decreases in predicted probabilities are tabulated instead. This method of calculating NRI is often referred to as continuous NRI (NRI > 0), which can be defined as follows:
NRI > 0 = [P(pnew > pold|Y = 1) − P(pnew < pold|Y = 1)]
– [P(pnew > pold|Y = 0) − P(pnew < pold|Y = 0)],
where pnew are the predicted probabilities from the new updated risk prediction model and pold are the predicted probabilities from the standard risk prediction model. Similar calculations as those shown above in the definition for NRI with categories can be used for estimating the components of NRI > 0; instead of counting the number of upward and downward movements, the number of observations that satisfy the above risk comparisons are assessed, for both events and nonevents.
The asymptotic standard errors can be estimated as
S E e N R I > 0 ^ = p ^ r i s k 1 > r i s k 0 , Y = 1 + p ^ r i s k 1 < r i s k 0 , Y = 1 n 1 ( p ^ r i s k 1 > r i s k 0 , Y = 1 p ^ r i s k 1 < r i s k 0 , Y = 1 ) 2 n 1
S E n e N R I > 0 ^ = p ^ r i s k 1 < r i s k 0 , Y = 0 + p ^ r i s k 1 > r i s k 0 , Y = 0 n 0 ( p ^ r i s k 1 < r i s k 0 , Y = 0 p ^ r i s k 1 > r i s k 0 , Y = 0 ) 2 n 0
S E N R I > 0 ^ = S E ( e N R I > 0 ^ ) 2 + S E ( n e N R I > 0 ^ ) 2
where each p ^ represents the empirical proportions of individuals satisfying their respective criteria and the subscripts risk1 and risk0 represent the predicted probability of event in the updated model and the standard model, respectively [15].
NRI was later extended to survival analysis. Chambless et al. express event NRI (NRIY=1) and nonevent NRI (NRIY=0) as:
N R I Y = 1 = P Z u p Y = 1 P ( Z d o w n | Y = 1 )
N R I Y = 0 = P Z d o w n Y = 0 P ( Z u p | Y = 0 )
where overall NRI is estimated as the weighted average of the two quantities:
N R I w = w · N R I Y = 1 + 1 w · N R I Y = 0 ,
with w between 0 and 1 [16]. A reminder that Z = X’β, and Zup represents upward movement in risk category or increased risk by the new model while Zdown represents downward movement in risk category or decreased risk by the new model. The authors point out that using this notation, the NRI considered by Pencina et al. was 2∙NRI0.5, and that the choice of w can be used to reflect relative costs and benefits of improvement in risk prediction [12,16].
Chambless et al. made some modifications to NRI for use in survival analysis, conditioning on time being less than some t [16]:
N R I Y = 1 ( t ) = P Z u p Y t = 1 P ( Z d o w n | Y ( t ) = 1 ) ,
N R I Y = 0 ( t ) = P Z d o w n Y t = 0 P ( Z u p | Y ( t ) = 0 ) ,
N R I ( t ) w = w · N R I Y = 1 ( t ) + 1 w · N R I Y = 0 ( t ) .
A count estimator for NRI(t) uses the following proportions for event NRI(t) and nonevent NRI(t):
N R I ^ Y = 1 ( t ) c o u n t = I ( Z u p , Y t = 1 ) I ( Z d o w n , Y t = 1 ) I ( Y t = 1 ) ,
N R I ^ Y = 0 ( t ) c o u n t = I ( Z d o w n , Y t = 0 ) I ( Z u p , Y t = 0 ) I ( Y t = 0 ) .
We obtain the components of NRI(t) by applying Bayes’ theorem:
N R I Y = 1 t = P ( Y t = 1 | Z u p ) · P ( Z u p ) P ( Y t = 1 ) P ( Y t = 1 | Z d o w n ) · P ( Z d o w n ) P ( Y t = 1 ) ,
N R I Y = 0 t = P ( Y t = 0 | Z d o w n ) · P ( Z d o w n ) P ( Y t = 0 ) P ( Y t = 0 | Z u p ) · P ( Z u p ) P ( Y t = 0 ) .
Assuming w = 0.5 and p = P(Y = 1), we obtain
N R I = 2 · N R I 0.5 t = P Y t = 1 Z u p p · P Z u p P Y t = 1 Z d o w n p · P Z d o w n p · ( 1 p ) .
Kaplan–Meier estimates are used to estimate P Y t = 1 Z u p , P Y t = 1 Z d o w n , and p = P(Y(t) = 1) to account for censoring, thus giving us our estimate for NRIw(t) [16]. This expression can be used for both NRI with categories (with upward and downward movement in categories of predicted risks) or for continuous NRI (where predicted risks between new and old models are compared directly), which corresponds nicely to the original derivation of NRI [12,15].

2.3.3. IDI and IDI(t)

Pencina et al. also introduced the integrated discrimination improvement (IDI), which does not require categories, and instead accounts for the size of movements up and down [12]. Using the differences between integrals of sensitivities (IS) and integrals of “one minus specificities” (IP) for models with and without the new variable (i.e., new model vs. standard model), the IDI quantifies jointly the overall improvement in sensitivity and specificity over all possible cutoffs on the (0,1) interval. The IDI is defined as
IDI = (ISnew − ISold) − (IPnew − IPold),
where ISnew is the integrated sensitivity over all possible cutoffs for the new model (standard plus a new added variable) and IPnew is the integrated “one minus specificity” over all possible cutoffs for the new model. ISold and IPold are the same quantities for the standard model. Thus, the IDI is the difference between improvement in average sensitivity and any potential increase in average “one minus specificity”.
It has been shown that to estimate IDI using sample data we can use the following equation:
I D I ^ = ( p ^ ¯ n e w , e v e n t s p ^ ¯ o l d , e v e n t s ) ( p ^ ¯ n e w , n o n e v e n t s p ^ ¯ o l d , n o n e v e n t s ) ,
where
p ^ ¯ e v e n t s = i   i n   e v e n t s p ^ i #   e v e n t s
is the mean of the model-based predicted probabilities of the event for those who develop the event and
p ^ ¯ n o n e v e n t s = j   i n   n o n e v e n t s p ^ j #   n o n e v e n t s
is the mean of the model-based predicted probabilities of the event for those who do not have the event. These quantities are estimated for both the standard (old) model and the added variable (new) model.
The standard error can be estimated as
S E I D I ^ = S E ^ e v e n t s 2 + S E ^ n o n e v e n t s 2
where S E ^ e v e n t s is the standard error of paired differences in new and old model-based predicted probabilities among those with the event and S E ^ n o n e v e n t s is the corresponding standard error among those without the event [12].
In the extension of IDI for survival analysis, IS(t) is the average sensitivity at time t and IP(t) is the average (1-specificity) at time t. Chambless et al. show that
I S t I P t = R 2 ( t ) = V a r [ S t Z ] S t · [ 1 S t ]
where S(t|Z) is the survival function, S(t) = E[S(t|Z)], and S(t)[1-S(t)] is the variance of the binomial variable ‘event have time t’ [16]. This ratio is interpreted as the proportion of variance explained by the model and the authors also show that R2(t) is bounded on the interval [0, 1]. This quantity can be estimated from the fitted survival function, such that
R ^ 2 ( t ) = V a r ^ [ S t Z ] S ^ t · [ 1 S ^ t ]
where S ^ t is obtained through Kaplan–Meier estimates. By rearranging the definition of IDI presented by M. Pencina et al. such that
IDI = (ISnew − IPnew) − (ISold − IPold),
we can see the following holds:
IDI(t) = R2(t)new − R2(t)old,
or the difference in the proportion of variance explained by the new and old models [12,16]. We can estimate IDI(t) by taking the difference between the estimators for R2(t)new and R2(t)old.

2.4. Confidence Interval Estimation

Various methods for confidence interval (CI) estimation exist. The most common estimation methods are intervals derived under normal theory assumptions, or asymptotic CIs [29]. We assume there is a one-sample situation with sample data obtained by random sampling from an unknown distribution F, where F → x = (x1, x2, …, xn). Suppose we have a parameter of interest θ such that θ = t(F). Let θ ^ = t( F ^ ) be the plug-in estimate of θ and let s e ^ be a reasonable estimate of the standard error for θ ^ . In most circumstances, as the sample size n increases and grows large, the distribution of θ ^ converges to the normal distribution:
Z = θ ^ θ s e ^ N 0,1
Let z α indicate the 100α quantile and z 1 α indicate the 100(1 − α) quantile of a N(0,1) distribution. The asymptotic CI with coverage probability equal to 1 − 2α (where α is equal to the significance level) for a parameter θ is given by:
θ ^ z 1 α s e ^ ,   θ ^ + z 1 α s e ^   .
SE represents the standard error and z is the critical value from the standard normal distribution [29].
Alternative methods for estimating CIs that do not employ asymptotic theory are CIs estimated with bootstrap techniques. The bootstrap is a data-based simulation method used for statistical inference [29]. The bootstrap distribution of a parameter estimator can be used to calculate a variety of 100(1 − α)% CIs without having to make normal theory assumptions [29]. We explore several types of CIs based on the bootstrap distribution: bootstrap-t, percentile, and bias-corrected percentile (BC), as well as hybrid and bias-corrected and accelerated percentile (BCa) in limited scenarios [29,30,31].

3. Simulation Study Results

For simulations utilizing logistic regression, we defined a vector X of five predictor variables conditional on Y following a multivariate normal distribution such that for nonevents X|Y = 0~N(µ0, Σ) and for events X|Y = 1~N(µ1, Σ), µ0 = (0.0, 0.0, 0.0, 0.0, 0.0) and µ1 = (0.7, 0.0, 0.2, 0.5, 0.8) respectfully. Thus, both x1 and x5 are strong predictors of the risk of Y, x2 is a null predictor, x3 is a weak predictor, and x4 is a moderate predictor. We assumed that x1 is the standard predictor and the four other predictor variables to be new candidate markers. Without loss of generality, we assumed that Σ is the identity matrix with order 5 × 5, which ensures independence among the predictors conditional on event status. We considered event rates of 10% and 50%. We assumed no missing data among N observations and that all N observations were independent. We employed logistic regression to estimate model-based predicted probabilities of the event Y using a function of linear combinations of the predictors described above. We defined the standard model as predicting the risk of event Y from x1 only, resulting in a baseline AUC close to 0.7. We further defined four updated models, each predicting the risk of event Y from x1 plus an additional predictor. For each model, we estimated predicted probabilities of event Y for each observation to estimate the measures of prediction increment quantifying the improvement of the new risk prediction models: ∆AUC, NRI, NRI > 0, and IDI. We considered NRI versions with two and three categories. For 2-category NRI (2catNRI), we chose the event rate as the category cutoff. For 3-category NRI (3catNRI), we used the categories [0.00, 0.05), [0.05, 0.20), and [0.20+) for 10% event rate and [0.00, 0.40), [0.40, 0.60), and [0.60+) for 50% event rate [28].
For the survival analysis, we considered two incidence rates, 10% and 50% (with 90% and 50% expected survival, respectively), over a 10-year follow-up period. We used two different censoring distributions: (1) Type I censoring (time-terminated) and (2) Type II (random censoring; failure-terminated) [8]. We simulated random censoring with the uniform distribution. Failure time assumed a Weibull distribution with shape parameter = 2. The scale parameter varied depending on the target incidence rate and effect size for each model. We defined a vector of five continuous, uncorrelated predictor variables following a multivariate normal distribution, with effect sizes of strong to null. For predictor variables x1, x2, x3, x4, and x5, hazard ratio (HR) values were 2.0, 1.0, 1.2, 1.5, and 2.0, respectively. We assumed x1 to be the established standard predictor, while the four remaining predictors were the candidate variables for model inclusion consideration. We considered the following four comparisons of the standard model (HR = 2.0): (1) standard (HR = 2.0) + null (HR = 1.0), (2) standard (HR = 2.0) + weak (HR = 1.2), (3) standard (HR = 2.0) + moderate (HR = 1.5), (4) standard (HR = 2.0) + strong (HR = 2.0). We assumed that the event status, failure time (or censoring time), and predictor variables were available for all N observations, as well as independence among observations. After assessing validity of the PH assumption, we employed Cox PH regression to estimate model-based predicted probabilities based on the models described above; for each comparison, we estimated the predicted probability according to the standard model and the predicted probability according to the added variable model. Using these pairs of predicted probabilities, we estimated five measures of prediction increment quantifying the improvement of the new risk prediction models: ∆C, 2catNRI, 3catNRI, NRI > 0, and IDI. We applied the same categories for 2catNRI and 3catNRI that were used in the logistic regression simulations [28].
For both logistic regression and survival analysis, we used n = 100,000 (with 1000 iterations) to construct empirical distributions and samples sizes of n = 2000 and n = 300 to construct bootstrapped 95% CIs with 2000 replications. We repeated the sampling 1000 times to assess coverage probabilities. Since derivations for standard error estimates were not previously provided in the survival analysis extensions, we omitted estimation of CIs assuming asymptotic theory in the survival analysis framework. Data simulation and analyses were completed using SAS software, version 9.4 (Copyright 2002–2012 SAS Institute Inc., Cary, NC, USA). We performed bootstrap analysis with the SAS macro jackboot.sas at http://support.sas.com/kb/24982.html (accessed on 15 October 2016).
Figure 1, Figure 2, Figure 3 and Figure 4 illustrate the empirical distributions of ΔAUC, IDI, ∆C, and IDI(t) for three of the added-variable regression models. The empirical distributions of 3catNRI, 2catNRI, and NRI > 0 are in Supplemental Figures S1–S11. Each figure illustrates the empirical distribution as the added marker moves closer to the null. The added marker with moderate effect size is omitted, as there is very little change in the shape of distribution until the null added marker. The empirical distributions under the null hypothesis are right-skewed. The histograms for IDI under the logistic framework demonstrate that the standard error appears to be underestimated. To assess where the issues with degeneration began, we also looked at empirical distributions for added markers with effect sizes of 0.1 and 0.05. We observed no change in the shape of the empirical distributions as compared with those for the weak effect size.
Table 1 and Table 2 summarize coverage probabilities for considered scenarios in logistic regression. Defining acceptable coverage as 94–96%, acceptable coverage occurs for 26% of asymptotic intervals, 36% for BC intervals, 23% for percentile intervals, and 18% for bootstrap-t intervals. Asymptotic, BC, and bootstrap-t interval methods exhibit low coverage (percentage low: 66%, 57%, and 68%, respectively) whereas the percentile method provides excess coverage (73% high). Likewise, none of the prediction increment measures had at least 50% of scenarios with acceptable coverage: ∆AUC with 47%, 3catNRI and 2catNRI each with 9%, NRI > 0 with 44%, and IDI with 34%. Considering combinations of CI type and prediction measures, the BC method has the best performance for NRI > 0 (69% with acceptable coverage), percentile-based for IDI and ∆AUC (63% and 56%, respectively), and BC for IDI and ∆AUC (56% and 50%, respectively). Coverage was generally too low when adding a null marker, except for percentile CIs, and all CI types for ∆AUC, for which it was high. When adding a moderate or strong marker, most CI methods have appropriate coverage for ∆AUC. There is consistently low coverage by the asymptotic CIs for IDI and 3catNRI, exemplifying the potential underestimation of their standard errors. We also note that the bootstrap-t intervals perform consistently poorly for all versions of NRI, except for 3catNRI with the strongest added marker in the largest sample size. Finally, while the percentile intervals appear to exhibit ≥95% coverage in nearly all scenarios, we do note that this method does not seem to perform well for IDI with a null added marker in the 50% event rate situation; the bootstrap-t interval seems to be the better choice. We also considered hybrid bootstrap intervals and BCa intervals in limited scenarios, and the results are summarized in Supplemental Table S1. We did not observe adequate improvements in coverage with either interval type.
Table 3 and Table 4 summarize coverage probabilities of four bootstrapped CI types assuming Type I censoring. Acceptable coverage occurs for 31% of BC intervals, 30% of percentile intervals, 24% of bootstrap-t intervals, and 13% of hybrid intervals. BC, bootstrap-t, and hybrid intervals tend to exhibit low coverage (60%, 63%, and 78% of intervals, respectively) while 68% of percentile intervals tend to have excess coverage. ∆C and IDI have the best chance of acceptable coverage (41%) while both categorical NRIs have the lowest percentages (8% for 3catNRI and 5% for 2catNRI). The BC method has the best performance for NRI > 0 (38% with acceptable coverage) and the percentile method and BC method are best for IDI and ∆C (57% and 50%, respectively, for both measures). In general, coverage is better as the adjusted HR of the added marker increases, with instances of excess or too low of coverage in the weak and null added marker scenarios. Similar results are observed assuming random censoring (Supplemental Tables S2 and S3). The hybrid intervals exhibit some unexpected behavior. The coverage of the hybrid intervals decreases as the adjusted HR of the added marker decreases, with a large increase in coverage under the null. The hybrid interval method adjusts for both bias and skewness, and this interval type has been shown to suffer from low coverage in unexpected situations [29]. In simulation, the expected coverage probability is not achieved under the alternative, and the method appears to over-correct for bias and skewness under the null. For Type I censoring, BC intervals, percentile intervals, and bootstrap-t intervals perform well for ∆C, NRI > 0, and IDI in scenarios with a moderate or strong added marker. For the 50% incidence rate, the hybrid interval also works well for these measures, but there is decreased coverage for the smaller sample size. Overall, issues with acceptable coverage exist for both 3catNRI and 2catNRI. For the null added marker, there are instances of acceptable coverage—most notably hybrid intervals for 3catNRI (assuming large sample size), BC intervals for NRI > 0 (assuming large sample size), and bootstrap-t for IDI (assuming small sample size). For random censoring, there are more noticeable issues with coverage. The BC intervals, percentile intervals, and bootstrap-t intervals still perform well in the moderate to strong added marker scenarios, but there are gaps in good performance. The hybrid interval appears to be a good option for categorical NRI in the null added marker case. Finally, the percentile interval had ≥95% coverage in most scenarios.
Due to the results of the empirical distributions and CI coverage probability analysis, we evaluated bias. Some of the bootstrap methods correct for bias, while others may amplify bias in situations where bias exists and is not properly accounted for. For example, the percentile method performs well for unbiased statistics, but may not perform well when bias is present or for asymmetric sampling distributions. In contrast, the BC method corrects for median bias, not mean bias. Bias of a point estimator θ ^ for a parameter θ is the difference between the expected value of θ ^ and θ; that is [32]:
B i a s θ θ ^ = E θ θ ^ θ
Thus, positive bias exists when the sample estimate is larger than the population estimate and negative bias exists when the opposite holds. Supplementary Figures S12–S23 show boxplots of bias estimates for all simulated scenarios under both the logistic and survival frameworks. The reference line represents zero bias. In general, the bias distributions are skewed to the right. In the logistic framework, variability in bias increases as the effect size increases for all five measures, with NRI > 0 having the largest variability overall. We also see positive bias with larger magnitudes for all measures in the smaller sample size scenarios. In the survival framework, we observe mean bias of zero for the strongest added markers, with positive bias as the adjusted hazard ratio of the added marker decreases towards the null.
Regarding the CI coverage probabilities, the percentile CIs exhibit coverage of more than 95%, while the other methods yield a coverage too low in more than half of scenarios. Potential issues arise for both coverage extremes. Coverage probabilities below 95% indicate that CIs are too narrow and that Type-I error may be inflated. Coverage probabilities substantially above 95% indicate conservative inference and may be wider than expected. We explored the relationship between the observed coverage probabilities and the widths of the CIs. We constructed boxplots demonstrating the distributions of CI widths. The data for the sample size of 300 are presented in Supplemental Figures S24–S38. Overall, the excess coverage of percentile CIs has not been achieved at the expense of their width. Their width is very similar to the width of the bootstrap BC intervals, yet with better coverage. The asymptotic CIs have the lowest width, but this is also reflected in their low coverage probabilities, further evidence of the potential underestimation of the standard error for some metrics. The bootstrap-t intervals not only demonstrate the most variability in interval width, but they also have consistently poor coverage except with strong added markers in large sample sizes. Additionally, we see an increasing trend in the width of the CIs as the effect size of the added marker increases, with this trend appearing to be most apparent for ∆AUC and IDI. This trend is not present for NRI > 0. It appears that, in general, while the percentile intervals frequently had more than 95% coverage, the widths were shorter when compared with the other CIs we examined. If investigators are looking for a “one size fits all” interval method, the percentile bootstrap interval may be an appropriate choice, as this interval performed the strongest in simulation without exhibiting excess width as compared with the other CI methods. This recommendation requires bootstrapping computations, which can easily be performed in SAS or R. The percentile interval method takes some of the least amount of extra time as compared with the other bootstrap methods. We also observed that the BC method works well for ∆AUC, 3catNRI, jsNRI, NRI > 0, and IDI in situations with a strong added marker and some scenarios with a moderate added marker, depending on event rate and sample size. For CI and standard error estimation from bootstrapping, we recommend that a sufficient number of bootstrap resamples is used; we used B = 2000 for 95% CIs. The number of bootstrap resamples required increases as the desired confidence level increases, which also increases computation time if an interval with more that 95% confidence is desired. However, if the investigator has limited time and wishes to use the asymptotic CI formulas, these formulas would best be used in cases with at least moderate added effect size and with using bootstrap-estimated standard errors, most importantly for IDI and NRI as the asymptotic standard error formulas for these metrics produce CIs with low coverage.

4. Conclusions

Discrimination improvement is an important metric to estimate when developing prediction models for assessing cancer risk and establishing new risk factors in oncology research. We summarized several measures used for estimating discrimination improvement: ΔAUC, NRI, NRI > 0, and IDI. We also discussed motivations for their development, definitions, and potential issues with their estimation and use. The motivation for this study was to improve the estimation of these measures, which was studied through simulations targeting confidence interval estimation. The simulation studies performed demonstrate that the assumed normal empirical distributions of these measures break down under the null hypothesis with varying levels of degeneration. This is problematic in estimation of confidence intervals assuming normal theory, which all of these measures assume. Additionally, the asymptotic standard errors for NRI and IDI under the logistic framework appear to be consistently underestimated, decreasing the coverage probability for the asymptotic confidence intervals in nearly all studied scenarios. The primary finding in both the logistic framework and the survival analysis framework is that the percentile CIs performed well regarding coverage, without compromise to their width; this finding was robust in most scenarios. For the simplest recommendation, we suggest the use of the percentile bootstrap interval with enough bootstrap resamples B, with increased B for larger confidence levels of (1 − α)%. The percentile interval method takes the least amount of extra computing time as compared with the other bootstrap methods, and it can be easily estimated with use of SAS, R, or other statistical computing software. However, if investigators wish to apply other bootstrap CI methods, there are other options depending on the desired performance metric, marker strength, incidence rate, and sample size.
There are limitations to this work that should be addressed in future research. While we currently recommend the percentile CIs as the method to use over intervals assuming normal therapy, we acknowledge that further methodological refinements should be considered to possibly enhance the performance of the bootstrapped methods, including a more comprehensive study of bias. We also plan to study the Jackknife technique, acknowledging that it may perform better for small data samples. Additionally, one strength of this work is that we assessed a wide range of possible scenarios to study CI performance in extreme cases. However, we acknowledge that additional simulations should be considered in future work. These should include additional sample size options, other incidence rates, risk models with more predictor variables included, correlated predictors, other censoring distributions, and models accounting for competing risks. We also acknowledge that these recommendations should be tested and validated in real data; at the time writing, we have applied for the use of specific clinical trial data to address this.

Supplementary Materials

Supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers17081259/s1.

Author Contributions

D.M.E., study design, study analysis, writing, review and editing, final approval; A.M., review and editing, final approval. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gail, M.H.; Brinton, L.A.; Byar, D.P.; Corle, D.K.; Green, S.B.; Schairer, C.; Mulvihill, J.J. Projecting individualized probabilities of developing breast cancer for white females who are being examined annually. J. Natl. Cancer Inst. 1989, 81, 1879–1886. [Google Scholar] [CrossRef]
  2. D’Agostino, R.B.; Griffith, J.L.; Schmidt, C.H.; Terrin, N. Measures for evaluating model performance. In Proceedings of the Biometrics Section; American Statistical Association: Alexandria, VA, USA, 1997; pp. 253–258. [Google Scholar]
  3. Steyerberg, E.W. Clinical Prediction Models: A Practical Approach to Development, Validation, and Updating; Springer Science+Business Media: New York, NY, USA, 2010. [Google Scholar]
  4. Spiegelman, D.; Colditz, G.A.; Hunter, D.; Hertzmark, E. Validation of the Gail et al. model for predicting individual breast cancer risk. J. Natl. Cancer Inst. 1994, 86, 600–607. [Google Scholar] [CrossRef]
  5. Rockhill, B.; Spiegelman, D.; Byrne, C.; Hunter, D.J.; Colditz, G.A. Validation of the Gail et al. model of breast cancer risk prediction and implications for chemoprevention. J. Natl. Cancer Inst. 2001, 93, 358–366. [Google Scholar] [CrossRef]
  6. Nickson, C.; Procopio, P.; Velentzis, L.S.; Carr, S.; Devereux, L.; Mann, G.B.; James, P.; Lee, G.; Wellard, C.; Campbell, I. Prospective validation of the NCI Breast Cancer Risk Assessment Tool (Gail Model) on 40,000 Australian women. Breast Cancer Res. BCR 2018, 20, 155. [Google Scholar] [CrossRef] [PubMed]
  7. Kumar, N.; Singh, V.; Mehta, G. Assessment of common risk factors and validation of the Gail model for breast cancer: A hospital-based study from Western India. Tzu Chi Med. J. 2020, 32, 362–366. [Google Scholar] [CrossRef] [PubMed]
  8. Harrell, F.E., Jr. Regression Modeling Strategies: With Applications to Linear Models, Logistic Regression, and Survival Analysis; Springer: New York, NY, USA, 2001. [Google Scholar]
  9. Pepe, M.S.; Kerr, K.F.; Longton, G.; Wang, Z. Testing for improvement in prediction model performance. Stat. Med. 2013, 32, 1467–1482. [Google Scholar] [CrossRef] [PubMed]
  10. DeLong, E.R.; DeLong, D.M.; Clarke-Pearson, D.L. Comparing the areas under two or more correlated receiver operating characteristic curves: A nonparametric approach. Biometrics 1988, 44, 837–845. [Google Scholar] [CrossRef]
  11. Pencina, M.J.; D’Agostino, R.B. Overall C as a measure of discrimination in survival analysis: Model specific population value and confidence interval estimation. Stat. Med. 2004, 23, 2109–2123. [Google Scholar] [CrossRef]
  12. Pencina, M.J.; D’Agostino, R.B., Sr.; D’Agostino, R.B., Jr.; Vasan, R.S. Evaluating the added predictive ability of a new marker: From area under the ROC curve to reclassification and beyond. Stat. Med. 2008, 27, 157–212. [Google Scholar] [CrossRef]
  13. Greenland, P.; O’Malley, P.G. When is a new prediction marker useful? A consideration of lipoprotein-associated phospholipase A2 and C-reactive protein for stroke risk. Arch. Intern. Med. 2005, 165, 2454–2456. [Google Scholar] [CrossRef]
  14. Cook, N.R. Use and misuse of the receiver operating characteristic curve in risk prediction. Circulation 2007, 115, 928–935. [Google Scholar] [CrossRef] [PubMed]
  15. Pencina, M.J.; D’Agostino, R.B., Sr.; Steyerberg, E.W. Extensions of net reclassification improvement calculations to measure usefulness of new biomarkers. Stat. Med. 2011, 30, 11–21. [Google Scholar] [CrossRef]
  16. Chambless, L.E.; Cummiskey, C.P.; Cui, G. Several methods to assess improvement in risk prediction models: Extension to survival analysis. Stat. Med. 2011, 30, 22–38. [Google Scholar] [CrossRef] [PubMed]
  17. Pepe, M.S. Problems with risk reclassification methods for evaluating prediction models. Am. J. Epidemiol. 2011, 173, 1327–1335. [Google Scholar] [CrossRef]
  18. Kerr, K.F.; Wang, Z.; Janes, H.; McClelland, R.L.; Psaty, B.M.; Pepe, M.S. Net reclassification indices for evaluating risk prediction instruments: A critical review. Epidemiology 2014, 25, 114–121. [Google Scholar] [CrossRef]
  19. Hilden, J.; Gerds, T.A. A note on the evaluation of novel biomarkers: Do not rely on integrated discrimination improvement and net reclassification index. Stat. Med. 2014, 33, 3405–3414. [Google Scholar] [CrossRef]
  20. Gerds, T.A.; Hilden, J. Calibration of models is not sufficient to justify NRI. Stat. Med. 2014, 33, 3419–3420. [Google Scholar] [CrossRef] [PubMed]
  21. Leening, M.J.; Steyerberg, E.W.; Van Calster, B.; D’Agostino, R.B., Sr.; Pencina, M.J. Net reclassification improvement and integrated discrimination improvement require calibrated models: Relevance from a marker and model perspective. Stat. Med. 2014, 33, 3415–3418. [Google Scholar] [CrossRef]
  22. McKeague, I.W.; Qian, M. An adaptive resampling test for detecting the presence of significant predictors. J. Am. Stat. Assoc. 2015, 110, 1422–1433. [Google Scholar] [CrossRef]
  23. Demler, O.V.; Pencina, M.J.; D’Agostino, R.B., Sr. Misuse of DeLong test to compare AUCs for nested models. Stat. Med. 2012, 31, 2577–2587. [Google Scholar] [CrossRef]
  24. Kerr, K.F.; McClelland, R.L.; Brown, E.R.; Lumley, T. Evaluating the incremental value of new biomarkers with integrated discrimination improvement. Am. J. Epidemiol. 2011, 174, 364–374. [Google Scholar] [CrossRef] [PubMed]
  25. Harrell, F.E., Jr.; Lee, K.L.; Mark, D.B. Multivariable prognostic models: Issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors. Stat. Med. 1996, 15, 361–387. [Google Scholar] [CrossRef]
  26. Kalbfleisch, J.D.; Prentice, R.L. The Statistical Analysis of Failure Time Data; Wiley: Hoboken, NJ, USA, 1980; pp. 70–118. [Google Scholar]
  27. Nam, B.H.; D’Agostino, R.B. Goodness-of-Fit Tests and Model Validity; Birkhauser: Boston, MA, USA, 2002; Chapter 20. [Google Scholar]
  28. Pencina, M.J.; Steyerberg, E.W.; D’Agostino, R.B., Sr. Net reclassification index at event rate: Properties and relationships. Stat. Med. 2017, 36, 4455–4467. [Google Scholar] [CrossRef]
  29. Efron, B.; Tibshirani, R.J. An Introduction to the Bootstrap; Chapman & Hall: New York, NY, USA, 1993. [Google Scholar]
  30. Efron, B. Nonparametric standard errors and confidence intervals (with discussions). Can. J. Stat. 1981, 9, 139–172. [Google Scholar] [CrossRef]
  31. Hall, P. Theoretical comparison of bootstrap confidence intervals. Ann. Stat. 1988, 16, 927–953. [Google Scholar] [CrossRef]
  32. Casella, G.; Berger, R.L. Statistical Inference; Duxbury: Pacific Grove, CA, USA, 2002. [Google Scholar]
Figure 1. Empirical distribution of ∆AUC. Visual representation of the empirical distribution as the added marker effect moves closer to the null.
Figure 1. Empirical distribution of ∆AUC. Visual representation of the empirical distribution as the added marker effect moves closer to the null.
Cancers 17 01259 g001
Figure 2. Empirical distribution of IDI. Visual representation of the empirical distribution as the added marker effect moves closer to the null.
Figure 2. Empirical distribution of IDI. Visual representation of the empirical distribution as the added marker effect moves closer to the null.
Cancers 17 01259 g002
Figure 3. Empirical distribution of ∆C assuming Type I censoring. Visual representation of the empirical distribution as the added marker effect moves closer to the null.
Figure 3. Empirical distribution of ∆C assuming Type I censoring. Visual representation of the empirical distribution as the added marker effect moves closer to the null.
Cancers 17 01259 g003
Figure 4. Empirical distribution of IDI assuming Type I censoring. Visual representation of the empirical distribution as the added marker effect moves closer to the null.
Figure 4. Empirical distribution of IDI assuming Type I censoring. Visual representation of the empirical distribution as the added marker effect moves closer to the null.
Cancers 17 01259 g004
Table 1. Coverage probabilities for 95% confidence intervals with 10% event rate in logistic framework.
Table 1. Coverage probabilities for 95% confidence intervals with 10% event rate in logistic framework.
n = 2000
Strong New Marker (μY=1 = 0.8)Moderate New Marker (μY=1 = 0.5)
∆AUC3catNRI2catNRINRI > 0IDI∆AUC3catNRI2catNRINRI > 0IDI
Asymptotic93.784.493.494.476.194.381.295.194.575.2
Bias Corrected94.694.591.094.695.594.892.892.494.794.2
Percentile94.797.097.595.595.494.797.499.196.294.0
Bootstrap-t95.194.489.393.795.595.390.789.894.196.1
Weak New Marker (μY=1 = 0.2)Null New Marker (μY=1 = 0.0)
∆AUC3catNRI2catNRINRI > 0IDI∆AUC3catNRI2catNRINRI > 0IDI
Asymptotic90.686.493.595.471.0100.094.694.195.894.0
Bias Corrected93.388.087.693.892.798.490.589.796.095.5
Percentile95.599.599.497.794.898.9100.0100.098.495.9
Bootstrap-t87.484.883.688.787.998.688.186.090.098.4
n = 300
Strong New Marker (μY=1 = 0.8)Moderate New Marker (μY=1 = 0.5)
∆AUC3catNRI2catNRINRI > 0IDI∆AUC3catNRI2catNRINRI > 0IDI
Asymptotic92.680.993.394.276.488.778.692.293.571.8
Bias Corrected94.391.291.895.395.492.387.187.092.392.9
Percentile94.497.699.297.294.995.999.199.798.494.8
Bootstrap-t96.088.887.593.697.386.482.883.087.390.5
Weak New Marker (μY=1 = 0.2)Null New Marker (μY=1 = 0.0)
∆AUC3catNRI2catNRINRI > 0IDI∆AUC3catNRI2catNRINRI > 0IDI
Asymptotic88.479.191.896.475.399.589.993.293.996.8
Bias Corrected91.083.185.589.785.698.488.087.494.694.5
Percentile97.199.9100.097.797.398.8100.0100.098.495.2
Bootstrap-t86.185.082.882.273.396.989.786.789.097.8
3catNRI categories: [0.00, 0.05), [0.05, 0.20), and [0.20+).
2catNRI categories: [0.00, 0.10) and [0.10+).
Table 2. Coverage probabilities for 95% confidence intervals with 50% event rate in logistic framework.
Table 2. Coverage probabilities for 95% confidence intervals with 50% event rate in logistic framework.
n = 2000
Strong New Marker (μY=1 = 0.8)Moderate New Marker (μY=1 = 0.5)
∆AUC3catNRI2catNRINRI > 0IDI∆AUC3catNRI2catNRINRI > 0IDI
Asymptotic95.589.793.895.770.494.590.493.993.669.9
Bias Corrected95.594.294.296.396.094.293.693.094.494.4
Percentile95.596.396.996.496.193.997.196.895.794.3
Bootstrap-t95.593.493.494.596.694.691.791.294.595.2
Weak New Marker (μY=1 = 0.2)Null New Marker (μY=1 = 0.0)
∆AUC3catNRI2catNRINRI > 0IDI∆AUC3catNRI2catNRINRI > 0IDI
Asymptotic95.292.095.394.570.199.995.592.993.891.8
Bias Corrected96.091.891.494.196.197.690.591.295.879.4
Percentile95.499.699.396.195.897.8100.0100.098.078.4
Bootstrap-t95.488.888.494.096.297.888.389.190.696.9
n = 300
Strong New Marker (μY=1 = 0.8)Moderate New Marker (μY=1 = 0.5)
∆AUC3catNRI2catNRINRI > 0IDI∆AUC3catNRI2catNRINRI > 0IDI
Asymptotic94.489.896.193.871.793.189.895.094.371.4
Bias Corrected94.993.594.094.094.394.990.691.594.994.5
Percentile95.098.498.795.993.895.298.299.196.594.8
Bootstrap-t95.991.090.793.796.195.688.688.094.496.4
Weak New Marker (μY=1 = 0.2)Null New Marker (μY=1 = 0.0)
∆AUC3catNRI2catNRINRI > 0IDI∆AUC3catNRI2catNRINRI > 0IDI
Asymptotic86.690.093.396.368.499.994.892.294.193.6
Bias Corrected89.986.889.890.887.098.290.292.194.376.9
Percentile98.8100.0100.099.198.198.4100.0100.097.372.1
Bootstrap-t78.782.185.084.477.398.686.989.089.097.9
3catNRI categories: [0.00, 0.40), [0.40, 0.60), and [0.60+).
2catNRI categories: [0.00, 0.50) and [0.50+).
Table 3. Coverage probabilities for 95% confidence intervals with 10% incidence rate and type I censoring in survival framework.
Table 3. Coverage probabilities for 95% confidence intervals with 10% incidence rate and type I censoring in survival framework.
n = 2000
Strong New Marker (Adjusted HR = 2.0)Moderate New Marker (Adjusted HR = 1.5)
∆C3catNRI2catNRINRI > 0IDI∆C3catNRI2catNRINRI > 0IDI
Bias Corrected95.194.393.193.995.795.692.190.595.194.8
Percentile95.296.998.495.595.895.397.598.496.594.8
Bootstrap-t95.593.292.493.596.296.290.890.294.196.2
Hybrid93.691.892.593.392.889.989.690.593.789.1
Weak New Marker (Adjusted HR = 1.2)Null New Marker (Adjusted HR = 1.0)
∆C3catNRI2catNRINRI > 0IDI∆C3catNRI2catNRINRI > 0IDI
Bias Corrected93.286.089.292.593.298.389.190.195.091.2
Percentile96.999.699.997.695.298.8100.0100.098.292.0
Bootstrap-t91.785.286.991.795.097.984.985.191.998.7
Hybrid75.786.890.091.176.6100.094.290.990.7100.0
n = 300
Strong New Marker (Adjusted HR = 2.0)Moderate New Marker (Adjusted HR = 1.5)
∆C3catNRI2catNRINRI > 0IDI∆C3catNRI2catNRINRI > 0IDI
Bias Corrected94.390.491.092.694.387.985.490.091.791.6
Percentile94.297.899.297.194.595.699.799.898.996.7
Bootstrap-t95.390.989.193.195.287.380.585.590.092.1
Hybrid83.684.788.290.481.377.881.588.988.373.1
Weak New Marker (Adjusted HR = 1.2)Null New Marker (Adjusted HR = 1.0)
∆C3catNRI2catNRINRI > 0IDI∆C3catNRI2catNRINRI > 0IDI
Bias Corrected92.682.589.189.284.598.585.784.393.484.8
Percentile98.3100.0100.098.197.599.099.8100.097.783.2
Bootstrap-t82.173.379.087.382.296.978.177.990.895.2
Hybrid92.491.087.986.566.399.793.384.189.699.7
3catNRI categories: [0.00, 0.05), [0.05, 0.20), and [0.20+).
2catNRI categories: [0.00, 0.10) and [0.10+).
Table 4. Coverage probabilities for 95% confidence intervals with 50% incidence rate and type I censoring in survival framework.
Table 4. Coverage probabilities for 95% confidence intervals with 50% incidence rate and type I censoring in survival framework.
n = 2000
Strong New Marker (Adjusted HR = 2.0)Moderate New Marker (Adjusted HR = 1.5)
∆C3catNRI2catNRINRI > 0IDI∆C3catNRI2catNRINRI > 0IDI
Bias Corrected95.694.195.294.594.595.093.492.994.595.7
Percentile95.996.097.094.994.694.696.997.396.595.3
Bootstrap-t95.793.495.394.294.794.993.092.694.895.4
Hybrid95.993.895.094.394.294.093.292.794.995.0
Weak New Marker (Adjusted HR = 1.2)Null New Marker (Adjusted HR = 1.0)
∆C3catNRI2catNRINRI > 0IDI∆C3catNRI2catNRINRI > 0IDI
Bias Corrected94.891.892.793.194.399.392.886.196.297.1
Percentile94.898.299.194.994.499.4100.0100.098.597.9
Bootstrap-t95.791.790.593.495.397.985.879.491.998.6
Hybrid90.990.592.193.290.1100.095.487.091.9100.0
n = 300
Strong New Marker (Adjusted HR = 2.0)Moderate New Marker (Adjusted HR = 1.5)
∆C3catNRI2catNRINRI > 0IDI∆C3catNRI2catNRINRI > 0IDI
Bias Corrected95.292.792.894.995.095.492.593.493.994.8
Percentile94.897.699.196.095.495.398.899.495.794.9
Bootstrap-t95.192.991.694.595.295.892.591.993.996.3
Hybrid93.392.091.894.192.688.391.392.393.688.5
Weak New Marker (Adjusted HR = 1.2)Null New Marker (Adjusted HR = 1.0)
∆C3catNRI2catNRINRI > 0IDI∆C3catNRI2catNRINRI > 0IDI
Bias Corrected90.487.091.492.890.499.488.776.295.096.9
Percentile98.099.899.998.996.899.5100.0100.098.697.6
Bootstrap-t86.885.386.291.790.397.977.067.791.598.2
Hybrid80.289.689.391.473.199.993.674.291.2100.0
3catNRI categories: [0.00, 0.40), [0.40, 0.60), and [0.60+).
2catNRI categories: [0.00, 0.50) and [0.50+).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Enserro, D.M.; Miller, A. Improving the Estimation of Prediction Increment Measures in Logistic and Survival Analysis. Cancers 2025, 17, 1259. https://doi.org/10.3390/cancers17081259

AMA Style

Enserro DM, Miller A. Improving the Estimation of Prediction Increment Measures in Logistic and Survival Analysis. Cancers. 2025; 17(8):1259. https://doi.org/10.3390/cancers17081259

Chicago/Turabian Style

Enserro, Danielle M., and Austin Miller. 2025. "Improving the Estimation of Prediction Increment Measures in Logistic and Survival Analysis" Cancers 17, no. 8: 1259. https://doi.org/10.3390/cancers17081259

APA Style

Enserro, D. M., & Miller, A. (2025). Improving the Estimation of Prediction Increment Measures in Logistic and Survival Analysis. Cancers, 17(8), 1259. https://doi.org/10.3390/cancers17081259

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop