Skip to main content
Biometrika logoLink to Biometrika
. 2012 Apr 29;99(3):615–630. doi: 10.1093/biomet/ass018

Predictive accuracy of covariates for event times

Li Chen 1, D Y Lin 2, Donglin Zeng 3
PMCID: PMC3635702  PMID: 23843671

Summary

We propose a graphical measure, the generalized negative predictive function, to quantify the predictive accuracy of covariates for survival time or recurrent event times. This new measure characterizes the event-free probabilities over time conditional on a thresholded linear combination of covariates and has direct clinical utility. We show that this function is maximized at the set of covariates truly related to event times and thus can be used to compare the predictive accuracy of different sets of covariates. We construct nonparametric estimators for this function under right censoring and prove that the proposed estimators, upon proper normalization, converge weakly to zero-mean Gaussian processes. To bypass the estimation of complex density functions involved in the asymptotic variances, we adopt the bootstrap approach and establish its validity. Simulation studies demonstrate that the proposed methods perform well in practical situations. Two clinical studies are presented.

Some key words: Censoring, Negative predictive value, Positive predictive value, Prognostic accuracy, Receiver operating characteristic curve, Recurrent event, Survival data, Transformation model

1. Introduction

In cohort studies, potentially censored times until the occurrence of a particular event, such as death or disease recurrence, are often recorded, along with data on certain covariates, including demographic, clinical and genetic factors. An important objective of such studies is to determine which set of covariates is most predictive of event times and thus can be used to select patients for targeted therapy or to refine disease prevention strategies. For example, the gene expression breast cancer study of van de Vijver et al. (2002) was designed to assess whether a 70-gene signature can improve the prediction of survival times and the selection of patients for adjuvant chemotherapy over standard clinical factors. Clinicians are extremely interested in the percentage of patients who can be spared adjuvant chemotherapy, and in the long-term survival probabilities for those patients. To address this type of question, a statistical measure directly related to survival probabilities is needed to quantify the predictive accuracy of covariates and to compare the predictive accuracy of different sets of covariates.

Several methods exist for assessing the predictive accuracy of covariates for survival time or a single event. One approach extends the proportion of variation explained by covariates, R2, for the linear regression model to the proportional hazards model (Korn & Simon, 1990; Schemper, 1990; Schemper & Henderson, 2000), and a second approach is to use a suitable loss function, such as the Brier score, to measure residual variation (Graf et al., 1999). These measures, however, may lack clinical relevance. A more recent approach is to extend the receiver operating characteristic curve for binary response to survival time (Heagerty & Zheng, 2005; Uno et al., 2007). This measure is focused on classification rather than prediction and is not directly related to survival probabilities.

For a binary disease status D and a binary covariate Z, each taking values 0 and 1, standard measures of predictive accuracy include the positive predictive value and negative predictive value, which are defined as ppv = pr(D = 1 | Z = 1) and npv = pr(D = 0 | Z = 0), respectively. The predictive values pertain to the probability of disease or nondisease given a positive or negative result for a dichotomous test and are thus of direct interest to the end users of the test. When Z is continuous and a larger value of Z corresponds to a higher risk of disease, Moskowitz & Pepe (2004) defined the positive and negative predictive curves ppv(υ) = pr{D = 1 | FZ (Z) > υ} and npv(υ) = pr{D = 0 | FZ (Z) ⩽ υ} (0 ⩽ υ ⩽ 1), where FZ (·) is the distribution function of Z. Using FZ (Z) instead of Z yields two appealing properties. First, the positive and negative predictive curves do not depend on the scales upon which the continuous covariates are measured and are thus comparable among different covariates. Second, it can be shown that ppv(υ) = 1 − {pr(D = 0) − υnpv(υ)}(1 − υ)−1, which means that a higher negative predictive curve corresponds to a higher positive predictive curve, so either curve is sufficient to quantify predictive accuracy. These measures were recently extended to survival data by Zheng et al. (2008), who defined time-dependent predictive curves as ppv(t, υ) = pr{Tt | FZ (Z) > υ} and npv(t, υ) = pr{T > t | FZ (Z) ⩽ υ}, where T is the survival time; the extension pertains to a single event with a single covariate. Uno et al. (2007) considered the predictive curves at a fixed time-point for a single event with multiple covariates by standardizing the linear predictor with a known transformation function and proposed inverse probability weighted estimators under completely-at-random censoring.

In the present paper, we propose a new concept, the generalized negative predictive function, to quantify the predictive accuracy of multiple covariates for both survival time and recurrent event times. In the special case of a single event with a single covariate, our function reduces to that of Zheng et al. (2008). We show that the proposed function is maximized at the set of covariates that is truly related to event times and thus can be used to quantify and compare the predictive accuracy of covariates. No such theoretical results are available even for a binary outcome or a single event with a single covariate. We develop nonparametric estimators for the proposed function, allowing censoring to depend on covariates, and establish their large-sample properties. Because a linear combination of covariates is obtained from a possibly misspecified working model to predict event times, the asymptotic distributions of the estimators for the case of multiple covariates are considerably more complicated than the case of a single covariate.

2. Generalized negative predictive function

Let N*(t) be the counting process which records the number of events the subject has experienced by time t. Let X be the vector of all potential covariates, and Z be a p × 1 subset of X. We wish to quantify the ability of Z in predicting N*(·).

We first consider the case where Z is a single covariate. Suppose that a larger value of Z corresponds to a higher event rate. Then at each time-point u ∈ (0, t], we can define a negative predictive curve, pr{dN*(u) = 0 | Y*(u) = 1, FZ (Z) ⩽ υ}, for the subjects who are at risk at that time, where Y*(u) is a process indicating whether the subject is at risk for experiencing an event at time u. By taking the product integral of these curves in the time interval (0, t], we define the generalized negative predictive function of Z as

G(t,υ)=0utpr{dN*(u)=0|Y*(u)=1,FZ(Z)υ}. (1)

For single-event data, N*(·) has a single jump at the event time T and Y*(t) = I (Tt), so that equation (1) reduces to npv(t, υ) = pr{T > t | FZ (Z) ⩽ υ} of Zheng et al. (2008). For recurrent event data, equation (1) is equivalent to

G(t,υ)=exp[0tE{dN*(u)|Y*(u)=1,FZ(Z)υ}],

which can be rewritten as exp[−E{N*(t) | FZ (Z) ⩽ υ}] because Y* (·) = 1; therefore, − log{G(t, υ)} can be interpreted as the predictive number of events by time t in the subgroup with FZ (Z) ⩽ υ. If the counting process has a Poisson structure such that the occurrence of an event is independent of the prior event history conditional on Z, then (1) can also be interpreted as the probability of no event by time t in the subgroup with FZ (Z) ⩽ υ.

We now consider the case that Z includes multiple covariates. To use Z to predict N*(·), a common approach is to obtain a linear combination of Z under a regression model. We consider a semiparametric working model in the form of

E{N*(t)|Z}=g(t,βTZ), (2)

where β is a p × 1 vector of unknown regression parameters, and g(t, υ) is a strictly increasing function of υ for all t > 0. For single-event data, we can use the class of semiparametric transformation models (Dabrowska & Doksum, 1988; Zeng & Lin, 2006), which includes the proportional hazards and proportional odds models as special cases. For recurrent event data, we can use the class of semiparametric transformation mean models (Lin et al., 2001).

Let β̂ be the estimator of β described in Zeng & Lin (2006). For single-event data and recurrent events with a Poisson structure, β̂ is the nonparametric maximum likelihood estimator. For recurrent events without a Poisson structure, β̂ is viewed as a maximum pseudo-likelihood estimator. It can be shown that β̂ converges to some constant vector β* under mild regularity conditions even when model (2) is misspecified. The predictive accuracy of Z can be quantified by the generalized negative predictive function of β*TZ,

GZ,β*(t,υ)=0utpr{dN*(u)=0|Y*(u)=1,Fβ*TZ(β*TZ)υ}, (3)

which is equal to pr{T > t | Fβ*TZ (β*TZ) ⩽ υ} for single-event data, and to exp[−E{N*(t) | Fβ*TZ (β*TZ) ⩽ υ}] for recurrent event data.

As a function of both t and υ, the generalized negative predictive function can be plotted against the cut-off percentage υ for a specific t. By comparing such plots for two sets of covariates, one can tell the difference in the event-free probability at time t between the two groups that include υ percentage of low-risk subjects as defined by the two sets of covariates. The low-risk group defined by the more predictive set of covariates has a higher survival probability. As in the case of the receiving operating characteristic curve, the area under the curve can be used as a summary measure. One can set a cut-off of survival probability to define low risk. For breast cancer patients, the low-risk group is usually defined to have a 5-year survival probability of at least 0.95. Then drawing a horizontal line at 0.95 would yield the percentage of patients who can be defined as low risk. For this type of application, the function is required to be monotone. The proposed function can also be plotted against t for a specific υ. By comparing such plots for two sets of covariates, one can tell how the difference in the event-free probability changes over time for the two low-risk groups defined by the two sets of covariates.

We now provide a formal justification for using the generalized negative predictive function to compare the predictive accuracy of different sets of covariates. Using Lemma A1 of Appendix 1, we can obtain an optimality property of GZ,β*(t, υ) that this function is maximized at the true linear combination of covariates and that the maximization holds uniformly for all t and υ. Specifically, suppose that Z0 is the subset of X truly related to N*(·) in that E{N*(t) | X} depends on Z0 only and that β0 is the true value of β associated with Z0 under model (2). Then GZ0,β0 (t, υ) ⩾ GZ,β (t, υ) for all t > 0 and υ ∈ (0, 1).

3. Inference procedures

3.1. Preliminaries

Let C denote the censoring time. In the presence of censoring, the observed counting process is N(t) = N*(tC) and the at-risk process is Y (t) = Y*(t)I (Ct), where ab = min(a, b), and I (·) is the indicator function. The data consist of n independent replicates {Ni (t), Yi (t), Xi : t ∈ [0, τ]} (i = 1, …, n), where τ is the endpoint of the study. We use the estimator β̂ from the working model (2) to construct nonparametric estimators for GZ,β* (t, υ). We consider the situation that C is independent of N*(·) and X, as well as the situation that C is independent of N*(·) conditional on X. We refer to these two censoring mechanisms as completely-at-random censoring and covariate-dependent censoring, respectively.

3.2. Single-event data under completely-at-random censoring

Under completely-at-random censoring, pr{dN*(u) = 0 | Y* (u) = 1, Fβ*TZ (β*T Z) ⩽ υ} = pr{dN(u) = 0 | Y (u) = 1, Fβ*TZ (β*T Z) ⩽ υ}. Thus, we estimate GZ,β* (t, υ) by

G^(t,υ)=0<ut[1i=1nI{Fn(β^TZi)υ}dNi(u)i=1nI{Fn(β^TZi)υ}Yi(u)],

where Fn(·) is the empirical distribution of β̂TZ. Note that Ĝ(t, υ) is the Kaplan–Meier estimator of the survival function in the subgroup with Fn(β̂TZ) ⩽ υ.

In Appendix 2, we show that the process n1/2{Ĝ(t, υ) − GZ,β* (t, υ)} converges weakly to a zero-mean Gaussian process and is asymptotically equivalent to the process

n1/2i=1nξ1i(t,υ)+n1/2{G^(t;c^υ,β^)G^(t;cυ,β*)}, (4)

where

ξ1i(t,υ)=GZ,β*(t,υ)I(β*TZicυ)0tdNi(u)Yi(u)dΛ(t;cυ,β*)E{I(β*TZcυ)Y(u)},
G(t;c,β)=0<ut[1E{I(βTZc)dN*(u)}E{I(βTZc)Y*(u)}],
G^(t;c,β)=0<ut{1i=1nI(βTZic)dNi(u)i=1nI(βTZic)Yi(u)},

cυ=Fβ*TZ1(υ), c^υ=Fn1(υ), and dΛ(t; c, β) = E{dN*(t) | Y*(t) = 1, βTZc}. The second term in (4) is asymptotically equivalent to n1/2i=1nξ2i(t,υ), where ξ2i (t, υ) is the value of ξ2(t, υ), defined in equation (A5) of Appendix 2, for the ith subject. Note that ξ2(t, υ) involves conditional density functions given multi-dimensional covariates, which are difficult to estimate directly. To avoid this difficulty, we approximate the asymptotic distribution by bootstrapping the observations {Ni (t), Yi (t), Xi : t ∈ [0, τ]} (i = 1, …, n). We recommend using 1000 bootstrap samples. The validity of the bootstrap is proved in Appendix 4.

3.3. Single-event data under covariate-dependent censoring

When censoring is not completely at random, the Kaplan–Meier type estimator for the generalized negative predictive function presented in § 3.2 is no longer consistent because pr{dN*(u) = 0 | Y*(u) = 1, Fβ*TZ (β*TZ) ⩽ υ} ≠ pr{dN(u) = 0 | Y (u) = 1, Fβ*TZ (β*TZ) ⩽ υ}. We decompose the set of all potential covariates X into two parts, U and W, where U is known to be independent of C, and C is independent of N*(·) conditional on W. Let q be the number of continuous components in W. Note that pr{T > t | Fβ*TZ (β*T Z) ⩽ υ} = E[E{I (T > t, β*T Zcυ) | W}]/υ = E{S(t; cυ, β*, W)I (β*T Zcυ)}/υ, where S(t; c, β, w) = pr(T > t | βT Zc, W = w). Since the cumulative hazard function corresponding to S(t; c, β, w), denoted by Λ(t; c, β, w), is equal to 0tE{dN(u)I(βTZc)|W=w}/E{I(βTZc)Y(u)|W=w}, we propose to estimate S(t; c, β, w) by a kernel-smoothed Kaplan–Meier type estimator

S^(t;c,β,w)=0<ut[1i=1nK{(Wiw)/h}I(βTZic)dNi(u)i=1nK{(Wiw)/h}I(βTZic)Yi(u)],

where K(x) = (||xc||) I (xd = 0), xd and xc are the discrete and continuous components of x, respectively, || · || is the Euclidean norm, (·) is a kernel function for which there exists some integer l such that ∫ yj (y) dy = 0 (j = 1, …, l − 1), ∫ yl (y) dy ≠ 0 and 2lq > 0, and h is a bandwidth such that nhq → ∞ and nh2l → 0 as n → ∞. When q ⩽ 3, we can use a symmetric smooth probability density function as the kernel function. In light of the relationship between S(t; cυ, β*, W) and GZ,β* (t, υ), we can then estimate GZ,β* (t, υ) by

G˜(t,υ)=υ1n1i=1nS^(t;c^υ,β^,Wi)I(β^TZic^υ).

In Appendix 3, we show that the process n1/2{(t, υ) − GZ,β* (t, υ)} converges weakly to a zero-mean Gaussian process. As in § 3.2, the asymptotic distribution can be approximated by the bootstrap method.

Due to the curse of dimensionality and the difficulty in choosing an appropriate kernel function for q > 3, the proposed method may not be very useful when q is large. For q > 3, we suggest obtaining a linear combination of W under a proportional hazards model for the censoring time and then applying the proposed method to the linear combination of W.

3.4. Recurrent event data

Under completely-at-random censoring, the inference procedures developed in § 3.2 can be readily applied to recurrent event data because the formulae do not involve specific forms of single-event data. Under covariate-dependent censoring, the fact that Y* (·) = 1 for recurrent event data yields

GZ,β*(t,υ)=exp[υ1E{μ(t;cυ,β*,W)I(β*TZcυ)}],

where μ(t;c,β,w)=0tE{dN*(u)|W=w,βTZc}, which can be consistently estimated by

μ^(t;c,β,w)=uti=1nK{(Wiw)/h}I(βTZic)dNi(u)i1nK{(Wiw)/h}I(βTZic)Yi(u).

Thus, we propose to estimate GZ,β* (t, υ) by

G˜(t,υ)=exp{υ1n1i=1nμ^(t;c^υ,β^,Wi)I(β^TZic^υ)}.

We show in Appendix 3 that the process n1/2{ (t, υ) − GZ,β* (t, υ)} converges weakly to a zero-mean Gaussian process. The asymptotic distribution can again be approximated by the bootstrap method.

4. Simulation studies

The first set of simulation studies concerns the performance of the proposed estimators under completely-at-random censoring. To mimic the breast cancer data of § 5.1, we generated survival times from the Weibull regression model λ(t | X) = −1(t/λ)k−1 exp(βT X), where λ(t | X) is the conditional hazard function of T given X, X is zero-mean trivariate normal with a covariance matrix that is the sample covariance matrix of age, tumor size and gene score of the breast cancer data, and the parameters k, λ, β are estimated from the breast cancer data. We also generated recurrent event times from the random-effect intensity model: λ(t | X, ξ) = ξkλ−1(t/λ)k−1 exp(βTX), where λ(t | X, ξ) is the conditional intensity function of N*(t) given X and ξ, and ξ is a gamma random variable with mean 1 and standard deviation 0.5. For completely-at-random censoring, we generated censoring times from a Weibull distribution whose parameters are estimated from the breast cancer data. Table 1 summarizes the results for the estimator Ĝ(t, υ) at υ = 0.25, 0.5, 0.75 and 1.0 for t = 5 years with n = 300. The parameter estimator has little bias. The standard error estimator reflects the true variability very well, and the confidence intervals have proper coverage probabilities. We also evaluated (t, υ), the estimator allowing covariate-dependent censoring, and found it to be nearly as efficient as Ĝ(t, υ). For survival data, we also evaluated the inverse probability weighted estimator of Uno et al. (2007). The relative efficiencies of Ĝ(t, υ) to Uno et al.’s estimator were found to be 3.80, 2.02, 1.34 and 1.0 for υ = 0.25, 0.5, 0.75 and 1.0, respectively.

Table 1.

Simulation results for the estimator Ĝ(t, υ) for data under completely-at-random censoring

υ Survival data Recurrent event data
True Bias SE SEE CP True Bias SE SEE CP
0.25 97.0 0.1 2.0 2.0 93 97.0 −0.0 2.1 2.0 91
0.50 94.9 0.1 1.8 1.9 96 94.8 −0.0 1.9 1.9 96
0.75 92.0 0.1 1.8 1.9 96 91.9 −0.0 1.9 2.0 96
1.0 85.6 −0.0 2.1 2.1 94 83.8 −0.0 2.5 2.5 94

True, the true value (×100) of G(t = 5 years, υ); Bias, the sampling bias (×100); SE, the sampling standard error (×100); SEE, the sampling mean (×100) of the standard error estimator; CP, the coverage probability (×100) of the 95% confidence interval. Each entry is based on 2000 replicates.

To assess the performance of the proposed estimators under covariate-dependent censoring, we adopted the above set-up but generated censoring times from the transformation model Λ(t|X)=H{0.1texp(βci=1mXi)}, where Λ(t | X) is the cumulative hazard function of C conditional on X, H(·) is a specific increasing function, and X1, …, Xm are standard normal variables with 0.3 correlations with the gene score. To compare the estimators under completely-at-random censoring and covariate-dependent censoring, we chose H(x) = x, m = 3 and βc = 0.5 to produce a censoring rate of 70.2%, which is close to that of the breast cancer data. We used the 3-dimensional independent Gaussian kernel density and set the bandwidth to σn−7/24, where σ is the maximum of the standard deviations of X1, X2 and X3. As shown in Table 2, the estimator (t, υ) has very small bias; the variance estimator is accurate and the confidence intervals have proper coverage probabilities. In contrast, the estimator Ĝ(t, υ) has severe bias and improper confidence intervals. We also evaluated the inverse probability weighted estimator of Uno et al. (2007), which turned out to be severely biased and inefficient, the bias being −0.239, −0.138, −0.066, and −0.009 for υ = 0.25, 0.5, 0.75 and 1.0, respectively, and the corresponding standard errors being 0.086, 0.052, 0.033 and 0.025.

Table 2.

Simulation results for the estimators G̃ (t, υ) and Ĝ(t, υ) for survival data under covariate-dependent censoring

n υ True (t, υ) Ĝ(t, υ)
Bias SE SEE CP Bias SE SEE CP
300 0.25 97.0 −0.1 3.3 2.9 96 −0.1 2.7 2.6 87
0.50 94.9 −0.0 2.7 2.6 96 −0.2 2.4 2.4 96
0.75 92.0 −0.1 2.5 2.5 96 −0.4 2.3 2.4 96
1.0 85.6 −0.2 2.6 2.5 94 −1.5 2.5 2.5 92
500 0.25 97.0 −0.0 2.5 2.2 95 −0.1 2.1 2.1 93
0.50 94.9 −0.0 2.1 2.0 95 −0.2 1.8 1.9 96
0.75 92.0 0.0 2.0 1.9 95 −0.5 1.8 1.8 96
1.0 85.6 −0.1 2.1 2.0 94 −1.4 2.0 2.0 90

True, the true value (×100) of G(t = 5 years, υ); Bias, the sampling bias (×100); SE, the sampling standard error (×100); SEE, the sampling mean (×100) of the standard error estimator; CP, the coverage probability (×100) of the 95% confidence interval. Each entry is based on 2000 replicates.

Finally, we considered the situation that censoring depends on a large number of covariates. We increased the number of covariates in the above censoring model to 10 or 20 and set H(x) = x or H(x) = log(1 + x). We used the one-dimensional Gaussian kernel density and set the bandwidth to σn1/3, where σ is the standard deviation of the linear predictor in the proportional hazards model for the censoring time. The results for βc = 0.1 and m = 10 and 20 are summarized in the Supplementary Material. The proposed method performs well even when the censoring model is incorrectly specified.

5. Examples

5.1. Gene expressions in breast cancer patients

van de Vijver et al. (2002) developed a database of 295 breast cancer patients. In this dataset, the time to death along with a 70-gene signature and some clinical factors were recorded. A total of 276 patients had complete information on the gene signature and clinical factors, and the survival times for 73.6% of them were censored. The main purpose of this study was to determine whether the 70-gene signature can improve the prediction of survival times above the clinical factors and thus be used to guide the selection of patients for adjuvant chemotherapy. We use this dataset to illustrate the generalized negative predictive function.

For our illustration, we compare three sets of covariates: the combination of age and tumor size, the 70-gene signature alone, and the combination of the gene signature, age and tumor size. We consider the proportional hazards models with these three sets of covariates. Because the censoring time depends on tumor size, the p-value being 0.0009, we allow censoring to depend on tumor size in our analysis. Figure 1 displays the 5-year negative predictive curves for the three sets. The gene signature indeed improves predictive accuracy above clinical factors and the gene signature alone is enough for prediction. If clinicians wish to spare 50% of the breast cancer patients adjuvant chemotheraphy, then those patients selected by the combination of the gene signature, age and tumor size will have a 5-year survival probability of 0.962, with a 95% confidence interval of (0.920, 0.992), whereas those patients selected by age and tumor size will have a 5-year survival probability of only 0.89. The area under the curve for the combination of the gene signature, age and tumor size is 0.945, which is significantly larger than that of the combination of age and tumor size and is very close to that of the gene signature alone, the two differences being 0.050 and 0.001, respectively, and the corresponding 95% confidence intervals being (0.020, 0.083) and (−0.008, 0.013). Because the effect of dependent censoring is fairly small in this case, the curves under completely-at-random censoring are very similar to their counterparts in Fig. 1 and are omitted.

Fig. 1.

Fig. 1

Estimation of G(t, υ) at t = 5 years in the breast cancer data. The black solid, dashed and dotted curves pertain to the combination of the gene signature, age and tumor size, gene signature alone, and the combination of age and tumor size, respectively. The grey curves are the 95% pointwise confidence limits corresponding to the black solid curve.

We now demonstrate how to use the negative predictive curve to guide the selection of patients for adjuvant chemotherapy. Because of the toxicity of chemotherapy, low-risk patients should be spared adjuvant chemotherapy. Suppose that we define low risk to be a 5-year survival probability of at least 95%. Then what percentage of patients will be assigned to the low-risk group and thus spared adjuvant chemotherapy based on the 70-gene signature? The answer is given by G−1 (t = 5, 0.95) = 58.0%, with a 95% confidence interval of (34.4%, 70.7%). In contrast, this percentage is only 10.5%, with a 95% confidence interval of (1.1%, 31.2%) based on age and tumor size. Thus, the use of the 70-gene signature can spare substantially more patients adjuvant chemotherapy.

Finally, we compare our negative predictive curves with those of Uno et al. (2007). Figure 2 displays the estimates of the curves of Uno et al. (2007), for which the linear predictor is standardized by the transformation function g(y) = 1 − exp{− exp(y)}. Figure 2 shows that the x-values of the curves for different linear predictors have different ranges, although the transformation function restricts the linear predictors to the interval (0, 1). In contrast, the x-values of our curves, as shown in Fig. 1, always have the range (0, 1). In Fig. 2, the curve for the combination of the gene signature, age and tumor size is considerably lower than that of the gene signature alone. This is not sensible because the combination of the gene signature, age and tumor size should at least have the same predictive accuracy as the gene signature alone. Therefore, the negative predictive curves standardized by a known transformation function cannot be used to directly compare the predictive accuracy of different covariates. Figure 2 also shows that the inverse probability weighted estimator can yield inappropriate estimates for small cut-off values.

Fig. 2.

Fig. 2

Estimates of the npv (t = 5 years, υ) of Uno et al. (2007) in the breast cancer data. The black solid, dashed and dotted curves pertain to the combination of the gene signature, age and tumor size, gene signature alone, and the combination of age and tumor size, respectively. (a) is based on the inverse probability weighted estimator of Uno et al. (2007); (b) is based on the Kaplan–Meier type estimator.

5.2. Exacerbations of respiratory symptoms in patients with cystic fibrosis

Patients with cystic fibrosis often suffer from repeated exacerbations of respiratory symptoms. A randomized clinical trial was conducted to evaluate the efficacy of rhDNase, a highly purified recombinant enzyme, in reducing the rate of of exacerbations (Therneau & Hamilton, 1997). By the end of the trial, 139 of 324 untreated patients and 104 of 321 treated patients had experienced at least one exacerbation; 42 untreated patients and 39 treated patients had at least two exacerbations. In addition to the times to exacerbations, the baseline level of forced expiratory volume in 1 second was recorded for each patient.

We use the proposed methods to evaluate the importance of forced expiratory volume in predicting the number of exacerbations. We compare the predictive accuracy of forced expiratory volume and treatment to that of treatment alone. We consider the class of transformation mean models for these two sets of covariates. The choices of the Box–Cox transformations H(x) = {(1 + x)ρ − 1}/ρ with ρ = 0.5, 1, 2 and logarithmic transformations H(x) = log(1 + rx)/r with r = 0.5, 1, 2 yield highly similar estimates for G(t, υ). Since treatment is binary, G(t, υ) for the two sets of covariates are comparable only at the threshold υ = 0.498, the proportion of treated patients. Figure 3 displays the estimated G(t, υ) versus t at υ = 0.498 based on the proposed estimator under completely-at-random censoring. The model with forced expiratory volume and treatment has much higher generalized negative predictive values than the model with treatment alone, especially for large t, indicating that forced expiratory volume plays an important role in predicting the number of exacerbations. The results based on the estimator allowing the censoring to depend on forced expiratory volume and treatment are virtually the same and omitted.

Fig. 3.

Fig. 3

Estimates of G(t, υ) at υ = 0.498 in the rhDNase study. The black solid and black dash-dotted curves pertain to the point estimates for the model with forced expiratory volume and treatment and the model with treatment alone, respectively. The grey curves pertain to the 95% pointwise confidence limits associated with the black solid curve.

6. Remarks

In the proposed methods, the working model is used only to obtain the coefficients for the linear combination of multiple covariates. The estimation of the generalized negative predictive function itself is purely nonparametric, so the predictive accuracy of different sets of covariates can be compared in a fair manner. Additional simulation studies revealed that the performance of the proposed methods is robust to misspecification of the transformation function. In real studies, the estimates of the proposed function, especially the relative magnitudes of the estimates for different sets of covariates, have been found to be similar under different transformation functions.

We have focused on the class of transformation models and its nonparametric maximum likelihood estimators (Zeng & Lin, 2006). We can also consider other survival models and other estimators provided that the influence functions for the regression parameters exist. Indeed, our methods are designed to assess the predictive accuracy of any model, whether it holds or not. If there is only a single covariate or the regression parameters are estimated from a previous study, then the influence functions for the regression parameters are set to zero in the proposed methods. Our work was partly motivated by the desire to compare the predictive accuracy of five gene-expression signatures for breast cancer (Fan et al., 2006), in which case the coefficients for each linear predictor were determined by a previous study and thus considered fixed in our estimation of the generalized negative predictive function; the results will be reported elsewhere.

Uno et al. (2007) considered the positive and negative predictive curves by thresholding the linear predictor with a known transformation function. As shown in § 5.1, such curves cannot be used to directly compare the predictive accuracy of different covariates. In addition, the inverse probability weighted estimator requires completely-at-random censoring and can yield severe bias under covariate-dependent censoring. Even under completely-at-random censoring, the inverse probability weighted estimator is less efficient than our Kaplan–Meier type estimator, especially for small cut-off values. Our proposed function thresholds the linear predictor by its distribution function and can be used to compare the predictive accuracy of different covariates. The use of the distribution function of the linear predictor creates additional challenges in deriving the asymptotic properties since the distribution function needs to be estimated and the additional variation needs to be accounted for.

The proposed function has nice connections with the receiving operating characteristic curve and overall misspecification rate. It can be shown that, for a given time t if the predictive curve of β1TZ1 is uniformly higher than that of β2TZ2, then the corresponding receiving operating characteristic curve is also uniformly higher, and vice versa. In addition, a uniformly higher predictive curve implies a lower overall misspecification rate. The key difference of the predictive curve from the receiving operating characteristic curve is that it depends on the patient’s survival probabilities and is thus more relevant for prediction. A main purpose of developing a prediction rule is to divide patients with low versus high survival probabilities. The prediction rule based on the receiving operating characteristic curve is not directly connected to survival probabilities, so it is difficult to determine whether the low/high risk groups determined by the prediction rule have high/low survival probabilities. As for the overall misspecification rate, it has the same problem as the receiving operating characteristic curve and does not distinguish between false positive and false negative errors.

There is a growing interest in assessing the predictive accuracy of time-dependent covariates. Henderson et al. (2002) proposed a modification of the explained variation measure of Schemper & Henderson (2000), Schoop et al. (2008) extended the Brier score, and Zheng & Heagerty (2007) extended the receiving operating characteristic method. All these extensions are confined to single-event data. The generalized negative predictive function proposed in this paper can easily incorporate time-dependent covariates by replacing Z in (3) by Z(u). It is straightforward to extend the two estimators for recurrent event data and the estimator for single-event data under completely-at-random censoring to the setting of time-dependent covariates, but the extension is difficult for single-event data under dependent censoring because our estimator relies on the equation GZ,β* (t, υ) = pr{T > t | Fβ*TZ (β*T Z) ⩽ υ}, which does not hold for time-dependent covariates.

Acknowledgments

This research was supported by the National Institutes of Health, U.S.A. The authors thank two referees and an associate editor for their helpful comments.

Appendix 1. Optimality of the generalized negative predictive function

Lemma A1. Let ξ(·) be a stochastic process, and Z0 and Z be subvectors of the random vector X. Assume that the linear combinations β0TZ0 and βTZ are continuous and that E{ξ(t)|X}=g0(t,β0TZ0), where g0(t, υ) is a strictly decreasing function of υ for all t > 0. Then E{ξ(t)|Fβ0TZ0(β0TZ0)υ}E{ξ(t)|FβTZ(βTZ)υ} for all t > 0 and υ ∈ (0, 1). In addition, if there exists some t such that the equality holds for all υ, then pr{Fβ0TZ0(β0TZ0)=FβTZ(βTZ)}=1.

Proof. Let Y0=Fβ0TZ0(β0TZ0), Y = FβTZ (βTZ), and f(t,y)=g0{t,Fβ0TZ01(y)}. Then E{I (Y0υ)} = E{I (Yυ)} = υ, E{ξ(t) | X} = f (t, Y0), and f(t, y) is strictly decreasing in y for all t > 0. Clearly, υE{ξ(t)|Fβ0TZ0(β0TZ0)υ}=E{f(t,Y0)I(Y0υ)}, which can be written as

E{f(t,Y0)I(Y0υ,Yυ)}+E{f(t,Y0)I(Y0υ,Y>υ)}. (A1)

Since f(t, y) is strictly decreasing in y for all t > 0, the second term in (A1) is greater than or equal to f(t, υ)E{I (Y0υ, Y > υ)}, where the equality holds if and only if pr(Y0υ, Y > υ) = 0. Since E{I (Y0υ)} = E{I (Yυ)} = υ, we have E{I (Y0υ, Y > υ)} = E{I (Y0 > υ, Yυ)}. This equality, together with the fact that f(t, y) is strictly decreasing in y for all t > 0, implies that the second term in (A1) is greater than or equal to E{f(t, Y0) I (Y0 > υ, Yυ)}. Thus, (A1) is greater than or equal to E{f(t, Y0) I (Yυ)} or υ E{ξ(t) | FβTZ (βTZ) ⩽ υ}. Hence, the results in Lemma A1 hold.

Remark 1. If β0TZ0 or βTZ is not continuous, then E{ξ(t)|Fβ0TZ0(β0TZ0)υ}E{ξ(t)|FβTZ(βTZ)υ} for all t > 0 and υ in the common range of the distribution functions of β0TZ0 and βTZ.

Appendix 2. Weak convergence of n1/2{Ĝ (t, υ) − GZ,β* (t, υ)}

In this section, we prove the weak convergence of n1/2{Ĝ (t, υ) − GZ,β* (t, υ)} for single-event data. The proof for recurrent events is similar and is omitted.

Let 𝒫n and 𝒫 denote the empirical measure and the distribution under the true model, respectively. For a measurable function f and measure Q, the integral ∫ f dQ is abbreviated as Qf. We assume that there exists a continuous covariate in Z, denoted as Z(k). Let Z(−k) be the remaining components of Z, and βk and β(−k) be the components of β corresponding to Z(k) and Z(−k), respectively. Let h1(t, x; y) and h2(x; y) be, respectively, the conditional densities of (T, Z(k)) and Z(k) given Z(−k) = y. We require that h1(t, x; y) and h2(x; y) are continuously differentiable. Let ψ be the influence function of β̂ in that n1/2(β̂β*) = n1/2 𝒫nψ + op(1).

Clearly,

n1/2{G^(t,υ)GZ,β*(t,υ)}=n1/2{G^(t;cυ,β*)G(t;cυ,β*)}+n1/2{G^(t;c^υ,β^)G^(t;cυ,β*)}. (A2)

The first term in (A2) pertains to the Kaplan–Meier estimator among the subjects with β*TZcυ and is asymptotically equivalent to n1/2(𝒫n𝒫) (t; cυ, β*), where

ξ(t;c,β)=[G(t;c,β)I(βTZc)0tdN(u)Y(u)dΛ(u;c,β)E{I(βTZc)Y(u)}].

By the result of Exercise 14 of van der Vaart & Wellner (1996, p. 152), the class of functions xI (βTxc) with β and c ranging over p and , respectively, is VC-subgraph and is thus 𝒫-Donsker by Theorems 2.6.7 and 2.5.2 of van der Vaart & Wellner (1996). In addition, the classes of functions {N(t) : t ∈ [0, τ]}, {Y (t) : t ∈ [0, τ]}, and {Λ(t; c, β) : t ∈ [0, τ]} are 𝒫-Donsker by Lemma 4.1 of Kosorok (2008). Thus, {ξ(t; c, β) : t ∈ [0, τ], c, βp} is a 𝒫-Donsker class by the preservation properties of 𝒫-Donsker classes (van der Vaart & Wellner, 1996, § 2.10). This result, together with the fact that n1/2{Ĝ (t; ĉυ, β̂) − Ĝ (t; cυ, β*)} = n1/2{G(t; ĉυ, β̂) − G(t; cυ, β*)} + n1/2{Ĝ(t; ĉυ, β̂) − G(t; ĉυ, β̂)} + n1/2{Ĝ(t; cυ, β*) − G(t; cυ, β*)}, implies that the second term in (A2) is asymptotically equivalent to

n1/2{G(t;c^υ,β^)G(t;cυ,β*)}+n1/2(𝒫n𝒫){ξ(t;c^υ,β^)ξ(t;cυ,β*)}. (A3)

As to be shown later, n1/2(ĉυcυ) is weakly convergent. Thus, the second term in (A3) converges uniformly to zero in probability by Lemma 19.24 of van der Vaart (1998). The first term in (A3) is asymptotically equivalent to

G(t;c,β*)c|c=cυn1/2(c^υcυ)+G(t;cυ,β)β|β=β*n1/2(β^β*),

where ∂G(t; c, β)/∂β can be obtained as follows, and ∂G(t; c, β)/∂c in the same manner. Let g1(t, β, c) = E{I (T > t)I (βTZc)}, and g2(β, c) = E{I (βTZc)}. Then for βk > 0, g1(t,β,c)=E[E{I(T>t)I(βTZc)|Z(k)}]=Eβk1(cβ(k)Z(k))th1(s,x;Z(k)) ds dx. Simple algebraic manipulations yield

G(t;c,β)β=g1(t,β,c)/βE{I(T>t)|βTZc}g2(β,c)/βE{I(βTZc)},

where

g1(t,β,c)βk=cβk2Eth1{s,βk1(cβ(k)Z(k));Z(k)}ds,g1(t,β,c)β(k)=βk1E[Z(k)th1{s,βk1(cβ(k)Z(k));Z(k)}ds],g2(β,c)βk=cβk2E[h2{βk1(cβ(k)Z(k));Z(k)}],g2(β,c)β(k)=βk1E[Z(k)h2{βk1(cβ(k)Z(k));Z(k)}].

We now establish the weak convergence of n1/2(ĉυcυ). Clearly,

n1/2{Fn(c)Fβ*TZ(c)}=n1/2(𝒫n𝒫)I(β*TZc)+n1/2𝒫{I(β^TZc)I(β*TZc)}+n1/2(𝒫n𝒫){I(β^TZc)I(β*TZc)}. (A4)

The second term on the right-hand side of (A4) is asymptotically equivalent to ∂g2(β*, c)/∂β n1/2(β̂β*), and the third term converges to zero in probability. It then follows from Lemma 12.8 of Kosorok (2008) that n1/2(ĉυcυ) is asymptotically equivalent to −{dFβ*TZ (c)/dc|c=cυ}−1n1/2(𝒫n𝒫){I (β*TZcυ) + ∂g2(β*, cυ)/∂βψ}.

Combining the above results, we conclude that n1/2{Ĝ(t, υ) − GZ,βL* (t, υ)} is asymptotically equivalent to n1/2𝒫n{ξ1(t, υ) + ξ2(t, υ)}, where ξ1(t, υ) = ξ(t; cυ, β*) and

ξ2(t,υ)={I(β*TZcυ)υ}G(t;c,β*)/cdFβ*TZ(c)/dc|c=cυ{G(t;c,β*)/cdFβ*TZ(c)/dc|c=cυg2(β,cυ)β|β=β*G(t;cυ,β)β|β=β*}ψ. (A5)

Appendix 3. Weak convergence of n1/2{G̃ (t, υ) − GZ,β* (t, υ)}

In this section, we prove the weak convergence of n1/2{ (t, υ) − GZ,β* (t, υ)} for single-event data. The proof for recurrent events is similar and is omitted. We express υn1/2{ (t, υ) − GZ,β* (t, υ)} as

n1/2(𝒫n𝒫){S(t;cυ,β*,W)I(β*Tcυ)}+n1/2𝒫[{S^(t;cυ,β*,W)S(t;cυ,β*,W)}I(β*TZcυ)]+n1/2𝒫{S^(t;c^υ,β^,W)I(β^TZc^υ)S^(t;cυ,β*,W)I(β*TZcυ)}+n1/2(𝒫n𝒫)[{S^(t;c^υ,β^,W)S(t;c^υ,β^,W)}I(β^TZc^υ)]+n1/2(𝒫n𝒫){S(t;c^υ,β^,W)I(β^TZc^υ)S(t;cυ,β*,W)I(β*TZcυ)}. (A6)

To study the second to the fourth terms in (A6), we first study

Λ^(t;c,β,w)=0ti=1nK{(Wiw)/h}I(βTZic)dNi(s)i=1nK{(Wiw)/h}I(βTZic)Yi(s).

Write = TC and Δ = I (TC). Then Λ̂(t; c, β, w) − Λ(t; c, β, w) can be written as

𝒫n[K(Wwh)I(βTZc)ΔI(T˜t)𝒫n{K(Wwh)I(βTZc)Y(u)}|u=T˜]𝒫[I(βTZc)I(T˜t)𝒫{I(βTZc)Y*(u)|W}|u=T˜|W]=(𝒫n𝒫)[K(Wwh)I(βTZc)ΔI(T˜t)𝒫n{K(Wwh)I(βTZc)Y(u)}|u=T˜]𝒫[K(Wwh)I(βTZc)ΔI(T˜t)(𝒫n𝒫){K(Wwh)I(βTZc)Y(u)}|u=T˜𝒫{K(Wwh)I(βTZc)Y(u)}𝒫n{K(Wwh)I(βTZc)Y(u)}|u=T˜]+(𝒫[K(Wwh)I(βTZc)ΔI(T˜t)𝒫{K(Wwh)I(βTZc)Y(u)}|u=T˜]𝒫[I(βTZc)I(T˜t)𝒫{I(βTZc)Y*(u)|W}|u=T˜|W]). (A7)

The third term on the right-hand side of (A7) is O(hl) by a simple transformation and Taylor expansion, together with the fact that 0tE{dN(u)I(βTZc)|W}/E{I(βTZc)Y(u)|W}=0tE{dN*(u)|Y*(u)=1,βTZc,W}.. Thus, by the Duhamel equation and the condition that nh2l = o(1), the second term in (A6) is asymptotically equivalent to n1/2(𝒫n𝒫)ξn(N, Y, Z, W; t, cυ, β*), where

ξn(N,Y,Z,W;t,c,β)=I(βTZc)(𝒫Z,W[I(βTZc)S(t;c,β,W)hqK{wW)/h}×Mn(N,Y;t,c,β,W)])|w=W,Mn(N,Y;t,c,β,w)=0tdN(u)Y(u)dΛ(u;c,β,w)𝒫n[hqK{(Ww)/h}I(βTZc)Y(u)],

and 𝒫Z,W denotes the expectation with respect to Z and W. By the condition that nhq → ∞, it can be shown that, uniformly in t, c and β, ξn(t; c, β) converges in probability to

ξ(t;c,β)=E{I(βTZc)|W}S(t;c,β,W)I(βTZc)0tdN(u)Y(u)dΛ(u;c,β,W)E{I(βTZc)Y(u)|W}.

This result, together with the fact that the class of functions ξn( N, Y, Z, w; t, c, β) indexed by w, t, c and β is 𝒫-Donsker, implies that, conditional on W1, …, Wn, the process n1/2(𝒫n𝒫)ξn(t; c, β) is asymptotically equivalent to n1/2(𝒫n𝒫)ξ(t; c, β) by Theorem 2.11.1 of van der Vaart & Wellner (1996), which further implies the unconditional asymptotic equivalence. Therefore, the second and third terms in (A6) are asymptotically equivalent to n1/2(𝒫n𝒫)ξ(t; cυ, β*) and

n1/2𝒫[S(t;c^υ,β^,W)I(β^TZc^υ)S(t;cυ,β*,W)I(β*TZcυ)]+n1/2(𝒫n𝒫){ξ(t;c^υ,β^)ξ(t;cυ,β*)}, (A8)

respectively. By the arguments of Appendix 2, (A8) is asymptotically equivalent to

υ×{G(t;c,β*)c|c=cυn1/2(c^υcυ)+G(t;cυ,β)β|β=β*n1/2(β^β*)}.

As in the case of n1/2𝒫 [{Ŝ(t; ĉυ, β̂, W) − S(t; ĉυ, β̂, W)}I (β̂TZĉυ)], we can show that n1/2𝒫n[{Ŝ(t; ĉυ, β̂, W) − S(t; ĉυ, β̂, W)}I (β̂TZĉυ)] is asymptotically equivalent to n1/2(𝒫n𝒫) (t; ĉυ, β̂), which implies that the fourth term in (A6) converges uniformly to zero in probability. The fifth term in (A6) converges uniformly to zero in probability by Lemma 19.24 of van der Vaart (1998).

Combining the above results, we obtain the weak convergence of n1/2{ (t, υ) − GZ,β* (t, υ)}.

Appendix 4. Validity of the bootstrap method

In this section, we prove the validity of the bootstrap method for Ĝ(t, υ) for survival data. The proofs for other cases are similar and thus omitted. Define

G^(b)(t;c,β)=0<ut{1i=1nMniI(βTZic)dNi(u)i=1nMniI(βTZic)Yi(u)},

where Mni is the number of times that {Ni (t), Yi (t), Xi : t ∈ [0, τ]} is redrawn from the original sample. Let c^υ(b) and β̂(b) be the bootstrap counterparts of ĉυ and β̂. Our goal is to show that the distribution of n1/2{Ĝ(t, υ) − GZ,β* (t, υ)} can be approximated by the conditional distribution of n1/2{G^(b)(t;c^υ(b),β^(b))G^(t,υ)} given the data.

By the arguments of Appendix 2, n1/2{G^(b)(t;c^υ(b),β^(b))G^(t,υ)} is equal to

n1/2{G^(b)(t;c^υ,β^)G^(t;c^υ,β^)}+n1/2{G^(b)(t;c^υ(b),β^(b)G^(b)(t;c^υ,β^)}. (A9)

Lemma S1 of the Supplementary Material implies that, given the data, the first term in (A9) has the same asymptotic distribution as n1/2(𝒫n𝒫)ξ(t; cυ, β*), where ξ(t; c, β) is defined in Appendix 2. It can also be shown that, given the data, n1/2(β̂(b)β̂) = n1/2(𝒫̂n𝒫n)ψ + op(1) under mild regularity conditions, where 𝒫̂n is the bootstrap empirical distribution (van der Vaart & Wellner, 1996, p. 345). Combining the arguments of Appendix 2 with those of Lemma S1, we can show that the limiting distribution of n1/2{G^(b)(t;c^υ(b),β^(b))G^(t,υ)} given the data is the same as the limiting distribution of n1/2(𝒫n𝒫) {ξ1(t, υ) + ξ2(t, υ)}, where ξ1(t, υ) and ξ2(t, υ) are defined in Appendix 2.

Supplementary material

Supplementary material available at Biometrika online includes additional simulation results and Lemma S1 of Appendix 4.

References

  1. Dabrowska DM, Doksum KA. Partial likelihood in transformation models with censored data. Scand J Statist. 1988;15:1–23. [Google Scholar]
  2. Fan C, Oh DS, Wessels L, Weigelt B, Nuyten DSA, Nobel AB, van’t Veer LJ, Perou CM. Concordance among gene-expression-based predictors for breast cancer. New Engl J Med. 2006;355:560–9. doi: 10.1056/NEJMoa052933. [DOI] [PubMed] [Google Scholar]
  3. Graf E, Schmoor C, Sauerbrei W, Schumacher M. Assessment and comparison of prognostic classification schemes for survival data. Statist Med. 1999;18:2529–45. doi: 10.1002/(sici)1097-0258(19990915/30)18:17/18<2529::aid-sim274>3.0.co;2-5. [DOI] [PubMed] [Google Scholar]
  4. Heagerty PJ, Zheng Y. Survival model predictive accuracy and ROC curves. Biometrics. 2005;61:92–105. doi: 10.1111/j.0006-341X.2005.030814.x. [DOI] [PubMed] [Google Scholar]
  5. Henderson R, Diggle P, Dobson A. Identification and efficacy of longitudinal markers for survival. Biostatistics. 2002;3:33–50. doi: 10.1093/biostatistics/3.1.33. [DOI] [PubMed] [Google Scholar]
  6. Korn EL, Simon R. Measures of explained variation for survival data. Statist Med. 1990;9:487–503. doi: 10.1002/sim.4780090503. [DOI] [PubMed] [Google Scholar]
  7. Kosorok MR. Introduction to Empirical Processes and Semiparametric Inference. New York: Springer; 2008. [Google Scholar]
  8. Lin DY, Wei LJ, Ying Z. Semiparametric transformation models for point processes. J Am Statist Assoc. 2001;96:620–8. [Google Scholar]
  9. Moskowitz CS, Pepe MS. Quantifying and comparing the predictive accuracy of continuous prognostic factors for binary outcomes. Biostatistics. 2004;5:113–27. doi: 10.1093/biostatistics/5.1.113. [DOI] [PubMed] [Google Scholar]
  10. Schemper M. The explained variation in proportional hazards regression. Biometrika. 1990;77:216–8. (Correction: (1994). Biometrika81, 631.) [Google Scholar]
  11. Schemper M, Henderson R. Predictive accuracy and explained variation in Cox regression. Biometrics. 2000;56:249–55. doi: 10.1111/j.0006-341x.2000.00249.x. [DOI] [PubMed] [Google Scholar]
  12. Schoop R, Graf E, Schumacher M. Quantifying the predictive performance of prognostic models for censored survival data with time-dependent covariates. Biometrics. 2008;64:603–10. doi: 10.1111/j.1541-0420.2007.00889.x. [DOI] [PubMed] [Google Scholar]
  13. Therneau TM, Hamilton SA. rhDNase as an example of recurrent event analysis. Statist Med. 1997;16:2029–47. doi: 10.1002/(sici)1097-0258(19970930)16:18<2029::aid-sim637>3.0.co;2-h. [DOI] [PubMed] [Google Scholar]
  14. Uno H, Cai T, Tian L, Wei LJ. Evaluating prediction rules for t-year survivors with censored regression models. J Am Statist Assoc. 2007;102:527–37. [Google Scholar]
  15. van der Vaart AW. Asymptotic Statistics. New York: Cambridge University Press; 1998. [Google Scholar]
  16. van der Vaart AW, Wellner JA. Weak Convergence and Empirical Processes. New York: Springer; 1996. [Google Scholar]
  17. van de Vijver MJ, He YD, van’t Veer LJ, Dai H, Hart AAM, Voskuil DW, Schreiber GJ, Peterse JL, Roberts C, Marton MJ, et al. A gene-expression signature as a predictor of survival in breast cancer. New Engl J Med. 2002;347:1999–2009. doi: 10.1056/NEJMoa021967. [DOI] [PubMed] [Google Scholar]
  18. Zeng D, Lin DY. Efficient estimation of semiparametric transformation models for counting processes. Biometrika. 2006;93:627–40. [Google Scholar]
  19. Zheng Y, Cai T, Pepe MS, Levy WC. Time-dependent predictive values of prognostic biomarkers with failure time outcome. J Am Statist Assoc. 2008;103:362–8. doi: 10.1198/016214507000001481. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Zheng Y, Heagerty PJ. Prospective accuracy for longitudinal markers. Biometrics. 2007;63:332–41. doi: 10.1111/j.1541-0420.2006.00726.x. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary material available at Biometrika online includes additional simulation results and Lemma S1 of Appendix 4.


Articles from Biometrika are provided here courtesy of Oxford University Press

RESOURCES