Abstract
High‐dimensional data are becoming increasingly common in the medical field as large volumes of patient information are collected and processed by high‐throughput screening, electronic health records, and comprehensive genomic testing. Statistical models that attempt to study the effects of many predictors on survival typically implement feature selection or penalized methods to mitigate the undesirable consequences of overfitting. In some cases survival data are also left‐truncated which can give rise to an immortal time bias, but penalized survival methods that adjust for left truncation are not commonly implemented. To address these challenges, we apply a penalized Cox proportional hazards model for left‐truncated and right‐censored survival data and assess implications of left truncation adjustment on bias and interpretation. We use simulation studies and a high‐dimensional, real‐world clinico‐genomic database to highlight the pitfalls of failing to account for left truncation in survival modeling.
Keywords: Cox model, high‐dimensional data, lasso, left truncation, penalized regression, survival analysis
1. INTRODUCTION
The recent development and accessibility of genomic (DNA sequencing, microarrays), proteomic (immuno‐histology, tissue arrays), clinical (electronic health records), digital (wearables), and imaging (PET, fMRI) technologies have led to an increasing complexity and dimensionality of patient health information. The amount of person‐level and disease‐specific detail captured by these technologies offers an exciting new opportunity to understand patient history and prognosis by combining these molecular data with macro‐scale clinical data on patient characteristics, treatment, and outcomes. Specifically, these data may offer insights into risk factors affecting patient‐related outcomes leading to interest in developing prognostic tools to improve clinical decision‐making and patient care. 1 , 2 , 3 , 4 Given the hundreds or even thousands of molecular, patho‐physiological, and clinical parameters that may be readily available for each patient, developing an appropriate statistical model that is robust to overfitting is critical to ensure good model performance and generalizability. Survival analysis is of particular interest given that overall survival is often a primary outcome of interest.
In a high‐dimensional setting, computationally efficient regularization and shrinkage techniques have been developed 5 , 6 , 7 for the Cox model. Similar to ordinary least squares regression, these penalized models estimate regression coefficients by minimizing the sum of squared errors, but place a constraint in the equation to minimize the sum of coefficients. This constraint (controlled by a parameter, ) penalizes a high number of variables in the model and shrinks the coefficients of less influential features towards zero—acting, in cases when some coefficients are shrunk fully to zero, as automatic feature selection. Common penalized approaches include lasso, 8 ridge, 9 , 10 and elastic net, 11 with extensions such as the adaptive lasso. 12
A common challenge in survival data is that patients are often included in the data only after having been at risk (eg, diseased) for some time so that patients with shorter survival times are unobserved. In other words, the data are left‐truncated because patients are only included if they survive beyond some initial milestone that marks entry into the observed data. An analysis of the observed patients is subject to an immortal time bias because observed patients cannot die prior to entering the study cohort. 13 , 14 An example in medicine is a dataset describing patients who receive a specific diagnostic or genetic test; patients who die before having the chance to be tested are excluded from the data, and consequently their survival times do not contribute to overall survival estimates. Traditional survival models are inappropriate in this case and predicted survival probabilities are overestimated because of the exclusion of short survival times, biasing the study sample and consequently the estimation of the baseline risk and regression coefficients. Such bias has been observed in large oncology cohorts marked by a milestone entry, including patients who receive genomic tests 15 and specialized treatments. 16 These databases are increasingly used to understand patient survival to support clinical decision‐making, motivating the need to identify and address this survivor selection bias in practice.
Left truncation is a well‐known feature of medical data and there are methods to adjust for it. 17 The Kaplan‐Meier estimator of the survival function requires almost no modifications as it is only necessary to change the risk set so that individuals are only at risk of an event after study entry. 18 Similarly, with left truncation, the likelihood function in a parametric model can be adjusted so that it is conditional on survival to the entry time.
Here, we apply a penalized Cox proportional hazards model for left‐truncated and right‐censored (LTRC) survival data and critically assess scenarios in which bias arises from and is exacerbated by left truncation. We apply the method to simulated data and a real‐world use case from a deidentified clinico‐genomic database (CGDB). Our results highlight the importance of adjusting for left truncation for both model interpretation and prediction. Furthermore, we show that use of inappropriate metrics for model evaluation can result in overly optimistic conclusions about the performance of predictive models. For implementation of the proposed method, we use the left truncation extension to survival modeling introduced in version 4.1 of the R package glmnet. 19
The remainder of this article is organized as follows: In Section 2, we introduce the penalized regression and left truncation framework and present the method that incorporates both. Section 3 presents the simulation results and Section 4 considers a real‐world dataset that combines electronic health records (EHRs) and genomic testing information. Finally, we discuss general takeaways and survival modeling recommendations in Section 5.
2. METHODS
2.1. Data and notation
Consider a typical survival analysis framework in the absence of truncated survival times. For each patient i, we have data of the form , , where is the observed survival time, defined as the time between baseline and event, is the event or failure of interest (eg, death), and is a vector of predictors of length p used in the regression model. The event or failure of interest is 1 if the event is observed or 0 if the event is right‐censored. Suppose there are m unique event times, and let be the ordered, unique event times. In this framework, all subjects are assumed to be at risk of the event starting at a commonly defined “time zero” (a milestone that serves as a reference point for survival such as date of diagnosis) and observed over the interval where T is the maximum follow‐up time. Let be the index of the observation with an event at time . A Cox proportional hazards model is commonly used to model the relationship between survival times and predictors assuming a semiparametric hazard of the form
(1) |
where is the hazard for patient i at time t, is the baseline hazard shared across subjects, and is a length p vector of predictor coefficients.
Cox 20 proposed a partial likelihood for without involving the baseline hazard
(2) |
where denotes the set of indices who are “at risk” for failure at time , called the risk set. Maximizing the partial likelihood solves for and allows for inferences on the relationship between survival time and the set of predictors.
2.2. Left truncation
In the absence of left truncation, the risk set assumes that every individual is at risk of an event at his or her respective time 0 and continues to be at risk until the event occurs or the individual is censored. In this sense, all individuals enter the risk set at the same time 0 and leave only at death or censoring. However, when survival times are left‐truncated, individuals are truly at risk prior to entering the study cohort (eg, prior to being observed for follow‐up). Take Figure 1 as an example.
Survival is defined as the time from diagnosis until death, but patients are not eligible to enter the cohort until a milestone is reached (black triangle). Patients 1 and 2 (in red) are excluded from the analysis because of death and right censoring, respectively, prior to study eligibility. A naive analysis on this data would therefore be biased because it only contains survival time for “super‐survivors” (those who have survived a long time, such as patients 3 and 7) and the recently diagnosed.
Fortunately, the partial likelihood can be easily modified to make the set of individuals who are at risk (have entered the study and are not censored or failed) at that time. Let individual i have survival data of the form , where individual i enters the study at time . Time is the left truncation time, also referenced in this text as entry time. The left‐truncated risk set . In other words, subjects are only considered at risk at a given event time after they have been observed to be at risk (ie, exceeded their left truncation time/passed milestone entry time, shown with a solid black line in Figure 1 ). For example, at the time of death of Patient 3 in Figure 1, only Patient 4 has reached the milestone for study entry and is at risk. Patients 5 to 7 reach their milestones at later time points.
The partial likelihood in Equation (2) can then be evaluated by substituting to make inferences on in the case of left truncation.
2.3. Penalized regression with LTRC data
Penalized regression methods such as (lasso) 8 and (ridge) 21 regression, elastic net regression, 11 and smoothly clipped absolute deviation 22 have been developed to overcome the limitation of fitting a model to a large number of predictors. In a high‐dimensional feature space, where the number of predictors p is large relative to the sample size n, a model may exhibit low bias but suffer from poor prediction accuracy and limited generalizability, a consequence of overfitting the data. Penalized approaches, which add a regularization penalty to constrain the size of the , reduce the complexity of the model and mitigate the negative effects of overfitting. Note that in high‐dimensional settings where the number of predictors is larger than the sample size (ie, ), a standard regression cannot be fit at all.
The extensions of penalized regression methods to survival data have been extensively described and applied, for example in microarray gene expression data, 1 transcriptomes, 23 and injury scale codes. 24 Like linear regression, the performance and fit of a Cox proportional hazards model is known to deteriorate with large p, providing an advantage to penalized approaches in high‐dimensional, time‐to‐event feature sets.
Simon et al 7 introduced a fast, pathwise algorithm to fit a regularized Cox proportional hazards model solving for
(3) |
where is the vector of regression coefficients and is maximized over Equation (3) subject to a constraint defined by , and is the (lasso), (ridge), or elastic net penalty on the sum of coefficients described elsewhere. 7 , 8 , 11 , 21 The algorithm by Simon et al fits over a path of values via cyclical coordinate descent. Instead of solving a least squares problem at each iteration, the algorithm solves a penalized reweighted least squares problem, in which observation weights w are defined in the coordinate descent least squares minimization step. The parameter can be identified through cross‐validation. Tied survival times are handled via the Breslow approximation 25 and details provided in Reference 7. Note that the Efron approximation will be added to the next version of glmnet.
Applying the standard penalized Cox implementation in the software glmnet to left‐truncated survival data will result in biased estimates. However, an extension to LTRC data is straightforward. We solve for Equation (3) with a modification to the risk set :
(4) |
where the left‐truncated risk set ; that is, subjects are only considered at risk after their left truncation times. This implementation is available in the release of glmnet version 4.1. 19
2.4. Real‐world use case
We consider the CGDB offered jointly by Flatiron Health and Foundation Medicine. 26 The CGDB is a US‐based deidentified oncology database that combines real‐world, patient‐level clinical data and outcomes with patient‐level genomic data to provide a comprehensive patient and tumor profile. The deidentified data originated from approximately 280 US cancer clinics (roughly 800 sites of care). Patients in the CGDB have received at least one Foundation Medicine comprehensive genomic profiling test as well as have had their EHRs captured by Flatiron Health. The retrospective, longitudinal EHR‐derived database comprises patient‐level structured and unstructured data, curated via technology‐enabled abstraction, and were linked to Foundation Medicine genomic data by deidentified, deterministic matching. Genomic alterations were identified via comprehensive genomic profiling of 300 cancer‐related genes on FMI's next‐generation sequencing based FoundationOne panel. 27 , 28 To date, over 400 000 samples have been sequenced from patients across the US. The data are deidentified and subject to obligations to prevent reidentification and protect patient confidentiality. Altogether, the CGDB represents thousands of potential predictors per patient, ranging from demographic, laboratory, and treatment information to the specific mutations detected on each biomarker included in a Foundation Medicine baitset. Institutional Review Board approval of the study protocol was obtained prior to study conduct, and included a waiver of informed consent.
We take a subset of 4429 patients with stage IV, nonsmall cell lung cancer (NSCLC) in the CGDB as a use case for this study. The outcome of interest is survival time predicted from the date of stage IV diagnosis. Survival times are left‐truncated because patients who die before receiving a Foundation Medicine test are logically not included in the cohort of Foundation Medicine test recipients.
As shown in Figure 2, the distribution of left truncation times (time from diagnosis to genomic testing) is right‐tailed with approximately 21.5% of patients diagnosed more than 1 year prior to receiving the Foundation Medicine genomic test. Moreover, the correlation between diagnosis year and left truncation time is strongly negative (), suggesting the presence of left truncation survivor bias as patients with records in the CGDB diagnosed longer ago needed to survive more years before the availability of genomic testing.
Next, we use these data as the basis for a simulation study and a real‐world application.
3. SIMULATION STUDY
Simulations were conducted to assess the performance of the method with LTRC data. To set notation, let , , and be latent survival, right censoring, and study entry times for patient i, respectively. The simulation proceeds by simulating , , and from suitable survival distributions and assumes that , , and are uncorrelated. The observed survival time is given by and a patient is right‐censored if . Patients are left‐truncated (ie, excluded from the sample) if .
3.1. Data generating process
We generated n patients and p predictors to construct a matrix, X. The first 11 predictors were from the CGDB and the remaining variables were simulated binary variables to represent a comprehensive set of CGDB binary alterations. The CGDB variables were the same 11 variables used in the small p model in Section 4 and generated by randomly sampling rows of the original data with replacement. The matrix of binary variables, was simulated using a (latent) multivariate probit approach
(5) |
with the kth binary variable if and 0 otherwise. Each diagonal element of equals 1 and nondiagonal elements determine the correlation between variables. The probability, that variable k was equal to 1 was randomly sampled from a uniform distribution ranging from 0.2 to 0.8; that is, we sampled and set where is the cumulative density function of the normal distribution. was generated by first constructing a positive semidefinite matrix by filling the matrix, , with random variables. S was then scaled into a correlation matrix by setting where D is a vector containing the diagonal elements of S.
Latent survival time was simulated from a Weibull distribution based on evidence from the CGDB as detailed in Section 3.2. We used a proportional hazards parameterization of the Weibull distribution with shape parameter a, scale parameter m, and for , probability density function given by,
(6) |
The predictors were used to model latent survival time through their effect on the scale parameter,
(7) |
(8) |
where is an intercept and is a vector of coefficients. and values of for the CGDB variables were estimated from the CGDB data as described in Section 3.2. The probability that the coefficient of a binary variables was nonzero was 0.5 and drawn from a Bernoulli distribution. Nonzero coefficients were drawn from a uniform distribution with lower bound of and upper bound of 0.25 so that the minimum and maximum hazard ratios were 0.78 and 1.28, respectively.
Time to right censoring was also sampled from a Weibull distribution but did not depend on predictors,
(9) |
Finally, a two‐part model was used to simulate time to study entry so that a proportion of patients could enter the study at while a proportion would enter at . The probability of an entry time greater than 0 was drawn from a Bernoulli distribution and positive entry dates were modeled using a lognormal distribution.
(10) |
(11) |
(12) |
3.2. Calibration
The simulation was calibrated to be consistent with the CGDB NSCLC data. The Nelson‐Aalen estimator of the cumulative hazard of overall survival was generally monotonically decreasing, which suggested that a Weibull distribution could adequately capture the baseline hazard. This was further supported by the similarity in fit between a spline‐based survival model with one internal knot at the median of uncensored survival time and the Weibull model (Figure S1). As such, an intercept only Weibull model was used to estimate a and in the simulation. Similarly, a Weibull distribution was used to model time to right censoring by reversing the censoring indicator in the model for overall survival. Both the model for overall survival and right censoring adjusted for left truncation.
The CGDB variables were standardized to have mean 0 and SD 1 and coefficients were estimated using a Cox proportional hazards model. The full coefficient vector combined the coefficients from the CGDB variables with the coefficients from the simulated binary variables. The natural logarithm of the scale parameter was set to so that predicted survival averaged across patients was equal to predicted survival from the intercept only Weibull model.
Latent time to study entry is inherently unobserved, but we believe a two‐part model can characterize the large fraction of tests in the observed data that occur near the diagnosis date (Figure 2). The probability of a positive entry time was set equal to 0.2 and the lognormal distribution for positive entry times was derived using a right skewed distribution with a median value of 1 and a mean value of 1.6, both measured in years. We found that this parameterization generated median survival times in the simulation that were similar to median survival in Kaplan‐Meier estimates that did not adjust for left truncation in the CGDB (Figure S2).
3.3. Implementation
Datasets were simulated where each row contained information on a patient's observed survival time (, a censoring indicator (), the time of study entry (). 75% of patients were assigned to a training set and the remaining 25% were assigned to a test set.
Within the training set, Cox models with lasso penalties were fit among the “observed” (ie, nontruncated) patients. Cross‐validation with 10 folds was used to tune the model and the value of that minimized the partial likelihood deviance was selected. Models were fit with and without adjusting for left truncation. All models adjusted for right censoring.
Predictions and model evaluations were performed on the test set. The models were evaluated using the “complete” sample consisting of both truncated and nontruncated patients. The use of the complete sample for model evaluation is a key advantage of a simulation study since it is unobservable in real‐world applications.
This process was repeated 200 times. For each iteration, models were evaluated based on a dataset consisting of patients. Simulations were performed in both small and large p scenarios. In the small p scenario, 10 binary variables were simulated in addition to the 11 CGDB variables so that ; in the large p scenario, 1000 binary variables were simulated so that .
3.4. Results
We used calibration curves 29 to compare models with and without a left truncation adjustment. Each point in the plot consists of simulated patients in a given decile of predicted survival at a given time point. The x‐axis is the averaged predicted survival probability and the y‐axis is the pseudo observed survival probability computed using the Kaplan‐Meier estimator. Perfectly calibrated predictions are those that lie along the 45 degree line so that predicted and observed survival probabilities are equal. The plots consist of patients in the complete sample and are therefore a good test of the effectiveness of the left truncation adjustment.
Calibration curves are displayed in Figure 3 for both the small and large p scenarios. Predicted survival probabilities from the model that does not adjust for left truncation are too high because the model is fit on the observed sample consisting only of patients that survived long enough to enter the sample. In other words, failing to adjust for left truncation results in an overestimation of survival probabilities. By contrast, the left‐truncated adjusted model corrects for this bias. The small p model is almost perfectly calibrated with points along the 45 degree line.
The results of the large p scenario are similar, although calibration is more difficult than in the lower dimensional setting with points lying farther from the 45 degree line for patients with higher survival probabilities. Predictions tend to be more accurate for patients with predicted survival close to and below 0.5, but slightly underestimated for patients with higher survival probabilities. Probability rescaling has been effective in such cases, for example with the use of Platt scaling 30 or isotonic regression, 31 though implementation is less straightforward in a survival context. 32
A common metric used to evaluate survival models is the C‐index, 33 which provides a measure of how well a model will discriminate prognosis between patients. We computed the C‐index using predictions from the Cox model with lasso penalty on high‐dimensional data. The results are presented in Table 1 and suggest that care should be taken when using the C‐index to evaluate survival models in the presence of left‐truncated data. When using the observed sample to evaluate the model, the C‐index is higher in a model that does not adjust for left truncation, even though it is poorly calibrated. Furthermore, even when using the complete sample, the C‐indices are very similar.
TABLE 1.
Left truncation adjustment | Test sample | C‐index |
---|---|---|
No | Observed | 0.72 |
Yes | Observed | 0.67 |
No | Complete | 0.69 |
Yes | Complete | 0.69 |
Note: Cox models with lasso penalty were fit using the high‐dimensional simulated data.
These results are perhaps not surprising. For example, when computing the C‐index in the observed sample, the model is evaluated in a test set that suffers from the same bias as the training set. Moreover, the C‐indices may be nearly identical in the complete sample because we did not simulate a dependence between the predictors and entry time. In other words, the bias in the coefficients is based only on the indirect correlation between the predictors and left truncation induced by (i) the correlation between the predictors and survival and (ii) the correlation between survival and the probability a patient is left‐truncated. As we will see in Section 4, the impact of left truncation adjustment on the estimated values of the coefficients (and by extension the C‐index) can be considerably larger if there is a more direct correlation.
4. REAL‐WORLD DATA APPLICATION: PATIENTS WITH NSCLC IN THE CGDB
We fit models using 4429 patients with stage IV, NSCLC. 75% of patients were assigned to a training set and the remaining 25% were assigned to a test set. The training set was then used to train and tune both small p and large p models with 11 and 631 predictor variables, respectively. Both unpenalized and lasso penalized Cox models were fit, with and without a left truncation adjustment. In the penalized models, the optimal value of the shrinkage parameter, , was selected using the value that minimized the partial likelihood deviance in 10‐fold cross‐validation. The models were evaluated by assessing predictions on the test set.
Both the small and large p models were designed to highlight left truncation in a real‐world dataset and its impact on models, as opposed to optimizing predictive performance and the prognostic value of predictors. For this reason, model specifications were relatively simple, for example, modeling linear and additive effects, applying no transformations to predictors (for instance, modeling mutations as binary), and including a variable, year of diagnosis, that is highly correlated with left truncation time but not necessarily with survival (Figure 2). Survival was predicted from diagnosis. For an evaluation of meaningful prognostic clinical and genomic factors in NSCLC, see, for example, Lai et al. 34
The smaller model consisted of the following variables: year of diagnosis (ie, index year), age at diagnosis, race, practice type (academic or community center), and the mutational status (mutated/not mutated) of three known prognostic biomarkers for NSCLC (TP53, KRAS, and EGFR). Short variants (SV) were considered for each biomarker while copy number alterations and rearrangements were also considered for TP53. 3002 patients remained in the training set after dropping patients with missing values on at least one covariate.
The larger model included all nonbiomarker variables from the smaller model and added nearly all mutations tested by Foundation Medicine. As the data come from both historic and current clinico‐genomic testing products, the set of tested genes has changed over time. To resolve the resultant missing data problem, we constrained the mutation data to biomarkers with less than 30% missingness and imputed the remaining mutation data using k nearest neighbors, 35 with k = 20. Biomarkers with zero variance or that were present in less than 0.1% of patients were also excluded.
Figure 4 displays hazard ratios for each model. The plot for the large p model only contains coefficients ranked in the top 10 by absolute value of the hazard ratio in either the left truncation adjusted or nonadjusted model. The arrows denote the impact of adjusting for left truncation on the hazard ratio. In some cases the adjustment can have a considerable impact on the hazard ratios, often moving a large hazard ratio all the way to 1 (ie, to a coefficient of 0). This is most notable with diagnosis year, which has a hazard ratio of approximately 1.5 in both the small and large p models but is reduced to less than 1 after the left truncation adjustment. This effect is likely due to the strong negative correlation between calendar time and the probability that a patient is left‐truncated.
Table 2 and Figure 5 evaluate the models based on predictions on the test set. Table 2 compares models using the C‐index. The C‐index is relatively low in all of these illustrative models, though is artificially higher in the unadjusted models that do not account for left truncation—sometimes considerably so—and decreases after adjustment as there are fewer differences in predicted relative risks across patients and the discriminatory ability of the model is low. As expected, the lasso model has a higher C‐index in the larger p setting as the Cox model is considerably overfitted.
TABLE 2.
C‐index | |||
---|---|---|---|
Model | Covariate size | No adjustment | Adjustment |
Cox | Small | 0.649 | 0.580 |
Cox (lasso) | Small | 0.648 | 0.580 |
Cox | Large | 0.626 | 0.556 |
Cox (lasso) | Large | 0.663 | 0.600 |
Note: “Adjustment” indicates models with a left truncation adjustment.
Figure 5 displays the same survival calibration curves presented in the simulation section. As in the simulation study, the unadjusted models overpredict survival probabilities; however, in contrast to the simulations, the adjusted models are poorly calibrated.
The adjusted models tend to be better calibrated for patients with the lowest survival probabilities but poorly calibrated for patients with higher survival probabilities. One possible reason for this result is that the adjusted models tended to push hazard ratios—especially for the influential “year of diagnosis” predictor—toward 1, resulting in less separation of low and high risk patients. Probability rescaling such as Platt scaling or isotonic regression may be effective on the adjusted probabilities here, which resemble the characteristic sigmoid‐shaped calibration curves produced by different learning algorithms, including the lasso; however the implementation is less straightforward in a survival context. 32
5. DISCUSSION
In this article, we applied an approach for estimating penalized Cox proportional hazards model with LTRC survival data and compared it to penalized models designed for survival data that is right‐censored but not left‐truncated. Using simulation studies and examples from real‐world EHR and genomics data, we showed that there is a need for approaches that can both adjust for left truncation and model high‐dimensional data. In particular, our simulation shows that predictions from models that fail to adjust for left truncation will overestimate true survival probabilities whereas models that properly adjust can yield well calibrated survival predictions, even with high‐dimensional data. Moreover, in our real‐world use case, there were significant differences in coefficient estimates between models that did and did not adjust for left truncation.
Our results have important implications for specification, interpretation, and evaluation of prognostic survival models. Much of the difficulty arises because training and test sets are both subject to the same selection bias. In other words, a careful understanding of the data generating process is crucial and naive approaches can lead analysts astray. For example, our simulations suggest that the commonly used C‐index cannot differentiate between models that mimic the true data generating process and those that do not, and may therefore lead to misleading conclusions in the presence of LTRC data. On the other hand, metrics that focus on predicted survival probabilities such as calibration curves are better able to determine whether a model can accurately predict the risk of an event. In our real‐world use case, an analyst could have mistakenly concluded that the C‐index in the nonadjusted model was indicative of good model performance when it was in fact driven in part by a high correlation between predictors and left truncation. For instance, in our large p model the hazard ratio for the predictor “year of diagnosis” moved from being the largest in absolute value to less than 1 after adjusting for left truncation. The strong positive relationship between “year of diagnosis” on the hazard in the nonadjusted model was artificial because patients diagnosed earlier appeared to survive longer due to immortal time bias. This relationship reversed completely after adjusting for left truncation, and illustrates that biases to coefficient estimates are driven by the association between predictors and left truncation time. In the case of “year of diagnosis,” this association was negative and biased the hazard ratio upwards. While some of the other hazard ratios in the large p model were pushed towards 1 after left truncation adjustment, the direction of this bias depends on the direction of the association with left truncation time and varies by predictor. In cases where predictors and left truncation time are positively associated, hazard ratios are biased downwards because the predictors are associated with more immortal time. Because the high C‐index misleadingly suggested good model performance, we recommend that models should be evaluated using both discrimination and calibration based metrics to facilitate interpretation.
Although not the primary focus of this article, our approach generated findings that were both consistent with prior literature in genomics and clinically plausible. One of the mutation features that had the highest importance was the presence of SV in the gene KEAP1. This finding is supported by similar reports of risk‐conferring status in an independent lung cancer cohort. 36 Conversely, variants in EGFR and ALK were found to be protective in the CGDB data, putatively because mutations in these genes qualify patients to receive targeted therapies and therefore benefit from improved response and longer survival.
When making predictions with survival data is of interest, it is critical to define the risk set appropriately, as our work demonstrates. To do this requires recognizing the underlying process by which individuals are captured in the data in the first place; that is, when and how individuals came to be included in the dataset of interest. One of the motivations for the current work derives from the advent of biobank and genomic test databases that store information on patient illness—presumably after the patient has been diagnosed and “at risk” for some time. In these data, overlooking the “delayed entry” of at‐risk patients could result in misleading conclusions, as we demonstrate in this work. In addition, patient cohorts marked by a milestone entry, such as genomic testing, may not only be subject to immortal time bias but also to a temporal selection bias. The latter may occur if left truncation time is correlated with survival time, for example, if genomic testing is ordered selectively for patients with worse prognosis. This was uncovered in a cohort of multistage cancer patients who received genomic testing 15 and can affect inference even in the presence of a left truncation adjustment. While the cohort of lung cancer patients in our real‐world use case was restricted to stage IV diagnoses, mitigating this potential selection bias, 15 it can be adjusted for by modeling the association between left truncation time and survival. 37 , 38
COMPETING INTERESTS
R. Tibshirani and B. Narasimhan are paid consultants for Roche.
Supporting information
ACKNOWLEDGEMENTS
We thank colleagues at Flatiron Health, Inc. and Foundation Medicine, Inc. and the anonymous reviewers for their thoughtful reviews and constructive feedback to improve the quality of this work. We thank T. Hastie and K. Tay for their work in implementing left truncation‐adjusted Cox models in glmnet.
McGough SF, Incerti D, Lyalina S, Copping R, Narasimhan B, Tibshirani R. Penalized regression for left‐truncated and right‐censored survival data. Statistics in Medicine. 2021;40(25):5487–5500. 10.1002/sim.9136
DATA AVAILABILITY STATEMENT
The data that support the findings of this study have been originated by Flatiron Health, Inc. and Foundation Medicine, Inc. These deidentified data may be made available upon request, and are subject to a license agreement with Flatiron Health and Foundation Medicine; interested researchers should contact cgdb‐fmi@flatiron.com to determine licensing terms. Code used to produce the analysis is publicly available at https://github.com/phcanalytics/coxnet‐ltrc.
REFERENCES
- 1. Gui J, Li H. Penalized Cox regression analysis in the high‐dimensional and low‐sample size settings, with applications to microarray gene expression data. Bioinformatics. 2005;21(13):3001‐3008. [DOI] [PubMed] [Google Scholar]
- 2. Wishart GC, Azzato EM, Greenberg DC, et al. PREDICT: a new UK prognostic model that predicts survival following surgery for invasive breast cancer. Breast Cancer Res. 2010;12(1):R1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Ow GS, Kuznetsov VA. Big genomics and clinical data analytics strategies for precision cancer prognosis. Sci Rep. 2016;6:36493. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Yousefi S, Amrollahi F, Amgad M, et al. Predicting clinical outcomes from large scale cancer genomic profiles with deep survival models. Sci Rep. 2017;7(1):1‐11. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. Tibshirani R. The lasso method for variable selection in the Cox model. Stat Med. 1997;16(4):385‐395. [DOI] [PubMed] [Google Scholar]
- 6. Friedman J, Hastie T, Tibshirani R. Regularization paths for generalized linear models via coordinate descent. J Stat Softw. 2010;33(1):1. [PMC free article] [PubMed] [Google Scholar]
- 7. Simon N, Friedman J, Hastie T, Tibshirani R. Regularization paths for Cox's proportional hazards model via coordinate descent. J Stat Softw. 2011;39(5):1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Tibshirani R. Regression shrinkage and selection via the lasso. J Royal Stat Soc Ser B (Methodol). 1996;58(1):267‐288. [Google Scholar]
- 9. Hoerl AE, Kennard RW. Ridge regression: applications to nonorthogonal problems. Technometrics. 1970;12(1):69‐82. [Google Scholar]
- 10. Hoerl AE, Kennard RW. Ridge regression: biased estimation for nonorthogonal problems. Technometrics. 1970;12(1):55‐67. [Google Scholar]
- 11. Zou H, Hastie T. Regularization and variable selection via the elastic net. J Royal Stat Soc Ser B (Stat Methodol). 2005;67(2):301‐320. [Google Scholar]
- 12. Zou H. The adaptive lasso and its oracle properties. J Am Stat Assoc. 2006;101(476):1418‐1429. [Google Scholar]
- 13. Lévesque LE, Hanley JA, Kezouh A, Suissa S. Problem of immortal time bias in cohort studies: example using statins for preventing progression of diabetes. BMJ. 2010;340:b5087. [DOI] [PubMed] [Google Scholar]
- 14. Giobbie‐Hurder A, Gelber RD, Regan MM. Challenges of guarantee‐time bias. J Clin Oncol. 2013;31(23):2963. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Kehl KL, Schrag D, Hassett MJ, Uno H. Assessment of temporal selection bias in genomic testing in a cohort of patients with cancer. JAMA Netw Open. 2020;3(6):e206976. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Newman NB, Brett CL, Kluwe CA, et al. Immortal time bias in national cancer database studies. Int J Radiat Oncol Biol Phys. 2020;106(1):5‐12. [DOI] [PubMed] [Google Scholar]
- 17. Kalbfleisch JD, Prentice RL. The Statistical Analysis of Failure Time Data. Vol 360. Hoboken, NJ: John Wiley & Sons; 2011. [Google Scholar]
- 18. Tsai WY, Jewell NP, Wang MC. A note on the product‐limit estimator under right censoring and left truncation. Biometrika. 1987;74(4):883‐886. [Google Scholar]
- 19. Friedman J, Hastie T, Tibshirani R, et al. glmnet: lasso and elastic‐net regularized generalized linear models. R package version 4.1; 2021.
- 20. Cox DR. Regression models and life‐tables. J Royal Stat Soc Ser B (Methodol). 1972;34(2):187‐202. [Google Scholar]
- 21. Tikhonov AN. Solution of incorrectly formulated problems and the regularization method. Soviet Math. 1963;4:1035‐1038. [Google Scholar]
- 22. Xie H, Huang J. SCAD‐penalized regression in high‐dimensional partially linear models. Ann Stat. 2009;37(2):673‐696. [Google Scholar]
- 23. Wu TT, Gong H, Clarke EM. A transcriptome analysis by lasso penalized Cox regression for pancreatic cancer survival. J Bioinform Comput Biol. 2011;9(1):63‐73. [DOI] [PubMed] [Google Scholar]
- 24. Mittal S, Madigan D, Cheng J, Burdc R. Large‐scale parametric survival analysis. Stat Med. 2013;32(23):3955‐3971. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Breslow NE. Contribution to the discussion of the paper by D.R. Cox. J Royal Stat Soc Ser B (Methodol). 1972;34(2):216‐217. [Google Scholar]
- 26. Singal G, Miller PG, Agarwala V, et al. Development and validation of a real‐world clinicogenomic database. J Clin Oncol. 2017;35(15):2514. [Google Scholar]
- 27. Birnbaum B, Nussbaum N, Seidl‐Rathkopf K, et al. Model‐assisted cohort selection with bias analysis for generating large‐scale cohorts from the EHR for oncology research; 2020.
- 28. Ma X, Long L, Moon S, Adamson BJ, Baxi SS. Comparison of population characteristics in real‐world clinical oncology databases in the US: flatiron health, SEER, and NPCR; 2020. 10.1101/2020.03.16.20037143 [DOI]
- 29. Harrell FE Jr. Regression Modeling Strategies: With Applications to Linear Models, Logistic and Ordinal Regression, and Survival Analysis. New York, NY: Springer; 2015. [Google Scholar]
- 30. Platt J. Probabilistic outputs for support vector machines and comparison to regularized likelihood methods. Adv Large Margin Classif. 1999;10(3):61‐74. [Google Scholar]
- 31. Niculescu‐Mizil A, Caruana R. Predicting good probabilities with supervised learning. Paper presented at: Proceedings of the ICML' 05: Proceedings of the 22nd International Conference on Machine Learning; 2005.
- 32. Goldstein M, Han X, Puli A, Perotte AJ, Ranganath R. X‐CAL: Explicit Calibration for Survival Analysis. Adv Neural Inf Process Syst. 2020;33:18296‐18307. [PMC free article] [PubMed] [Google Scholar]
- 33. Harrell FE Jr, Lee KL, Mark DB. Multivariable prognostic models: issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors. Stat Med. 1996;15(4):361‐387. [DOI] [PubMed] [Google Scholar]
- 34. Yu‐Heng L, Chen WN, Hsu TC, Lin C, Tsao Y, Wu S. Overall survival prediction of non‐small cell lung cancer by integrating microarray and clinical data with deep learning. Sci Rep. 2020;10(1):4679. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Troyanskaya O, Cantor M, Sherlock G, et al. Missing value estimation methods for DNA microarrays. Bioinformatics (Oxford, England). 2001;17(6):520‐525. 10.1093/bioinformatics/17.6.520. [DOI] [PubMed] [Google Scholar]
- 36. Shen R, Martin A, Ni A, et al. Harnessing clinical sequencing data for survival stratification of patients with metastatic lung adenocarcinomas. JCO Precis Oncol. 2019;3:PO.18.00307. 10.1200/PO.18.00307. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Thiébaut AC, Bénichou J. Choice of time‐scale in Cox's model analysis of epidemiologic cohort data: a simulation study. Stat Med. 2004;23(24):3803‐3820. [DOI] [PubMed] [Google Scholar]
- 38. Chiou SH, Austin MD, Qian J, Betensky RA. Transformation model estimation of survival under dependent truncation and independent censoring. Stat Methods Med Res. 2019;28(12):3785‐3798. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The data that support the findings of this study have been originated by Flatiron Health, Inc. and Foundation Medicine, Inc. These deidentified data may be made available upon request, and are subject to a license agreement with Flatiron Health and Foundation Medicine; interested researchers should contact cgdb‐fmi@flatiron.com to determine licensing terms. Code used to produce the analysis is publicly available at https://github.com/phcanalytics/coxnet‐ltrc.