Abstract
Longitudinal and high‐dimensional measurements have become increasingly common in biomedical research. However, methods to predict survival outcomes using covariates that are both longitudinal and high‐dimensional are currently missing. In this article, we propose penalized regression calibration (PRC), a method that can be employed to predict survival in such situations. PRC comprises three modeling steps: First, the trajectories described by the longitudinal predictors are flexibly modeled through the specification of multivariate mixed effects models. Second, subject‐specific summaries of the longitudinal trajectories are derived from the fitted mixed models. Third, the time to event outcome is predicted using the subject‐specific summaries as covariates in a penalized Cox model. To ensure a proper internal validation of the fitted PRC models, we furthermore develop a cluster bootstrap optimism correction procedure that allows to correct for the optimistic bias of apparent measures of predictiveness. PRC and the CBOCP are implemented in the R package pencal, available from CRAN. After studying the behavior of PRC via simulations, we conclude by illustrating an application of PRC to data from an observational study that involved patients affected by Duchenne muscular dystrophy, where the goal is predict time to loss of ambulation using longitudinal blood biomarkers.
Keywords: Duchenne muscular dystrophy, high‐dimensionality, longitudinal data analysis, optimism correction, penalized regression calibration, risk prediction modeling, survival analysis
1. INTRODUCTION
High‐dimensionality and longitudinal measurements have become more and more common in biomedical datasets. This evolution has been fostered by several factors, including the increasing availability of clinical registry data and the diffusion of microarray and high‐throughput sequencing technologies. One of the consequences of these changes is that when developing a prediction model for a time to event outcome, biostatisticians are increasingly confronted with the availability of a larger number of possible covariates (often hundreds, or even thousands). More often than not, such covariates can vary over time, leading to scenarios where survival can be predicted using covariates that are both longitudinal and high‐dimensional. Additional features that might need to be dealt with in such scenarios are the presence of potentially strong correlations between the covariates, an unbalanced repeated measurements design, and missing data at certain measurement times for one or more covariates.
An example of such scenarios is represented by the MARK‐MD study, 1 , 2 an observational study on patients affected by Duchenne muscular dystrophy (DMD) that constitutes the motivating example of this article. DMD is a rare neuromuscular disorder whose consequences include progressive loss of muscular tissue and muscle mass, loss of ambulation (LoA) around the age of 12, and premature death. To date no cure for DMD has been found, and identification of non‐invasive biomarkers that could be employed to monitor disease progression and to predict disease milestones such as LoA is urgent. The MARK‐MD project aimed to study the dynamic evolution of 118 proteins in a longitudinal cohort of DMD patients, and it yielded an unbalanced dataset comprising between 1 and 5 repeated measurements per patient. Protein expression was measured using a suspension bead array platform, whereby the expression levels of a protein are not measured directly, but they are instead implicitly inferred from the relative abundance of multiple antibodies that bind to the protein of interest. In particular, between 1 and 5 antibodies per protein were measured, yielding a total of 240 antibodies matching 118 unique proteins; in most cases, high correlations were found between antibodies that targeted the same protein. 1 Statistical analysis of the longitudinal biomarker data 1 yielded evidence that several proteins are associated with disease progression and with the wheelchair dependency status, raising the question of whether blood biomarkers may be employed to predict the age at which a clinically meaningful milestone such as LoA will occur.
When developing a model to predict time to LoA for the MARK‐MD dataset, three main features should be taken into account:
High‐dimensionality: The number of possible predictors (240) largely outweighs the number of subjects (93);
The availability of repeated measurements (between 1 and 5) from the same subject;
The presence of several groups of strongly correlated predictors, each of which comprises multiple items (the antibodies) that are measured to reconstruct the same biological process of interest (the protein that the antibodies bind to).
While several statistical methods are available to tackle separately each of these three features, developing a model that can address them jointly is more challenging. To date, several approaches to predict survival using a high‐dimensional set of cross‐sectional covariates have been proposed, 3 however extensions to high‐dimensional settings with longitudinal covariates are currently lacking. On the other hand, study of the association between one or more longitudinal covariates and a time‐to‐event outcome is typically based on the specification of a joint model for longitudinal and survival data. 4 , 5 Estimation of joint models becomes computationally prohibitive when more than a handful of longitudinal covariates are included, restricting their application to low‐dimensional settings. 6 , 7
In practice, the lack of methods to predict survival using covariates that are both longitudinal and high‐dimensional often forces a choice between selecting only a small subset of the available predictors, or using only the baseline measurements of all variables. It is apparent, however, that both choices are suboptimal: if predictions are based on a limited number of pre‐specified covariates, important predictors may be missed; if the longitudinal measurements are discarded, important information about the way in which the predictors change over time will be missed. Therefore, a method that avoids these simplifications and does not sacrifice part of the available information would be preferable.
In this article, we propose penalized regression calibration (PRC), a method that makes it possible to predict survival times using a high‐dimensional set of predictors that are measured repeatedly over time, and which may additionally display strong correlations (as in the case of the MARK‐MD study). PRC comprises three modeling steps, which respectively consist of (1) the specification and estimation of a model for the longitudinal biomarkers, (2) the computation of subject‐specific summaries of the trajectories described by the biomarkers, and (3) the estimation of a model for the survival outcome.
More in detail, in step 1, we specify multivariate latent process mixed models (MLPMMs) 8 that allow us to model jointly all the items (ie, the antibodies) which are employed to measure a latent biological process of interest (the protein targeted by those antibodies). By doing so, we are able to model the trajectories described by highly correlated antibodies using a latent variable model that comprises two sets of random effects, one set for the observed items and another for the latent processes. Moreover, we can easily handle unbalanced designs or missing values thanks to the flexibility of mixed effects models. In step 2, we proceed to the computation of subject‐specific summaries of the longitudinal biomarkers, which in practice are the predicted random effects from the MLPMM. In step 3, we employ these summaries as predictors of the time to event outcome (time to LoA in the MARK‐MD study) by estimating a penalized Cox model with ridge or elasticnet penalty, which can then be used to compute predictions of survival probabilities.
In addition to this first model based on the MLPMM, we also consider a simpler model where in step 1 the biomarkers are modeled independently using linear mixed models (LMMs). 9 Such an approach, which performs univariate modeling of the biomarkers and is thus computationally less intensive, might be preferred in situations where no distinction between observed items and latent biological processes of interest exists. In terms of relationship with existing statistical methods, PRC might be regarded as a high‐dimensional extension of the ordinary regression calibration 10 (ORC) approach. However, ORC and PRC differ in their purposes: while the goal of ORC is to estimate the association between a single longitudinal biomarker and survival, PRC aims to predict survival based on a high‐dimensional set of longitudinal biomarkers.
A further contribution of our work is the implementation of a computational procedure that allows to perform the internal validation of PRC. The internal validation of prediction models based on penalized regression is often a neglected step, possibly due to the belief that the penalization of the model parameters will suffice to prevent overfitting, so that no formal assessment of the optimism associated to the reported performance is then needed. In this article, we do not take this belief for granted, and instead we explicitly explore the topic of the internal validation of PRC, which is a penalized prediction model. In particular, we implement a cluster bootstrap optimism correction procedure (CBOCP hereafter) that allows to estimate the optimism of naive predictiveness measures, thereby making it possible to properly assess whether overfitting is present. The CBOCP requires the re‐estimation of the PRC model on a sequence of clustered bootstrap samples, and the computation of estimates of the optimism associated to the apparent model performance, which are obtained by comparing the predictive performance of the newly fitted models on the bootstrapped dataset and on the original dataset. When averaged over all cluster bootstrap samples, these estimates yield a bootstrap optimism correction that is subtracted from the apparent measure of model predictiveness (ie, the measure of predictiveness computed on the original dataset where PRC was fitted). In Figure 1, we provide a schematic representation of the different steps involved in the estimation of PRC, and in the computation of the CBOCP.
Lastly, we provide an implementation of functions to compute PRC and its associated CBOCP using R, and make them available through the R package pencal, 11 which can be freely downloaded from CRAN.
The remainder of this article is organized as follows: in Section 2, we introduce the methodological framework for PRC, discussing in detail the three steps involved in its estimation and the CBOCP. In Section 3, we employ simulations to compare PRC to a simpler prediction approach that employs information on the baseline biomarker levels, but ignores the repeated measurements taken after baseline; additionally, we compare the predictive performance of PRC to that of a joint modeling approach in a low‐dimensional setting with three longitudinal covariates. In Section 4, we employ PRC to predict time to LoA for the patients involved in the MARK‐MD study. Lastly, in Section 5, we provide some conclusions and discuss possible extensions of PRC.
2. METHODS
We consider a situation in which a set of p latent biological processes (indexed by ) are considered as potential predictors of an event of interest, but they are not measured directly. Instead, items (indexed by ) that are related to the sth biological process are employed to reconstruct it, so that in total, items are employed as proxies for the p latent biological processes. In simpler situations where each biological process of interest can be measured directly, or where for all , the distinction between items and latent processes is unnecessary, and the subscript q can be dropped.
The distinction between a latent biological process of interest and the items used to measured it is motivated by the setup of proteomic studies, where typically the expression of a protein is not measured directly, but it is instead reconstructed by measuring the abundance of multiple antibodies that bind to the protein itself. In the case of the MARK‐MD study, the latent biological processes are proteins, and the items are the antibodies that were measured to reconstruct the level of the proteins. In particular, between 1 and 5 antibodies were measured for each protein, so that . This distinction is not unique to proteomic studies: it may as well arise in different settings, for example, when different psychometric items are employed to reconstruct a latent cognitive process of interest.
We assume that the items are measured longitudinally following either a balanced or an unbalanced design, and that for each individual , repeated measurements are collected before the event of interest occurs. We denote by the value of the qth item related to the sth latent biological process that is measured on individual i at their jth repeated measurement (). We summarize all measurements of the qth item referring to the sth biological process in a vector .
Moreover, we use to denote the age of subject i when the jth measurement is collected, and we let be the event time (measured starting from the baseline age ) for individual i, and be a censoring indicator such that if the event of interest is observed at , or in case of right‐censoring.
Let be the vector containing all repeated measurements of all items obtained from subject i. In Sections 2.1 to 2.3, we propose PRC, a method that allows to predict the time to event outcome using as predictor , which comprises a high‐dimensional set of longitudinally measured items. Specifically, in Section 2.1, we model the trajectories described by the longitudinal biomarkers using the MLPMM. 8 In Section 2.2, we proceed to the computation of subject‐specific summaries of the longitudinal trajectories that refer both to the measured items, and to the latent biological process. In Section 2.3, we specify a penalized Cox model that relates the time to event outcome to these summaries, and we employ it to compute predictions of survival probabilities.
In Section 2.4, we turn our attention to the problem of correctly assessing the predictive performance of PRC, discussing the CBOCP that we have implemented to obtained unbiased estimates of in‐sample predictiveness. Lastly, in Section 2.5, we provide details about the implementation of the proposed methodology in R, and we shortly introduce the R package pencal. 11
2.1. Modeling the longitudinal biomarkers
The first step of PRC involves the estimation and specification of a statistical model capable of describing the evolution over time of the longitudinal biomarkers. We propose to jointly model the dynamic evolution of all the items that refer to the sth latent biological process with the following MLPMM: 8
(1) |
where and are fixed effect parameters, is a vector of shared random effects that comprises a shared random intercept () and a shared random slope (), are item‐specific random intercepts, and is a measurement‐error term. As identifiability condition, we set . If there are some processes that are measured using only one item, that is, there exist some s for which , the MLPMM of Equation (1) can be simplified into , which is a LMM with correlated random intercept and random slope.
We employ maximum likelihood to estimate model (1), using the R package lcmm. 12 Note that for simplicity we have included only age as covariate and random slope in model (1), but depending on the features of the problem at hand one may choose to add to model (1) further covariates (eg, to account for relevant confounders), and to include further shared random effects.
The advantage of using the MLPMM specified in (1) is 2‐fold. First, it allows us to model jointly all the items that refer to the sth latent biological process, so that we can properly account for the possibility that the items may be strongly correlated with each other. Second, the MLPMM enables us to separate the within‐item variability from the between‐item variability, thanks to the inclusion of a first set of random effects that refer to the observed items, and of a second set of random effects that are linked to the unobserved biological process of interest. In practice, this means that although we cannot directly measure the relevant biological process, we can describe its underlying dynamic evolution using the shared random effects in (1).
This multivariate modeling approach is motivated by the setup of the MARK‐MD study, where proteins were not measured directly, but through multiple antibodies, and it can be applied to any other situation in which one measures multiple items that can be related to a common latent process. While our framework is general enough to account for such complex designs, it can be flexibly adapted to simpler situations where the biological processes of interest are directly measured. In such scenarios, one might prefer to model the longitudinal biomarkers independently from each other using separate LMMs. 9 Since in such cases for all , we can drop the subscript q from , and simplify the notation to . The general formulation of the LMM in matrix notation is
(2) |
where and are design matrices that respectively refer to the fixed effects coefficients and to the random effects , is a vector of normally distributed random effects, and is a Gaussian error term.
The LMM framework makes it possible to specify a wide range of random effect structures in (2); in particular, to keep the analogy with model (1) we hereafter consider the following LMM with correlated random intercepts and random slopes:
(3) |
where comprises a random intercept and a random slope , and . We employ maximum likelihood for the estimation of model (3), resorting in particular to the R package nlme. 13
2.2. Derivation of the predicted random effects
In the second step of PRC, we compute summaries of the longitudinal trajectories described by the biomarkers based on the mixed models estimated in the previous step. These summaries are the predicted random effects, 14 and their computation (and interpretation) differs depending on the type of model fitted in the first step.
Computation of the predicted random effects in R is typically straightforward when performed for the very same dataset on which the mixed effects model was fitted. However, it is more complicated when one needs to compute the predicted random effects for a new dataset, because this is not implemented neither in the lcmm package that estimates the MLPMM, nor in the popular R packages nlme and lme4 that are commonly used to estimate LMMs. Nevertheless, the ability to compute the predicted random effects on a dataset different from that on which a MLPMM or a LMM is fitted is fundamental for the implementation of the optimism correction that we will describe in Section 2.4, as well as for the computation of predictions for a new individual that was not included in the original dataset on which PRC was fitted. Therefore, hereafter we illustrate how to compute the predicted random effects for the MLPMM and for the LMM.
For the MLPMM, we can compute the predicted random effects and by adapting the formulas provided by Ebrahimpoor et al 15 to the MLPMM of Equation (1) as follows:
(4) |
where , is the equivalent of with as entries, is the random‐effects design matrix associated to in (1), , , , and where I denotes identity matrices and “all‐ones” matrices (ie, matrices whose entries are all equal to 1) of dimension .
For the LMM, instead, the predicted random effects are given by
(5) |
where is the marginal covariance matrix of subject i.
2.3. Prediction of the survival outcome
In the last step of the PRC approach, we employ the summary measures obtained in Section 2.2 to compute predictions of the individual survival probabilities. To achieve this goal, we specify a penalized Cox model where we include baseline age and the predicted random effects as covariates. When the longitudinal biomarkers are modeled using the MLPMM of Equation (1), we can consider two alternative prediction approaches.
Let and denote the vectors of shared and item‐specific predicted random effects for subject i computed using (4). The first approach that we consider is proposed having in mind a data generating mechanism where the event times are influenced only by the latent biological processes: with such an approach, the items are viewed just as an instrument to reconstruct the unobservable biological process, and they are not believed to have an additional predictive value of their own. Thus, in this first approach the hazard function depends only on the shared random effects:
(6) |
where denotes the hazard function, and is the baseline hazard. Hereafter, we refer to this modeling approach as PRC MLPMM(U).
On the contrary, the second approach that we consider allows for the possibility that besides being useful to reconstruct the latent biological processes, the items may carry additional information that is useful for the prediction of the time to event outcome. Thus, in this second approach the item‐specific random effects are included as predictors alongside with the shared :
(7) |
We refer to this modeling approach as PRC MLPMM(U+B).
Lastly, in the case in which the longitudinal markers are modeled with univariate LMMs we let
(8) |
where is the vector containing all the predicted random effects for subject i, as computed with (5). We refer to this modeling approach as PRC LMM.
In models (6) to (8), we have included baseline age as covariate, assuming that subjects entered the study at different ages. If this is not the case, can be dropped. Furthermore, in practice one may want or need to include in such models relevant predictors or confounders that are not measured longitudinally, such as, for example, gender, ethnicity, hospital, etc.; obviously, this can be easily done by adding such covariates to the linear predictors.
Models (8), (6), and (7) may (and typically will) comprise a large, high‐dimensional set of predictors. For example, in the MARK‐MD study the number of predictors is 237 for the PRC MLPMM(U) approach of Equation (6), 355 for the PRC MLPMM(U+B) in (7), and 481 for the PRC LMM approach in (8). In such high‐dimensional scenarios, maximum likelihood estimation is unfeasible, and some form of regularization is needed. In this article, we consider two penalized likelihood estimation methods, respectively based on the ridge 16 and on the elasticnet 17 penalties. The elasticnet penalty is a linear combination of an and an penalty that comprises two tuning parameters and . determines the overall level of regularization imposed on the regression coefficients by the penalty, whereas determines the relative contributions of the two penalties. When , the ridge penalty is obtained, whereas when , we obtain the lasso penalty. For model (7), the elasticnet penalty is given by
(9) |
where we do not penalize the regression coefficient associated to the baseline age. The ridge penalty can be obtained from (9) by fixing . For models (6) and (8), the penalties can be obtained from Equation (9) by dropping the terms containing .
We employ the R package glmnet 17 to estimate the penalized Cox models described above. We use the cross‐validation procedure implemented in glmnet to select the optimal value of tuning parameter when the ridge penalty is used. Moreover, we implement a nested cross‐validation procedure 18 to select the optimal values of the two tuning parameters when the elasticnet penalty is used.
Once the penalized Cox model of step 3 is estimated, we can proceed with the computation of the predicted survival probabilities. As usual with the Cox model, these can be obtained as follows:
where is a nonparametric estimate of the baseline hazard function, and denotes the linear predictor in models (6) to (8). Conditional survival probabilities can be obtained as , where .
2.4. Cluster bootstrap optimism correction procedure
We employ two measures of predictiveness to evaluate the predictive performance of the PRC‐LMM and PRC‐MLPMM methods: the time‐dependent area under the ROC curve (tdAUC hereafter), which we estimate using the nearest neighbor estimator proposed by Heagerty et al, 19 and the concordance (or C) index, which we estimate using the method proposed by Pencina and D'Agostino. 20
A common problem when developing prediction models is that of correctly evaluating the predictive performance of a method. Typically, the naive estimate of an index of predictiveness will be optimistically biased, due to the overfitting caused by the fact that the naive measure of performance is computed using the same data on which the model is estimated. For this reason, with low‐dimensional problems it is customary to implement an internal validation strategy that allows to correct for this bias. 21 However, internal validation is often a neglected step in the development of high‐dimensional risk prediction models, probably due to the belief that the penalization introduced in the estimation phase will by itself prevent any overfitting. In this article, we do not subscribe to this belief, and instead we develop a CBOCP that makes it possible to estimate the bias of the naive measures of performance, thus quantify the extent of the overfitting.
Our CBOCP adapts the bootstrap optimism correction method 22 to the case of longitudinal data. Hereafter we illustrate how the correction is computed in the case of the C index, but the computation proceeds in the same manner for the tdAUC. Let be the set of subject indicators contained in the dataset at hand, which we denote by . Note that for each subject i, contains repeated measurements. The CBOCP proceeds through the following steps:
Draw from a sample with replacement comprising n (possibly replicated) subject ids. Denote by the bootstrap sample thus obtained, and by the corresponding longitudinal dataset;
Estimate the PRC model of interest using as data. Denote by the model thus fitted;
Estimate the value of the C index of when applied to , , and to the original dataset , ;
Repeat steps 1 to 3 times (typically, or 200);
Compute the optimism correction .
Let denote the naive estimate of the C index. The optimism‐corrected C index, , is obtained by subtracting the optimism correction from the naive C index:
(10) |
2.5. The R package pencal
We have implemented the methodology described in Sections 2.1 to 2.4 in the R package pencal, 11 which can be freely downloaded from CRAN at the link https://cran.r‐project.org/package=pencal. The package comprises functions to estimate the PRC MLPMM(U), PRC MLPMM(U+B), and PRC LMM models, to compute the CBOCP, and to obtain the predicted survival probabilities from the fitted models. A vignette illustrating how to use the package is also available on CRAN. The scripts used for the simulations described in Section 3 are available at https://github.com/mirkosignorelli/pencal.
3. SIMULATIONS
As mentioned in Section 1, the current lack of methods that can predict survival using predictors that are both longitudinal and high‐dimensional forces the use of simpler prediction strategies, which either ignore the longitudinal information, or focus on a limited number of predictors and ignore the remaining ones. In this section, we describe the results of simulations studies designed to compare PRC to a simpler prediction approach where the baseline biomarker levels are employed as predictors in a penalized Cox model. This approach, which we refer to as baseline pCox model hereafter, can handle the large number of available biomarkers, but it ignores the repeated measurements that carry information on their evolution over time. In Sections 3.1 and 3.2, we compare the baseline pCox approach to the PRC LMM, PRC MLPMM(U), and PRC MLPMM(U+B) methods, showing that by properly modeling the dynamic evolution of the biomarkers we can improve the accuracy of predictions in a high‐dimensional setting. In Section 3.3, we compare the distribution of the naive and optimism‐corrected C index and tdAUC, showing the importance of the CBOCP in evaluating overfitting (and correcting for it). Lastly, in Section 3.4, we turn our attention to computing time.
3.1. Evaluation of the PRC LMM and the effect of the number of repeated measurements
We begin our simulation studies by evaluating how the level of improvement in predictive accuracy that can be obtained using PRC can depend on the number of available repeated measurements and on the type of relationship existing between the longitudinal biomarker profiles and survival. For simplicity, here we consider a situation where no distinction between items and latent biological processes is present (namely, ), and we compare the PRC LMM and baseline pCox approaches. In Section 3.2, we will consider more complex scenarios where such distinction is present, and we will employ the PRC MLPMM instead of the PRC LMM.
We first consider a low‐dimensional setting with biomarkers. Without loss of generality, we assume that all patients entered the study at the same age. We generate the longitudinal profiles of each biomarker from the LMM of Equation (3), considering two cases that differ for the number of available repeated measurements per patient (we will refer to these two cases as scenarios with “few” and “many” repeated measurements), and we simulate survival times from a Weibull model. 23 We consider three simulations: in simulation 1, we assume that the random intercepts and random slopes are uncorrelated and have the same variance, that is, and , and that the random intercept and slopes of 6 of the 30 biomarkers contribute to determining the survival times, that is, for , and otherwise, in Equation (8). Thus, in simulation 1, the random intercepts and slopes play an equal role in determining the time‐to‐event outcome. In simulations 2 and 3, instead, we consider cases where the slopes play a more prominent role. In particular, in simulation 2, we still let , but we assume that only the random slopes have non‐null coefficients in Equation (8). Instead, in simulation 3, we set and , and similar to in simulation 1 we assume that both the random intercepts and random slopes of six biomarkers contribute to determining the survival times. Within each simulation, we consider two different sample sizes ( and ). We compute the optimism‐corrected tdAUC (with ) and C index using bootstrap samples of size , and repeat each simulation setting 100 times. Hereafter we present the results obtained using the ridge penalty, and in the Supplementary Material, we show the equivalent results when the elasticnet penalty is used.
Figure 2 shows the distribution of the tdAUC and C index in simulations 1, 2, and 3 for the baseline pCox model, and for the PRC LMM model when few or many repeated measurements are available. In each of the simulation scenarios, we can observe that use of the PRC LMM produces an improvement over the baseline pCox; in particular, the improvement is stronger with the larger sample size (), and when more repeated measurements are available. Furthermore, the improvement is somewhat stronger in simulations 2 and 3, where the survival times are more strongly determined by the progression rates of the biomarkers.
Similar results are obtained when using elasticnet instead of ridge as penalty (Supplementary Figure 1). We can further observe that the distribution of the performance measures across random replicates of the same simulation has somewhat higher variability with elasticnet than with ridge. This result is probably due to the fact that the elasticnet penalty is a linear combination of an and an penalty that comprises two tuning parameters, and . When elasticnet reduces to ridge, and when to the lasso. Because with elasticnet the cross‐validation of the tuning parameters can yield different choices of across random replicates whereas with ridge is fixed, we can indeed expect the distribution of performance measures to be more variable with elasticnet than with ridge.
Next, we consider the same simulation scenarios in a high‐dimensional setting with predictors. We assume that survival time depends on the first 10 predictors, and in simulations 4, 5, and 6 we make the same assumptions on the variances of the random effects and on the relationship between random effects and survival times made in simulations 1, 2, and 3. The results of these simulations using the ridge penalty are presented in Figure 3. When we consider the optimism‐corrected tdAUC estimates, the results are largely in line with those obtained in the low‐dimensional simulations: the gain associated to the use of the PRC LMM model is stronger with larger sample sizes, when more repeated measurements are available, and when (in simulations 5 and 6) the survival times are more strongly associated with the random slopes. When looking at the C index, instead, we see that the gain associated with the PRC‐LMM is apparent with , but not with . This latter result seems to indicate that obtaining performance improvements can be more challenging in a high‐dimensional setting, and that larger sample sizes might be required to accurately evaluate predictiveness in high‐dimensions.
Although similar remarks hold for elasticnet (Supplementary Figure 2), once again we observe a higher variability in the distribution of the performance measurements; furthermore, while the improvement with the PRC LMM method is still apparent in terms of tdAUC, it tends to vanish when looking at the C index.
The primary motivation for the development of PRC is that estimation of joint models becomes computationally unfeasible when more than a handful of longitudinal predictors are available. As a matter of fact, PRC represents a computationally feasible alternative to joint models for the prediction of survival from high‐dimensional longitudinal predictors. Nevertheless, it might be still be interesting to compare the predictive performance of PRC to that of joint models in low‐dimensional settings, where both approaches can be pursued. In Section 3 of the Supplementary Material, we present the results of a simulation study designed to compare the performance of PRC LMM to a joint modeling approach in a situation with only three longitudinal predictors.
3.2. Evaluation of the PRC MLPMM(U) and PRC MLPMM(U+B) methods
We now turn our attention to the validation of the PRC MLPMM(U) and PRC MLPMM(U+B) methods, considering simulations where we have multiple correlated items that refer to the same latent biological process.
Similarly to what we have done in Section 3.1, we begin by considering a low‐dimensional scenario where there are latent biological processes, each measured through items. We generate the longitudinal items using the MLPMM of Equation (1), and survival times using a Weibull model where survival depends on the shared random effects, that is, on . In simulation 7, we set and let survival depend on the shared random effects of 4 of the 10 latent biological processes; in other words, random intercepts and random slopes have the same weight both in the definition of the longitudinal profiles and in the generation of survival times (note that this is a multivariate equivalent of simulations 1 and 4). In simulation 8, we keep the same , but we let survival depend only on the random slopes of the latent processes (in analogy with simulations 2 and 5). Lastly, in simulation 9, we set , and we let survival depend both on the shared random intercepts and on the shared random slopes of the latent processes (this is the equivalent of simulations 3 and 6). Within each simulation, we consider two different sample sizes ( and ). We compute the optimism‐corrected tdAUC at and the C index using bootstrap samples of size 100, and for each simulation setting we consider 100 random replicates. Here we present the results obtained using the ridge penalty, and in the Supplementary Material, the equivalent results when using elasticnet.
The distributions of the tdAUC and C index in simulations 7, 8, and 9 using the ridge penalty are presented in Figure 4. In all simulations, we observe that the two PRC MLPMM methods outperform the baseline pCox both with respect to the tdAUC and to the C index, and with both sample sizes. Moreover, we can observe that the PRC‐MLCMM(U) and PRC‐MLCMM(U+B) models achieve a similar predictive performance, despite the fact that the data are generated in line with the PRC‐MLCMM(U) model, and not with the PRC MLPMM(U+B); as a matter of fact, PRC‐MLCMM(U) performs slightly better, as expected given the way in which the survival times are simulated. Supplementary Figure 3 shows the results of simulations 7, 8, and 9 using the elasticnet penalty; in this case, the results are substantially the same as those obtained with ridge.
Lastly, we assess the same simulation scenarios in a high‐dimensional setting with predictors, where in particular there are latent biological processes, each measured using three items. We assume that survival time depends on the shared random effects of the first 10 latent variables, and in simulations 10, 11, and 12, we make the same assumptions on the variances of the random effects and on the relationship between random effects and survival times made in simulations 7, 8, and 9.
The results of these simulations using the ridge penalty are presented in Figure 5. Similarly to what we observed in simulations 4, 5, and 6 (in Section 3.1), the improvement of the PRC methods over the baseline pCox is apparent with both and when looking at the tdAUC, and with the C index; it is less apparent when we consider the C index in the scenarios. Moreover, in line with what already observed in simulations 7, 8, and 9, the difference between PRC‐MLPMM(U) and PRC‐MLPMM(U+B) is relatively small. Similar remarks hold for elasticnet (Supplementary Figure 4), where, however, we can once again observe a higher variability in the distribution of the performance measurements, similar to simulations 4, 5, and 6.
3.3. Effect of the cluster bootstrap optimism correction
To showcase the effect that the CBOCP has on the level of the reported measures of predictiveness, we have compared the distribution of the naive estimates of the tdAUC and C index to that of the optimism‐corrected estimates obtained with the CBOCP introduced in Section 2.4.
Figure 6 compares the naive and optimism‐corrected estimates of the tdAUC and C index in simulation 10 (similar results hold for the other simulations described in Sections 3.1 and 3.2). It is apparent that the effect of the optimism correction is rather sizeable for all models. For example, in the case of the PRC‐MLPMM(U+B) model with the median of the naive estimates is 0.978 for tdAUC(5) and 0.881 for the C index, whereas the median of the optimism‐corrected estimates is 0.908 for tdAUC(5) and 0.801 for the C index. These results highlight the importance of properly evaluating predictive performance by implementing an internal validation procedure that can estimate the optimism associated to the naive performance estimators.
3.4. Computing time
In this section, we study the typical computing time of the PRC‐LMM and PRC‐MLPMM models and its relation to sample size and number of longitudinal predictors. For this benchmarking, we employ the data generated in simulations 7 and 10, and monitor the mean computing time of a penalized Cox model with baseline predictors, of the PRC‐LMM model, and of the PRC‐MLPMM(U) and PRC‐MLPMM(U+B)) models. Computations were performed using an Intel E7‐4890 processor with 2.2 GHz CPU, and version 0.4.1 of the R package pencal; for each model, computation of the CBOCP was parallelized over 10 cores.
Supplementary Table 1 reports the computing time associated to the estimation of the different models. We can observe that both for the PRC‐LMM and the PRC‐MLPMM models, computing time increases less than linearly with respect to the sample size n. Moreover, computing time increases more than linearly with respect to the number of longitudinal predictors for the PRC‐LMM approach, and approximately linearly for the two PRC‐MLPMM approaches. Similar observations hold for the computation of the CBOCP, whose mean computing times are presented in Supplementary Table 2.
By comparing the computing times in Supplementary Tables 1 and 2, we can observe that while the estimation of PRC is not particular demanding in terms of computing time, the computation of the CBOCP is typically slower. This is due to the fact that bootstrap validation procedures require the repetition of the same computations for a rather large number (typically hundreds) of bootstrap samples. Due to their repetitiveness, these computations can be easily parallelized: for this reason, in the R package pencal we offer the possibility to easily parallelize computations over multiple cores, making it easy for users to reduce computing time through parallelization.
4. APPLICATION
In this section, we illustrate an example application of PRC to the data gathered in the MARK‐MD study, the motivating dataset that we briefly introduced in Section 1. The MARK‐MD study 1 , 2 was an observational study that focused on the validation of a set of blood biomarkers as longitudinal markers of disease progression for DMD, a severe neuromuscular disorder whose consequences include progressive loss of muscular tissue and muscle mass, LoA around the age of 12, and premature death. One of the reasons that motivate the research interest in blood biomarkers for DMD is that the methods currently used to monitor disease progression in DMD (namely, timed tests that measure the distance and speed that a patient can walk, and magnetic resonance imaging) can be motivation‐dependent and rather noisy, so that measuring the level and dynamic change of blood biomarkers would represent a desirable alternative.
Hereafter we focus on the problem of predicting the age at which DMD patients will experience LoA, which is a crucial disease milestone for DMD. The MARK‐MD study led to the collection of 303 longitudinal blood samples from a total of 157 patients. However, 64 of these patients were not ambulant where they entered the study, so here we restrict our attention to the 93 patients who were ambulant when they entered the MARK‐MD study. Of these 93 patients, 55 experienced the event of interest (LoA), whereas 38 are right‐censored. Figure 7 shows the Kaplan‐Meier estimator associated to the LoA event.
An interesting feature of the MARK‐MD study is the fact that the biological markers of interest, that is, the proteins, were not measured directly; instead, between 1 and 5 antibodies, with different affinity and specificity towards a given target protein, were employed to measure each protein. More in detail, an antibody‐based array was employed to measure the abundance of 240 antibodies that targeted 118 proteins in total; 37 proteins were measured using a single antibody, 52 using two different antibodies, 18 with 3 distinct antibodies, 10 with 4 antibodies, and 1 with 5 different antibodies. The 118 proteins partitioned the antibodies into non‐overlapping sets (ie, each antibody targeted a single protein). In other words, we are interested in latent biological process, each of which is measured using between 1 and 5 items, that is, .
In most cases, a moderate/strong positive correlation can be observed between antibodies that target the same protein (see Figure 8 for an example based on the three antibodies used to measure the protein CFH, and Supplementary File 1 of the original study 1 for a more detailed overview), indicating that a multivariate modeling of the antibodies that match the same protein might be more appropriate than modeling each antibody separately.
Table 1 shows the number of repeated blood measurements taken before LoA that are available for each patient. Note that while between 2 and 5 repeated measurements before LoA are available for 48 patients, for 45 patients only 1 measurement can be used: this feature makes the longitudinal modeling of the biomarkers challenging, because the fact that no longitudinal information is available for almost half of the patients makes it is difficult to derive accurate predicted random effects, and in particular to obtain accurate random slopes.
TABLE 1.
Number of repeated measurements before LoA | 1 | 2 | 3 | 4 | 5 |
---|---|---|---|---|---|
Frequency | 45 | 23 | 18 | 6 | 1 |
We estimated four different models that can be employed to predict time to LoA: a penalized Cox model that employs the baseline biomarkers levels as covariates, the PRC‐LMM model that in step 1 models each antibody separately with a LMM with random intercepts and random slopes, and the PRC‐MLCMM(U) and PRC‐MLCMM(U+B) models, which jointly model all antibodies targeting the same protein using the MLPMM. Each of the models was fitted using both the ridge, and the elasticnet penalty. The predictive accuracy of each model was evaluated by estimating the optimism‐corrected C index, and the optimism‐corrected tdAUC evaluated every half year, up to 5 years from baseline. The optimism correction was performed using the CBOCP with bootstrap samples.
The optimism‐corrected C index estimates of the aforementioned models are compared in Table 2. Irrespective of the type of penalty used, the best performing model appears to be the PRC‐MLPMM(U+B). It is interesting to note that whereas this model outperforms the penalized Cox model that uses only the baseline marker values, the PRC‐LMM and PRC‐MLPMM(U) models do not. We believe that this somewhat surprising result may be due to the lack of repeated measurements for almost half of the patients: as a matter of fact, for those 45 patients for which only 1 measurement before LoA is available it is virtually impossible to reliably estimate the random slopes (progression rates). We can expect this complication to affect more severely the performance of the PRC‐LMM and PRC‐MLPMM(U) models than that of the PRC‐MLPMM(U+B) model, because whereas in the PRC‐LMM and PRC‐MLPMM(U) models half of the predictors are random intercepts and half random slopes, in the PRC‐MLPMM(U+B) the random intercepts ( and in Equation 1) employed for prediction outnumber the random slopes (; the total number of random intercepts is 237, vs 118 random slopes).
TABLE 2.
Optimism‐corrected C index | ||
---|---|---|
Model | Ridge penalty | Elasticnet penalty |
Penalized Cox with baseline markers | 0.696 | 0.692 |
PRC‐LMM | 0.668 | 0.666 |
PRC‐MLPMM(U) | 0.640 | 0.659 |
PRC‐MLPMM(U+B) | 0.710 | 0.733 |
Similar results hold for the optimism‐corrected estimates of the tdAUC, which are shown in Figure 9. Once again, the most predictive model appears to be the PRC‐MLPMM(U+B) model when fitted with elasticnet penalty, followed by the same model fitted with the ridge penalty and by the penalized Cox model with baseline measures as covariates.
Lastly, we note that for the baseline penalized Cox and the PRC‐LMM models, the predictive performance is substantially the same when using ridge or elastic net; this is due to the fact that for these two models, the value of the elastic net tuning parameter on the original dataset selected by nested cross‐validation is , meaning that the elastic net model is the same as the ridge model. Thus, the naive estimates of the C index and tdAUC are the same for the two different penalties. Slight differences in the estimation of the optimism correction are however expected, because with elasticnet, is tuned independently for each bootstrap sample during the CBOCP. This, in turn, yields slightly different optimism‐corrected estimates of performance. On the contrary, for the PRC‐MLPMM(U) and PRC‐MLPMM(U+B) models we observe a more substantial difference between ridge and elastic net: this is due to the fact that for these models, , so the models selected by ridge and elastic net are actually different, and have different predictive performance.
5. DISCUSSION
High‐dimensionality and repeated measurements are commonly encountered features of modern biomedical data. However, the current lack of statistical methods that can predict survival using a high‐dimensional set of longitudinally measured predictors limits the possibility to exploit the full predictive potential of complex biomedical datasets. When dealing with a large number of longitudinal predictors, this methodological gap makes it necessary to employ simpler prediction strategies that are either based on the pre‐selection of a limited number of longitudinal covariates as predictors, or on the use of the baseline measurements only. Such simplifications can be detrimental for the performance of risk prediction models, because they discard part of the available information.
In this article, we overcome this limitation by proposing PRC, a method that makes it possible to predict survival using covariates that are at the same time longitudinal and high‐dimensional. PRC begins with the specification of mixed effects models that describe the dynamic evolution of the biomarkers, it proceeds to the computation of subject‐specific summaries of the biomarker's trajectories, and finally it employs such summaries as predictors of the survival outcome (Figure 1). Use of mixed effects models makes it easy to handle unbalanced repeated measurement designs, as well as missing data. Additionally, PRC is capable of handling the presence of multiple items that are employed to measure the same underlying biological process, and which might be highly correlated with each other.
To illustrate PRC, we have employed data from the MARK‐MD study as motivating example. In this study, the measured predictors are antibodies that are employed to reconstruct the abundance of a smaller number of proteins, which cannot be measured directly and can thus be regarded as the latent biological processes of interest. To account for this peculiar feature, we have introduced PRC considering a general framework where multiple items can be employed to measure the same biological process of interest. However, here we would like to emphasize that the applicability of PRC is not limited to studies with such a complex setup: PRC can also be applied to simpler contexts where no distinction exists between the items and the biological processes of interest. In such situations, use of the PRC LMM approach instead of the PRC MLPMM might be preferable.
Moreover, we would like to remark that although we have illustrated PRC using an example where proteomic data are employed as predictors, the applicability of PRC is not restricted to proteomic datasets. As a matter of fact, there is no restriction on the type of longitudinal variables that might be used, so that in practice PRC can be applied to datasets comprising a highly heterogeneous set of variables, which might, for example, include clinical variables, laboratory measurements, different types of omic measurements, and performance tests. Moreover, applicability of PRC is not restricted to problems arising in medicine and biology, as the method can be employed with any type of survival outcomes and longitudinal predictors.
To estimate the penalized Cox models in step 3 of PRC (Section 2.3), we have focused on the use of the ridge and elasticnet penalties, which are commonly employed for prediction problems. If desired, one may consider the use of alternative forms of regularization (lasso, group lasso, SCAD, etc.). In the case of Equation (9), one might also consider a modification of elasticnet that combines the ridge and group lasso penalty to jointly select or exclude all summaries that refer to a single biological process. Although in the R package pencal we allow for the possibility to use the lasso penalty, in general we do not recommend its use, because lasso models typically have lower predictive performance in comparison to ridge and elastic net, and they additionally have problems when dealing with highly correlated predictors. 24
In the simulations presented in Section 3, we observed that ridge and elasticnet usually result in a similar prediction performance; however, the distribution of the optimism‐corrected tdAUC and C index over randomly replicated datasets is often more variable with elasticnet, indicating that use of ridge may yield more stable predictors. Moreover, the estimation of PRC is faster when employing the ridge penalty, because this is a special case of elasticnet. Thus, our practical recommendation is to estimate PRC using the ridge penalty, unless a sparse predictor is desired.
A further contribution of our work is the implementation of the CBOCP, which makes it possible to perform a proper internal validation of the fitted prediction model. As shown in Section 3.3, the optimism correction in high‐dimensional settings can be substantial, indicating that the naive measures of predictive performance may exhibit a strong bias, and that performing the CBOCP is thus necessary to correct for this bias.
Motivated by the presence in the MARK‐MD study of strong correlations between antibodies that target the same protein, in this article, we have proposed a multivariate modeling approach where all items targeting the same biological process are modeled jointly (step 1 of the PRC‐MLPMM model). Primary advantages of this approach are the facts that it can tackle the major source of correlation between the longitudinal antibodies, and that it allows to obtain summaries of the longitudinal trajectories that refer to the latent biological processes. On the other hand, this approach does not model the weaker correlations that exist between antibodies that target different proteins, as specifying a single multivariate mixed model for 240 longitudinal biomarkers is computationally prohibitive. If computationally feasible, an extension of PRC‐MLPMM whereby all 240 longitudinal biomarkers are jointly modeled in step 1 could yield prediction models with higher predictive accuracy.
Lastly, in this article, we have assumed that survival times have been systematically collected and reported as an outcome measured in continuous time. However, in clinical practice, it may happen that the survival status of a patient is only measured at discrete time points, for example, on the dates of a visit, resulting in interval censoring. Dealing with interval‐censored outcomes would require an extension of PRC whereby the penalized Cox model employed in step 3 of PRC is replaced with a model for interval‐censored data suitable for the analysis of high‐dimensional data.
AUTHOR CONTRIBUTIONS
Mirko Signorelli and Roula Tsonaka conceived this study, developed the methodology described in Section 2, planned the simulations presented in Section 3, and analyzed their results. Mirko Signorelli developed the R package pencal, implemented the simulations, and performed the statistical analyses presented in Section 4. Pietro Spitali, Cristina Al‐Khalili Szigyarto, and the MARK‐MD Consortium contributed the biomedical data from the MARK‐MD study that are analyzed in Section 4. Mirko Signorelli, Pietro Spitali, and Roula Tsonaka worked on the application of PRC to the MARK‐MD data. Mirko Signorelli and Roula Tsonaka wrote the manuscript; all coauthors revised the manuscript.
COLLABORATORS
The authors gratefully acknowledge the following collaborators from the MARK‐MD Consortium: Prof. Erik Niks, Department of Neurology, Leiden University Medical Center (NL); Prof. Volker Strauss, MRC Centre for Neuromuscular Diseases, Institute of Genetic Medicine, Newcastle University (UK); Prof. Francesco Muntoni, The Dubowitz Neuromuscular Centre, UCL Institute of Child Health (UK); Dr. Burcu Ayoglu, Department of Protein Science, KTH‐Royal Institute of Technology (SW); Dr. Camilla Johansson, Department of Protein Science, KTH‐Royal Institute of Technology (SW).
Supporting information
ACKNOWLEDGEMENTS
The authors gratefully acknowledge funding from the Association Francaise Contre les Myopathie (grant number 17724), the Stichting Duchenne Parent Project (project 16.006) and Spieren voor Spieren (grant number SvS15). Moreover, they acknowledge the MRC Centre for Neuromuscular Diseases Biobank London and EuroBioBank for giving access to the samples.
Signorelli M, Spitali P, Szigyarto CA‐K, The MARK‐MD Consortium , Tsonaka R. Penalized regression calibration: A method for the prediction of survival outcomes using complex longitudinal and high‐dimensional data. Statistics in Medicine. 2021;40(27):6178–6196. 10.1002/sim.9178
Funding information Association Francaise Contre les Myopathie, 17724; Stichting Duchenne Parent Project, 16.006; Stichting Spieren voor Spieren, SvS15
DATA AVAILABILITY STATEMENT
Data sharing is not applicable to this article as no new data were created or analyzed in this study.
REFERENCES
- 1. Signorelli M, Ayoglu B, Johansson C, et al. Longitudinal serum biomarker screening identifies MDH2 as candidate prognostic biomarker for Duchenne muscular dystrophy. J Cachexia Sarcopenia Muscle. 2020;11:505‐517. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Strandberg K, Ayoglu B, Roos A, et al. Blood‐derived biomarkers correlate with clinical progression in Duchenne muscular dystrophy. J Neuromuscul Dis. 2020;7(3):231‐246. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Van Wieringen WN, Kun D, Hampel R, Boulesteix AL. Survival prediction using gene expression data: a review and comparison. Comp Stat Data Anal. 2009;53(5):1590‐1603. [Google Scholar]
- 4. Rizopoulos D. Joint Models for Longitudinal and Time‐to‐Event Data: With Applications in R. Boca Raton, FL: Chapman & Hall/CRC Press; 2012. [Google Scholar]
- 5. Chen Y, Wang Y. Variable selection for joint models of multivariate longitudinal measurements and event time data. Stat Med. 2017;36(24):3820‐3829. [DOI] [PubMed] [Google Scholar]
- 6. Hickey GL, Philipson P, Jorgensen A, Kolamunnage‐Dona R. Joint modelling of time‐to‐event and multivariate longitudinal outcomes: recent developments and issues. BMC Med Res Methodol. 2016;16(1):117. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Mauff K, Steyerberg E, Kardys I, Boersma E, Rizopoulos D. Joint models with multiple longitudinal outcomes and a time‐to‐event outcome: a corrected two‐stage approach. Stat Comput. 2020;1‐16. [Google Scholar]
- 8. Proust‐Lima C, Amieva H, Jacqmin‐Gadda H. Analysis of multivariate mixed longitudinal data: a flexible latent process approach. Brit J Math Stat Psychol. 2013;66(3):470‐487. [DOI] [PubMed] [Google Scholar]
- 9. McCulloch CE, Searle SR, Neuhaus JM. Generalized, Linear, and Mixed Models. Hoboken, NJ: John Wiley & Sons; 2008. [Google Scholar]
- 10. Ye W, Lin X, Taylor JMG. Semiparametric modeling of longitudinal measurements and time‐to‐event data–a two‐stage regression calibration approach. Biometrics. 2008;64(4):1238‐1246. [DOI] [PubMed] [Google Scholar]
- 11. Signorelli M. pencal: penalized regression calibration (PRC). R package version 0.4.1; 2021. https://CRAN.R‐project.org/package=pencal.
- 12. Proust‐Lima C, Philipps V, Liquet B. Estimation of extended mixed models using latent classes and latent processes: the R package LCMM. J Stat Softw. 2017;78(2):1‐56. [Google Scholar]
- 13. Pinheiro J, Bates D, DebRoy S, Sarkar D, R Core Team . nlme: linear and nonlinear mixed effects models. R package version 3.1‐148; 2020. https://CRAN.R‐project.org/package=nlme.
- 14. Tsonaka R, Mil AHM, Houwing‐Duistermaat JJ. A two‐stage mixed‐effects model approach for gene‐set analyses in candidate gene studies. Stat Med. 2012;31(11‐12):1190‐1202. [DOI] [PubMed] [Google Scholar]
- 15. Ebrahimpoor M, Spitali P, Goeman JJ, Tsonaka R. Pathway testing for longitudinal metabolomics. Stat Med. 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Verweij PJM, Van Houwelingen HC. Penalized likelihood in Cox regression. Stat Med. 1994;13(23‐24):2427‐2436. [DOI] [PubMed] [Google Scholar]
- 17. Simon N, Friedman J, Hastie T, Tibshirani R. Regularization paths for Cox's proportional hazards model via coordinate descent. J Stat Softw. 2011;39(5):1‐13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Engebretsen S, Bohlin J. Statistical predictions with glmnet. Clin Epigenet. 2019;11(1):1‐3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Heagerty PJ, Lumley T, Pepe MS. Time‐dependent ROC curves for censored survival data and a diagnostic marker. Biometrics. 2000;56(2):337‐344. [DOI] [PubMed] [Google Scholar]
- 20. Pencina MJ, D'Agostino RB. Overall C as a measure of discrimination in survival analysis: model specific population value and confidence interval estimation. Stat Med. 2004;23(13):2109‐2123. [DOI] [PubMed] [Google Scholar]
- 21. Steyerberg EW. Clinical Prediction Models: A Practical Approach to Development, Validation, and Updating. Berlin, Germany: Springer Science & Business Media; 2008. [Google Scholar]
- 22. Efron B. Estimating the error rate of a prediction rule: improvement on cross‐validation. J Am Stat Assoc. 1983;78(382):316‐331. [Google Scholar]
- 23. Bender R, Augustin T, Blettner M. Generating survival times to simulate Cox proportional hazards models. Stat Med. 2005;24(11):1713‐1723. [DOI] [PubMed] [Google Scholar]
- 24. Zou H, Hastie T. Regularization and variable selection via the elastic net. J Royal Stat Soc Ser B. 2005;67(2):301‐320. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Data sharing is not applicable to this article as no new data were created or analyzed in this study.