Abstract
Modern health data science applications leverage abundant molecular and electronic health data, providing opportunities for machine learning to build statistical models to support clinical practice. Time-to-event analysis, also called survival analysis, stands as one of the most representative examples of such statistical models. We present a deep-network-based approach that leverages adversarial learning to address a key challenge in modern time-to-event modeling: nonparametric estimation of event-time distributions. We also introduce a principled cost function to exploit information from censored events (events that occur subsequent to the observation window). Unlike most time-to-event models, we focus on the estimation of time-to-event distributions, rather than time ordering. We validate our model on both benchmark and real datasets, demonstrating that the proposed formulation yields significant performance gains relative to a parametric alternative, which we also propose.
1. Introduction
Time-to-event modeling is one of the most widely used statistical analysis tools in biostatistics and, more broadly, health data science applications. For a given subject, these models estimate either a risk score or the time-to-event distribution, from a pre-specified point in time at which a set of covariates (predictors) are observed. In practice, the model is parameterized as a weighted, often linear, combination of covariates. Time is estimated parametrically or nonparametrically, the former by assuming an underlying time distribution and the latter as proportional to observed event times. These models have been widely used in risk profiling (Hippisley-Cox & Coupland, 2013; Cheng et al., 2013), treatment planning, and drug development (Fischl et al., 1987). Time-to-event modeling, and in a larger context, point processes, constitute the fundamental analytical tools in applications for which the future behavior of a system or individual is to be characterized statistically.
The principal time-to-event modeling tool is the Cox Proportional Hazards (Cox-PH) model (Cox, 1992). Cox-PH is a semi-parametric model that assumes the effect of covariates is a fixed, time-independent, multiplicative factor on the hazard rate, which characterizes the instantaneous death rate of the surviving population. By optimizing a partial likelihood formulation, Cox-PH circumvents the difficulty of specifying the unknown, time-dependent, baseline hazard function. Consequently, Cox-PH results in point-estimates proportional to the event times. Further, estimation of Cox-PH models depends heavily on event ordering and not the time-to-event itself, which is known to compromise the scalability of the estimation procedure to large datasets. This poor scaling behavior is manifested because the formulation is not amenable to stochastic training with minibatches.
It is well accepted that the fixed-covariate-effects assumption made in Cox-PH is strong, and unlikely to hold in reality (Aalen, 1994). For instance, individual heterogeneity and other sources of variation, often likely to be dependent on time, are rarely measured or totally unobservable. This unobservable variation has been gradually recognized as a major concern in survival analysis and cannot be safely ignored (Collett, 2015; Aalen et al., 2001). When these sources of variation are independent of time, they can be modeled via fixed or random effects (Aalen, 1994; Hougaard, 1995). However, in cases for which they render the hazard rate time-dependent, such variation is difficult to control, diagnose, or model parametrically. Cox-PH is known to be sensitive to such assumption violations (Aalen et al., 2001; Kleinbaum & Klein, 2010). Moreover, Cox-PH focuses on the estimation of the covariate effects rather than the survival time distribution, i.e., time-to-event prediction. The motivation behind Cox-PH and its shortcomings make it less appealing in applications where prediction is of highest importance.
An alternative to the Cox-PH model is the Accelerated Failure Time (AFT) model (Wei, 1992). AFT makes the simplifying assumption that the effect of covariates either accelerates or delays the event progression, relative to a parametric baseline time-to-event distribution. However, by not making the baseline hazard a constant, as in standard Cox-PH, AFT is often a more reasonable assumption in clinical settings when predictions are important (Wei, 1992). AFT also encompasses a wide range of popular parametric proportional hazards models and proportional odds models, when the event baseline time distribution is specified properly (Klein & Moeschberger, 2005). Learning in AFT models falls into the category of maximum likelihood estimation, and therefore it scales well to large datasets, when trained via stochastic gradient descent. Further, AFT is also more robust to unobserved variation effects, relative to Cox-PH (Keiding et al., 1997).
From a machine learning perspective, recent advances in deep learning are starting to transform clinical practice. Equipped with modern learning techniques and abundant data, machine-learning-driven diagnostic applications have surpassed human-expert performance in a wide array of health care applications (Cheng et al., 2013; 2016; Havaei et al., 2017; Djuric et al., 2017; Gulshan et al., 2016). However, applications involving time-to-event modeling have been largely under-explored. From the existing approaches, most focus on extending Cox-PH with nonlinear neural-network-based covariate mappings (Katzman et al., 2016; Faraggi & Simon, 1995; Zhu et al., 2016), casting the time-to-event modeling as a discretized-time classification problem (Yu et al., 2011; Fotso, 2018), or introducing a nonlinear map between covariates and time via Gaussian processes (Fernández et al., 2016; Alaa & van der Schaar, 2017). Interestingly, all of these approaches focus their applications toward relative risk, fixed-time risk (e.g., 1-year mortality) or competing events, rather than event-time estimation, which is key to individualized risk assessment.
Generative Adversarial Networks (GANs) (Goodfellow et al., 2014) have recently demonstrated unprecedented potential for generative modeling, in settings where the goal is to estimate complex data distributions via implicit sampling. This is done by specifying a flexible generator function, usually a deep neural network, whose samples are adversarially optimized to match in distribution to those from real data. Succesful examples of GAN include generation of images (Radford et al., 2015; Salimans et al., 2016), text (Yu et al., 2017; Zhang et al., 2017) and data conditioned on covariates (Isola et al., 2016; Reed et al., 2016). However, ideas from adversarial learning are yet to be exploited for the challenging task that is time-to-event modeling.
Previous work often represents time-to-event distributions using a limited family of parametric forms, i.e., log-normal, Weibull, Gamma, Exponential, etc. It is well understood that parametric assumptions are often violated in practice, largely because of the model is unable to capture unobserved (nuisance) variation. This fundamental shortcoming is one of the main reasons why non-parametric methods, e.g., Cox proportional hazards, are so popular. Adversarial learning leverages a representation that implicitly specifies a time-to-event distribution via sampling, rather than learning the parameters of a pre-specified distribution. Further, GAN-learning penalizes unrealistic samples, which is a known issue in likelihood-based models (Karras et al., 2018).
The work presented here seeks to improve the quality of the predictions in nonparametric time-to-event models. We propose a deep-network-based nonparametric time-to-event model called a Deep Adversarial Time-to-Event (DATE) model. Unlike existing approaches, DATE focuses on the estimation of time-to-event distributions, rather than event ordering, thus emphasizing predictive ability. Further, this is done while accounting for missing values, high-dimensional data and censored events. The key contributions associated with the DATE model are: (i) The first application of GANs to nonlinear and nonparametric time-to-event modeling, conditioned on covariates. (ii) A principled censored-event-aware cost function that is distribution-free and independent of time ordering. (iii) Improved uncertainty estimation via deep neural networks with stochastic layers. (iv) An alternative, parametric, non-adversarial time-to-event AFT model to be used as baseline in our experiments. (v) Results on benchmark and real data demonstrate that DATE outperforms its parametric counterpart by a substantial margin.
2. Background
Time-to-event datasets are usually composed of three variables , the covariates xi = [xi1, …, xip] ∈ ℝp and event pairs (ti, li), where ti is the time-to-event of interest and li ∈ {0, 1} is a binary censoring indicator. Typically, li = 1 indicates the event is observed, alternatively, li = 0 indicates censoring at ti (when li = 0, an event, if it occurs, transpires after the observation window that ends at time ti). N denotes the size of the dataset.
Time-to-event models characterize the survival function:
defined as the fraction of the population that survives up to time t, or the conditional hazards rate function h(t|x) (see below), defined as the instantaneous rate of occurrence of an event at time t given covariates x. We derive the relationship between h(t|x) and S(t) from standard definitions (Kleinbaum & Klein, 2010):
(1) |
where f(t|x) is the conditional survival density function and S(t|x) is formally the complement of the cumulative conditional density function F(t|x). To solve the hazard-rate differential equation (1), we establish the following relationship between f(t|x), h(t|x) and S(t|x) via the cumulative conditional hazard as
(2) |
(3) |
See Supplementary Material for examples of parametric time-to-event characterizations. Time-to-event models, e.g., Cox-PH and AFT, leverage the results in (2) and (3), to characterize the relationship between covariates x and time-to-event t, when estimating the conditional hazard function h(t|x). Two popular frameworks, Cox-PH and AFT, approach the estimation of h(t|x) using nonparametric and parametric techniques, respectively.
Cox Proportional Hazard
The Cox-PH (Cox, 1992) model is a semi-parametric, linear model where the conditional hazard function h(t|x) depends on time through the baseline hazard h0(t), and independent of covariate x as
(4) |
Provided with the N observation tuples in , Cox-PH estimates the regression coefficients, β ∈ ℝp that maximize the partial likelihood (Cox, 1992):
(5) |
where is independent of the baseline hazard in (4). Note also that (5) only depends on the ordering of ti, for i, …, N, and not their actual values. Cox-PH is nonparametric in that it estimates the ordering of the events, not their times, thus avoiding the need to specify a distribution for h0(t). Several techniques have been developed that assume a parametric distribution for h0(t), in order to estimate the actual time-to-event, however, not nearly as widely adopted as standard Cox-PH. See Bender et al. (2005) for specifications of h0(t) that result in an exponential, Weibull or Gompertz survival density functions. For example, when f(t) = λ exp(−λt), i.e., exponential survival density, thus h0(t) = λ.
Accelerated Failure Time
The AFT model (Wei, 1992) is a popular alternative to the widely used Cox-PH model. In this model, similar to Cox-PH, it is assumed that h(t|x)) = ψ(x)h0(ψ(x)t) where ψ(x) is the total effect of covariates, x, usually through a linear relationship ψ(x) = exp(−x⏉β) wher β represents the regression coefficients. If the conditional survival density function satisfies, f(t|x) = ψ(x)f0(t), i.e., S(t) independent of x like in Cox-PH, then we can write
(6) |
where t0 ∼ p0(t) is the unmoderated time, thus ξ characterizes the baseline survival density distribution. Note the similarity between (4) and (6) despite the differences in their motivation. Different choices of baseline distribution yield a variety of AFT distributions, including Weibull, log-normal, gamma and inverse Gaussian (Klein & Moeschberger, 2005). Intuitively, AFT assumes the effect of the covariates, ψ(x), accelerates or delays the life course, which is often meaningful in a clinical or pharmaceutical setting, and sometimes easier to interpret compared with Cox-PH (Wei, 1992). Empirical evidence has shown that AFT is more robust to missing values and misspecification of the survival function than Cox-PH (Keiding et al., 1997). Gaussian-process-based AFT models (Fine & Gray, 1999; Alaa & van der Schaar, 2017) have been used to model competing risk applications with success.
3. Adversarial Time-to-Event
We develop a nonparametric model for p(t|x), where t is the (non-censored) time-to-event from the time at which covariates x were observed. More precisely, we learn the ability to sample from p(t|x) via approximation q(t|x). Further, we do so without specifying a distribution for the marginal (baseline survival distribution), p0(t), which in AFT is usually assumed log-normal. Like in Cox-PH and AFT, we assume that p0(t) is independent of covariates x.
For censored events, li = 0, we wish the model to have a high likelihood for p(t > ti|xi), while for non-censored events, li = 1, we wish that the pairs {xi, ti} be consistent with data generated from p(t|x)p0(x), where p0(x) is the (empirical) marginal distribution for covariates, from which we can sample but whose explicit form is unknown.
We consider a conditional generative adversarial network (GAN) (Goodfellow et al., 2014), in which we draw approximate samples from p(t|x), for li = 1, as
(7) |
where pϵ(ϵ) is a simple distribution, e.g., isotropic Gaussian or uniform (discussed below). The generator, Gθ(x, ϵ; l = 1) is a deterministic function of x and ϵ, specified as a deep neural network with model parameters θ, that implicitly defines qθ(t|x, l = 1) in a nonparametric manner. We explicitly note that l = 1, to emphasize that all t drawn from this model are event times (non-censored times). Ideally the pairs {x, t} manifested from the model in (7) are indistinguishable from the observed data i.e., the non-censored samples.
Let and be the disjoint subsets of noncensored and censored data, respectively. Given a discriminator function Dϕ(x, t) specified as a deep neural network with model parameters ϕ. The cost function based on the non-censored data has the following form:
(8) |
where pnc (t, x) is the empirical joint distribution responsible for , and the expectation terms are estimated through samples {t, x} ~ pnc(t, x) and ϵ ~ pϵ(ϵ) only. We seek to maximize wrt discriminator parameters ϕ, while seeking to minimize it wrt generator parameters θ. For non-censored data, the formulation in (8) is the standard conditional GAN.
We also leverage the censored data to inform the parameters θ of generative model Gθ (x, ϵ; l = 0). We therefore consider the additional cost function
(9) |
where max(0, ·) encodes that Gθ (x, ϵ; l = 0) incurs no cost as long as the sampled time is larger than the censoring point. Further, pc (t, x) is the empirical joint distribution responsible for , from which samples {t, x} are drawn to approximate the expectation in (9). Note that max(0, ·) is one of many choices; smoothed or margin-based alternatives may be considered, but are not addressed here, for simplicity.
For cases in which the proportion of observed events is low, the costs in (8) and (9) underrepresent the desire that time-to-events must be as close as possible to the ground truth, t. For this purpose, we also impose a distortion loss d( ·, · )
(10) |
that penalizes Gθ (x, ϵ; l = 1) for not being close to the event time t for non-censored events only. In the experiments, we set d(a, b) = ||a – b||1.
The complete cost function is
(11) |
where {λ2, λ3} > 0 are tuning parameters controlling the trade-off between non-censored and censored cost functions relative to the discriminator objective in (8). In our experiments we set λ2 = λ3 = 1, provided that (9) and (10) are written in terms of expectations, thus already account for the proportion differences in and . However, this may not be sufficient in heavily imbalanced cases or when the time domains for and are very different.
The cost function in (11) is optimized using stochastic gradient descent on minibatches from . We maximize wrt ϕ and minimize it wrt θ. The terms and reward Gθ (x, ϵ; l = 1) if it synthesizes data that are consistent , and the term encourages Gθ (x, ϵ; l = 1) to generate event times that are consistent with the data , i.e., larger than censoring times.
Time-to-event uncertainty
The generator in (7) has a single source of stochasticity, ϵ, which in GAN-based models has been traditionally applied as input to the model, independent of covariates x. In a Multi-Layer Perceptron (MLP) architecture, , where h1 denotes the vector of layer-1 hidden units, g(·) is the activation function (RELU in the experiments), W10 and W11 are weight matrices for covariates and noise, respectively, and we have omitted the bias term for clarity.
In a model with multiple layers, the noise term applied to the input tends to have a small effect on the distribution of sampled event times (see experiments). More specifically, samples from qθ(t|x) tend to have small variance. This results in a model with underestimated uncertainty, hence overconfident predictions. This is due to many factors, including compounding effects of activation nonlinearities, layer-wise regularizers (e.g., dropout), and cancelling terms when the support of the noise distribution is the real line (both positive and negative). Although the cost function in (11) rewards the generator for producing (non-censored) event times close to the ground-truth, thus in principle encouraging event time distributions to cover it, this rarely happens in practice. This issue is well-known in the GAN literature (Salimans et al., 2016).
Here we take a simple approach consisting on adding sources of stochasticity to every layer of the generator as hj = g(Wj0hj–1+Wj1ϵj), where j = 1, …, L and L is the number of layers. By doing this, we encourage increased coverage on the event times produced by the generator, without substantially changing the model or the learning procedure. In the experiments, we use a multivariate uniform distribution, ϵj ~ Uniform(0, 1), for j = 1, …, L, over Gaussian to reduce cancelling effects. As we show empirically, this approach produces substantially better coverage compared to having noise only on the input layer and with out the convergence issues associated with the additional stochasticity.
In Figure 1 we illustrate the contribution of the noise on each layer to the distribution of event times. In this example we show 10 test-set estimated time-to-event distributions using a 2-layer model with noise sources in all layers, including the input. We see that ground-truth times are nicely covered by the estimated distributions. Also, that the combination of noise sources, rather than any individual source jointly contribute to the desired distribution coverage. Additional examples, including censored times, are found in the Supplementary Material.
4. Baseline AFT Model
In the experiments, we will demonstrate the capabilities of the DATE model. However, since there are not many scalable nonlinear time-to-event models focused on event time estimation, below we present a parametric (non-adversarial) model to be used in the experiments as baseline. We do so with the goal of providing a fair comparison and, at the same time, to highlight the distinguishing factors of the DATE model. Starting from (6), we consider the following MLP-based log-normal AFT:
(12) |
where μβ(x) and are MLPs parameterized by β representing the mean and variance of the log-transformed time-to-event as a function of covariates x. For convenience we adopt the log-normal distribution for event time t, mainly because we found that it is considerably more stable during optimization, compared to other popular survival distributions, e.g., Weibull or Gamma.
Note that (12) is very different from (7), where the generator implicitly defines the distribution q(t|x). The likelihood function of the log-normal AFT model in (12) for all events (censored and non-censored) is then
(13) |
where ϕ(·) and Φ(·) are the Gaussian density and cumulative density functions, respectively, and v(ti, xi) = (log ti – μβ(xi))/σβ(xi). The likelihood in (13) is convenient, because it allows estimation of time-to-event, while seemly accounting for censored events. The latter comes as benefit of having a parametric model with closed-form cumulative density function.
The cost function for the Deep Regularized AFT (DRAFT) model is written as:
(14) |
where the first term is the negative log-likelihood loss from (13), η > 0 is a tuning parameters, and is a regularization loss that encourages event times to be properly ordered. Specifically, we use the following cost function adapted from Steck et al. (2008):
where is the set of all pairs {i, j} in {1, …, N} for which the second argument is observed, i.e., li = 1, and σ(·) is the sigmoid function. The cost function above is a lower bound on the Concordance Index (CI) (Harrell et al., 1984), which constitutes a difficult-to-optimize discrete objective, that is widely used as performance metric for survival analysis, precisely because it captures time-to-event order. Further, it is reminiscent of the partial-likelihood of Cox-PH in (5), but is more amenable to stochastic training.
5. Related Work
Deep learning models, specifically MLPs, have been successfully integrated with Cox-PH-based objectives to improve risk estimation in time-to-event models. Faraggi & Simon (1995) proposed an neural-network-based model optimized using the standard partial-likelihood cost function from Cox-PH. Katzman et al. (2016) is similar to Faraggi & Simon (1995), but leverages modern deep learning techniques such as weight decay, batch normalization and dropout. Luck et al. (2017) replaced the partial-likelihood formulation in (5) with Efron’s approximation (Efron, 1977) and an isotonic regression cost function adapted from Menon et al. (2012) to handle censored events. Zhu et al. (2016) proposed a time-to-event model for image covariates based on convolutional networks.
From the Gaussian process literature, Fernández et al. (2016) proposed a time-to-event model inspired by a Poisson process, where the nonlinear map between covariates and time is modeled as a Gaussian process on the Poisson rate. More recently, Alaa & van der Schaar (2017) proposed a deep multi-task Gaussian process model for survival analysis with competing risks, and learned via variational inference.
Following a different path, other approaches recast the timeto-event problem as a classification task. Yu et al. (2011) proposed a linear model where (discretized) time is estimated using a sequence of dependent regressors. More recently, Fotso (2018) extended their approach to a nonlinear mapping of covariates using deep neural networks. Generative approaches have also been proposed to infer survival-time distributions with variational inference. Deep Survival Analysis (DSA) (Ranganath et al., 2016) specifies a latent model that leverages deep exponential family distribution, however their approach does not handle censored events.
All the above methods focus on relative risk quantified as the CI on the ordering of event times, or fixed-time risk, e.g., 1-year mortality. However, relative risk is most useful when associated with covariate effects, which is difficult in nonlinear models based either on neural networks or Gaussian processes. Fixed-time risk, although very useful in practice, can be recast as a classification problem rather than a substantially more complex time-to-event model. Importantly, none of these approaches consider the task of time-to-event estimation, despite the fact that Gaussian process and generative approaches can be repurposed for such task.
6. Experiments
The loss functions in (11) and (14) for DATE and DRAFT, respectively, are minimized via stochastic gradient descent. At test time, we draw 200 samples from (7) and (12) for DATE and DRAFT, respectively, and use medians for quantitative results requiring point estimates, i.e., , where ts is a sample from the trained model. Detailed network architectures, optimization parameters and initialization settings are in the Supplementary Material. TensorFlow code to replicate experiments can be found at https://github.com/paidamoyo/adversarial_time_to_event.
Comparison Methods
For non-deep learning based models, we considered arguably the two most popular approaches to time-to-event modeling, namely, (regularized) Cox-Efron and RSF. For deep learning models, we considered DRAFT, which generalizes existing neural-network-based methods by using both a parametric log-normal AFT objective and a non-parametric ordering cost function. Extending DRAFT to a mixture of log-normal distributions with different variances but shared mean, did not result in improved performance. We did not consider (variational) models for non-censored events only because learning from censored events is one of the main defining characteristics of time-to-event modeling (otherwise, the model essentially becomes non-negative regression). Further, in most practical situations the proportion of non-censored events is low, e.g., 24% in the EHR.
Datasets
Our model is evaluated on 4 diverse datasets: i) FLCHAIN: a public dataset introduced in a study to determine whether non-clonal serum immunoglobin free light chains are predictive of survival time (Dispenzieri et al., 2012). ii) support: a public dataset introduced in a survival time study of seriously-ill hospitalized adults (Knaus et al., 1995). iii) seer: a public dataset provided by the Surveillance, Epidemiology, and End Results Program. See (Ries et al., 2007) for details concerning the definition of the 10-year follow-up breast cancer subcohort used in our experiments. iv) EHR: a large study from Duke University Health System centered around inpatient visits due to comorbidities in patients with Type-2 diabetes.
The datasets are summarized in Table 1, where p denotes the number of covariates to be analyzed, after one-hot-encoding for categorical (cat) variables. Events indicates the proportion of the observed events, i.e., those for which li = 1. NAN indicates the proportion of missing entries in the N × p covariate matrix and tmax is the time range for both censored and non-censored events.
Table 1.
ehr | flchain | support | seer | |
---|---|---|---|---|
Events (%) | 23.9 | 27.5 | 68.1 | 51.0 |
N | 394,823 | 7,894 | 9,105 | 68,082 |
p (cat) | 729 (106) | 26 (21) | 59 (31) | 789 (771) |
NaN (%) | 1.9 | 2.1 | 12.6 | 23.4 |
t max | 365days | 5,215days | 2,029days | 120months |
Details about the public datasets: FLCHAIN, SUPPORT and SEER, including preprocessing procedures, can be found in the references provided above. EHR is a study designed to track primary care encounters of 19, 064 Type-2 diabetes patients over a period of 10 years (2007–2017). The purpose of the analysis is to predict diabetes-related causes of hospitalization within 1 year of an EHR-recorded primary care encounter. Data is processed and analyzed at the patient encounter-level. The total number of encounters is N = 394, 823. To avoid bias due to multiple encounters per patient, we split the training, validation and test sets so that a given patient can only be in one of the sets. The covariates, collected over a period of a year before the primary care encounter of interest, consist of a mixture of continuous and categorical summaries extracted from electronic health records: vitals and labs (minimum, maximum, count and mean values); comorbidities, medications and procedures ICD-9/10 codes (binary indicators and counts); and demographics (age, gender, race, language, smoking indicator, type of insurance coverage, as either continuous or categorical variables).
6.1. Qualitative results
First, we visually compare the test-set time-to-event distributions by DATE and DRAFT on FLCHAIN data. In Figure 2 we show the top best (left) and worst (middle-left) predictions on censored events, and the top best (middle- right) and worst (right) predictions on non-censored events. Circles denote ground-truth events or censoring points, while box-plots represent distributions over 200 samples for both DATE and DRAFT models. We see that: (i) in nearly every case, DATE is more accurate than DRAFT. (ii) DRAFT tends to make predictions outside the event range (tmax = 5, 215 days), denoted as a horizontal dashed line. (iii) DRAFT tends to overestimate the variance of its predictions, approximately by one order of magnitude relative to DATE. This is not very surprising as DRAFT has an MLP dedicated to estimate, conditioned on the covariates, the variance of the time-to-event distribution. However, note that variances estimated well over the domain of the events (tmax) are not necessarily meaningful or desirable. Figures with similar findings for the other 3 datasets can be found in the Supplementary Material.
To provide additional insight into the performance of DATE compared to DRAFT, we report the Normalized Relative Error (NRE) defined as and for non-censored and censored events, respectively, where t, and tmax denote the ground-truth time-to-event, median time estimated (from samples) and event range, as indicated in Table 1. The NRE distribution provides a visual representation of the extent of test-set errors, while revealing whether the models are biased toward either overestimating or underestimating the event times. Although models with unbiased NREs are naturally preferred, in most clinical applications where being conservative is important, overestimated time-to-events must be avoided as much as possible. Figure 3 shows NRE distributions for test-set non-censored events on support and ehr data. We see that DRAFT results in a considerable amount of errors beyond the event range (|NRE| > 1), tmax = 120 months or tmax = 365 days for support and ehr, respectively. Further, we see that the NRE distribution for DRAFT is heavily skewed toward NRE > 1, thus tending to overestimate event times. On the other hand, DATE produces errors substantially more concentrated around 0 and within |NRE| < 1, relative to DRAFT. This demonstrates the advantage of the adversarial method DATE over the likelihood-based method DRAFT in generating realistic samples. Similar results were observed on the other datasets for both censored and non-censored events. See Supplementary Material for additional figures.
6.2. Quantitative results
Relative absolute error
The perfromance of DATE is evaluated in terms of absolute error relative to the event range, i.e., . For censored events, the relative error is defined as , to account for the fact that no error is made as long as . Table 2 shows median and 50% empirical intervals for relative absolute errors on non-censored events, on all test-data. Results on censored data are small and comparable across approaches, and are thus presented in the Supplementary Material. Specifically, we see that DATE outperforms DRAFT in 3 our of 4 cases by a substantial margin, and is comparable on the support data. For instance, on ehr data 75% of all DATE test-set predictions have a relative absolute error less than 43% (approx. 156 days) which is substantially better than the 81% (approx. 295 days) by DRAFT.
Table 2.
DATE | DATE-AE | DRAFT | |
---|---|---|---|
ehr | 23.6 (11.1,43.0) | 24.5(12.4,44.0) | 36.7(16.1,81.3) |
flchain | 19.5(9.5,31.1) | 19.3 (8.9,32.4) | 26.2(9.0,53.5) |
support | 2.7(0.4,16.1) | 1.5 (0.4,19.2) | 2.0(0.2,35.3) |
seer | 18.6 (8.3,34.1) | 20.2(10.3,35.8) | 23.7(9.9,51.2) |
Missing data
Since missing data are common in clinical data, e.g., seer data contains 23.4% missing values, we also consider a modified version of DATE, where the generator in (7) takes the form t = Gθ (z, ϵ, l = 1), where z is modeled as an adversarial autoencoder (Dumoulin et al., 2016; Li et al., 2017; Pu et al., 2017) with an encoder/decoder pair specified similar to DATE. See Supplementary Material for additional details. This model, denoted in Table 2 as DATE-AE, does not require missing covariates to be imputed before hand, which is the case of DATE and DRAFT as specified originally. The results show no substantial performance improvement by DATE-AE, relative to DATE; results indicate that all of these approaches (DRAFT, DATE and DATE-AE) are robust to missing data. As a benchmark, we took FLCHAIN and support, to then artificially introduced missing values ranging in proportion from 20% to 50%. Results in Supplementary Material support the idea of robustness, since all three approaches resulted in median relative absolute errors within 1% of those in Table 2.
We also tried to quantify statistically the match between time-to-event samples generated from DATE and those from the empirical distribution of the data, using the distribution-free two-sample test based on Maximum Mean Discrepancy (MMD) proposed by Sutherland et al. (2017). Due to sample size limitations (number of non-censored events in the test-set) and high-variances on the p-value estimates, we could not reliably reject the hypothesis that real and DATE samples are drawn from the same distribution. We did confirm it for DRAFT, which is not surprising considering both qualitative and quantitative results discussed above.
Concordance Index
The concordance Index (CI) (Harrell et al., 1984), which quantifies the degree to which the order of the predicted times is consistent with the ground truth, is the most well-known performance metric in survival analysis. Although not the focus of our approach, we compared DATE to DATE-AE, DRAFT, Random Survival Forests (Ishwaran et al., 2008), and Cox-PH (with Efron’s approximation (Efron, 1977)). We found all of these models to be largely comparable. The results, presented in the Supplementary Material, show DATE(-AE) and DRAFT being the best-performing models on EHR and support, respectively. On SEER, DATE(-AE) and DRAFT outperform Cox-PH and RSF. Finally, on FLCHAIN, the smallest dataset, all methods perform about the same. Note that unlike Cox-PH and DRAFT in (5) and (14), DATE(-AE) does not explicitly encourage proper time-ordering on the objective function; it is consequently deemed a strength of the proposed GAN-based DATE model that it properly learns ordering, without needing to explicitly impose this condition when training. DATE does not have a clear advantage in terms of learning the correct order, however, we verified empirically that adding an ordering cost function, in (14), to DATE does not improve the results.
Distribution coverage
We now demonstrate that the DATE model, with noise sources on all layers, has time-to-event distributions with larger variances than versions of DATE with noise only on the input of the neural network. Table 3 shows the median of the 95% intervals for all test-set time-to-event distributions on support data. DATE with Uniform(0,1) has larger variance and coverage compared to the other alternatives, while keeping relative absolute errors and CIs largely unchanged (see Supplementary Material for details). We did not run models with Uniform(−1,1) and Gaussian(0,1) only on the output layer, because from the other results presented above it is clear that these two options are not nearly as good as having Uniform(0,1) noise on all layers. Note also that we did not include DRAFT in these comparisons. DRAFT has naturally good coverage due to the variance of the time-to-event distributions being modeled independent for each observation as a function of the covariates (see for instance Figure 2). However, DRAFT has difficulties keeping good coverage while maintaining good performance, i.e,, small absolute relative error.
Table 3.
Uniform(−1,1) | Uniform(0,1) | Gaussian(0,1) | |
---|---|---|---|
Non-censored | |||
All | 60.0(3.9,176.5) | 149.9(8.5,926.8) | 37.9(3.5,237.4) |
Input | 28.9(1.8,114.8) | 22.4(1.5,91.2) | 33.7(1.6,127.6) |
Output | —- | 168.8(16.6,844.3) | —- |
| |||
Censored | |||
All | 231.3(177.2,332.1) | 1397.3(990.9,2000.1) | 350.5(254.4,539.3) |
Input | 137.3(99.4,205.0) | 86.9(64.4,135.0) | 155.8(106.7,229.3) |
Output | —- | 1158.6(873.8,1670.4) | —- |
7. Conclusions
We have presented an adversarially-learned time-to-event model that leverages a distribution-free cost function for censored events. The proposed approach extends GAN models to time-to-event modeling with censored data, and it is based on deep neural networks with stochastic layers. The model yields improved uncertainty estimation relative to alternative approaches. As a baseline model for our experiments, we also proposed a parametric AFT-based with a parametric log-normal distribution on the time of event. To the best of our knowledge, this work is the first to leverage adversarial learning to improve estimation of time-to-event distributions, conditioned on covariates. Experimental results on challenging time-to-event datasets showed that DATE, our adversarial solution, consistently outperforms DRAFT, its parametric (log-normal) counterpart. As future work, we will extend DATE to models with competing risks and longitudinally measured covariates.
Supplementary Material
Acknowledgements
The authors would like to thank the anonymous reviewers for their insightful comments. This research was supported in part by DARPA, DOE, NIH, ONR, NSF and SANOFI. The authors would also like to thank Shuyang Dai, Liqun Chen, Rachel Ballantyne Draelos, Xilin Cecilia Shi, Yan Zhao and Dr. Neha Pagidipati, for fruitful discussions.
References
- Aalen OO Effects of frailty in survival analysis. Statistical Methods in Medical Research, 1994. [DOI] [PubMed] [Google Scholar]
- Aalen OO, Gjessing HK, et al. Understanding the shape of the hazard rate: A process point of view (with comments and a rejoinder by the authors). Statistical Science, 2001. [Google Scholar]
- Alaa AM and van der Schaar M Deep Multi-task Gaussian Processes for Survival Analysis with Competing Risks. In NIPS, 2017. [Google Scholar]
- Bender R, Augustin T, and Blettner M Generating survival times to simulate Cox proportional hazards models. Statistics in medicine, 2005. [DOI] [PubMed] [Google Scholar]
- Cheng J-Z, Ni D, Chou Y-H, Qin J, Tiu C-M, Chang Y-C, Huang C-S, Shen D, and Chen C-M Computer-aided diagnosis with deep learning architecture: applications to breast lesions in US images and pulmonary nodules in CT scans. Scientific reports, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cheng W-Y, Yang T-HO, and Anastassiou D Development of a prognostic model for breast cancer survival in an open challenge environment. Science translational medicine, 2013. [DOI] [PubMed] [Google Scholar]
- Collett D Modelling survival data in medical research. CRC press, 2015. [Google Scholar]
- Cox DR Regression models and life-tables. In Breakthroughs in statistics. 1992. [Google Scholar]
- Dispenzieri A, Katzmann JA, Kyle RA, Larson DR, Therneau TM, Colby CL, Clark RJ, Mead GP, Kumar S, Melton LJ, et al. Use of nonclonal serum immunoglobulin free light chains to predict overall survival in the general population. In Mayo Clinic Proceedings, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Djuric U, Zadeh G, Aldape K, and Diamandis P Precision histology: how deep learning is poised to revitalize histomorphology for personalized cancer care. npj Precision Oncology, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dumoulin V, Belghazi I, Poole B, Lamb A, Arjovsky M, Mastropietro O, and Courville A Adversarially learned inference. arXiv, 2016. [Google Scholar]
- Efron B The efficiency of Cox’s likelihood function for censored data. JASA, 1977. [Google Scholar]
- Faraggi D and Simon R A neural network model for survival data. Statistics in medicine, 1995. [DOI] [PubMed] [Google Scholar]
- Fernández T, Rivera N, and Teh YW Gaussian processes for survival analysis. In NIPS, 2016. [Google Scholar]
- Fine JP and Gray RJ A proportional hazards model for the subdistribution of a competing risk. JASA, 1999. [Google Scholar]
- Fischl MA, Richman DD, Grieco MH, Gottlieb MS, Volberding PA, Laskin OL, Leedom JM, Groopman JE, Mildvan D, Schooley RT, et al. The efficacy of azidothymidine (AZT) in the treatment of patients with AIDS and AIDS-related complex. New England Journal of Medicine, 1987. [DOI] [PubMed] [Google Scholar]
- Fotso S Deep neural networks for survival analysis based on a multi-task framework. arXiv, 2018. [Google Scholar]
- Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, and Bengio Y Generative adversarial nets. In NIPS, 2014. [Google Scholar]
- Gulshan V, Peng L, Coram M, Stumpe MC, Wu D, Narayanaswamy A, Venugopalan S, Widner K, Madams T, Cuadros J, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA, 2016. [DOI] [PubMed] [Google Scholar]
- Harrell FE, Lee KL, Califf RM, Pryor DB, and Rosati RA Regression modelling strategies for improved prognostic prediction. Statistics in medicine, 1984. [DOI] [PubMed] [Google Scholar]
- Havaei M, Davy A, Warde-Farley D, Biard A, Courville A, Bengio Y, Pal C, Jodoin P-M, and Larochelle H Brain tumor segmentation with deep neural networks. Medical image analysis, 2017. [DOI] [PubMed] [Google Scholar]
- Hippisley-Cox J and Coupland C Predicting risk of emergency admission to hospital using primary care data: derivation and validation of QAdmissions score. BMJ open, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hougaard P Frailty models for survival data. Lifetime data analysis, 1995. [DOI] [PubMed] [Google Scholar]
- Ishwaran H, Kogalur UB, Blackstone EH, and Lauer MS Random survival forests. The annals of applied statistics, 2008. [Google Scholar]
- Isola P, Zhu J-Y, Zhou T, and Efros AA Image-to-image translation with conditional adversarial networks. arXiv, 2016. [Google Scholar]
- Karras T, Aila T, Laine S, and Lehtinen J Progressive growing of GANs for improved quality, stability, and variation. In ICLR, 2018. [Google Scholar]
- Katzman J, Shaham U, Bates J, Cloninger A, Jiang T, and Kluger Y Deep survival: A deep cox proportional hazards network. arXiv, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Keiding N, Andersen PK, and Klein JP The role of frailty models and accelerated failure time models in describing heterogeneity due to omitted covariates. Statistics in medicine, 1997. [DOI] [PubMed] [Google Scholar]
- Klein JP and Moeschberger ML Survival analysis: techniques for censored and truncated data. Springer Science & Business Media, 2005. [Google Scholar]
- Kleinbaum DG and Klein M Survival analysis. Springer, 2010. [Google Scholar]
- Knaus WA, Harrell FE, Lynn J, Goldman L, Phillips RS, Connors AF, Dawson NV, Fulkerson WJ, Califf RM, Desbiens N, et al. The support prognostic model: objective estimates of survival for seriously ill hospitalized adults. Annals of internal medicine, 1995. [DOI] [PubMed] [Google Scholar]
- Li C, Liu H, Chen C, Pu Y, Chen L, Henao R, and Carin L Alice: Towards understanding adversarial learning for joint distribution matching. In NIPS, 2017. [Google Scholar]
- Luck M, Sylvain T, Cardinal H, Lodi A, and Bengio Y Deep Learning for Patient-Specific Kidney Graft Survival Analysis. arXiv, 2017. [Google Scholar]
- Menon AK, Jiang XJ, Vembu S, Elkan C, and Ohno-Machado L Predicting accurate probabilities with a ranking loss. In ICML, 2012. [PMC free article] [PubMed] [Google Scholar]
- Pu Y, Chen L, Dai S, Wang W, Li C, and Carin L Symmetric Variational Autoencoder and Connections to Adversarial Learning. arXiv, 2017. [Google Scholar]
- Radford A, Metz L, and Chintala S Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv, 2015. [Google Scholar]
- Ranganath R, Perotte A, Elhadad N, and Blei D Deep survival analysis. In Machine Learning for Healthcare Conference, 2016. [Google Scholar]
- Reed S, Akata Z, Yan X, Logeswaran L, Schiele B, and Lee H Generative adversarial text to image synthesis. arXiv, 2016. [Google Scholar]
- Ries LAG, Young JL Jr, Keel GE, Eisner MP, Lin YD, and Horner M-JD Cancer survival among adults: US SEER program, 1988–2001. Patient and tumor characteristics SEER Survival Monograph Publication, 2007. [Google Scholar]
- Salimans T, Goodfellow I, Zaremba W, Cheung V, Radford A, and Chen X Improved techniques for training gans. In NIPS, 2016. [Google Scholar]
- Steck H, Krishnapuram B, Dehing-oberije C, Lambin P, and Raykar VC On ranking in survival analysis: Bounds on the concordance index. In NIPS, 2008. [Google Scholar]
- Sutherland DJ, Tung H-Y, Strathmann H, De S, Ramdas A, Smola A, and Gretton A Generative models and model criticism via optimized maximum mean discrepancy. ICLR, 2017. [Google Scholar]
- Wei L-J The accelerated failure time model: a useful alternative to the Cox regression model in survival analysis. Statistics in medicine, 1992. [DOI] [PubMed] [Google Scholar]
- Yu C-N, Greiner R, Lin H-C, and Baracos V Learning patient-specific cancer survival distributions as a sequence of dependent regressors. In NIPS, 2011. [Google Scholar]
- Yu L, Zhang W, Wang J, and Yu Y Seqgan: Sequence generative adversarial nets with policy gradient. In AAAI, 2017. [Google Scholar]
- Zhang Y, Gan Z, Fan K, Chen Z, Henao R, Shen D, and Carin L Adversarial feature matching for text generation. In ICML, 2017. [Google Scholar]
- Zhu X, Yao J, and Huang J Deep convolutional neural network for survival analysis with pathological images. In Bioinformatics and Biomedicine (BIBM), 2016 IEEE International Conference on, 2016. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.