ABSTRACT
Bayesian predictive probabilities of success (PPoS) use interim trial data to calculate the probability of trial success. These quantities can be used to optimise trial size or to stop for futility. In this paper, we describe a simulation‐based approach to compute the PPoS for clinical trials with competing event data, for which no specific methodology is currently available. The proposed procedure hinges on modelling the joint distribution of time to event and event type by specifying Bayesian models for the cause‐specific hazards of all event types. This allows the prediction of outcome data at the conclusion of the trial. The PPoS is obtained by numerically averaging the probability of success evaluated at fixed parameter values over the posterior distribution of the parameters. Our work is motivated by two randomised clinical trials: the I‐SPY COVID phase II trial for the treatment of severe COVID‐19 (NCT04488081) and the STHLM3 prostate cancer diagnostic trial (ISRCTN84445406), both of which are characterised by competing event data. We present different modelling alternatives for the joint distribution of time to event and event type and show how the choice of the prior distributions can be used to assess the PPoS under different scenarios. The role of the PPoS analyses in the decision‐making process for these two trials is also discussed.
Keywords: Bayesian predictive probability of success, clinical trials, competing events, interim analysis
Abbreviations
- DMC
data monitoring committee
- HR
hazard ratio
- PCa
prostate cancer
- PCH
piecewise constant cause‐specific hazard
- PCM
prostate cancer mortality
- PPoS
predictive probability of success
- PSA
prostate‐specific antigen
- RR
risk ratio
1. Introduction
The interim monitoring of clinical trials has become an essential component of trial design [1]. Bayesian methods are particularly well suited for this purpose, as they enable the prediction of the probability of trial success at its conclusion based on interim data, averaged over the posterior distribution of the treatment effect and other nuisance parameters [2]. By predicting the probability that the trial will provide sufficient evidence to support the study's primary hypothesis if the trial continues to its end, the Bayesian predictive probability of success (PPoS) can inform decisions regarding the continuation, modification, or early termination of the trial. If the PPoS is too low, and the trial is therefore predicted to be unlikely to achieve its objective, the trial may be stopped early for futility [3, 4, 5], reducing the risk of exposing participants to ineffective or harmful therapies and conserving resources that can be reallocated to more promising trials [1]. Conversely, if completing the prespecified follow‐up period for participants who have already been enrolled is predicted to lead to success with a sufficiently high probability, enrolment of new participants can be stopped, and the primary outcome can be analysed at the conclusion of the follow‐up period [3, 6]. Furthermore, the PPoS can be employed to adjust the sample size at interim analyses [7] or to inform the timing of the primary analysis, by determining the minimum duration of the follow‐up period required to achieve a certain PPoS [8].
Predictive probabilities of success have been primarily employed to monitor clinical trials with continuous or binary endpoints, e.g., in the I‐SPY 2 trial [9] and the REST trial [10]. In contrast, the use of PPoS in clinical trials with survival data is considerably less common [6, 11, 12]. One possible approach to deriving PPoS relies on asymptotic approximations of the posterior distribution or the test statistic for the parameter of interest, as well as the approximation of the information fraction at interim analysis [13, 14, 15]. PPoS based on these asymptotic arguments has been compared with those based on more flexible and general simulation‐based approaches [13, 14, 16], also referred to as imputed predictive probabilities [15]. Results showed that the PPoS computed using these approximations do not always align with the corresponding simulation‐based PPoS. In fact, simulation may still be necessary to assess the performance of approximation‐based PPoS [15]. Existing simulation‐based methods for survival data are limited to those settings where the endpoint is defined as the time span from randomisation until the occurrence of a single event type [7, 13, 14, 17, 18]. Furthermore, these approaches adopt simple Bayesian survival models for the distribution of the time‐to‐event outcome, such as exponential [7, 13, 17] or piecewise exponential models [14], and leverage conjugate priors to derive closed‐form posterior distributions. While these choices allow for tractable analytical solutions, they limit the flexibility and applicability of these methods. In these simulation‐based approaches, the follow‐up times for the patients still at risk at interim monitoring are censored and these data contribute to the fit of the Bayesian survival model. The predicted survival times for these patients are then sampled from the posterior predictive distribution. To the best of our knowledge, no specific methodology has been proposed to account for competing event data when computing the PPoS.
To fill this gap, in this paper we take the approach of computing PPoS using statistical simulation [19] and extend it to the setting of survival data with competing events. Competing events occur frequently in survival data from clinical trials, where participants may experience different types of events and the occurrence of one event precludes the occurrence of all other event types [20]. For example, in randomised clinical trials evaluating potential treatments for hospitalised patients diagnosed with COVID‐19, hospital discharge due to recovery and death without prior recovery are mutually exclusive (i.e., competing) events [21]. We propose to model the joint distribution of time to event and event type by specifying parametric Bayesian models for the cause‐specific hazard functions, allowing the prediction of the outcome data at the conclusion of the trial and thus the computation of the PPoS. Our suggested procedure provides a general approach that can be used for interim monitoring of newly designed trials or to extend existing designs, for example for phase II [22] or phase III clinical trials [7], to competing event data.
Our work is motivated by two randomised clinical trials: the I‐SPY COVID phase II trial for the treatment of severe COVID‐19 [23] (NCT04488081) and the STHLM3 prostate cancer diagnostic trial [24] (ISRCTN84445406). Both trials are characterised by a time‐to‐event outcome, namely time to recovery from COVID and time to prostate cancer mortality. In the I‐SPY COVID trial, death from any cause is a competing event and the PPoS was used at interim monitoring to inform the decision to continue randomisation of patients to a specific investigational agent. In the STHLM3 trial, death from causes other than prostate cancer is a competing event and the PPoS was employed to predict the shortest follow‐up period necessary to accumulate enough prostate cancer deaths in order to reach sufficient power for the final analysis.
2. Methods
2.1. Notation and Preliminaries
For each trial participant , where , let be the absolutely continuous variable representing the true event time with support on the positive real half‐line, and let be the competing event indicator. The right‐censored event time is denoted by , where is the censoring time, and is the observed event indicator, where 0 indicates right censoring, and 1(·) is the indicator function. Lastly, let be a vector of baseline covariates, including at least the randomisation arm . In total, we define the random vector . Hereafter, we suppress the subscript to simplify notation.
The cause‐specific hazard for the ‐th competing event is defined as
The cause‐specific hazards for all events collectively define the joint distribution of [25]. The crude cumulative probabilities (risks) of experiencing the competing event by the time are defined as .
At interim monitoring, the goal is to predict the probability of trial success at the final analysis given the interim data and the assumed prior distributions on the parameter vector for the data‐generating process used to predict future data. Let be the observed data at the time of interim analysis, be the future data available only at the time of final analysis, which encompasses both observed and yet‐to‐be‐observed data, and let be the critical region such that if the trial will be declared a success. The PPoS, introduced by Spiegelhalter et al. [2], is defined as:
In the PPoS, the conditional power , defined as the probability of trial success given the interim data and a fixed , is averaged over the posterior distribution of , denoted . The definition of trial success can be based on Bayesian posterior probabilities, as in the I‐SPY COVID trial, or on frequentist p‐values, as in the STHLM3 trial [3].
Three possible types of data must be considered when computing the PPoS at interim monitoring with time‐to‐event data. Firstly, there are completely observed data, where the event type, the time to the event, and the baseline covariates are known (). Secondly, there are partially observed data, where the baseline covariates are known, but the follow‐up time is right‐censored at the time of interim monitoring (). Thirdly, there are completely unobserved data, which will be observed if new participants are enrolled in the future (). In order to relate these types of data to the PPoS formula, note that is given by stacking and (). The data include , , but also , representing the updated data at the conclusion of the trial for the participants in (). At interim monitoring, and are not fully observed and to calculate the simulation‐based PPoS they need to be predicted from statistical models.
2.2. Predictive Probability of Success With Competing Event Data
The framework to predict the probability of trial success via statistical simulation can be divided into three phases: modelling, prediction, and analysis [19]. The proposed approach specialises the first and second phases to allow the modelling and prediction of competing event data, which are then analysed in the third phase. In this section, we describe each step of the simulation process and summarise it in a compact form in Algorithm 1.
ALGORITHM 1. Simulation‐Based Algorithm to Approximate the Bayesian Predictive Probability of Success (PPoS) in Trials With Competing Event Data.

Note: denotes the true event time; is the competing event indicator; is a vector of baseline covariates; and are the model parameters; indicates the simulated trials; indexes the observations to be predicted at final analysis; is the data observed at interim analysis; denotes the set of participants censored at interim analysis; and represents the set of future participants enrolled in the trial.
2.2.1. Modelling Phase
The goal of the modelling phase is to specify a Bayesian model for the joint distribution of the competing event data, the baseline data, and the parameters [26]:
| (1) |
To facilitate modelling, we decompose into a product of two distributions: one for the competing event data given the baseline data and one for the baseline data. We assume that these two distributions have no shared parameters and write . We also assume that the priors on the parameters of the two distributions are independent, i.e., that . As a result, the joint distribution in Equation (1) can be rewritten as follows:
![]() |
(2) |
We propose to specify the distribution for the competing event data in terms of cause‐specific hazards for all event types [27]:
where is the event‐free survival function. Assuming an independent right‐censoring mechanism, the contribution to likelihood for the ‐th observation is proportional to [25]:
Patients censored at the interim analysis () are still at risk of experiencing any of the competing events. They contribute to the likelihood with the event‐free survival function evaluated at their censoring time.
The model for the distribution of the baseline data depends on the specific type of data to be modelled. The contribution to likelihood for the ‐th observation is written generically as:
Lastly, the prior distributions on the model parameters, and , complete the specification of the joint distribution in Equation (2).
The model is then fitted to the interim data and samples are drawn from the posterior distribution:
for example, using Markov Chain Monte Carlo (MCMC) machinery.
2.2.2. Prediction Phase
In this phase, predictions for future observations are made by sampling from the posterior predictive distribution [26]:
As the posterior predictive distribution is not available in closed form, posterior predictive samples can be generated by first decomposing into the product and then employing the following forward sampling algorithm:
Make a draw from the posterior distribution:
-
2
Simulate the baseline covariate data for the future trial participants, if any:
-
3
Simulate the event times:
For future trial participants:
Event times can be simulated by applying the probability integral transformation [28]. First, draw a random value and then generate an event time as . If the event‐free survival function cannot be inverted analytically, root finding is used to numerically solve for .
-
b
For participants censored at the time of interim analysis:
In this case, the event times are simulated conditionally on being still at risk at the observed censoring time and given the observed baseline data . This is done by treating as a left‐truncation time and simulating the event times from the corresponding truncated distribution, where is defined as .
-
4
Simulate the event types:
For future trial participants:
where is a distribution with [29].
-
b
For participants censored at the time of interim analysis:
where the conditioning is on the observed baseline data , instead of the predicted data .
Lastly, generate censoring times (e.g., at the maximum allowed follow‐up time) and compute and .
Note that if no further participants are to be recruited after interim monitoring, the Prediction phase is simplified because sampling from is unnecessary. The Modelling phase is also simplified, as modelling the distribution is sufficient.
2.2.3. Analysis Phase
The data completely observed at the time of interim monitoring and the data generated in the Prediction phase are stacked into and analysed using the prespecified frequentist or Bayesian analytical method, , which maps to the space where the critical region is defined. The binary variable indicating whether the simulated trial reaches success at final analysis is defined as:
To predict the probability of trial success, the Prediction and Analysis phases are repeated a sufficiently large number of times, and the value is recorded for each iteration . The proportion of simulated trials that achieve success approximates the PPoS [19]:
2.3. Software
All the analyses were performed in R version 4.3.2 [30]. Bayesian models were fitted to the data in Stan version 2.32.6 using full Bayesian statistical inference with MCMC sampling [31]. The brms package version 2.21.0 was used as the interface to Stan [32]. For Bayesian models with Weibull hazard, we used a custom response distribution available at https://github.com/anddis/brms‐weibullPH. The non‐parametric Aalen–Johansen estimates of the cumulative incidence functions were obtained using the survival package version 3.7–0 [33]. We provide the R code for reproducing the analyses on a synthetic dataset at https://github.com/cmicoli/PPoS‐CompetingRisks.
3. Motivational Example 1: The I‐SPY COVID Trial
3.1. Trial Description and Motivation
The I‐SPY COVID (NCT04488081) is a multicentre phase‐II open‐label adaptive platform trial aimed at identifying potential therapeutics for severe COVID‐19 [23] (Table A1). Patients with a positive SARS‐CoV‐2 test by PCR or rapid antigen testing on ≥ 6 L/min oxygen or intubated were screened for eligibility. Patients meeting the eligibility criteria were randomly assigned to either a common control arm, consisting of a backbone regimen of dexamethasone and remdesivir, or to backbone plus one of up to 4 investigational agents in the trial at any one time. Patients who provided informed consent were included in the modified intention‐to‐treat (mITT) population. The maximum number of patients who could be randomised to each investigational arm was 125.
The primary endpoints of the trial were time to durable recovery and time to death due to any cause. Time to recovery was defined as the time elapsed between randomisation and the earliest second consecutive day with WHO COVID level 4 (i.e., < 6 L/min oxygen) or below. In the analyses, follow‐up was censored at 60 days if patients were still at risk. An agent would graduate if the vector consisting of the posterior probability of the cause‐specific hazard ratio (csHR) for recovery being and the posterior probability of the HR for all‐cause mortality being was in the 2‐dimensional critical region . Separate Bayesian Weibull models with weakly informative priors were used to model the cause‐specific hazard for recovery and the hazard for all‐cause mortality in the mITT population and to derive the posterior probabilities of graduation. These posterior probabilities were updated every 2 weeks and reviewed by an independent data monitoring committee (DMC). The primary role of the DMC was to ensure the safety of the patients in the trial, but also to recommend to the principal investigators that an agent should graduate or be dropped for futility, based on unblinded review of trial results. A detailed description of the study design, operating characteristics, and statistical analysis plan of the trial can be found elsewhere [23, 34, 35]. The analysis presented here replicates that performed at one of the interim reviews, with minor differences in statistical software and data, and it focuses solely on the time‐to‐recovery endpoint.
At the interim analysis approximately 15 weeks after the first patient was randomised to the backbone plus cyclosporine arm (2021‐07‐19), there were 58 patients in the cyclosporine arm and 75 in the control arm. A total of 34 patients in the cyclosporine arm and 47 in the control arm had recovered, while the corresponding numbers of patients who died before recovery were 7 and 12, respectively. The posterior median for cyclosporine versus control was 0.97 (95% quantile credible interval: 0.63, 1.47), indicating that the instantaneous recovery rate was similar in the two arms. The corresponding posterior probability of graduation was 0.44, outside the critical region . Given that the interim data on the recovery endpoint was not in favour of cyclosporine after almost 50% of the maximum allowed number of patients had been randomised to cyclosporine, the DMC asked to predict the probability that cyclosporine would demonstrate sufficient benefit to meet the graduation criterion at the maximum sample size. The PPoS analysis was not prespecified in the trial protocol.
3.2. Results
Let the binary variable be the baseline WHO COVID level of 6 or 7 () versus 5 () and the binary variable be the cyclosporine () versus control arm (). Due to randomisation, we assumed to be independent of . In the modelling phase, we specified a Bayesian model for the joint distribution of the competing event data, WHO COVID level, randomisation arm and the parameters, . The model was specified in terms of the cause‐specific hazards for the two competing events, recovery from COVID () and death from any cause (). In particular, we specified Weibull models stratified by randomisation arm so as not to impose any proportionality constraint between the cause‐specific hazards for the cyclosporine and the control arm: . We let the scale parameter depend on baseline WHO COVID level, such that , while the shape parameter did not depend on any covariates. The following priors completed the model specification: , , . The baseline WHO COVID level followed a distribution with a noninformative prior . Lastly, we fitted the model to the interim data .
In the prediction phase, we fixed the number of new patients in the cyclosporine arm to , and given that the randomisation probability to the cyclosporine arm was 0.45, we sampled the number of patients in the control arm from a distribution. For these new patients, we simulated the baseline WHO COVID level, time to event, and event type (). For the 33 patients still at risk at the interim analysis (), the outcome data were simulated conditional on the observed baseline WHO COVID level, randomisation arm, and their censoring times. Predicted follow‐up times were censored if they exceeded 60 days. The interim and predicted data were stacked in the dataset .
In the analysis phase, we analysed interim and predicted data together using Bayesian proportional hazard models and then derived the posterior probability over 4000 draws, as prespecified in the SAP. We repeated the prediction and analysis phase times and approximated the PPoS as
indicating that Cyclosporine would not likely have shown a benefit on the COVID recovery rate even at maximum sample size. Combining the information from the prespecified posterior probability analysis and the PPoS, the DMC recommended stopping randomisation of patients to the cyclosporine arm.
As the PPoS was low, we assessed in a sensitivity analysis how more enthusiastic priors about the for cyclosporine versus control for the remainder of the trial would affect the PPoS [36]. The rationale for this analysis was to reassure the DMC that their decision to stop the cyclosporine arm due to futility would be made even under optimistic assumptions about the effect of cyclosporine on the recovery rate in future patients. To this end, we specified the model so that the cause‐specific hazards for the two randomisation arms were proportional with a proportionality constant of . In particular, we defined , . The priors for , , and were defined as in the main PPoS analysis, while . We let the prior take on different values for the mean and standard deviation, covering different scenarios ranging from weakly neutral to strongly optimistic priors. We explored 15 different scenarios by setting and . Some of these priors were extremely optimistic, for example the equal‐tailed 95% prior limits for the under the ‐prior were . These values were considerably larger than what was considered plausible at the time of the trial design, where the largest considered in the sample size calculations was 1.80.
Figure 1 shows the distribution of the posterior probability over 2500 simulations for the 15 different scenarios. The PPoS varied from 0.034 to 0.498, indicating that even an unrealistically optimistic and strong prior for resulted in a relatively low PPoS. This sensitivity analysis confirms that cyclosporine is unlikely to reach the graduation threshold for recovery at maximum sample size even under an extremely optimistic scenario.
FIGURE 1.

Distributions of the posterior probability of graduation over 2500 simulations for the 15 different scenarios considered in the sensitivity analysis for the I‐SPY COVID trial. The red bar corresponds to the PPoS, i.e., to the proportion of simulations for which . Each panel corresponds to a different prior specified in the modelling phase. The values specified for the prior's parameters and are reported along the rows and columns, respectively.
4. Motivational Example 2: Long‐Term Follow‐Up of the STHLM3 Trial
4.1. Trial Description and Motivation
The population‐based diagnostic STHLM3 trial (ISRCTN84445406) was conducted between May 2012 and January 2015. A random sample of men aged 50–69 years living in Stockholm, Sweden, was invited to participate in a one‐time prostate cancer screening intervention using prostate‐specific antigen (PSA) and the Stockholm3 test, a prediction model based on clinical variables, plasma protein biomarkers, and a polygenic risk score [24]. One of the secondary aims of the STHLM3 trial was to evaluate the efficacy of screening invitation on long‐term prostate cancer mortality (PCM) in the presence of death from other causes as a competing event, which is the focus of the analysis reported here. The population eligible for random invitation to participate in the screening intervention consisted of 240,494 men. Of these, 169,621 (70%) were randomly invited and 59,088 (35%) participated.
An interim analysis was performed using follow‐up data until the end of 2018 to compare the crude cumulative risk of PCM between the two groups at 6.5 years after invitation [37]. Crude cumulative risks were estimated using the Aalen‐Johansen estimator and the difference in PCM risk was summarised by a risk ratio (RR) with the non‐invited group as the reference. A p‐value for the two‐sided alternative hypothesis versus the null was calculated as , where is the asymptotic standard error for the and is the standard normal cumulative distribution function. We considered a two‐sided alternative hypothesis because we could not rule out a priori that screening invitation might increase PCM risk. At interim analysis, the significance level was set to 0.02. During a maximum of 6.6 years of follow‐up, we observed 101 prostate cancer deaths in the invited group and 54 in the non‐invited group, while the estimated PCM risks were 8.6 and 13.2 per 10,000 men, respectively (Table 1). The for PCM was 0.66 (98% confidence interval (CI): 0.41, 1.06), , which favoured the invited group but did not reach formal statistical significance.
TABLE 1.
Interim analysis results of the STHLM3 trial. Number of randomised men (N), deaths from prostate cancer (PCa), deaths from other causes and corresponding mortality risks for the invited and non‐invited randomisation groups.
| Randomisation group | N | Deaths from PCa | Deaths from other causes | PCa mortality risk | Other causes mortality risk |
|---|---|---|---|---|---|
| Invited | 169,621 | 101 | 6644 | 8.6 per 10,000 men | 525 per 10,000 men |
| Not invited | 70,873 | 54 | 2814 | 13.2 per 10,000 men | 536 per 10,000 men |
Note: Risks at the interim analysis were estimated using the Aalen–Johansen estimator treating the other event as a competing event.
The time point for the final comparison of PCM risk between the two invitation arms was prespecified to be between 10 and 15 years after study initiation. The PPoS was used to guide the selection of the specific time point for the final analysis (). We first predicted the probability of finding a statistically significant RR after 10 years given the interim data and then we repeated the prediction process five other times, adding 1 year at a time until a maximum of 15 years, thus allowing for more prostate cancer deaths to accumulate. The alpha level for the final analysis used in the PPoS estimation was set at 0.035 (critical region ) so that the overall type I error rate, considering both interim and final analyses, would be maintained at 0.05 [38]. A formal threshold for the PPoS was not prespecified in the original protocol and the decision on when to perform the final analysis was agreed upon at a consensus meeting that included statisticians and principal investigators of the STHLM3 trial.
4.2. Results
In the modelling phase, we specified a Bayesian model for , where the numeric variable represents age at baseline (years), and the binary variable represents the randomisation group: invited () versus non‐invited (). As enrolment in the trial was completed, modelling of baseline covariates was not necessary. To model , we assumed piecewise constant cause‐specific hazards (PCH) for the two competing events, prostate cancer death () and death from other causes (). We stratified the model by invitation group and included age at baseline as a covariate. The cause‐specific hazards were piecewise constant on time intervals defined by the boundary knots and by the internal knots :
where [39]. For the cause‐specific hazard for PCM, we chose years, while for the cause‐specific hazard for death from other causes we chose years, giving and time intervals, respectively. The interim data contained more information on other cause mortality than for PCM (Table 1), allowing a finer grid for the internal knots. We completed the models by specifying a lag‐1–autoregressive prior on the parameters for the baseline hazard () and a weakly informative prior on the parameter for baseline age () (Modelling Strategy A in Table 2). We then fitted the model to the interim data .
TABLE 2.
Modelling strategies used in the modelling stage of the STHLM3 trial PPoS analysis.
| Modelling strategy | Specification | Priors and hyperpriors | PPoS at 10.6 years | PPoS at 11.6 years | PPoS at 12.6 years | PPoS at 13.6 years | PPoS at 14.6 years | PPoS at 15.6 years | |
|---|---|---|---|---|---|---|---|---|---|
| A | PCH model |
|
0.783 | 0.826 | 0.856 | 0.864 | 0.885 | 0.892 | |
| , | for | ||||||||
|
|
|
||||||||
|
|
|
||||||||
| B | PCH model |
|
0.758 | 0.796 | 0.832 | 0.868 | 0.877 | 0.892 | |
|
|
for | ||||||||
|
|
|
||||||||
|
|
|||||||||
| C | PCH model |
|
0.783 | 0.825 | 0.850 | 0.863 | 0.884 | 0.894 | |
|
|
for | ||||||||
|
|
|
||||||||
|
|
|||||||||
| D | Weibull model |
|
0.913 | 0.946 | 0.958 | 0.967 | 0.976 | 0.980 | |
| Scale parameter: |
|
||||||||
| Shape parameter: |
|
Note: PPoS at different time points for the final analysis are reported. Subscript is the event type: death due to prostate cancer, death due to other causes. Subscript identifies the randomisation group: not invited, invited.
Abbreviation: PCH, piecewise constant cause‐specific hazard.
In the prediction phase, we simulated the time to event and event type for those men who were still alive at the end of their follow‐up period in the interim data (), conditioning the predicted event time to be greater than the censoring time. To mimic the fact that the outcome data provided yearly by national registries cover the entire calendar year, the final analysis after 10 years will use data until the end of 2022. To this end, the predicted event times were censored at subject‐specific times given by the difference between 2022‐12‐31 and each subject's invitation date (i.e., time 0), giving a maximum follow‐up of 10.6 years. The same approach was applied for predictions at later time points. As enrolment in the STHLM3 trial is complete, the predicted data used in the final analysis consisted only of and .
For each final analysis time point, we repeated times the prediction and analysis phase. At each iteration, we estimated the cumulative PCM risks in the two invitation groups (Figure 2) and derived the RR together with the associated p‐value at the time of final analysis, corresponding to the largest follow‐up time ( years). The PPoS was approximated as:
![]() |
which ranged from 0.783 for the final analysis at 10.6 years of follow‐up to 0.892 for the final analysis at 15.6 years (Table 2, Modelling strategy A).
FIGURE 2.

Red and blue thick step lines are the non‐parametric crude cumulative risks of death from prostate cancer and of death due to other causes estimated on interim data () for men invited and not invited to participate in the STHLM3 trial, respectively. Grey step lines are a random sample of 500 out of the 2500 non‐parametric crude cumulative risk estimates based on interim and predicted data together (). Predicted data was simulated using Bayesian piecewise constant cause‐specific hazard models with 5 and 12 time intervals for death from prostate cancer and death from other causes, respectively (Modelling strategy A).
We performed a sensitivity analysis to assess the extent to which the choices made in the modelling phase could affect the PPoS. In particular, we first explored the consequences of changing the number and location of the internal knots of the PCH models, and then used Weibull models for the cause‐specific hazards. The specification of the alternative modelling strategies (B–D) is shown in Table 2 together with the corresponding PPoS, which ranged from 0.758 to 0.913 for the earliest time point for the final analysis (10.6 years), and was greater than 0.80 for later time points, regardless of the modelling strategy (Table 2). Since the PPoS was considered sufficiently high and robust across all modelling strategies for all time points, the statisticians and principal investigators of the STHLM3 trial decided that the PCM risk in the two groups would be compared at 11.6 years after invitation, i.e., with follow‐up data until the end of 2023.
5. Discussion
In this paper, we have described a simulation‐based approach to compute the PPoS for interim monitoring of clinical trials when the survival outcome is subject to competing events. Our work was motivated by two randomised clinical trials, both of which had a primary survival outcome with death as a competing event. In the I‐SPY COVID trial, the PPoS analysis complemented the prespecified efficacy analysis, which led the DMC to recommend stopping the randomisation in the backbone‐plus‐cyclosporine arm due to futility. In the STHLM3 trial, the PPoS analysis was used to inform the decision to perform the final analysis using follow‐up data covering until the end of 2023.
In the modelling phase, we have proposed the use of cause‐specific hazard models to model the joint distribution of time to event and event type given covariates. However, alternative approaches are possible and include multivariate latent failure time models [40], vertical models [41], and direct modelling of the cumulative incidence functions [42]. We have chosen cause‐specific hazard models because they provide a convenient and flexible modelling framework [43]. In addition, as the cause‐specific hazards allow the estimation of the joint distribution based on observable quantities alone, unverifiable assumptions about the dependence structure of the latent event times can be avoided. Different models can be specified to explore how sensitive the PPoS is to modelling assumptions, as we did in Section 4. However, in general, we recommend stratifying the cause‐specific hazard models over the randomisation groups, as done in previous applications [7, 14]. The use of baseline covariates other than randomisation arm can in principle improve the prediction of the outcome data [44]. This comes at the cost of an increased complexity due to the need to model their joint distribution, if new trial participants' baseline data are to be simulated. A simpler approach that however ignores some variability in the data generating process is to sample with replacement the observed baseline covariate data, i.e., from their empirical joint distribution.
The specification of the prior distributions represents a common concern with any approach based on Bayesian methodology, and the one proposed here is no exception. Weakly informative priors can be specified so that the predictions can be considered to be based on interim data only [2]. Informative priors may be otherwise specified to incorporate external information about the data‐generating process, including the treatment effect [45]. For PPoS sensitivity analyses, informative priors may be used to anticipate or explore the consequences of changes in the population or in the treatment effect over time [46], or to counterbalance the prior opinions of someone who would doubt the observed results [36], as we did in Section 3. Ultimately, the choice of the priors should be aligned with the objectives of the analysis. If necessary, the choice of the prior distributions can be discussed and agreed upon with the regulators at the time of trial design. Finally, it should be noted that the priors used in the modelling phase do not need to be the same as those used in the analysis phase.
As the proposed approach is simulation‐based, it can be computationally intensive, especially as the complexity of the prediction and analysis phase increases. Depending on the specific Bayesian models used in the modelling phase, predicting survival times can be a time‐consuming task. This is the case when the event‐free survival function cannot be inverted analytically or when it has to be evaluated by numerical integration [28]. The analysis phase can also be computationally demanding, especially when Bayesian methods are used. For reference, computing the PPoS for the I‐SPY COVID example took approximately 9 min on a high‐performance computing cluster using 40 cores, while for the larger STHLM3 trial it took over 10 h. The computational burden increases at the design stage, when the trial's operating characteristics are usually assessed for a possibly large number of scenarios. The number of simulations can be guided not only by time and computational constraints, but also by the accuracy with which the PPoS needs be approximated [47]. For example, our choice of simulations was motivated by setting the upper bound on the standard error of the PPoS to 0.01. On the other hand, our approach offers full flexibility in the modelling and analysis phase and does not require resorting to asymptotic approximations [15] or to the use of conjugate priors [7]. In particular, the proposed framework can be applied to both frequentist and Bayesian trials, and any analytical method for competing event data—whether frequentist or Bayesian—can be used as the basis to define trial success in the analysis phase.
The use of Bayesian PPoS for interim monitoring of clinical trials is appealing because it is easily interpretable by trial stakeholders, including physicians, trial participants, and DMC board members. Furthermore, it directly addresses a relevant question: how likely is it that the trial will achieve its objective, given the current data, if it continues to its end? Importantly, this predictive question is addressed taking into account two sources of variability: the variability in the data not yet observed at interim monitoring and the uncertainty in the parameter estimates for the data‐generating process. At the same time, the calculation of the PPoS involves a number of assumptions that may lead to an inaccurate prediction of the success probability if they do not hold true [44, 46]. As we have done in our motivational examples, we recommend performing sensitivity analyses to evaluate how different assumptions may affect the PPoS [45]. The thresholds for the stopping rules that will trigger early termination of the trial are often specified at the design stage, including those based on the PPoS. This allows the trial's frequentist operating characteristics to be evaluated, for example using a simulation study, and the thresholds to be selected according to the desired long‐run properties of the resulting design [9]. Bayesian operating characteristics can also be used to select thresholds for early stopping in an analogous fashion [48].
In conclusion, our work extends the existing literature on the PPoS by proposing a simulation‐based approach that allows the PPoS to be used for interim monitoring of clinical trials in the common setting where the trial outcome is subject to competing events.
Conflicts of Interest
M.E. reports four pending prostate cancer diagnostic‐related patents: method for indicating a presence or non‐presence of aggressive prostate cancer (WO2013EP7425920131120); prognostic method for individuals with prostate cancer (WO2013EP7427020131120); method for indicating a presence of prostate cancer in individuals with particular characteristics (WO2018EP5247320180201); and method for indicating the presence or non‐presence of prostate cancer (WO2013SE5055420130516). The Karolinska Institutet collaborates with A3P Biomedical in developing the technology for the Stockholm3 test. M.E. owns shares in A3P Biomedical. All other authors declare no conflicts of interest.
Acknowledgements
I‐SPY COVID Funding: The I‐SPY COVID trial was funded by COVID R&D Consortium, Allergan, Amgen Inc., Takeda Pharmaceutical Company, Ingenus Pharmaceuticals LLC, Implicit Bioscience, Johnson Johnson, Pfizer Inc., Roche/Genentech, Apotex Inc., FAST Grant from Emergent Venture George Mason University, The DoD Defense Threat Reduction Agency (DTRA), The Department of Health and Human Services Biomedical Advanced Research and Development Authority (BARDA), and The Grove Foundation.
STHLM3 Funding: The STHLM3 trial was funded by the Swedish Cancer Society (Cancerfonden), the Swedish Research Council (Vetenskapsrådet), Region Stockholm, and Svenska Druidorden, Åke Wibergs Stiftelse, the Swedish e‐Science Research Center, the Karolinska Institutet, and Prostatacancerförbundet.
Appendix A.
COVID Executive Committee
Laura J. Esserman; Carolyn S. Calfee; Michael A. Matthay; Kathleen D. Liu; D. Clark Files; Martin Eklund; Karl W. Thomas.
ARDS Executive Committee
Laura J. Esserman; Kathleen D. Liu; D. Clark Files; Kevin Gibbs; Derek W. Russell; Martin Eklund; Karl W. Thomas.
TABLE A1.
List of sites, principal investigators, and affiliations of the I‐SPY COVID consortium.
| Site | Principal investigators (PI) | Affiliation |
|---|---|---|
| Columbia NYC | Jeremy R. Beitler, MD, MPH | Centre for Acute Respiratory Failure, Columbia University |
| Emory | Sara C. Auld, MD | Department of Medicine, Emory University |
| Georgetown | Nathan Cobb, MD | Georgetown University Medical Centre |
| Hoag (Irvine) | Philip Robinson, MD | Hoag Memorial Hospital Presbyterian Centre for Research and Education |
| Hoag (Newport Beach) | Philip Robinson, MD | Hoag Memorial Hospital Presbyterian Centre for Research and Education |
| KP LAMC | Kenneth K. Wei, MD | Division of Pulmonary and Critical Care Medicine, Kaiser Permanente Los Angeles Medical Center |
| Logan Health | Timothy Obermiller, MD | Critical Care Medicine, Logan Health Research Institute |
| Long Beach Memorial | Fady A. Youssef, MD | Department of Internal Medicine, Pulmonary Division, Long Beach Memorial Medical Centre |
| Main Line Health | Eliot Friedman, MD | Department of Medicine, Main Line Health |
| Mercy Hospital | Sandya Samavedam, MD | Critical Care Medicine, Mercy Hospital |
| Northwestern | Richard G. Wunderink, MD | Department of Medicine, Pulmonary and Critical Care Division, Northwestern University Feinberg School of Medicine |
| Sanford Health | Paul Berger, DO | Department of Critical Care, Sanford Health |
| Spectrum Health | Malik M. Khan, MD | Pulmonary and Critical Care, Spectrum Health |
| Stamford Health | Michael Bernstein, MD | Pulmonary Medicine, Stamford Health |
| U Penn | Nuala J. Meyer, MD | Department of Medicine, University of Pennsylvania Perelman School of Medicine |
| UAB | Derek W. Russell, MD | Pulmonary, Allergy, and Critical Care Medicine, University of Alabama, Birmingham |
| Sheetal Gandotra, MD | Pulmonary, Allergy, and Critical Care Medicine, University of Alabama, Birmingham | |
| UC Davis | Timothy Albertson, MD | Division of Pulmonary, Critical Care and Sleep Medicine, University of California Davis |
| UC Irvine | Richard A. Lee, MD | Division of Pulmonary Diseases and Critical Care Medicine, University of California |
| UCSF | Kathleen D. Liu, MD, PhD | Divisions of Nephrology and Critical Care Medicine, University of California San Francisco |
| Laura J. Esserman, MD | Department of Surgery | |
| Carolyn S. Calfee, MD | Division of Pulmonary, Department of Medicine, Critical Care, Allergy and Sleep Medicine | |
| Michael A. Matthay, MD | Division of Pulmonary, Department of Medicine, Critical Care, Allergy and Sleep Medicine | |
| University of Colorado | Ellen L. Burnham, MD | Department of Medicine, University of Colorado |
| University of Miami | Christopher P. Jordan, MD | Department of Pulmonary Care and Sleep Medicine |
| Daniel H. Kett, MD | Department: Medicine, Division: Pulmonary, Critical Care and Sleep Medicine | |
| University of Michigan | Robert Hyzy, MD | Division of Pulmonary and Critical Care Medicine, Department of Medicine, University of Michigan |
| USC | Santhi Kumar, MD | Division of Pulmonary, Critical Care and Sleep Medicine, Keck School of Medicine, USC |
| Wake Forest | D. Clark Files, MD | Department of Internal Medicine, Wake Forest University |
| Kevin W. Gibbs, MD | Pulmonary Critical Care, Allergy and Immunologic Disease | |
| West Virginia University | Jeremiah Hayanga, MD | Department of Cardiovascular and Thoracic Surgery, Heart and Vascular Institute, West Virginia University |
| Yale | Jonathan L. Koff, MD | Section of Pulmonary, Critical Care, and Sleep Medicine, Yale University |
Micoli C., Crippa A., Connor J. T., I‐SPY COVID Consortium , Eklund M., and Discacciati A., “Simulation‐Based Bayesian Predictive Probability of Success for Interim Monitoring of Clinical Trials With Competing Event Data: Two Case Studies,” Pharmaceutical Statistics 25, no. 1 (2026): e70050, 10.1002/pst.70050.
Funding: This work was supported by the Cancerfonden, the Vetenskapsrådet, and the Prostatacancerförbundet.
Martin Eklund and Andrea Discacciati contributed equally to the work.
Data Availability Statement
The I‐SPY COVID trial data reported in this manuscript are maintained by the study sponsor, Quantum Leap Healthcare Collaborative (QLHC). Data can be shared with investigators external to the trial following approval by QLHC and the ISPY COVID Data Access and Publications Committee. This request can be initiated by contacting QLHC. The STHLM3 trial data reported in this manuscript are available from the corresponding author upon request after ethical approvals have been obtained from the Swedish Ethical Review Board.
References
- 1. Jennison C. and Turnbull B. W., “Statistical Approaches to Interim Monitoring of Medical Trials: A Review and Commentary,” Statistical Science 5, no. 3 (1990): 299–317, 10.1214/ss/1177012099. [DOI] [Google Scholar]
- 2. Spiegelhalter D. J., Freedman L. S., and Blackburn P. R., “Monitoring Clinical Trials: Conditional or Predictive Power?,” Controlled Clinical Trials 7, no. 1 (1986): 8–17, 10.1016/0197-2456(86)90003-6. [DOI] [PubMed] [Google Scholar]
- 3. Saville B. R., Connor J. T., Ayers G. D., and Alvarez J., “The Utility of Bayesian Predictive Probabilities for Interim Monitoring of Clinical Trials,” Clinical Trials 11, no. 4 (2014): 485–493, 10.1177/1740774514531352. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Connor J. T., Elm J. J., Broglio K. R., and ESETT and ADAPT‐IT Investigators , “Bayesian Adaptive Trials Offer Advantages in Comparative Effectiveness Trials: An Example in Status Epilepticus,” Journal of Clinical Epidemiology 66, no. 8 Suppl (2013): S130–S137, 10.1016/j.jclinepi.2013.02.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. Sevransky J. E., Rothman R. E., Hager D. N., et al., “Effect of Vitamin C, Thiamine, and Hydrocortisone on Ventilator‐ and Vasopressor‐Free Days in Patients With Sepsis: The VICTAS Randomized Clinical Trial,” JAMA 325, no. 8 (2021): 742–750, 10.1001/jama.2020.24505. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. White W. B., Grady D., Giudice L. C., Berry S. M., Zborowski J., and Snabes M. C., “A Cardiovascular Safety Study of LibiGel (Testosterone Gel) in Postmenopausal Women With Elevated Cardiovascular Risk and Hypoactive Sexual Desire Disorder,” American Heart Journal 163, no. 1 (2012): 27–32, 10.1016/j.ahj.2011.09.021. [DOI] [PubMed] [Google Scholar]
- 7. Broglio K. R., Connor J. T., and Berry S. M., “Not Too Big, Not Too Small: A Goldilocks Approach to Sample Size Selection,” Journal of Biopharmaceutical Statistics 24, no. 3 (2014): 685–705, 10.1080/10543406.2014.888569. [DOI] [PubMed] [Google Scholar]
- 8. Choi S. C., Smith P. J., and Becker D. P., “Early Decision in Clinical Trials When the Treatment Differences Are Small. Experience of a Controlled Trial in Head Trauma,” Controlled Clinical Trials 6, no. 4 (1985): 280–288, 10.1016/0197-2456(85)90104-7. [DOI] [PubMed] [Google Scholar]
- 9. Park J. W., Liu M. C., Yee D., et al., “Adaptive Randomization of Neratinib in Early Breast Cancer,” New England Journal of Medicine 375, no. 1 (2016): 11–22, 10.1056/NEJMoa1513750. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. McNamee J. J., Gillies M. A., Barrett N. A., et al., “Effect of Lower Tidal Volume Ventilation Facilitated by Extracorporeal Carbon Dioxide Removal vs Standard Care Ventilation on 90‐Day Mortality in Patients With Acute Hypoxemic Respiratory Failure: The REST Randomized Clinical Trial,” JAMA 326, no. 11 (2021): 1013–1023, 10.1001/jama.2021.13374. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Wilber D. J., Pappone C., Neuzil P., et al., “Comparison of Antiarrhythmic Drug Therapy and Radiofrequency Catheter Ablation in Patients With Paroxysmal Atrial Fibrillation: A Randomized Controlled Trial,” JAMA 303, no. 4 (2010): 333–340, 10.1001/jama.2009.2029. [DOI] [PubMed] [Google Scholar]
- 12. Alexander B. M., Trippa L., Gaffey S., et al., “Individualized Screening Trial of Innovative Glioblastoma Therapy (INSIGhT): A Bayesian Adaptive Platform Trial to Develop Precision Medicines for Patients With Glioblastoma,” JCO Precision Oncology 3 (2019): 1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Wang Y., Fu H., Kulkarni P., and Kaiser C., “Evaluating and Utilizing Probability of Study Success in Clinical Development,” Clinical Trials 10, no. 3 (2013): 407–413, 10.1177/1740774513478229. [DOI] [PubMed] [Google Scholar]
- 14. Berry S. M., Carlin B. P., Lee J. J., and Muller P., Bayesian Adaptive Methods for Clinical Trials (CRC Press, 2010). [Google Scholar]
- 15. Marion J., Lorenzi E., Allen‐Savietta C., Berry S., and Viele K., “Predictive Probabilities Made Simple: A Fast and Accurate Method for Clinical Trial Decision‐Making,” Statistics in Medicine 44, no. 13–14 (2025): e70120, 10.1002/sim.70120. [DOI] [PubMed] [Google Scholar]
- 16. Tang Z., “Optimal Futility Interim Design: A Predictive Probability of Success Approach With Time‐To‐Event Endpoint,” Journal of Biopharmaceutical Statistics 25, no. 6 (2015): 1312–1319, 10.1080/10543406.2014.983646. [DOI] [PubMed] [Google Scholar]
- 17. Yin G., Chen N., and Lee J. J., “Bayesian Adaptive Randomization and Trial Monitoring With Predictive Probability for Time‐To‐Event Endpoint,” Statistics in Biosciences 10, no. 2 (2018): 420–438, 10.1007/s12561-017-9199-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Schmidli H., Bretz F., and Racine‐Poon A., “Bayesian Predictive Power for Interim Adaptation in Seamless Phase II/III Trials where the Endpoint is Survival up to Some Specified Timepoint,” Statistics in Medicine 26, no. 27 (2007): 4925–4938, 10.1002/sim.2957. [DOI] [PubMed] [Google Scholar]
- 19. Trzaskoma B. and Sashegyi A., “Predictive Probability of Success and the Assessment of Futility in Large Outcomes Trials,” Journal of Biopharmaceutical Statistics 17, no. 1 (2007): 45–63, 10.1080/10543400601001485. [DOI] [PubMed] [Google Scholar]
- 20. Austin P. C., Ibrahim M., and Putter H., “Accounting for Competing Risks in Clinical Research,” JAMA 331, no. 24 (2024): 2125–2126, 10.1001/jama.2024.4970. [DOI] [PubMed] [Google Scholar]
- 21. Dodd L. E., Follmann D., Wang J., et al., “Endpoints for Randomized Controlled Clinical Trials for COVID‐19 Treatments,” Clinical Trials 17, no. 5 (2020): 472–482, 10.1177/1740774520939938. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Yin G., Chen N., and Lee J. J., “Phase II Trial Design With Bayesian Adaptive Randomization and Predictive Probability,” Journal of the Royal Statistical Society. Series C, Applied Statistics 61, no. 2 (2012): 219–235, 10.1111/j.1467-9876.2011.01006.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Files D. C., Matthay M. A., Calfee C. S., et al., “I‐SPY COVID Adaptive Platform Trial for COVID‐19 Acute Respiratory Failure: Rationale, Design and Operations,” BMJ Open 12, no. 6 (2022): e060664, 10.1136/bmjopen-2021-060664. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Grönberg H., Adolfsson J., Aly M., et al., “Prostate Cancer Screening in Men Aged 50‐69 Years (STHLM3): A Prospective Population‐Based Diagnostic Study,” Lancet Oncology 16, no. 16 (2015): 1667–1676, 10.1016/S1470-2045(15)00361-7. [DOI] [PubMed] [Google Scholar]
- 25. Andersen P. K., Abildstrom S. Z., and Rosthøj S., “Competing Risks as a Multi‐State Model,” Statistical Methods in Medical Research 11, no. 2 (2002): 203–215, 10.1191/0962280202sm281ra. [DOI] [PubMed] [Google Scholar]
- 26. Gelman A., Carlin J. B., Stern H. S., Dunson D. B., Vehtari A., and Rubin D. B., Bayesian Data Analysis, 3rd ed. (Chapman & Hall/CRC, 2013). [Google Scholar]
- 27. Andrillon A., Chevret S., Lee S. M., and Biard L., “Surv‐CRM‐12: A Bayesian Phase I/II Survival CRM for Right‐Censored Toxicity Endpoints With Competing Disease Progression,” Statistics in Medicine 41, no. 29 (2022): 5753–5766, 10.1002/sim.9591. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Crowther M. J. and Lambert P. C., “Simulating Biologically Plausible Complex Survival Data,” Statistics in Medicine 32, no. 23 (2013): 4118–4134, 10.1002/sim.5823. [DOI] [PubMed] [Google Scholar]
- 29. Beyersmann J., Latouche A., Buchholz A., and Schumacher M., “Simulating Competing Risks Data in Survival Analysis,” Statistics in Medicine 28, no. 6 (2009): 956–971, 10.1002/sim.3516. [DOI] [PubMed] [Google Scholar]
- 30. R Core Team , R: A Language and Environment for Statistical Computing (R Foundation for Statistical Computing, 2023). [Google Scholar]
- 31. Carpenter B., Gelman A., Hoffman M. D., et al., “Stan: A Probabilistic Programming Language,” Journal of Statistical Software 76 (2017): 1–32, 10.18637/jss.v076.i01. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32. Bürkner P. C., “brms: An R Package for Bayesian Multilevel Models Using Stan,” Journal of Statistical Software 80 (2017): 1–28, 10.18637/jss.v080.i01. [DOI] [Google Scholar]
- 33. Therneau T. M., A Package for Survival Analysis in R, Package Version 3.5‐8., 2024.
- 34. Files D. C., Aggarwal N., Albertson T., et al., “Report of the First Seven Agents in the I‐SPY COVID Trial: A Phase 2, Open Label, Adaptive Platform Randomised Controlled Trial,” Eclinicalmedicine 58 (2023): 101889, 10.1016/j.eclinm.2023.101889. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Calfee C. S., Liu K. D., Asare A. L., et al., “Clinical Trial Design During and Beyond the Pandemic: The I‐SPY COVID Trial,” Nature Medicine 28, no. 1 (2022): 9–11, 10.1038/s41591-021-01617-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36. Spiegelhalter D. J., Freedman L. S., and Parmar M. K. B., “Bayesian Approaches to Randomized Trials,” Journal of the Royal Statistical Society. Series A, Statistics in Society 157, no. 3 (1994): 357–416, 10.2307/2983527. [DOI] [Google Scholar]
- 37. Micoli C., Crippa A., Discacciati A., et al., “A0890: The Stockholm3 Prostate Cancer Screening Trial (STHLM3): An Interim Analysis of Mortality Results After 6.5 Years of Follow‐Up,” European Urology 83 (2023): S1266–S1267, 10.1016/S0302-2838(23)00930-2. [DOI] [Google Scholar]
- 38. DeMets D. L. and Lan K. K., “Interim Analysis: The Alpha Spending Function Approach,” Statistics in Medicine 13, no. 13–14 (1994): 1341–1352, 10.1002/sim.4780131308. [DOI] [PubMed] [Google Scholar]
- 39. Friedman M., “Piecewise Exponential Models for Survival Data With Covariates,” Annals of Statistics 10, no. 1 (1982): 101–113, 10.1214/aos/1176345693. [DOI] [Google Scholar]
- 40. Meng C., Esserman D., Li F., et al., “Simulating Time‐To‐Event Data Subject to Competing Risks and Clustering: A Review and Synthesis,” Statistical Methods in Medical Research 32, no. 2 (2023): 305–333, 10.1177/09622802221136067. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Nicolaie M. A., vHC H., and Putter H., “Vertical Modeling: A Pattern Mixture Approach for Competing Risks Modeling,” Statistics in Medicine 29, no. 11 (2010): 1190–1205, 10.1002/sim.3844. [DOI] [PubMed] [Google Scholar]
- 42. Jeong J. H. and Fine J., “Direct Parametric Inference for the Cumulative Incidence Function,” Journal of the Royal Statistical Society: Series C: Applied Statistics 55, no. 2 (2006): 187–200. [Google Scholar]
- 43. Kipourou D. K., Charvat H., Rachet B., and Belot A., “Estimation of the Adjusted Cause‐Specific Cumulative Probability Using Flexible Regression Models for the Cause‐Specific Hazards,” Statistics in Medicine 38, no. 20 (2019): 3896–3910. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44. Avalos‐Pacheco A., Ventz S., Arfè A., et al., “Validation of Predictive Analyses for Interim Decisions in Clinical Trials,” JCO Precision Oncology 7 (2023): e2200606, 10.1200/PO.22.00606. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45. Rufibach K., Jordan P., and Abt M., “Sequentially Updating the Likelihood of Success of a Phase 3 Pivotal Time‐To‐Event Trial Based on Interim Analyses or External Information,” Journal of Biopharmaceutical Statistics 26, no. 2 (2016): 191–201, 10.1080/10543406.2014.972508. [DOI] [PubMed] [Google Scholar]
- 46. Saville B. R., Detry M. A., and Viele K., “Conditional Power: How Likely Is Trial Success?,” JAMA 329, no. 6 (2023): 508–509, 10.1001/jama.2022.25080. [DOI] [PubMed] [Google Scholar]
- 47. Koehler E., Brown E., and Haneuse S. J. P. A., “On the Assessment of Monte Carlo Error in Simulation‐Based Statistical Analyses,” American Statistician 63, no. 2 (2009): 155–162, 10.1198/tast.2009.0030. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48. Emerson S. S., Kittelson J. M., and Gillen D. L., “Bayesian Evaluation of Group Sequential Clinical Trial Designs,” Statistics in Medicine 26, no. 7 (2007): 1431–1449. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The I‐SPY COVID trial data reported in this manuscript are maintained by the study sponsor, Quantum Leap Healthcare Collaborative (QLHC). Data can be shared with investigators external to the trial following approval by QLHC and the ISPY COVID Data Access and Publications Committee. This request can be initiated by contacting QLHC. The STHLM3 trial data reported in this manuscript are available from the corresponding author upon request after ethical approvals have been obtained from the Swedish Ethical Review Board.


