Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2020 Mar 2;62(5):1264–1283. doi: 10.1002/bimj.201900020

Adaptive seamless clinical trials using early outcomes for treatment or subgroup selection: Methods, simulation model and their implementation in R

Tim Friede 1,, Nigel Stallard 2, Nicholas Parsons 2
PMCID: PMC8614126  PMID: 32118317

Abstract

Adaptive seamless designs combine confirmatory testing, a domain of phase III trials, with features such as treatment or subgroup selection, typically associated with phase II trials. They promise to increase the efficiency of development programmes of new drugs, for example, in terms of sample size and/or development time. It is well acknowledged that adaptive designs are more involved from a logistical perspective and require more upfront planning, often in the form of extensive simulation studies, than conventional approaches. Here, we present a framework for adaptive treatment and subgroup selection using the same notation, which links the somewhat disparate literature on treatment selection on one side and on subgroup selection on the other. Furthermore, we introduce a flexible and efficient simulation model that serves both designs. As primary endpoints often take a long time to observe, interim analyses are frequently informed by early outcomes. Therefore, all methods presented accommodate interim analyses informed by either the primary outcome or an early outcome. The R package asd, previously developed to simulate designs with treatment selection, was extended to include subgroup selection (so‐called adaptive enrichment designs). Here, we describe the functionality of the R package asd and use it to present some worked‐up examples motivated by clinical trials in chronic obstructive pulmonary disease and oncology. The examples both illustrate various features of the R package and provide insights into the operating characteristics of adaptive seamless studies.

Keywords: adaptive design, clinical trials, closed test procedure, combination test, Dunnett test

1. INTRODUCTION

There is a long history of application of sequential methods in clinical trials to allow the monitoring of accumulating data at a series of interim analyses (Jennison & Turnbull, 1999; Whitehead, 1997). Whilst most early work focussed on the aim of stopping the trial as soon as sufficient evidence has been obtained, this body of work has rapidly expanded to include the use of interim data for other design adaptations, including sample size reestimation (Friede & Kieser, 2006; Proschan, 2009). Over the past years, there has been considerable interest in using interim analyses for selection of treatments, with less effective treatments dropped from the study, or for selection of patient subgroups, with recruitment following an interim analysis limited to subgroup(s) in which a promising effect is indicated (Bauer, Bretz, Dragalin, König, & Wassmer, 2016; Pallmann et al., 2018).

A major statistical challenge in the development of such methods is the control of the type I error rate when adaptations are made on the basis of data that will also be included in the final analysis. This can be achieved by the combination test approach of Bauer and Köhne (1994), which yields flexible designs for treatment selection (see, e.g., Bauer and Kieser, 1999; Posch et al., 2005; Bretz, Schmidli, König, Racine, and Maurer, 2006) and subgroup selection (see Brannath et al., 2009; Jenkins, Stone, and Jennison, 2011; Wassmer and Dragalin, 2015).

Earlier work has assumed that the adaptations would be informed by the pre‐specified primary outcome, which is also used for hypothesis testing. From a practical perspective, however, this can be a strong limitation. In particular, in chronic diseases clinically meaningful endpoints might take some time to observe, which means that most or all patients are recruited by the time the primary outcome is observed for the first patients. This is illustrated, for example, by Chataway et al. (2011) in the context of secondary progressive multiple sclerosis. As a consequence, adaptations need to be based on early outcomes for adaptive designs to be feasible in these situations. Therefore, some adaptive seamless designs have been extended to allow the use of short‐term endpoint data for decision‐making at interim while the pre‐specified primary endpoint is used for hypothesis testing (Friede et al., 2011; Friede, Parsons, & Stallard, 2012; Jenkins et al., 2011; Kunz, Friede, Parsons, Todd, & Stallard, 2014, 2015; Stallard, 2010; Stallard, Kunz, Todd, Parsons, & Friede, 2015).

It is acknowledged that these more complex designs require intensive simulation studies in the planning to evaluate their operating characteristics (Benda, Branson, Maurer, & Friede, 2010; Friede et al., 2010). A limitation to the use of adaptive methods in practical applications is often the availability of software to enable construction and evaluation of appropriate study designs and to conduct the final analysis. A number of commercial software packages including ADDPLAN and EAST are available for this purpose. Although some R packages for group‐sequential designs including including gsDesign and gscounts and adaptive group‐sequential designs such as rpact are available from CRAN, there is still a shortage of comprehensive freely available software for adaptive seamless designs with treatment or subgroup selection, with the exception of the R package asd developed to plan clinical trials with treatment selection (Parsons et al., 2012).

Using the combination test approach, the aim of this paper is to present, to our knowledge for the first time, designs for treatment and subgroup selection in a unified notation and to present an efficient simulation model for their evaluation. By expressing the subgroup selection problem in a similar setting to that of treatment selection, the R package asd, originally developed for treatment selection, was extended to include designs of both types, that is, with subgroup or treatment selection. The methods implemented are based on the combination testing approach with at most two design stages. The designs obtained are fully flexible, controlling the type I error rate for any data‐driven adaptation including treatment or subgroup selection as well as sample size adaptation.

Although the methods we propose can be based on p‐values obtained from either summary statistics or individual patient data, in order to provide a general and efficient simulation model, we base this on the simulation of standardized sufficient statistics. The simulation model assumes multivariate normal distributions for the test statistics, but not for the individual observations. Therefore, the simulation model is not only widely applicable but the simulations are also fast. Furthermore, early outcomes informing the interim decisions are incorporated, since this is often important in practice as explained above. The application of the methods using the R package asd will be illustrated by clinical trials in chronic obstructive pulmonary disease (COPD) and oncology with treatment and subgroup selection, respectively.

2. METHODS

2.1. Notation and hypotheses

We consider first the setting of treatment selection designs. The study is conducted in up to J2 stages. In the first stage, patients are randomized between K experimental treatments and a control treatment. In a general setting, suppose that observations of the pre‐specified primary outcome from treatment group k, k=0,,K, where k=0 corresponds to the control treatment, have a distribution depending on some parameter μk, and that it is desired to test the family of null hypotheses Hk:θk=0, k=1,,K, where θk=μkμ0. Let pkj denote a p‐value for the test of the null hypothesis Hk based on the data from patients first observed in stage j. These data may not be available at the time of the interim analysis following stage j, but become available only later on in the trial (Friede et al., 2011). As we will explain in Section 2.2 below, the interim decisions need not necessarily be based on the primary outcome but could, in principle, make use of any available outcome while still maintaining control of the type I error rate.

Next we consider the setting of subgroup selection designs. Suppose that a single experimental treatment is to be compared with a control treatment but that patients can be categorized as belonging to one or more predefined subgroups. Interest is focused on the treatment effect in the full population and each subgroup, in particular in the subgroup with the largest treatment effect.

As part of the aim of this paper is to make explicit the similarities between methodology for treatment and subgroup selection designs, we will use similar notation for both settings when this does not cause confusion. Thus suppose that observations come from patients in K subgroups (including the full population as a special case) labeled k,k=1,,K. Suppose that data for patients in subgroup k receiving treatment r,r=0,1 have a distribution depending on some parameter μkr. Setting θk=μk1μk0, it is desired to test the family of hypotheses Hk:θk=0, k=1,,K. As above, pkj denotes a p‐value for the test of null hypothesis Hk based on the data from patients with an outcome first observed in stage j.

2.2. Interim selection rules

Although in both the treatment and subgroup selection settings, the testing strategies described here control the type I error rate for any selection rule, it is good practice to specify selection rules in advance to enable calculation of operating characteristics including sample size and power. In the following, we introduce some possible examples of interim selection rules based on the primary outcome, though these could equally be applied to an early outcome as we will explain below.

One obvious way to proceed would be to select the treatment or subgroup which performs best in terms of some statistic Zj,k (where Zj,k is some estimator of θk following stage j). However, this might not be wise in situations where sample sizes are relatively small given the differences between the treatments or subgroups, since there is a rather high risk of picking some treatment or subgroup that is not optimal. The so‐called ε‐rule proposed by Kelly et al. (2005) selects all treatments or subgroups with statistics Zj,k for which Zj,kmaxiZj,iε (assuming that larger values of Zj,k are better). For ε=0, this rule reduces to selecting the maximum only. For large ε, no selection takes place as all treatments or subgroups are carried forward. Otherwise varying numbers of treatments or subgroups are carried forward into the next stage. For practical applications, it is advisable to study the operating characteristics of the design for a range of values for ε. Comments on the optimal choice of ε are provided in Section 4 of Friede and Stallard (2008). The authors conclude that “selecting only the best treatment is a simple selection rule which is often optimal or close to optimal” (Friede & Stallard, 2008).

Multi‐arm studies including several doses of an experimental drug motivate another selection rule. In some indications, it is not uncommon to select not one but two doses for confirmatory testing in phase III. The COPD study discussed in more detail in Section 5 is a good example for this. A more generalized version of this rule would be to select the best K out of K treatments or subgroups where K would be specified in advance.

The selection of treatments or subgroups in interim analyses could be informed by the primary outcome or, if this is not feasible, by an early outcome. In situations where an early outcome informs an interim adaptation and the primary endpoint is used for hypothesis testing, we refer to the primary endpoint also as the final outcome to distinguish it more clearly from the early outcome. The early outcome need not necessarily fulfill all requirements of a surrogate endpoint (Burzykowski, Molenberghs, & Buyse, 2005) as weaker conditions might suffice. Chataway et al. (2011) used the phrase of a “biologically plausible” outcome that “gives some indication as to whether the mechanism of action of a test treatment is working as anticipated”. Nevertheless, for the operating characteristics of the adaptive seamless design the correlation between the early and the final outcomes on an individual patient level as well as the treatment effects on both the early and the final outcome (population level) are relevant as we will see below.

2.3. Error rate control via the closed testing procedure

In either the treatment selection or the subgroup selection setting, the problem has been posed in such a way that it is desired to test the null hypotheses Hk:θk=0,k=1,,K. It is desirable to conduct these hypotheses tests so as to control the familywise error rate in the strong sense, that is to control the probability of rejection of any true null hypothesis within this family, at some specified level, α. Strong error rate control may be achieved through a closed testing procedure in which, denoting by HK the intersection hypothesis kKHk, all hypotheses HK for K{1,,k} are tested at nominal level α, and Hk rejected if and only if HK is rejected at this level for all Kk (Marcus, Peritz, & Gabriel, 1976).

Application of the closed testing procedure requires a test of the intersection hypothesis HK for each K{1,,k}. These hypotheses tests must also combine evidence from the different stages in the trial. This may be achieved through the use of a combination testing method, as described in detail in the next subsection.

2.4. The combination testing method and early stopping

Although the treatment or subgroup selection might be informed by an early outcome, hypothesis testing is for the primary outcome. Extending the notation introduced above, let pKj denote a p‐value for a test of the null hypothesis HK based on data from patients with an outcome first observed at stage j=1,,J. By construction, under HK, pKjU[0,1], or if a conservative test is used, pKj is stochastically larger than or equal to U[0, 1], for all j and K (Brannath, Posch, & Bauer, 2002). We also assume that the conditional distribution of pKj given pK1,,pKj1 is stochastically no smaller than a U[0, 1] for all pK1,,pKj1 for all j, which is also referred to as the p‐clud condition (Brannath et al., 2002). The condition is satisfied if the p‐values from different stages are independent. In practical applications, the p‐clud condition is only satisfied asymptotically and referred to as asymptotically p‐clud (Brannath et al., 2009). In the following, we assume that all p‐values are (at least asymptotically) p‐clud.

The p‐values from the different stages can be combined using a number of combination functions (Bauer & Köhne, 1994; Lehmacher & Wassmer, 1999) to give test statistics CKj(pK1,,pKj) which, under the assumptions above regarding the distributions of the p‐values, have known distributions under HK irrespective of adaptations made to the study design. These test statistics may thus be used to test hypothesis HK (Bauer & Kieser, 1999; Bretz et al., 2006).

Although a number of combination functions have been proposed, here we use the inverse normal combination function (Lehmacher & Wassmer, 1999), which is equivalent to the method of Cui, Hung, and Wang (1999). This gives test statistics CKj=j=1jwjΦ1(1pKj) where w1,,wJ are specified in advance and the sum of their squares is equal to 1, that is, j=1Jwj2=1. Given the distributional assumptions under HK, CKj are distributed as, or are stochastically no larger than, a multivariate normal distribution with mean zero, var(CKj)=w12++wj2 and cov(CKj,CKj)=w12++wmin{j,j}2.

Following Posch et al. (2005), we assume that hypotheses cannot be rejected once they are dropped, resulting in conservative tests. Furthermore, applying the closed testing principle outlined in Section 2.3 in an adaptive design the p‐value pKj is replaced by pKIjj where Ij is the set of hypotheses carried forward into stage j (Posch et al., 2005).

If interim analyses are used for adaptation of the design but not for stopping the trial, a final test of HK may be based on CKJ. More generally, a sequential test of HK may be conducted based on the joint distribution of CK1,,CKJ, rejecting HK if CKjuKj for some critical values uK=(uK1,,uKJ)T. To simplify the notation, uK and uKj will generally be written as u and uj when it is clear which hypothesis is tested. The single constraint that the overall error rate should be at most α is insufficient to determine u uniquely. A common approach in sequential analysis is to specify the type I error to be spent at each interim analysis, and to find uK1,,uKJ to satisfy

prHK(CKjuKj,somejj)=αj, (1)

where α1αJ=α are either specified in advance (Slud & Wei, 1982) or depend on the observed information in some predetermined way (Lan & DeMets, 1983). Critical values uK1,,uKJ satisfying (1) can be found recursively with uKj found directly from the joint distribution of CHK1,,CHKJ via a numerical search once uK1,,uKj1 are known. Computational details are given by, for example, Jennison and Turnbull (1999).

Construction of the sequential test statistics CK1,,CKJ requires specification of p‐values pK1,,pKJ for testing HK. For elementary hypotheses, that is when |K|=1, pKj can be obtained from a standard test, such as a t‐test for normally distributed data or a chi‐squared test for binary data. When |K|>1, pKj should be calculated so as to allow for the multiple comparisons implicit in testing HK. In the treatment selection setting, as comparisons are with a common control, a Dunnett test (Dunnett, 1955) may be used. In the subgroup selection setting, the simple Bonferroni procedure, Simes' procedure (Brannath et al., 2009) or the Spiessens–Debois test (Spiessens & Debois, 2010), a Dunnett‐type test with a generalized covariance structure, may be used. In each case, the level of adjustment depends on the size of K.

3. SIMULATION MODEL

As eluded to in Section 1, simulations are often required in the planning phase of a study to choose design options such as sample sizes or interim selection rules. Here, we propose a simulation model that is efficient in the sense that population statistics are generated rather than individual patient data. Therefore, computation times do not increase with larger sample sizes.

In a wide variety of settings, it is possible to obtain statistics that are, at least asymptotically, normally distributed with known variance. When a series of interim analyses is conducted, these statistics follow a multivariate normal distribution with non‐zero correlations, since data obtained at earlier stages are also used at later ones. In the setting of treatment or subgroup selection, these multivariate normal distributions may be extended to give the joint distribution of statistics corresponding to different treatment comparisons or for treatment effects in different subgroups. To be clear, these statistics are not actually necessarily the test statistics used to test the hypotheses but are rather some estimators of θk. Similarly, the distributional forms assumed in the simulations need not be used as the basis of hypotheses tests, since these, as described in Section 2.4, can use the combination testing approach.

When considering adaptations to the trial design based on the observation of short‐term endpoint data, it is helpful to further extend the simulation models to include test statistics calculated based on different endpoints. The distributions of the resulting test statistics are described briefly in this section. Additional detail is given in Appendix.

3.1. Treatment selection

Considering treatment selection designs, let θ^kj denote the estimate of the treatment effect, θk, for treatment k based on the data available at the jth interim analysis, and Ikj1 denote the variance of this estimate. As these estimates are often, at least asymptotically normally distributed, our model will be based on an assumption of this distributional form. Assume that Ikj does not depend on k, and so may be denoted by Ij, and that the correlation between different treatment comparisons with a common control group is 1/(1+λ), as will often be the case if we have 1:λ randomization. Setting Skj=θ^kjIkj, we then have

(S11,,SK1,,S1J,,SKJ)NIθ,Σ(GS)(I)ΣK(CS)11+λ,

where I and θ denote, respectively, the vectors (I1,,IJ) and (θ1,,θK) and

ΣK(CS)(r)=1rrr1rrr1

and

Σ(GS)(I)(1)=I1(1)I1(1)I1(1)I1(1)I2(1)I2(1)I1(1)I2(1)IJ(1)

denote variance matrices of the K×K complex symmetric form and of that obtained in the usual group‐sequential setting.

3.2. Subgroup selection

In the subgroup selection setting, let θ^kj denote the estimate of the treatment effect, θk, for subgroup k based on the data available at the jth interim analysis. Again, assume that these estimates are normally distributed, let Ikj1 denote the variance of this estimate and assume that Ikj does not depend on k, and so may be denoted by Ij. In the simplest case, in which K=2 and subgroup 2 is the whole population and subgroup 1 a proportion of size τ, setting Skj=θ^kjIkj, we have

(S11,S21,,S1J,S2J)NIθ,Σ(GS)(I)Σ(GS)τ1.

Extensions to the case when a short‐term endpoint is also considered, and to more complex subgroup selection settings are discussed in the Appendix.

4. SOFTWARE IMPLEMENTATION IN R

4.1. R package asd

The simulation models for two‐stage treatment selection and subgroup selection designs, described in Section 3, can be implemented in the R (R Development Core Team, 2009) package asd (Parsons et al., 2012), which is available from the Comprehensive R Archive Network (CRAN) at http://cran.r‐project.org/package=asd). This package comprises a number of functions that allow the properties of seamless phase II/III clinical trial designs, potentially using early outcomes, for treatment or subgroup selection to be explored and evaluated prior to a study commencing. An earlier version of this package, without the extension to subgroup designs, enhanced options for a wider range of outcome measures or more complete output model description was described previously by Parsons et al. (2012). The general structure of the code comprises a set of base functions that implement lower level tasks such as hypothesis testing, treatment selection, and closed testing procedures. The base functions, which have been described previously in the setting of treatment selection designs (Parsons et al., 2012), have been modified in the latest version of the asd package to work with both subgroup and treatment selection designs. The higher level user facing function asd.sim has been replaced by two new functions treatsel.sim and subpop.sim that are called directly by the user to implement hypothesis testing and simulations. The more general functions, gtreatsel.sim and gsubpop.sim can in principle be called directly by the user, although due to the more complex input structure this is generally not recommended.

4.2. Function subpop.sim

Friede et al. (2012) extended the previously described combination test (CT) approaches, for co‐primary analyses in a single pre‐defined subgroup and the full population, using the methods proposed by Spiessens and Debois (2010) to control the familywise error rate (FWER) in the subgroup and the full population, and also proposed a novel method to obtain a critical value for the definitive test using a conditional error function (CEF) approach; full details of these methods are given by Friede et al. (2012). The function subpop.sim implements all the methods for subgroup selection in adaptive clinical trials reported by Friede et al. (2012) and subsequent correspondence (Friede, Parsons, & Stallard, 2013). The authors described and explored the performance of a number of methods in the setting described here, distinguishing between two distinct approaches to control the FWER, a CT method (Brannath et al., 2009; Jenkins et al., 2011) and a CEF method.

4.2.1. Input arguments

An overview and brief description of the input arguments available for subpop.sim is shown in Table 1. The CT methodology described by Friede et al. (2012) can be implemented in subpop.sim using either the Spiessens and Debois (2010) (SD), Simes or Bonferroni testing procedures to control the FWER. These and the CEF approach can be selected using the method argument to subpop.sim. The following options are available: (i) CT‐SD (method="CT-SD"), (ii) CT‐Simes (method="CT-Simes"), (iii) CT‐Bonferroni (method="CT-Bonferroni"), and (iv) CEF (method ="CEF").

TABLE 1.

Brief description and available input arguments to R functions subpop.sim and treatsel.sim

Implementation
Argument Description Treatment selection (treatsel.sim) Subgroup selection (subpop.sim)
n Sample sizes for each treatment group at stage 1 (interim) and stage 2 (final) analyses List of sample sizes; for example, 32 in each group in stage 1 and 64 in stage 2, list(stage1=32, stage2=64) An additional list option can be used to increase (enrich) the sample size in stage 2; for example, enrich sample size to 128 in stage 2, list(stage1=32, enrich=128, stage2=64)
effect Effect sizes for early and final outcomes List of effect sizes for the control (first) and each treatment group; for example, for control (0) and two effects of size 0.1 and 0.2 for early and 0.2 and 0.3 for final outcome, list(early=c(0,0.1,0.2), final=c(0,0.2,0.3)) The first element of each vector is the effect size in the subgroup and the second is the effect size in the full population; for example, for an effect size of 0.4 in subgroup and 0.2 in the full population for both early and final outcomes list(early=c(0.4,0.2), final=c(0.4,0.2)). An optional argument can be included to set the effect size for the control group; for example, default is zero, control=list(early=0, final=NULL)
outcome Outcome type for early and final outcomes List of outcome types, options for normal (N), time‐to‐event (T), and binary (B) are currently available; for example, normal for early and final outcomes list(early="N", final="N").
nsim Number of simulations An integer <1×107
sprev Subgroup prevalence The prevalence of the subgroup in the main population; 0<sprev<1. Subgroup prevalence can be either fixed or allowed to vary at each simulation; default is sprev.fixed=TRUE
corr Correlation between outcomes Correlation between early and final outcomes; 1<corr<1
seed Seed number Seed number to ensure repeatability of simulations
select Method for treatment selection Seven available options (see Section 4.3.1), with default, select=0, to select all treatments. For select=4, the epsilon rule, can be used to allow more flexibility, with default epsilon=1 and for select=6, all treatments greater than a threshold can be selected, with default thresh=1 Two options are available. The default threshold selection rule (select="thresh"), for which limits must be set; for example, selim=c(-1,1). A futility selection rule (select="futility"; See Section 5.2) is also available, for which limits must be set; selim=c(0,0).
ptest Treatments for counts of the number of rejections A vector of valid treatment numbers for treatment specific counts of rejections; for example, for four treatments, ptest=c(1), ptest=c(1, 2) or ptest=c(1,2,3,4) are all valid options
method Methodology used for simulations; and for subpop.sim either Either method="invnorm" or "fisher" to select inverse normal or Fisher combination test Select method="CT-SD", "CT-Simes", "CT-Bonferroni" or "CEF" (see Section 4.2.1)
fu Follow‐up options Subjects in the dropped treatment groups can followed‐up (fu=TRUE), with default fu=FALSE (see Section 4.3.1.)
weight Stage 1 weight Stage 1 weight be defined with this option (0 ⩽ weight ⩽1); default weight=NULL
level Test level Test level must be set (0level1); default level=0.025
file File name for output If unset will default to R console. For example, set file="output.txt", direct output to file in R working directory (getwd())

The syntax providing the group sample sizes, for stages 1 and 2, for a putative trial design and effect sizes for early and final outcomes is consistent across the available outcome types and comprises a list of the selected options. For instance, a design where the required sample size per treatment arm is 100 for stage 1 and 200 for stage 2 would be implemented using the following expression, n=list(stage1= 100,stage2=200). The assumption, in the current version of asd, is that the sample size in the control arm of the study is the same as in the treatment arm. The default setting is that if the subgroup only is selected at the interim analysis at stage 1, then the subgroup prevalence remains the same after stage 1. If an increase in the sample size in the subgroup was planned, if this group only was selected (enrichment), then this can be implemented by adding an additional item to the list to indicate this. For instance, if we wanted a sample size of 200 in the subgroup in this setting, irrespective of the subgroup prevalence, then we would modify the previous argument to the following n=list(stage1=100,enrich=200,stage2=200).

Effect sizes for early and final outcomes are also given as a list using expressions of the following structure, where the first element of each vector is the effect size in the subgroup and the second element is the effect size in the full population. Setting effect=list(early=c(0.3,0.1),final=c(0.3,0.1)) specifies an effect size of 0.3 in the subgroup and 0.1 in the full population; the effect size in the control group is set by default to be zero. The default setting is for normal outcomes for both early and final outcomes, outcome=list(early="N",final="N"), and the effects are interpreted given these options. The available options for outcome types are normal (N), time‐to‐event (T), and binary (B), and all combinations of these are allowed for early and final outcome measures. Generally, it is assumed that higher means (N) and lower event rates (B or T) are better. A detailed description of these options is left for the following section describing function treatsel.sim; the options described there are analogous to those available for subpop.sim. In the simpler setting where group selection is based purely on the final outcome, this can be implemented by setting the effect sizes for the early and final outcomes to be equal and the correlation between the early and final outcomes to one (i.e., corr=1).

The subgroup prevalence is set by a single argument, namely (sprev). For instance, sprev=0.5 indicates that the subgroup comprises half of the full population. The function subpop.sim randomly generates test statistics (with a seed number set using seed) and accumulates results from usually a large number of simulations that must be set using the nsim option (default setting, nsim=1000). The prevalence can be either fixed at the set value (sprev.fixed=TRUE) or allowed to vary (sprev.fixed=FALSE) using a single realization of the binomial random variate generation function rbinom, at the set values for the sample size and subgroup prevalence, at each simulation.

Subgroup selection at interim is implemented using the so‐called threshold selection rule (Friede & Stallard, 2008; Friede et al., 2012) (select="thresh"). If for the difference Δ of the test statistics for the full population and the subgroup holds Δl1, then the subgroup only is tested at the end of stage 2. If Δ>l2 the full population only is tested. Otherwise both subgroup and full populations are tested. The thresholds (l1,l2) are set using the argument selim, which is a vector of standard deviation multiples. If, for instance, large limits are set (e.g., selim=c(-10,10)) then both subgroup and full populations will always be tested at the trial endpoint, whereas if selim=c(0,0), only the test regarding the population with the largest test statistic at interim is taken into stage 2. Intermediate values for selim between these extremes provide more flexible selection options. The weight for the CT approaches, if unset, is given by nstage1/(nstage1+nstage2) with nstage1 and nstage2 referring to the (full) population sizes at stage 1 and stage 2, respectively. The test level is set by default to 0.025 (level=0.025).

4.2.2. Output

The output to subpop.sim first gives a summary of the simulation model, including expected values for the test statistics at each stage of the study. The main summary table reports the number of times that hypotheses H0{S}, H0{F}, H0{S,F} were selected for testing and rejected when the subgroup (S), the full population (F), or both were tested. Output from subpop.sim is by default directed to the usual R console, but to save more detailed summaries of the simulation model, output can be directed to a file using this as an argument to the file function (e.g., file="output.txt"). Section 5 shows how subpop.sim is used in a practical setting with example data.

4.3. Function treatsel.sim

Function treatsel.sim replaces and generalizes the previous function asd.sim, with the name change made to make it much more explicit that the code implements only simulations for treatment selection designs for multi‐arm studies. Much of the syntax and model setup is consistent between treatsel.sim and subpop.sim.

4.3.1. Input arguments

An overview and brief description of the input arguments available for treatsel.sim is shown in Table 1. One aspect of the design setup that differs considerably between treatsel.sim and subpop.sim is the coding of the treatment effects. Treatment effect sizes are given for the control group μ0 and the test treatment or treatments μk and as vectors for both early and final outcomes; for instance, effect=list(early=c(0,0.1,0.2,0.1), final=c(0,0.1,0.2,0.3)) indicates that there are three test treatments with effect sizes 0.1, 0.2, and 0.1 for the early outcome and 0.1, 0.2, and 0.3 for the final outcome, respectively; the control is set to 0 for both outcomes. The null hypotheses tested are Hk:θk=μkμ0=0. There is no limit to the number of test treatment groups K. However, in practice, our experience is that the code runs slowly for designs with eight or more treatment groups, that is, K8. The setting of the effect sizes can be clarified further by considering the available options for the outcome types (normal N, time to event T, and binary B), set using the option outcome, with the default being to have both early and final outcomes normal (outcome=list(early="N",final="N")), with all nine combinations available. Generally, it is assumed that higher means (N) and lower event rates (B or T) are better.

For normal outcomes, the test statistics for the simulation model for the K test treatments, relative to the control group, are given by n/2×(μkμ0). For time‐to‐event outcomes, effects are interpreted as minus log hazard rates, with the control μ0 set to zero and test statistics are given by ok/4×(μkμ0), where the expected total number of events in the control and treatment groups ok for treatment k is calculated under an assumed exponential model to be n×(1exp(exp(μ0)))+n×(1exp(exp(μk))). Binary effects are characterized by log odds ratios, 1/ok+1/(nok)+1/o0+1/(no0)×(μkμ0), where μk is minus the log odds of the event and the observed number of events in treatment group k is ok=n×1/(1+exp(μk)). Some care must be taken when setting‐up the simulation model, as clearly the interpretation of the effect argument to treatsel.sim is dependent on the options selected for outcome.

The method argument to treatsel.sim allows either the inverse normal (invnorm) or Fisher's (fisher) combination test and the logical follow‐up argument (fu) determines whether (i) patients in the dropped treatment groups are removed from the trial and unknown test statistics in the dropped treatments are set to at stage 2 (fu=FALSE) or (ii) patients are kept in the trial and followed‐up to the final outcome, in the same manner as the patients recruited in stage 1 in the selected treatment groups (fu=TRUE); Friede et al. (2011) called option (i), the default setting, discontinued follow‐up and option (ii) complete follow‐up. Seven treatment selection rules based on stage 1 test statistics are available in treatsel.sim, and are chosen with the select argument; (i) select all treatments (select=0), (ii) select the maximum (select=1), (iii) select the maximum two (select=2), (iv) select the maximum three (select=3), (v) flexible treatment selection using the ε‐rule (Friede & Stallard, 2008; Friede et al., 2011), with additional argument (epsilon) (select=4), (vi) randomly select a single treatment (select=5) or (vii) select all treatments greater than a threshold, with the additional argument (thresh) (select=6).

The only additional argument available for treatsel.sim, that has not been covered in the section describing subpop.sim, is ptest. This is a vector of valid treatment numbers for determining specific counts for the number of simulations that reject the null hypothesis; for instance, for three test treatments and ptest=c(1,3), treatsel.sim will count and report the number of rejections of one or both hypotheses for testing treatments 1 and 3 against the control, in addition to the number of rejections of each of the elementary hypotheses.

4.3.2. Output

The output from treatsel.sim first gives a summary of the simulation model, including expected values for the test statistics at each stage of the study. The main summary tables report (i) the number of treatments selected at stage 1, (ii) treatment selection at stage 1, that is, how often each treatment was selected, (iii) counts of hypotheses rejected at study endpoint, for each of the elementary hypotheses (H0{1},H0{2},,H0{K}), and (iv) the number of times that one or more than one of the hypotheses identified in ptest are rejected. Section 5 shows how treatsel.sim is used in a practical setting with example data.

5. EXAMPLES

In this section, we illustrate the methods described above by two example studies using the R package asd. The first is a multi‐arm randomized controlled trial in COPD with treatment selection; which will be considered in Section 5.1. As a second example, we consider trials in oncology with time‐to‐event outcomes and subgroup selection; this will be considered in Section 5.2.

5.1. Clinical trial in COPD with treatment selection

Barnes et al. (2011) and Donohue et al. (2010) report a seamless adaptive design with dose selection, which was also discussed elsewhere (Cuffe, Lawrence, Stone, & Vandemeulebroecke, 2014). Patients were randomized to four doses of indacaterol (75μg, 150μg, 300μg and 600μg), active controls, and placebo control. For the purpose of illustration, we ignore the active controls in the following and use the observed results to illustrate the design process. The primary outcome was the percentage of days of poor control over 26 weeks. As recruitment for the entire study took only 6 months, the interim treatment selection could not be based on the primary endpoint. Trough forced expiratory volume in 1 s (FEV1) at 15 days was identified as a suitable early outcome to inform the interim analysis. From Figure 1 of Barnes et al. (2011), difference in trough FEV1 compared to placebo at 15 days is 150 ml, 180 ml, 210 ml, and 200 ml for the indacaterol doses 75μg, 150μg, 300μg, and 600μg, respectively. Reported 95% confidence intervals suggest two standard errors of treatment difference is approximately 60 ml, so assuming equal stage 1 sample sizes n1=110, then the standard deviation of the measurements is approximately 220 ml. Standardized effect sizes are thus approximately 0.68, 0.82, 0.95, and 0.91 for the indacaterol doses 75μg, 150μg, 300μg, and 600μg, respectively. Regarding the final outcome days of poor control (%) over 26 weeks, we gather from the report of stage 2 results at http://clinicaltrials.gov/show/NCT00463567 that the placebo rate was 35%. Let us assume rates of 31%, 30%, 28%, and 29% for the four doses of indacaterol (75μg, 150μg, 300μg and 600μg). Based on reported standard errors, we estimate the standard deviation to be approximately 30%. Hence, the standardized effect sizes are approximately 0.13 (75μg), 0.17 (150μg), 0.23 (300μg), and 0.20 (600μg). The approximate sample sizes per arm were 100 patients in stage 1 and 300 patients in stage 2. Since the aim was to select two doses of indacaterol at the interim analysis to take into stage 2, we consider an overall sample size for the two stage of 5×100+3×300=1,400 patients. Furthermore, a moderate positive correlation between early and final outcomes of 0.4 is assumed.

FIGURE 1.

FIGURE 1

Probability of rejection of at least one elementary hypothesis (○; solid line), and selection probabilities (dashed lines) for stage 1 sample sizes in range 10–175 per group

In the following, we will consider three settings: (i) continuous early and final outcomes and selecting always two doses for the second stage, as in the original study; (ii) continuous early and final outcomes, as in the original study, but for the purpose of illustration, we vary the selection rule using the threshold rule to allow varying numbers of doses being taken forward for confirmatory testing in the second stage; (iii) as final outcomes are non‐normal and early and final outcomes are on different scales, we assume in another setting that the final is binary and not continuous.

5.1.1. Continuous early and final outcomes: Selecting always two doses for the second stage

As an example, we consider a fixed total sample size of 1,400 patients, which we can allocate between the two stages, with either more or less resources for each stage. We consider the following combinations of n1=10,25,,175 and 5×n1+3×n2=1,400. Each one of these sample size options can be tested using the following implementation of the treatsel.sim function for the setting with 100 patients per arm in stage 1 and 300 patients per arm in stage 2:


treatsel.sim(n=list(stage1=100,stage2=300), effect=list(early=c(0,0.68,0.82,0.95,0.91),

 final=c(0,0.13,0.17,0.23,0.20)), outcome=list(early="N",final="N"),

 nsim=10000,corr=0.4,seed=145514,select=2, level=0.025,ptest=c(3,4))


This code sets the sample sizes for each stage, and provides the effect estimates as described above, for a normal early outcome (“N”) and a normal (“N”) final outcome. The correlation is set to 0.4, and the number of simulations to 10,000. The select=2 option implements the rule that chooses the two treatments with the largest test statistics an interim. The test level is set to 0.025 and the ptest options allows us to count rejections for either or both of the doses 300μg and 600μg. Results from running this code are as follows (omitting the parts describing the setup):


simulation of test statistics:

expectation early = 4.8 5.8 6.7 6.4

expectation final stage 1 = 0.9 1.2 1.6 1.4 and stage 2 = 1.6 2.1 2.8 2.4

weights: stage 1 = 0.5 and stage 2 = 0.87



number of treatments selected at stage 1:

 n %

 1 0 0.00

 2 10000 100.00

 3 0 0.00

 4 0 0.00

Total 10000 100.00



treatment selection at stage 1:

 n %

 1 383 3.83

 2 3282 32.82

 3 8661 86.61

 4 7674 76.74



hypothesis rejection at study endpoint:

 n %

 H1 183 1.83

 H2 2067 20.67

 H3 7206 72.06

 H4 5541 55.41



reject H3 and/or H4 = 8469 : 84.69%


The first part of the output provides a summary of the model setup, and the second part values of the test statistics used in the simulations. The squared weights are calculated, in this case, as 100/400 and 300/400. The results are summarized in the three lower tables. The first indicates that two treatments were always selected at stage 1, the second gives the number of simulations in which each treatment was selected and the third gives the number of simulations in which the elementary hypotheses were rejected. The final statement gives the number of simulations in which at least one of the treatments picked using the ptest options were rejected. Assigning the output for this function to an object that we for the sake of illustration call simply output, then the summaries described here can be accessed directly, for instance, for plotting data or other analysis, using the syntax output$count.total, output$select.total, output$reject.total, and output$sim.reject.

The results of the simulations suggest that given the large treatment effects on the early outcome (trough FEV1 at 15 days), and the modest effects on the final outcome (days of poor control (%) over 26 weeks), the greatest power would have been achieved by using a sample size of around 50 per group in stage 1 (see Figure 1).

5.1.2. Continuous early and final outcomes: Selecting varying numbers of doses for the second stage

As an alternative to always selecting the best two performing treatments at interim, we now consider a threshold rule, where all treatments with test statistics at interim analysis above a fixed threshold are taken into stage 2. If no treatments reach the threshold, the study is stopped for futility. Simulations are implemented using the same effect sizes as in setting 1, a stage 1 sample size of 40, and a stage 2 sample size of 400 in the following code for a threshold of 3:


treatsel.sim(n=list(stage1=40,stage2=400),

 effect=list(early=c(0,0.68,0.82,0.95,0.91),

 final= c(0,0.13,0.17,0.23,0.20)),

 outcome=list(early="N",final="N"),

 nsim=10000,corr=0.4,seed=145514,select=6,

 thresh=3,level=0.025,ptest=c(3,4))


The select=6 option implements the threshold rule, with the fixed early outcome test statistic threshold set using the thresh option. Results from running this code are as follows (omitting the parts describing the setup):


simulation of test statistics:

expectation early = 3 3.7 4.2 4.1

expectation final stage 1 = 0.6 0.8 1 0.9 and stage 2 = 1.8 2.4 3.3 2.8

weights: stage 1 = 0.3 and stage 2 = 0.95



number of treatments selected at stage 1:

 n %

 1 800 8.00

 2 1634 16.34

 3 3098 30.98

 4 4175 41.75

Total 9707 97.07



treatment selection at stage 1:

 n %

 1 5083 50.83

 2 7469 74.69

 3 8914 89.14

 4 8596 85.96



hypothesis rejection at study endpoint:

 n %

 H1 2480 24.80

 H2 4882 48.82

 H3 7769 77.69

 H4 6642 66.42



reject H3 and/or H4 = 8600 : 86%


Although the probability for futility stopping is not given explicitly, it can be easily derived. The considered design was stopped for futility only 3% (=100%− 97.07%) of the time; with all four experimental treatments being taken into stage 2 more than 41% of the time. Running the above code for thresholds in the range 0 to 6 (at intervals of a half) and extracting output for each option gives the results summarized in Figure 2.

FIGURE 2.

FIGURE 2

Probability of rejection of at least one elementary hypothesis (○; solid line), futility stopping (dotted line), selection probabilities (dashed lines), and expected total sample size (•; solid line) for thresholds on a standardized z‐scale in the range 0–6

The expected overall sample sizes for the scenarios in Figure 2, based on the simulated number of treatments selected at stage 1 and the fixed stage 1 and stage 2 sample sizes of 40 and 400, are as follows; 2199.5, 2197.3, 2188.1, 2164.3, 2109.0, 1991.2, 1802.5, 1548.2, 1264.0, 1004.2, 807.9, 688.2, and 631.0. The power drops off rapidly as the threshold increases from 3 to 5, as futility stopping increases from 3% to 65%. When, on average, two test treatments are taken into stage 2, that is when the fixed threshold is somewhere between 3.5 and 4 (overall sample size between 1548.2 and 1264.0), the power is lower (between 78.7% and 65.4%) than the analogous setting in Figure 1 (86.6%). The early outcome effect sizes and distributions are such that a fixed threshold that on average picks two treatments at interim, also stops for futility so often that it reduces the power considerably compared to a design that always takes exactly two treatments.

5.1.3. Continuous early and binary final outcome

The implementation described here has focused on normal outcome measures for both early and final outcomes. However, if one or other outcome were binary, the changes necessary to implement this new scenario are relatively straightforward. For instance, instead of using days of poor control over 26 weeks as the final outcome measure, we might use a threshold based on this outcome. If this was the case, then success or failure of the treatment could be determined for each study participant, based on some a priori threshold for the number of days that one might expect to maintain control over the 26 week period. For the sake of example, let us consider the failure rate to be 50% in the control group, and 45% (75μg), 45% (150μg), 40% (300μg), and 40% (600μg), at each dose, respectively. Given that every other aspect of the design is the same as the first setting in the COPD example, then this design can be implemented using the following code.


treatsel.sim(n=list(stage1=100,stage2=300), effect=list(early=c(0,0.68,0.82,0.95,0.91),

 final=c(0.50,0.45,0.45,0.40,0.40)), outcome=list(early="N",final="B"),

 nsim=10000,corr=0.4,seed=145514,select=2, level=0.025,ptest=c(3,4))


This gives an overall rejection probability of 76.99%, by changing the outcome argument for the treatsel.sim function of Section 5.1.2 to list(early="N",final="B") and the vector of final effect sizes to c(0.50,0.45,0.45,0.40,0.40).

5.2. Clinical trials in oncology with subgroup selection

Jenkins et al. (2011) suggested designs for adaptive seamless phase II/III designs for oncology trials using correlated survival endpoints. Designs of this type can be implemented relatively straightforwardly using the subpop.sim function. Here, we explore some design properties for a typical scenario from amongst the many that Jenkins et al. (2011) explored. We assume early and final time‐to‐event outcomes with a hazard ratio of 0.6 in the subgroup and 0.9 in the full population for both, and a correlation between endpoints of 0.5; in the setting of an oncology trial, the endpoints might be progression free and overall survival. We set the stage 1 sample size to 100 patients per arm and the stage 2 sample size to 300 patients per arm, if we progress in the full population, and to 200 patients per arm if we progress in the subgroup only. The subgroup prevalence is fixed at 0.3. Using the futility rule for selection at interim, with limits for the subgroup and full population both set to 0, this scenario can be implemented using the following code:


subpop.sim(n=list(stage1=100,enrich=200,stage2=300),

 effect=list(early=c(0.6,0.9),final=c(0.6,0.9)),

 sprev=0.3,outcome=list(early="T",final="T"),

 nsim=10000,corr=0.5,seed=1234,select="futility",

 selim=c(0,0),level=0.025,method="CT-SD")


The method="CT-SD" option implements the combination test method with Spiessens and Debois testing procedure. Given test statistics at interim of S 1 and S 2 for the subgroup and the full population, respectively, and selection rule limits (l1,l2), the futility rule implements the following options: (i) continue with a co‐primary analysis if S1<l1 and S2<l2, (ii) continue in the subgroup alone if S1<l1 and S2>=l2, (iii) continue in the full population alone if S1>=l1 and S2<l2, and (iv) stop for futility if S1l1 and S2l2. Results from running this code are as follows (omitting the parts describing the setup):


simulation of test statistics:

expectation early: sub-pop = -1.46 : full-pop = -0.58

expectation final stage 1: sub-pop = -1.46 : full-pop = -0.58

expectation final stage 2: sub-pop only = -3.76 : full-pop only = -1.01

expectation final stage 2, both groups selected: sub-pop = -2.52 : full-pop = -1.01

weights: stage 1 = 0.5 and stage 2 = 0.87



hypotheses rejected and group selection options at stage 1 (n):

 Hs Hf Hs+Hf Hs+f n n%

sub 2225 0 0 2225 2309 23.09

full 0 48 0 54 227 2.27

both 5370 1658 1636 5407 6987 69.87

total 7595 1706 1636 7686 9523 -

% 75.95 17.06 16.36 76.86 95.23 -

reject Hs and/or Hf = 76.65%


The output reports that in 75.95%, 17.06%, and 16.36% of the simulations, the null hypothesis was rejected in the subgroup, full population and both, respectively. The final two columns give a breakdown of the selections made at the interim analysis; 23.09% of the simulations were continued in the subgroup only, 2.27% in the full population only, 69.87% in both, and 4.77% were stopped for futility. A concise summary of this table can be obtained for further analysis, by assigning to an output object and accessing the results using the syntax output$results.

It is informative in understanding the futility rule to run the above code for a grid of futility rule limits in the range 0 to −3; the results of this for each option are summarized in Table 2, where the notation lS and lF indicate the limits for the subgroup and full population, respectively.

TABLE 2.

Selection probabilities, probability of futility stopping and power (probability of rejecting at least one elementary null hypothesis) for a range of values of lS and IF

Selection (%)
lF
lS
Subgroup Full Both Futility (%) Power (%)
0 0 23.1 2.3 69.9 4.8 76.7
0 −1 11.4 16.2 55.8 16.7 58.8
0 −2 2.3 45.1 26.5 26.1 34.2
0 −3 0.1 66.0 6.0 27.9 20.7
−1 0 60.0 0.4 32.3 7.3 83.9
−1 −1 37.4 4.0 29.7 29.0 61.4
−1 −2 12.3 16.5 16.9 54.2 30.3
−1 −3 1.5 28.4 4.8 65.4 13.8
−2 0 84.9 0.0 7.4 7.7 88.6
−2 −1 60.1 0.3 7.2 32.4 65.0
−2 −2 24.1 2.4 5.6 68.0 29.2
−2 −3 4.4 5.3 2.1 88.2 8.0
−3 0 91.6 0.0 0.7 7.7 89.7
−3 −1 66.7 0.0 0.7 32.6 66.0
−3 −2 28.6 0.1 0.6 70.7 28.8
−3 −3 5.6 0.4 0.3 93.7 6.1

From Table 2, it is clear that as lS becomes more negative, the subgroup is selected progressively less often and similarly as lF becomes more negative, it is selected progressively more often. The balance between the two limits determines overall power in this setting. With a strong effect in the subgroup and a much weaker effect in the full population, the best strategy is to always, unless stopping for futility, test in the subgroup at the final analysis. For the grid of values tested here, this is best achieved when lF=3 and lS=0.

6. DISCUSSION

Adaptive seamless designs are recognized as a tool to increase the efficiency of clinical development programs by combining features of learning and confirming in a single trial, while traditional development programs would have investigated these in separate trials. However, their implementation is more involved than traditional designs (see, e.g., Quinlan and Krams (2006) for a discussion). One aspect is the planning which is more complex, often requiring extensive Monte Carlo simulations (Benda et al., 2010; Friede et al., 2010). Here, we presented a unified framework for adaptive seamless designs with treatment or subgroup selection. Furthermore, we developed a flexible and yet efficient simulation model. This, as all other methods discussed, can accommodate interim selection informed by an early outcome rather than the final one. Furthermore, we demonstrate how the R package asd, freely available from CRAN, can be used to evaluate and compare operating characteristics of various designs.

Here, we employed the combination test approach to achieve type I error rate control. For treatment selection, alternative approaches build on the work of Thall, Simon, and Ellenberg (1988, 1989) using the group‐sequential method, but require that a single treatment continues along with a control beyond the first stage (Stallard & Todd, 2003) or that the number of treatments at each stage is specified in advance (Stallard & Friede, 2008). Methods based on the combination test approach are more flexible (see, e.g., Bauer and Kieser (1999), Posch et al. (2005), and Bretz et al. (2006)), but may be less powerful in some settings (Friede & Stallard, 2008). Magirr, Jaki, and Whitehead (2012) proposed a group‐sequential method that does allow completely flexible treatment selection, though this may be at the cost of conservatism, that is, decreased type I error rate below the nominal level, and an associated loss in power. Koenig et al. (2008) showed how the conditional error principle of Müller and Schäfer (2001) may be used to extend the Dunnett test (Dunnett, 1955) to a two‐stage design with flexible treatment selection. This has been shown to compare well in terms of power with competing methods (Friede & Stallard, 2008).

There is a smaller body of work on clinical trials with subgroup selection, which has mainly used the combination testing approach, although methods based on the conditional error principle (Friede et al., 2012; Placzek & Friede, 2019; Stallard, Hamborg, Parsons, & Friede, 2014) and the group‐sequential approach (Magnusson & Turnbull, 2013) have also been proposed. The number of subgroups considered is usually small and different assumptions are made regarding the subgroup structure. For instance, Placzek and Friede (2019) consider nested subgroups which might arise from using different thresholds on a continuous (or at least ordinal) biomarker. An overview is provided in Ondra et al. (2016).

Although the development of some of the methods was motivated by applications in later development phases, that is, the seamless progression from phase II to phase III, the methods are more widely applicable of course. The focus on later development phases arises mainly because type I error rate is often not considered an issue in early development phases. However, in development programmes that are different from traditional ones in that they do not include at least two independent confirmatory trials, this might be different as then evidence from earlier trials will also be considered by regulators for replication of the effect demonstrated in a single phase III trial. These situations typically arise with rare diseases, in particular, rare cancers where there is a move toward platform trials including basket and umbrella trials (Woodcock & LaVange, 2017). The application of the simulation model and its implementation in the R package asd are subject to future research.

As the title of the paper by Woodcock and LaVange (2017) indicates, sometimes the application of both treatment and population selection might be of interest, in particular, in early clinical development. Motivated by late stage confirmatory trials, here we considered only designs with either treatment or subgroup selection, but the testing procedure, simulation model, and the implementation could be extended to designs where both treatments and subgroups are selected. Again, this could be subject to further research.

Throughout the manuscript, we assume that the hypothesis tests are based on a single primary outcome and that this is not changed during the trial. We consider this the likely scenario for confirmatory trials, although the adaptation of endpoints has been discussed in the literature (Kieser, 2005). However, additional information arising from other outcomes can be used to support the interim decisions. As discussed by Chataway et al. (2011), these need not be surrogate endpoints.

The testing procedures, the simulation model, and their implementation in the R package asd are quite general and are applicable to a variety of outcomes including time‐to‐event endpoints. However, the assumptions made including the p‐clud condition must be satisfied at least asymptotically for the testing procedure and the simulation model to be valid. For time‐to‐event endpoints, these can be easily violated resulting in substantial inflation of the type I error rate (Bauer & Posch, 2004). The main reason lies in the fact that patients with censored observations in one data look will contribute data to the next look, which is not an issue as long as the information is restricted to the censored observation times and does not include other information such as baseline characteristics or other outcomes. This complication motivated some authors including Wassmer (2006) and Jahn‐Eimermacher and Ingel (2009) to propose extensions of the combination test approach which deal with the specific issues of time‐to‐event endpoints. An overview and a discussion is provided in Chapter 9 of Wassmer and Brannath (2015).

There are of course some further limitations. Here, we focused very much on hypothesis testing, although the estimation of the treatment effects is equally important. There has been some interest in improved estimators in adaptive seamless designs with treatment (Bowden & Glimm, 2014; Brannath et al., 2009; Posch et al., 2005; Stallard & Kimani, 2018) or subgroup selection (Kimani et al., 2015, 2018), in particular, in more recent years.

With regard to interim decisions, we considered here only fairly straightforward rules although Bayesian statistics (e.g., predictive probabilities (Brannath et al., 2009)) are also used as the basis for interim decisions. Currently, we consider expanding the R package asd in this direction.

CONFLICT OF INTEREST

The authors have declared no conflict of interest.

Open Research Badges

This article has earned an Open Data badge for making publicly available the digitally‐shareable data necessary to reproduce the reported results. The data is available in the Supporting Information section.

This article has earned an open data badge “Reproducible Research” for making publicly available the code necessary to reproduce the reported results. The results reported in this article could fully be reproduced.

Supporting information

Supporting Information

ACKNOWLEDGMENTS

All authors gratefully acknowledge support by the UK Medical Research Council (grant number G1001344). Furthermore, the authors are grateful to the associate editor and the reviewer for their helpful comments.

DERIVATION OF DATA SIMULATION MODELS

We consider first treatment selection designs. Let μ^kj(i) be the estimate for treatment k, k=0,1,,K with k=0 corresponding to control, at stage j , j=1,,J, for endpoint i, i=1,2.

Assume, as often holds at least asymptotically, that μ^kj(i) is normally distributed with

μ^kj(i)N(μk(i),(Ikj(i))1),

with

cov(μ^kj(i),μ^kj(i))=0,kk,
cov(μ^kj(i),μ^kj(i))=(Ikmax{j,j}(i))1

and

cov(μ^kj(i),μ^kj(i))=ρ(Ikmax{j,j}(i))1(Ikmax{j,j}(i))1,ii.

For independent normal random variables with known variance σ2, we have Ikmax{j,j}(i)=σ2/nkj(i) where nkj(i) is the number of observations at look j for treatment k on endpoint i.

Let θ^kj(i)=θ^kj(i)θ^0j(i),k=1,,K, Ikj(i)=((Ikj(i))1+(I0j(i))1)1 and Skj(i)=θ^kj(i)Ikj(i).

Let λkj(i)=(Ikj(i))1/(I0j(i))1. If λkj(i) is constant, say equal to λ, and Ikj(i)=Ik(i) for all k (for a single sample of known‐variance normals, this is equivalent to assuming that sample sizes are the same for all experimental treatments). In this case, considering first a single endpoint, we get

S11(1)SK1(1)S1J(1)SKJ(1)Nθ1(1)I1(1)θK(1)I1(1)θ1(1)IJ(1)θK(1)IJ(1),Σ (A1)

where the variance matrix is given by

Σ=I1(1)I1(1)1+λI1(1)1+λI1(1)1+λI1(1)I1(1)1+λI1(1)1+λI1(1)1+λI1(1)I1(1)I1(1)1+λI1(1)1+λI1(1)1+λI1(1)I1(1)1+λI1(1)1+λI1(1)1+λI1(1)I1(1)I1(1)1+λI1(1)1+λI1(1)1+λI1(1)I1(1)1+λI1(1)1+λI1(1)1+λI1(1)I1(1)I1(1)1+λI1(1)1+λI1(1)1+λI1(1)I1(1)1+λI1(1)1+λI1(1)1+λI1(1)I2(1)I2(1)1+λI2(1)1+λI2(1)1+λI2(1)I2(1)1+λI2(1)1+λI2(1)1+λI2(1)I2(1)I2(1)1+λI2(1)1+λI2(1)1+λI2(1)I2(1)1+λI2(1)1+λI2(1)1+λI2(1)I1(1)I1(1)1+λI1(1)1+λI1(1)1+λI1(1)I1(1)1+λI1(1)1+λI1(1)1+λI1(1)I2(1)I2(1)1+λI2(1)1+λI2(1)1+λI2(1)I2(1)1+λI2(1)1+λI2(1)1+λI2(1)IJ(1)IJ(1)1+λIJ(1)1+λIJ(1)1+λIJ(1)IJ(1)1+λIJ(1)1+λIJ(1)1+λIJ(1).

Note that we can rewrite the right hand side of (A1) as

NI(1)θ(1),Σ(GS)(I(1))ΣK(CS)(11+λ)

where I (1) and θ(1) denote, respectively, the vectors (I1(1),,IJ(1)), and (θ1(1),,θK(1)) and

ΣK(CS)(r)=1rrr1rrr1

and

Σ(GS)(I(1))=I1(1)I1(1)I1(1)I1(1)I2(1)I2(1)I1(1)I2(1)IJ(1)

denote variance matrices of the K×K complex symmetric form and of that obtained in the usual group‐sequential setting.

When both endpoints are considered, the vector

(S11(1),,SK1(1),,S1J(1),,SKJ(1),S11(2),,SK1(2),,S1J(2),,SKJ(2))

is normally distributed with mean

θ(1)I(1)θ(2)I(2)

and variance

Σ(GS)(I(1))ρΣ(GS)(I(12))ρΣ(GS)(I(12))Σ(GS)(I(2))ΣK(CS)(11+λ)

where is a J‐dimensional vector with Ij(12)=Ij(1)Ij(2),j=1,,J.

We consider next designs with subgroup selection. In analogy to the notation used above, now let θ^kj(i) be the estimate of the treatment effect in subgroup k, k=1,,K at stage j , j=1,,J, for endpoint i, i=1,2. We will typically consider the case of K=2, with k=2 corresponding to the entire population and k=1 corresponding to a subgroup comprising a proportion τ of the entire population.

Let (Ikj(i))1 denote the variance of θ^kj(i) and let Skj(i)=θ^kj(i)Ikj(i). As above, we will assume that the θ^kj(i) are normally distributed with mean θk(i) and variance (Ikj(i))1 with

cov(θ^kj(i),θ^kj(i))=(Imax{k,k}max{j,j}(i))1

and

cov(θ^kj(i),θ^kj(i))=ρ(Imax{k,k}max{j,j}(i)Imax{k,k}max{j,j}(i))1/2,ii.

Denote Ij2(i) by Ij(i) and assume further that Ij1(i)=τIj(i) for i=1,2,j=1,,J (see Spiessens and Debois), then

(S11(1),S21(1),,S1J(1),S2J(1),S11(2),S21(2),,S1J(2),S2J(2))

is normally distributed with mean

θ(1)I(1)θ(2)I(2)

and variance

Σ(GS)(I(1))ρΣ(GS)(I(12))ρΣ(GS)(I(12))Σ(GS)(I(2))Σ(GS)τ1

where I(i)=(I1(i),,IJ(i)), I12=(I1(12),,IJ(12)) with Ij(12)=Ij(1)Ij(2), θ(i)=(θ1(i),θ2(i)), and Σ(GS)(I) and ΣK(CS)(r) are as defined above.

If more than two subgroups are considered, or if these are nested in some other way than the second being a subset of the first, the last matrix can be modified accordingly to reflect this.

Friede T, Stallard N, Parsons N. Adaptive seamless clinical trials using early outcomes for treatment or subgroup selection: Methods, simulation model and their implementation in R. Biometrical Journal. 2020;62:1264–1283. 10.1002/bimj.201900020

REFERENCES

  1. Barnes, P. J. , Pocock, S. J. , Magnussen, H. , Iqbal, A. , Kramer, B. , Higgins, M. , & Lawrence, D. (2011). Integrating indacaterol dose selection in a clinical study in COPD using an adaptive seamless design. Pulmonary Pharmacology & Therapeutics, 23, 165–171. [DOI] [PubMed] [Google Scholar]
  2. Bauer, P. , & Kieser, M. (1999). Combining different phases in the development of medical treatments within a single trial. Statistics in Medicine, 18, 1833–1848. [DOI] [PubMed] [Google Scholar]
  3. Bauer, P. , & Köhne, K. (1994). Evaluation of experiments with adaptive interim analyses. Biometrics, 50, 1029–1041. [PubMed] [Google Scholar]
  4. Bauer, P. , Bretz, F. , Dragalin, V. , König, F. , & Wassmer, G. (2016). Twenty‐five years of confirmatory adaptive designs: Opportunities and pitfalls. Statistics in Medicine, 35, 325–347. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bauer, P. , & Posch, M. (2004). Letter to the editor: Modification of the sample size and the schedule of interim analyses in survival trials based on data inspections. Statistics in Medicine, 23, 1333–1335. [DOI] [PubMed] [Google Scholar]
  6. Benda, N. , Branson, M. , Maurer, W. , & Friede, T. (2010). Aspects of modernizing drug development using scenario planning and evaluation. Drug Information Journal, 44, 299–315. [Google Scholar]
  7. Bowden, J. , & Glimm, E. (2014). Conditionally unbiased and near unbiased estimation of the selected treatment mean for multistage drop‐the‐loser trials. Biometrical Journal, 56, 332–349. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Brannath, W. , Posch, M. , & Bauer, P. (2002). Recursive combination tests. Journal of the American Statistical Association, 97, 236–244. [Google Scholar]
  9. Brannath, W. , Zuber, E. , Branson, M. , Bretz, F. , Gallo, P. , Posch, M. , & Racine‐Poon, A. (2009). Confirmatory adaptive designs with Bayesian decision tools for a targeted therapy in oncology. Statistics in Medicine, 28, 1445–1463. [DOI] [PubMed] [Google Scholar]
  10. Bretz, F. , Schmidli, S. , König, F. , Racine, A. , & Maurer, W. (2006). Confirmatory seamless phase II/III clinical trials with hypotheses selection at interim: General concepts. Biometrical Journal, 48, 623–634. [DOI] [PubMed] [Google Scholar]
  11. Burzykowski, T. , Molenberghs, G. , & Buyse, M. E. (2005). The Evaluation of Surrogate Endpoints. New York: Springer. [Google Scholar]
  12. Chataway, J. , Nicholas, R. , Todd, S. , Miller, D. , Parsons, N. , Valdés‐Márquez, E. , Stallard, N. , & Friede, T. (2011). A novel adaptive design strategy increases the efficiency of clinical trials in secondary progressive multiple sclerosis. Multiple Sclerosis, 17, 81–88. [DOI] [PubMed] [Google Scholar]
  13. Cuffe, R. L. , Lawrence, D. , Stone, A. , & Vandemeulebroecke, M. (2014). When is a seamless study desirable? Case studies from different pharmaceutical sponsors. Pharmaceutical Statistics, 13, 229–237. [DOI] [PubMed] [Google Scholar]
  14. Cui, L. , Hung, H. M. J. , & Wang, S. J. (1999). Modification of sample size in group sequential clinical trials. Biometrics, 55, 853–857. [DOI] [PubMed] [Google Scholar]
  15. Donohue, J. F. , Fogarty, C. , Lötvall, J. , Mahler, D. A. , Worth, H. , Yorgancioglu, A. , Iqbal, A. , Swales, J. , Owen, R. , Higgins, M. , & Kramer, B. , for the INHANCE Study Investigators. (2010). Once‐daily bronchodilators for chronic obstructive pulmonary disease: Indacaterol versus tiotropium. American Journal of Respiratory and Critical Care Medicine, 182, 155–162. [DOI] [PubMed] [Google Scholar]
  16. Dunnett, C. W. (1955). A multiple comparison procedure for comparing several treatments with a control. Journal of the American Statistical Association, 50, 1096–1121. [Google Scholar]
  17. Friede, T. , & Kieser, M. (2006). Sample size recalculation in internal pilot study designs: A review. Biometrical Journal, 48, 537–555. [DOI] [PubMed] [Google Scholar]
  18. Friede, T. , & Stallard, N. (2008). A comparison of methods for adaptive treatment selection. Biometrical Journal, 50, 767–781. [DOI] [PubMed] [Google Scholar]
  19. Friede, T. , Nicholas, R. , Stallard, N. , Todd, S. , Parsons, N. , Valdés‐Márquez, E. , & Chataway, J. (2010). Refinement of the clinical scenario evaluation framework for assessment of competing development strategies with an application to multiple sclerosis. Drug Information Journal, 44, 713–718. [Google Scholar]
  20. Friede, T. , Parsons, N. , Stallard, N. , Todd, S. , Valdés‐Márquez, E. , Chataway, J. , & Nicholas, R. (2011). Designing a seamless phase II/III clinical trial using early outcomes for treatment selection: An application in multiple sclerosis. Statistics in Medicine, 30, 1528–1540. [DOI] [PubMed] [Google Scholar]
  21. Friede, T. , Parsons, N. , & Stallard, N. (2012). A conditional error function approach for subgroup selection in adaptive clinical trials. Statistics in Medicine, 31, 4309–4320. [DOI] [PubMed] [Google Scholar]
  22. Friede, T. , Parsons, N. , & Stallard, N. (2013). Correction: ”A conditional error function approach for subgroup selection in adaptive clinical trials”. Statistics in Medicine, 32, 2513–2514. [DOI] [PubMed] [Google Scholar]
  23. Jahn‐Eimermacher, A. , & Ingel, K. (2009). Adaptive trial design: A general methodology for time to event data. Contemporary Clinical Trials, 30, 171–177. [DOI] [PubMed] [Google Scholar]
  24. Jenkins, M. , Stone, A. , & Jennison, C. (2011). An adaptive seamless phase II/III design for oncology trials with subpopulation selection using correlated survival endpoints. Pharmaceutical Statistics, 10, 347–356. [DOI] [PubMed] [Google Scholar]
  25. Jennison, C. , & Turnbull, B. W. (1999). Group Sequential Methods with Applications to Clinical Trials. Boca Raton, FL: CRC Press. [Google Scholar]
  26. Kelly, P. J.S tallard., N., & Todd, S. (2005). An adaptive group sequential design for phase II/III clinical trials that select a single treatment from several. Journal of Biopharmaceutical Statistics, 15, 641–658. [DOI] [PubMed] [Google Scholar]
  27. Kieser, M. (2005). A note on adaptively changing the hierarchy of hypotheses in clinical trials with flexible design. Drug Information Journal, 39, 215–222. [Google Scholar]
  28. Kimani, P. K. , Todd, S. , & Stallard, N. (2015). Estimation after subpopulation selection in adaptive seamless trials. Statistics in Medicine, 34, 2581–2601. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Kimani, P. K. , Todd, S. , Renfro, A. L. , & Stallard, N. (2018). Point estimation following two‐stage adaptive threshold enrichment clinical trials. Statistics in Medicine, 34, 2581–2601. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Koenig, F. , Brannath, W. , Bretz, F. , & Posch, M. (2008). Adaptive Dunnett tests for treatment selection. Statistics in Medicine, 27, 1612–1625. [DOI] [PubMed] [Google Scholar]
  31. Kunz, C. U. , Friede, T. , Parsons, N. , Todd, S. , & Stallard, N. (2014). Data‐driven treatment selection for seamless phase II/III trials incorporating early‐outcome data. Pharmaceutical Statistics, 13, 238–246. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Kunz, C. U. , Friede, T. , Parsons, N. , Todd, S. , & Stallard, N. (2015). A comparison of methods for treatment selection in seamless phase II /III clinical trials incorporating information on short‐term endpoints. Journal of Biopharmaceutical Statistics, 25, 170–189. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Lan, K. K. G. , & DeMets, D. L. (1983). Discrete sequential boundaries for clinical trials. Biometrika, 70, 659–663. [Google Scholar]
  34. Lehmacher, W. , & Wassmer, G. (1999). Adaptive sample size calculation in group‐sequential trials. Biometrics, 55, 1286–1290. [DOI] [PubMed] [Google Scholar]
  35. Magirr, D. , Jaki, T. , & Whitehead, J. (2012). A generalized Dunnett test for multi‐arm multi‐stage clinical studies with treatment selection. Biometrika, 99, 494–501. [Google Scholar]
  36. Magnusson, B. P. , & Turnbull, B. W . (2013). Group sequential enrichment design incorporating subgroup selection. Statistics in Medicine, 32, 2695–2714. [DOI] [PubMed] [Google Scholar]
  37. Marcus, R. , Peritz, E. , & Gabriel, K. R. (1976). On closed testing procedures with special reference to ordered analysis of variance. Biometrika, 63, 655–660. [Google Scholar]
  38. Müller, H.‐H. , & Schäfer, H. (2001). Adaptive group sequential designs for clinical trials: Combining the advantages of adaptive and of classical group sequential approaches. Biometrics, 57, 886–891. [DOI] [PubMed] [Google Scholar]
  39. Ondra, T. , Dmitrienko, A. , Friede, T. , Graf, A. , Miller, F. , Stallard, N. , & Posch, M. (2016). Methods for identification and confirmation of targeted subgroups in clinical trials: A systematic review. Journal of Biopharmaceutical Statistics, 26, 99–119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Pallmann, P. , Bedding, A. W. , Choodari‐Oskooei, B. , Dimairo, M. , Flight, L. , Hampson, L. V.... Jaki, T. (2018). Adaptive designs in clinical trials: Why use them, and how to run and report them. BMC Medicine, 16, 29. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Parsons, N. , Friede, T. , Todd, S. , Valdes Marquez, E. , Chataway, J. , Nicholas, R. , & Stallard, N. (2012). An R package for implementing simulations for seamless phase II/III clinical trials using early outcomes for treatment selection. Computational Statistics and Data Analysis, 56, 1150–1160. [Google Scholar]
  42. Placzek, M. , & Friede, T . (2019). A conditional error function approach for adaptive enrichment designs with continuous endpoints. Statistics in Medicine, 38, 3105–3122. [DOI] [PubMed] [Google Scholar]
  43. Posch, M. , Koenig, F. , Branson, M. , Brannath, W. , Dunger‐Baldauf, C. , & Bauer, P . (2005). Testing and estimation in flexible group sequential designs with adaptive treatment selection. Statistics in Medicine, 24, 3697–3714. [DOI] [PubMed] [Google Scholar]
  44. Proschan, M. A. (2009). Sample size re‐estimation in clinical trials. Biometrical Journal, 51, 348–357. [DOI] [PubMed] [Google Scholar]
  45. Quinlan, J. A. , & Krams, M. (2006). Implementing adaptive designs: Logistical and operational considerations. Drug Information Journal, 40, 437–444. [Google Scholar]
  46. R Development Core Team (2009). R: A language and environment for statistical computing. Vienna, Austria: R foundation for statistical computing. Available from http://www.R‐project.org
  47. Slud, E. , & Wei, L. J. (1982). Two‐sample repeated significance tests based on the modified Wilcoxon statistic. Journal of the American Statistical Association, 77, 862–868. [Google Scholar]
  48. Spiessens, B. , & Debois, M . (2010). Adjusted significance levels for subgroup analysis in clinical trials. Contemporary Clinical Trials, 31, 647–656. [DOI] [PubMed] [Google Scholar]
  49. Stallard, N. (2010). A confirmatory seamless phase II/III clinical trial design incorporating short‐term endpoint information. Statistics in Medicine, 29, 959–971. [DOI] [PubMed] [Google Scholar]
  50. Stallard, N. , & Friede, T. (2008). A group‐sequential design for clinical trials with treatment selection. Statistics in Medicine, 27, 6209–6227. [DOI] [PubMed] [Google Scholar]
  51. Stallard, N. , & Kimani, P. K. (2018). Uniformly minimum variance conditionally unbiased estimation in multi‐arm multi‐stage clinical trials. Biometrika, 105, 495–501. [Google Scholar]
  52. Stallard, N. , & Todd, S . (2003). Sequential designs for phase III clinical trials incoporating treatment selection. Statistics in Medicine, 22, 689–703. [DOI] [PubMed] [Google Scholar]
  53. Stallard, N. , Hamborg, T. , Parsons, N. , & Friede, T. (2014). Adaptive designs for confirmatory clinical trials with subgroup selection. Journal of Biopharmaceutical Statistics, 24, 168–187. [DOI] [PubMed] [Google Scholar]
  54. Stallard, N. , Kunz, C. U. , Todd, S. , Parsons, N. , & Friede, T. (2015). Flexible selection of a single treatment incorporating short‐term endpoint information in a phase II/III clinical trial. Statistics in Medicine, 34, 3104–3115. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Thall, P. F. , Simon, R. , & Ellenberg, S. S. (1988). Two‐stage selection and testing designs for comparative clinical trials. Biometrika, 75, 303–310. [Google Scholar]
  56. Thall, P. F. , Simon, R. , & Ellenberg, S. S. (1989). A two‐stage design for choosing among several experimental treatments and a control in clinical trials. Biometrics, 45, 537–547. [PubMed] [Google Scholar]
  57. Wassmer, G. (2006). Planning and analyzing adaptive group sequential survival trials. Biometrical Journal, 48, 714–729. [DOI] [PubMed] [Google Scholar]
  58. Wassmer, G. , & Brannath, W. (2015). Group Sequential and Confirmatory Adaptive Designs in Clinical Trials. Switzerland: Springer. [Google Scholar]
  59. Wassmer, G. , & Dragalin, V. (2015). Designing issues in confirmatory adaptive population enrichment trials. Journal of Biopharmaceutical Statistics, 25, 651–669. [DOI] [PubMed] [Google Scholar]
  60. Whitehead, J. (1997). The Design and Analysis of Sequential Clinical Trials (Revised 2nd ed.). Chichester, UK: Wiley. [Google Scholar]
  61. Woodcock, J. , & LaVange, L. M. (2017). Master protocols to study multiple therapies, multiple diseases, or both. New England Journal of Medicine, 377, 62–70. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supporting Information


Articles from Biometrical Journal. Biometrische Zeitschrift are provided here courtesy of Wiley

RESOURCES