Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2013 Jan 1.
Published in final edited form as: J Biopharm Stat. 2012;22(2):312–328. doi: 10.1080/10543406.2010.536873

Phase II Cancer Clinical Trials with Heterogeneous Patient Populations

Sin-Ho Jung 1, Myron N Chang 2, Sun J Kang 3
PMCID: PMC3324125  NIHMSID: NIHMS361921  PMID: 22251176

SUMMARY

The patient population for a phase II trial often consists of multiple subgroups in terms of risk level. In this case, a popular design approach is to specify the response rate and the prevalence of each subgroup, to calculate the response rate of the whole population by the weighted average of the response rates across subgroups, and to choose a standard phase II design such as Simon’s optimal or minimax design to test on the response rate for the whole population. In this case, although the prevalence of each subgroup is accurately specified, the observed prevalence among the accrued patients to the study may be quite different from the expected one because of the small sample size, which is typical in most phase II trials. The fixed rejection value for a chosen standard phase II design may be either too conservative (i.e., increasing the false rejection probability of the experimental therapy) if the trial accrues more high-risk patients than expected or too anti-conservative (i.e., increasing the false acceptance probability of the experimental therapy) if the trial accrues more low-risk patients than expected. We can avoid such problem by adjusting the rejection values depending on the observed prevalence from the trial. In this paper, we investigate the performance of the flexible designs compared with the standard design with fixed rejection values under various settings.

Keywords: Conditional power, Conditional type I error, Minimax design, Optimal design, Prevalence

1 Introduction

A typical single-arm phase II trial is to evaluate the efficacy of an experimental therapy compared to a historical control before it proceeds to a large scale phase III trial to be compared to a prospective control. The patient population for a phase II trial often consists of multiple subgroups, also called cohorts, with different prognosis although the study therapy is expected to be similarly beneficial for all subgroups. In this case, the final decision on the study treatment should adjust for the heterogeneity of the patient population.

Suppose that we want to evaluate the tumor response of CD30 antibody, SGN-30, combined with GVD (Gemcitabine, Vinorelbine, Pegylated Liposomal Doxorubicin) chemotherapy in patients with relapsed or refractory classical Hodgkin lymphoma (HL) through a phase II trial. In a previous study, GVD only has led to responses in 65% of patients with relapsed or refractory HL patients who never had a transplant and 75% in the transplant group. About 50% of patients in the previous study never had a transplant. Combining the data from the two cohorts, the response rate (RR) for the whole patient population is estimated as 70%(= 0.5 × 0.65 + 0.5 × 0.75).

Using this outcome as historical control data, the new study is designed as a single-arm trial for testing

H0:p70%againstHa:p>70%, (1)

where p denotes the true RR of the combination therapy in the patient population combining the two subgroups, one for those with prior transplants and the other for those without one.

A standard design to account for the heterogeneity of the patient population is a single-arm trial based on a specified prevalence for each cohort for testing hypotheses (1). For the example study, we consider an increase in RR by 15% or larger clinically significant for each cohort. So, we will not be interested in the combination therapy if the true RR, p, is lower than p0 = 70% and will be strongly interested if the true RR is higher than pa = 85%. Then, the Simon’s (1989) two-stage optimal design for testing

H0:p0=70%againstHa:pa=85%

with type I error no larger than α* = 0.1 and power no smaller than 1−β* = 0.9 is described as follows.

  • Stage 1

    Accrue n1 = 20 patients. If ā1 = 14 or fewer patients respond, then we stop the trial concluding that the combination therapy is inefficacious. Otherwise, the trial proceeds to stage 2.

  • Stage 2

    Accrue an additional n2 = 39 patients. If more than ā = 45 patients out of the total n = 59(= n1 + n2) respond, then the combination therapy will be accepted for further investigation.

Assuming that the number of responders from the two stages are independent binomial random variables with probability of ‘success’ pl under Hl (l = 0, a), we obtain the exact type I error and power of the two-stage design as 0.0980 and 0.9029, respectively.

In developing such a standard design, an accurate specification of the prevalence of each cohort is critical. If the prevalence is erroneously specified, the type I error of the statistical testing can not be accurately controlled. Even when the prevalence is accurately specified, the observed prevalence from the new study may be quite different from the true one. This can easily happen in phase II trials with mostly small sample sizes. If a new study accrues a larger number of high-risk (low-risk) patients than expected, then the trial will have a higher false negativity (positivity).

London and Chang (2005) resolve this issue by choosing rejection values based on a stratified analysis method. They adopt early stopping boundaries for both low and high efficacy cases based on a type I error rate and power spending function approach. Sposto and Gaynon (2009) propose a two-stage design with a lower stopping value only based on large sample approximations that may not hold well for phase II trials with small sample sizes. Wathen et al. (2008) propose a Bayesian method to test on the efficacy for each subgroup.

In this paper, we consider a similar design situation to those by London and Chang (2005) and Sposto and Gaynon (2009). The sample sizes are determined by a standard design, such as Simon’s minimax or optimal, based on a specified prevalence of each cohort, but the rejection value is adjusted depending on the observed prevalence from the trial. Conditioning on the observed prevalence, the rejection values are chosen by calculating the type I error rate and the power using the accurate probability distributions accounting for the small sample sizes of typical phase II trials. Using some real examples we show that our design accurately controls the conditional type I error rate and power in a wide range of the prevalence. In contrast, the conditional type I error of a standard design with fixed rejection values wildly fluctuates around the pre-specified level depending on the observed prevalence. Furthermore, the marginal type I error and power of the standard design can be heavily biased if the specified prevalence is different from the true prevalence.

The rejection values for a multi-stage phase II trial are chosen for the pre-determined sample size at each stage. However, due to dropouts or over-accrual, the realized sample size of the trial may be slightly different from the planned sample sizes. In this case, the chosen rejection values with respect to the planned sample size are not valid any more. To tackle this issue, we propose to calculate a p-value conditioning on the realized sample sizes and to conduct a statistical testing by comparing the conditional p-value with the pre-specified type I error rate. We also present a single-stage design for stratified phase II study analysis.

2 Single-Stage Designs

Suppose that we want to design a phase II trial on a new therapy with respect to a patient population with two cohorts of patients, called the high-risk cohort and the low-risk cohort. Cases with more than two cohorts will be discussed later. For cohort j(= 1, 2), let pj denote the RR of the therapy and γj denote the prevalence (γ1 + γ2 = 1). The RR for the combined population is given as p = γ1p1 + γ2p2. Based on some historical control data, we will not be interested in the new therapy if its RR for cohort j is p0j or lower, and will be highly interested in it if its RR is paj(= p0j + Δj for Δj > 0) or higher. Let p0 = γ1p01 + γ2p02 and pa = γ1pa1 + γ2pa2.

2.1 Unstratified Designs

A standard single-stage design to test hypotheses H0: pp0 vs. Ha: p > p0 is to accrue a certain number of patients, say n, and to reject the therapy, i.e. failing to reject H0, if the observed number of responders is smaller than or equal to a chosen rejection value ā. Given pre-specified type I error rate α*, power 1 − β*, and clinically significant difference Δj for cohort j(=1,2), we choose the smallest n together with an integer ā satisfying

α=P(X>a¯p=p0)α

and

1β=P(X>a¯p=pa)1β, (2)

where X denotes the number of responders among n patients. Given ā, we usually calculate the exact type I error α and power 1−β by regarding X as a binomial random variable with n independent Bernoulli trials with probability of success p = γ1p1 + γ2p2, i.e. α = B(ā|n, p0) and 1 − β = B(ā|n, pa), where B(xn,p)=i=x+1nb(xn,p) and

b(xn,p)=(nx)px(1p)nxforx=0,1,,n.

We call (n, ā) a standard or unstratified design.

Let b(n, p) denote the binomial distribution with n independent Bernoulli trials with probability of success p. Let Mj be a random variable denoting the number of patients from cohort j among n patients. Assuming that the population consists of infinitely many patients, we have M1 ~ b(n, γ1) and M2 = nM1. Conditioning on M1 = m1, the number of responders Xj among mj patients from cohort j follows b(mj, p0j) under H0. Hence, it is easy to show that above type I error for a standard design can be calculated also as

α=EM1P(X1+X2>a¯p01,p02,M1)=m1=0nx1=0m1x2=0nm1I(x1+x2>a¯)b(x1m1,p01)b(x2nm1,p02)b(m1n,γ1).

Power (2) can be calculated similarly.

2.2 Stratified Designs

For a stratified single-stage design, we propose to choose a value a satisfying the α*-condition given the observed m1 value while fixing n(= m1 + m2) at the sample size of a standard design. Given M1 = m1 (m2 = nm1), the conditional type I error for a rejection value a is calculated as

α(m1)=P(X1+X2>ap01,p02,m1)=x1=0m1x2=0m2I(x1+x2>a)b(x1m1,p01)b(x2m2,p02).

Given m1, we want to choose the maximal a = a(m1) such that α(m1) ≤ α*. For the chosen rejection value a = a(m1), the conditional power is calculated as

1β(m1)=P(X1+X2>apa1,pa2,m1)=x1=0m1x2=0m2I(x1+x2>a)b(x1m1,pa1)b(x2m2,pa2). (3)

In summary, a stratified single-stage design for a population with two cohorts is chosen as follows:

  • Step 1

    Specify γ1, (p01, p02, pa1, pa2), and (α*, 1 − β*).

  • Step 2

    Choose a reasonable n as follows.

    1. Calculate p0 = γ1p01 + γ2p02 and pa = γ1pa1 + γ2pa2.

    2. Choose a standard single-stage design (n, ā) for testing
      H0:p=p0vs.Ha:p=pa

      under the (α*, 1 − β*)-condition. We choose this n (or a little larger number) as the sample size of the stratified design.

  • Step 3

    For m1 ∈ [0, n], choose the maximum a = a(m1) satisfying α(m1) ≤ α*.

  • Step 4

    Given (n, m1, a), calculate the conditional power 1 − β(m1) by (3).

The study protocol using a stratified design may provide a table of {a(m1), α(m1), 1−β(m1)} for 0 ≤ m1n. When the study is over, we observe m1 and x(= x1 + x2), and reject the study therapy if xa(m1).

Noting that M1 ~ b(n, γ1), we can calculate the marginal type I error and power of the stratified design by

α=E{α(M1)}=m1=0nα(m1)b(m1n,γ1)

and

1β=E{1β(M1)}=m1=0n{1β(m1)}b(m1n,γ1),

respectively. Since, for each m1 ∈ [0, …, n], we choose a = a(m1) so that its conditional type I error does not exceed α*, the marginal type I error will not exceed α*.

2.3 Example 1

Let’s consider the study discussed in Section 1 using Δ1 = Δ2 = 0.15. Under γ1 = γ2 = 0.5 and response rates (p01, p02) = (0.65, 0.75), the hypotheses in terms of the population RR are expressed as H0: p0 = 0.7 and Ha: p1 = 0.85. For (α*, 1 − β*) = (0.1, 0.9), the standard (unstratified) design with the minimal sample size is (n, ā) = (53, 41) which has α = 0.0906 and 1 − β = 0.9093. The type I error and power are valid only when the true prevalence is γ1 = γ2 = 0.5.

Suppose that the study observed (x1, x2) = (28, 13) and m1 = 36. Note that the observed prevalence for the high-risk cohort, γ̂1 = 36/53 = 0.68, is much larger than the expected γ1 = 0.5. By the unstratified design, x = 41 equals the rejection value ā = 41, so that the therapy will be rejected. However, noting that m1 = 36 is much larger than expected, the stratified design lowers the rejection value to a = 40, so that, with observation x = 41, the therapy will be accepted for further investigation. Similarly, the unstratified Simon’s design may falsely accept the therapy if γ̂1 is much lower than the specified prevalence γ1 = 0.5.

Table 1 lists the conditional type I error and power of the standard unstratified design for each m1 ∈ [0, n]. Note that if m1 is much larger than 1, i.e. too many cohort 1 (high risk) patients are accrued, then the standard rejection value ā = 41 is so anti-conservative that the conditional type I error and power become smaller than the specified α* = 0.1 and 1 − β* = 0.9, respectively. On the other hand, if m1 is too small compared to 1, i.e. too many cohort 2 (low-risk) patients are accrued, then the standard rejection value ā = 41 is so conservative that the conditional type I error becomes larger than the specified α* = 0.1 level. Figure 1(a) displays the conditional type I error rate and power of the standard (unstratified) design. We observe that the conditional type I error of the standard design widely varies between 0.0182 for m1 = 53 and 0.2961 for m1 = 0. Its conditional power also widely varies around 1 − β* = 0.9.

Table 1.

Conditional type I error and power of single-stage standard (unstratified) and stratified designs with n = 53 for (p01, p02, Δ) = (0.65, 0.75, 0.15) and (α*, 1 − β*) = (0.1, 0.9). The standard design has a fixed critical value ā = 41.

m1 Unstratified
Stratified
m1 Unstratified
Stratified
α 1 − β a α 1 − β α 1 − β a α 1 − β
0 0.2961 0.9947 44 0.0606 0.9215 27 0.0869 0.9081 41 0.0869 0.9081
1 0.2852 0.9939 44 0.0569 0.9142 28 0.0823 0.9011 41 0.0823 0.9011
2 0.2746 0.9930 44 0.0535 0.9065 29 0.0780 0.8938 41 0.0780 0.8938
3 0.2641 0.9919 43 0.0972 0.9517 30 0.0738 0.8862 41 0.0738 0.8862
4 0.2540 0.9908 43 0.0920 0.9467 31 0.0699 0.8782 41 0.0699 0.8782
5 0.2440 0.9896 43 0.0870 0.9414 32 0.0661 0.8699 41 0.0661 0.8699
6 0.2343 0.9882 43 0.0822 0.9357 33 0.0625 0.8612 41 0.0625 0.8612
7 0.2249 0.9866 43 0.0776 0.9296 34 0.0590 0.8523 41 0.0590 0.8523
8 0.2157 0.9849 43 0.0733 0.9232 35 0.0557 0.8430 41 0.0557 0.8430
9 0.2067 0.9830 43 0.0691 0.9163 36 0.0526 0.8334 40 0.0961 0.9049
10 0.1980 0.9810 43 0.0651 0.9091 37 0.0496 0.8236 40 0.0913 0.8981
11 0.1895 0.9787 43 0.0614 0.9015 38 0.0468 0.8134 40 0.0867 0.8909
12 0.1813 0.9763 43 0.0578 0.8935 39 0.0441 0.8029 40 0.0822 0.8835
13 0.1734 0.9736 43 0.0544 0.8851 40 0.0415 0.7922 40 0.0780 0.8757
14 0.1656 0.9707 42 0.0969 0.9368 41 0.0391 0.7812 40 0.0739 0.8677
15 0.1582 0.9676 42 0.0919 0.9311 42 0.0367 0.7699 40 0.0701 0.8593
16 0.1510 0.9642 42 0.0870 0.9250 43 0.0346 0.7584 40 0.0664 0.8506
17 0.1440 0.9606 42 0.0823 0.9186 44 0.0325 0.7467 40 0.0628 0.8417
18 0.1373 0.9566 42 0.0779 0.9119 45 0.0305 0.7347 40 0.0594 0.8325
19 0.1308 0.9525 42 0.0736 0.9048 46 0.0287 0.7225 40 0.0562 0.8230
20 0.1245 0.9480 42 0.0696 0.8973 47 0.0269 0.7102 39 0.0955 0.8886
21 0.1185 0.9432 42 0.0657 0.8895 48 0.0252 0.6977 39 0.0908 0.8813
22 0.1127 0.9382 42 0.0620 0.8813 49 0.0237 0.6850 39 0.0864 0.8738
23 0.1071 0.9328 42 0.0585 0.8727 50 0.0222 0.6722 39 0.0820 0.8659
24 0.1017 0.9271 42 0.0551 0.8638 51 0.0208 0.6592 39 0.0779 0.8578
25 0.0966 0.9211 41 0.0966 0.9211 52 0.0195 0.6462 39 0.0739 0.8495
26 0.0916 0.9148 41 0.0916 0.9148 53 0.0182 0.6330 39 0.0701 0.8408

Figure 1.

Figure 1

Conditional type I error and power of standard (unstratified) and stratified designs with n = 53 for (α*, 1 − β*, Δ) = (0.1, 0.9, 0.15). The standard design has a fixed critical value ā = 41. The upper lines are conditional powers and the lower lines are conditional type I error.

The second part of Table 1 reports the conditional rejection value a(m1) and its {α(m1), 1− β(m1)} for each m1 ∈ [0, n]. The conditional rejection value a(m1) decreases from 44 to 39 as m1 increases. Note that ā = a(m1) = 41 for m1 values around 1 = 26.5. Figure 1(a) also displays the conditional type I error rate and power of the stratified design. While the conditional type I error of the stratified design α(m1) is closely controlled below α*, the conditional power is also well controlled around 1 − β* = 0.9. If we want 1 − β to be larger than 1 − β* for all m1 ∈ [0, n], we have to choose a slightly larger n than 53.

If the difference of the response probabilities between two cohorts |p01p02| is larger, then the range of the rejection values for the stratified design will be wider, and the conditional type I error and power of the standard design will vary more widely. Let’s consider (p01, p02) = (0.6, 0.8) and Δ1 = Δ2 = 0.15. Under (γ1, α*, 1 − β*) = (0.5, 0.1, 0.9), the standard design will be the same as above, (n, ā) = (53, 41), but the stratified rejection value a(m1) decreases from 47 to 37 as m1 increases from 0 to 53. Figure 1(b) displays the conditional type I error rate and power of the standard and stratified designs. Comparing Figures 1(a) and (b), we observe that the stratified design controls its conditional type I error rate and power closely to their nominal levels regardless of |p01p02| value, but those of the standard design change further away from their specified levels with a larger difference. We also observe that, with a larger |p01p02| value, the conditional type I error and power of the stratified design fluctuate more often because the conditional critical value changes more frequently, see Figure 1(b).

Let’s investigate the impact of an erroneously specified prevalence on the study design. Suppose that the true prevalence is γ1 = 0.3, but the study is designed under a wrong specification of γ1 = 0.5. Let’s assume (p01, p02) = (0.65, 0.75), Δ1 = Δ2 = 0.15, and (α*, 1 − β*) = (0.1, 0.9) as above. Under the erroneously specified prevalence, the standard and stratified designs will be the same as above, as shown in Table 1. The standard design has marginal type I error and power (α, 1 − β) = (0.1530, 0.9631) and the stratified design has (α, 1 − β) = (0.0767, 0.9116). Under the true γ1 = 0.3, p0 = γ1p01 + γ2p02 = 0.72 and p1 = γ1pa1 + γ2pa2 = 0.87 are farther away from 1/2 than those under the specified γ1 = 0.5, so that the marginal power for the stratified design is still larger than 1 − β* even though the marginal type I error is much below α* = 0.1. The marginal type I error for the standard design is much larger than the specified α * = 0.1. Under a wrong projection of the prevalence, the type I error of a standard design can be heavily biased, but that of the stratified design will be always controlled below α*.

Now, suppose that the true prevalence is γ1 = 0.7, but the study is designed under a erroneously specified γ1 = 0.5. In this case, the standard design has marginal type I error and power (α, 1−β) = (0.0501, 0.8209) and the stratified design has (α, 1−β) = (0.0768, 0.8762). The power for the stratified design is slightly smaller than 1−β* because of the conservative adjustment of conditional type I error. However, the power for the standard design is much smaller than 1 − β*. The impact of erroneously specified prevalence on the bias of marginal type I error and power will be larger with a larger difference between p01 and p02.

3 Two-Stage Designs

Because of ethical and economical issues, two-stage designs have been more popular for phase II cancer clinical trials than single-stage designs. We may stop a trial early when the RR of a study treatment turns out to be either too low or too high (e.g., London and Chang, 2005), but we consider the more popular design with an early stopping due to a low RR only here. If the experimental treatment is efficacious, we usually do not have a compelling ethical argument to stop the trial early and want to continue collecting more data to be used in designing a future phase III trial. Furthermore, this simplifies the computations and makes the statistical testing easier when the final sample size is different from a pre-determined one. Under a two-stage design, we accrue nk patients during stage k(= 1, 2). Let n = n1 + n2. For stage k(= 1, 2) and cohort j(= 1, 2), let Mkj and Xkj be random variables denoting the number of patients and the number of responders, respectively. Note that nk = mk1 + mk2.

3.1 Unstratified Designs

An example standard (unstratified) two-stage design is demonstrated in Section 1. Given (α*, 1−β*), a standard design (n1, n2, ā1, ā) is chosen among the two-stage designs to satisfy αα* and 1 − β ≥ 1 − β*, where α and 1 − β are obtained assuming that X1 = X11 + X12 and X2 = X21 + X22 are independent binomial random variables with probability of success p0 under H0 and pa under Ha, respectively, refer to e.g. Simon (1989) and Jung et al. (2001).

3.2 Stratified Designs

Given (M11, M21) = (m11, m21), a design (n1, n2, a1, a) has conditional type I error

α(m11,m21)=P(X11+X12>a1,X11+X12+X21+X22>ap01,p02)=x11=0m11x12=0m12x21=0m21x22=0m22I(x11+x12>a1,x11+x12+x21+x22>a)×b(x11m11,p01)b(x12m12,p02)b(x21m21,p01)b(x22m22,p02) (1)

and power

1β(m11,m21)=P(X11+X12>a1,X11+X12+X21+X22>apa1,pa2)=x11=0m11x12=0m12x21=0m21x22=0m22I(x11+x12>a1,x11+x12+x21+x22>a)×b(x11m11,pa1)b(x12m12,pa2)b(x21m21,pa1)b(x22m22,pa2). (4)

We want to find a two-stage stratified design {n1, n2, a1(m11), a(m11, m21)} whose conditional type I error is smaller than or equal to α* for each combination of (m11, m21) in mk1 ∈ [0, nk]. In order to simplify the computation associated with the search procedure, we fix (n1, n2) at the first and second stage sample sizes for a standard two-stage design based on a specified prevalence γ1, such as Simon’s (1989) minimax or optimal design, or admissible design by Jung et al. (2004). Given M11 = m11, we also propose to fix a1 = a1(m11) at [m11p01 + m12p02], where [c] denotes the largest integer not exceeding c. In other words, we reject the experimental therapy early if the observed number of responders from stage 1 is no larger than the expected number of responders under H0. Now, the only design parameter we need to choose is a, the rejection value for stage 2. Given {α*, n1, m11, n2, m21, a1(m11)}, we choose the largest a = a(m11, m21) satisfying α(m11, m21) ≤ α*. Its conditional power, 1 − β(m11, m21), is calculated by (4).

If the observed prevalence is close to the specified one (i.e., m11/n1γ1 and m21/n2γ1), then the conditional rejection values {a1(m11), a(m11, m21)} will be the same as the unstratified rejection values (ā1, ā). As in single-stage designs, the conditional power may be smaller than 1 − β* for some (m11, m21). If we want to satisfy 1 − β ≥ 1 − β* for all combinations of {(m11, m21), 0 ≤ m11n1, 0 ≤ m21n2}, then we have to choose a slightly larger n than that of a standard design.

When the true prevalence of cohort 1 is γ1, Mk1 for k = 1, 2 are independent random variables following b(nk, γ1). Given (M11, M21) = (m11, m21), let α(m11, m21) and 1 −β(m11, m21) denote the conditional type I rate and power for conditional rejection values {a1(m11), a(m11, m21)}, respectively. Then, the marginal (unconditional) type I error and power are obtained by

α=m11=0n1m21=0n2α(m11,m21)b(m11n1,γ1)b(m21n2,γ1)1β=m11=0n1m21=0n2{1β(m11,m21)}b(m11n1,γ1)b(m21n2,γ1),

respectively. In summary, a phase II trial with a stratified two-stage design is conducted as follows.

  • Step 1

    Specify (p01, p02, pa1, pa2) and (α*, 1 − β*).

  • Step 2

    Choose sample sizes for two stages (n1, n2) by:

    1. Specify γ1, the prevalence for cohort 1.

    2. For p0 = γ1p01 + γ2p02 and pa = γ1pa1 + γ2pa2, choose a standard (unstratified) two-stage design for testing
      H0:p=p0vs.Ha:p=pa

      that satisfies the (α*, 1 − β*)-condition. We use (n1, n2) for the chosen standard design as the stage 1 and 2 sample sizes of the stratified design.

  • Step 3

    After stage 1, calculate a1 = a1(m11) = [m11p01 + m12p02] based on the observed m11. We reject the therapy if x1 = x11 + x12 is smaller than or equal to a1(m11). Otherwise, we proceed to stage 2.

  • Step 4

    After stage 2, choose the maximum a = a(m11, m21) satisfying α(m11, m21) ≤ α* based on (m11, m21). Accept the therapy if x = x11 +x12 +x21 +x22 is larger than a(m11, m21).

  • Step 5

    The conditional power 1 − β(m11, m21) for a two-stage design (n1, m11, n2, m21, a1, a) is calculated by (4).

Note that the description of the whole procedure and the design parameters listed in Steps 1 and 2 should be included in the study protocol.

3.3 Example 2

Let’s consider the design setting of Example 1 with (p01, p02, Δ1, Δ2) = (0.65, 0.75, 0.15, 0.15), γ1 = 0.5 and (α*, 1−β*) = (0.1, 0.9). Under the setting, the Simon’s optimal two-stage design is given as (n1, n, ā1, ā) = (20, 59, 14, 45). We choose (n1, n) = (20, 59) for our stratified two-stage design.

Suppose that the study observed (x1, x) = (15, 45) and (m11, m21) = (14, 28). Note that much larger number of patients than expected are accrued from the high-risk group, cohort 1. By the Simon’s design, x = 45 equals ā = 45, so that the therapy will be rejected. However, the stratified critical values for (m11, m21) = (14, 28) are given as (a1, a) = (13, 44), so that, with observations (x1, x) = (15, 45), the therapy will be accepted for further investigation.

Figure 2(a) displays the conditional type I error and power of the Simon’s optimal design (marked as ‘Unstratified’) and the stratified design under the design settings. While the conditional type I error of the stratified design is closely controlled below α*, that of the unstratified design wildly fluctuates between 0.0185 and 0.3110 depending on (m11, m21). Also, the conditional power of the stratified design is closely maintained around 1 − β*, but that of the Simon’s design widely changes between 0.6447 and 0.9876. In the x-axis of Figure 2(a) (Figure 2(b) also), only m11 values are marked, but actually m21 values run from 0 to n2 = 39 between consecutive m11 values. Consequently, the conditional type I error rate and power, especially for the standard unstratified design, regularly fluctuate between consecutive m11 values.

Figure 2.

Figure 2

Conditional type I error and power of two-stage standard (unstratified) and stratified designs under (α*, 1 − β*, Δ) = (0.1, 0.9, 0.15). The unstratified design has (n1, n, ā1, ā) = (20, 59, 14, 45). The upper lines are conditional powers and the lower lines are conditional type I error.

Figure 2(b) displays the conditional type I error and power of the two designs when the two cohorts have a larger difference in RR, (p01, p02) = (0.6, 0.8), with other parameters fixed at the same values as above. Note that, with γ1 = 0.5, Simon’s optimal design will be identical to that for (p01, p02) = (0.65, 0.75). As in the single-stage design case (Figure 2(b)), we observe that the conditional type I error and power of the the unstratified design fluctuate more wildly than those with (p01, p02) = (0.65, 0.75), whereas the performance of the stratified design is almost the same.

If the true prevalence is accurately specified, then the Simon’s optimal design has marginal type I error and power of (α, 1 − β) = (0.0954, 0.9010), and the stratified design has (α, 1 − β) = (0.0792, 0.9044) if (p01, p02) = (0.65, 0.75) and (α, 1 − β) = (0.0788, 0.9159) if (p01, p02) = (0.6, 0.8). Both designs satisfy (α*, 1 − β*) = (0.1, 0.9). However, if the true prevalence of cohort 1 is γ1 = 0.3 but γ1 = 0.5 is specified in designing the study, then the Simon’s design has (α, 1 − β) = (0.1618, 0.9521) if (p01, p02) = (0.65, 0.75) and (α, 1 − β) = (0.2548, 0.9798) if (p01, p02) = (0.6, 0.8). Note that the Simon’s design has a more biased marginal type I error when two cohorts are more different in RR. On the other hand, the stratified design always controls the marginal type I error below α* and power close to 1 − β* even under an erroneously specified prevalence, e.g. (α, 1 − β) = (0.0776, 0.9203) if (p01, p02) = (0.65, 0.75) and (α, 1 − β) = (0.0782, 0.9481) if (p01, p02) = (0.6, 0.8).

4 Some Extensions

In this section, we present some extended concepts of stratified designs that are discussed above.

4.1 Conditional P-value

In the previous sections, a stratified two-stage design is determined by the sample sizes (n1, n2) and the rejection value a1, a) conditioning on the number of patients from each cohort at each stage at the design stage. When the trial is completed, however, the number of patients accrued to the study may be slightly different from the predetermined sample size. This happens since often some patients drop out or turn out to be ineligible after registration. Because of this, we usually accrue a slightly larger number of patients than the planned sample size, say 5% more. So, the total number of eligible patients at the end of a trial may be different from the planned n. In this case, sample size is a random variable, and the rejection value chosen for the planned sample size may not be valid any more. As a flexible testing method for two-stage phase II trials, we propose to calculate the p-value conditioning on the observed sample size as well as the observed prevalence from each cohort, and to reject H0 when the conditional p-value is smaller than the pre-specified α* level.

If a trial is stopped due to lack of efficacy after stage 1, then usually we are not interested in p-value calculation. Suppose that the trial has proceeded to stage 2 to observe (x1, x) together with (n1, m11, n2, m21). Then, the interim testing after stage 1 will be conducted using the rejection value a1 = [m11p01 + m12p02]. Given mkj (mk1 + mk2 = nk), Xkj ~ b(mkj, p0j) under H0. Hence, the p-value for an observation (x11, x12, x21, x22) conditioning on (n1, m11, n2, m21) is obtained by

pvalue=i11=0m11i12=0m12i21=0m21i22=0m22I(i11+i12>a1,i11+i12+i21+i22x)j=12k=12b(ikjmkj,p0j).

We reject H0 if p-value < α*. Note that the calculation of a conditional p-value does not require specification of the true prevalence. In order to avoid informative sampling issue, the final sample size should be determined without knowing the number of responders from the study.

Let’s revisit Example 2 with (p01, p02) = (0.65, 0.75). Suppose that, at the design stage, we chose (n1, n) = (20, 59) based the Simon’s optimal design, but the study accrued a slightly larger number of patients (n1, n2) = (20, 40), among whom (m11, m21) = (12, 25) were from cohort 1 and (x1, x) = (15, 46) responded. For the original sample size (n1, n) = (20, 59), the stratified rejection values are (a1, a) = (13, 45) with respect to (m11, m21) = (12, 24) or (12, 25). Hence, we could accept the therapy if the number of responders (x1, x) = (15, 46) was observed from the design as originally planned, (n1, n) = (20, 59). However, by having one more eligible patient from stage 2, it became unclear whether we should accept the therapy or not. To resolve this issue, we calculate the p-value for (x1, x) = (15, 46) conditioning on (n1, n2) = (20, 40) and (m11, m21) = (12, 25), p-value = 0.1089. The conditional p-value is marginally larger than α* = 0.1, so that we may consider accepting the therapy for further investigation.

Jung et al. (2006) propose an exact p-value calculation method for two-stage phase II designs with homogeneous patient populations.

4.2 When there are more than two cohorts

Suppose that there are J(≥ 2) cohorts with RR pj for cohort j(= 1, …, J). We consider two-stage designs here. We accrue n1 and n2 patients for stages 1 and 2, respectively. The response rates for J cohorts are specified as p0 = (p01, …, p0J ) under H0 and pa = (pa1, …, paJ ) under Ha. Let (Mk1, …, MkJ ) denote the random vector representing the numbers of patients from the J cohorts among nk patients accrued during stage k(j=1JMkj=nk,k=1,2), and (mk1, …, mkJ ) denote their observed values. Let Xkj denote the number responders among Mkj patients from cohort j during stage k. Then, given mkj, Xkj is a random variable with b(mkj, pj).

Let Mk = (Mk1, …, MkJ ) and mk = (mk1, …, mkJ ) for stage k = 1, 2. Given (m1, m2), the conditional type I error and power for chosen rejection values (a1, a) are calculated as

α(m1,m2)=P{j=1JX1j>a1,j=1J(X1j+X2j)>ap0,m1,m2}=x11=0m11x1J=0m1Jx21=0m21x2J=0m2JI{j=1Jx1j>a1,j=1J(x1j+x2j)>a}×j=1Jb(x1jm1j,p0j)b(x2jm2j,p0j)

and

1β(m1,m2)=P{j=1JX1j>a1,j=1J(X1j+X2j)>apa,m1,m2}=x11=0m11x1J=0m1Jx21=0m21x2J=0m2JI{j=1Jx1j>a1,j=1J(x1j+x2j)>a}×j=1Jb(x1jm1j,paj)b(x2jm2j,paj), (5)

respectively. A phase II trial with a stratified two-stage design is conducted as follows.

  • Step 1

    Specify p0, pa and (α*, 1 − β*).

  • Step 2

    Choose the sample sizes for two stages (n1, n2) by:

    1. Specify the prevalence for each cohort, (γ1, …, γJ ).

    2. For p0=j=1Jγjp0j and pa=j=1Jγjpaj, choose a standard (unstratified) two-stage design for testing
      H0:p=p0vs.Ha:p=pa

      that satisfies the (α*, 1−β*) condition. We choose (n1, n2) for the standard design as the stage 1 and 2 sample sizes of our stratified design.

  • Step 3

    After stage 1, calculate a1=[j=1Jm1jp0j] based on the observed m1. We reject the therapy and stop the trial if x1a1, where x1=j=1Jx1j. Otherwise, proceed to stage 2.

  • Step 4

    After stage 2, choose the maximum integer a satisfying α(m1, m2) ≤ α* based on the observed (m1, m2). We accept the therapy if x > a for further investigation, where x=k=12j=1Jx1j.

  • Step 5

    Calculate the conditional power 1 − β(m1, m2) for the two-stage testing defined by (n1, n2, m1, m2, a1, a) using (5).

Given sample sizes n1 and n2, M1 and M2 are independent multinomial random vectors with probabilities of ‘success’ for the J cohorts (γ1, …, γJ ) and n1 and n2 independent trials, respectively. The marginal type I error and power can be calculated by taking the expectations of α(M1, M2) and 1 − β(M1, M2) with respect to M1 and M2.

5 Discussions

Patient heterogeneity is an immensely important problem that arises frequently in most clinical trials. Accounting for the randomness of observed subgroup sample sizes arises in any test where patients are heterogeneous, so this issue is not limited to phase II. While this issue is well addressed in phase III clinical trials by stratified randomization, the issue in phase II trials that are traditionally conducted as single-arm trials has been widely ignored. We can avoid such issue by randomizing patients between the experimental arm and a prospective control (Jung, 2008; Rubinstein et al. 2005; Thall et al. 1989) as in phase III trials. Adopting a stratified randomization, a randomized phase II trial definitely guarantees an unbiased comparison, but it requires up to four times larger sample size compared to a single-arm trial under the same type I error rate and power.

When historical control data come from a small prior study, we may use the estimates from the prior study as the design parameters p0j, but we should incorporate the variation of these control parameters in designing a new study (Makuch and Simon, 1980; Thall and Simon, 1990). If the RR for the control is very low, e.g. 0.05 or 0.1, then we may be comfortable with a single-arm trial even with design parameters with some variation. However, for stable diseases with a larger RR, we may consider a randomized phase II trial when the estimated design parameters are unreliable.

In this paper, we assume that there exist reliable historical data for each cohort of a study population as in a standard single-arm phase II trial case. Under the assumption, we propose a stratified design method for single-arm phase II trials with heterogeneous patient populations. Rejection values are chosen to control the conditional type I error depending on the numbers of patients accrued from different subpopulations. As a result, a stratified design always controls the type I error under the pre-specified level regardless of the realized prevalence of each cohort, while its power is maintained around the pre-specified level. In contrast, the standard (unstratified) designs can have much biased type I error if the prevalence is erroneously projected or the realized prevalence is far from the true one. The bias can be more serious if the difference in RR between cohorts is larger. A stratified single-arm phase II trial does not require a larger sample size than a standard single-arm phase II trial. For each cohort, we can measure the effect size between the historical control and the experimental therapy in terms of odds ratio, i.e. paj(1 − p0j)/{p0j(1 − paj)} for cohort j. If too many patients are entered from the cohort with a smaller effect size, then the conditional power may be lower than the specified level. So, in the end of study, we may calculate the conditional power with respect to the observed number of patients from each cohort, and slightly increase the sample size for an appropriate conditional power if necessary. In this case, the proposed conditional p-value can be useful to draw the final testing result based on the revised sample size.

In designing a two-stage trial, we consider early stopping for lack of efficacy as in the usual phase II clinical trial designs (Simon, 1989). The critical values are calculated using exact binomial distributions conditioning on the number of patients from each subpopulation at each stage. London and Chang (2005) propose similar two-stage designs with early stopping boundaries for both low and high efficacy cases. The stopping values are selected based on alpha- and power-spending criteria, so that their design algorithm requires a search for the upper and lower stopping values satisfying the assigned alpha and power levels allocated to the first stage. On the other hand, we fix the lower stopping value at the expected number of responders under the null hypothesis, so that our design algorithm requires much simpler computation. Although not reporting in this paper, we compared the conditional power between our designs and those by London and Chang (2005), and found that the two methods have almost identical powers. Sposto and Gaynon (2009) propose a two-stage design with a lower stopping value only as in our designs. They derive the critical values by minimizing an ad hoc objective function based on large sample approximations. Typically phase II trials have small sample sizes, so that a statistical testing based on large sample approximations may not control the type I error rate accurately. In Example 2 with (p01, p02, Δ1, Δ2) = (0.6, 0.8, 0.15, 0.15) and α = 0.1, the Sposto and Gaynon method controls the conditional type I error rate between 0.022 and 0.126 while ours control it between 0.054 and 0.100 (see Figure 2(b)). Sposto and Gaynon (2009) method repeatedly calculates double integrals using a numerical approach, so that it takes much more computing time than our exact method requiring computation of binomial probabilities.

Acknowledgments

This research was supported by a grant from the National Cancer Institute, CA142538.

References

  1. Jung SH. Randomized phase II trials with a prospective control. Statistics in Medicine. 2008;27:568–583. doi: 10.1002/sim.2961. [DOI] [PubMed] [Google Scholar]
  2. Jung SH, Carey M, Kim KM. Graphical search for two-stage phase II clinical trials. Controlled Clinical Trials. 2001;22:367–372. doi: 10.1016/s0197-2456(01)00142-8. [DOI] [PubMed] [Google Scholar]
  3. Jung SH, Lee TY, Kim KM, George S. Admissible two-stage designs for phase II cancer clinical trials. Statistics in Medicine. 2004;23:561–569. doi: 10.1002/sim.1600. [DOI] [PubMed] [Google Scholar]
  4. Jung SH, Owzar K, George SL, Lee TY. P-value calculation for multistage phase II cancer clinical trials (with discussion) Journal of Biopharmaceutical Statistics. 2006;16:765–783. doi: 10.1080/10543400600825645. [DOI] [PubMed] [Google Scholar]
  5. London WB, Chang MN. One- and two-stage designs for stratified phase II clinical trials. Statistics in Medicine. 2005;24:2597–2611. doi: 10.1002/sim.2139. [DOI] [PubMed] [Google Scholar]
  6. Makuch RW, Simon RM. Sample size considerations for non-randomized comparative studies. Journal of Chronic Disease. 1980;33:175–181. doi: 10.1016/0021-9681(80)90017-x. [DOI] [PubMed] [Google Scholar]
  7. Rubinstein LV, Korn EL, Freidlin B, Hunsberger S, Ivy SP, Smith MA. Design issues of randomized phase II trials and a proposal for phase II screening trials. Journal of Clinical Oncology. 2005;23(28):7199–7206. doi: 10.1200/JCO.2005.01.149. [DOI] [PubMed] [Google Scholar]
  8. Simon R. Optimal two-stage designs for phase II clinical trials. Controlled Clinical Trials. 1989;10:1–10. doi: 10.1016/0197-2456(89)90015-9. [DOI] [PubMed] [Google Scholar]
  9. Sposto R, Gaynon PS. An adjustment for for patient heterogeneity in the design of two-stage phase II trials. Statistics in Medicine. 2009;28:2566–2579. doi: 10.1002/sim.3624. [DOI] [PubMed] [Google Scholar]
  10. Thall PF, Simon R. Incorporating historical control data in planning phase II clinical trials. Statistics in Medicine. 1990;9:215–228. doi: 10.1002/sim.4780090304. [DOI] [PubMed] [Google Scholar]
  11. Thall PF, Simon R, Ellenberg SS. A two-stage design for choosing among several experimental treatments and a control in clinical trials. Biometrics. 1989;45:537–547. [PubMed] [Google Scholar]
  12. Wathen JK, Thall PF, Cook JD, Estey EH. Accounting for patient heterogeneity in phase II clinical trials. Statistics in Medicine. 2008;27:2802–2815. doi: 10.1002/sim.3109. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES