Summary
In many phase I trials, the design goal is to find the dose associated with a certain target toxicity rate. In some trials, the goal can be to find the dose with a certain weighted sum of rates of various toxicity grades. For others, the goal is to find the dose with a certain mean value of a continuous response. In this article, we describe a dose-finding design that can be used in any of the dose-finding trials described above, trials where the target dose is defined as the dose at which a certain monotone function of the dose is a prespecified value. At each step of the proposed design, the normalized difference between the current dose and the target is computed. If that difference is close to zero, the dose is repeated. Otherwise, the dose is increased or decreased, depending on the sign of the difference.
Keywords: Continual reassessment method, Maximum tolerated dose, Phase I trials, Up-and-down designs
1. Introduction
Methodology for dose-finding trials in oncology is an evolving area. When cytotoxic drugs are being investigated, the primary outcome is usually toxicity, measured either as a binary or an ordinal outcome. For cytostatic drugs, the therapeutic response, measured as a continuous outcome, may be of primary interest. Cytostatic agents often target one specific process in the malignant transformation of cells and usually result in growth inhibition rather than tumor regression. Based on their specific mechanism of action, cytostatic agents are expected to have a more favorable toxicity profile (Hoekstra, Verweij, and Eskens, 2003). Because side effects are often less frequent with these agents, the dose of interest in this type of dose-finding trial may be defined by the biological activity of an agent, often measured as a continuous variable, rather than toxicity. For example, the primary endpoint in a trial of O6-benzylguanine for patients undergoing surgery for malignant glioma (Friedman et al., 1998) was inhibition of O6-alkylguanine-DNA-alkyltransferase (AGT) in brain tumors, measured as a continuous variable in fmol/mg. This inhibition was believed to decrease with dose. Several dose-finding problems where at least one of the outcomes is continuous have been previously studied in the literature (Eichhorn and Zacks, 1973; Berry et al., 2001; Bekele and Shen, 2005; Whitehead et al., 2006). Berry et al. (2001) and Bekele and Shen (2005) considered situations where the mean outcome was not necessarily a monotone function of dose. Here, we propose a design for the case where monotonicity of the outcome is assumed. Using the knowledge of monotonicity in the design of the study, and in the estimation procedure, can improve considerably the chances of selecting the correct dose at the conclusion of the study compared to situations where the assumption of monotonicity cannot be made.
The goal of most dose-finding trials for cytotoxic drugs is to find the dose that has a certain probability of dose-limiting toxicity. Sometimes the target dose is defined by using toxicity measured on an ordinal scale. Bekele and Thall (2004) presented an intuitive and flexible way to define the target dose in a trial with ordinal outcomes. The target dose was defined as the dose where a weighted sum of probabilities of the outcome categories is equal to a certain value. This effective way of aggregating different levels of toxicity is becoming popular. Ivanova (2006) described an example of a trial where an ordinal toxicity outcome was used to define the maximum tolerated dose (MTD). Bekele and Thall (2004) and Yuan, Chappell, and Bailey (2007) proposed Bayesian designs for such trials.
In this article, we will present a unified approach to dose finding in studies for which the outcome of interest is believed to be a monotone function of dose. The target dose will be defined as the dose at which the outcome of interest is equal to some prespecified value. We start in Section 2 by outlining the proposed method. In Section 3, the effect of design parameters on performance is investigated. In Sections 4, 5, and 6, we give examples of how to apply this new design to trials with continuous outcomes, ordinal toxicity outcomes, and binary outcomes. We also compare the new design with other existing methods. In Section 7, we present a discussion.
2. Notation and Methods
2.1 Defining the New Design
Let D = {d1, …, dK} denote the set of ordered dose levels selected for a trial. A subject's response at dk has distribution function , where (μ1, …, μK) and are vectors of means and variances corresponding to D. Observations from different subjects are independent. Only one observation per subject is taken. The goal is to find dose dm ∈ D such that μm = μ*. If there is no such dose, the goal is to find the dose dm with μm closest to μ*. We refer to μ* as the target value and dm as the target dose.
Subjects are assigned sequentially starting with the lowest dose. The total number of subjects is equal to M. Let n(t) = (n1(t), …, nK (t)) be the number of subjects at each of the K doses right after subject t, t ≤ M, has been assigned, that is, n1(t) + ⋯ + nK (t) = t. Let Yji be the observation from the ith subject assigned to dose dj, i = 1, …, nj (t). Let
be the sample mean and variance computed from all available observations at dj, nj (t) = 2, 3 …. Define Tj(nj(t)), nj (t) = 2, 3 …, to be the t-statistic
If sj(nj(t)) = 0, as it might be when F is discrete, Tj(nj(t)) is equal to +∞ or −∞ depending on the sign of Ȳj(nj(t)) − μ*. Subjects are assigned in cohorts or one at a time. Suppose the most recent subject t was assigned to dose dj. The next subject is assigned as follows:
(i) if Tj(nj(t)) ≤ − Δ, the next subject is assigned to dose dj +1;
(ii) if Tj(nj(t)) ≥ Δ, the next subject is assigned to dose dj −1; and
(iii) if −Δ < Tj(nj(t)) < Δ, the next subject is assigned to dose dj.
Here Δ > 0 is the design parameter. The choice of the design parameter and its implications on design performance will be investigated in Section 3.
Several authors have indicated the usefulness and necessity of a start-up rule in toxicity dose-finding studies (Ivanova et al., 2003; Cheung, 2005). We recommend assigning at least two subjects to any untried dose before the dose can be escalated.
2.2 Computing the Distribution for Subject Allocation
One of the important characteristics of a dose-finding design is the allocation distribution achieved for small and moderate sample sizes. In this section, we describe how to compute the distribution of n(M), the number of subjects assigned to each dose by the time M subjects have been assigned, for a certain dose–response scenario. Without loss of generality assume that subjects are assigned in cohorts of equal size s, making the total number of cohorts equal to M/s. Denote by rj the number of cohorts assigned to dose dj, j = 1, …, K, at the time that all but one cohort has been assigned, r1 + ⋯ + rK = R = M/s − 1. Let X(r) ∈ D be the dose assigned to cohort r, r = 1, …, R+1. Let Ωr be the outcome data accumulated after r cohorts, and let , j = 1, …, K, be the outcome data obtained at dj after r assignments. To compute the distribution of n(M) one needs to compute the probabilities of all possible paths of the process {X(r), r = 1, …, R + 1}. In dose-finding trials X(1) is usually fixed, and therefore
The decision whether to escalate, repeat, or decrease the dose depends on the data accumulated at the current dose only, and thus
(1) |
Define Υ = {1, …, R} and let Sj = {r ∈ Υ : xr = dj} be the set cohort numbers in Υ assigned to dose dj, j = 1, …, K, with elements . Because Υ = S1 υ … υ SK, and sets Sj are disjoint, the product over assignments in (1) can be replaced by a product over S1, …, SK. So we have
(2) |
.
Define the intervals corresponding to the three possible decisions to increase, repeat, or decrease the dose as C1 = (−∞, −Δ], C0 = (−Δ, Δ), and C−1 = [Δ, +∞), so that
(3) |
Substituting (3) in (2), we obtain
If F is the normal distribution function, these probabilities can be computed based on the joint distribution of sequential t-statistics, which can be obtained using the approach of Jennison and Turnbull (1991). The last step is to compute the distribution of n(M) using the probabilities of the paths {X(r), r = 1, …, R + 1}.
3. Choosing the Design Parameter
If the design parameter Δ is small, the current dose is repeated if the average response at it is very close to the target and changed otherwise. A narrow window (−Δ, Δ) results in a small probability of repeating a dose even if the true mean value of the quantity of interest at that dose is equal to μ*. In dose-finding trials it is often desirable to assign as many subjects as possible to the target dose. In this section, we will find Δ that maximizes the number of subjects assigned to the target dose and study how the choice of Δ affects design performance. The number of subjects assigned to the target dose depends on the dose–response model. We consider a set of dose–response models and compute the number of subjects assigned to the target dose averaged over the models in the set.
Consider a dose–response model where Yj ~ N(μj, σ2) with μj = μ* + (j − k)β, j = 1, …, K, that is μj − μj −1 = β for j = 2, …, K. When k = 1, …, K, we obtain a set of K dose–response scenarios with the target dose-taking values d1, …, dK. Because the operational characteristics of the design depend on β and σ only through β/σ, we set σ = 1. For each K, the number of doses, N, the total number of subjects in the trial, and β/σ, the goal is to find Δ that maximizes the number of subjects assigned to the target dose averaged over the K scenarios. The process of computing the expected number of subjects was described in Section 2.2.
We considered scenarios with four and six doses and present the case with six doses in detail. For example, suppose we are interested in finding the best Δ to use in a trial with N = 8 cohorts of size s = 3 and β/σ = 0.3. Figure 1 presents a plot of the average number of subjects assigned to the target dose for several values of Δ. The optimal value of Δ is 0.54 yielding an average of 7.86 subjects assigned to the target dose out of the total of 24 subjects. Values of Δ in the range (0, 1) yield good averages for the number of subjects assigned to the target dose (Figure 1). Figure 2 displays the optimal values of Δ for a range of values β/σ for trials with N = 8 cohorts and s = 3 subjects per cohort (solid line). The optimal value of Δ was computed using function optim in R (R Development Core Team, 2007; http://www.R-project.org). Figure 2 also displays the optimal values of Δ for trials with N = 10 and 12. The optimal Δ increases as β/σ increases and as the number of cohorts N increases. As β/σ increases, doses become more spread out, and therefore it is beneficial to increase the window (−Δ, Δ). When the total number of “moves” N is small, it is advantageous to use a small window (−Δ, Δ) to be able to “move” faster from the lowest dose to the target dose. As N becomes larger the optimal window becomes larger. To approximate the best Δ, we fitted a linear model to the data presented in Figure 2 resulting in the formula . Note that, for large values of , Δopt does not increase linearly in . When is large, computations similar to those described above should be repeated.
We investigated whether varying Δ in the course of a trial can lead to substantial improvements in design performance. We examined a design with two parameters Δ1 and Δ2. The first parameter Δ1 was used for i = 1, …, k, whereas Δ2 was used for i = k + 1, …, K. We maximized the number of subjects allocated to the target dose over (Δ1, Δ2, k). For example, for a trial with β/σ = 0.3 with eight cohorts and three patients per cohort, the optimal triple was (0.45, 1.05, 2) yielding 7.92 subjects assigned to the target dose. For comparison, a single Δ trial yielded 7.86 subjects at the target dose. In all the setups considered, the optimal pair was such that Δ1 < Δ2. It is logical to have Δ1 < Δ2, because a smaller Δ in the beginning results in a higher probability of changing the dose, thus increasing the chance of finding the target dose quickly. Because varying Δ did not lead to considerable gain, we recommend using a single Δ.
The optimal Δ depends on a dose–response scenario. We computed Δ that is optimal over a set of scenarios and is a function of β/σ. Obviously β/σ is not known before the trial to compute Δopt. Also the variances of Yk can vary over doses and doses are not equally spaced. Our goal here was to give some guidelines for selecting Δ, keeping in mind that parameter values in a certain range of optimal values yield a similar average number of subjects assigned to the target dose (see, e.g., Figure 1). Suppose that we would like to compute the best Δ for a trial with binary toxicity outcomes targeting doses with toxicity rate of Γ = 0.2. The trial will accrue N = 8 cohorts with s = 3 subjects per cohort. Assume that the difference in toxicity rates between adjacent doses is 0.12, as that at the target dose , yielding Δopt = 0.54. In general, in toxicity trials with binary outcomes σ varies from 0.3 to 0.5 and the difference in toxicity rates between adjacent doses is about 0.05–0.20, yielding a range β/σ from 0.10 to 0.67. The best Δ for trials with N = 8, 10, 12 cohorts and s = 3 will be in the range of 0.20–1.40 (Figure 2). For example, a single toxicity observed in three subjects in a trial with Γ = 0.2 yields a value of the test statistic of 0.49, whereas two out of three toxicities yield 1.71; therefore a decision rule using any Δ in (0.29,1.51), will lead to repeating the dose if one out of three toxicities are observed and decreasing if two toxicities are observed. To illustrate the design performance we used Δ = 1 in the simulation study reported in Sections 5–7 with all models and outcome types. Our simulation study confirms that this is a reasonable choice of the design parameter.
We also investigated the impact of the cohort size on the average number of subjects assigned to the target dose in trials with the total sample size of 24. For β/σ = 0.3, the trial with N = 12, s = 2, and Δopt = 0.71 yielded 8.73 subjects assigned to the target dose; the trial with N = 8, s = 3, and Δopt = 0.54 yielded 7.86 assigned to the target dose; the trial with N = 6, s = 4, and Δopt = 0.40 yielded 7.16 subjects assigned to the target dose. Smaller cohort sizes yielded more subjects assigned to the target dose on average. This was expected, especially in this example where the number of doses is relatively large (six doses) compared to the number of cohorts (6, 8, and 12 cohorts).
4. Application to Dose Finding with Continuous Outcomes
4.1 A Numerical Example
To illustrate how to apply the new design in a trial with continuous outcomes we use data from Friedman et al. (1998). The goal of the trial was to find a dose that produces undetectable AGT activity. Friedman et al. (1998) dichotomized the continuous outcome at 10 fmol/mg for the purpose of constructing the dose-escalation rule. The trial accrued 30 patients; 3, 3, 13, and 11 patients to the doses 40, 60, 80, and 100 mg/m2, respectively. As we illustrate here, it is easy to conduct the trial using a continuous outcome and therefore dichotomization is not needed. Suppose the goal is to find the dose with AGT activity equal to 5 fmol/mg protein. To construct an example of a trial with continuous outcomes, outcomes yij (Table 1) were selected randomly from the corresponding dose cohorts reported by Friedman et al. (1998). For example, y31, y32, y33 were randomly selected from 13 outcomes reported at 80 mg/m2. The hypothetical trial was stopped when the data at the 100 mg/m2 dose-cohort were exhausted. Patients with undetectable AGT activity were assigned a value of 5 fmol/mg. Because the AGT activity was believed to be decreasing with dose, the dose was escalated for positive values of the t-statistics and deescalated for negative ones. Table 1 gives the averages of AGT activity, t-statistic, and decision for each dose cohort.
Table 1.
Ȳj | t-statistic | Decision | |
---|---|---|---|
Dose 1, 40 mg/m2 | |||
(y11, y12, y13) = (26.35, 42.00, 15.00) | 27.78 | 2.91 | Increase the dose to 60 mg/m2 |
Dose 2, 60 mg/m2 | |||
(y21, y22, y23) = (23.00, 13.50, 10.83) | 15.78 | 2.92 | Increase the dose to 80 mg/m2 |
Dose 3, 80 mg/m2 | |||
(y31, y32, y33) = (11.70, 9.03, 5.00) | 8.58 | 1.84 | Increase the dose to 100 mg/m2 |
Dose 4, 100 mg/m2 | |||
(y41, y42, y43) = (4.07, 5.00, 8.70) | 5.92 | 0.65 | Repeat the dose at 100 mg/m2 |
(y44, y45, y46) = (2.50, 4.07, 6.13) | 5.08a | 0.09a | Repeat the dose at 100 mg/m2 |
(y47, y48, y49) = (3.60, 5.00, 5.00) | 4.90a | −0.18a | Repeat the dose at 100 mg/m2 |
(y410, y411) = (6.80, 6.60) | 5.22a | 0.43a | Repeat the dose at 100 mg/m2 |
Calculated based on the combined sample.
4.2 Comparison with the Feasible Sequential Procedure of Eichhorn and Zacks (1973)
For continuous outcomes, Eichhorn and Zacks (1973) proposed a feasible sequential search algorithm for finding the largest dose x such that Pr{Y (x) < μ* | x} ≥ γ, where γ is a tolerance probability. The main assumption was that response Y (x) given x is normal with mean bx and variance σ2x2, where σ2 is known and b > 0 is unknown. The dose x0 = 0 was assumed to be the dose below which the response is negligible. It was also assumed that a dose such that is below the target dose is known. Eichhorn and Zacks' sequential search algorithm operates on a continuous dose space and is from a class of feasible procedures. A procedure is feasible if, for each new assignment, the probability that the dose is higher than the target does not exceed α. Eichhorn and Zacks' sequential algorithm can be briefly described as follows. If observations y1, …, yn at doses x1, …, xn have been obtained, the next dose is
and Zγ denotes the upper 100γ% point of the standard normal distribution.
We compared our proposed design with the method of Eichhorn and Zacks (1973) for normally distributed outcomes. We set γ = 0.5, so that the problem considered by Eichhorn and Zacks (1973) is the same as the problem we consider (because the distribution of Y is symmetric). The value of is set to 1. The first comparison is made for a model where Y (x) given x has a normal distribution with mean 0.1x and variance 0.12x2 with dose levels from the set (1, 2, 3, 4, 5, 6) (model I in Table 2). For model I we considered six different values for the target, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, which correspond to the target dose being dose 1, 2, 3, 4, 5, and 6, respectively. The second comparison was made for a model where Y (x) given x is normal with mean 0.5 + 0.1x and variance 0.12x2 (model II in Table 2). For model II we considered six different target values, 0.6, 0.7, 0.8, 0.9, 1.0, 1.1, corresponding to the target dose being dose 1, 2, 3, 4, 5, and 6. To accommodate a discrete set of dose levels, the dose from the set 1, 2, 3, 4, 5, 6 closest to xn+1 is chosen in Eichhorn and Zacks' procedure. When the target dose was dose 1, a sample size of 15 was used; for all other targets a sample size of 60 was used. Simulation results were obtained by simulation with 4000 replicates.
Table 2.
Percent recommendation |
Allocation proportion |
|||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
d1 | d2 | d3 | d4 | d5 | d6 | d1 | d2 | d3 | d4 | d5 | d6 | |
Target at d1 | ||||||||||||
ND, I, II | 0.91 | 0.09 | 0.00 | 0.00 | 0.00 | 0.00 | 11.9 | 2.9 | 0.2 | 0.0 | 0.0 | 0.0 |
EZ, α = 0.05, I | 0.99 | 0.01 | 0.00 | 0.00 | 0.00 | 0.00 | 15.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 |
EZ, α = 0.5, I | 0.89 | 0.10 | 0.01 | 0.00 | 0.00 | 0.00 | 12.1 | 1.8 | 0.4 | 0.1 | 0.1 | 0.6 |
EZ, α = 0.5, II | 0.99 | 0.01 | 0.00 | 0.00 | 0.00 | 0.00 | 14.8 | 0.2 | 0.0 | 0.0 | 0.0 | 0.0 |
Target at d2 | ||||||||||||
ND, I, II | 0.04 | 0.89 | 0.07 | 0.00 | 0.00 | 0.00 | 8.8 | 39.2 | 11.0 | 1.0 | 0.0 | 0.0 |
EZ, α = 0.05, I | 0.08 | 0.90 | 0.02 | 0.00 | 0.00 | 0.00 | 28.6 | 31.2 | 0.2 | 0.0 | 0.0 | 0.0 |
EZ, α = 0.5, I | 0.00 | 0.93 | 0.07 | 0.00 | 0.00 | 0.00 | 5.2 | 45.0 | 7.2 | 1.1 | 0.4 | 1.1 |
EZ, α = 0.5, II | 0.90 | 0.10 | 0.00 | 0.00 | 0.00 | 0.00 | 0.0 | 54.4 | 5.5 | 0.2 | 0.0 | 0.0 |
Target at d3 | ||||||||||||
ND, I, II | 0.01 | 0.12 | 0.73 | 0.13 | 0.01 | 0.00 | 2.4 | 11.6 | 33.0 | 11.2 | 1.6 | 0.2 |
EZ, α = 0.05, I | 0.00 | 0.32 | 0.64 | 0.04 | 0.00 | 0.00 | 4.3 | 38.6 | 16.5 | 0.5 | 0.0 | 0.0 |
EZ, α = 0.5, I | 0.00 | 0.06 | 0.80 | 0.13 | 0.01 | 0.00 | 1.4 | 9.5 | 35.0 | 9.9 | 2.0 | 2.2 |
EZ, α = 0.5, II | 0.64 | 0.29 | 0.07 | 0.00 | 0.00 | 0.00 | 39.1 | 17.4 | 3.4 | 0.1 | 0.0 | 0.0 |
Target at d4 | ||||||||||||
ND, I, II | 0.00 | 0.04 | 0.21 | 0.59 | 0.14 | 0.12 | 2.0 | 3.5 | 15.0 | 26.8 | 10.5 | 2.2 |
EZ, α = 0.05, I | 0.00 | 0.00 | 0.47 | 0.46 | 0.06 | 0.00 | 1.9 | 9.1 | 37.6 | 10.6 | 0.0 | 0.0 |
EZ, α = 0.5, I | 0.00 | 0.00 | 0.13 | 0.67 | 0.12 | 0.02 | 1.1 | 1.2 | 12.9 | 27.8 | 11.0 | 6.1 |
EZ, α = 0.5, II | 0.08 | 0.32 | 0.52 | 0.09 | 0.00 | 0.00 | 9.0 | 24.7 | 22.9 | 3.2 | 0.1 | 0.0 |
Target at d5 | ||||||||||||
ND, I, II | 0.00 | 0.01 | 0.05 | 0.28 | 0.48 | 0.18 | 2.0 | 2.4 | 5.2 | 17.3 | 22.3 | 10.9 |
EZ, α = 0.05, I | 0.00 | 0.00 | 0.02 | 0.53 | 0.36 | 0.08 | 1.3 | 3.0 | 15.7 | 31.2 | 7.8 | 1.0 |
EZ, α = 0.5, I | 0.00 | 0.00 | 0.00 | 0.19 | 0.57 | 0.25 | 1.1 | 0.3 | 2.2 | 14.4 | 22.9 | 19.1 |
EZ, α = 0.5, II | 0.00 | 0.01 | 0.34 | 0.56 | 0.08 | 0.01 | 1.6 | 7.9 | 27.2 | 20.2 | 2.9 | 0.2 |
Target at d6 | ||||||||||||
ND, I, II | 0.00 | 0.00 | 0.03 | 0.06 | 0.31 | 0.60 | 2.0 | 2.1 | 3.1 | 7.2 | 18.2 | 27.4 |
EZ, α = 0.05, I | 0.00 | 0.00 | 0.00 | 0.08 | 0.52 | 0.40 | 1.1 | 1.4 | 5.4 | 20.7 | 24.1 | 7.3 |
EZ, α = 0.5, I | 0.00 | 0.00 | 0.00 | 0.00 | 0.24 | 0.76 | 1.0 | 0.1 | 0.7 | 3.4 | 14.8 | 40.0 |
EZ, α = 0.5, II | 0.00 | 0.00 | 0.01 | 0.41 | 0.49 | 0.09 | 1.1 | 3.0 | 10.7 | 26.6 | 15.7 | 2.9 |
Table 2 displays the proportion of runs each dose was recommended as the target dose and the dose allocation. Eichhorn and Zacks (1973) recommended using a feasibility level of α = 0.05, because the procedure was proposed for an outcome reflecting very poor toxicity. When α = 0.05, the procedure is conservative and yields low percentages of correct recommendation (lines titled EZ, α = 0.05, I) when the target is at doses 4, 5, and 6. For that reason, we also simulated Eichhorn and Zacks' procedure with α = 0.5 (lines titled EZ, α = 0.5, I). When α = 0.5 and γ = 0.5, the formula for next dose assignment simplifies to , that is, no feasibility restrictions on assignments are imposed. As expected, when the true dose–response model is linear in dose the procedure of Eichhorn and Zacks performs very well (lines titled EZ, α = 0.5, I). However, for model II where the mean is 0.5 + 0.1x the procedure of Eichhorn and Zacks does not perform well (lines titled EZ, α = 0.5, II). The main conclusion from this simulation study is that in the case of linear dose response with no intercept, our new method performed reasonably well detecting the correct target dose in 9% fewer cases than the method of Eichhorn and Zacks that was specifically designed for this dose–response relationship (compare lines ND, I, II, and EZ, α = 0.5, I). The procedure of Eichhorn and Zacks performed poorly for model II. At the same time, our method is robust to the shift in the mean and performs exactly the same in models I and II.
The estimate of the target dose in our method was obtained from the isotonic estimate of the dose–response curve, whereas in Eichhorn and Zacks' procedure the estimate was obtained from fitting a linear model without an intercept to the data. In simulations with model I, if a linear model estimate was used in our method, the percentages of correct selection were nearly identical to these of Eichhorn and Zacks' procedure. The reverse was also true. This suggests that for model I both methods provide good quality data for estimation with many assignments at and around the target dose.
5. Application to Dose-Finding Trials with Ordinal Outcomes
We applied our method to a trial with ordinal toxicity outcome described in Bekele and Thall (2004). The outcome of interest is a toxicity score whose expected value is a weighted sum of rates of different toxicity grades and types. The goal was to estimate the dose with mean toxicity score 3.04. The weights and the target value were determined by physicians and we refer the reader to Bekele and Thall (2004) for the description of their Bayesian design. Thirty-six subjects were assigned in groups of four. The starting dose was d2. We simulated the new design for all six scenarios considered in Bekele and Thall (2004) and show the results of these simulations in Table 3. Results for the new design were obtained by simulation with 4000 replicates. To estimate the target dose after the trial for the new design, we first obtained the isotonic estimate of the mean function. The dose with the estimated value of the function closest to 3.04 was declared the estimated target dose. If there were two or more such doses, the highest dose with the estimated value below 3.04 was chosen. If all the estimated values at these doses were higher than 3.04, the lowest of these doses was chosen.
Table 3.
d1 | d2 | d3 | d4 | d5 | d6 | d7 | d8 | d9 | d10 | ||
---|---|---|---|---|---|---|---|---|---|---|---|
Scenario 1 | |||||||||||
ND | Proportion recommended | 0.00 | 0.86 | 0.14 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |
Allocation | 1.2 | 17.3 | 12.5 | 5.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | |
BTD | Proportion recommended | 0.01 | 0.94 | 0.05 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |
Allocation | 2.2 | 21.8 | 5.3 | 5.6 | 1.1 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | |
Scenario 2 | |||||||||||
ND | Proportion recommended | 0.04 | 0.95 | 0.01 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |
Allocation | 5.7 | 18.0 | 8.0 | 4.3 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | |
BTD | Proportion recommended | 0.04 | 0.92 | 0.04 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |
Allocation | 4.5 | 22.0 | 4.5 | 4.6 | 0.4 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | |
Scenario 3 | |||||||||||
ND | Proportion recommended | 0.00 | 0.00 | 0.00 | 0.00 | 0.85 | 0.15 | 0.00 | 0.00 | 0.00 | 0.00 |
Allocation | 0.0 | 0.0 | 0.0 | 8.2 | 20.6 | 6.8 | 0.4 | 0.0 | 0.0 | 0.0 | |
BTD | Proportion recommended | 0.00 | 0.00 | 0.00 | 0.06 | 0.85 | 0.09 | 0.00 | 0.00 | 0.00 | 0.00 |
Allocation | 0.0 | 0.0 | 0.0 | 5.6 | 18.5 | 9.7 | 2.1 | 0.1 | 0.0 | 0.0 | |
Scenario 4 | |||||||||||
ND | Proportion recommended | 0.00 | 0.00 | 0.00 | 0.03 | 0.94 | 0.03 | 0.00 | 0.00 | 0.00 | 0.00 |
Allocation | 0.0 | 0.0 | 0.1 | 11.3 | 20.6 | 4.0 | 0.0 | 0.0 | 0.0 | 0.0 | |
BTD | Proportion recommended | 0.00 | 0.00 | 0.00 | 0.05 | 0.87 | 0.08 | 0.00 | 0.00 | 0.00 | 0.00 |
Allocation | 0.0 | 0.0 | 0.1 | 6.6 | 20.6 | 7.7 | 1.0 | 0.0 | 0.0 | 0.0 | |
Scenario 5 | |||||||||||
ND | Proportion recommended | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.02 | 0.07 | 0.85 | 0.06 | 0.00 |
Allocation | 0.0 | 0.0 | 0.0 | 4.3 | 4.4 | 4.7 | 7.9 | 11.8 | 2.8 | 0.1 | |
BTD | Proportion recommended | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.10 | 0.81 | 0.09 | 0.00 |
Allocation | 0.0 | 0.0 | 0.0 | 4.1 | 4.1 | 4.1 | 5.7 | 11.4 | 5.6 | 1.0 | |
Scenario 6 | |||||||||||
ND | Proportion recommended | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.01 | 0.06 | 0.90 | 0.03 | 0.00 |
Allocation | 0.0 | 0.0 | 0.0 | 4.1 | 4.1 | 4.4 | 7.9 | 12.9 | 2.6 | 0.0 | |
BTD | Proportion recommended | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.14 | 0.80 | 0.06 | 0.00 |
Allocation | 0.0 | 0.0 | 0.0 | 4.0 | 4.0 | 4.1 | 5.7 | 11.7 | 5.8 | 0.7 |
The methods were compared with respect to the percentage of time the target dose was correctly selected and the proportion of subjects assigned to each dose. A comparison of the new design with Δ = 1 with the design of Bekele and Thall (2004) slightly favors the new design overall. The new design yields a higher probability of correctly selecting the target dose in four out of six scenarios. More importantly, it assigns fewer patients to doses higher than the target dose in four out of six scenarios.
Some outcomes, for example, myelosuppression, were measured on a continuous scale. If our method is used, there is no need to convert a continuous outcome to an ordinal one to score the outcome. Instead, the continuous outcome can be scored, and this and scores from ordinal and binary variables can be used to compute the average score for a patient.
6. Note on the Application to Toxicity Studies
The new design was developed for trials with ordinal or continuous outcomes. However, it can be used with binary outcomes. Numerous designs have been developed for dose-finding trials with binary outcomes (Wetherill, 1963; O'Quigley, Pepe, and Fisher, 1990; Babb, Rogatko, and Zacks, 1998). The continual reassessment method (CRM) has been shown to perform very well in dose finding (O'Quigley et al., 1990; O'Quigley and Chevret, 1991; Faries, 1994). Cheung and Chappell (2000) simulated the CRM for five scenarios from O'Quigley et al. (1990). We compared the new design with the CRM using all five scenarios. The goal was to estimate the dose with the probability of toxicity of Γ = 0.2. Subjects were assigned one at a time starting with the lowest dose. The total number of subjects was either 25 or 48. As suggested in Section 2, at least three subjects were assigned to any untried dose before it was escalated. Estimation of the target dose after the trial was carried out as described in Section 5. We simulated the new design with Δ = 1 and show the results of these simulations in Table 4. Results were obtained by simulation with 4000 replicates. As seen from Table 4, the performance of the new design is worse than that of the CRM for scenarios 1, 3, and 4; however, the new design performs substantially better in scenario 5. In scenario 5, the CRM fails to converge to the target dose (Cheung and Chappell, 2002) and therefore its performance is poor in this scenario. Allocations of subjects as far as overdosing are comparable for the new design and the CRM across all five scenarios.
Table 4.
Percent recommendation |
Allocation |
|||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
d1 | d2 | d3 | d4 | d5 | d6 | d1 | d2 | d3 | d4 | d5 | d6 | |
Scenario 1 | 0.05 | 0.10 | 0.20 | 0.30 | 0.50 | 0.70 | ||||||
n = 25 | ||||||||||||
ND | 0.06 | 0.20 | 0.45 | 0.26 | 0.03 | 0.00 | 5.1 | 7.4 | 7.8 | 3.8 | 0.9 | 0.1 |
CRM | 0.01 | 0.20 | 0.49 | 0.29 | 0.02 | 0.00 | 4.0 | 6.4 | 8.6 | 5.2 | 0.7 | 0.0 |
n = 48 | ||||||||||||
ND | 0.01 | 0.16 | 0.56 | 0.26 | 0.01 | 0.00 | 5.2 | 11.5 | 18.8 | 10.5 | 1.9 | 0.1 |
CRM | 0.00 | 0.14 | 0.63 | 0.23 | 0.00 | 0.00 | 4.2 | 10.3 | 21.3 | 11.4 | 0.8 | 0.0 |
Scenario 2 | 0.30 | 0.40 | 0.52 | 0.61 | 0.76 | 0.87 | ||||||
n = 25 | ||||||||||||
ND | 0.91 | 0.08 | 0.01 | 0.00 | 0.00 | 0.00 | 20.8 | 3.7 | 0.5 | 0.0 | 0.0 | 0.0 |
CRM | 0.92 | 0.07 | 0.01 | 0.00 | 0.00 | 0.00 | 20.5 | 3.6 | 1.8 | 0.1 | 0.0 | 0.0 |
n = 48 | ||||||||||||
ND | 0.98 | 0.02 | 0.00 | 0.00 | 0.00 | 0.00 | 42.5 | 4.8 | 0.6 | 0.1 | 0.0 | 0.0 |
CRM | 0.98 | 0.02 | 0.00 | 0.00 | 0.00 | 0.00 | 42.5 | 4.6 | 0.8 | 0.1 | 0.0 | 0.0 |
Scenario 3 | 0.05 | 0.06 | 0.08 | 0.11 | 0.19 | 0.34 | ||||||
n = 25 | ||||||||||||
ND | 0.03 | 0.05 | 0.12 | 0.30 | 0.34 | 0.16 | 4.8 | 5.0 | 5.0 | 4.7 | 3.7 | 1.8 |
CRM | 0.00 | 0.03 | 0.12 | 0.39 | 0.41 | 0.05 | 3.7 | 4.0 | 5.2 | 6.7 | 4.4 | 1.0 |
n = 48 | ||||||||||||
ND | 0.00 | 0.01 | 0.04 | 0.20 | 0.56 | 0.18 | 4.8 | 5.4 | 6.9 | 9.6 | 13.5 | 7.7 |
CRM | 0.00 | 0.01 | 0.05 | 0.40 | 0.51 | 0.04 | 3.7 | 4.5 | 7.0 | 15.8 | 15.1 | 1.9 |
Scenario 4 | 0.06 | 0.08 | 0.12 | 0.18 | 0.40 | 0.71 | ||||||
n = 25 | ||||||||||||
ND | 0.05 | 0.10 | 0.25 | 0.46 | 0.14 | 0.00 | 5.4 | 5.8 | 6.11 | 5.2 | 2.3 | 0.3 |
CRM | 0.01 | 0.07 | 0.26 | 0.54 | 0.12 | 0.00 | 4.0 | 4.8 | 6.7 | 7.4 | 2.1 | 0.2 |
n = 48 | ||||||||||||
ND | 0.00 | 0.04 | 0.19 | 0.64 | 0.12 | 0.00 | 5.7 | 7.3 | 10.5 | 16.2 | 7.7 | 0.7 |
CRM | 0.00 | 0.03 | 0.22 | 0.68 | 0.07 | 0.00 | 4.1 | 5.9 | 12.0 | 21.7 | 4.1 | 0.2 |
Scenario 5 | 0.00 | 0.00 | 0.03 | 0.05 | 0.11 | 0.22 | ||||||
n = 25 | ||||||||||||
ND | 0.00 | 0.00 | 0.01 | 0.10 | 0.34 | 0.56 | 3.0 | 3.1 | 4.0 | 4.5 | 5.3 | 5.1 |
CRM | 0.00 | 0.00 | 0.00 | 0.10 | 0.61 | 0.29 | 3.0 | 3.0 | 3.2 | 4.9 | 7.5 | 3.4 |
n = 48 | ||||||||||||
ND | 0.00 | 0.00 | 0.00 | 0.01 | 0.28 | 0.71 | 3.0 | 3.1 | 3.9 | 5.2 | 11.6 | 21.2 |
CRM | 0.00 | 0.00 | 0.00 | 0.02 | 0.60 | 0.38 | 3.0 | 3.0 | 3.2 | 6.1 | 21.6 | 11.1 |
7. Discussion
In this article, we present a unified approach to dose-finding trials with a monotone objective function. The method is general and can be used with a variety of response types. Possible applications of our proposed methodology include trials with continuous and ordinal outcomes. Also, the design can be used in trials with bivariate outcomes, as in Ivanova (2003), as long as the function of interest is monotone in dose.
In this article, we have assumed that the mean response was monotone in dose. We used isotonic estimates of the mean responses at the end of the trial. One can compute the iso-tonic estimates of the mean responses at each step, and then use the isotonic estimate at the current dose in place of a sample mean. Using isotonic estimates improves the design performance somewhat for binary and ordinal outcomes, and more so for continuous outcomes, especially if the variability of the outcome is high. Leung and Wang (2001) used isotonic estimates of toxicity rates in their design, but proposed a different decision rule. To make an assignment, they considered the current dose, as well as the two adjacent doses, and then selected the dose with the estimated mean response closest to the target. Let us consider an example of a trial where the current dose is dose dj and dose dj+1 has two subjects assigned, with one of them having toxic response and the other having a nontoxic response. The probability of observing one or more toxicities out of two assignments if the true toxicity rate is 0.2 is rather high (0.36). If the target toxicity rate is 0.2, Leung and Wang's (2001) decision rule would not assign any more subjects to dj+1 and higher dose levels. This would result in a very conservative assignment rule. Ivanova and Flournoy (2008) have compared several designs based on the isotonic estimation of dose-toxicity curve in trials with binary outcomes. Their simulation results gave strong evidence that the decision rule in an isotonic design, which uses the current dose only, is far superior to decision rules that use more doses.
On the flip side, the question of how to handle cases where the lowest dose is very safe but the next higher dose is unsafe, needs to be considered. Our proposed design would oscillate between the two doses, thereby possibly exposing subjects to a highly toxic dose. This problem is a problem of the initial incorrect choice of dose levels, and as such, it is hard to remedy with any design. For trials where there is a danger of assigning subjects to highly toxic doses, we suggest putting an additional rule in place that prevents assigning subjects to a dose if there is evidence that its mean toxicity score is higher than allowable. One can then use the sequential probability ratio test, or some other type of sequential stopping boundary. For example, Ivanova, Qaqish, and Schell (2005) described how to set up a boundary for continuous monitoring of toxicity, where the toxicity outcome was binary. Such stopping boundaries are especially useful in cases where the lowest dose appears to be highly toxic and therefore these boundaries should be used routinely in conjunction with either the CRM, or our proposed design, in oncology dose-finding trials. Another way to avoid rapid escalation when using our proposed design, which has been suggested by the associate editor, is to use an asymmetric interval in the decision rule. Here, the dose is repeated if the test statistic is in the interval (−Δ1, Δ2) with Δ1 ⪢ Δ2, and is changed if it is not.
In the examples presented in this article, the design parameter was chosen before the trial. However, and especially in the case of continuous outcomes, it might be worth tuning the design parameter during the trial as information on the spacing between doses, and the variance of the outcome, becomes available. The optimal parameter would be chosen to maximize the number of subjects assigned to the target dose. It might be even more important to maximize the proportion of trials that select the target dose correctly. Our extensive simulation study shows that the same, or nearly the same parameter Δ, maximizes the proportion of correct recommendations. (Simulation study results are available from the authors on request.) In some trials, the goal might be to estimate a dose with mean response μ*, where the estimate d* is not necessarily in D (Ivanova, Bolognese, and Perevozskaya, 2008). Because d* is estimated by interpolation, Δ = 0 would work well because it maximizes the assignments to the two doses adjacent to d*.
Acknowledgements
This work was supported in part by NIH grant RO1 CA120082-01A1. The authors thank Steve Coad and Dominic Moor for helpful comments and discussions. The authors thank an associate editor and an anonymous reviewer for helpful comments.
References
- Babb J, Rogatko A, Zacks S. Cancer phase I clinical trials: Efficient dose escalation with overdose control. Statistics in Medicine. 1998;17:1103–1120. doi: 10.1002/(sici)1097-0258(19980530)17:10<1103::aid-sim793>3.0.co;2-9. [DOI] [PubMed] [Google Scholar]
- Bekele BN, Shen Y. A Bayesian approach to jointly modeling toxicity and biomarker expression in a phase I/II dose-finding trial. Biometrics. 2005;61:344–354. doi: 10.1111/j.1541-0420.2005.00314.x. [DOI] [PubMed] [Google Scholar]
- Bekele BN, Thall PT. Dose-finding based on multiple toxicities in a soft tissue sarcoma trial. Journal of the American Statistical Association. 2004;99:26–35. [Google Scholar]
- Berry DA, Müller P, Grieve AP, Smith M, Parke T, Blazek R, Mitchard N, Krams M. Adaptive Bayesian designs for dose-ranging drug trials. In: Gatsonis C, Kass RE, Carlin B, Carriquiry A, Gelman A, Verdinelli I, West M, editors. Case Studies in Bayesian Statistics V. Springer-Verlag; New York: 2001. pp. 99–181. [Google Scholar]
- Cheung YK. Coherence principles in dose finding studies. Biometrika. 2005;92:863–873. [Google Scholar]
- Cheung YK, Chappell R. Sequential designs for phase I clinical trials with late-onset toxicities. Biometrics. 2000;56:1177–1182. doi: 10.1111/j.0006-341x.2000.01177.x. [DOI] [PubMed] [Google Scholar]
- Cheung YK, Chappell R. A simple technique to evaluate model sensitivity in the continual reassessment method. Biometrics. 2002;58:671–674. doi: 10.1111/j.0006-341x.2002.00671.x. [DOI] [PubMed] [Google Scholar]
- Eichhorn BH, Zacks S. Sequential search of an optimal dosage, I. Journal of the American Statistical Association. 1973;68:594–598. [Google Scholar]
- Faries D. Practical modifications of the continual reassessment method for phase I cancer clinical trials. Journal of Biopharmaceutical Statistics. 1994;4:147–164. doi: 10.1080/10543409408835079. [DOI] [PubMed] [Google Scholar]
- Friedman HS, Kokkinakis DM, Pluda J, Friedman AH, Cokgor I, Haglung MM, Ashley D, Rich J, Dolan ME, Pegg A, Moschel RC, Mclendon RE, Kerby T, Herndon JE, Bigner DD, Schold SC. Phase I trial of O6-benzylguanine for patients undergoing surgery for malignant glioma. Journal of Clinical Oncology. 1998;16:3570–3575. doi: 10.1200/JCO.1998.16.11.3570. [DOI] [PubMed] [Google Scholar]
- Hoekstra R, Verweij J, Eskens FA. Clinical trial design for target specific anticancer agents. Investigational New Drugs. 2003;21:243–250. doi: 10.1023/a:1023581731443. [DOI] [PubMed] [Google Scholar]
- Ivanova A. A new dose-finding design for bivariate outcomes. Biometrics. 2003;59:1003–1009. doi: 10.1111/j.0006-341x.2003.00115.x. [DOI] [PubMed] [Google Scholar]
- Ivanova A. Escalation, up-and-down and A+B designs for dose-finding trials. Statistics in Medicine. 2006;25:3668–3678. doi: 10.1002/sim.2470. [DOI] [PubMed] [Google Scholar]
- Ivanova A, Flournoy N. Comparison of isotonic designs for dose-finding. Statistics in Biopharmaceutical Research. 2008 doi: 10.1198/sbr.2009.0010. in press. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ivanova A, Haghighi AM, Mohanti SG, Durham SD. Improved up-and-down designs for phase I trials. Statistics in Medicine. 2003;22:69–82. doi: 10.1002/sim.1336. [DOI] [PubMed] [Google Scholar]
- Ivanova A, Qaqish BF, Schell MJ. Continuous toxicity monitoring in phase I trials in oncology. Biometrics. 2005;61:540–545. doi: 10.1111/j.1541-0420.2005.00311.x. [DOI] [PubMed] [Google Scholar]
- Ivanova A, Bolognese J, Perevozskaya I. Adaptive design based on t-statistic for dose-response trials. Statistics in Medicine. 2008;27:1581–1592. doi: 10.1002/sim.3209. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jennison C, Turnbull BW. Exact calculations for sequential t, χ2, and F tests. Biometrika. 1991;78:133–141. [Google Scholar]
- Leung DH, Wang YG. Isotonic designs for phase I trials. Controlled Clinical Trials. 2001;22:126–138. doi: 10.1016/s0197-2456(00)00132-x. [DOI] [PubMed] [Google Scholar]
- O'Quigley J, Chevret S. Methods for dose-finding studies in cancer clinical trials: A review and results of a Monte Carlo study. Statistics in Medicine. 1991;10:1647–1664. doi: 10.1002/sim.4780101104. [DOI] [PubMed] [Google Scholar]
- O'Quigley J, Pepe M, Fisher L. Continual reassessment method: A practical design for phase I clinical trials in cancer. Biometrics. 1990;46:33–48. [PubMed] [Google Scholar]
- R Development Core Team R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. 2007 Available at http://www.R-project.org.
- Wetherill GB. Sequential estimation of quantal response curves. Journal of the Royal Statistical Society, Series B. 1963;25:1–48. [Google Scholar]
- Whitehead J, Zhou Y, Mander A, Ritchie S, Sabin A, Wright A. An evaluation of Bayesian designs for dose-escalation studies in healthy volunteers. Statistics in Medicine. 2006;25:433–445. doi: 10.1002/sim.2213. [DOI] [PubMed] [Google Scholar]
- Yuan Z, Chappell R, Bailey H. The continual reassessment method for multiple toxicity grades: A Bayesian quasi-likelihood approach. Biometrics. 2007;63:173–179. doi: 10.1111/j.1541-0420.2006.00666.x. [DOI] [PubMed] [Google Scholar]