Skip to main content
Taylor & Francis Open Select logoLink to Taylor & Francis Open Select
. 2018 Jan 5;48(3):616–627. doi: 10.1080/03610926.2017.1417429

Conditionally unbiased estimation in the normal setting with unknown variances

David S Robertson a,, Ekkehard Glimm b,c
PMCID: PMC6540744  PMID: 31217751

ABSTRACT

To efficiently and completely correct for selection bias in adaptive two-stage trials, uniformly minimum variance conditionally unbiased estimators (UMVCUEs) have been derived for trial designs with normally distributed data. However, a common assumption is that the variances are known exactly, which is unlikely to be the case in practice. We extend the work of Cohen and Sackrowitz (Statistics & Probability Letters, 8(3):273-278, 1989), who proposed an UMVCUE for the best performing candidate in the normal setting with a common unknown variance. Our extension allows for multiple selected candidates, as well as unequal stage one and two sample sizes.

KEYWORDS: Selection bias, Two-stage sample, Uniformly minimum variance conditionally unbiased estimation

MATHEMATICS SUBJECT CLASSIFICATION: 62-07

1. Introduction

Two-stage adaptive trial designs offer an efficient way of selecting and validating multiple candidate treatments within a single trial. A common strategy is to select the best performing treatment (according to some ranking criteria) after an interim analysis, and to then validate its properties in an independent sample in stage 2.

However, selecting and ranking candidates in this way can induce bias into the naïve estimates that combine data from both stages. If the selection rules are not properly taken into account by the estimation strategy, then intuitively one might expect overly optimistic estimates of the performance of the selected candidate, given that it had to perform ‘well’ in stage 1 in order to proceed to stage 2.

In order to efficiently and completely correct for this selection bias, the technique of Rao-Blackwellisation can be used, where the unbiased stage 2 data is conditioned on a complete, sufficient statistic. The resulting estimator is the uniformly minimum variance conditionally unbiased estimator (UMVCUE). An appealing feature of the UMVCUE is that, as the name suggests, it has the smallest variance (or equivalently, mean squared error (MSE)) amongst all possible unbiased estimators.

This two-stage estimation framework was introduced by Cohen and Sackrowitz (1989), who derived the UMVCUE for normally distributed data. Their work has since been extended to a variety of other trial settings with normal (or asymptotically normal) data. Bowden and Glimm (2008) extended the UMVCUE to apply to unequal stage one and two sample sizes, and when the parameter of interest belongs to the j-th best candidate out of k. In related work, Kimani, Todd, and Stallard (2013) derived UMVCUEs for the means of the selected and control treatments in the seamless phase II/III trial setting with early stopping. Meanwhile, Bowden and Dudbridge (2009) derived the UMVCUE for a two-stage genome-wide association study with ranking based on p-values. In a recent development, Robertson, Prevost, and Bowden (2016a, 2016b) generalised all of these approaches to normal data with an arbitrary correlation structure.

However, apart from the original Cohen and Sackrowitz paper, all of the above analyses assume that the variances of the parameter of interest are known exactly. This may not be a reasonable assumption to make. In practice, the variance of a treatment effect (say) will often be estimated directly from the data of the trial itself. Alternatively, the variance will be based on the results from a previous trial or pilot study on the same treatments.

Cohen and Sackrowitz did derive the UMVCUE for the mean of the selected normal population with a common unknown variance. However, the estimator is only valid for the highest ranked population when the stage 1 sample sizes are all equal, and the stage 2 sample size is equal to one. In this paper, we aim to address these limitations.

In Section 2 we present the model framework and the (corrected) form of the UMVCUE given by Cohen and Sackrowitz (1989). We extend their approach in Section 3, and in Section 4 compare the resulting UMVCUE with the one derived by Bowden and Glimm (2008) that assumes a known variance. A case study based on the INHANCE trial is presented in Section 5, and we conclude with a discussion in Section 6.

2. Equal sample sizes

To start with, we consider the setting of Cohen and Sackrowitz (1989) with a common unknown variance. Suppose there are k experimental treatments, with each tested on n subjects in stage 1. Let the stage 1 data Xij, i = 1, 2, …, k; j = 1, 2, …, n, be normally distributed with means μi and common unknown variance σ2. Denote the stage 1 sample means by X¯1,X¯2,,X¯k.

At the end of stage 1, the treatment with the largest sample mean is selected for confirmatory analysis in stage 2. Let Y be a single observation taken from the highest ranked treatment group in stage 2. Also let Q be the event {X:X¯1>X¯2>>X¯k}. Without loss of generality, we condition on Q, as this can be viewed as simply a relabelling of the treatments.

Cohen and Sackrowitz derived the following UMVCUE for the mean μ1 of the highest ranked treatment:

Theorem 2.1.

The UMVCUE of μ1 given Q is

Zn+1-nn+1×S˜(1-r_2)c22ccB(c,c)Fβ(c,c)12(r_+1) (1)

where

Z=nX¯1+Y,c=k(n-1)/2S2=i=1kj=1nXij2+Y2S˜2=S2-(n+1)Zn+12-ni=2kX¯i2r=n(n+1)S˜Zn+1-X¯2,r_=min(r,1)B(a,b)=Γ(a)Γ(b)Γ(a+b)Fβ(v1,v2)isthecdfofaBetadistributionwithparametersv1,v2.

Remark.

Note that there are two errors in the formula for the UMVCUE as presented in Cohen and Sackrowitz (1989). Firstly, the summation in the definition of S˜2 should start from i = 2, and not i = 1. In addition, the denominator of the second term of the UMVCUE should have a factor of 22c, and not 2c + 1.

3. Extending the UMVCUE

A natural extension to the setting of Cohen and Sackrowitz is to consider unequal sample sizes. Suppose now that treatment i is tested on ni subjects in stage 1. The stage 1 data Xij are again normally distributed with mean μi and common unknown variance σ2. The stage 1 sample means X¯i are given by X¯i=1nijniXij. We also now allow there to be more than one subject in stage 2.

Sometimes the properties of treatments that were not the highest ranked are of interest. Hence an additional extension is to find the UMVCUE of the mean of the l-th ranked treatment (l = 1, …, k). To this end, let Y¯l be the mean of ml additional observations (denoted by Ylj) taken from the l-th ranked treatment group in stage 2.

Again conditioning on Q, we have the following UMVCUE for the mean μl.

Theorem 3.1.

The UMVCUE of μl given Q is

Zlnl+ml-nlml(nl+ml)×S˜l(1-r_l2)c-(1-q_l2)c22ccB(c,c)Fβ(c,c)12(r_l+1)-Fβ(c,c)12(q_l+1) (2)

where

Zl=nlX¯l+mlY¯l,c=(N-k)/2,N=i=1kniSl2=i=1kj=1niXij2+mlY¯l2S˜l2=Sl2-(nl+ml)Zlnl+ml2-ilniX¯i2rl=nl(nl+ml)/mlS˜lZlnl+ml-X¯l+1,r_l=min(rl,1)forl=1,,k-1ql=nl(nl+ml)/mlS˜lZlnl+ml-X¯l-1,q_l=max(ql,-1)forl=2,,k

and we define q 1 = −1 and rk = 1.

Proof.

The proof is similar to Cohen and Sackrowitz’s. We let

Ul=ml(nl+ml)nl×Y¯l-Zlnl+mlS˜l.

The joint density of (Xij,Y¯l) given Q is

K(μ,σ2)IQ(x¯1,,x¯k)1(2πσ2/ml)1/2exp-ml2σ2(y¯l-μl)2i,j1(2πσ2)1/2×exp-12σ2(xij-μi)2μ,σ2exp-12σ2i,j(xij-μi)2+ml(y¯l-μl)2μ,σ2exp-12σ2i,jxij2+mly¯l2-2j=1nlxlj+mlylμl-2ilj=1nixijμi=exp-12σ2sl2-2zlμl-2ilnix¯iμi

where μ,σ2 indicates that we are ignoring terms proportional to μ or σ2, and K(μ,σ2)=Eμ,σ2IQ(x¯1,,x¯k) with IQ denoting the indicator function of the ordering condition Q. Hence by the factorisation criterion, (Zl,X¯c,Sl2) is a sufficient statistic, where X¯c is the set of k − 1 variables {X¯i,i=1,,k,il}. The statistic is also complete, since it is from the exponential family of distributions (Lehmann and Romano 2005).

The UMVCUE of μl is the Rao-Blackwellisation of the unbiased estimator Y¯l, conditional on the complete, sufficient statistic (Zl,X¯c,Sl2). That is, we seek the estimator E(Y¯l|Zl,X¯c,Sl2,Q). Given the form of Ul, showing that E(Y¯l|Zl,X¯c,Sl2,Q) is equal to Equation (2) is equivalent to showing

EUl|Zl,X¯c,Sl2,Q=-1-r_l2c-(1-q_l2)c22ccB(c,c)Fβ(c,c)12(r_l+1)-Fβ(c,c)12(q_l+1). (3)

This in turn is equivalent to showing that the conditional density f(ul|zl,x¯c,Sl2,Q) of (Ul|Zl,X¯c,Sl2,Q) is

(1-u2)c-1q_lr_l(1-u2)c-1duIQ'ul,zl,x¯c,Sl2 (4)

where Q'={q_l<Ul<r_l,X¯1>>X¯l-1,X¯l+1>>X¯k}. This is because integrating uf(ul|zl,x¯c,Sl2,Q) gives the right hand side of Equation (3).

Consider a new sampling model, where Y¯l* is the mean of ml additional observations from population l, without having observed X¯1>X¯2>>X¯k. Let S*l, Zl*, U*l and S**l be equal to Sl, Zl, Ul and S˜l with Y¯* replacing Y¯ in all formulae. Rearranging terms gives the following expression for U*l 2:

Ul*2=ml(nl+ml)nlY¯l*-Zl*nl+ml2ml(nl+ml)nlY¯l*-Zl*nl+ml2+i=1kj=1ni(Xij-X¯i)2. (5)

Hence U*l 2 follows a β(12,c) distribution. Note that clearly U*l 2 ⩽ 1⇒ − 1 ⩽ Ul* ⩽ 1. Additional restrictions on U*l come from conditioning on Q. Firstly, X¯l>X¯l+1Y¯l*<1ml(Zl*-nlX¯l+1). Hence U*l < rl*, where

rl*=nl(nl+ml)/mlSl**Zl*nl+ml-X¯l+1.

Secondly, X¯l-1>X¯lY¯l*>1ml(Zl*-nlX¯l-1). Hence U*l > ql*, where

ql*=nl(nl+ml)/mlSl**Zl*nl+ml-X¯l-1.

Note that U*l is an ancillary statistic, and so by Basu's theorem (Basu 1955) is independent of Zl*,X¯c,Sl*2. Hence the joint density of (Ul*,Zl*,X¯c,Sl*2) is

g(ul*)h(zl*,x¯c,Sl*2) (6)

where g(u*l) is the density of U*l and h(zl*,x¯c,Sl*2) is the density of (Zl*,X¯c,Sl*2). The numerator of the conditional density of Ul*|(Zl*,X¯c,Sl*2,Q) is then Equation (6) times

K-1(μ,σ2)IQ*(ul*,zl*,x¯c,Sl*2)

where

Q*={q_l*<Ul*<r_l*,X¯1>>X¯l-1,X¯l+1>>X¯k},r_l*=min(rl*,1),q_l*=max(ql*,-1),

with I Q* being the indicator functions of the ordering condition Q*.

The denominator of the conditional density of Ul*|(Zl*,X¯c,Sl*2,Q) is the integral of Equation (6) with respect to u*l from q_l* to r_l*. This can be calculated using the fact that U*l 2 follows a β(12,c) distribution. Putting everything together gives the conditional density equivalent to Equation (4).

Remark 1.

As an important special case, setting l = 1 in Equation (2) gives the following UMVCUE for the mean of the highest-ranking treatment:

Z1n1+m1-n1m(n1+m1)×S˜(1-r_2)c22ccB(c,c)Fβ(c,c)12(r_+1) (7)

where

Z1=n1X¯1+m1Y¯1,S2=i=1kj=1niXij2+m1Y¯12S˜2=S2-(n1+m1)Zn1+m12-i=2kniX¯i2r=n1(n1+m1)/m1S˜Z1n1+m1-X¯2,r_=min(r,1)

Remark 2.

As a simple check of consistency with Section 2, set ni = n (for i = 1, …, k) and m 1 = 1 in Equation (7). Then we recover the corrected Equation (1) of Cohen and Sackrowitz.

4. Comparison with the known variance setting

Assuming a common known variance σ2, we can use the results of Bowden and Glimm (2008) to find the UMVCUE. Letting ϕ and Φ denote the pdf and cdf respectively of a standard normal distribution, the UMVCUE of μl is as follows.

Theorem 4.1.

The UMVCUE of μl given Q is

Zlnl+ml-nlml(nl+ml)×σ[φ(Wl,l+1)-φ(Wl,l-1)][Φ(Wl,l+1)-Φ(Wl,l-1)] (8)

where

Wl,l+1=nl(nl+ml)/mlσZlnl+ml-X¯l+1Wl,l-1=nl(nl+ml)/mlσZlnl+ml-X¯l-1

and we define X 0 = ∞ and X k + 1 = −∞.

Remark.

This is structurally the same as the UMVCUE with unknown variances given by Equation (2). Firstly, both estimators are in the form of the MLE Zlnl+ml minus a bias correction term, where the latter has a multiplicative factor of nlml(nl+ml).

The role of the known standard deviation σ in Equation (8) is played by the estimate S˜l in Equation (2). Hence rl and ql in Equation (2) can be seen to be exact analogues of W l, l + 1 and  W l, l − 1 respectively in Equation (8). Finally, the standard normal density and distribution functions in Equation (8) are replaced by (transformed) beta density and distribution functions in Equation (2).

4.1. Simulation study

We now conduct simulation studies to assess how the UMVCUE with a known variance performs when the assumed known variance σ^2 is not exactly equal to the true variance σ2. We compare various estimators for the largest mean μ1:

  • Stage 2 estimator, which is independent of σ2.

  • MLE, which is independent of σ2.

  • U^known, the UMVCUE which assumes that σ2 is known, and equal to some value σ^2.

  • U^unknown, the UMVCUE with unknown σ2.

To evaluate the performance of a generic estimator for μ1, say μ*1, we use the following definitions of the bias and MSE, as in Bowden and Glimm (2008):

bselμ1*=i=1KEμ1*-μiX1=XiPX1=Xi,MSEselμ1*=i=1KE(μ1*-μi)2X1=XiPX1=Xi.

We first compare characteristics of the estimators for trials with different target powers. Consider testing the hypothesis H 0: μ1 ⩽ 0 against the alternative H 1: μ1 > 0 at the end of the trial. To do so, we use the usual t-statistic

Td=Z1σ^n1+m1

where σ^2 is the pooled sample variance of the stage 1 and stage 2 data, i.e.

σ^2=1N+m1-k-1i=1kj=1ni(Xij-X¯i)2+j=1m1(Y1j-Y¯1)2. (9)

Under H 0, Td follows a t-distribution with d = N + m 1k − 1 degrees of freedom. Hence, when testing at significance level α, we reject H 0 if Td > t − 1 d(1 − α), where t − 1 d denotes the inverse cdf of a t-distribution with d degrees of freedom.

Suppose we are targeting a power of at least (1 − β) to detect a λ–fold standard deviation increase in the mean, i.e. under the alternative hypothesis H1:μ1=λσ^. This implies that

n1+m11λtd-1(1-α)+td-1(1-β)2. (10)

For simplicity, suppose we have equal numbers of subjects for each treatment and each stage, denoted by M, so that n 1 = m 1 = M, N = kM and hence d = M(k + 1) − k − 1. For given values of α, β and λ, Equation (10) can be solved for M numerically.

In our simulation study, α = 0.05 and we vary β from 0.05 to 0.5, which corresponds to trials with powers between 50% to 95%. We set k = 3, with μ = (0, 0, 0), σ2 = 1 and λ = 0.5. For the estimator U^known, we calculate σ^2 as the pooled sample variance given in Equation (9) and simply plug it in to Equation (8).

Figure 1 shows the mean bias of the MLE and U^known as the power varies from 50% (which corresponds to M = 4) to 95% (which corresponds to M = 12). We also simulated the bias of U^unknown and the stage 2 estimator, but as expected, this resulted in no bias apart from simulation error. Hence these simulation results are not shown here, and we simply plot where the bias equal to zero. The MLE is positively biased, with a mean bias ranging from 0.21 (for a power between 50% and 58%) to 0.12 (for a power of 95%). In contrast, U^known is essentially unbiased, although there seems to be a very small positive bias when the power is less than 58%.

Figure 1.

Figure 1.

Mean bias of the MLE and U^known as the power varies from 50% to 95%. Results are based on 105 simulations for each value of the per-group sample size M.

Figure 2 shows the MSE of the four estimators. The stage 2 estimator has the highest MSE, while the MLE has the lowest. The estimators U^unknown and U^known have approximately the same MSE, which is 29%–30% higher than the MSE for the MLE, but 26%–27% lower than the MSE for the stage 2 estimator.

Figure 2.

Figure 2.

MSE of the four estimators as the power varies from 50% to 95%. Results are based on 105 simulations for each value of the sample size M.

To demonstrate the differences further, Figure 3 shows the boxplots for the estimators when M = 10, which corresponds to a power of 90%. The boxplots demonstrate the severity of the positive bias of the MLE, and the unbiasedness of the other three estimators. The distribution of the estimators U^unknown and U^known seem to be almost identical.

Figure 3.

Figure 3.

Boxplots of the four estimators for a sample size M = 10, based on 105 simulations.

To explore what sort of estimates of σ^ lead to substantial differences between U^unknown and U^known, we conduct a further simulation study which does not assume that we are using the pooled sample variance as our estimate of σ^. Keeping M = 10 as before, we vary the value of σ^2 from 0.25 to 4 and simply plug it in to Equation (8). In order to get a handle on the likely values of σ^2 that could be estimated from the data, note that σ^2σ2dχd2, and hence var(σ^2)=2σ2d.

For M = 10, σ2 = 1 and k = 3, this gives var(σ^2)0.0556. Hence in our simulation setting, a very simple approximate 95% confidence interval for σ^2 is given by 1±2×0.0556, which we show on the plots below as a shaded gray region.

Figure 4 shows the mean bias of the MLE and U^known. The MLE has a positive bias of 3410π=0.1338... for all values of σ^2. In contrast U^known is positively biased for σ^2<1 and negatively biased for σ^2>1. The symmetry of the positive and negative bias of U^known around σ^2=1 explains why, on average, using a pooled sample variance estimator (which would be expected to under and overestimate σ2 almost equally) results in an essentially unbiased estimator.

Figure 4.

Figure 4.

Mean bias of the MLE and U^known as σ^2 varies from 0.25 to 4. Results are based on 105 simulations for each value of σ^2. The gray region shows an approximate 95% confidence interval for σ^2.

Figure 5 shows the MSE of the four estimators. The stage 2 estimator has the highest MSE of 0.1, while the MLE has the lowest MSE of 120(1+34π)=0.05689... for all values of σ^2. The estimator U^unknown has a MSE of approximately 0.074. This is a 26% decrease compared to the stage 2 estimator, but a 43% increase compared to the MLE. Finally, U^known has a lower MSE than U^unknown for σ^2<1, which can be explained by regarding U^known as a sort of shrinkage estimator. Again, the symmetry of the difference between the MSEs of U^known and U^unknown around σ^2=1 explains why, on average, the MSEs of these two estimators are so similar.

Figure 5.

Figure 5.

MSE of the four estimators as σ^2 varies from 0.25 to 4. Results are based on 105 simulations for each value of σ^2. The gray region shows an approximate 95% confidence interval for σ^2.

5. Case study

Finally, we illustrate our methodology using data based on the INHANCE study (Lawrence, Bretz, and Pocock 2014), which evaluated the use of inhaled indacaterol for the treatment of patients suffering from chronic obstructive pulmonary disease (COPD). The study followed a two-stage adaptive seamless design, where the first stage was a dose-finding stage with dose selection at the interim analysis, and the second stage was a confirmatory analysis of the efficacy and safety of the selected doses. The primary outcome was the 24-hour post-dose FEV1 (which is the Forced Expiratory Volume after one second).

In the study, four doses of indacaterol were tested in stage 1 (75 μg, 150 μg, 300 μg and 600 μg), and two doses were selected based on a set of dose selection guidelines (see pages 84–87 in Lawrence, Bretz, and Pocock (2014) for further details). For simplicity, and in order for the selected doses to be the same as those that would be selected when ranking by sample means, we only consider the three doses 75 μg, 150 μg and 300 μg of indacaterol.

Table 1 gives the observed least-squares mean treatment difference versus placebo at week 2 for each dose of indacaterol for the two stages, with the stage 1 results as given in Barnes et al. (2010), and the stage 2 results as given in Donohue et al. (2010). On the basis of ranking by the sample means, doses 150 μg and 300 μg would be selected for confirmatory analysis in stage 2.

Table 1.

Sample sizes used for the case study, with the observed treatment difference from placebo at week 2 as observed in the INHANCE study.

  Stage 1
Stage 2
Indacaterol dose (μg) ni Treatment difference vs placebo (L) mi Treatment difference vs placebo (L)
75 10 0.15
150 9 0.18 9 0.18
300 7 0.21 9 0.18

The INHANCE study had over 100 patients randomised to each of the three doses in stage 1, and over 400 patients randomised to each of the two selected doses in stage 2. With these large sample sizes, we would not expect there to be appreciable differences between the estimators. Hence, for the purposes of our illustrative case study, we consider a trial with only n = (10, 9, 7) patients in stage 1, and 9 patients for each dose in stage 2, as shown in Table 1.

We use the observed mean treatment differences from each stage of the INHANCE study to simulate a realisation of the data with the sample sizes given above. For each dose at each stage, the data was simulated from a normal distribution with mean equal to that observed in the INHANCE study, and a standard deviation of 0.3L (as given in the results of Donohue et al. 2010).

Table 2 shows the various estimators for the treatment difference from placebo (L) at week 2 for the two selected doses. Note that for U^known, we use the pooled sample variance as given in Equation (9) as our estimate for σ^2. For both doses, there is a drop in the mean between the stage 1 and stage 2 data, with the MLE in-between since it is a weighted mean. For the 300 μg dose, there are only very small differences between the MLE, U^known and U^unknown, while for the 150 μg dose these three estimators are in fact identical. Hence for these data, there would be no practical difference in using any of these three estimators to correct for selection bias.

Table 2.

Estimators for the treatment difference from placebo (L) at week 2, based on results from the INHANCE study.

Indacaterol dose (μg) Stage 1 Stage 2 MLE U^known U^unknown
300 0.218 0.179 0.196 0.194 0.195
150 0.192 0.180 0.186 0.186 0.186

6. Discussion

In two-stage adaptive trials with normally distributed data, it is unlikely that the variance will be known exactly. Our modified estimator allows for efficient unbiased estimation of multiple selected treatments with a common unknown variance, where the stage one and two sample sizes are arbitrary.

As the simulation studies demonstrated, when using the pooled sample variance, on average the UMVCUE U^known that assumes a known variance will be essentially unbiased, with a very similar MSE to our modified UMVCUE. However, if the variance is under or overestimated, then U^known is no longer unbiased, with a higher MSE if the variance is overestimated. We note that our proposed estimator is most useful when the sample sizes are small. Indeed, if σ^2 is estimated using the pooled sample variance, then for d > 200, var(σ^2)<0.01σ2 and there is unlikely to be any substantial variation of σ^2 from the true value σ2.

In this paper, we only looked at ranking by the treatment means. However, it should be reasonably straightforward to adapt our result to ranking by other criteria. One example would be ranking by one-sided ‘p-value’, where the conditioning would change to the event Q'={X:n1X¯1>>nkX¯k}. Then the proof follows through identically to give the formula (2), except that the X¯l+1 in rl is replaced by nl+1X¯l+1/nl, and the X¯l-1 in ql is replaced by nl-1X¯l-1/nl.

The proof of the UMVCUE is quite a delicate one, and we do not see a straightforward way to generalise the result to the setting where we no longer assume a common unknown variance but instead let treatment i have variance σ2 i say. The problem is that U*l 2 as given in Equation (5) is no longer a ratio of independent chi-squared distributions, and would now depend on the unknown parameters σ2 1, …, σk 2.

Finally, the focus of this paper was on point estimation, but it is natural to try to derive confidence intervals for the UMVCUE as well. One approach is to use a bootstrap procedure, similar to that described in e.g., Bowden and Glimm (2008) and Robertson, Prevost, and Bowden (2016a). Alternatively, the approach of Sampson and Sill (2005) can possibly be extended to give exact confidence intervals in the unknown variance setting.

Acknowledgments

The authors would like to thank the two anonymous reviewers, whose comments greatly improved the paper. David S. Robertson was supported by the Medical Research Council (grant number MC_UP_1302/4).

References

  1. Barnes P. J., Pocock S. J., Magnussen H., Iqbal A., Kramer B., Higgins M., and Lawrence D. 2010. Integrating indacaterol dose selection in a clinical study in COPD using an adaptive seamless design. Pulmonary Pharmacology & Therapeutics 23 (3):165–71. [DOI] [PubMed] [Google Scholar]
  2. Basu D. 1955. On statistics independent of a complete sufficient statistic. Sankhyā: The Indian Journal of Statistics (1933–1960) 15 (4):377–80. [Google Scholar]
  3. Bowden J., and Dudbridge F. 2009. Unbiased estimation of odds ratios: Combining genomewide association scans with replication studies. Genetic Epidemiology 33 (5):406–18. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bowden J., and Glimm E. 2008. Unbiased estimation of selected treatment means in two-stage trials. Biometrical Journal 50 (4):515–27. [DOI] [PubMed] [Google Scholar]
  5. Cohen A., and Sackrowitz H. B. 1989. Two stage conditionally unbiased estimators of the selected mean. Statistics & Probability Letters 8 (3):273–8. [Google Scholar]
  6. Donohue J. F., Fogarty C., Lotvall J., Mahler D. A., Worth H., Yorgancioglu A., Iqbal A., Swales J., Owen R., Higgins M., and Kramer B. 2010. Once-daily bronchodilators for chronic obstructive pulmonary disease: Indacaterol versus tiotropium. American Journal of Respiratory and Critical Care Medicine 182 (2):155–62. [DOI] [PubMed] [Google Scholar]
  7. Kimani P. K., Todd S., and Stallard N. 2013. Conditionally unbiased estimation in phase II/III clinical trials with early stopping for futility. Statistics in Medicine 32 (17):2893–910. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Lawrence D., Bretz F., and Pocock S. 2014. INHANCE: An adaptive confirmatory study with dose selection at interim In Indacaterol, 77–92. Basel: Springer. [Google Scholar]
  9. Lehmann E. L., and Romano J. P. 2005. Testing statistical hypotheses. New York: Springer. [Google Scholar]
  10. Robertson D. S., Prevost A. T., and Bowden J. 2016a. Accounting for selection and correlation in the analysis of two-stage genome-wide association studies. Biostatistics 17:634–49. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Robertson D. S., Prevost A. T., and Bowden J. 2016b. Unbiased estimation in seamless phase II/III trials with unequal treatment effect variances and hypothesis-driven selection rules. Statistics in Medicine 35:3907–22. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Sampson A. R., and Sill M. W. 2005. Drop-the-losers design: Normal case. Biometrical Journal 47 (3):257–68. [DOI] [PubMed] [Google Scholar]

Articles from Communications in Statistics are provided here courtesy of Taylor & Francis

RESOURCES