Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2023 Dec 4.
Published in final edited form as: Biometrics. 2022 Aug 2;79(3):2116–2126. doi: 10.1111/biom.13711

A general framework for subgroup detection via one-step value difference estimation

Dana Johnson 1, Wenbin Lu 2, Marie Davidian 2
PMCID: PMC10694635  NIHMSID: NIHMS1945184  PMID: 35793474

Abstract

Recent statistical methodology for precision medicine has focused on either identification of subgroups with enhanced treatment effects or estimating optimal treatment decision rules so that treatment is allocated in a way that maximizes, on average, predefined patient outcomes. Less attention has been given to subgroup testing, which involves evaluation of whether at least a subgroup of the population benefits from an investigative treatment, compared to some control or standard of care. In this work, we propose a general framework for testing for the existence of a subgroup with enhanced treatment effects based on the difference of the estimated value functions under an estimated optimal treatment regime and a fixed regime that assigns everyone to the same treatment. Our proposed test does not require specification of the parametric form of the subgroup and allows heterogeneous treatment effects within the subgroup. The test applies to cases when the outcome of interest is either a time-to-event or a (uncensored) scalar, and is valid at the exceptional law. To demonstrate the empirical performance of the proposed test, we study the type I error and power of the test statistics in simulations and also apply our test to data from a Phase III trial in patients with hematological malignancies.

Keywords: exceptional law, optimal treatment rule, precision medicine, subgroup test, survival analysis, value function

1 |. INTRODUCTION

Consider the setting in which an active treatment and a control are to be compared on the basis of some outcome—which may be continuous, discrete, or a possibly censored time-to-event. Due to patient treatment effect heterogeneity, it is possible that a subgroup of patients benefit from the active treatment, but the overall estimated treatment effect is nonsignificant or borderline significant. When this happens, one may want to “salvage” the treatment by testing whether at least some subgroup benefits from treatment, and, if the null hypothesis of no benefit is rejected, identifying the subgroup(s) with an enhanced treatment effect (Lipkovich et al., 2017). For example, Lipkovich et al. (2017) describe a Phase III clinical trial in patients with hematological malignancies that resulted in a hazard ratio (treatment vs. control) for overall survival that was borderline significant (one-sided p=0.0367). Due to this borderline result, the sponsor decided to search for subgroups with improved benefit/risk ratios. To handle situations similar to those described above, statistical methods from the precision medicine literature are often applied. This literature mainly focuses on one of two things: estimation of an optimal treatment regime, or identification of subgroups with enhanced treatment effects. Various parametric or nonparametric methods have been developed for the estimation of optimal treatment regimes with both uncensored outcomes (eg, Watkins and Dayan, 1992; Murphy, 2003; Zhang et al., 2012a, b; Zhao et al., 2012) and censored time-to-event outcomes (eg, Bai et al., 2017; Goldberg and Kosorok, 2012; Jiang et al., 2017; Zhao et al., 2015). Likewise, intensive research has been done for identifying subgroups with enhanced/differential treatment effects (eg, Cai et al., 2010; Foster et al., 2011; Lipkovich et al., 2011; Li et al., 2019; Wang et al., 2019; Zhao et al., 2013). However, less work has been done in the area of subgroup testing. Here the focus is to test whether or not one treatment is better than the other for at least some subgroup of the population. As argued by Shi et al. (2020), this is a foundational question that needs to be answered prior to either estimating an optimal treatment regime or searching for subgroups with enhanced treatment effects. Existing subgroup testing literature includes parametric mixture models (Shen and He, 2015; Wu et al., 2016) and change-plane approaches (Fan et al., 2017; Kang et al., 2017; Wei and Kosorok, 2018). Some limitations of these approaches are that the subgroup is assumed to have some parametric form, such as defined by a linear function, and/or the treatment effect is assumed to be constant within the subgroup. Such assumptions may be restrictive in practical applications. These challenges motivate us to develop a general framework for testing for the existence of a subgroup with enhanced treatment effects, which does not require specification of the form of the subgroup and allows heterogeneous treatment effects within the subgroup.

Our proposed test statistics are built based on estimators of the value difference between an estimated optimal regime and a fixed regime that assigns everyone to the same treatment. Both uncensored outcomes and censored time-to-event outcomes are considered. To derive the asymptotic distributions of the test statistics under the null hypothesis, we adapt the one-step estimation procedure proposed by Luedtke and van der Laan (2016, LL16 henceforth), which can lead to valid inference even under the exceptional law, that is, when the treatment effect of individual patients can be zero with a positive probability. In addition, within the one-step estimation framework, we consider two schemes to partition observations into smaller chunks of data and show that one particular partition can enhance the power of the considered test statistic under certain settings. In Section 2, we introduce notation and assumptions. In Section 3, we propose several value difference estimators for both uncensored and censored outcomes. In Section 4, we introduce our test procedure. We study the performance of the proposed test statistics via extensive simulations in Section 5 and apply the proposed test to the Phase III trial data described in Lipkovich et al. (2017) in Section 6.

2 |. STATISTICAL FRAMEWORK FOR SUBGROUP TESTING

We first introduce the framework in the case of an uncensored outcome. Let X be a p×1 vector of baseline covariates taking values in 𝒳. Let Y be an observed, continuous outcome coded so larger values are preferred, and let A be a binary treatment indicator, where A=0 corresponds to control/standard of care, and A=1 corresponds to the active treatment. The observed data consist of independent and identically distributed (iid) copies of O={X,A,Y}, indexed by i=1,,n. Let Y*(a) be the potential outcome (Rubin, 1974) a patient could experience if he/she were to receive treatment option a,a=0,1. We assume that

E{Y*(a)X}=C0X+aτX, (1)

where C0(X) is the expected conditional outcome if the control were assigned to all patients with X in the population, and τ(X)=E{Y*(1)X}-E{Y*(0)X} is the conditional average treatment effect (CATE).

We assume the stable unit treatment value assumption (SUTVA) holds, which implies that Y=Y*(A). In addition, we assume both the positivity assumption and the no unmeasured confounders assumption hold, which state that pr(A=aX=x)>0 for all x𝒳 and that {Y*(0),Y*(1)}AX, respectively, where means “independent” and the notation X refers to conditional independence given X. Under these assumptions, it can be shown that τ(X)=E(YX,A=1)-E(YX,A=0).

We are interested in testing whether or not the active treatment is beneficial, relative to control, for at least some patients in the population. Therefore, we are interested in testing

H0:τ(x)0forallx𝒳vs.H1:𝒳1𝒳withprX𝒳1>0suchthatτx>0forallx𝒳1; (2)

where, importantly, we allow a portion of the population to not experience a treatment effect, that is, we allow pr{τ(X)=0}>0. This is referred to as the exceptional law. Let d be a treatment rule that maps information in X to a treatment decision in {0, 1}. The potential outcome under d is defined as Y*{d(X)}=Y*(1)1{d(X)=1}+Y*(0)1{d(X)=0}, where 1(⋅) is the indicator function with 1(B)=1 if the event is true and = 0 otherwise. The value of a treatment rule is defined as V(d)=E[Y*{d(X)}], and an optimal treatment rule, doptdopt(X)=argmaxdV(d). Under our considered model and assumptions, the optimal treatment rule is given by dopt(X)=1{τ(X)>0}. Note that when τ(X)=0 the defined optimal rule selects the control for unique representation. Let d(X)0 denote the regime that gives everyone treatment 0. Define Ψdopt=Vdopt-V(0), which is the value difference under the optimal rule and the fixed regime d(X)0. It can be shown that Ψdopt=0 under H0, while it is positive under H1. This motivates us to use estimators of the value difference as test statistics for testing the considered hypothesis in (2). If the outcome of interest is a right-censored time-to-event, the above framework can be modified. The observed data become O={X,A,U,Δ}, where T is the event time of interest, C is the censoring time, U=min(T,C), and Δ=1(TC). The definition of the CATE changes from a difference in conditional means to a difference in conditional restricted mean survival times. Let T*(a) denote the potential time to event had treatment a been administered, and let L be such that pr(UL)>0. The restricted mean survival time (RMST) under treatment a, conditional on X, is defined as E[min{T*(a),L}X]=0LS*a(tX)dt, where S*a(tX)=P(T*(a)tX) is the conditional survival function under treatment a. Then, the CATE is τ(X)=E[min{T*(1),L}X]-C0(X), where C0(X)=E[min{T*(0),L}X]. In addition to the positivity assumption and extensions of the SUTVA and no unmeasured confounders assumption to time-to-event data, we assume that censoring is noninformative, that is, that C{T*(0),T*(1)}X. Under these four assumptions, τ(X) can be identified as a function of the observed data (see Web Appendix A in the Supporting Information). We consider the same hypothesis as defined in (2), but the value function is now defined based on the restricted mean survival time, that is, V(d)=E(min[T*{d(X)},L]).

3 |. VALUE DIFFERENCE ESTIMATORS

3.1 |. Value difference estimators for an uncensored outcome

To construct estimators for the value difference Ψdopt, we use the inverse propensity score weighted estimators proposed by Zhang et al. (2012b) for the value under the optimal rule and the fixed regime d0, respectively. Specifically, define

SIPWOi=1Ai=1τXi>0πAiXiYi-1Ai=1τXi>0πAiXi-1×C0Xi+τXi1τXi>0-1Ai=01-πXiYi, (3)
SAIPWOi=SIPWOi+1Ai=01-πXi-1C0Xi, (4)

and Ψ^Ψ^(O)=n-1i=1nSOi, where π(X)=pr(A=1X) is the propensity score, πA(X)=Aπ(X)+(1-A){1-π(X)}, and S serves as a placeholder for either SIPW or SAIPW. In both Ψ^IPW and Ψ^AIPW the estimator for Vdopt, which corresponds to the first three terms of (3) and the first term of (4), is an augmented inverse probability weighted (AIPW) estimator, whereas the estimator for V(0), which corresponds to the last term in (3) and (4), is an inverse probability weighted (IPW) estimator in Ψ^IPW, but an augmented inverse probability weighted estimator in Ψ^AIPW. When constructing the value difference estimators Ψ^IPW and Ψ^AIPW, the functions π,C0, and τ are estimated from the data. The details will be given in the next section.

Remark 1.

Inspection of (4) reveals that Ψ^AIPW is exactly equal to 0 whenever 1τXi>0=0 for all i. In the regular setting, that is, pr{τ(X)=0}=0, we have τ(X)<0 with probability 1 under H0. Therefore, Ψ^AIPW under H0 would be exactly equal to 0 asymptotically because τ^Xi will become negative for all i as n goes to infinity, as long as τ^ is a consistent estimator of τ. This implies that the limiting null distribution of Ψ^AIPW will be degenerate under the regular setting. On the contrary, the limiting null distribution of Ψ^IPW will still be normal because the IPW estimator of V(0) used in Ψ^IPW does not exactly cancel with the AIPW estimator of Vdopt under the null.

3.2 |. Value difference estimators for censored outcome

Let UiL=minUi,L and ΔiL=1UiLCi=Δi+(1-Δi1UiL. Define

SIPWSOi=ΔiL1Ai=1τXi>0KcUiLXi,AiπAiXiUiL-1Ai=1τXi>0πAiXi-1ζiC0Xi+τXi1τXi>0-ΔiL1Ai=0KcUiLXi,Ai1-πXiUiL, (5)
SAIPWSOi=SIPWSOi+1Ai=01-πXi-1C0Xi, (6)
SCAIPWSOi=SAIPWSOi+1Ai=1τXi>0πAiXi0LdMcrXi,AiKcrXi,AimrXi,Ai-1Ai=01-πXi0LdMcrXi,AiKcrXi,AimrXi,Ai, (7)

and Ψ^SΨ^S(O)=n-1i=1nSSOi, where Kc(tXi,Ai=prCitXi,Ai is the conditional survival function of censoring times, ΛcrXi,Ai is the associated conditional cumulative hazard function, dMcrXi,Ai=dNc(r)-Y(r)dΛcrXi,Ai,Nc(r)=1Uir,Δi=0 is the censoring counting process, Y(r)=1Uir is the at risk process, mrXi,Ai=Emin(T,L)Tr,Xi,Ai, and ζi equals 1 in (5) and is a mean one perturbation term in (6) and (7); see Remark 2. We use SS as a placeholder for SIPWS, SAIPWS, or SCAIPWS. The estimator in (7) is called a “CAIPW” estimator because it is an AIPW estimator with additional censoring augmentation terms (the “C” refers to censoring). Specifically, the second and third terms on the right-hand side of (7) are censoring augmentation terms under dopt and d(X)0, respectively, which can improve the efficiency of the value difference estimator by recovering information on censored but regime-consistent patients. It can be shown that the CAIPW estimator is the locally efficient estimator for the value difference (Bai et al., 2013) and has a desirable double robustness property—meaning if either the propensity model and the model for the censoring distribution are correctly specified, or if the conditional distribution of TiXi,Ai is correctly specified, Ψ^CAIPWS is a consistent estimator for the value difference. If the data come from a randomized trial, so that π and the censoring distribution can be correctly specified, then all of Ψ^IPWS, Ψ^AIPWS, and Ψ^CAIPWS are consistent estimators for Ψdopt (Bai et al., 2017).

Remark 2.

For reasons described in Web Appendix B of the Supporting Information, ζi needs to be included in (6) and (7) because without ζi it is likely for Ψ^j,AIPWS(Cj,𝒪j-1*) and Ψ^j,CAIPWS(𝒞j,𝒪j-1*) (described in Section 4.2) to be exactly equal to zero even at the exceptional law—unlike for an uncensored outcome, for which this is a concern only in the regular setting.

4 |. TEST PROCEDURES

4.1 |. Estimation of nuisance functions

The value difference estimators presented in Sections 3.1 and 3.2 are not directly useful because τ(X),C0(X),π(X),KCrXi,Ai,dMcrXi,Ai, and mrXi,Ai are usually unknown functions. For an uncensored outcome, we propose estimating C0(X) and τ(X) using random forests (Breiman, 2001). Specifically, we use the counterfactual synthetic random forest approach described by Lu et al. (2018). Refer to Web Appendix C for the justification for using this approach and for details on how this estimator is computed. While choosing a nonparametric method places less assumptions on the form of C0(X) and τ(X), a concern with using these methods is the “curse of dimensionality.” Random forests tend to be more robust to the curse of dimensionality (can handle larger p) than, say, kernel regression. For a censored outcome, we propose estimating C0(X) and τ(X) using random survival forests (Ishwaran et al., 2008). Refer to Web Appendix D for details. We estimate Kc(tX,A) by fitting a Cox proportional hazards model with U as the observed time and 1-Δ as the status. The censoring augmentation term (the part that is an integral), which is a function of nuisance functions, can be rewritten as

0LdMcrXi,AiKcrXi,AimrXi,Ai=(1-ΔiL)mUiLXi,AiKcUiLXi,Ai-0UiLdΛcrXi,AiKcrXi,AimrXi,Ai. (8)

Refer to Web Appendix E for the proof. In (8), the estimator for the conditional cumulative censoring hazard, which we denote Λ^c(rX,A), is obtained from the same Cox proportional hazards model used to estimate Kc(tX,A). Finally, as discussed in Chapter 8 of Tsiatis et al. (2020), m(rX,A) can be written as

m(rX,A)=rLz-dSzXi,AiSrXi,Ai+LSrXi,AiSLXi,Ai. (9)

Thus, an estimator for mrXi,Ai, mˆrXi,Ai, is obtained by replacing S(X,A) in (9) with the random survival forest estimator S^(X,A). For both uncensored and censored outcomes, we estimate the propensity score using logistic regression.

4.2 |. One-step framework

We replace the nuisance functions in the expressions for the value difference estimators in Section 3 with the estimators described in Section 4.1, and we follow the one-step estimation framework in LL16 to handle the nonregularity resulting from estimating dopt under the exceptional law. We describe the one-step framework and associated data partitioning methods (Section 4.3) in the censored outcome context. The steps and notation for the uncensored outcome setting are analogous and are given in the Web Appendix F. We first estimate the nuisance functions using a “chunk” of ln observations. We partition the remaining data into rn=n-ln/m chunks of size m, where m1, and ln is chosen such that n-ln/m as n. Let 𝒞j=Oi:Oijthdatachunk}, so that each 𝒞j for j=1,,rn contains m observations. We also let 𝒪j*=Oi:Oik=0jck, where 𝒞0𝒪0* corresponds to all observations in the initial data chunk of size ln. Refer to Web Figure 1. Now define

Sj,CAIPWS(Oi,𝒪j-1*)=Sj,AIPWS(Oi,𝒪j-1*)+1Ai=1τˆjXi>0πˆj,AiXi×0LdM^j,crXi,AiK^j,crXi,Aim^jrXi,Ai-1Ai=01-π^jXi0LdM^j,crXi,AiK^j,crXi,Aim^jrXi,Ai, (10)

and Ψˆj,S(𝒞j,𝒪j-1*)=m-1cjSj,S(Oi,𝒪j-1*), where πˆjX,πˆj,AX,τ^jX,C^j,0X,K^crXi,Ai,dM^crXi,Ai, and m^rXi,Ai are estimators for π(X),πA(X),τ(X),C0(X),KcrXi,Ai,dMcrXi,Ai, and mrXi,Ai, respectively, based on the “historical” data in 𝒪j-1*. At step j, these estimators are known, fixed functions. For brevity, expressions for Sj,IPWS(Oi,𝒪j-1*) and Sj,AIPWS(Oi,𝒪j-1*) are not given here, but they are analogous to (10).

We consider the test statistic

T,SBTS=rn1/2j=1rnσ^j1Ψ^j,S𝒞j,𝒪j1*=rn1/2j=1rnσ^j1Ψ^S(O),Ψ^S(O)=j=1rnσ^j1Ψ^j,S𝒞j,𝒪j1*j=1rnσ^j1. (11)

Here, Ψ^S(O) is a weighted average of the chunk-specific value difference estimators, so that Ψ^S(O) is an estimator for Ψdopt. The σ^j are estimators for the standard deviation of Ψ^j,S(𝒞j,𝒪j-1*) given 𝒪j-1* (see Section 4.3). The asymptotic distribution of Ψ^S(O) can be derived using conditions similar to (C1)—(C5) in LL16. Thus, the rejection region for an asymptotically valid, α level test is T,SBT>z1-α, where z1-α denotes the 1-α quantile of the standard normal distribution. Details on how random forests may achieve the convergence rate required by (C4) and (C5) of LL16 are provided in the Web Appendix G.

4.3 |. Data chunking approaches

There are two ways one might partition the observed data into chunks. One way is to randomly assign observations to a chunk without paying attention to the treatment label, which is the approach in LL16 referred to here as the sequential, blinded-treatment (SBT) approach. Another option is to assign observations to chunks such that the proportion of treatment observations in each chunk preserves the marginal probability of treatment, pr(A=1), in the study. For example, consider a randomized trial where π=0.4 is known, and we set the chunk size m=10. Then observations are first allocated to 𝒞1 through 𝒞rn such that each of these chunks contains four treatment observations and six control observations. The remaining ln observations are assigned to 𝒞0; although the proportion of treatment observations in 𝒞0 may not equal π, it should be close in large samples. We call this second approach the sequential, averaged-propensity matching (SAP-match) approach. For data from an observational study, the SAP-match approach can be implemented by replacing π with the sample proportion of treatment observations, pˆ. Thus, for each value difference estimator, there exist two versions of the test statistic—the SBT version and the SAP-match version. For time-to-event outcomes, we use Ψ^j,S,mπ(𝒞j,𝒪j-1*) to denote the SAP-match estimator, whereas Ψ^j,S(𝒞j,𝒪j-1*) denotes the SBT estimator. The form of the test statistic in (11), under the SAP-match chunking method, is T,SAP-matchS=rn-1/2j=1rnσ^j,mπ-1Ψ^j,S,mπ(𝒞j,𝒪j-1*), where σ^j,mπ is an estimator for var{Ψ^j,S,mπ(𝒞j,𝒪j-1*)𝒪j-1*}1/2 formed using only the data in Oj-1*. Notation for uncensored outcomes is analogous.

When π is known, it is shown in Web Appendix H that E{Ψ^j,S,mπ(𝒞j,𝒪j-1*)𝒪j-1*}=E{Ψ^j,S(𝒞j,𝒪j-1*)𝒪j-1*}, so that Ψ^j,S,mπ is also a consistent estimator for Ψ(d^jopt), where d^jopt is a fixed estimator for the optimal rule. A similar result holds in the uncensored outcome setting.

When observations are randomly allocated to chunks, which is the case under the SBT approach, σj can be estimated as described in LL16. Specifically, taking 𝒪j-1* to represent the number of elements in 𝒪j-1*, let

σ^j2=maxqj,m1𝒪j1*1Oi𝒪j1*Sj,SOi,𝒪j1*2{𝒪j1*1Oi𝒪j1*Sj,SOi,𝒪j1*}2 (12)

for any qj0, as j, which adapts (14) of LL16 to our setting. The sequence qj ensures that σ^j-1 is finite for all j. It can depend on both j and Sj,S(Oi,𝒪j-1*), for Oi𝒪j-1*. When π is known, the variance of Ψ^j,S,mπ(𝒞j,𝒪j-1*), given 𝒪j-1*, is

varΨ^j,S,mπ(𝒞j,𝒪j-1*)𝒪j-1*=m-1πvarSj,S(Oi,𝒪j-1*)𝒪j-1*,Ai=1+(1-π)varSj,S(Oi,𝒪j-1*)𝒪j-1*,Ai=0, (13)

which naturally suggests that we can estimate σj,mπ by the square root of

σ^j,mπ2=maxqj,m1π𝒪1,j1*1oi𝒪1,j1*Sj,SOi,𝒪j1*2𝒪1,j1*1×Oi𝒪1,j1*Sj,SOi,𝒪j1*2+1π𝒪0,j1*1Oi𝒪0,j1*Sj,SOi,𝒪j1*2𝒪0,j1*1Oi𝒪0,j1*Sj,SOi,𝒪j1*2 (14)

for any qj0, as j. In the above, 𝒪0,j-1* and 𝒪1,j-1*, respectively, denote the sets of observations with A=0 and A=1 that are in the historical data 𝒪j-1*. When π is unknown, pˆ=n-1i=1nAi can be substituted for π in (14). Theorem 1 shows that σ^j2 and σˆj,mπ2 cannot be used interchangeably; the proof is given in Web Appendix I. Similar arguments can be used to show this relationship also holds for the uncensored outcome setting.

Theorem 1.

Under the conditions given in Web Appendix I,

varΨ^j,S(𝒞j,𝒪j-1*)𝒪j-1*varΨ^j,S,mπ(𝒞j,𝒪j-1*)𝒪j-1*.

According to Theorem 1, using the SAP-match approach to construct the data chunks may give the proposed test more power to detect departures from H0 than using the SBT approach. The simulations in Section 5.1 show that this power difference is remarkable in the randomized controlled trial setting with (SBT/SAP-match) value difference estimators based on Sj,IPW(Oi,𝒪j-1*); however, when (SBT/SAP-match) value difference estimators are based on Sj,AIPW(Oi,𝒪j-1*), the difference in power is indistinguishable across chunking method. Perhaps this is because m-1varE{Sj,AIPW(Oi,𝒪j-1*)𝒪j-1*,Ai}𝒪j-1*, which appears in the proof of Theorem 1, is sufficiently small under our choice of m=10.

5 |. SIMULATION STUDIES

5.1 |. Real-valued, uncensored health outcome

We evaluate the performance of the four versions of the proposed test statistic (TIPW,SAP-match,TAIPW,SAP-match,TIPW,SBT, and TAIPW,SBT) by generating data from three models of the following form: Yi=C0Xi+AiτXi+εi,εi are iid N(0,0.25). For all three models, Ai are Bernoulli with mean π for randomized study simulations that used the SBT approach; i=1nAi=nπ for randomized study simulations that used the SAP-match approach; and Bernoulli with mean πXi for all observational data simulations. Across the three models, the proportion of the population with a treatment effect of exactly zero ranged from around 0.26–0.89. Each simulation involved 500 Monte Carlo data sets with n=600 or 1000. In all simulations, ln=n/2 and m=10. We defined c to be the treatment effect among the patients who do not have a treatment effect of 0. When c0,H0 is true; when c>0,H1 is true. We considered a range of values for c, with π={0.4,0.5,0.6} for the randomized study simulations, and for the observational study simulations propensity score models that, marginally, yield similar probabilities. We used synthetic regression forests from the R package randomForest-SRC (Ishwaran and Kogalur, 2019) to estimate C0(X) and τ(X).

We present results under Model 1 with π=0.5 and n=1000, where Model 1 takes C0Xi=3.18+0.2X1i+X2i+0.5X3i, and τXi=c1-0.5X2i2+X4i>0. We generated X1, X2, X3 as independent N(0,1) and X4, X5 as independent Bernoulli with mean 0.5. Under Model 1 (scheme 1), about 58.9% of the population had a treatment effect of 0. Results under the other two models and settings are similar, and can be found in Web Tables 415 of the Supporting Information.

As shown in Table 1, the type I error (α=0.05) under the randomized setting is controlled under all four versions of the test statistic. In the observational setting, it is controlled under all versions except for TIPW,SAP-match (see Web Table 2). The test statistics are slightly, negatively biased, which is to be expected when this one-step procedure is used in finite samples. The power using TAIPW, is superior to that using TIPW,•. Remarkably, under the randomized setting, TIPW,SAP-match has much higher power than TIPW,SBT. This can be explained by Theorem 1. For TAIPW,•, the power is similar across the SBT and SAP-match chunking methods. We empirically validated the result in Theorem 1 by running simulations that computed TIPW,SAP-match by using σ^j in place of σ^j,mπ. Refer to Web Table 1. We also considered a second data-generating scheme (scheme 2) for Model 1, where X1-X5 are as in scheme 1, and 20 additional iid N(0, 1) covariates were generated. As anticipated, the type I error remains controlled, and the power is almost the same as that under scheme 1. See Web Table 3 in the Supporting Information.

TABLE 1.

Randomized study results under Model 1 scheme 1, with π=0.5,α=0.05,n=1000

c Method μIPW μAIPW σIPW σAIPW 1βIPW/αIPW 1βAIPW/αAIPW
0.30 SAP-match 1.61 2.50 1.01 1.01 0.51 0.80
0.30 SBT 0.62 2.55 1.00 0.99 0.15 0.81
0.40 SAP-match 2.29 3.39 1.01 1.00 0.73 0.96
0.40 SBT 0.92 3.44 1.00 1.01 0.21 0.97
0.50 SAP-match 2.92 4.40 0.94 1.02 0.92 1.00
0.50 SBT 1.17 4.43 1.00 1.01 0.30 1.00
0.00 SAP-match 0.03 −0.03 1.00 0.99 0.04 0.05
0.00 SBT −0.03 0.02 1.01 1.00 0.06 0.06
−1.00 SAP-match 0.02 −0.06 0.99 1.00 0.04 0.05
−1.00 SBT −0.05 −0.07 1.02 1.04 0.06 0.05
−2.00 SAP-match 0.03 −0.03 0.98 0.99 0.05 0.05
−2.00 SBT −0.04 −0.00 1.02 1.03 0.06 0.06

Note:Mean (μ), standard deviation (σ), and power or type I error (1β/α) of the one-step value difference test statistic based on 500 simulated data sets. Subscripts “IPW” and “AIPW” correspond to TIPW, and TAIPW,, respectively. “Method” refers to chunking method. Largest standard error for μ, σ, and 1β/α is 0.05, 0.07, and 0.02, respectively.

5.2 |. Right-censored time-to-event

We examine the type I error rate and power of TIPW,SBTS, TAIPW,SBTS, and TCAIPW,SBTS. For brevity, we only ran simulations under the SBT chunking approach. For each simulation, data sets were generated according to the following accelerated failure time (AFT) model: logTi=1.75+0.5X1i+X1i2+0.3X2i+0.2X3i+0.3X4i+0.6X5i+c*Ai*𝟙X2+3X4>0+ei, where c controls whether H0(c0) or H1(c>0) is true, and X1-X5 were generated as in Model 1 scheme 1 of Section 5.1. Under the above model, approximately 25% of the population had a treatment effect of exactly zero (based on 10,000 Monte Carlo samples). We considered two distributions for ei:ei generated as N(0,0.25) (Model 1), or ei=logZi, where Zi is generated as Exponential(1) (Model 2). Model 1 is not a proportional hazards model, whereas Model 2 is. We also varied the propensity model and the model for the censoring time. Specifically, Ai is a Bernoulli random variable with mean πXi=0.5 or πXi=exp(-0.3+0.2X1i+0.6X5i/1+exp-0.3+0.2X1i+0.6X5i. The latter model, used to emulate observational data, is such that the marginal probability that A=1 is approximately 0.5. We considered three models for the censoring time Ci: (a) U(0,50), (b) U(0,100), and (c) exponential with rate exp-13.5+0.1X1i+0.2Ai. We use the notation “Model (1/2) (a/b/c) “ to refer to the combination of failure time model and censoring time model that is of interest. The cutoff time for computing RMST was chosen such that approximately 80–90% of the data are observed by time L. Censoring models (a) and (c) induce higher levels of censoring (by time L), compared with censoring model (b) (See Web Table 16).

We chose n, ln, m, α, and the number of Monte Carlo data sets per simulation as in Section 5.1. The perturbation term, ζi, used to compute TAIPW,SBTS and TCAIPW,SBTS, was generated as Exponential(1). Random survival forest estimates were computed using the R package random-ForestSRC (Ishwaran and Kogalur, 2019). We present results under the observational data setting for Model 1 b and Model 1c. Results from the other models are similar and can be found in the Supporting Information. The type I error is controlled for all simulation scenarios and test statistics considered (see Tables 2, 3, and Web Tables 1727). Within each scenario, TCAIPW demonstrates more power to detect the alternative than TIPW and TAIPW, and TIPW and TAIPW tend to yield identical power. Thus, it appears that any potential efficiency gains that may result from using TAIPW instead of TIPW are negated by the random perturbation term in the specification of TAIPW. By comparing Table 2 with Table 3 (Web Tables 19, 22, and 26 with Web Tables 20, 23, and 27), one can see that the power gain from using TCAIPW instead of either TIPW or TAIPW becomes more pronounced as the amount of censoring increases. By comparing Model 1a with Model 1b (Web Tables 18 and 24 with Web Table 19 and Table 2) and Model 2a versus 2b (Web Tables 21 and 25 with Web Tables 22 and 26), one can see that as the percentage of censoring increases, the power of the test decreases. Finally, as in the uncensored outcome setting, the proposed test remains valid even as p increases from 5 to 25 (see Web Table 17). In Web Appendix L, we examine in both uncensored and censored data scenarios how the choice of m may affect the power of the proposed test via simulations with m={2,5,10,20}. The results in Web Table 29 show that power is insensitive to the choice of m.

TABLE 2.

Results under Model 1b, with πXi=exp-0.3+0.2X1i+0.6X5i/1+exp-0.3+0.2X1i+0.6X5i and L = 42

c n μIPW μAIPW μCAIPW σIPW σAIPW σCAIPW 1βIPW/αIPW 1βAIPW/αAIPW 1βCAIPW/αCAIPW
1.25 600 2.82 3.08 5.79 1.08 1.06 1.09 0.86 0.91 1.00
1.25 1000 3.82 4.03 7.56 1.03 1.02 1.14 0.98 0.99 1.00
0.75 600 1.85 2.09 3.71 1.08 1.06 1.05 0.57 0.67 0.97
0.75 1000 2.50 2.73 4.92 1.03 1.05 1.08 0.82 0.86 1.00
0.50 600 1.18 1.38 2.42 1.08 1.06 1.04 0.33 0.42 0.77
0.50 1000 1.68 1.87 3.27 1.06 1.05 1.04 0.53 0.58 0.95
0.00 600 −0.18 −0.03 −0.09 1.10 1.06 1.00 0.06 0.05 0.03
0.00 1000 −0.15 −0.03 −0.04 1.08 1.03 0.98 0.05 0.06 0.05
−0.50 600 −0.40 −0.22 −0.27 1.05 1.03 1.01 0.03 0.04 0.03
−0.50 1000 −0.33 −0.14 −0.19 1.03 0.97 0.98 0.03 0.03 0.03
−1.00 600 −0.32 −0.13 −0.14 1.07 1.10 1.07 0.02 0.06 0.05
−1.00 1000 −0.21 0.05 0.04 1.01 0.99 0.98 0.02 0.05 0.05

Note: Mean (μ), standard deviation (σ), and power or type I error (1β/α) of the one-step value difference test statistic is based on 500 simulated data sets of n = {600, 1000}. Subscripts IPW, AIPW, and CAIPW specify whether the results are based on TIPW,SBTS,TAIPW, SBTS, or TCAIPW, SBTS, respectively. Largest standard error for μ and σ, and 1β/α is 0.05 and 0.08, and 0.02, respectively.

TABLE 3.

Results under Model 1c, with πXi=exp-0.3+0.2X1i+0.6X5i/1+exp-0.3+0.2X1i+0.6X5i and L=33

c n μIPW μAIPW μCAIPW σIPW σAIPW σCAIPW 1βIPW/αIPW 1βAIPW/αAIPW 1βCAIPW/αCAIPW
1.25 600 1.84 1.99 4.86 1.05 1.07 1.10 0.58 0.63 1.00
1.25 1000 2.62 2.72 6.48 1.08 1.06 1.12 0.81 0.86 1.00
0.75 600 1.24 1.39 3.24 1.06 1.07 1.03 0.35 0.39 0.93
0.75 1000 1.84 1.96 4.40 1.08 1.07 1.05 0.57 0.64 1.00
0.50 600 0.78 0.94 2.20 1.09 1.10 1.02 0.23 0.27 0.70
0.50 1000 1.27 1.38 2.97 1.08 1.08 1.04 0.37 0.40 0.89
0.00 600 −0.21 −0.08 −0.04 1.08 1.11 1.01 0.05 0.06 0.05
0.00 1000 −0.09 0.02 −0.02 1.08 1.09 1.06 0.05 0.06 0.06
−0.50 600 −0.42 −0.29 −0.31 1.06 1.08 1.07 0.03 0.04 0.03
−0.50 1000 −0.28 −0.10 −0.17 1.00 1.00 0.99 0.02 0.04 0.03
−1.00 600 −0.30 −0.15 −0.17 1.08 1.10 1.06 0.04 0.05 0.05
−1.00 1000 −0.19 0.04 0.03 1.03 0.98 0.99 0.02 0.06 0.05

Note: Entries are as in Table 2.

6 |. APPLICATION TO A PHASE III TRIAL

In the Phase III trial referenced in Lipkovich et al. (2017), 599 patients with hematological malignancies were randomized to either experimental therapy plus best supporting care (active treatment, n=303) or best supporting care (control, n=296). The hazard ratio (treatment vs. control) for overall survival was borderline significant (hazard ratio = 0.85, one-sided p=0.0367). We use the proposed one-step value difference test to test whether at least some subgroup benefits from the active treatment. Specifically, we are testing whether at least some subgroup has improved restricted mean survival time under active treatment, compared to control. We considered the same 14 covariates described in Lipkovich et al. (2017); refer to Web Appendix J for a list of the covariates. Eight patients had unknown prognostic score for myelodysplastic syndromes risk assessment, which was one of the covariates collected. Those eight patients were excluded from the proposed test. We chose L=510 days such that about 90% of the Ui are observed by L. This is similar to how we chose L in our simulations, as it prevents K^c,jUiLXi,Ai from being too close to zero. We let ln=291 and m=10, and we randomly allocated the observed data to each of the 31 chunks. Results based on TIPW,SBTS, TAIPW,SBTS, and TCAIPW,SBTS are presented in Table 4. Using the CAIPW estimator, we find evidence (α=0.05) that at least some patients have a larger RMST under active treatment versus control; however, we do not find such evidence when using the IPW or AIPW estimator. This emphasizes the importance of using the CAIPW estimator, especially when the data are heavily censored. For these data, approximately 83% of the event times were censored. Given that there is evidence of treatment benefit for at least some patients (based on CAIPW estimator), the next step would be to identify the subgroup(s) and/or estimate an optimal treatment regime. While not the focus of this paper, this can be done under a variety of approaches. For example, Laber and Zhao (2015), Tao et al. (2018), and Sun and Wang (2021) propose various tree-based methods for estimating optimal treatment rules. Refer to Web Appendix K (and corresponding outputs in Web Figures 2 and 3 and Web Table 28) for an additional example, this time involving an uncensored endpoint.

TABLE 4.

Results from testing the hypothesis in (2) on the Phase III trial data

Estimator Test statistic p-Value
IPW −2.91 0.998
AIPW −2.94 0.998
CAIPW 1.66 0.0481

7 |. DISCUSSION

We have proposed a one-step value difference test for the existence of a subgroup with a beneficial treatment effect that yields root-n-rate inference for Ψdopt=Vdopt-V(0). The test is applicable to cases when the outcome of interest is an uncensored, continuous variable or when the outcome of interest is a right-censored time-to-event. While our implementation uses random forests to estimate C0(X) and τ(X), provided (C4) and (C5) of LL16 are met, alternative regression estimation techniques can be used. One of the benefits to using random forests is that they are more robust to the “curse of dimensionality” than some other nonparametric estimation methods. Extension of the proposed framework to accommodate more than two treatments may be feasible. For example, if two active treatments (“active 1” and “active 2”) are to be compared to a control, one may consider redefining the test statistic to be the arg max of two value difference test statistics—one based on a value difference estimator for active 1 versus control and the other based on a value difference estimator for active 2 versus control. Furthermore, building off the ideas in Zhang et al. (2013) for the uncensored outcomes setting, and Hager et al. (2018) and Jiang et al. (2017) for the censored outcome setting, one may consider extending the proposed subgroup test to multiple treatment stages. These are areas of future work.

Supplementary Material

supplement
code

ACKNOWLEDGMENTS

The authors would like to thank two referees and the associate editor for their invaluable suggestions.

Footnotes

SUPPORTING INFORMATION

Web Appendices, Tables, and Figures referenced in Sections 26 are available with this paper at the Biometrics website on Wiley Online Library along with R code for performing RCT and observational study simulations for an uncensored outcome under Model 1, and for a censored outcome under Model (1/2) (a/b/c).

Data S1

Figure 1. Diagram of how the data are partitioned into chunks. “obs.” stands for observations.

Figure 2. Classification tree for subgroup membership under the ACTG175 data analysis with zidovudine + didanosine as treatment.

Figure 3. Density plot of age among patients predicted to benefit from zidovudine + didanosine (dark grey fill), and among patients predicted not to benefit from zidovudine + didanosine (light grey fill).

Table 1. Model 1 randomized study results for TIPW,SAP-match computed using σ^j in place of σ^j,mπ under scheme 1, π=0.5, α=0.05, n=1000.

Table 2. Model 1 observational data results for TIPW,SAP-match, under scheme 1, π0.5, α=0.05, n=1000.

Table 3. Model 1 randomized study results for TAIPW,SBT under scheme 2, π=0.5, α=0.05, n=1000. Largest standard error for μ, σ, and 1-β/α is 0.05, 0.07, and 0.02 respectively.

Table 4. Model 1 randomized study results under scheme 1, α=0.05, n=600.

Table 5. Model 1 randomized study results under scheme 1, α=0.05, n=1000.

Table 6. Model 2 randomized study results under scheme 1, α=0.05, n=600.

Table 7. Model 2 randomized study results under scheme 1, α=0.05, n=1000.

Table 8. Model 3 randomized study results under scheme 1, α=0.05, n=600.

Table 9. Model 3 randomized study results under scheme 1, α=0.05, n=1000.

Table 10. Model 1 observational study results under scheme 1, α=0.05, n=600.

Table 11. Model 1 observational study results under scheme 1, α=0.05, n=1000.

Table 12. Model 2 observational study results under scheme 1, α=0.05, n=600.

Table 13. Model 2 observational study results under scheme 1, α=0.05, n=1000.

Table 14. Model 3 observational study results under scheme 1, α=0.05, n=600.

Table 15. Model 3 observational study results under scheme 1, α=0.05, n=1000.

Table 16. Estimated range of % censoring before time L for c=-1,-0.5,0,0.5,0.75,1.25 in the models listed above.

Table 17. Results under high-dimensional version (scheme 2) of Model 1c, with πXi=0.5 and L=33.

Table 18. Results under Model 1a, with πXi=0.5 and L=30.

Table 19. Results under Model 1b, with πXi=0.5 and L=42.

Table 20. Results under Model 1c, with πXi=0.5 and L=33.

Table 21. Results under Model 2a, with πXi=0.5 and L=30.

Table 22. Results under Model 2b, with πXi=0.5 and L=42.

Table 23. Results under Model 2c, with πXi=0.5 and L=33.

Table 24. Results under Model 1a, with πXi=exp-0.3+0.2X1i+0.6X5i/1+exp-0.3+0.2X1i+0.6X5i and L=30.

Table 25. Results under Model 2a, with πXi=exp-0.3+0.2X1i+0.6X5i/1+exp-0.3+0.2X1i+0.6X5i and L=30.

Table 26. Results under Model 2b, with πXi=exp-0.3+0.2X1i+0.6X5i/1+exp-0.3+0.2X1i+0.6X5i and L=42.

Table 27. Results under Model 2c, with πXi=exp-0.3+0.2X1i+0.6X5i/1+exp-0.3+0.2X1i+0.6X5i and L=33.

Table 28. Results from application to ACTG175 data

Table 29. Results are based on 500 simulated data sets.

Data S2

DATA AVAILABILITY STATEMENT

The data that support the findings in this paper are available as follows. The data analyzed in Section 6 are available by request from the authors of Lipkovich et al. (2017). The data from AIDS Clinical Trials Group 175 analyzed in Web Appendix K are available on request from the AIDS Clinical Trials Group (https://actgnetwork.org) and from the National Technical Information Service (https://ntis.gov).

REFERENCES

  1. Bai X, Tsiatis AA, Lu W and Song R (2017) Optimal treatment regimes for survival endpoints using a locally-efficient doubly-robust estimator from a classification perspective. Lifetime Data Analysis, 23(4), 585–604. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Bai X, Tsiatis AA and O’Brien SM (2013) Doubly-robust estimators of treatment-specific survival distributions in observational studies with stratified sampling. Biometrics, 69(4), 830–839. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Breiman L (2001) Random forests. Machine Learning, 45(1), 5–32. [Google Scholar]
  4. Cai T, Tian L, Wong P and Wei L (2010) Analysis of randomized comparative clinical trial data for personalized treatment selection. Biostatistics, 12, 270–282. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Fan A, Song R and Lu W (2017) Change-plane analysis for subgroup detection and sample size calculation. Journal of the American Statistical Association, 112(518), 769–778. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Foster JC, Taylor JMG and Ruberg SJ (2011) Subgroup identification from randomized clinical trial data. Statistics in Medicine, 30(24), 2867–2880. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Goldberg Y and Kosorok MR (2012) Q-learning with censored data. Annals of Statistics, 40(1), 529–560. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Hager R, Tsiatis AA and Davidian M (2018) Optimal two-stage dynamic treatment regimes from a classification perspective with censored survival data. Biometrics, 74, 1180–1192. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Ishwaran H and Kogalur U (2019) Fast Unified Random Forests for Survival, Regression, and Classification (RF-SRC). R Package Version 2.9.1. [Google Scholar]
  10. Ishwaran H, Kogalur UB, Blackstone EH and Lauer MS (2008) Random survival forests. Annals of Applied Statistics, 2(3), 841–860. [Google Scholar]
  11. Jiang R, Lu W, Song R and Davidian M (2017) On estimation of optimal treatment regimes for maximizing t-year survival probability. Journal of the Royal Statistical Society Series B Statistical Methodology, 79, 1165–1185. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Jiang R, Lu W, Song R, Hudgens M and Naprvavnik S (2017) Doubly robust estimation of optimal treatment regimes for survival data-with application to an HIV/AIDS study. Annals of Applied Statistics, 11, 1763–1786. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Kang S, Lu W and Song R (2017) Subgroup detection and sample size calculation with proportional hazards regression for survival data. Statistics in Medicine, 36, 4646–4659. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Laber EB and Zhao YQ (2015) Tree-based methods for individualized treatment regimes. Biometrika, 102, 501–514. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Li J, Yue M and Zhang W (2019) Subgroup identification via homogeneity pursuit for dense longitudinal/spatial data. Statistics in Medicine, 38, 3256–3271. [DOI] [PubMed] [Google Scholar]
  16. Lipkovich I, Dmitrienko A and D’Agostino RB, Sr, (2017) Tutorial in biostatistics: data-driven subgroup identification and analysis in clinical trials. Statistics in Medicine, 36, 136–196. [DOI] [PubMed] [Google Scholar]
  17. Lipkovich I, Dmitrienko A, Denne J and Enas G (2011) Subgroup identification based on differential effect search—a recursive partitioning method for establishing response to treatment in patient subpopulations. Statistics in Medicine, 30(21), 2601–2621. [DOI] [PubMed] [Google Scholar]
  18. Lu M, Sadiq S, Feaster DJ and Ishwaran H (2018) Estimating individual treatment effect in observational data using random forest methods. Journal of Computational and Graphical Statistics, 27(1), 209–219. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Luedtke AR and van der Laan MJ (2016) Statistical inference for the mean outcome under a possibly non-unique optimal treatment strategy. Annals of Statistics, 44(2), 713–742. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Murphy S (2003) Optimal dynamic treatment regimes (with discussions). Journal of the Royal Statistical Society Series B (Statistical Methodology), 65, 331–366. [Google Scholar]
  21. Rubin D (1974) Estimating causal effects of treatments in randomized and nonrandomized studies. Journal of Educational Psychology, 66(5), 688–701. [Google Scholar]
  22. Shen J and He X (2015) Inference for subgroup analysis with a structured logistic-normal mixture model. Journal of the American Statistical Association, 110(512), 303–312. [Google Scholar]
  23. Shi C, Lu W and Song R (2020) A sparse random projection-based test for overall qualitative treatment effects. Journal of the American Statistical Association, 115(531), 1201–1213. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Sun Y and Wang L (2021) Stochastic tree search for estimating optimal dynamic treatment regimes. Journal of the American Statistical Association, 116, 421–432. [Google Scholar]
  25. Tao Y, Wang L and Almirall D (2018) Tree-based reinforcement learning for estimating optimal dynamic treatment regimes. Annals of Applied Statistics, 12, 1914–1938. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Tsiatis A, Davidian M, Holloway S and Laber E (2020) Dynamic Treatment Regimes: Statistical Methods for Precision Medicine. Boca Raton, FL: Chapman & Hall/CRC Press. [Google Scholar]
  27. Wang J, Li J, Li Y and Wong W (2019) A model-based multithreshold method for subgroup identification. Statistics in Medicine, 38, 2605–2631. [DOI] [PubMed] [Google Scholar]
  28. Watkins C and Dayan P (1992) Q-learning. Machine Learning, 8, 279–292. [Google Scholar]
  29. Wei S and Kosorok M (2018) The change-plane Cox model. Biometrika, 105, 891–903. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Wu R, Zheng M and Yu W (2016) Subgroup analysis with time-to-event data under a logistic-Cox mixture model. Scandinavian Journal of Statistics, 43, 863–878. [Google Scholar]
  31. Zhang B, Tsiatis A, Davidian M, Zhang M, and Laber E (2012a). Estimating optimal treatment regimes from a classification perspective. Stat, 1, 103–114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Zhang B, Tsiatis A, Laber E, and Davidian M (2012b). A robust method for estimating optimal treatment regimes. Biometrics, 68, 1010–1018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Zhang B, Tsiatis AA, Laber EB and Davidian M (2013) Robust estimation of optimal dynamic treatment regimes for sequential treatment decisions. Biometrika, 100, 681–694. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Zhao L, Tian L, Cai T, Claggett B and Wei LJ (2013) Effectively selecting a target population for a future comparative study. Journal of the American Statistical Association, 108(502), 527–539. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Zhao Y, Zeng D, Rush A and Kosorok M (2012) Estimating individualized treatment rules using outcome weighted learning. Journal of the American Statistical Association, 107, 1106–1118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Zhao YQ, Zeng D, Laber EB, Song R, Yuan MH and Kosorok MR (2015) Doubly robust learning for estimating individualized treatment with censored data. Biometrika, 102, 151–168. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

supplement
code

Data Availability Statement

The data that support the findings in this paper are available as follows. The data analyzed in Section 6 are available by request from the authors of Lipkovich et al. (2017). The data from AIDS Clinical Trials Group 175 analyzed in Web Appendix K are available on request from the AIDS Clinical Trials Group (https://actgnetwork.org) and from the National Technical Information Service (https://ntis.gov).

RESOURCES