Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2022 Dec 4.
Published in final edited form as: Biometrics. 2021 Jun 15;78(4):1503–1514. doi: 10.1111/biom.13504

Estimating mean potential outcome under adaptive treatment length strategies in continuous time

Hao Sun 1, Ashkan Ertefaie 1, Brent A Johnson 1
PMCID: PMC9482146  NIHMSID: NIHMS1826298  PMID: 34086345

Abstract

An adaptive treatment length strategy is a sequential stage-wise treatment strategy where a subject’s treatment begins at baseline and one chooses to stop or continue treatment at each stage provided the subject has been continuously treated. The effects of treatment are assumed to be cumulative and, therefore, the effect of treatment length on clinical endpoint, measured at the end of the study, is of primary scientific interest. At the same time, adverse treatment-terminating events may occur during the course of treatment that require treatment be stopped immediately. Because the presence of a treatment-terminating event may be strongly associated with the study outcome, the treatment-terminating event is informative. In observational studies, decisions to stop or continue treatment depend on covariate history that confounds the relationship between treatment length on outcome. We propose a new risk-set weighted estimator of the mean potential outcome under the condition that time-dependent covariates update at a set of common landmarks. We show that our proposed estimator is asymptotically linear given mild assumptions and correctly specified working models. Specifically, we study the theoretical properties of our estimator when the nuisance parameters are modeled using either parametric or semiparametric methods. The finite sample performance and theoretical results of the proposed estimator are evaluated through simulation studies and demonstrated by application to the Enhanced Suppression of the Platelet Receptor IIb/IIIa with Integrilin Therapy (ESPRIT) infusion trial data.

Keywords: adaptive treatment strategies, causal inference, informative eligibility, survival analysis, treatment competing events, treatment discontinuation

1 |. INTRODUCTION

An adaptive treatment length strategy is a sequential dynamic treatment strategy to give treatment continuously from baseline to a point in time when a subject chooses to stop treatment or until a competing treatment-terminating event occurs, whichever comes first. In both the adaptive treatment length strategy and the observational study, treatment length is determined sequentially as decisions to stop or continue treatment are repeated over time, as long as the subject has been treated continuously up to that time when a decision to stop or continue is made. In the observational data, the decision to stop or continue treatment also depends on covariate history up to the current time. In short, the broad scientific objective is to describe how mean outcome changes as a function of treatment length when the decision to stop or continue throughout the treatment window depends on a history of tailoring information that induces confounding by indication in the observational study. Applications of adaptive treatment length strategies include informative treatment discontinuation in infusion studies (Johnson and Tsiatis, 2004, 2005; Zhang et al., 2011; Yang et al., 2018), switching antiretroviral therapies (Johnson et al., 2013), and palliative care for patients suffering from neuromuscular diseases (Lu and Johnson, 2017). By swapping treatment initiation for treatment discontinuation and the role of intermediate events, Hu et al. (2018) showed how to extend these ideas to evaluate strategies for initiating combination antiretroviral therapies in the presence of comorbidities, such as tuberculosis (see also, Robins et al., 2008).

Estimation of causal estimands for adaptive treatment strategies (ATSs), or dynamic treatment regimes, from longitudinal data with time-varying confounders has well-known challenges with, by now, equally well-known methods to overcome such challenges. The successful application of existing methods to adaptive treatment length strategies depends critically on the observed data and methodological assumptions. To fix ideas, a dynamic treatment regime d¯K is a sequence of decision rules d¯K=(dt,tT), where the decision rule dt : dom(Ht) → dom(At), Ht is the combined treatment and covariate histories up to time t, At is treatment selection at time t, dom(·) is short-hand for the domain of the argument and the time domain is T. Compared with static regimes that are fixed a priori at the beginning of the study, dynamic regimes are adaptive in that treatment selection at time t depends on a subject’s history of data up to that point. If AK is the collection of all possible treatment vectors a¯K=(a1,,aK), then the potential outcome of interest under regime d¯K is

Y(d¯K)=a¯KAKY(a¯K)I{d1(H1)=a1}I{dK(HK)=aK}, (1)

where Y(a¯K) is the potential response had a subject followed treatment regime a¯K. More generally, the potential outcome Y(d¯K) is defined through the sequence of stochastic rules, say p¯d¯K=(p1,d¯K,,pK,d¯K) with treatment assignment at the t-th stage in the dynamic regime given by pt,d¯K(at|Ht),t=1,,K. Without loss of generality, we will refer to the potential outcome Y(d¯K) under both general stochastic treatment assignment mechanisms and the degenerate case when pt,d¯K(at|Ht)=I{dt(Ht)=at}. The objective is to estimate the value function, E[Y(d¯K)]. In our application to adaptive treatment length strategies in Section 5, treatment assignment At is a binary indicator of whether treatment has been stopped by time t, and Ct is the indicator that treatment has been stopped due to involuntary adverse treatment-terminating events by time t. Then our objective in discrete time may be summarized as estimating E[Y(d¯K)] for ATS d¯K defined by the rule, dt(Ht) = 1 − I(tto)I(Ct = 0), that treats continuously until provider stops treatment at to or until a treatment-terminating event, whichever comes first.

In the observational treatment length study, decisions to stop treatment by choice depend on time-varying covariate information. Because time-dependent covariates occur downstream of earlier treatment choices but prior to the final outcome of interest, this leads to a classic case of confounding and one challenge to overcome in estimating the causal estimand. A second challenge that our proposed method attempts to overcome is to estimate the causal estimand when the time domain T is an interval. When the domain T is a finite number of time points, other authors (Johnson and Tsiatis, 2004; Lu and Johnson, 2017) have shown how to apply causal inference methods for dynamic treatment regimes (Robins, 1997; Robins et al., 2000; Murphy et al., 2001) to estimate the causal estimand from observational data. However, when the treatment assignment is a continuous random variable or, in our case, the domain T is an interval, the literature of relevant methodology is less robust. Johnson and Tsiatis (2005) proposed an inverse probability weighted estimator of E[Y(d¯K)] under a continuous treatment assignment mechanism when the value function is assumed to be a known parametric function in treatment length. Lu and Johnson (2015) proposed a direct estimator of E[Y(d¯K)] assuming baseline confounding but no time-varying confounders. Methods for continuous treatment assignment mechanisms in related problems have been proposed elsewhere in the literature (e.g., Lok, 2008; Zhang et al., 2011; Díaz and van der Laan, 2013; Kennedy et al., 2017). However, these methods are not directly applicable to our problem as they ignore the informative nature of the treatment-terminating event.

In this paper, we propose an estimator that is initially motivated by Robins’ (1986) G-computation formula for estimating the mean potential outcome of a static treatment regime from longitudinal data extended to the case of continuously varying covariates and treatments (Gill and Robins, 2001). One practical difficulty with G-estimation in continuous time is modeling general covariate histories that can change in infinitesimals of time. To avoid this challenge, our method uses a variation of risk-set weighting and, as motivated by our application, assumes that covariate histories are updated at a finite number of landmarks. Then, our estimator is constructed as stochastic integrals of weighted predictions from regression models for the set of potential outcomes as functions of covariate histories. Because our estimator does not model the treatment assignment mechanism, our method may be applied to data irrespective of the domain T for the treatment length random variable. We study the asymptotic behavior of the proposed estimator and show that, under regularity conditions, our estimator is asymptotically linear.

2 |. BACKGROUND

While the text in Section 1 introduced basic ideas of adaptive treatment length strategies using notational conventions for ATSs in discrete time, the methods here are tailored for causal inference in continuous time for which appropriate modifications are needed. Section 2.1 gives the notation used in the sequel followed by g-formula principles in Sections 2.2 and 2.3 used to motivate our proposed approach in Section 3.

2.1 |. Notation

In our setup, the presumption is that treatment decisions are made repeatedly in infinitesimals [t, t + dt) throughout the treatment window T=[0,τ], provided that the subject has been treated up to time t. Hence, let At represent the counting process that treatment is stopped by time t, A¯t={As,s[0,t]} is treatment history up to and including time t, a¯t is a sample path of treatment assignment to time t, and let Aτ be the collection of all sample paths. Similarly, define Ct(a¯t) as the random indicator that a treatment-terminating event has occurred by time t if a patient is following treatment sequence a¯t; in particular, let Ct(a¯t) realize the value +1 if treatment has stopped by time t due to treatment-terminating event. Let Xt(a¯t) be a potential auxiliary time-dependent covariate if following treatment sequence a¯t and X¯t(a¯t) is the covariate history through time t. The potential outcome Y(a¯τ) is defined as the outcome if a patient follows the treatment sequence a¯τ. The complete set of potential treatment-terminating events and outcome for the adaptive treatment length study is {Ct(a¯t),Y(a¯τ),a¯τAτ} and the complete collection of auxiliary covariates is {Xt(a¯t),a¯τAτ}. Thus, if we let Lt(a¯t)={Ct(a¯t),Xt(a¯t)}, then the complete set of potential random variables is Opo={Lt(a¯t),Y(a¯τ),a¯τAτ,t[0,τ]}. By the stable unit treatment value assumption (SUTVA; Rubin, 1978), a subject’s treatment assignment is not affected by the treatment assignment of others. Hence, the observed data are O={Lt(A¯t),At,Y(A¯τ),t[0,τ]} and written concisely as O = {Lt, At, Y, t ∈ [0, τ]}. The observed data for all n subjects are O1, … , On and assumed to be independent random vectors from a common distribution.

In order to identify the causal estimand, we require that treatment assignment at time t be a function of covariate and treatment histories up to time t but not a function of information beyond time t. This is known as the sequential randomization assumption and is summarized in (2),

At{Lt+dt(a¯t),,Lτ(a¯τ),Y(a¯τ)}|Ht, (2)

where Ht=(L¯t,A¯t). To describe the treatment assignment mechanism in the observational study, we define the instantaneous probability of stopping a subject’s treatment by choice at time t, provided the subject is still receiving treatment at time t, as

λt,obs(Ht)=limdt0P(dAt=1|At=0,Ct=0,X¯t)/dt. (3)

Using our above notation where dAt is a Bernoulli event for stopping treatment by choice in [t, t + dt), we heuristically define the intensity of treatment assignment in the observational study as P(dAt=1|Ht)=dΛ˜t,obs(Ht), where dΛ˜t,obs(Ht)=I(At=0,Ct=0)λt,obs(Ht)dt. Using this notation, the treatment assignment mechanism in the interval [t, t + dt) from the observational study is written as pt,obs(dAt|Ht)=dΛ˜t,obs(Ht)dAt{1dΛ˜t,obs(Ht)}, and under the conditional independence assumption in (2), pt,obs(|Ht)=pt,obs(|Hτ,Y(a¯τ)).

Analogous to treatment assignment in the observational study, we define treatment assignment in the ATS through integrated intensity processes, that is, pt,d¯τ(dAt|Ht)=dΛ˜t,d¯τ(Ht)dAt{1dΛ˜t,d¯τ(Ht)}, where Pd¯τ(dAt=1|Ht)=dΛ˜t,d¯τ(Ht) and Pd¯τ is probability measure in the ATS. Consequently, using the treatment assignment mechanism in the ATS, {pt,d¯τ(|Ht),t[0,τ]}, we may define the potential outcome

Y(d¯τ)=a¯τAτY(a¯τ)π[0,τ]pt,d¯τ(at(dt)|Ht),

where π[0,τ] is the product integral from 0 to τ and the target causal estimand in continuous time is the value function E[Y(d¯τ)].

2.2 |. G-computation in continuous time

Robins’ (1986) G-computation method is a powerful technique that is widely used to infer the causal effect from longitudinal observational data. In principle, Robins’ G-computation formula is used to express the distribution of potential outcome on a dynamic treatment regime as nested functions of the observed data. A plug-in estimator is derived by replacing unknown distributions with parametric families of statistical models, estimating unknown parameters from the observed data, and evaluating the causal estimand with fitted quantities. In our setting, where the time-varying covariates are measured only up to the treatment stopping time, the G-computation method is challenging because it requires predictions from correctly specified models of complex stochastic processes in continuous time.

To gain a better insight into the challenge of G-computation in our setting and the proposed remedy, we describe the causal estimand and likelihood for the adaptive treatment length problem. Now, for a specified ATS, d¯τ, the causal estimand E[Y(d¯τ)] may be written as

a¯τAτl¯τyyPA¯τ(d¯τ)(hτ+)×{fY(y|a¯τ,l¯τ)fL¯τ(l¯τ|a¯τ)}dydv(l¯τ), (4)

where a¯τAτ indicates all treatment paths a¯τAτ consistent with d¯τ, PA¯τd¯τHτ+=π[0,τ]pt,d¯τdAtHt with Hτ+(Aτ,Hτ), fY is the conditional density of the endpoint Y given the entire covariate and treatment histories, and fL¯τ is the conditional density of L¯τ given treatment history A¯τ=a¯τ, and ν is the dominating measure for L¯τ. Without loss of generality, we write fL¯τ as an infinite product of joint conditional densities of dLt in the infinitesimal [t, t + dt) given L¯t=l¯t and that a subject has been continuously treated up to time t, that is,

fL¯τ(l¯τ|a¯τ)=π[0,τ]P(dLtdlt|A¯t=a¯t,L¯t=l¯t), (5)

where P(dL0dl0|A0,L0)=P(dL0dl0). The likelihood of the data is

L=PA¯τ(obs)(Hτ+){fY(Y|A¯τ,L¯τ)fL¯τ(L¯τ|A¯τ)}, (6)

where PA¯τ(obs)(Hτ+)=π[0,τ]pt,obs(dAt|Ht). Because the estimand in (4) is a function of parameters in curly brackets in (6) and randomization probabilities pt,d¯τ(|Ht) in PA¯τ(d¯τ) that are known a priori by design, the causal estimand can be estimated by replacing the unknown parameters with estimates informed by the likelihood L in (6). This is, in principle, Robins’ (1986) G-computation method for the adaptive treatment length strategy d¯τ in continuous time.

When Lt is measured at a finite number of grid points, fL¯τ in (5) reduces to a finite product of conditional densities (e.g., Murphy et al., 2001, (4.4)). But if any part of Lt is measured in continuous time, then fL¯τ in (5) does not simplify and is difficult to model using familiar stochastic processes, in general. First, L¯τ={Lt,t[0,τ]} is the combined histories of the terminating event C¯τ, that is, a point process, and covariate history X¯τ. Second, even if we could develop a method using a convenient Gaussian process, for example, to model covariate history X¯τ, it is rarely the case that the covariate history is a vector of continuous variables only and such a method would be of limited practical value. Positing a flexible model for the event history L¯τ and estimating the parameters of such a model from a selective biased sample of subjects are two practical challenges for any direct plug-in estimator of (4) via G-computation formula. In Section 2.3, we show how a common set of landmark updates to covariate history simplifies the expression in (5).

2.3 |. fL¯τ under common landmark updates to covariate histories

Now we assume that while treatment could be stopped at any time in the interval (0, τ], either by choice or treatment-terminating event, covariate information updates only at a finite, common number of landmarks, 0 < t1 < t2 < ⋯ < tK < τ. To avail ourselves of this special simplifying structure on the covariate process, we split L¯t=(C¯t,X¯t) into covariate history X¯t and history of terminating events C¯t. We write P(dLt|A¯t,L¯t) as the product P(dXt|A¯t,L¯t)P(dCt|A¯t,L¯t,Xt), where P(dCt|A¯t,L¯t,Xt) is the probabilistic mechanism for treatment-terminating event in the interval [t, t + dt). By assuming that the covariate history is left-continuous, we have that P(dCt|A¯t,C¯t,X¯t)=dΛ˜tξ(Ht)dCt{1dΛ˜tξ(Ht)}, where P(dCt=1|A¯t,L¯t,Xt)=dΛ˜tξ(Ht), dΛ˜tξ(Ht)=I(At=0,Ct=0)λtξ(Ht)dt, and λtξ(Ht) is the hazard of a treatment-terminating event at time t given covariate history and that the subject has been continuously treated to time t. Furthermore, a piece-wise constant assumption on covariate history implies that P(dXt|A¯t,L¯t) is a density function at the landmark tj, given covariate history up to tj and that a subject is continuously treated to time tj for j = 1, … , K, and is degenerate otherwise. Define P(dXtdxt|X¯t=x¯t)=qt(xt|x¯t) and P(dXtdxt|A¯t=a¯t,L¯t=l¯t)=q˜t(xt|a¯t,l¯t) with q˜t(xt|a¯t,l¯t)=qt(xt|x¯t)I(at=0,ct=0). Then, the density fL¯τ(l¯τ(t)|a¯τ) simplifies to

fL¯τ*(l¯τ|a¯τ)=π[0,τ]{dΛ˜tξ(Ht)}dct{1dΛ˜tξ(Ht)}×j=1Kq˜tj(xtj|a¯tj,l¯tj). (7)

Therefore, under a piece-wise constant assumption on covariate history, we rewrite the likelihood L in (6) as:

L=PA¯τ(obs)(Hτ+)fY(Y|A¯τ,L¯τ)fL¯τ*(L¯τ|A¯τ), (8)

where fL¯τ* is given in (7). Our landmark condition on covariate history similarly affects our expression (4) for the causal estimand E[Y(d¯τ)], that is,

a¯τAτl¯τyyPA¯τ(d¯τ)(Hτ+)fY(y|a¯τ,l¯τ)×fL¯τ**(l¯τ|a¯τ)dydν*(l¯τ), (9)

where ν* is the dominating measure for L¯τ under a piece-wise constant assumption on covariate history. Expression (9) leads directly to a plug-in G-computation estimator for general dynamic treatment regime d¯τ given by the randomization probabilities pt,d¯τ(|Ht), where the treatment-terminating event can occur at any time in the interval (0, τ] and the covariate history is piece-wise constant with updates at landmarks t1, … , tK. Namely, if the functions in (9) can be estimated consistently, then replacing the unknown functions with their estimated counterparts leads to a plug-in estimator of E[Y(d¯τ)].

3 |. METHODS

In order to derive a plug-in estimator from expression (9), each conditional density qt(xt|x¯t) must be modeled for each landmark tj. Such models are arguably easier in discrete time than in continuous time but can still be a formidable practical challenge because of the arbitrary nature of covariate history X¯t. Also, because the covariate process fails to update at landmarks after treatment has stopped, we have missing data for all subjects with observed treatment length less than than target treatment length. Therefore, we have incomplete data to estimate the parameters of qt(xt|x¯t) and, because of informative treatment discontinuation, data get both sparser and more selective as the landmark tj increases toward τ, the right-hand endpoint of the treatment window. Applying a simple degenerate data structure—for example, projection via last observation carried forward (Petersen et al., 2014) or projection through eligible subjects only (Lu and Johnson, 2017)—eases the computational burden but can lead to inconsistent estimation of qt(xt|x¯t), and hence E[Y(d¯τ)], when such assumptions are not true. The proposed estimator circumvents some challenges in the modeling of covariate histories through the idea of risk-set weighting.

Let the decision rule dt(Ht) = 1 − I(tto)I(Ct = 0) define the adaptive treatment length strategy d¯τ. We evaluate the causal estimand in (9) by considering each a¯τAτ on a case-by-case basis. Write mY(a¯τ,c¯τ,x¯τ)=E(Y|A¯τ=a¯τ,C¯τ=c¯τ,X¯τ=x¯τ), Gt(Ht)=π[0,t]{1dΛSξ(HS)}.

  • (Case 1) Subject was continuously treated until the provider stopped treatment by choice at time to, that is, a¯τ(1)={as:=I{s>t°},s[0,τ]}. The estimand in (9) evaluates to E[mY(a¯τ(1),c¯τ(1),X¯to)Gto(H˜to(1))], where c¯τ(1)={cs:=0,s[0,τ]}, H˜t(1)=(a¯τ(1),c¯τ(1),X¯t), the constrained treatment and covariate history Ht under (a¯τ(1),c¯τ(1)).

  • (Case 2) Subject was continuously treated until they experienced a random treatment-terminating event at some t < to. For each infinitesimal [t, t + dt), t < to, let a¯τ(2).t={as:=I{s>t},s[0,τ]}, c¯τ(2)t={cs:=I{s>t},s[0,τ]}. Then, at time t, t < to, (9) is E[mY(a¯τ(2)t,c¯τ(2)t,X¯t)dΛtξ(H˜t(2)t)Gt(H˜t(2)t)], where H˜t(2)t=(a¯τ(2)t,c¯τ(2)t,X¯t) is the constrained treatment and covariate history Ht under (a¯τ(2)t,c¯τ(2)t). The full expression under Case 2 is the sum over all infinitesimals [t, t + dt) from 0 to to.

Then, the target estimand (9) is the sum of the marginal expectations in Case 1 and Case 2. Writing the contribution to the summand for Case 2 as a stochastic integral with integrator dGt(Ht)=Gt(Ht)dΛtξ(Ht), the target estimand is

E[mY(a¯τ(1),c¯τ(1),X¯to)Gto(H˜to(1))0tomY(a¯τ(2)t,c¯τ(2)t,X¯t)dGt(H˜t(2)t)], (10)

where the expectation in (10) is taken with respect to the distribution of covariate history X¯τ. Importantly, the conditional mean mY in expression (10) assumes that covariate history affects the outcome Y when the subject is being treated but not after treatment is stopped. This assumption is consistent with available data for the motivating example in Section 5 and would be true, for example, if the reason treatment stopped was due to death. However, if covariate data were collected after treatment was stopped and did affect outcome, the methods proposed here could be extended by replacing X¯τ with X¯τ in mY in (10). We also note that covariate information collected after treatment has stopped cannot affect treatment assignment and cannot, therefore, be a confounding variable regardless of whether it affects the outcome or not.

Unfortunately, the covariate history X¯τ is only observed for those subjects who are still receiving treatment at the most recent landmark, by the landmark assumption in Section 2.3, and so (10) cannot be estimated with an ordinary sample average. For this purpose, let Rt=1At=πst(1AS) be the indicator that a subject is still being treated at time ⌊t⌋ = max{tj|tjt, j = 1, … , K} and let Wt(Ht)=πst{1λsU(Hs)}, where λtU(Ht) is the all-cause hazard of treatment discontinuation at time t given treatment and covariate history Ht. Then, under the sequential randomization assumption, a standard argument in the analysis of observational data using Horwitz–Thompson weighting applied to the “risk-set” indicator Rt leads us to conclude that (10) is equal to the expression

E[RtomY(a¯τ(1),c¯τ(1),X¯to)Gto(H˜to(1))Wto(Hto)0toRtmY(a¯τ(2)t,c¯τ(2)t,X¯t)dGt(H˜t(2)t)Wt(Ht)]. (11)

Substituting consistent estimators m^Y, G^, and W^ for mY, G, and W, respectively, and taking a sample average of the expression in square brackets in (11) leads to our estimator

n[Rtom^Y(a¯τ(1),c¯τ(1),X¯to)G^to(H˜to(1))W^to(Hto)0toRtm^Y(a¯τ(2)t,c¯τ(2)t,X¯t)dG^t(H˜t(2).t)W^t(Ht)]. (12)

Naturally, the performance and behavior of the statistic in (12) is tied closely to the flexibility of the regression functions used to model mY, G, and W. We adopt Cox’s (1972) proportional hazards model for λtξ(Ht) and λtU(Ht) and consider different strategies for modeling mY with varying amounts of flexibility. The large sample behavior of the proposed estimator in (12) is outlined in the Appendix with details given in the Supporting Information. We use bootstrap resampling to estimate the asymptotic variance of our estimator in (12) for its ease in practical implementation by data analysts. Resampling can be justified formally using techniques described in Kosorok (2008, Theorem 10.2), for example. The finite sample behavior of the point and interval estimator are evaluated in Section 4.

Remark 1.

When there is confounding through baseline factors only and those factors are observed on the entire study sample, the risk-set weighting is superfluous and the estimand in (10) can be estimated directly with a sample average as shown by Lu and Johnson (2015). Our method generalizes the estimator by Lu and Johnson (2015) to the case of time-dependent confounding, when covariate history updates at landmark time points t1, … , tK throughout the study follow-up period.

4 |. SIMULATION STUDIES

We conducted simulation studies to evaluate the proposed estimator E[Y(d¯τ)] in finite samples under the ATS d¯τ given by dt(Ht) = 1 − I(tto)I(Ct = 0). The performance of our proposed estimator depends on the choice of modeling and estimation strategies leading to estimators of the regression functions mY, λtξ, and λtU. To facilitate connections and comparison to earlier works, we write the observed data as O=(Y,U,Δ,X¯U), where U=0τ(1At)dt is the observed treatment length, Δ=10τdCt is a binary indicator that realizes the value 0 if treatment was stopped due to treatment-terminating event and 1 otherwise, and X¯U is covariate history through time U (e.g., Johnson and Tsiatis, 2004, 2005; Lu and Johnson, 2015). Through this parameterization, the conditional mean mY(A¯τ,C¯τ,X¯τ)=mY(U,Δ,X¯U),I(Ut,Δ=0) is the counting process for the treatment-terminating event, the at-risk indicator is Rt = I(U ≥ ⌊t⌋), and λtU(Ht) is the hazard function for the at-risk process given history Ht.

Our simulation scenarios are motivated by the data in Section 5. There are a total of five prespecified landmarks are tj ∈ {6, 12, 18, 24, 30} and t0 = 0. We consider two covariates Xt=(X0(1),Xt(2)), where X0(1) is a time-independent standard normal random variable and Xt(2) is a time-dependent covariate. The time-dependent covariate Xt(2) follows a first-order Markov model: Xtk(2) is normal with mean Xtk1(2) and unit variance for k = 1, … , 5, and Xt0(2) is standard normal. The cumulative covariate history is X¯t=(X0,,Xt) and the upper limit of the treatment length window is τ = 30.

Let λt(Ht) be the cause-specific hazard function for treatment discontinuation time with Δ = 1, and λtξ(Ht) be the cause-specific hazard function for treatment discontinuation time with Δ = 0. We assume that both hazard of treatment length choice and λt(Ht) and λtξ(Ht) follow proportional hazard models, λt(Ht;γ)=λ0,texp(γX¯t), and λtξ(Ht;η)=λ0,tξexp(ηX¯t). We use piecewise constant baseline hazard functions for λ0,t and λ0,tξ such that λ0,t=j=1KκjI{t(tj1,tj]} and λ0,tξ=j=1KκjξI{t(tj1,tj]}, where κj and κjξ are the baseline rates in the jth interval for T and ξ, respectively. We assume the outcome models follow the general semiparametric models, mY(U,Δ,X¯t)=Δm1(U,X¯t)+(1Δ)m0(U,X¯t), m0(t,X¯t)=Φ{ϕ0(t)+βX¯t} and m1(t,X¯t)=Φ{ϕ1(t)+αX¯t}, where ϕ1(t)=log(t), ϕ0(t)=1+t1/2. The observed data are thus (Y,U,Δ,X¯U). Note, under this setup, the induced hazard function for U is λtU(Ht)=λtξ(Ht)+λt(Ht). In the interest of space, the results for continuous endpoint with identity link Φ(·) are presented here while the the results for binary endpoint with logit link function are relegated to Web Appendix D. Specific parameters for the simulation results are presented below:

  • Scenario 1 (time-dependent confounding): We set α = β = (0.5, −0.5), and:
    • for a roughly 45% adverse event rate: γ = (−0.3, 0.5), η = (0.3, −0.5), (κ1, … , κ5) = (1.55, 2, 2.8, 4.4, 14)/100, (κ1ξ,,κ5ξ)=(1.35,1.6,2.4,3.8,11)/100;
    • for a roughly 20% adverse event rate: γ = (−0.1, 0.5), (κ1, … , κ5) = (2.2, 3, 4, 6.2, 23), (κ1ξ,,κ5ξ)=(0.6,0.75,1,1,4)/100.
  • Scenario 2 (baseline confounding only): We set α = β = (0, −0.5), and:
    • for a roughly 45% adverse event rate: γ = (0, 0.5), η = (0, −0.5), (κ1, … , κ5) = (1.75, 2.1, 3.3, 4.7, 14)/100, (κ1ξ,,κ5ξ)=(1.4,1.6,2.5,4,11)/100;
    • for a roughly 20% adverse event rate: (κ1, … , κ5) = (2.5, 3, 4, 6.2, 23)/100, and (κ1ξ,,κ5ξ)=(0.6,0.75,1,1,4)/100.

In all scenarios, we generate 500 Monte Carlo data sets of sizes n = 250 and n = 500. The chosen parameters lead to observed treatment length distribution with approximate quantiles of 20%, 40%, 60%, and 80% at the landmarks 6, 12, 18, and 24, respectively. True values of the estimand were computed via numerical integration.

Figure 1 summarizes simulation results for estimating E[Y(d¯τ)] with different to in the presence of time-dependent confounding. For both the proposed estimator and Lu and Johnson (LJ, 2015), mY is fit through parametric models with the correct function of time (see also Web Appendix D for the simulation results where mY is fit by partly linear models). In the presence of time-dependent confounding, LJ is not consistent and the bias does not shrink to zero as the sample size increases as seen in the top row of Figure 1. The bias is especially severe in the middle part of the investigated time interval due to the misspecification of the outcome model mY. On the other hand, the proposed estimator performs well across a range of adverse events rates, 45% or 20%. When the sample size is modest, for example, n = 250, there is some bias in the tails of the treatment window. However, this extraneous finite sample bias diminishes substantially as the sample size increases to n = 500.

FIGURE 1.

FIGURE 1

Simulation results based on 500 Monte Carlo data set. The empirical bias, standard deviation, and coverage rates for two different adverse events rates, 45% and 20%, under time-dependent confounding. The outcome is Gaussian. MCSD is the Monte Carlo standard deviation and SEE is the standard error estimate

The panels in the second row describe the empirical standard error (MCSD, gray line) and the bootstrap-based standard error estimates (SEE, black line) for our proposed estimator (i.e., we do not present the variance estimation results for LJ method, but see Lu and Johnson, 2015). These two curves agree closely over the entire treatment window for both sample sizes suggesting that bootstrap resampling performs well. As exemplified in Figure 1, realizations of the estimator over treatment length can appear as nonsmooth functions in time. But this is a direct consequence of how the proposed estimator adjusts for time-dependent confounding, with informative updates at the landmarks. The last row shows that the empirical coverage rate of LJ may be much lower than nominal level when important time-dependent confounders are ignored. The empirical coverage rate curve of our proposed estimator is close to nominal level, which justifies our asymptotic arguments.

When there is only baseline confounding, LJ showed that their estimator was consistent and asymptotically normal. It is important to compare estimators in this setting to evaluate how much precision is lost due to our estimation scheme to accommodate time-dependent confounders. In simulation scenario 2, only the baseline covariate X0(1) has a nontrivial effect on the outcome. As predicted, Figure 2 shows that both the proposed estimator and LJ are relatively unbiased when time-dependent confounding plays no role. Finite sample bias in the tails disappears as the sample size increases. The second row in Figure 2 shows that the MCSD curve of LJ is 25% smaller than the proposed estimator for treatment duration greater than or equal to 25 with continuous outcomes. This result suggests LJ is more efficient than the proposed estimator when there is confounding through baseline factors only, and one explanation for this result is because there is additional uncertainty in risk-set weighting or any method that attempts to adjust for time-dependent confounding. As seen in Figure 1, the bootstrap-based variance estimation performs well here and the empirical coverage probability is close to the nominal level, which support our theoretical results in Web Appendix A.

FIGURE 2.

FIGURE 2

Simulation results based on 500 Monte Carlo data set. The empirical bias, standard deviation, and coverage rates for two different adverse events rates, 45% and 20%, without time-dependent confounding. The outcome is Gaussian. MCSD is the Monte Carlo standard deviation

Web Appendix D provides additional simulation results to show the performance of the proposed estimator when mY is modeled through the general semiparametric model and estimated via profile least squares. We also give results for the case where mY is misspecified by both parametric and semiparametric models but can be learned by flexible machine learning methods. In particular, we show that the point-wise bias tends to zero as the sample size increases and thus supports our claim of consistency using nonregular methods.

5 |. ANALYSIS OF THE INFUSION TRIAL DATA

The proposed methods are motivated by the Enhanced Suppression of the Platelet Receptor IIb/IIIa with Integrilin Therapy (ESPRIT), a double-blind, randomized, placebo-controlled clinical trial of a novel dosing regimen (O’Shea et al., 2001). Briefly, patients undergoing elective stent percutaneous coronary intervention were given two bolus doses of eptifibatide followed by eptifibatide or placebo infusion just prior to coronary intervention for patients on the intervention or standard of care arm, respectively. If an adverse event occurred, the protocol advised that infusion be stopped immediately; in the absence of such adverse events, providers decided when to stop the infusion. Then, the adaptive treatment length policy in this context is to infuse for a fixed amount of time t or until an infusion-terminating event occurs, whichever comes first.

To facilitate a comparison to earlier work, we considered similar clinical definitions for adverse event and clinical endpoint. We define the outcome Y as the composite binary endpoint of death, myocardial infarction, or urgent target vessel revascularization within 30 days. Furthermore, an adverse event necessitating infusion termination included any combination of abrupt closure, no reflow, or coronary thrombosis. The baseline covariate data include indicators of diabetes, percutaneous transluminal coronary angioplasty, angina, heparin use, and the continuous variable weight. Enzyme level is the one time-dependent covariate and was collected at landmarks of 6, 12, 18, and 24 h. There were 1040 patients on the integrilin arm but 18 patients (1.73%) are missing enzyme levels. We analyze the complete-case data and thus the effective sample size is n = 1022.

Table 1 gives estimates of E[Y(d¯τ)] at five time points using variations of the newly proposed estimator and the estimator by Lu and Johnson (2015), where the latter method does not adjust for time-dependent confounding through enzyme level. Both methods use proportional hazards models to model the hazard of treatment-terminating event and allow for flexible definitions of regression functions mY. We note that point estimates for LJ in Table 1 and in Lu and Johnson (2015, Section 4.2) are not identical because the sample size is different; that is, if time-dependent enzyme level is not used, one can retain data for 18 patients with missing enzyme levels. Nevertheless, the point estimates are rather similar.

TABLE 1.

Thirty-day event proportions (%) from the ESPRIT infusion study. Proportional hazards models for λtξ and λtU functions while different statistical models for mY include linear models (LM), partially linear models (PLM), and super learner (SL) ensembles with gam, glm, MARS, and random forest. The estimator that adjusts for baseline confounders only is the direct estimator proposed by Lu and Johnson (LJ, 2015) whereas the new proposed estimator adjusts for time-varying confounders. Table entries include point estimates (Est.) and bootstrap standard errors (SE) at five time points

Time (h) Potential confounding
Time-dependent
Baseline only
SL
PLM
LM
SL
PLM
LM
Est. SE Est. SE Est. SE Est. SE Est. SE Est. SE
16 5.98 0.92 5.67 1.43 5.88 1.11 6.92 1.01 5.00 1.38 5.43 1.03
18 6.56 0.97 6.60 0.91 6.53 0.82 7.18 1.21 6.44 0.94 6.50 0.83
20 7.11 1.07 7.71 1.24 7.45 0.78 7.40 1.14 7.88 1.25 7.56 0.79
22 7.20 0.92 8.25 1.34 7.99 1.02 7.70 1.16 8.46 1.37 8.69 1.05
24 8.32 1.27 8.81 2.31 8.53 1.48 8.14 2.26 10.64 2.28 9.89 1.58

Table 1 shows the numerical results of the proposed method and LJ method when outcome parts are modeled by super learner (SL), partially linear model (PLM), and linear model (LM). All variations of the two methods return similar patterns of estimates of the expected event proportion across different treatment strategies. In general, the expected event proportion is gently increasing for infusion length strategies ranging from 16 to 24 h. Hence, there is some evidence to suggest that infusing with integrilin for shorter durations is better than infusing for longer duration (Johnson and Tsiatis, 2004). Point estimates of the proposed method are relatively comparable under different working outcome models. When adjusting for baseline confounding only, the results from PLM and LM are similar but they differ from the SL results. We interpret this as mild evidence, though not statistically significant, that a parametric outcome working model may be misspecified when time-dependent enzyme level is not included among the set of potential confounders in the method. When time-dependent enzyme level is included in the analysis, the agreement among the point estimates is very good using different outcome working models. Standard error estimates using a parametric working outcome model are uniformly smaller compared to those using a general semiparametric working model, as expected, and the same pattern is observed regardless of whether one adjusts for time-dependent confounders or not. The bootstrap-based standard error estimates for SL may be invalid but are shown for completeness.

6 |. DISCUSSION

Motivated by the ESPRIT trial, we developed a method to evaluate the effect of continuous adaptive treatment length subject to eligibility criteria in the presence of potential time-dependent covariates. The proposed approach combines a likelihood-based method and risk-set weighting method to address the incomplete covariate history problem and flexible working models can be utilized to construct the estimation. Additionally, simulation studies show that the adjustment of time-dependent covariates is necessary to understand the effect of a nonrandomized treatment on outcome in the presence of time-dependent confounders.

The consistency of the proposed estimator relies on correctly specified models and assumptions for the nuisance parameters, including but not limited to the piece-wise constant assumption on the time-dependent covariate process. One future research topic is to construct a doubly robust estimator for adaptive treatment length policies so that the consistency follows if at least one of the nuisance parameters is modeled correctly (Kennedy et al., 2017). We have proposed to use machine learning methods (e.g., superlearner) to reduce the chance of obtaining a biased estimate caused by some model misspecification. This, however, may lead to estimators that are not asymptotically linear, which can complicate deriving a variance formula for the corresponding estimator. One can overcome this problem by using a targeted minimum loss–based estimator where the nuisance parameters are targeted (van der Laan, 2014).

In our framework, covariate history is not collected beyond the observed treatment length U. This is an artifact regarding the data made available to our team. In the clinic, it is reasonable to believe that a subset of time-varying covariates Xt might be collected beyond treatment termination, especially for a patient admitted for heart surgery as in our data application in Section 5. Under a different set of circumstances, say, if we were investigating a dynamic treatment regime where treatment could be restarted, covariate information collected after the first stoppage would be important for the data analysis.

Supplementary Material

Supp

ACKNOWLEDGMENTS

We gratefully acknowledge the associate editor and two anonymous reviewers whose comments improved the presentation substantially. The authors thank Dr. James Tcheng for permission to use the ESPRIT trial data in Section 5. The project was supported in part by the University of Rochester CTSA award number UL1 TR002001 R21AA027571 and R01DA048764 from the National Institutes of Health, and EHSC award number P30 ES001247 from the Environmental Health Sciences Center. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

APPENDIX: OUTLINE OF OPERATING CHARACTERISTICS

We outline the operating characteristics of the proposed estimator and describe the theoretical results formally in Web Appendix A. We make the following assumptions to derive a new expression for the causal estimand that leads to our proposed estimator.

A1. Consistency of the observed outcome (SUTVA);

A2. Sequential randomization assumption in (2);

A3. Positivity (or overlap) of the treatment assignment and treatment-terminating events:

A4. The probability that treatment is stopped by choice and also stopped by treatment-terminating event in the same infinitesimal [t, t + dt) has zero probability;

A5. X¯t is left-continuous, piece-wise constant with jumps at the landmarks such that L¯τ has density (7).

Assumptions A1 and A2 are the consistency and identifiability assumptions used in causal inference to estimate causal estimands. Assumption A3 ensures that statistical inference of the treatment effect is possible. Assumption A4 prevents ties between competing events while Assumption A5 makes clear our landmark condition on covariate history.

Let μto=E[Y(d¯τ)] under d¯τ given by dt(Ht) = 1 − I(tto)I(Ct = 0) when G is the entire population. Assume conditions A1–A5 above, conditions C1–C3 in Web Appendix C at fixed target time to, and that λtξ(Ht) and λtU(Ht) are correctly modeled through a proportional hazards assumption. If m^Y is a parametric estimator of mY and the model mY is correct, then μ^to is consistent, asymptotically linear, and n1/2(μ^toμto) converges in distribution to a mean-zero normal random variable with variance E[φ(X¯to,to)2], where φ(X¯t,t) is the influence curve at time t and detailed in Web Appendix C. We prove a similar result where we allow mY to follow a general semiparametric model, estimate the nonparametric component using local polynomials and the parametric components via profile least squares. Those details are provided in Web Appendix D.

Additionally, we provide conditions on the regression function mY under which the proposed estimator μ^to will be consistent but the Donsker condition used in Theorem 10.2 in Kosorok (2008) for inference procedures may be violated. Assume assumptions A1–A5, λtξ(Ht) and λtU(Ht) are correctly modeled via PH, and conditions C1–C3 in Web Appendix C at fixed target time to. At fixed target time to, if mY is fit by models such that m^YmY=op(1), then |μ^toμto|=op(1) for to ∈ (0, τ].

Footnotes

SUPPORTING INFORMATION

Web Appendix A, referenced in Section 3, Web Appendix C, referenced in Appendix, Web Appendix D, referenced in Section 4, R programs implementing the proposed method are available at the Biometrics website on Wiley Online Library.

DATA AVAILABILITY STATEMENT

The data that support the findings in this paper are available upon request from Dr. James Tcheng, principal investigator of the Enhanced Suppression of the Platelet Receptor IIb/IIIa with Integrilin Therapy (ESPRIT) trial. Restrictions apply to the availability of these data, which were used in this paper with the explicit written permission of the principal investigator. Data are available from the authors (brent_johnson@urmc.rochester.edu) with the permission of Dr. James Tcheng (tchen001@mc.duke.edu).

REFERENCES

  1. Cox DR (1972) Regression models and life-tables. Journal of the Royal Statistical Society: Series B (Methodological), 34, 187–202. [Google Scholar]
  2. Díaz I and van der Laan MJ (2013) Targeted data adaptive estimation of the causal dose–response curve. Journal of Causal Inference, 1, 171–192. [Google Scholar]
  3. Gill RD and Robins JM (2001) Causal inference for complex longitudinal data: The continuous case. Annals of Statistics, 29, 1785–1811. [Google Scholar]
  4. Hu L, Hogan JW, Mwangi AW and Siika A (2018) Modeling the causal effect of treatment initiation time on survival: Application to HIV/TB co-infection. Biometrics, 74, 703–713. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Johnson BA, Ribaudo H, Gulick RM and J. EJ, (2013) Modeling clinical endpoints as a function of time of switch to second-line ART with incomplete data on switching times. Biometrics, 69, 732–740. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Johnson BA and Tsiatis AA (2004) Estimating mean response as a function of treatment duration in an observational study, where duration may be informatively censored. Biometrics, 60, 315–323. [DOI] [PubMed] [Google Scholar]
  7. Johnson BA and Tsiatis AA (2005) Semiparametric inference in observational duration-response studies, with duration possibly right-censored. Biometrika, 92, 605–618. [Google Scholar]
  8. Kennedy EH, Ma Z, McHugh MD and Small DS (2017) Nonparametric methods for doubly robust estimation of continuous treatment effects. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 79, 1229–1245. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Kosorok MR (2008) Introduction to Empirical Processes and Semiparametric Inference. New York: Springer. [Google Scholar]
  10. Lok JJ (2008) Statistical modeling of causal effects in continuous time. The Annals of Statistics, 36, 1464–1507. [Google Scholar]
  11. Lu X and Johnson BA (2015) Direct estimation of the mean outcome on treatment when treatment assignment and discontinuation compete. Biometrika, 102, 797–807. [Google Scholar]
  12. Lu X and Johnson BA (2017) Direct estimation for adaptive treatment length policies: methods and application to evaluating the effect of delayed peg insertion. Biometrics, 73, 981–989. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Murphy SA, van der Laan MJ, Robins JM and Group CPPR (2001) Marginal mean models for dynamic regimes. Journal of the American Statistical Association, 96, 1410–1423. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. O’Shea JC, Hafley GE, Greenberg S, Hasselblad V, Lorenz TJ, Kitt MM, et al. (2001) Platelet glycoprotein IIb/IIIa integrin blockade with eptifibatide in coronary stent intervention: the esprit trial: a randomized controlled trial. JAMA, 285, 2468–2473. [DOI] [PubMed] [Google Scholar]
  15. Petersen M, Schwab J, Gruber S, Blaser N, Schomaker M and Van der Laan M (2014) Targeted maximum likelihood estimation for dynamic and static longitudinal marginal structural working models. Journal of Causal Inference, 2, 147–185. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Robins J (1986) A new approach to causal inference in mortality studies with a sustained exposure period—application to control of the healthy worker survivor effect. Mathematical Modelling, 7, 1393–1512. [Google Scholar]
  17. Robins JM Causal inference from complex longitudinal data. In: Berkane M, ed. Latent Variable Modeling and Applications to Causality: Lecture Notes in Statistics 120. New York: Springer-Verlag, 1997;69–1. [Google Scholar]
  18. Robins J, Orellana L and Rotnitzky A (2008) Estimation and extrapolation of optimal treatment and testing strategies. Statistics in Medicine, 27, 4678–4721. [DOI] [PubMed] [Google Scholar]
  19. Robins JM, Hernán MÁ and Brumback B (2000) Marginal structural models and causal inference in epidemiology. Epidemiology, 11, 550–560. [DOI] [PubMed] [Google Scholar]
  20. Rubin DB (1978) Bayesian inference for causal effects: the role of randomization. The Annals of Statistics, 6(1), 34–58. [Google Scholar]
  21. van der Laan MJ (2014) Targeted estimation of nuisance parameters to obtain valid statistical inference. The International Journal of Biostatistics, 10, 29–57. [DOI] [PubMed] [Google Scholar]
  22. Yang S, Tsiatis AA and Blazing M (2018) Modeling survival distribution as a function of time to treatment discontinuation: a dynamic treatment regime approach. Biometrics, 74, 900–909. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Zhang M, Joffe MM and Small DS (2011) Causal inference for continuous-time processes when covariates are observed only at discrete times. Annals of Statistics, 39, 131–173. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Zhang M, Tsiatis AA, Davidian M, Pieper KS and Mahaffey KW (2011) Inference on treatment effects from a randomized clinical trial in the presence of premature treatment discontinuation: the synergy trial. Biostatistics, 12, 258–269. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supp

Data Availability Statement

The data that support the findings in this paper are available upon request from Dr. James Tcheng, principal investigator of the Enhanced Suppression of the Platelet Receptor IIb/IIIa with Integrilin Therapy (ESPRIT) trial. Restrictions apply to the availability of these data, which were used in this paper with the explicit written permission of the principal investigator. Data are available from the authors (brent_johnson@urmc.rochester.edu) with the permission of Dr. James Tcheng (tchen001@mc.duke.edu).

RESOURCES