Abstract
Precision medicine seeks to maximize the quality of healthcare by individualizing the healthcare process to the uniquely evolving health status of each patient. This endeavor spans a broad range of scientific areas including drug discovery, genetics/genomics, health communication, and causal inference all in support of evidence-based, i.e., data-driven, decision making. Precision medicine is formalized as a treatment regime which comprises a sequence of decision rules, one per decision point, which map up-to-date patient information to a recommended action. The potential actions could be the selection of which drug to use, the selection of dose, timing of administration, specific diet or exercise recommendation, or other aspects of treatment or care. Statistics research in precision medicine is broadly focused on methodological development for estimation of and inference for treatment regimes which maximize some cumulative clinical outcome. In this review, we provide an overview of this vibrant area of research and present important and emerging challenges.
Keywords: data-driven decision science, dynamic treatment regimes, machine learning, patient heterogeneity, statistical inference
1. INTRODUCTION
The idea of improving health outcomes by tailoring treatment to individual patient characteristics is centuries old and remains a core component of medical practice. The scientific method began to impact medical treatment through the use of statistical inference by the late 1700s, but advances began to dramatically increase after the success of the first randomized controlled clinical trial, conducted by Austin Bradford Hill in 1946, which demonstrated the efficacy of streptomycin for treating tuberculosis (Stusser 2006). Following Hill’s trial was a period of rapid methodological progress in the design and analysis of clinical trials as well as observational studies. Systematic study of the integration of data, experience, and clinical judgment into the clinical decision process led to the concept of ‘evidence-based medicine’ wherein clinical decision making is based on (to the extent possible) empirical evidence with randomized controlled trials being a gold-standard for generating such evidence (Eddy 1990). However, the primary scientific aim in most clinical trials is the identification of the best treatment for a given disease area with any heterogeneity in patient characteristics or outcomes being viewed as a nuisance to the research process.
Awareness that patient heterogeneity was important in evaluating treatments–and not just a nuisance–began to emerge late in the 20th century, among both clinicians (Sorensen 1996) and biostatisticians (Longford and Nelder 1999). That patient heterogeneity implied the need to individualize therapy in the context of evidence-based medicine was nicely articulated in Kravitz et al (2004). These constituent concepts, combined together, yield the modern concept of precision medicine, the paradigm wherein patient heterogeneity is leveraged through data-driven approaches to improve treatment decisions so that the right treatment is given to the right patient at the right time. Precision medicine became a national priority with President Obama’s announcement of the Precision Medicine Initiative in his 2015 State of the Union Address (The White House 2015). We note that precision medicine is conceptually the same as stratified medicine (Lonergan et al 2017) and personalized medicine (Kosorok and Moodie 2016).
The goal of this paper is to provide a review of the current state-of-the-art in statistical research for precision medicine. The chief priority of statistical research in precision medicine is to use data to inform decision making in healthcare; thus, this encompasses a wide range of tasks including drug-discovery, biomarker identification, estimation and inference for causal treatment effects, modeling health communication and shared decision making, and study design. However, our focus in this review is on estimation and inference for treatment regimes which prescribe interventions based on individual patient characteristics. An estimated optimal treatment regime might be used as part of a decision support system within a healthcare organization or to generate new clinical hypotheses for future study. Thus, it is critical that statistical methodology for precision medicine be rigorous, transparent, reproducible, and generalizable.
We view precision medicine as fitting within the broader concepts of precision public health and data-driven decision science. However, the focus on data-driven patient-centered care with its inherent challenges, e.g., patient heterogeneity, implementation cost, causal confounding, etc., distinguishes precision medicine as its own field of study. To this point, precision medicine has led to new methodologies and insights in semi-parametric modeling, causal inference, non-regular asymptotics, clinical trial design, and machine learning (see Murphy 2003; Robins 2004; Chakraborty and Moodie 2013; Laber et al 2014; Zhao et al 2015a; Kosorok and Moodie 2016, and references therein). There have also been major advancements in genetics driven by the vision for precision medicine (see, e.g., Torkamani et al 2018); however, our focus will be broader in that while the biomarkers used to inform treatment selection could be genetic or genomic factors, we also allow that these could be demographic and physiological measurements, co-morbid conditions, individual patient preferences, lifestyle, and so on.
The remainder of the paper is as follows. In Section 2, we formalize the goals of precision medicine and catalog different decision settings. In Section 3, we discuss biomarker discovery as an important supporting task in constructing an optimal treatment regime. In Section 4, we review the methodological underpinnings of precision medicine including regression-based and direct-search estimation, managing multiple outcomes, and inference for estimated optimal regimes. We conclude with a summary and discussion of pressing open problems.
2. GOALS OF PRECISION MEDICINE
In this section, we formalize an optimal dynamic treatment regime and contrast estimation of an optimal regime with subgroup identification and causal effect estimation. We also discuss generalizations of the dynamic treatment regime framework and how decision problems in precision medicine fit into precision public health and data-driven decision science. Note that we view decision support—especially the estimation of optimal or near-optimal treatment regimes and those endeavors which directly support this—to be the primary goal of precision medicine. Accordingly, we view modeling the disease process per se to be of secondary importance in precision medicine, except when it is directly supportive of dynamic treatment regime discovery.
2.1. Discovering dynamic treatment regimes
As stated previously, the goal in precision medicine is to use data to improve decision making in healthcare. Dynamic treatment regimes formalize decision making as a sequence of decision rules, one per decision point, that map available information to a recommended intervention. The decision points may either be fixed in calendar time or driven by patient outcomes. Thus, the timing and number of decision points may be random and can vary considerably across patients in some application domains. Furthermore, the set of allowable interventions at any given time point may vary according to a patient’s health status, availability, or other factors (Bembom and van der Laan 2008; Schulte et al 2014; Laber and Staicu In press; Laber et al 2018); however, for simplicity, in describing methods for estimation and inference in Section 4, we will not do so in complete generality. We formalize the notion of an optimal treatment regime using the language of potential outcomes (Rubin 1978, 2005; Dawid 2015).
2.1.1. The single-decision setting.
In the single decision setting the observed data are assumed to be of the form which comprise n i.i.d. triples (X, A, Y) where: denotes baseline patient characteristics; denotes the assigned treatment; and denotes the outcome coded so that higher values are better. For each define to be the set of allowable treatments for a patient presenting with X = x. A dynamic treatment regime in this context is a map which satisfies d(x) ∈ ψ(x) for all ; under d, patients presenting with X = x would be assigned treatment d(x). An optimal treatment regime yields the maximal mean outcome if applied to select treatments in the population of interest. Let Y* (a) denote the potential outcome under treatment and subsequently for any regime, d, define the potential outcome under d to be . The optimal regime, say dopt, satisfies: (i) dopt(x) ∈ ψ(x) for all ; and (ii) EY* (dopt) ≥ EY* (d) for all d such that d(x) ∈ ψ(x) and all . There are other notions of optimality including maximizing quantiles of the outcome distribution (Linn et al 2017; Wang et al In press); maximizing efficacy subject to constraints on risk/harm (Linn et al 2015; Laber et al 2018; Wang et al In press); and maximizing the mean outcome subject to cost or logistical constraints (Luedtke and van der Laan 2016a; Lakkaraju and Rudin 2017).
The optimal dynamic treatment regime, dopt, is defined in terms of potential outcomes; however, in order to construct an estimator of dopt we need to identify it in terms of the data-generating model. In Section 4, we present causal conditions under which such identifiability holds. We note that the term ‘dynamic’ refers to the individualization of treatment to patient characteristics, not time; hence, the term ‘dynamic treatment regime’ is used even with a single decision point. One could also assume that the clinicians assigning treatment are already making nearly optimal choices, and thus a potentially good dynamic treatment is one obtained by mimicking the clinicians’ decisions Wallace et al (2018). We will consider this in Section 5 for the special case where physicians make the optimal decision at least for a small portion of the time.
2.1.2. The multi-decision setting.
In the multi-decision setting there are two or more opportunities for treatment change for some subset of the population. Examples include the treatment of small-cell lung cancer where two or more lines of chemotherapy may be required (Zhao et al 2011). A key concept in the multi-decision setting is that interventions can affect a patient’s health status in multiple ways including: (immediate effects) it may make an immediate impact on their health status; (moderating effects) it may generate information that is useful for subsequent decisions, e.g., failure to respond to a drug in a given class may indicate that other drugs belonging to the same class may also perform poorly; and (delayed effects) it may change the patient’s health status so as to set them up for future success, e.g., providing a patient with cognitive behavioral therapy at one stage may allow them to reap greater benefits from tele-therapy at later time points. Thus, applying a treatment that leads to a suboptimal proximal effect may lead to better longterm outcomes if there are strong delayed or prognostic effects (Thall et al 2007; Kidwell 2016).
In the multi-decision setting, we assume that the observed data are of the form which comprise n i.i.d. replicates of (X1, A1, Y1, … , XT, AT, YT) where: denotes baseline information and denotes interim information collected during the course of the stage t − 1 treatment for t = 2, …, T; denotes the assigned treatment; and Yt denotes a proximal outcome measured after the treatment at stage t for t = 1, … , T . Define H1 = X1 and Ht = (Ht−1, At−1, Yt−1, Xt) so that Ht is the available patient history at time t. Let and let denote the set of allowable treatments for a patient presenting with Ht = ht at time t. A dynamic treatment regime in this setting is a sequence of functions d = (d1, … , dT) such that satisfies dt(ht) ∈ ψt(ht) for all ht for t = 1, … , T. An optimal treatment regime maximizes the expectation of some (prespecified) cumulative outcome measure Y = y(Y1, … , YT ), e.g., or y(v1, … , vT ) = maxt vt, or y(v1, … , vT ) = vT, etc.
Interventions applied at time t affect both the proximal outcomes Yt and interim measurements Xt+1; thus, to define the optimal treatment regime we need to consider both potential proximal outcomes and potential interim measurements. We use an overline to denote history so that . Define to be the potential interim measurements at time t under treatment sequence , where products of sets are interpreted as cartesian products, and define to be the potential proximal outcome under . The potential outcome under is therefore . For any regime, d, the potential outcome is
where we have defined and 1B is the indicator of B. An optimal regime, dopt, satisfies: (i) for all and t = 1,…,T; and (ii) EY*(dopt) ≥ EY* (d) for all d satisfying dt(ht) ∈ ψt(ht) for all and t = 1, … , T.
As in the single-decision setting, the optimal regime is defined in terms of potential outcomes and is only identifiable in terms of the data-generating model under additional assumptions; we discuss such assumptions in Section 4. The preceding development assumes that future patients will be treated over the same time horizon as the patients in the sample; however, this need not be the case. In some settings, e.g., diabetes (Ertefaie 2015; Luckett et al 2017) or cystic fibrosis (Tang and Kosorok 2012), interventions are applied over an indefinite time horizon. This structure would also be applicable to precision screening for cancer (Olsen and Lund 2017), among other applications. Thus, the objective in these settings is often to estimate a treatment regime that can potentially be applied beyond the time horizon over which the training data are collected. Such extrapolation typically requires additional structure to be imposed on the data generating model; a common assumption is that the decision process is (perhaps after some suitable transformation) heterogeneous and Markov (Puterman 2005). We discuss this further in Section 4.
2.1.3. Other decision settings.
The decision settings described above assume: (S1) the observed data consist of i.i.d. replicates; (S2) the data-generating model is fixed and indifferent to the actions of the decision maker; and (S3) the observed data are used to construct treatment regimes for application with yet unobserved future patients. However, these assumptions may need to be relaxed in some application domains. For example, in the context of managing the spread of an emerging infectious disease over a network of individuals, spatial dependence and spillover effects preclude treating the individuals as independent replicates; furthermore, in this setting, one must manage the disease in real-time (Laber et al In press). Thus, neither (S1) nor (S3) hold in this setting. In the context of adversarial decision making wherein one faces an intelligent (and adaptive) opponent, the data-generating model can change in response to the decision makers actions, e.g., imagine playing poker repeatedly against the same shrewd player (Cesa-Bianchi and Lugosi 2006).
There are potentially other decision settings not covered by the above structures, however, the methods and ideas presented in this review are readily extensible to new domains. There are also additional complications associated with estimation of optimal treatment regimes that occur more generally in statistical modeling and thus we will not discuss them here; e.g., missing data (Shortreed et al 2011; Shortreed and Moodie 2012; Shortreed et al 2014; Kosorok and Moodie 2016); measurement error (Shani et al 2013), and model-building (Biernot and Moodie 2010; Rich et al 2010; Henderson et al 2010; Gunter et al 2011; Lu et al 2013; Laber et al 2014; Song et al 2015; Luedtke and van der Laan 2016b).
2.2. Treatment effect estimation and subgroup identification
Estimation of an optimal treatment regime is closely related to subgroup identification and estimation of the conditional average treatment effect (CATE). For simplicity, we use a single-decision problem with binary treatments, so that the data for a generic subject are (X, A, Y ), where denotes baseline patient information; denotes the assigned treatment; and denotes the outcome coded so that higher values are better. We assume that for all . In this setup, the CATE is defined as Δ(x) = E {Y*(1) – Y*(−1)|X = x}. For any regime d it can be seen that
whence it can be seen that dopt(x) = sign {Δ(x)}is optimal, where sign(x) = ±1 according to whether x > 0 or < 0. Thus, a natural approach to estimating an optimal treatment regime in this context is to first construct an estimator of Δ(x) and subsequently to use the plug-in estimator of dopt(x); we discuss this approach in more detail in Section 4. We note this can be developed for more than two treatments, but we focus here on two treatment for simplicity.
Subgroup identification seeks to find a subgroup in the target population with an en hanced treatment effect; sometimes this is referred to as finding ‘the right patient for the right treatment.’ Given a threshold, , one way to operationalize a subgroup is through the level set . Thus, one could estimate this level set using the plug-in estimator , where is an estimator of Δ(x). We note that while much of clinical research focuses on estimating the CATE, precision medicine seeks to go further by tailoring treatments to subgroups (as reflected in x) to benefit each subgroup and thereby further benefit the population on average.
To illustrate the differences between estimation of an optimal treatment regime, estimation of the CATE, and subgroup identification consider the following generative model: X ~ Normal(0, τ2), A ⊥ X and A ~ Uniform (−1, 1}, ∈ ⊥ (X, A), ∈ ~ Normal(0, σ2), and Y = g(X, A) + ∈, where and ⊥ denotes independence. The optimal treatment can be seen to equal , the CATE is given by , the subgroup corresponding to the level-set with threshold η is given by which, provided it is nonempty, is an interval [ℓ, u] with −∞ ≤ ℓ ≤ u ≤ ∞. The optimal regime in this toy example is linear. Suppose that one attempted to estimate an optimal linear decision rule by postulating a linear model of the form g(x, a; β) = β0 + β1a + β2x + β3ax so that Δ(x; β) = 2β1 + 2 β3x which is to be estimated using least squares. It can be shown that the projection of g(x, a) onto g(x, a; β) is given by g(x, a; β*), where
Thus, it can be seen that β*can be far from α* and subsequently the linear rule obtained by estimating g(x, a) with a linear model need not lead to a high-quality linear decision rule, e.g., if τ = 1 and α*= (4.176, 1.720, −4.704, 0.320)T then the optimal rule and the rule disagree on more than 60% of the population.
2.3. Dynamic data-driven decision science
Estimation of an optimal treatment regime is an example of a reinforcement learning problem in that one must learn about optimal decision making using data on the interactions between one or more decision makers and the environment. There is an expansive literature on reinforcement learning in computer science and engineering (e.g., Sutton and Barto 1998; Si 2004; Powell 2007; Szepesvari 2010; Busoniu et al 2010). This literature was developed with a focus on algorithmic efficiency, computational scalability, and empirical performance; state-of-the-art reinforcement learning algorithms are expected to identify complex and subtle patterns from massive data sets. In contrast, in the precision medicine literature, methodology was developed with a focus on casual validity, generalizability, and interpretability within a domain context; state-of-the-art methodologies for precision medicine are expected to be transparent, rigorous, and to generate new scientific knowledge from (relatively) small data sets. However, cross-pollination is rapidly increasing due to technological advancements facilitating the collection and curation of massive amounts of patient-level data in real (or near-real) time and the emergence of mobile-health (Tewari and Murphy 2017; Nahum-Shani et al 2017; Luckett et al 2017).
Precision medicine is also closely connected with control theory and operations research. Within these areas there is a rich history of modeling the underlying system dynamics (i.e., the generative model) and using simulation-optimization to inform decision making; standard texts include Hillier (1990) and Macia and Thaler (2005). Introducing stochasticity into dynamic systems to address various forms of uncertainty has led to many rich developments in stochastic differential equations (Nisio 2015) and in Markov Decision Processes (Puterman 2005). Simulation-based approaches which use interactions between complex agents, such as agent-based modeling, have also been developed which allow for studying situations of greater complexity than normally achievable through systems of mathematical equations (Wilensky and Rand 2015). These approaches can be effective when there is rich scientific theory to inform the construction of the underlying models; however, such information is rarely available in the context of precision medicine making these methods diffcult to apply directly.
3. BIOMARKERS
In precision medicine research a common clinical goal is the identification of patient biomarkers that are important for choosing an optimal treatment. We use the term biomarker generically to represent a scalar feature constructed from current patient information; thus, a biomarker could be a single component of the available history or a composite measurement constructed from multiple components. A biomarker may provide valuable clinical information by being: (i) prognostic, i.e., the biomarker is useful in predicting the mean outcome of a patient; (ii) moderating, i.e., the biomarker is useful for predicting contrasts of the mean outcome across different candidate treatments; and (iii) prescriptive, i.e., the biomarker is useful in selecting the treatment that maximizes the mean outcome (see Teran Hidalgo et al 2016, for additional refinements on biomarker classification). We focus here on the mean for simplicity, but other distributional summaries, such as the median, could also be used. Figure 1 shows conceptual schematics for each of the three biomarker types. Below we formalize these notions and show that they are nested so that prescriptive biomarkers are moderating and prognostic while moderating biomarkers are prognostic but need not be prescriptive. We focus on the single-decision setting; analogs for the multi-decision setting can be derived based on the approximate dynamic programming methods described in Section 4.
3.1. Prognostic biomarkers
Consider the single-decision binary treatment setting with Then it follows that E {Y*(a)|X = x} = μ(x) + aΔ(x)/2, where μ(x) = E {Y*(1) + Y* (−1)|X = x}. We note that these ideas can be generalized to more than two treatments, but we restrict ourselves here to the two treatment situation for ease of exposition. To illustrate key concepts, we first assume models of the form and and biomarkers under consideration are the components of X = (X1, … , Xp)T; a more general definition is given below. Under this model, the biomarker Xj is a prognostic biomarker if either or is not zero. Under the causal conditions presented in Section 4, one can estimate the vectors and using ordinary least squares and whether a biomarker is prognostic using standard methods. The regression features found in many predictive models in medical research, such as, for example, a Cox regression model used to predict survival of non-Hodgkin’s lymphoma patients (Non-Hodgkin’s Lymphoma Prognostic Factors Project 1992), are examples of prognostic biomarkers.
Now consider the setting where one has a set of candidate biomarkers where is a possible composite summary of X.For any , we say that is sufficient for prognosis if , where σ(U) denotes the σ-algebra generated by U. We define a set of biomarkers, , to be minimal sufficient for prognosis if it is sufficient for prognosis and for any other sufficient set , where denotes the number of elements in .We define a biomarker Bj to be prognostic if for some minimal sufficient set . As in the linear setting, a general approach to identifying predictive biomarkers is to model μ(x) and Δ(x) as functions of the candidate markers and to apply standard methods for variable selection. One could also use methods for feature construction in regression to identify the set of candidate features (Cook and Ni 2005; Li 2007; Lee and Verleysen 2007).
3.2. Moderating biomarkers
Moderating biomarkers are predictive of the contrast between two treatments. In the linear model example, where and we say that Xj is a moderating biomarker if is not zero. Thus, it can be seen immediately that a moderating biomarker is also prognostic. As with identifying prognostic biomarkers, under appropriate causal conditions, one can use ordinary least squares to estimate and test whether is zero to identify moderating biomarkers. Under the postulated linear model with unbounded biomarkers, every moderating biomarker is also prescriptive; however, this does not hold in general. In the more general setting with candidate biomarkers we say that subset of biomarkers is sufficient for moderation if and furthermore minimal sufficient if for any other sufficient set .
3.3. Prescriptive biomarkers
To illustrate the difference between moderating and prescriptive biomarkers, consider the model and suppose that the support of X is . Then we say that Xj is a prescriptive biomarker if is not zero. It can be seen that if is zero but is nonzero then Xj is moderating but not prescriptive. More generally, under the setting with candidate biomarkers we say that is sufficient for prescription if and we say that is minimally sufficient for prescription it if further satisfies for all sufficient .The identification of prescriptive biomarkers has been studied as a subtopic within precision medicine with early approaches seeking to identify ‘qualitative interactions’ (Gail and Simon 1985; Gunter et al 2011) and more recent approaches focused on identification of variables that are informative for identification of an optimal treatment regime (Song et al 2015; Zhang and Zhang 2016; Fan et al 2016).
A concrete example of both moderating and prescriptive biomarkers can be found in Gail and Simon (1985) who analyze data from the National Surgical Adjuvant Breast and Bowel Project (Fisher et al 1983). They find that a patient’s age and progesterone receptor level (PR) are informative of the effect of adding tamoxifen to chemotherapy. Thus both age and PR are moderating biomarkers. Now define the composite biomarker D to be 1 if both age< 50 and PR< 10 and to be −1 otherwise. Gail and Simon found that if D = 1, tamoxifen should be added, but it should not be added otherwise. In this setting D is both a moderating and a prescriptive biomarker. We note that the FDA identifies a biomarker as predictive if it is predictive of the contrast between active treatment and a control; thus, a predictive variable is prescriptive for a contrast involving a control (FDA-NIH Biomarker Working Group 2016).
4. ESTIMATING DYNAMIC TREATMENT REGIMES
4.1. The single-decision setting
In the single-decision setting, as described in section 2.1.1., the goal is to estimate a regime, dopt, that satisfies EY* (dopt) ≥ EY* (d) for any other regime d. In order to construct an estimator, we need to express dopt in terms of the data-generating model. To do this, we make the following assumptions: (i) positivity, P (A =a |X = x) > 0 for all a ∈ ψ(x) for all ; (ii) consistency, Y = Y* (A); and (iii) strong ignorability, . These assumptions are standard (Kidwell 2016) though they are not as general as possible (Robins et al 2000; Petersen et al 2012).
Define Q(x, a) = E(Y|X = x, A = a), then under the preceding assumptions, dopt(x) = arg max a∈ψ(x) Q(x, a) is an optimal regime. This immediately suggests a regression-based estimator wherein one first constructs an estimator of Q(x, a) and subsequently uses the plug-in estimator . For example, one might posit a linear model of the form , where and xa, are features of x. Define to be the empirical measure. Let and subsequently so that the plug-in estimator is . Linear models are commonly used because they are regarded as being easy to interpret, e.g., the coefficients of can be used to identify what biomarkers are likely to impact a patient’s outcome under treatment a. However, while linearity lends itself to interpretation, this may come at the cost of misspecification; thus, as with any regression model, one should apply model diagnostics and interactive model-building techniques to ensure a high-quality model.
To mitigate misspecification, one can use a flexible class of models to estimate Q(x, a), e.g., trees (Taylor et al 2015; Zhang et al 2012b), boosting (Kang et al 2014), generalized additive models (Moodie et al 2014), or non-linear basis expansions (Qian and Murphy 2011). However, using a flexible model for Q(x, a) can render the estimated rule unintelligible—thereby obscuring the scientific content of the estimated regime and limiting its value as a decision support tool. This issue led to the development of regression-based policy-search methods wherein the class of regimes is decoupled from the class of estimators used for Q(x, a). For any regime d it follows under the preceding causal conditions that EY* (d) = EQ {X, d(X)}; thus, if is a pre-specified class of regimes then the optimal within this class is Let be an estimator of Q(x, a) then the plug-in estimator of Because the class is chosen independently of the class of models for Q(x, a), one can use nonparametric regression estimators while maintaining control of the form of the estimated optimal regime. Furthermore, because this approach is built upon regression, it is easily extensible to settings with complex data structures, censored data, measurement error or other settings for which regression models have been developed.
Regression-based estimators were derived from a regression based representation of the optimal treatment regime, an alternative representation based on importance sampling leads to another class of estimators termed direct-search or classification-based estimators. For simplicity, we assume that treatment is binary and coded so that. Under the causal conditions stated above, the marginal mean outcome under a regime d is
where π(a; x) = P (A = a|X = x) is the propensity score (Qian and Murphy 2011). Thus, the optimal regime satisfies dopt = arg maxd V (d). Given an estimator, which might be obtained using logistic regression, the inverse probability weighted estimator of V (d) is given by Given a class of regimes, , one could construct an estimator of by direct maximization, i.e.,; however, because of the discontinuous indicator function, direct optimization is not computationally feasible save for settings where is small. However, it can be see that
where sign(u) = 1u>0 − 1u<0 is the sign function and (u)+ = max(0, u) is the positive part function. Thus, given a class of regimes, , the optimal regime within this class satisfies
hence, it can be seen that minimizes a cost-sensitive classification problem (Elkan 2001; Zadrozny et al 2003; Zhou and Liu 2006; Pires et al 2013) with cost function |Y|/π(A; X), class label A sign(Y), and input, over the set of classifiers (see Zhao et al 2012; Zhang et al 2012a,b). Thus, one can estimate by applying off-the-shelf classification algorithms using the estimated cost in place of |Y|/π(A; X); for a discussion of converting cost-sensitive classification problems into standard (i.e., constant cost) classification problems see (Zadrozny 2003 and references therein). Outcome Weighted Learning (OWL, Zhao et al 2012) uses this framework with support vector machines (see, e.g., Chapter 12 of Hastie et al 2009) to estimate an optimal treatment regime; convergence rates for OWL were among the first to establish the mathematical underpinnings of direct-search estimation for optimal treatment regimes and led to a series of refinements including residual outcome weighted learning (Laber and Zhao 2015; Wang et al 2016; Zhou et al 2017) and improved bounds based on efficiency theory (Athey and Wager 2017). More generally, policy-search methods based on maximizing an estimator over a prespecified class of regimes have been extended to a wide range of settings including: ordinal treatments (Chen et al In press); right-censored outcomes (Zhao et al 2015b; Cui et al 2017) (see also (Bai et al 2017)); continuous treatments (Laber and Zhao 2015; Chen et al 2016b; Kallus 2018); and high-dimensional data (Song et al 2015; Jeng et al 2018).
4.2. The multi-decision setting
The multi-decision setting is complicated by the need to account for delayed treatment effects and prognostic effects, e.g., information gain that improves decision making at subsequent decision points. We consider two multi-decision settings: (i) a finite time horizon wherein the number of decision points is small and finite; and (ii) an indefinite time horizon wherein the number of decision points is large or indeterminate. There are, of course, many intermediate settings but these two encompass many commonly encountered settings in precision medicine. The methods we discuss apply to both observational and randomized studies.
4.2.1. Finite time horizon.
In the finite horizon setting, we can estimate the optimal dynamic treatment regimes through a variety of reinforcement learning techniques, including G-estimation (Robins 2004; Stephens 2016), Q- and A- learning (for an overview of these methods, see Schulte et al 2014, wherein G-learning is articulated as a special case of A-learning), modeling of the entire longitudinal process (Xu et al 2016), several extensions of outcome weighted learning (Zhao et al 2015a), among other approaches. Recall that in precision medicine, we are primarily interested in estimating the decision rule, and, in many settings, it is much more robust and feasible to not model the entire process if possible. Many of the learning methods listed above were motivated, at least in part, to obtain the dynamic treatment regime without needing to model the full process. Early seminal work in estimating dynamic treatment regimes includes (Robins 1986), (Robins 1997), (Murphy et al 2001), (van der Laan et al 2001), (Murphy 2003), and (Robins 2004). Because of its flexibility and relative ease in implementation, we will present Q-learning in some detail, followed by a brief discussion of OWL and several related methods, but we will not further discuss other approaches here.
As in the single-decision setting, we will derive regression-based and inverse-weighting or classification-based representations of the optimal treatment regime in terms of the data-generating model and subsequently use these representations to construct estimators of the optimal regime. Using the notation of Section 2.1.1, we make the following assumptions: (i) positivity, P (At = at|Ht = ht) > 0 for all; (ii) consistency, for t = 2,…,T and and; and (iii) sequential ignorabilityt =1,…,T, where ⊗ denote cartesian product taken over the specified range of indices; see Schulte et al (2014) for additional discussion of these assumptions.
Define QT(hT, aT) = E (Y|HT = hT, AT = aT) and for t = T − 1, … , .1 define Qt(ht, at) = E {maxat+1 Qt+1(Ht+1, at+1)|Ht = ht, At = at} then it follows from dynamic programming (Bellman 1957) that
1. |
which we term the regression-based representation of the optimal regime. Q-learning is an approximate dynamic programming algorithm based on (1) which proceeds as follows. Construct an estimator obtained by regressing Y on HT and AT, subsequently for t = T − 1, … , 1 let be an estimator of Qt(ht, at) obtained by regressing on Ht and At. The Q-learning estimator of dopt is thus for t=1,…,T (Murphy 2005b; Schulte et al 2014).
Because Q-learning relies on a series of regression models, it is easily extensible to a variety of models and data structures (Zhao et al 2009; Goldberg and Kosorok 2012; Moodie et al 2014) and allows the user to interactively construct and critique the models used for the Q-functions (Rich et al 2010; Laber et al 2014). However, as in the single-decision setting, in the above formulation the estimated optimal decision rule is tied to the class of models used for the Q-functions; thus, one may be forced with an unpleasant trade-off between severe model misspecification and an unintelligible black box. Instead one can use Q-learning with policy-search wherein the class of regimes is divorced from the class of models for the Q-functions (see Zhang et al In press; Laber et al 2018). An alternative representation of the optimal decision rule is based on inverse probability weighting. For any regime, d, it follows that
Thus, given estimators for t = 1,…,T, the plug-in estimator of V(d) is
and given a class of regimes , an estimator of is. Directly computing is diffcult except in small problems as the indicators make this into a discontinuous optimization problem. This computational issue is ameliorated in Zhao et al (2015a) through the use of a surrogate optimization function for which is smooth and has a global optimum.
However, computational issues aside, there is another difficulty with optimizing or one of its surrogates when the number of time points T is large as: (i) the product of indicators can rapidly become zero for the majority of subjects, e.g., with binary treatments assigned uniformly at random at each stage, the product of indicators will be zero for all but n(1/2)T of the original n subjects on average; and (ii) the product of the propensity scores can grow quite small leading to high-variance. For these reasons, direct search estimators based on work best for settings where T is small, e.g., T = 2; however, in such settings OWL and related methods can offer significant gains in terms of robustness and marginal mean outcome (see Zhang et al 2013; Zhao et al 2015a). As in the single-decision setting, OWL and other direct search estimators are based on converting or a more efficient augmented variant of this estimator into a cost sensitive classification problem. We note that products of indicators across time points share a similar structure to hierarchical classification problems (Gorden 1987; Wang et al 2009) though this connection has yet to be fully explored.
4.2.2. Infinite time horizon.
The infinite horizon setting applies when a sequence of similar decisions need to be made over an extended time as happens for example when treating diabetes or other chronic diseases as mentioned previously. In this setting, decision making is typically modeled as a Markov Decision Process (MDP, Puterman 2005) which encompasses a tremendously broad class of decision problems (Sutton 1997). We assume that the observed data are of the form which comprise n i.i.d. replicates of (S1, A1, S2, … , ST−1, AT−1, ST), where: St ∈ S denotes a summary of the patient’s health status at time t; and At ∈ A denotes the treatment applied at time t = 1, … , T; and T is the observed time horizon. We assume that there exists a momentary reward function so that y(s, a, s′) captures the momentary goodness for a patient with health status s who receives treatment a and subsequently transitions to a health status s′: we write Yt = y(St, At, St+1) to denote the observed momentary outcome at time t. While the observed data are collected over a horizon T the goal may be to estimate a treatment regime that can be applied indefinitely to a new patient, i.e., well beyond T decision points. To this end we assume that the observed data are Markov and Homogeneous so that for any and t ≥ 1
where the measure μst,at does not depend on time. In application, the raw data may not satisfy these conditions and one must judiciously construct the state St as a summary of the raw data to ensure that these conditions hold (at least approximately, see Wang et al 2018, and references therein).
For each let denote the set of allowable treatments for a patient with status St = st. A treatment regime in this context is a map that satisfies d(s) ∈ ψ(s) for all so that under d a patient presenting with state St = st at time t ≥ 1 would be recommended to receive treatment d(st); because d is stationary it can be applied for all t, even if t > T .1 Let 1 denote the potential patient status at time t under treatment sequence so that the potential status under a regime d is
The potential momentary outcome for regime d is . Define the conditional discounted marginal mean outcome under d to be
where γ ∈ [0, 1) is a discount factor that balances the trade-off between immediate and longterm outcomes. An optimal regime, dopt, satisfies V (s; dopt) ≥ V (s; d) for all and all regimes d. Given distribution R on we define the marginal mean outcome with respect to reference R to be VR(d) =∫ V (s; d)dR(s). One can think of the reference distribution R as the observed sample distribution or the distribution of a potential future population of patients, whichever is of primary interest. In the context of policy-search methods over a pre-specified class of regimes, , we define the optimal regime within with respect to reference R as.
Under causal assumptions analogous to those used in the finite horizon case the following recursion holds:
from which it can be seen that for any function that (see Precup 2000; Paduraru 2013; Luckett et al 2017)
The forgoing expression forms the basis for an estimating equation for V (s, d). Let V (s, d; λ) be a postulated class of models for V (s, d) indexed by; we assume that V (s, d; λ) is differentiable in λ for each s and d. Define
and let be a solution to Λn(d, λ) = 0. The estimated conditional marginal mean outcome is. Furthermore, given a reference distribution, R, a policy-search estimator of is. Note that this is an infinite horizon variant of OWL (see Luckett et al 2017, for additional description and an online version that uses stochastic regimes). Q-learning in the infinite horizon setting can be derived using analogous arguments (see Ertefaie 2015) and has been applied to manage infection from Pseudomonas aeruginosa in patients with cystic fibrosis (Tang and Kosorok 2012) and to control diabetes (see Ertefaie 2015).
We note that inference for V-learning has been derived based on assuming that the underlying dynamic process is constant or, more precisely, stationary, at least over moderately long stretches of time (see Luckett et al 2017). However, in some practical settings, such as in diabetes, it may be more realistic to expect the dynamics to gradually change as patients age. In this more complicated setting, inference is more challenging and remains on open research question.
4.2.3. Mobile health.
One important motivation for infinite horizon reinforcement learning is mobile health (mHealth) wherein it is feasible to both collect information and provide interventions to patients in real time. This is the case for the work in both Ertefaie (2015) and Luckett et al (2017). Such data can be collected retrospectively or by using SMART designs. A somewhat different approach to precision medicine in mHealth involves using data obtained from a micro-randomized clinical trial developed by (Klasnja et al 2015). These are particularly suited for developing interventions involving prompting people to take healthy actions to improve health behavior (Bekiroglu et al 2017). Often, these are designed to improve proximal outcomes and not necessarily longer term outcomes and can be framed as a contextual bandit problem (Tewari and Murphy 2017). This is an active and exciting area of precision medicine research.
4.3. Data sources and study design
Data for estimating dynamic treatment regimes can come from a range of sources including: convenience samples, planned observational studies, randomized clinical trials, and hybrid study designs (Zatzick et al 2016; Liu et al 2017). In many of these sources, including randomized clinical trials, the primary study objective is not estimation of an optimal treatment regime (Lavori and Dawson 2000, 2004; Murphy 2005a; Laber et al 2016); at best, estimation of an optimal treatment regime is a planned—but strictly exploratory—analysis. However, such data can still be a rich resource for estimation and inference for optimal treatment regimes. Electronic health records (EHR), for example, are collected for administrative or insurance purposes but have been shown to be useful for precision medicine research (see, e.g., Hripcsak et al 2016). Planned observational studies are usually conducted in epidemiology and other fields but usually involve careful design, planning, and execution so that the quality of data is high (see, e.g., (Thiese 2014)). These designs are frequently the inspiration for causal inference research as the absence of randomization treatment assignment complicates identification of causal relationships among treatments and risk factors. These designs can also include careful selection of subsets of convenience samples to improve quality of causal inference, as done, for example in Lund et al (2015).
Randomized clinical trials are a gold standard for data collection as they protect against unmeasured confounding and can be designed to ensure efficient estimation of the targeted estimand. For single-stage decision problems, a k-arm randomized clinical trial with equal randomization provides maximal information about average treatment effects across pairs of treatments. For multistage decision problems, Sequential Multiple Assignment Randomized Trials (SMARTs, Lavori and Dawson 2000, 2004; Murphy 2005a) allow for the efficient comparison of treatment sequences and fixed (i.e., not data-dependent) treatment regimes. In a SMART, a patient is randomized at each point in the treatment process where there is clinical equipoise and thus each patient may be randomized multiple times throughout the trial. Figure 2 shows a schematic for a two-stage SMART for evaluating behavioral interventions for cancer pain management (Kelleher et al 2017). In the first stage subjects were randomized with equal probability to one of two variants of Pain Coping Skills Training (PCST); in the second stage, responders were randomized to either a maintenance therapy or no further treatment whereas non-responders were randomized to either maintenance therapy or an intensified treatment. Two-stage SMARTs are among the most common though they are an extremely flexible design with many different variations (Kidwell 2014, 2016; Penn State Methodology Center 2018).
For various reasons, randomization at each decision point is not always possible, ethical, feasible or even scientifically optimal, and various hybrid designs can be considered. A pragmatic clinical trial (see, e.g., (Ford and Norrie 2016)) is one in which various design features are carefully incorporated to ensure similarity with treatment conditions to the real world. This can include randomizing clinics instead of patients to ensure that the treatment is the same throughout the clinic as would normally happen in practice, or recruiting a more heterogeneous patient population through liberal inclusion criteria. In a certain sense, a SMART clinical trial is more pragmatic than traditional clinical trials as the treatment decisions being evaluated are more similar to those utilized in practice. Hybrid designs have both randomization and observational components and often have a pragmatic motivation. One example is the enrichment design proposed by (Liu et al 2017) which allows the first treatment assignment to be non-randomized but has the second treatment assignment randomized. There are many other possibilities which we will not explore further here. Two important points we want to make are: first, that heterogeneity is good for precision medicine as this is needed to estimate an optimal treatment regime which is valid for a broad range of possible patients; and, second, that design of studies used for discovering precision medicine is a crucially important aspect of precision medicine research.
5. MANAGING MULTIPLE OUTCOMES
So far, we have considered optimal dynamic treatment regimes in terms of a single scalar outcome that we wish to optimize. In many clinical settings, there are multiple outcomes which need to be considered when managing treatment decisions. For example, in treating schizophrenia, there can be steep trade-off between side-effects and efficacy (Butler et al 2018). As another example, consider bipolar depression in which there is a trade-off between depressive symptoms and risk of mania (Luckett et al 2018). Recent work on precision medicine with multiple outcomes includes set-valuated treatment regimes which recommend a set of ‘acceptable’ treatments given a patient’s history (Fard 2009; Lizotte et al 2012; Laber et al 2014; Lizotte and Laber 2016; Wu 2016); and methods where a primary outcome is maximized while a secondary outcome is constrained to be within an acceptable region, using either regression based or outcome weighted learning methods (Linn et al 2015; Wang et al In press; Laber et al 2018).
In some cases, there is a trade-off between two or more endpoints which depends on patient preferences or other factors that depend on individual patient needs. Butler et al (2018) develop a method for the single-decision setting which uses item response theory to elicit patient preferences and then combines this with Q-learning to optimize the patient-preferred composite utility which is a convex combination of two outcomes. They show that under reasonable regularity and logic conditions, the optimal regime for any utility is equivalent to the optimal regime for a utility expressed as a convex combination of the available outcomes. They also show that the item-response model leads to the optimal patient-preferred treatment decisions under reasonable regularity conditions as the number of items grows and the sample size increases. One caveat of their approach is that the items in the patient preference instrument need to be appropriately calibrated which can be challenging. Extensions allowing calibration to be done empirically and which can be applied to both single- and multi- decision settings are given in Butler (2016).
Luckett et al (2018) study the situation where the trade-offs between two outcomes depend on complex individual-level factors about which clinicians have imperfect information. They consider observational data on clinicians prescribing anti-depressants to patients with bipolar depression and measures of both depression and mania outcomes are observed in the patients. They assume that the clinicians are trying to act optimally and that they sometimes succeed but not always. Based on the estimated Q-functions for each outcome (depression and mania), they estimate the weight in the combined utility of the convex combination of the two outcomes as a function of patient-level covariates as well as the probability of correct treatment assignment also as a function of patient-level covariates. They demonstrate that under reasonable regularity conditions, the asymptotic joint limiting distribution of the parameters are obtained at the rate. The limiting distribution is non-Gaussian and requires a non-standard bootstrap for inference. They demonstrate the validity of the inference through simulation studies and apply the method to the STEP-BD study data on bipolar disorder (Sachs et al 2007). They also demonstrate that applying the estimated dynamic treatment regime obtained from this data using the proposed method can lead to a statistically significant increased average patient-specific composite outcome for future patients. Generally speaking, addressing multiple outcomes in precision medicine is crucially important and there is much interesting work yet to be done in this area.
6. STATISTICAL INFERENCE
For many of the methods described above, asymptotic consistency, i.e., the property that the estimated quantities converge to the truth as the sample size grows, has been proven. For the sake of discussion, we refer to asymptotic consistency as zero order inference. However, typically with statistical procedures, it is valuable to be able to also provide first order inference consisting of confidence intervals, hypothesis tests, and sample size calculations. Generally speaking, first order inference is not yet known for many of the machine learning tools used in precision medicine and this is an open and active area of research. Because the focus is to inform decision making, a primary emphasis is on inference for performance of a treatment regime; note that a confidence or prediction set for the marginal mean outcome of an estimated optimal treatment regime is still meaningful even if models underlying estimation of the regime are misspecified.
Although first order inference for many of the machine learning approaches utilized has not been developed, some advances have been made in a number of settings, including for support vector machines (Laber and Murphy 2011) and for random forests (Wager and Athey In press). In addition, computation of error bounds can be a useful assessment of performance which is more precise than the presence or absence of consistency but not precise enough to obtain first order inference. These have been developed for many machine learning tools used in precision medicine, as in (Qian and Murphy 2011; Goldberg and Kosorok 2012; Zhao et al 2012; Cui et al 2017), and, more recently, have been improved for some settings in (Athey and Wager 2017). We also note the sample size formulas for the single-decision setting have been developed based on the value function (Laber et al 2016). Because regression-based approaches applied to the single-decision setting involve standard regression analyses, inference in this setting can sometimes be straightforward. However, in the multi-decision setting, for example with Q-learning involving two or more decision times, the inference is non-regular even if linear regression is used at each decision time (Chakraborty and Murphy 2010; Moodie and Richardson 2010; Chakraborty et al 2013, 2014; Laber et al 2014; Song et al 2015).
Inference for precision medicine is an active and important area of research, and we have only touched on it briefly here. We note that much of the inferential challenges follow from the use of complex machine learning procedures. One could argue that this is a reason to avoid machine learning methods in precision medicine. However, since the primary goal of precision medicine is to find dynamic treatment regimes that perform well on future patients, we need to use the best available tools, and this includes machine learning methods.
7. DISCUSSION
Precision medicine is beginning to emerge as a well-defined discipline with specific goals, areas of focus, and tailored methodology. Specifically, the primary goal is to discover treatment rules which leverage heterogeneity to improve clinical decision making in a manner that is reproducible, generalizable, and which can adapt as needed. We note that patient heterogeneity is a blessing for precision medicine although it may not be convenient for other areas of medical research. We also highlight the focus in precision medicine on discovery, as opposed to confirmatory research, and note that this makes the inferential aspects somewhat distinct from some areas of traditional medical research. Nevertheless, discovery of precision medicine should be confirmed rigorously just as with other medical discoveries. The emphasis on both discovery and heterogeneity makes machine learning tools particularly valuable in this quest, and this means that the inferential challenges are different and in many ways more difficult.
We also note that there are many other important supporting aspects of precision medicine which we have not discussed, including implementation, national policy questions, data storage and management, among many others. We also have not included numerous relevant research contributions to machine learning and other areas in many disciplines, both within statistics as well as outside, including many biomedical sciences, computer science, operations research, engineering, robotics, economics, and others.
Nevertheless, we hope that this review helps to clarify the goals of precision medicine and becomes a catalyst for bringing together the diverse disciplines and perspectives that are needed to make dramatic advances in precision medicine which will yield fundamental changes in human health and well being. This is an exciting and vibrant area of research with many open questions and tremendous opportunities.
ACKNOWLEDGMENTS
The authors thank the editor and reviewers for constructive feedback which led to an improved manuscript. Both authors were funded in part by grant P01 CA142538 from the National Cancer Institute, and the second author was also funded in part by grants DMS-1555141 from the National Science Foundation and R01 DK108073 from the National Institute of Diabetes and Digestive and Kidney Diseases.
Footnotes
DISCLOSURE STATEMENT
The authors are not aware of any affiliations, memberships, funding, or financial holdings that might be perceived as affecting the objectivity of this review.
There is little loss in generality in restricting attention to stationary regimes; under mild regularity conditions there exists a stationary regime that leads to a discounted marginal mean outcome at least as large as any other (possibly non-stationary) regime (Bertsekas 2005).
LITERATURE CITED
- Athey S, Wager S. 2017. Efficient policy learning. arXiv:1702.02896 [math.ST]
- Bai X, Tsiatis AA, Lu W, Song R. 2017. Optimal treatment regimes for survival endpoints using a locally-efficient doubly-robust estimator from a classification perspective. Lifetime Data Analysis 23:585–604 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bellman RE. 1957. Dynamic Programming Princeton NJ: Princeton University Press [Google Scholar]
- Bekiroglu K, Lagoa C, Murphy SA, Lanza ST. 2017. Control engineering methods for the design of robust behavioral treatments. IEEE Transactions on Control Systems Technology 25:979–90 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bembom O, van der Laan MJ. 2008. Analyzing sequentially randomized trials based on causal effect models for realistic individualized treatment rules. Statistics in medicine 27:3689–716. [DOI] [PubMed] [Google Scholar]
- Bertsekas DP. 2005. Dynamic programming and optimal control (Vol. 1, No. 3). Belmont MA: Athena scientific. [Google Scholar]
- Biernot P, Moodie EE. 2010. A comparison of variable selection approaches for dynamic treatment regimes. The International Journal of Biostatistics 6(1):Article 6 [DOI] [PubMed] [Google Scholar]
- Busoniu L, Babuska R, De Schutter B, Ernst D. 2010. Reinforcement Learning and Dynamic Programming Using Function Approximators New York: CRC press [Google Scholar]
- Butler EL. 2016. Using Patient Preferences to Estimate Optimal Treatment Strategies for Competing Outcomes. PhD Dissertation. University of North Carolina at Chapel Hill Department of Biostatistics [Google Scholar]
- Butler EL, Laber EB, Davis SM, Kosorok MR. 2018. Incorporating patient preferences into estimation of optimal individualized treatment rules. Biometrics 74:18–26 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cesa-Bianchi N, Lugosi G. 2006. Prediction, Learning, and Games Cambridge: Cambridge University Press [Google Scholar]
- Chakraborty B, Laber EB, Zhao YQ. 2013. Inference for dynamic treatment regimes using an m-out-of-n bootstrap scheme. Biometrics 69:714–23 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chakraborty B, Laber EB, Zhao YQ. 2014. Inference about the expected performance of a data-driven dynamic treatment regime. Clinical Trials, 11(4), 408–417. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chakraborty B, Moodie EE. 2013. Statistical Methods for Dynamic Treatment Regimes New York: Springer [Google Scholar]
- Chakraborty B, Murphy SA, Strecher V. 2010. Inference for non-regular parameters in optimal dynamic treatment regimes. Statistical Methods in Medical Research 19:317–343 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen J, Fu H, He X, Kosorok MR, Liu Y. Estimating individualized treatment rules for ordinal treatments. Biometrics In press [DOI] [PMC free article] [PubMed]
- Chen G, Zeng D, Kosorok MR. 2016. Personalized dose finding using outcome weighted learning (with discussion and rejoinder). Journal of the American Statistical Association 111:1509–47 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cook RD, Ni L. 2005. Sufficient dimension reduction via inverse regression: A minimum discrepancy approach. Journal of the American Statistical Association 100:410–428. [Google Scholar]
- Cui Y, Zhu R, Kosorok MR. 2017. Tree based weighted learning for estimating individualized treatment rules with censored data. Electronic Journal of Statistics 11:3927–53 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dawid AP. 2015. Statistical causality from a decision-theoretic perspective. Annual Review of Statistics and Its Application 2:273–303 [Google Scholar]
- Eddy DM. 1990. Practice policies: Guidelines for methods. Journal of the American Medical Association 263:1939–41 [Google Scholar]
- Elkan C 2001. The foundations of cost-sensitive learning. International Joint Conference on Artificial Intelligence 17:973–78 [Google Scholar]
- Ertefaie A 2015. Constructing dynamic treatment regimes in infinite-horizon settings. arX-ive:1406.0764v2 [stat.ME] [Google Scholar]
- Fan A, Lu W, Song R. 2016. Sequential advantage selection for optimal treatment regime. Annals of Applied Statistics 10:32–53 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fard MM. 2009. Non-Deterministic Policies In Markovian Processes Doctoral Dissertation, McGill University [Google Scholar]
- FDA-NIH Biomarker Working Group. 2016. BEST (Biomarkers, Endpoints, and Other Tools) Resource Silver Spring MD Food and Drug Administration Bethesda MD: National Institutes of Health; [PubMed] [Google Scholar]
- Fisher B, Redmond C, Brown A, Wickerham DL, Wolmark N, Allegra J, Escher G, Lippman M, Savlov E, Wittliff J, Fisher ER. 1983. Influence of tumor estrogen and progesterone receptor levels on the response to Tamoxifen and chemotherapy in primary breast cancer. Journal of Clinical Oncology 1:227–41 [DOI] [PubMed] [Google Scholar]
- Ford I, Norrie J. 2016. Pragmatic Trials. New England Journal of Medicine 375:454–63 [DOI] [PubMed] [Google Scholar]
- Gail M, Simon R. 1985. Testing for qualitative interactions between treatment effects and patient subsets. Biometrics 41:361–72 [PubMed] [Google Scholar]
- Goldberg Y, Kosorok MR. 2012. Q-learning with censored data. Annals of Statistics 40:529–60 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gordon AD. 1987. A review of hierarchical classification. Journal of the Royal Statistical Society, Series A 150:119–37 [Google Scholar]
- Gunter L, Zhu J, Murphy SA. 2011. Variable selection for qualitative interactions. Statistical methodology 8:42–55 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hastie T, Tisbhsirani R, Friedman J. 2009. The Elements of Statistical Learning: Data Mining, Inference, and Prediction New York: Springer [Google Scholar]
- Henderson R, Ansell P, Alshibani D. 2010. Regret-Regression for Optimal Dynamic Treatment Regimes. Biometrics 66:1192–1201 [DOI] [PubMed] [Google Scholar]
- Hillier FS, Liberman GJ. 1990. Introduction to Operations Research 5th Edition. New York: McGraw-Hill, Inc. [Google Scholar]
- Hripcsak G, Ryan PB, Duke JK, Shah NH, Park RW, Huser V, Suchard MA, Schuemie MJ, DeFalco FJ, Perotte A, Banda JM, Reich CG, Schilling LM, Matheny ME, Meeker D, Pratt N, Madigan D. 2016. Characterizing treatment pathways at scale using the OHDSI network. Proceedings of the National Academic of Sciences 113:7329–36 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jeng XJ, Lu W, Peng H. 2018. High-dimensional inference for personalized treatment decision. Electronic Journal of Statistics 12:2074–89 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kallus N 2018. Policy evaluation and optimization with continuous treatment. arXiv:1802.06037 [stat.ML] [Google Scholar]
- Kang C, Janes H, Huang Y. 2014. Combining biomarkers to optimize patient treatment recommendations (with discussion and rejoinder). Biometrics 70:695–707 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kelleher SA, Dorfman CS, Vilardaga JCP, Majestic C, Winger J, Gandhi V, Nunez C, Van Denberg A, Shelby RA, Reed SD, Murphy SA, Davidian M, Laber EB, Kimick GG, Westbrook KW, Abernathy AP, Somers TJ. 2017. Optimizing delivery of a behavioral pain intervention in cancer patients using a sequential multiple assignment randomized trial SMART. Contemporary Clinical Trials 57:51–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kidwell KM. 2014. SMART designs in cancer research: Past, present, and future. Clinical Trials 11:445–56 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kidwell KM. 2016. DTRs and SMARTs: Definitions, designs, and applications. In: Adaptive Treatment Strategies in Practice: Planning Trials and Analyzing Data for Personalized Medicine. Eds: Kosorok MR, Moodie EEM. Pp. 7–24. Philadelphia PA Alexandria VA ASA-SIAM Series on Statistics and Applied Probability [Google Scholar]
- Klasnja P, Hekler EB, Shiffman S, Boruvka A, Almirall D, Tewari A, Murphy SA. 2015. Micro-randomized trials: An experimental design for developing just-in-time adaptive interventions. Health Psychology 34(Suppl):1220–28 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kosorok MR, Moodie EEM. 2016. Adaptive Treatment Strategies in Practice: Planning Trials and Analyzing Data for Personalized Medicine. Philadelphia PA Alexandria VA ASA-SIAM Series on Statistics and Applied Probability
- Kravitz RL, Duan N, Braslow J. 2004. Evidence-based medicine, heterogeneity of treatment effects, and the trouble with averages. The Milbank Quarterly 82:661–87 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Laber EB, Linn KA, Stefanski LA. 2014. Interactive model building for Q-learning. Biometrika 101:831–47 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Laber EB, Lizotte DJ, Ferguson B. 2014. Set-valued dynamic treatment regimes for competing outcomes. Biometrics 70:53–61 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Laber EB, Lizotte DJ, Qian M, Pelham WE, Murphy SA. 2014. Dynamic Treatment Regimes: Technical Challenges and Applications. Electronic Journal of Statistics 8:1225–72 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Laber EB, Meyer NJ, Reich BJ, Pacifici K, Collazon JA, Drake J. 2018. Optimal treatment allocations in space and time for on-line control of an emerging infectious disease. Journal of the Royal Statistical Society, Series C In press [PMC free article] [PubMed] [Google Scholar]
- Laber EB, Murphy SA. 2011. Adaptive confidence intervals for the test error in classification (with discussion and rejoinder). Journal of the American Statistical Association 106:904–45 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Laber EB, Staicu AM. Functional feature construction for personalized dynamic treatment regimes. Journal of the American Statistical Association In press [DOI] [PMC free article] [PubMed] [Google Scholar]
- Laber EB, Wu F, Munera C, Lipkovich I, Colucci S, Ripa S. 2018. Identifying optimal dosage regimes under safety constraints: An application to long term opioid treatment of chronic pain. Statistics in medicine 37:1407–18 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Laber EB, Zhao YQ. 2015. Tree-based methods for individualized treatment regimes. Biometrika 102:501–14 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Laber EB, Zhao YQ, Regh T, Davidian M, Tsiatis A, Stanford JB, Zeng D, Song R, Kosorok MR. 2016. Using pilot data to size a two-arm randomized trial to find a nearly optimal personalized treatment strategy. Statistics in Medicine 35:1245–56 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lakkaraju H, Rudin C. 2017. Learning cost-effective and interpretable treatment regimes. In: Proceedings of the International 20th Conference on Artificial Intelligence and Statistics 54:166–75 [Google Scholar]
- Lavori PW, Dawson R. 2000. A design for testing clinical strategies: Biased adaptive within-subject randomization. Journal of the Royal Statistical Society, Series A 163:29–38 [Google Scholar]
- Lavori PW, Dawson R. Dynamic treatment regimes: Practical design considerations. Clinical Trials 1:9–20 [DOI] [PubMed] [Google Scholar]
- Lee JA, Verleysen M. 2007. Nonlinear dimensionality reduction. New York: Springer Science & Business Media. [Google Scholar]
- Li L (2007). Sparse sufficient dimension reduction. Biometrika 94:603–13 [Google Scholar]
- Linn KA, Laber EB, Stefanski LA. 2015. Estimation of dynamic treatment regimes for complex outcomes: Balancing benefits and risks. In: Adaptive Treatment Strategies in Practice: Planning Trials and Analyzing Data for Personalized Medicine. Eds: Kosorok MR, Moodie EEM. Pp. 249–62. Philadelphia PA Alexandria VA ASA-SIAM Series on Statistics and Applied Probability [Google Scholar]
- Linn KA, Laber EB, Stefanski LA. 2017. Interactive Q-learning for quantiles. Journal of the American Statistical Association 112:638–49 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Liu Y, Wang Y, Zeng D. 2017. Sequential multiple assignment randomized trials with enrichment design. Biometrics 73:378–90 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lizotte DJ, Bowling M, Murphy SA. 2012. Linear fitted-Q iteration with multiple reward functions. Journal of Machine Learning Research 13:3252–95 [PMC free article] [PubMed] [Google Scholar]
- Lizotte DJ, Laber EB. 2016. M ulti-objective Markov decision processes for data-driven decision support. The Journal of Machine Learning Research 17:7378–405 [PMC free article] [PubMed] [Google Scholar]
- Lonergan M, Senn SJ, McNamee C, Daly AK, Sutton R, Hattersley A, Pearson E, Pirmohamed M. Defining drug response for stratified medicine. 2017. Drug Discovery Today 22:173–79 [DOI] [PubMed] [Google Scholar]
- Longford NT, Nelder JA. 1999. Statistics versus statistical science in the regulatory process. Statistics In Medicine 18:2311–20 [DOI] [PubMed] [Google Scholar]
- Lu W, Zhang HH, Zeng D. 2013. Variable selection for optimal treatment decision. Statistical Methods in Medical Research 22:493–504 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Luckett DJ, Laber EB, Kahkoska AR, Maahs DM, Mayer-Davis E, Kosorok MR. 2017. Estimating dynamic treatment regimes in mobile health using V-learning. arXiv:1611.03531v2 [stat.ML] [DOI] [PMC free article] [PubMed] [Google Scholar]
- Luckett DJ, Laber EB, Kosorok MR. 2018. Estimation and optimization of composite outcomes. arXiv:1711.10581v2 [stat.ML] [PMC free article] [PubMed] [Google Scholar]
- Luedtke AR, van der Laan MJ. 2016a. Optimal individualized treatments in resource-limited settings. The International Journal of Biostatistics 12:283–303 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Luedtke AR, van der Laan MJ. 2016b. Super-learning of an optimal dynamic treatment rule. The International Journal of Biostatistics 12:305–32 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lund JL, Richardson DB, Stürmer T. 2015. The active comparator, new user study design in pharmacoepidemiology: historical foundations and contemporary application. Current Epidemiological Reports 2:221–8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Macia NF, Thaler GJ. 2005. Modeling and Control of Dynamic Systems New York: Thompson Delmar Learning [Google Scholar]
- Moodie EEM, Dean N, Sun YR. 2014. Q-learning: Flexible learning about useful utilities. Statistics in Biosciences 6:223–43 [Google Scholar]
- Moodie EEM, Richardson TS. 2010. Estimating optimal dynamic regimes: Correcting bias under the null. Scandinavian Journal of Statistics 37:126–46 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Murphy SA. 2003. Optimal dynamic treatment regimes (with discussion and rejoinder). Journal of the Royal Statistical Society, Series B 65:331–66 [Google Scholar]
- Murphy SA. 2005a. An experimental design for the development of adaptive treatment strategies. Statistics in Medicine 24:1455–81 [DOI] [PubMed] [Google Scholar]
- Murphy SA. 2005b. A generalization error for Q-learning. Journal of Machine Learning Research 6:1073–97 [PMC free article] [PubMed] [Google Scholar]
- Murphy SA, van der Laan MJ, Robins JM, Conduct Problems Prevention Research Group. 2002. Marginal mean models for dynamic regimes. Journal of the American Statistical Association 96:1410–23 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nahum-Shani I, Smith SN, Spring BJ, Collins LM, Witkiewitz K, Tewari A, Murphy SA. 2017. Justin-time adaptive interventions (JITAIs) in mobile health: key components and design principles for ongoing health behavior support. Annals of Behavioral Medicine 52:446–62 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nisio M. Stochastic Control Theory: Dynamic Programming Principle. 2nd Edition. Japan: Springer; 2015. [Google Scholar]
- Non-Hodgkin’s Lymphoma Prognostic Factors Project. 1993. A prediction model for aggressive non-Hodgkin’s lymphoma: The international non-Hodgkin’s lymphoma prognostic factors project. New England Journal of Medicine 329:987–94 [DOI] [PubMed] [Google Scholar]
- Olsen KS, Lund E. 2017. Population-based precision cancer screening-Letter. Cancer Epidemiology, Biomarkers and Prevention [DOI] [PubMed]
- Paduraru C. Off-policy Evaluation in Markov Decision Processes. PhD thesis, McGill University. 2013. [Google Scholar]
- Penn State Methodology Center. 2018. Projects Using SMARTs. https://methodology.psu.edu/ra/adap-inter/projects.
- Petersen ML, Porter KE, Gruber S, Wang Y, van der Laan MJ. 2012. Diagnosing and responding to violations in the positivity assumption. Statistical methods in medical research 21:31–54 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pires BA, Szepesvari C, Ghavamzadeh M. 2013. Cost-sensitive multiclass classification risk bounds. International Conference on Machine Learning 1391–9 [Google Scholar]
- Powell WB. 2007. Approximate Dynamic Programming: Solving the Curses of Dimensionality New York: Wiley and Sons [Google Scholar]
- Precup D 2000. Temporal abstraction in reinforcement learning. PhD thesis, University of Massachusetts Amherst. [Google Scholar]
- Puterman ML. 2005. Markov Decision Processes: Discrete Stochastic Dynamic Programming. 2nd Edition Hoboken NJ: Wiley and Sons [Google Scholar]
- Qian M, Murphy SA. 2011. Performance guarantees for individualized treatment rules. Annals of Statistics 39:1180–2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rich B, Moodie EE, Stephens DA, Platt RW. 2010. Model checking with residuals for g-estimation of optimal dynamic treatment regimes. The International Journal of Biostatistics 6(2):Article 12 [DOI] [PubMed] [Google Scholar]
- Robins JM. 1986. A new approach to causal inference in mortality studies with sustained exposure periods—application to control of the health worker survivor effect. Computers and Mathematics with Applications 7:1393–1512 [Google Scholar]
- Robins JM. 1997. Causal inference from complex longitudinal data. Lecture Notes in Statistics 120:69–117 [Google Scholar]
- Robins JM. 2004. Optimal structural nested models for optimal sequential decisions In: Proceedings of the Second Seattle Symposium in Biostatistics, Lecture Notes in Statistics 179:189–326. New York: Springer [Google Scholar]
- Robins JM, Hernan MA, Brumback B. 2000. Marginal structural models and causal inference in epidemiology. Epidemiology 11:550–60 [DOI] [PubMed] [Google Scholar]
- Rubin DB. 1978. Bayesian inference for causal effects: The role of randomization. Annals of Statistics 6:34–58 [Google Scholar]
- Rubin DB. 2005. Causal inference using potential outcomes: Designs, modeling, decisions. Journal of the American Statistical Association 100:322–31 [Google Scholar]
- Sachs GS, Nierenberg AA, Calabrese JR, Marangell LB, Wisniewski SR, Gyulai L, Friedman ES, Bowden CL, Fossey MD, Ostacher MJ, Ketter TA, Patel J, Hauser P, Rapport D, Martinez JM, Allen MH, Miklowitz DJ, Otto MW, Dennehy EB, Thase ME. 2007. Effectiveness of adjunctive antidepressant treatment for bipolar depression. New England Journal of Medicine 356:1711–22 [DOI] [PubMed] [Google Scholar]
- Schulte PJ, Tsiatis AA, Laber EB, Davidian M. 2014. Q- and A-learning methods for estimating optimal dynamic treatment regimes. Statistical Science 29:640–61 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shani G, Pineau J, Kaplow R. 2013. A survey of point-based POMDP solvers. Autonomous Agents and Multi-Agent Systems 27:1–51. [Google Scholar]
- Shortreed SM, Laber EB, Lizotte DJ, Stroup TS, Pineau J, Murphy SA. 2011. Informing sequential clinical decision-making through reinforcement learning: an empirical study. Machine learning 84:109–36 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shortreed SM, Laber EB, Scott ST., Pineau J, Murphy SA. 2014. A multiple imputation strategy for sequential multiple assignment randomized trials. Statistics in Medicine 33:4202–14 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shortreed SM, Moodie EE. 2012. Estimating the optimal dynamic antipsychotic treatment regime: evidence from the sequential multiple assignment randomized Clinical Antipsychotic Trials of Intervention and Effectiveness schizophrenia study. Journal of the Royal Statistical Society: Series C 61:577–99 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Si J (Ed.). 2004. Handbook of Learning and Approximate Dynamic Programming Volume 2 New York: John Wiley and Sons [Google Scholar]
- Song R, Kosorok MR, Zeng D, Zhao YQ, Laber EB, Yuan M. 2015. On sparse representation for optimal individualized treatment selection with penalized outcome weighted learning. Stat 4:59–68 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Song R, Wang W, Zeng D, Kosorok MR. 2015. Penalized Q-learning for dynamic treatment regimens. Statistica Sinica 25:901–20 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sorensen TIA. 1996. Which patients may be harmed by good treatments? The Lancet 348:351–2 [DOI] [PubMed] [Google Scholar]
- Stephens DA. 2016. G-estimation for dynamic treatment regimes in the longitudinal setting. In: Adaptive Treatment Strategies in Practice: Planning Trials and Analyzing Data for Personalized Medicine. Eds: Kosorok MR, Moodie EEM. Pp. 89–117. Philadelphia PA Alexandria VA ASA-SIAM Series on Statistics and Applied Probability [Google Scholar]
- Stusser RJ. 2006. Reflections on the scientific method in medicine In Medical Sciences. Eds: Verhasselt YLG, Jablensky AV, Pellegrini A, Sayegh C, Kazanjian A, Rantanen J, Mansourian BP, Wojtczak AM, Sayers BM, Szczerban J, Aluwihare APR, Napalkov NP, Brauer GW, Davies AM, Mahfouz SM, Manciaux MRG, Kitney R, Arata AA. Pp. 23–47. In Encyclopedia of Life Support Systems. Developed under the auspices of the UNESCO, Eolss Publishers, Oxford [Google Scholar]
- Sutton RS. 1997. On the significance of Markov decision processes International Conference on Artificial Neural Networks, Springer, Berlin, Heidelberg, Pp. 273–82 [Google Scholar]
- Sutton RS, Barto AG. 1998. Reinforcement Learning: An Introduction MIT Press; Cambridge, MA [Google Scholar]
- Szepesvari C 2010. Algorithms for reinforcement learning. Synthesis Lectures on Artificial Intelligence and Machine Learning 4:1–103 [Google Scholar]
- Tang Y, Kosorok MR. 2012. Developing adaptive personalized therapy for cystic fibrosis using reinforcement learning. The University of North Carolina at Chapel Hill Department of Biostatistics Technical Report Series. Working Paper 30.
- Taylor JM, Cheng W, Foster JC. 2015. Reader reaction to “A robust method for estimating optimal treatment regimes” by Zhang et al (2012). Biometrics 71:267–73 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Teran Hidalgo SJ, Lawson MT, Luckett DJ, Chaudhari M, Chen J, Choudhury A, Di Flrio A, Jiang X, Nguyen CT, Kosorok MR. 2016. A master pipeline for discovery and validation of biomarkers In: Machine Learning for Health Informatics: State-of-the-Art and Future Challenges Ed Holzinger A. Springer International Publishing; Pp. 259–88 [Google Scholar]
- Tewari A, Murphy SA. 2017. From ads to interventions: Contextual bandits in mobile health In: Mobile Health – Sensors, Analytic Methods, and Applications. Eds Regh J, Murphy S, Kumar S. Springer [Google Scholar]
- Thall PF, Leiko HW, Logothetis CJ, Millikan RE, Tannir NM. 2007. Bayesian and frequentist two-stage treatment strategies based on sequential failure times subject to interval censoring.Statistics in medicine 26:4687–702. [DOI] [PubMed] [Google Scholar]
- Thiese MS. 2014. Observational and interventional study design types: an overview. Biochemia Medica 24:199–210 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tian L, Alizadeh AA, Gentles AJ, Tibshirani R. 2014. A simple method for estimating interactions between a treatment and a large number of covariates. Journal of the American Statistical Association 109:1517–32 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Torkamani A, Wineinger NE, Topol EJ. 2018. The personal and clinical utility of polygenic risk scores. Nature Review Genetics [DOI] [PubMed]
- van der Laan MJ, Murphy SA, Robins JM. 2001. Analyzing dynamic regimes using structural nested mean models. Unpublished. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wager A, Athey S. Estimation and inference of heterogeneous treatment effects using random forests. Journal of the American Statistical Association In press [Google Scholar]
- Wallace MP, Moodie EEM, Stephens DA. 2018. Reward ignorant modeling of dynamic treatment regimes. Biometrical Journal [DOI] [PubMed] [Google Scholar]
- Wang Y, Fu H, Zeng D. Learning optimal personalized treatment rules in consideration of benefit and risk: with an application to treating type 2 diabetes patients with insulin therapies. Journal of the American Statistical Association In press [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang L, Laber EB, Witkiewitz K. (2018). Sufficient Markov Decision Processes. arXiv:1704.07531v2 [stat.ME] [Google Scholar]
- Wang J, Shen X, Pan W. 2009. On large margin hierarchical classification with multiple paths. Journal of the American Statistical Association 104:1213–23 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang Y, Wu P, Liu Y, Weng C, Zeng D. 2016. Learning optimal individualized treatment rules from electronic health record data. International Conference on Health care Informatics 65–71 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang L, Zhou Y, Song R, Sherwood B. Quantile-optimal treatment regimes. Journal of the American Statistical Association In press [DOI] [PMC free article] [PubMed] [Google Scholar]
- The White House: Office of the Press Secretary. 2015. FACT SHEET: President Obama’s Precision Medicine Initiative
- Wilensky U, Rand W. 2015. An Introduction to Agent-Based Modeling: Modeling Natural, Social, and Engineered Complex Systems with NetLogo Cambridge MA: Massachusetts Institute of Technology Press [Google Scholar]
- Wu T 2016. Set Valued Dynamic Treatment Regimes. PhD Dissertation, University of Michigan. [Google Scholar]
- Xu Y, Müller P, Wahed AS, Thall PF. 2016. Bayesian nonparametric estimation for dynamic treatment regimes with sequential transition times (with discussion and rejoinder). Journal of the American Statistical Association 111:921–50 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zadrozny B, Langford J, Abe N. 2003. Cost-sensitive learning by cost-proportionate example weighting. Third IEEE International Conference on Data Mining 435–442 [Google Scholar]
- Zatzick DF, Russo J, Darnell D, Chambers DA, Palinkas L, Van Eaton E, Wing J, Ingraham LM, Guiney R, Heagerty P, Comstock B, Whiteside LK, Jurkovich G. 2016. An effectiveness-implementation hybrid trial study protocol targeting posttraumatic stress disorder and comorbidity. Implementation Science 11:58. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang Y, Laber EB, Davidian M, Tsiatis AA. Decision lists for optimal dynamic treatment regimes. Journal of the American Statistical Association In press [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang B, Tsiatis BB, Davidian M, Zhang M, Laber EB. 2012. Estimating optimal treatment regimes from a classification perspective. Stat 1:103–14 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang B, Tsiatis AA, Laber EB, Davidian M. 2012. A robust method for estimating optimal treatment regimes. Biometrics 68:1010–18 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang B, Tsiatis AA, Laber EB, Davidian M. 2013. Robust estimation of optimal dynamic treatment regimes for sequential treatment decisions. Biometrika 100:681–94 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang B, Zhang M. (2016). Variable Selection for Estimating the Optimal Treatment Regimes in the Presence of a Large Number of Covariate. Technical Report: University of Michigan School of Public Health. Number 120
- Zhao YF, Kosorok MR, Zeng D. 2009. Reinforcement learning design for cancer clinical trials.Statistics in Medicine 28:3294–315 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhao YQ, Zeng D, Laber EB, Kosorok MR. 2015. New statistical learning methods for estimating optimal dynamic treatment regimes. Journal of the American Statistical Association 110:583–98 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhao YQ, Zeng D, Laber EB, Song R, Yuan M, Kosorok MR. 2015. Doubly robust learning for estimating individualized treatment with censored data. Biometrika 102:151–68 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhao YQ, Zeng D, Rush AJ, Kosorok MR. 2012. Estimating individualized treatment rules using outcome weighted learning. Journal of the American Statistical Association 107:1106–18 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhao YF, Zeng D, Socinski MA, Kosorok MR. 2011. Reinforcement learning strategies for clinical trials in non-small cell lung cancer. Biometrics 67:1422–33 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhou ZH, Liu XY. 2006. On multi-class cost-sensitive learning. AAAI 567–72 [Google Scholar]
- Zhou X, Mayer-Hamblett N, Khan U, Kosorok MR. 2017. Residual weighted learning for estimating individualized treatment rules. Journal of the American Statistical Association 112:169–87 [DOI] [PMC free article] [PubMed] [Google Scholar]