Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2013 Dec 1.
Published in final edited form as: Psychol Methods. 2012 Oct 1;17(4):478–494. doi: 10.1037/a0029373

Q-Learning: A Data Analysis Method for Constructing Adaptive Interventions

Inbal Nahum-Shani 1, Min Qian 2, Daniel Almirall 3, William E Pelham 4, Beth Gnagy 4, Greg Fabiano 4, Jim Waxmonsky 4, Jihnhee Yu 4, Susan Murphy 5
PMCID: PMC3747013  NIHMSID: NIHMS422586  PMID: 23025434

Abstract

Increasing interest in individualizing and adapting intervention services over time has led to the development of adaptive interventions. Adaptive interventions operationalize the individualization of a sequence of intervention options over time via the use of decision rules that input participant information and output intervention recommendations. We introduce Q-learning, which is a generalization of regression analysis to settings in which a sequence of decisions regarding intervention options or services are made. The use of Q is to indicate that this method is used to assess the relative quality of the intervention options. In particular, we use Q-learning with linear regression to estimate the optimal (i.e., most effective) sequence of decision rules. We illustrate how Q-learning can be used with data from sequential multiple assignment randomized trials (SMART; Murphy, 2005) to inform the construction of a more deeply tailored sequence of decision rules than those embedded in the SMART design. We also discuss the advantages of Q-learning compared to other data analysis approaches. Finally, we use the Adaptive Interventions for Children with ADHD SMART study (Center for Children and Families, SUNY at Buffalo, William E. Pelham as PI) for illustration.

Keywords: Decision Rules, Adaptive Interventions, Q-Learning, Regression

Introduction

The advantages of adaptive interventions are widely acknowledged in the behavioral and social sciences. In adaptive interventions the composition and/or the intensity of the intervention are individualized based on individuals’ characteristics or clinical presentation, and then adjusted in response to their ongoing performance (see, e.g., Bierman, Nix, Maples, & Murphy, 2006; Connell, Dishion, Yasui, & Kavanagh, 2007; Marlowe et al., 2008; Schaughency & Ervin, 2006). The conceptual idea of an adaptive intervention can be operationalized by using decision rules (Bierman et al., 2006) that link subjects’ characteristics and ongoing performance with specific subsequent intervention options (i.e., the type and the intensity/dosage of the intervention). The assignment of particular intervention options is based on participants’ values on tailoring variables—baseline and time-varying variables which strongly moderate the effect of certain intervention options, such that the type or intensity of the intervention should be tailored according to these moderators (see the companion paper, Nahum-Shani et al., 2011, for more details and examples of decision rules).

High-quality adaptive interventions are constructed by selecting good decision rules, namely decision rules that are expected to optimize the overall effectiveness of the sequence of tailored intervention options. In recent years intervention scientists have become increasingly interested in experimental designs and data analysis methods that are specifically suited for selecting high-quality decision rules for intervention development (Brown et al., 2009; Collins, Murphy & Strecher, 2007; Rivera, Pew & Collins, 2007). The sequential multiple assignment randomized trials (SMART) was developed to provide data specifically for the purpose of constructing adaptive interventions. In a SMART design, participants proceed through multiple intervention stages (corresponding to critical decision points) and at each stage, each individual is randomized among intervention options (see the companion paper, Nahum-Shani et al., 2011, for an overview of SMART designs). Nahum-Shani et al. (2011) provide methods that can be used to compare intervention options at different stages of the adaptive intervention as well as to compare the relatively simple adaptive interventions embedded in a SMART design. However, investigators are often interested in constructing interventions that are more complex than those embedded in the SAMRT design. For example, investigators often collect information concerning potential moderators (e.g., baseline characteristics of the individual and/or context, adherence to and/or side effects from prior intervention stages) and plan to use this information to investigate whether and how intervention options should be tailored according to these variables. In other words, investigators are interested in using additional information collected as part of the SMART study to explore ways to more deeply tailor the adaptive intervention. Accordingly, data analysis methods are needed in order to construct the best sequence of decision rules that employ additional potential useful tailoring variables at each intervention stage.

Here, we introduce Q-learning (Watkins, 1989)—a novel but straightforward methodology drawn from computer science that can be used for the construction of high-quality adaptive interventions from data. Our implementation of Q-learning will utilize a series of linear regressions to construct the sequence of decision rules (i.e., the sequence of adaptive intervention options) that maximizes a continuous outcome. Q-learning is used to assess the quality of the decision (i.e., the intervention option) at each critical decision stage (i.e., intervention stage), while appropriately controlling for effects of both past and subsequent adaptive decisions (i.e., adaptive intervention options).

We first provide a general framework for Q-learning with linear regression. Then, drawing on hypothetical examples from the area of goal-setting in organizational research (based on ideas from Erez, 1990 and Fried & Slowik, 2004), we illustrate how Q-learning can be employed to analyze data from four common types of SMART studies. We also compare Q-learning with other data analysis approaches that might be used in constructing adaptive interventions. Finally, we illustrate Q-learning using data from a SMART study aiming to develop an adaptive intervention for improving the school-based performance of children with attention deficit hyperactivity disorder (ADHD; Center for Children and Families, SUNY at Buffalo, William E. Pelham as PI).

Motivation for Q-learning

One way to develop a high-quality adaptive intervention is to use data to construct an optimal sequence of decision rules, namely an optimal sequence of adaptive or individualized intervention options. For example, assume an investigator is interested in finding the best way to set goal difficulty on a complex task in order to maximize employee performance. The investigator conducts a SMART study (see the companion paper Nahum-Shani et al., 2011 for more details concerning SMART designs) on N employee participants involving two critical goal-setting stages (i.e., intervention stages); at each stage there are two goal-setting options (i.e., two intervention options). At the first stage of the goal-setting process (e.g., the beginning of the year), employees were randomized with probability .5 to one of two goal-setting options. Let denote the randomized goal-setting options (i.e., the first-stage intervention options) at the first stage, coded -1 for a moderate goal (i.e., a goal with a moderate level of difficulty) and 1 for a difficult goal. At the second stage of the goal-setting process (e.g., the middle of the year), employees were re-randomized with probability .5 to one of two goal-setting options. Let A2 denote the randomized goal-setting options at the second stage of the goal-setting process, coded -1 for a moderate goal and 1 for a difficult goal. Let Y denote the supervisor’s annual assessment of the employee’s performance at the end of the second stage, coded so that high values are preferred.

Assume the investigator considers the baseline self-efficacy of the employee (i.e., the employee’s judgment of his/her capability to accomplish a certain level of performance; Bandura, 1986) as a candidate tailoring variable for the first-stage goal. Denote the employee’s baseline self-efficacy by O1. Assume the investigator also considers the attainment of the first-stage goal (i.e., whether or not the employee achieved the first-stage goal) as a candidate tailoring variable for the second-stage goal-setting options. Denote the attainment of the first-stage goal by O2. Accordingly, the data record for each of the N employees is O1,A1,O2,A2,Y1.

In general, Ot contains predictors of the primary outcome. O1 can be a vector of baseline measures (e.g., baseline performance, personality and demographic characteristics), and O2 can be a vector of intermediate outcomes measured prior to the second stage of the intervention (e.g., self-efficacy, affect, and goal commitment prior to the second stage of the goal-setting process). O1 and O2 might condition (moderate) the effects of the intervention options; additionally O2 might be affected by A1 and O1 (e.g., the attainment of the first-stage goal might depend on whether the first-stage goal was moderate or difficult, as well as on the employee’s baseline level of self-efficacy).

Now consider using the data on the N employees to construct a sequence of decision rules, that is, a sequence of goal-setting options that adapt to the employee’s baseline self-efficacy, as well as to the attainment of the first-stage goal. A decision rule at the first stage is denoted by d1, where the available information (employee’s baseline self-efficacy, denoted by O1) is the input, and the goal-setting option at the first-stage (a1) is the output. A decision rule at the second stage is denoted by d2. In this decision rule, the input is the available information on the employee’s baseline self-efficacy, the first-stage goal-setting option, and the attainment of the first-stage goal (O1,a1,O2), and the output is the second-stage goal-setting option (a2).

Suppose we are interested in using the data described above to construct an optimal adaptive intervention. Here, the word optimal means that if these decision rules were used to assign goals to the entire study population, then this would lead to the maximal expected annual assessment of the employee’s performance. Denote the optimal adaptive intervention by the sequence of decision rules (d1,d2). Q-learning is a method for using data to construct the decision rules (d1,d2) that operationalize the optimal adaptive intervention. Q-learning uses backwards induction (Bellman & Dreyfus, 1962) to construct a sequence of decision rules that map or link the observations (here captured by the tailoring variables) to the actions the agent (decision maker) ought to take in order to maximize a primary outcome. In terms of constructing an adaptive intervention, Q-learning can be used to find the sequence of decision rules that link the observed information concerning an individual (e.g., characteristics and responses to past intervention options) to the most efficient intervention type and intensity/dosage. Q-learning allows us to contrast the intervention options at each stage controlling for effects of both past and subsequent adaptive intervention options. This enables investigators to contrast intervention options when used as part of a sequence, as opposed to contrasting intervention options as stand-alone options for each stage (see the companion paper Nahum-Shani et al., 2011 for more details). In the following section we show how Q-learning can be used to construct an optimal sequence of decision rules.

Q-Learning

To illustrate the intuition behind Q-learning, it is useful to first consider the case in which an expert2 provides the multivariate distribution of O1,O2, and Y, for every sequence of decisions a1,a2. In this case, we construct the optimal sequence of decision rules using backwards induction as follows. We begin by finding the optimal decision rule at the second stage, namely d2(O1,a1,O2).

d2(O1,a1,O2)=argmaxa2Q2(O1,a1,O2,a2),

where Q2(O1,a1,O2,a2) = E[Y|O1,a1,O2,a2]. Accordingly, the optimal second-stage decision rule d2(O1,a1,O2) is the second-stage intervention option a2 for which Q2(O1,a1,O2,a2) (i.e., the expected primary outcome, conditional on information available up to the second stage) attains its maximal value. The use of Q to denote the conditional expectation is a mnemonic to indicate that this expectation is used to assess the quality of the intervention option. Q2 is the conditional expectation that provides the quality of choosing second-stage option a2, given the information available (O1,a1,O2).

Then, we move backwards in time to construct the optimal decision rule at the first stage, namely d1(O1).

d1(O1)=argmaxa1Q1(O1,a1)

where Q1(O1,a1) = E[maxaz QZ(O1,a1,O2,a2)|O1,a1] is the conditional expectation that provides the quality of choosing first-stage intervention option a1, controlling for the use of the best second-stage intervention option and given the information available (O1). Accordingly, the optimal first-stage decision rule d1(O1) equals the first-stage option a1 for which Q1(O1,a1) attains its maximal value (i.e., the first-stage intervention option that, given the information available up to the first stage, leads to the maximal expected mean outcome obtained by choosing the optimal second-stage intervention option). Q1 and Q2 are often called Q-functions (Sutton & Barto, 1998). Note that the optimal sequence of decision rules (d1,d2) output the first-stage and second-stage intervention options that maximize Q1,Q2 respectively.

Here, we focus on using SMART study data to construct the optimal sequence of decision rules, because we do not know the true multivariate distribution of O1,O2 and Y. We represent the study data by {O1i,A1i,O2i,A2i,Yi},i = 1,…,N, where N is the number of study participants. Throughout, for simplicity, we assume that participants are randomly assigned to one of two intervention options at each of the two intervention stages (e.g., A1 and A2 are randomized).

We estimate the Q-functions, from which we construct the optimal sequence of decision rules as described above, using Q-learning. In general Q-learning involves any one of a variety of regression techniques (linear regression, non-parametric regression, additive regression) and can be used with a variety of outcomes including longitudinal and/or binary, ordinal and continuous outcomes. For clarity we present Q-learning with linear regression and a continuous outcome Y. In this case, the second-stage Q-function might be modeled as

Q2(O1,A1,O2,A2;γ2,α2)=γ20+γ21O1+γ22A1+γ23O1A1+γ24O2+(α21+α22A1+α23O2)A2, (1)

where, γ2 = (γ20,γ21,γ22,γ23,γ24), and α2 = (α21,α22,α23). Notice that our main interest lies primarily in the parameters α2, because they contain information with respect to how the second-stage intervention (A2) should vary as a function of the candidate tailoring variables (here A1 and O2). Based on Equation (1) one can see that the second-stage intervention option (a2) that maximizes Q2 is the one that maximizes the term (α21 + α22A1 + α23O2)a2. If (α21 + α22A1 + α23O2) > 0, the term (α21 + α22A1 + α23O2)a2 attains its maximal value by a2 = 1; if (α21 + α22A1 + α23O2) < 0, the term (α21 + α22A1 + α23O2)a2 attains its maximal value by a2 = −1. We estimate the vector parameters γ2 and α2 by the following regression:

Yγ20+γ21O1+γ22A1+γ23O1A1+γ24O2+(α21+α22A1+α23O2)A2.

Next, we estimate the quality of the optimal second-stage option. This is

Y~i=maxa2Q2(O1i,A1i,O2i,a2;γ^2,α^2),i=1,,n.

Here, Y~i reduces to

Y~i=γ^20+γ^21O1i+γ^22A1i+γ^23O1iA1i+γ^24O2i+α^21+α^22A1i+α^23O2i. (2)

Y~i is the expected mean outcome obtained by choosing the optimal second-stage intervention option, given the information available (O1,a1,O2).

We use a linear regression for the first stage Q-function as well:

Q1(O1,A1,;γ1,α1)=γ10+γ11O1+(α11+α12O1)A1, (3)

where, γ1 = (γ10,γ11), and α = (α11,α12). Based on Equation (3), the first-stage option (a1) that maximizes Q1 is the value of a1 that maximizes the term (α11 + α12O1)a1; that is, if (α11 + α12O1) > 0, a1 = 1 maximizes the term (α11 + α12O1)a1, and if (α11 + α12O1) < 0, a1 = −1 maximizes the term (α11 + α12O1)a1. We again use regression to estimate γ1 and α1 as follows:

Y~γ10+γ11O1+(α11+α12O1)A1.

Notice that this time we regress the estimated quality of the optimal second-stage option (i.e., the maximal expected primary outcome obtained by taking the best second-stage intervention options) on A1,O1, and A1O1.

In summary, the estimated optimal sequence of decision rules (i.e., the best adaptive first-stage and second-stage intervention options) is

d^2(O1,A1,O2)=argmaxa2Q2(O1,A1,O2,a2,γ^2,α^2)=sign(α^21+α^22A1+α^23O2)d^1(O1)=argmaxa1Q1(O1,a1;γ^1,α^1)=sign(α^11+α^12O1),

where d^2(O1,A1,O2) is the estimated best second-stage intervention option (a2); that is, the second-stage intervention option that maximizes the mean of the primary outcome, given (O1,A1,O2); d^1(O1) is the estimated best first-stage intervention option (a1) that, given (O1), maximizes the mean of the maximal expected primary outcome (i.e., the maximal expected primary outcome obtained by taking the best second-stage intervention option).

Under the assumption that the linear models for Q1 and Q2 are correct and the observations from one individual to another are independent, the estimators of the regression coefficients are consistent (unbiased in large samples) for the true regression coefficients33. Also as is the case with all generalized linear models (e.g., logistic regression, ordinal regression, etc.), a crucial assumption is that the sample size is sufficiently large so that the distribution of the estimators for the regression coefficients can be well approximated by the normal distribution. Practically, the sample size must be larger if Y and/or any of the variables in O2 have highly non-continuous, skewed, or heavy tailed distributions than if all of these variables have continuous symmetric distributions.

Consider the construction of confidence intervals and/or hypothesis testing concerning the regression coefficients in Q2. Note that the second-stage linear regression for Q2 is an ordinary linear regression. Hence, in large samples bootstrap can be used to estimate standard errors, form confidence intervals and conduct hypothesis tests.4 Inference for the estimators of the regression coefficients in Q1 is less standard. To see this, consider Equation (2) and note that the formula for Y~ (i.e., the dependent variable for the first-stage regression) contains an absolute value function. Because the absolute value function is non-differentiable at the point 0, the distribution of the estimators of the regression coefficients in Q1 cannot be consistently approximated by standard methods such as the bootstrap. That is, in these cases, standard large-sample bootstrap-based tests and confidence intervals might perform poorly (Chakraborty, Murphy, & Strecher, 2010; Robins, 2004). To provide confidence intervals for the first-stage regression coefficients, methods that address the problem of non-differentiability are required. The software (provided at http://methodology.psu.edu/ra/adap-treat-strat) utilizes the Soft-thresholding with percentile bootstrap method in which Q-learning is applied to each bootstrap sample with one small adjustment to the formula for Y~. More specifically, the term α^21+α^22A1+α^23O22 is replaced by α^21+α^22A1+α^23O22(1λα^21+α^22A1+α^23O22)+, where λ is equal to 3(1,A1,O22)TΣ^2(1,A1,O22)N and Σ^2N is the estimated covariance matrix of α^2. This adjustment tests whether (α^21+α^22A1+α^23O22) is close to zero and, if so, shrinks α^21+α^22A1+α^23O22 to zero. Chakraborty et al. (2010) found that bootstrap intervals using this small adjustment achieve the desired confidence level across a wide variety of simulated settings.

Using Q-learning to Analyze Data from Four Different Types of SMART Studies

In the companion paper (Nahum-Shani et al., 2011) we describe four common SMART studies: (a) SMART designs that do not use intermediate outcomes as part of the experimental design (i.e., SMARTs with no embedded tailoring variables, as in Figure 2 of Nahum-Shani et al., 2011); (b) SMART designs in which whether or not to re-randomize depends on an intermediate outcome (as in Figure 1 of Nahum-Shani et al., 2011); (c) SMART designs in which participants are re-randomized to different second-stage intervention options depending on an intermediate outcome (as in Figure 3 of Nahum-Shani et al., 2011); and (d) SMART designs in which whether or not to re-randomize depends on an intermediate outcome and prior treatment (as in Figure 4 of Nahum-Shani et al., 2011). In the following, we illustrate the use of Q-learning with respect to data from each of these four types of SMART designs. In general, there are three main differences between the four designs in terms of the use of Q-learning. First, the regression models might differ. Second, the subsample of the SMART data used for estimating Q2 might differ. Third, there might be differences in the construction of the estimated quality of the optimal second-stage intervention option Y~. To clarify this, we use hypothetical examples from the area of goal-setting in organizational research.

SMARTs with no embedded tailoring variables

In these SMART designs all participants are re-randomized regardless of any observed information (e.g., intermediate outcomes such as response or adherence, or the intervention options offered in prior stages). Accordingly, assume an investigator obtained data from the goal-setting study described previously, in which at the first-stage (i.e., beginning of the year) and at the second stage (i.e., middle of the year) employees were randomly assigned (with .5 probability) to one of two goal-setting options (A1/A2= −1 for a moderate goal, A1/A2= 1 for a difficult goal). Recall that in addition the investigator obtained data on two candidate tailoring variables: self-efficacy at baseline (O1) and goal-attainment (O2). Assume self-efficacy is a continuous measure that the investigator has standardized (mean=0, STD=1). Also assume goal-attainment was measured in terms of whether (coded as 1) or not (coded as 0) the first-stage goal was achieved. The investigator is interested in using this data to obtain the optimal (i.e., in terms of the employee’s annual performance assessment) sequence of goal-setting options while assessing (a) whether and how to tailor the first-stage goal-setting options to the employee’s self-efficacy at baseline; and (b) whether and how to tailor the second-stage goal-setting options to the first-stage goal-setting option assigned to the employee, and to the employee’s attainment of the first-stage goal.

To apply Q-learning in this context, begin with the second stage of the goal-setting process, considering the first-stage goal-setting option (A1) as well as goal-attainment (O2), as candidate-tailoring variables for the second-stage goal-setting options. Here the model in Equation (1) can be used for Q2; regress the primary outcome Y on the predictors to obtain the parameter estimates γ^2,α^2. Based on the estimated regression coefficients, estimate the term (α^21+α^22A1+α^23O2) for every given level of A1 and O2. If (α21 + α22A1 + α23O2) > 0 the decision rule recommends assigning a difficult goal (A2 = 1) at the second stage; if (α21 + α22A1 + α23O22) < 0, the decision rule recommends assigning a moderate goal (A2 = −1) at the second stage of the goal-setting process. For example, if (α21 + α22 + α23) > 0, the decision rule recommends assigning a difficult goal at the second stage (A2 = 1) to employees who achieved (O2 = 1) a difficult goal (A1 = 1) at the first stage of the goal-setting process; if (α21 + α22) < 0 the decision rule recommends assigning a moderate goal at the second stage (A2 = −1) to employees who failed to achieve (O2 = 0) a difficult goal (A1 = 1) at the first stage of the goal-setting process.

Now move backwards in time to find the best first-stage goal-setting option (A1) controlling for the best second-stage goal-setting option (i.e., assuming all participants were assigned to the best second-stage goal-setting option, given their first-stage goal-setting option and goal-attainment). Use Equation (2) to estimate the quality, Y~, of the optimal second-stage goal-setting option. Then, model Q1 by Equation (3) and regress Y~ on the predictors to obtain γ^1 and α^1. Based on these estimated regression coefficients, estimate the term (α^11+α^12O1) for every value of O1 that is of special interest (e.g., the mean and mean ± 1 STD). If (α11 + α12O1) > 0, the decision rule recommends assigning a difficult goal (A1 = 1) at the first stage of the goal-setting process; and if (α11 + α12O1) < 0, the decision rule recommends assigning a moderate goal (A1 = −1) at the first stage of the goal-setting process. For example, if (α11 + α12) > 0, the decision rule recommends assigning a difficult first-stage goal (A1 = 1) to employees who reported relatively high levels of self-efficacy at baseline (i.e., when O1 = 1, that is the level of self-efficacy is one standard deviation above the sample mean); if (α11α12) < 0, the decision rule recommends assigning a moderate first-stage goal (A1 = −1) to employees who reported relatively low levels of self-efficacy at baseline (i.e., when O1 = −1; that is, the level of self-efficacy is one standard deviation below the sample mean).

SMARTs in which whether to re-randomized or not depends on an intermediate outcome

In these SMART designs an observed intermediate outcome (usually response or adherence to prior intervention option) is used to determine whether or not a participant should be re-randomized. Accordingly, consider data obtained from a two-stage goal-setting SMART study in which at the first-stage (e.g., the beginning of the year) employees were randomized (with .5 probability) to one of two goal-setting options: to receive a moderate goal (A1= −1), or to receive a difficult goal (A1= 1). At the second stage (e.g., the middle of the year) only employees who did not achieve the first-stage goal were re-randomized (with .5 probability) to one of two goal-setting options: to reduce the difficulty of the first-stage goal (A2= −1), or to maintain the level of difficulty of the first-stage goal (A2=1). Employees who achieved the first-stage goal were not re-randomized and received another goal, similar in its level of difficulty to the first-stage goal. Notice that in this SMART design, goal-attainment is a tailoring variable that is embedded in the design. It is used to determine whether the employee should be re-randomized at the second stage of the goal-setting process.

Assume that the investigator also obtained data on two candidate tailoring variables (that are not embedded in the design): the employee’s self-efficacy at baseline (O1); and the employee’s commitment to the first-stage goal, namely the employee’s unwillingness to abandon or change the initial goal (Donovan & Radosevich, 1998; O2). Assume both measures are continuous. The investigator is interested in using this data to estimate the optimal (i.e., in terms of the employee’s annual performance evaluation) sequence of goal-setting options that adapt to the employee’s goal-attainment, while assessing (a) whether and how the first-stage goal-setting options should be tailored to the employee’s level of self-efficacy at baseline; and (b) for employees who failed to achieve the first-stage goal, whether and how to tailor the second-stage goal-setting options to the first-stage goal-setting option assigned to the employee and to the employee’s commitment to the first-stage goal.

To apply Q-learning in this context, begin with the second stage of the goal-setting process, aiming to find the best second-stage goal-setting option for employees who failed to achieve the first-stage goal. Accordingly, use Equation (1) to model Q2 for employees who failed to achieve the first-stage goal and use only data from employees who failed to achieve the first-stage goal in the regression analysis for obtaining γ^2 and α^2. Based on the estimated regression coefficients obtain (α^21+α^22A1+α^23O2) for every given level of A1 and levels of O2 that are of special interest (e.g., mean, and mean ± 1SD). If (α21 + α22A1 + α23O2) > 0, the decision rule recommends maintaining the level of goal difficulty (A2 = 1) for employees who failed to achieve the first-stage goal; and if (α21 + α22A1 + α23O22) < 0 the decision rule recommends reducing the level of difficulty (A2 = −1) for employees who failed to achieve the first-stage goal.

Now, move backwards in time to find the best first-stage goal-setting option (A1), controlling for the best second-stage goal-setting option for employees who fail to achieve the first-stage goal. Accordingly, use Equation (2) to estimate Y~ (i.e., the quality of the optimal second-stage goal-setting option) for employees who failed to achieve the first-stage goal, and set Y~=Y for employees who achieved the first-stage goal (these employees were not re-randomized). Then, use Equation (3) to model Q1 and regress Y~ on the predictors to obtain γ^1 and α^1. Based on these estimated regression coefficients, estimate the term (α^11+α^12O1) for every value of O1 that is of special interest (e.g., the mean and mean ± 1 SD). If (α11 + α12O1) > 0, the decision rule recommends assigning a difficult goal (A1 = 1) at the first stage of the goal-setting process; and if (α11 + α12O1) < 0, the decision rule recommends assigning a moderate goal (A1 = −1) at the first stage of the goal-setting process.

SMARTs in which re-randomization to different second-stage intervention options depends on an intermediate outcome

In these SMART designs, an observed intermediate outcome is used to determine to which set of intervention options a participant should be re-randomized. Accordingly, consider data obtained from a two-stage goal-setting SMART study in which at the first stage of the goal-setting process employees were randomized (with .5 probability) to one of two goal-setting options: to receive a moderate goal (A1=−1), or to receive a difficult goal (A1=1). At the second stage of the goal-setting process employees who achieved the first-stage goal were re-randomized (with .5 probability) to one of two second-stage goal-setting options: to maintain the difficulty of the first-stage goal (A2= -1); or enhance the difficulty of the first-stage goal (A2=1). Employees who failed to meet the first-stage goal were re-randomized (with .5 probability) to one of two goal-setting options: to reduce the difficulty of the first-stage goal (A2= -1), or to maintain the difficulty of the first-stage goal (A2=1). Notice that the two second-stage goal-setting options (A2) are different depending on whether or not the first-stage goal was achieved.

Assume that the investigator also obtained data on three candidate tailoring variables: the employee’s self-efficacy at baseline (O1); the quality of the strategies the employee used (see Chesney & Locke, 1991) to achieve the first-stage goal (O21); and the employee’s commitment to the first-stage goal (O22). The investigator is interested in using this data to estimate the optimal (in terms of the employee’s annual performance evaluation) sequence of goal-setting options that adapt to an employee’s attainment of the first-stage goal, while assessing (a) whether and how to tailor the first-stage goal-setting options to the employee’s level of self-efficacy at baseline; (b) for employees who achieved the first-stage goal, whether and how to tailor the second-stage goal-setting options to the first-stage goal-setting option and to the quality of strategies the employee used to achieve the first-stage goal; and (c) for employees who failed to achieve the first-stage goal, whether and how to tailor the second-stage goal-setting options to the first-stage goal-setting option, and to the employee’s commitment to the first-stage goal. Notice that the candidate tailoring variables considered for employees who achieved the first-stage goal (A1,O21) are different from the candidate tailoring variables considered for employees who failed to achieve the first-stage goal (A1,O22).

To apply Q-learning in this context, begin with the second stage of the goal-setting process, aiming to find the best second-stage goal-setting option for employees who failed to achieve the first-stage goal and the best second-stage goal-setting options for employees who achieved the first-stage goal. Accordingly, model Q2 by

Q2(O1,A1,O21,O22,A2;γ2,α2)=γ20+γ21O1+γ22A1+γ23A1O1+γ24O21+γ25O22+[α21R+α22(1R)+α23A1R+α24A1(1R)+α25O21R+α26O22(1R)]A2,

where R indicates whether (R=1) or not (R=0) an employee achieved the first-stage goal5. That is, for employees who achieved the first-stage goal (R=1), Q2 is modeled by

Q2(O1,A1,O21,O22,A2;γ2,α2)=γ20+γ21O1+γ22A1+γ23A1O1+γ24O21+γ25O22+(α21+α23A1+α25O21)A2.

and for employees who failed to achieve the first-stage goal (R=0) Q2 is modeled by

Q2(O1,A1,O21,O22,A2;γ2,α2)=γ20+γ21O1+γ22A1+γ23A1O1+γ24O21+γ25O22+(α22+α24A1+α26O22)A2.

Accordingly, regress the primary outcome Y on the predictors to obtain γ^2 and α^2. Then, use these estimated regression coefficients to obtain (α^21+α^23A1+α^25O21) and (α22 + α24A1 + α26O22) for every given level of A1 and every value of O21 and O22 of special interest (e.g., mean, and mean ± 1 SD). For employees who achieved the first-stage goal, the decision rule recommends enhancing the difficulty (A2 = 1) of the goal if (α21 + α23A1 + α25O21) > 0 and to maintain the level of goal difficulty (A2 = −1) if (α21 + α23A1 + α25O21) < 0. For employees who failed to achieve the first-stage goal, the decision rule recommends maintaining the level of goal difficulty ( (A2 = 1) if (α22 + α24A1 + α26O22) > 0 and reducing the level of goal difficulty (A2= −1) if (α22 + α24A1 + α26O22) < 0.

Now, move backwards in time to find the best first-stage goal-setting option (A1), controlling for the best second-stage goal-setting option for employees who achieve the first-stage goal and for employee who fail to achieve the first-stage goal. Accordingly, use Y~=γ^20+γ^21O1+γ^22A1+γ^23A1O1+γ^24O21+α^21+α^23A1+α^25O21 to estimate the quality of the optimal second-stage goal-setting option for employees who achieved the first-stage goal, and Y~=γ^20+γ^21O1+γ^22A1+γ^23A1O1+γ^25O22+α^22+α^24A1+α^26O22 to estimate the quality of the optimal second-stage goal-setting option for employees who failed to achieve the first-stage goal. Then, use Equation (3) to model Q1 and regress Y~ on the predictors to obtain γ^1 and α^1. Based on these estimated regression coefficients, estimate the term (α^11+α^12O1) for every value of O1 that is of special interest (e.g., the mean and mean ± 1 STD). If (α11 + α12O1) > 0, the decision rule recommends assigning a difficult goal (A1 = 1) at the first stage of the goal-setting process; and if (α11 + α12O1) < 0, the decision rule recommends assigning a moderate goal (A1 = −1) at the first stage of the goal-setting process.

SMARTs in which whether to re-randomize or not depends on an intermediate outcome and prior treatment

In these SMART designs an intermediate outcome as well as the prior intervention option are used to determine whether or not a participant should be re-randomized. Accordingly, consider data obtained from a two-stage goal-setting SMART study in which at the first stage of the goal-setting process (i.e., beginning of the year) employees were randomized (with .5 probability) to one of two goal-setting options: to receive a moderate goal (A1=−1), or to receive a difficult goal (A1=1). At the second stage of the goal-setting process (e.g., middle of the year) only employees who failed to achieve a difficult goal at the first-stage were re-randomized (with .5 probability) to two options: to maintain the difficulty of the first-stage goal (A2= −1), or to reduce the difficulty of the first-stage goal (A2=1). Employees who achieved a difficult first-stage goal or employees who received a moderate first-stage goal were not re-randomized. More specifically, employees who achieved a difficult first-stage goal received another difficult goal. Employees who received a moderate first-stage goal received another moderate goal if they failed to achieve the first-stage goal, or a difficult goal if they achieved the first-stage goal.

Assume that in addition the investigator obtained information on two candidate tailoring variables: the employee’s self-efficacy at baseline (O1); and the employee’s commitment to the first-stage goal (O2). Assume the investigator is interested in using this data to estimate the optimal (i.e., in terms of the employee’s annual performance evaluation) sequence of goal-setting options that adapt to an employee’s attainment of the first-stage goal, while assessing (a) whether and how to tailor the first-stage goal-setting options to the employee’s level of self-efficacy at baseline; and (b) for employees who failed to achieve a difficult first-stage goal, whether and how to tailor the second-stage goal-setting options to the employee’s goal commitment.

To apply Q-learning in this context, begin with the second stage of the goal-setting process, aiming to find the best second-stage goal-setting option for employees who failed to achieve a difficult first-stage goal. Accordingly, model Q2 for employees who failed to achieve a difficult first-stage goal by

Q2(O1,O2,A2;γ2,α2)=γ20+γ21O1+γ22O2+(α21+α22O2)A2,

and use only data from employees who failed to achieve a difficult first-stage goal in the regression analysis for obtaining γ^2 and α^2. Then, use these estimated regression coefficients to obtain (α^21+α^22O2) for every value of O2 that is of special interest (e.g., mean, and mean ± 1 SD). The decision rule recommends maintaining the same goal (A2 = 1) for employees who failed to achieve a difficult first-stage goal if (α21 + α22O2) > 0 and to reduce the difficulty of the first-stage goal (A2= −1) if (α21 + α22O2) < 0.

Now, move backwards in time to find the best first-stage goal-setting option (A1), controlling for the best second-stage goal-setting option for employees who fail to achieve a difficult first-stage goal. Accordingly, use Y~=γ^20+γ^21O1+γ^22O2+α^21+α^22O2 to estimate the quality of the optimal second-stage goal-setting option for employees who failed to achieve a difficult first-stage goal. Set Y~=Y for employees who were not re-randomized (i.e., those who achieved a difficult first-stage goal or received a moderate goal at the first-stage). Then, use Equation (3) to model Q1 and regress Y~ on the predictors to obtain γ^1 and α^1. Based on these estimated regression coefficients, estimate the term (α^11+α^12O1) for every value of O1 that is of special interest (e.g., mean, and mean ± 1SD). If (α^11+α^12O1)>0, the decision rule recommends assigning a difficult goal (A1 = 1) at the first stage of the goal-setting process; and if (α^11+α^12O1)<0, the decision rule recommends assigning a moderate goal (A1 = −1) at the first stage of the goal-setting process.

Alternatives to Q-learning

A natural alternative to Q-learning is a single regression approach. More specifically, one might want to construct (d1,d2) using a single regression that includes the first-stage options, the second-stage options and the candidate tailoring variables. Consider, for example, the goal-setting SMART study with no embedded tailoring variables described above. The single regression equation might be

Yθ0+θ1O1+θ2A1+θ3O1A1+θ4O2+θ5A2+θ6A1A2+θ7A2O2. (4)

However, using estimates based on this equation to construct the optimal sequence of decision rules (d1,d2) is problematic in two main aspects. First, because O2 (e.g., goal-attainment) might be an outcome of A1 and a potential predictor of Y, O2 cuts off any portion of the effect of A1 on Y that occurs via O2. To clarify this, O2 can be conceptualized as a mediator in the relationship between A1 and Y (e.g., the effect of the first-stage goal-setting options on annual performance assessment can be transmitted through the attainment of the first-stage goal). Adding O2 to a regression in which A1 is used to predict Y will reduce the effect of A1. In the presence of O2, the coefficient for A1 no longer expresses the total effect of the first-stage goal-setting options on the outcome, but rather what is left of the total effect (the direct effect) after cutting off the part of the effect that is mediated by A1 (the indirect effect; Baron & Kenny, 1986; MacKinnon, Warsi, & Dwyer, 1995). Note that ascertaining the total effect of the intervention options (e.g., the goal-setting options) at a given stage (say, A1) is crucial to finding the best decision rule (e.g., d1), because it provides information concerning the overall effect of the these intervention options. Although the direct effect of the intervention options at a given stage might be helpful in identifying mechanisms or processes through which these intervention options might affect the outcome, it is not as helpful in deciding which intervention option is superior. Accordingly, any inference concerning the optimal intervention option at the first stage based on Equation (4) is likely to be misleading.

Second, even if O2 is not a mediator, the coefficients of the A1 terms (main effects and interactions) in Equation (4) can be impacted by unknown causes of both O2 and Y so that A1 might appear to be falsely less or more correlated with Y. This bias occurs when A1 affects O2 while O2 and Y are affected by the same unknown causes (see Figure 1).

Figure 1.

Figure 1

Illustration of unmeasured confounders affecting O2 and Y.

In order to demonstrate the way in which Q-learning reduces the bias resulting from unmeasured causes, consider again the goal-setting SMART trial with no embedded tailoring variables described above. For simplicity, assume there are no baseline variables O1. Also assume U~N(0,1) is an unmeasured cause (say a personality characteristic) that has an effect on the annual performance assessment (Y) and the attainment of the first-stage goal (O2). More specifically, suppose Y = 1 + 0.5U + εY, and O2 = 1 + 0.5U + 0.5A1 + ε0. For both models the error terms (ε’s) are independent and standard normally distributed. Notice that in this example, O2 does not mediate the relationship between A1 and Y; A1 affects O2, but neither A1 nor A2 affect Y. We generated 1,000 samples (N=500 each) using the above example. On each data set we used the single-regression approach and Q-learning. The single-regression model is

Yθ0+θ1A1+θ2O2+θ3A2+θ4A1A2. (5)

A natural approach to using Equation (5)6 to construct the optimal sequence of decision rules is as follows: we construct the optimal decision rule at the second stage by finding the value of A2 that maximizes Equation (5) (i.e., that maximizes the term [θ^3+θ^4A1]A2). That is, d^2(A1)=sign(θ^3+θ^4A1). Replacing A2 by sign(θ^3+θ^4A1), the estimated maximal expected outcome is

θ^0+θ^1A1+θ^2O2+θ^3+θ^4A1. (6)

Now, we rewrite the maximal expected outcome in Equation (6) as

θ^0+θ^1A1+θ^2O2+θ^3+θ^4A1=θ^0+θ^1A1+θ^2O2+A1+12θ^3+θ^4+1A12θ^3θ^4=θ^0+θ^2O2+12(θ^3+θ^4+θ^3θ^4)+[θ^1+12(θ^3+θ^4θ^3θ^4)]A1.

Next, we find the value of A1=1 that maximizes the above. Accordingly, if [θ^1+12(θ^3+θ^4θ^3θ^4)]>0, we can conclude that A1=1 (a difficult goal) is the best first-stage goal-setting option, given that we are going to choose the best second-stage goal-setting option. If [θ^1+12(θ^3+θ^4θ^3θ^4)]<0 we conclude that A1= −1 (a moderate goal) is the best first-stage goal-setting option, given that we are going to choose the best second-stage goal-setting option.

On the other hand, consider Q-learning. In analogy to Equation (5) we use the models

Q2(A1,O2,A2;γ2,α2)=γ20+γ21A1+γ22O2+α21A2+α22A1A2andQ1(A1;γ1,α1)=γ10+α11A1.

Applying the Q-learning algorithm, we obtain estimates of the parameters (γ^j,α^j),j=1,2. We estimate the best second-stage intervention options by choosing A2=sign(α^21+α^22A1), and the best first-stage intervention option by choosing A1=sign(α^11). Using this approach α^11 is the estimated effect of the first-stage goal-setting options, given that we are going to choose the best second-stage goal-setting option.

In conclusion, we know that the sign of [θ^1+12(θ^3+θ^4θ^3θ^4)] determines which first-stage goal-setting option is selected as best in the single-stage regression approach, whereas the sign of α^11 determines which first-stage goal-setting option is selected as best in Q-Learning. We compare the distribution of these two quantities across the 1000 generated samples. Recall that in our example the first-stage goal-setting options have no effect (i.e., the effect of A1 equals zero); thus both distributions should be centered at zero. Figure 2 presents the distribution of [θ^1+12(θ^3+θ^4θ^3θ^4)] and Figure 3 presents the distribution of α^11. It is easy to see that the distribution of the Q-learning-based estimate is centered around zero (SD = .06), while the distribution of the single-regression-based estimate has a mean of −.10 (SD = .06). Thus, if there are unobserved causes of both O2 and Y, the single-regression approach in Equation (6) might lead to erroneous conclusions concerning the best sequence of decision rules. In contrast, the Q-learning method provides unbiased estimators of the parameters needed to construct the optimal sequence of decision rules.

Figure 2.

Figure 2

Distribution of estimated coefficient [θ^1+12(θ^3+θ^4θ^3θ^4)]..

Figure 3.

Figure 3

Distribution of estimated coefficient of (α^11).

Data Example: Adaptive Interventions for Children with ADHD

Attention-Deficit Hyperactivity Disorder (ADHD) is a chronic disorder affecting 5-10% of school age children. It adversely impacts functioning at home, at school and in social settings (Pliszka 2007). The limited success of pharmacological and behavioral interventions in the treatment of childhood ADHD has led to the now-common clinical practice of combining these two modalities (see Pelham et al., 2000). While the literature clearly supports the efficacy of this combined approach to treatment, little is known about the optimal way to sequence pharmacological and behavioral interventions (Pelham & Fabiano, 2008; Pelham & Gnagy, 1999). Accordingly, a SMART study was conducted (William E. Pelham, PI) with the general aim to find the optimal sequence of intervention options to reduce ADHD symptoms and improve school performance among children.

Design and Research Questions

Recall that the observable SMART study data for one participant is denoted by {O1,A1,O2,A2,Y}. In the first stage of the ADHD SMART study (at the beginning of a school year) children were randomly assigned (with probability .5) to a low dose of medication (A1 coded as −1) or a low-intensity behavioral intervention (A1 coded as 1). Beginning at the eighth week, each child’s response to the first-stage intervention was evaluated monthly until the end of that school year. Monthly ratings from the Impairment Rating Scale (IRS; Fabiano et al., 2006; available from http://wings.buffalo.edu/adhd), and an individualized list of target behaviors (ITB) (e.g., Pelham et al., 2002; Pelham, Evans, Gnagy, & Greenslade, 1992) were used to evaluate response. At each monthly assessment, children whose average performance on the ITB was less than 75% and who were rated by teachers as impaired on IRS in at least one domain were designated as inadequate responders to the first stage of the intervention. If the child was classified as a responder, then he/she remained in the first stage of the intervention and continued the assigned first-stage intervention option. If the child was classified as an inadequate responder, he/she entered the second stage of the intervention. These children were re-randomized (with probability .5) to one of two second-stage intervention options, either to increasing the dose/intensity of the first-stage intervention option (A2 coded as −1) or to augmenting the first-stage intervention option with the other type of intervention (i.e., adding behavioral intervention for those who started with medication, or adding medication for those who started with behavioral intervention; A2 coded as 1). Note that there are only two key decisions in this trial: the first-stage intervention decision (A1), and then the second-stage intervention decision (A2) for those not responding satisfactorily to the first stage of the intervention. The structure of this study is illustrated in Figure 4.

Figure 4.

Figure 4

Sequential Multiple Assignment Randomized Trial for ADHD study.

By design, the only embedded tailoring variable in the ADHD study is whether or not the child responded to the first stage of the intervention. However, it is interesting to assess whether the data can be used to construct a more deeply tailored adaptive intervention. For example, the investigators might be interested in assessing whether and how (a) the first stage of the intervention should be tailored according to whether or not the child received medication prior to the first stage of the intervention; (b) the second stage of the intervention should be tailored according to the child’s level of adherence to the first stage of the intervention; and (c) the second stage of the intervention should be tailored according to the intervention option offered at the first stage. Q-learning can be used to estimate the best sequence of decision rules while evaluating these three candidate tailoring variables.

Sample

149 children (75% boys) between the ages of 5-12 (mean 8.6 years) participated in the study. Due to drop-out and missing data7, the effective sample used in the current analysis was 138. At the first stage of the intervention, 70 children were randomized to low dose of medication, and 68 were randomized to low dose of behavioral intervention. By the end of the school year, 81 children had met the criteria for non-response at one of the monthly evaluations and had been re-randomized to one of the two second-stage intervention options (40 non-responding children were assigned to increasing the dose of the first-stage intervention, and 41 non-responding children were assigned to augmenting the first-stage intervention with the other type of intervention).

Measures

Primary Outcome (Y)

The level of children’s classroom performance based on the Impairment Rating Scale (IRS) after an 8-month period is our primary outcome. This outcome ranges from 1 to 5, with higher values reflecting better classroom performance.

Medication Prior to First-Stage Intervention (O11)

This measure reflects whether the child did (coded as 1) or did not (coded as 0) receive medication at school during the previous school year (i.e., prior to the first stage of the intervention).

Baseline measures

(a) ADHD symptoms at the end of the previous school year, reflecting the mean of teacher’s evaluation on 14 ADHD symptoms (the Disruptive Behavior Disorders Rating Scale; Pelham et al., 1992), ranging from 0 to 3 and inverse coded so that larger values reflect fewer symptoms (i.e., better classroom performance; labeled O12); (b) oppositional defiant disorder (ODD) diagnosis indicator, reflecting whether the child was (coded as 1) or was not (coded as 0) diagnosed with ODD before the first-stage intervention (labeled O13).

Month of non-response (O21)

The month during the school year at which the child showed inadequate response to the first stage of the intervention, and hence entered the second stage of the intervention. This measure is relevant only for those who showed inadequate response during the school year (i.e., classified as non-responders to the first stage of the intervention).

Adherence to first-stage intervention (O22)

This measure reflects whether adherence to the first-stage intervention was high (coded as 1) or low (coded as 0). We constructed this indicator based on two other measures that express (a) the percentage of days the child received medication during the school year calculated based on pill counts (for those assigned to low-dose medication at the first stage of the intervention), and (b) the percentage of days the child received the behavioral intervention during the school year based on the teacher’s report of behavioral interventions used in the classroom (for those assigned to behavioral intervention at the first stage of the intervention). The distributions of these two measures are presented in Figures 5 and 6. Based on these distributions, we constructed O22, such that for those assigned to behavioral intervention at the first stage of the intervention, low adherence (O22 =0) means receiving less than 75% days of behavioral intervention, and for those assigned to medication at the first stage of the intervention, low adherence (O22=0) reflects receiving less than 100% days of medication8.

Figure 5.

Figure 5

Distribution for % days on behavioral intervention for those assigned to low-intensity behavioral intervention at the first stage of the intervention.

Figure 6.

Figure 6

Distribution for % days on medication for those assigned to low dose of medication at the first stage of the intervention.

Data Analysis Procedure

Using the Q-learning approach, the optimal sequence of decision rules can be estimated based on two regressions, one for each intervention stage. We start from the second stage, aiming to find the best second-stage intervention option for non-responding children, given the information we have up to the second stage (O11,O12,O13,A1,O21,O22). Because children were classified as non-responders at different time points along the school year, we included the month of non-response (O21) in this regression. We also included the two baseline measures (O12,O13) in the regression in order to reduce error variance. We consider the first-stage intervention (A1) as well as the level of adherence to the first-stage intervention (O22), as candidate tailoring variables for the second stage of the intervention. Q2 for non-responders is modeled by

Q2(O11,O12,O13,A1,O21,O22,A2;γ2,α2)=γ20+γ21O11+γ22O12+γ23O13+γ24A1+γ25A1O11+γ26O21+γ27O22+(α21+α22A1+α23O22)A2. (7)

In general, this regression might include further baseline variables or other potential tailoring variables such as negative/ineffective parenting styles and medication side effects. We obtained γ^2,α^2 by using regression on the data from the children who did not respond in the first stage. In this simple case, the decision rule recommends augmenting the first-stage intervention option with the alternative type of intervention (A2 = 1) for a child who does not respond to the first stage of the intervention if (α21 + α22A1 + α23O22) > 0 and increasing the dose of the first-stage intervention option (A2 = −1) if (α21 + α22A1 + α23O22) < 0. We used the GLM procedure in SAS to obtain estimated coefficients based on Equation (7). To obtain (α^21+α^22A1+α^23O22) for every combination of A1 and O22, we used the Estimate statement in GLM (this statement enables the researcher to estimate linear combinations of the regression parameters and their standard errors). More specifically, for A1 = 1 and O2=0, we obtained (α^21+α^22); for A1 = −1 and O22=0, we obtained (α^21+α^22); for A1 = 1 and O22=1, we obtained (α^21+α^22+α^23); and for A1 = −1 and O22=1, we obtained (α^21+α^22+α^23). The standard test statistic (t-test) provided by the GLM procedure was used to assess whether each of these estimates significantly differ from zero. Additionally, because A2 can obtain −1/1 values, we estimated the difference between the two second-stage intervention options conditional on A1 and O22 (e.g., the estimated simple effect of A2) by (α^21+α^22A1+α^23O22)(α^21α^22A1α^23O22)=2(α^21+α^22A1+α^23O22)..

Now we move backwards in time, aiming to find the best first-stage intervention option (A1) controlling for the best second-stage intervention option. Based on Equation (7) the estimated quality of the optimal second-stage intervention option for non-responders is

Y~=γ^20+γ^21O11+γ^22O12+γ^23O13+γ^24A1+γ^25A1O11+γ^26O21+γ^27O22+α^21+α^22A1+α^23O22.

Because responders remain on their first-stage intervention option, we set Y~=Y for responders. Q1 is modeled by

Q1(O11,O12,O13,A1;γ1,α1)=γ10+γ11O11+γ12O12+γ13O13+(α11+α12O11)A1

Again we used the SAS GLM procedure to regress Y~ on the predictors and obtain γ^1 and α^1. We used the Estimate statement in GLM to obtain (α^11+α^12O11) for every level of O11. If (α11 + α12O11) > 0, the best first-stage intervention option would be low-intensity behavioral intervention (A1 = 1). If (α11 + α12O11) < 0, the best first-stage intervention option would be low dose of medication (A1 = −1). Additionally, because A1 can obtain −1/1 values, the estimated difference between the two first-stage intervention options conditional on O11 (e.g., the estimated simple effect of A1) is (α^11+α^12O11)(α^11α^12O11)=2(α^11+α^12O11).. We used the soft-thresholding method (Chakraborty et al., 2009) to provide confidence intervals for the first-stage regression coefficients.

Results

Table 1 presents the results for the second-stage regression. Based on these estimates, we estimated the formula (α^21+α^22A1+α^23O22) for every given combination of A1 and O22 (see Table 2).

Table 1.

Estimated Coefficients for Q2 (N=81).

Effect Estimate SE Lower limit
95% CI
Upper limit
95% CI
Intercept 1.36 0.53
O11 (medication prior to first-stage intervention) −10.27 0.31
O12 (baseline: ADHD symptoms) 0.94 0.26
O13 (baseline: ODD diagnosis) 0.93 0.28
O21 (month of non-response) 0.02 0.10
O22 (adherence to first-stage intervention) 0.18 0.27
A1 (first-stage intervention options) 0.03 0.14
A2 (second-stage intervention options) −0.72 0.22 −1.15 −0.29
O22*A2 (adherence to first-stage
intervention*second-stage intervention options)
0.97 0.28 0.41 1.53
A1*A2 (first-stage intervention options*second-stage
intervention options)
0.05 0.13 −0.22 0.32

Table 2.

Estimates of (α^21+α^22A1+α^23O22) for every combination of A1 and O22 (N=81)

A1 O22 Estimated
(α^21+α^22A1+α^23O22)
SE Lower limit
95% CI
Upper limit
95% CI
−1 (medication) 1 (high adherence) 0.20 0.23 −0.26 0.67
−1 (medication) 0 (low adherence) −0.77 0.27 −1.30 −0.23
1 (behavioral
intervention)
1 (high adherence) 0.30 0.22 −0.13 0.74
1 (behavioral
intervention)
0 (low adherence) −0.67 0.24 −1.14 −0.19

The results in Table 1 show that the effect of the second-stage intervention options (A2) is negative and statistically significant α^21=.72, lower limit 95% CI= −1.15, upper limit 95% CI= −.29). Although the interaction between the first-stage intervention options (A1) and the second-stage intervention options (A2) is not statistically significant (α^22=0.05, lower limit 95% CI= −.22, upper limit 95% CI= .32), the interaction between adherence to the first stage of the intervention (O22) and the second-stage intervention options (A2) is statistically significant (α^23=.97, lower limit 95% CI=.41, upper limit 95% CI= 1.53).

The results in Table 2 indicate that when adherence to the first stage of the intervention is low (O22 = 0), the term (α^21+α^22A1) is negative and statistically significant, regardless of whether the first-stage intervention option was low dose of medication (α^21α^22=.77, lower limit 95% CI= −1.30, upper limit 95% CI= −0.32), or low-intensity behavioral intervention (α^21+α^22=.67, lower limit 95% CI= −1.14, upper limit 95% CI= −.19). Accordingly, when adherence to the first stage of the intervention is low, the term (α21 + α22A1)A2 is maximized when A2 = −1 (augment the first-stage intervention option with the alternative type of intervention). However, when adherence to the first stage of the intervention is high (O22 = 1), the term (α^21+α^22A1+α^23) was not found to be significantly different from zero, regardless of whether the first-stage intervention option was low dose of medication ((α^21α^22+α^23=.20; lower limit 95% CI= −.26, upper limit 95% CI= .67) or low-intensity behavioral intervention (α^21+α^22+α^23=.30, lower limit 95% CI= −.13, upper limit 95% CI= .74).

Overall, the results of the second-stage regression suggest that if a child does not respond to the first stage of the intervention (regardless of whether the first-stage intervention option was low dose of medication or low-intensity behavioral intervention), and if adherence to the first stage of the intervention is low, augmenting the first-stage intervention option with the alternative type of intervention (A2 = −1), leads to better classroom performance relative to increasing the dose/intensity of the first-stage intervention option (A2 = 1). However, if adherence to the first stage of the intervention is high, there is inconclusive evidence with respect to the difference between the two second-stage intervention options. Figure 7 presents the predicted means for each of the second-stage intervention options (A2), given the first-stage intervention options (A1) and adherence to the first stage of the intervention (O22).

Figure 7.

Figure 7

Predicted mean of classroom performance for each of the second-stage intervention options (A2), given the first-stage intervention (A1) and adherence to first-stage intervention (O22).

Table 3 presents the results for the first-stage regression. Based on these estimates, we estimated the term (α^11+α^12O11) for each value of O11 (see Table 4). The results in Table 4 indicate that the effect of the first-stage intervention options (A1) is positive and marginally significant (α^11=.17, lower limit 90% CI= −.01, upper limit 90% CI=.34), and the interaction between the first-stage intervention options (A1) and medication prior to the first stage of the intervention (O11) is negative and marginally significant (α^12=.32, lower limit 90% CI= −.59, upper limit 90% CI= −.06).

Table 3.

Estimated coefficients and soft-threshold confidence intervals for Q1 (N=138).

Effect Estimate SE Lower limit
90% CI
Upper limit
90% CI
Intercept 2.61 0.16
O11 (medication prior to first-stage intervention) −0.37 0.14
O12 (baseline: ADHD symptoms) 0.73 0.11
O13 (baseline: ODD diagnosis) 0.75 0.13
A1 (first-stage intervention options) 0.17 0.07 −0.01 0.34
O11*A1 (medication prior to first-stage
intervention*first-stage intervention options)
−0.32 0.14 −0.59 −0.06

Table 4.

Estimates of (α^11+α^12O11) for each level of O11.

O11 Estimated
(α^11+α^12O11)
SE Lower limit
90% CI
Upper limit
90% CI
1 (medication prior to first-stage
intervention)
−0.15 0.12 −0.44 0.11
0 (no medication prior to first-stage
intervention)
0.17 0.07 −0.01 0.34

Based on the mates in Table 4, we estimated the formula (α^11+α^12O11) for every given value of O11 (see Table 5). The results in Table 5 indicate that when O11 = 0, the term (α^11) is positive and marginally significt (Estimate = .17, anlower limit 90% CI= −.01, upper limit 90% CI= .34). However, when O11 =1 the term (α^11+α^12) is not significantly different from zero (estimate= −.15, lower limit 90% CI= −.44, upper limit 90% CI= .11). This means that controlling for the optimal second-stage intervention option (offered to non-responders), low dose of behavioral intervention (A1=1) leads to better classroom performance relative to low dose of medication (A1= −1), for children who did not receive medication prior to the first stage of the intervention (O11 = 0). However, there is inconclusive evidence with respect to the difference between the two first-stage intervention options for children who received medication at school prior to the first stage of the intervention. Figure 8 presents the predicted means for each of the first-stage intervention options (A1), given whether or not the child received medication at school prior to the first stage of the intervention (O11).

Figure 8.

Figure 8

Predicted estimated quality of the second-stage intervention for each of the first-stage intervention options (A1), given whether or not the child received medication at school prior to first-stage intervention (O11).

 Overall, the optimal sequence of decision rules based on this data analysis is as follows:
 IF the child received medication prior to the first stage of the intervention
  THEN offer low dose of medication or low-intensity behavioral intervention.
 ELSE IF the child did not receive medication prior to the first stage of the intervention
   THEN offer low-intensity behavioral intervention.
Then,
 IF the child shows inadequate response to the first stage of the intervention
  THEN IF child’s adherence to first stage of the intervention is low,
    THEN augment the first-stage intervention option with the
    other type of intervention.
   ELSE IF child’s adherence to the first stage of the intervention is high
    THEN augment the first-stage intervention option with the
    other type of intervention or intensify the first-stage
    intervention option.
 ELSE IF the child show adequate response to the first stage of the intervention,
  THEN continue first-stage intervention.

Discussion

In the current study, we introduced Q-learning: a novel regression-based data analysis method for constructing high-quality decision rules. We discussed how Q-learning can be used to investigate the possibility of more deeply tailored adaptive interventions than those embedded in the SMART study. We provided a general framework for Q-learning and also demonstrated how this framework can be generalized for the analysis of data from four common types of SMART designs. We then discussed three advantages of the Q-learning approach over a single-regression-based approach. First, Q-learning appropriately controls for the optimal second-stage intervention option when assessing the effect of the first-stage intervention. Second, the effects estimated by Q-learning incorporate both the direct and indirect effects of the first-stage intervention options, the combination of which is necessary for making intervention decision rules. Third, Q-learning reduces potential bias resulting from unmeasured causes of both the tailoring variables and the primary outcome. Finally, we illustrated the application of Q-learning using a simplified version of the Adaptive Interventions for Children with ADHD study, with the general aim to guide researchers who wish to apply this method to construct high-quality adaptive interventions.

Q-learning can be used to estimate the optimal sequence of decision rules in a straightforward and intuitive manner. Although in the current study we focused on only two intervention stages, and used dichotomous tailoring variables with only two values, Q-learning can be used for studies with more than two stages, and can be easily extended to continuous as well as categorical tailoring variables. Additionally, we used effect coding (1, -1) to denote the randomized intervention options at each stage. However, dummy coding (0,1) can also be used. The regression approach presented here can be generalized via a generalized linear model in cases of binary (more generally, categorical) outcomes. An R package for using Q-learning with data from a two-stage SMART design is available at http://methodology.psu.edu/ra/adap-treat-strat/qlearning.

Still, applying this method involves several challenges. First, when the data are from observational studies, direct implementation of this analysis might give biased results due to unmeasured confounding factors that predict the probability of being offered intervention options A1 or A2, given past intervention history. This reflects a selection bias caused by non-random treatment. For example, in the context of the ADHD example, assume that children’s family functioning affected whether they would be initially offered medication or behavioral intervention. In such cases, Q-learning should be implemented in combination with methodologies that adjust for confounding (see Robins, 1999).

Second, inferential challenges caused by non-differentiability should be taken into consideration when applying Q-learning. As noted previously, in Q-learning, non-differentiability arises because the formula for Y~ (i.e., the dependent variable for the first-stage regression) contains an absolute value function. Because the absolute value function is non-differentiable at the point 0, the distribution of the estimators of the regression coefficients in Q1 cannot be consistently approximated by standard methods such as the bootstrap. In the current analysis, we used the soft-threshold operation recommended by Chakraborty et al. (2010). Although the efficiency of this approach in reducing the bias of the intervention effects was documented (see Chakraborty et al., 2010), improved inferential methods are possible and are currently under development.

Third, in the current analysis we considered only two candidate tailoring variables. However, studies often collect information on a large set of covariates (e.g., multiple surveys of mental health status and functioning) from which a smaller subset of variables must be selected for any practical implementation of adaptive interventions. Accordingly, researchers might be interested in using the data to select a subset of tailoring variables that depicts the estimated optimal sequence of decision rules as closely as possible to the optimal rule which uses all variables. Biernot and Moodie (2010) discuss methods for selecting tailoring variables in randomized settings, comparing two selection methods (reducts-- a variable selection tool from computer sciences, and the S-score criterion proposed by Gunter Zhu, & Murphy, 2011). Still, additional research effort should be directed towards developing and exploring methods for the selection of tailoring variables.

Finally, because our data analysis was illustrative in nature, we handled missing data (resulting from dropout, unavailability during data collection period, or unwillingness/inability of teachers to respond), by using listwise deletion, ignoring subjects with incomplete information. Because this approach can have serious drawbacks (see, e.g.,Qin, Zhang & Leung, 2009; Little & Rubin, 1987; Schafer & Olsen, 1998), we recommend that researchers consider applying modern missing data techniques, such as multiple imputation (MI; Rubin, 1987), which allow more efficient estimation from incomplete data (see Shortreed, Laber, Pineau, & Murphy, 2010 for recent research on using MI to adjust for missing data in SMART studies).

Despite these challenges, our research demonstrates that the construction of the optimal sequence of decision rules from data can be achieved by a relatively simple regression-based procedure. In light of the growing interest in developing evidence-based individualized interventions in the behavioral sciences, the current research is part of an ongoing endeavor to advance methodological research relevant to adaptive interventions, hoping to further increase researchers’ awareness of the conceptual appeal and practical advantages of this approach.

Acknowledgments

The project described was supported by Award Number P50DA010075 from the National Institute on Drug Abuse. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institute on Drug Abuse or the National Institutes of Health. The authors would like to thank Ayala Cohen for providing valuable feedback on earlier versions of this manuscript.

Footnotes

1

Throughout we use uppercase letters to represent a random variable, and lowercase letters to represent a particular value, or realization, of that random variable.

2

Expert systems (or knowledge-based systems) are defined broadly as computer programs that mimic the reasoning and problem solving of a human ‘expert’. These systems use pre-specified knowledge about the particular problem area. They are based on theoretical models, employing deep knowledge systems as a basis for their operation (Velicer, James Prochaska & Redding, 2006).

3

This is a standard assumption for consistency in generalized linear models such as logistic regression, survival analysis and ordinal regression.

4

In small samples the t-test statistic can be used if the usual assumptions underpinning classical linear regression hold (normal residuals, homogeneous variance, etc.).

5

The indicator for response/non-response is part of the vector of intermediate outcomes measured prior to the second-stage of the intervention (O2), however we use the notation R, instead of the notation O23 for clarity.

6

We ensured that model (5) provides a correct description of the data given the formula we used for Y and O2 above.

7

In a full analysis one would want to use a modern missing data method to avoid bias.

8

Such relatively high adherence rates may result from obtaining adherence data only for the first 8 weeks of the school year. Moreover, study medication was to be taken only on school days, and was dispensed monthly.

Contributor Information

Inbal Nahum-Shani, Institute for Social Research, University of Michigan, Ann Arbor, MI.

Min Qian, Department of Biostatistics, Columbia University, New York, NY.

Daniel Almirall, Institute for Social Research, University of Michigan, Ann Arbor, MI.

Susan Murphy, Department of Statistics and Institute for Social Research, University of Michigan, Ann Arbor, MI.

References

  1. Baron RM, Kenny DA. The moderator–mediator variable distinction in social psychological research: Conceptual, strategic, and statistical considerations. Journal of Personality and Social Psychology. 1986;51:1173–1182. doi: 10.1037//0022-3514.51.6.1173. [DOI] [PubMed] [Google Scholar]
  2. Bellman RE, Dreyfus SE. Applied dynamic programming. Princeton University Press; Princeton, NJ: 1962. [Google Scholar]
  3. Bierman KL, Nix RL, Maples JJ, Murphy SA. Examining clinical judgment in an adaptive intervention design: The Fast Track Program. Journal of Consulting and Clinical Psychology. 2006;74:468–481. doi: 10.1037/0022-006X.74.3.468. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Biernot P, Moodie EM. A comparison of variable selection approaches for dynamic treatment regimes. The International Journal of Biostatistics. 2010;6(1) doi: 10.2202/1557-4679.1178. Article 6. [DOI] [PubMed] [Google Scholar]
  5. Brown CH, Ten Have TR, Jo B, Dagne G, Wyman PA, Muthen B, Gibbons RD. Adaptive designs for randomized trials in public health. Annual Review of Public Health. 2009;30:1–25. doi: 10.1146/annurev.publhealth.031308.100223. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Chakraborty B, Murphy SA, Strecher V. Inference for non-regular parameters in optimal dynamic treatment regimes. Statistical Methods in Medical Research. 2010;19(3):317–343. doi: 10.1177/0962280209105013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Chesney A, Locke E. An examination of the relationship among goal difficulty, business strategies, and performance on a complex management simulation task. Academy of Management Journal. 1991;34:400–424. [Google Scholar]
  8. Collins LM, Murphy SA, Strecher V. The multiphase optimization strategy (MOST) and the sequential multiple assignment randomized trial (SMART) new methods for more potent ehealth interventions. American Journal of Preventive Medicine. 2007;32(5):S112–S118. doi: 10.1016/j.amepre.2007.01.022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Connell AM, Dishion TJ, Yasui M, Kavanagh K. An adaptive approach to family intervention: Linking engagement in family-centered intervention to reductions in adolescent problem behavior. Journal of Consulting and Clinical Psychology. 2007;75(4):568–579. doi: 10.1037/0022-006X.75.4.568. [DOI] [PubMed] [Google Scholar]
  10. Donovan JJ, Radosevich DJ. The moderating role of goal commitment on the goal difficulty – performance relationship: A metaanalytic review and critical reanalysis. Journal of Applied Psychology. 1998;83:308–315. [Google Scholar]
  11. Erez M. Performance quality and work motivation. In: Kleinbeck U, Thierry H, Haecker H, Quast H, editors. Work motivation. Erlbaum; Hillsdale, NJ: 1990. pp. 53–65. [Google Scholar]
  12. Fabiano GA, Pelham WE, Waschbusch DA, Gnagy EM, Lahey BB, Chronis AM, Burrows-MacLean L. A practical measure of impairment: Psychometric properties of the impairment rating scale in samples of children with attention deficit hyperactivity disorder and two school-based samples. Journal of Clinical Child and Adolescent Psychology. 2006;35:369–385. doi: 10.1207/s15374424jccp3503_3. [DOI] [PubMed] [Google Scholar]
  13. Fried Y, Slowik LH. Enriching goal-setting theory with time: An integrated approach. Academy of Management Review. 2004;29:404–422. [Google Scholar]
  14. Gunter L, Zhu J, Murphy SA. Variable selection for qualitative interactions. Statistical Methodology. 2011;8:42–55. doi: 10.1016/j.stamet.2009.05.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Little RJA, Rubin DB. Statistical Analysis with Missing Data. Wiley; New York, NY: 1987. [Google Scholar]
  16. MacKinnon DP, Warsi G, Dwyer JH. A simulation study of mediated effect measures. Multivariate Behavioral Research. 1995;30:41–62. doi: 10.1207/s15327906mbr3001_3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Marlowe DB, Festinger DS, Arabia PL, Dugosh KL, Benasutti KM, Croft JR, McKay JR. Adaptive interventions in drug court: A pilot experiment. Criminal Justice Review. 2008;33:343–360. doi: 10.1177/0734016808320325. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Murphy SA. An experimental design for the development of adaptive treatment strategies. Statistics in Medicine. 2005;24:455–1481. doi: 10.1002/sim.2022. [DOI] [PubMed] [Google Scholar]
  19. Nahum-Shani I, Qian M, Pelham WE, Gnagy B, Fabiano G, Waxmonsky J, Murphy SA. Experimental design and primary data Analysis methods for comparing adaptive interventions. 2011. Manuscript submitted for publication. [DOI] [PMC free article] [PubMed]
  20. Pelham WE, Evans SW, Gnagy EM, Greenslade KE. Teacher ratings of DSM-III-R symptoms for the disruptive behavior disorders: Prevalence, factor analyses, and conditional probabilities in a special education sample. School Psychology Review. 1992;21:285–299. [Google Scholar]
  21. Pelham WE, Fabiano GA. Evidence-based psychosocial treatment for attentiondeficit/hyperactivity disorder. Journal of Clinical Child and Adolescent Psychology. 2008;37:184–214. doi: 10.1080/15374410701818681. [DOI] [PubMed] [Google Scholar]
  22. Pelham WE, Gnagy EM. Psychosocial and combined treatments for ADHD. Mental Retardation and Developmental Disabilities Research Reviews. 1999;5:225–236. [Google Scholar]
  23. Pelham WE, Gnagy EM, Greiner AR, Hoza B, Hinshaw SP, Swanson JM, McBurnett K. Behavioral vs. behavioral and pharmacological treatment in ADHD children attending a summer treatment program. Journal of Abnormal Child Psychology. 2000;28:507–525. doi: 10.1023/a:1005127030251. [DOI] [PubMed] [Google Scholar]
  24. Pelham WE, Hoza B, Pillow DR, Gnagy EM, Kipp HL, Greiner AR, Fitzpatrick E. Effects of methylphenidate and expectancy on children with ADHD: Behavior, academic performance, and attributions in a summer treatment program and regular classroom setting. Journal of Consulting and Clinical Psychology. 2002;70:320–335. [PubMed] [Google Scholar]
  25. Pliszka S. Practice parameter for the assessment and treatment of children and adolescents with attention-deficit/hyperactivity disorder. Journal of the American Academy of Child & Adolescent Psychiatry. 2007;46:894–921. doi: 10.1097/chi.0b013e318054e724. [DOI] [PubMed] [Google Scholar]
  26. Qin J, Zhang B, Leung DHY. Empirical likelihood in missing data problems. Journal of the American Statistical Association. 2009;104(488):1492–1503. [Google Scholar]
  27. Rivera DE, Pew MD, Collins LM. Using engineering control principles to inform the design of adaptive interventions: A conceptual introduction. Drug and Alcohol Dependence. 2007;88:S31–S40. doi: 10.1016/j.drugalcdep.2006.10.020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Robins JM. Association, causation, and marginal structural models. Synthese. 1999;121:151–179. [Google Scholar]
  29. Robins JM. Optimal structural nested models for optimal sequential decisions. In: Lin DY, Heagerty P, editors. Proceedings of the Second Seattle Symposium on Biostatistics. Springer; New York: 2004. pp. 189–326. [Google Scholar]
  30. Rubin DB. Multiple imputation for nonresponse in surveys. Wiley & Sons; New York: 1987. [Google Scholar]
  31. Schafer JL, Olsen MK. Multiple imputation for multivariate missing-data problems: A data analyst’s perspective. Multivariate Behavioral Research. 1998;33:545–571. doi: 10.1207/s15327906mbr3304_5. [DOI] [PubMed] [Google Scholar]
  32. Schaughency E, Ervin R. Building capacity to implement and sustain effective practices to better serve children. School Psychology Review. 2006;35(2):155–166. [Google Scholar]
  33. Shortreed SM, Laber E, Pineau J, Murphy SA. Imputations methods for the clinical antipsychotic trials of intervention and effectiveness study. School of Computer Science, McGill University; Montreal, Canada: 2010. (Technical Report SOCS-TR-2010.8) [Google Scholar]
  34. Sutton RS, Barto AG. Reinforcement Learning: An Introduction. MIT Press; Cambridge, Mass: 1998. [Google Scholar]
  35. Velicer WF, Prochaska JO, Redding CA. Tailored communications for smoking cessation: Past successes and future directions. Drug and Alcohol Review. 2006;25:47–55. doi: 10.1080/09595230500459511. [DOI] [PubMed] [Google Scholar]
  36. Watkins CJCH. Doctoral Thesis. University of Cambridge; England: 1989. Learning from delayed rewards. [Google Scholar]

RESOURCES