Abstract
Inference for parameters associated with optimal dynamic treatment regimes is challenging as these estimators are nonregular when there are non-responders to treatments. In this discussion, we comment on three aspects of alleviating this nonregularity. We first discuss an alternative approach for smoothing the quality functions. We then discuss some further details on our existing work to identify non-responders through penalization. Third, we propose a clinically meaningful value assessment whose estimator does not suffer from nonregularity.
1. Introduction
The authors are to be congratulated for their excellent and thoughtful paper on statistical inference for dynamic treatment regimens. They have addressed several important and long-standing issues in this area. As discussed by the authors, nonsmoothness of the problem in some of the parameters of interest leads to estimators that are not smooth in the data. This in turn makes inference for these parameters challenging. In the following, we comment on a few additional strategies to alleviate the resulting nonregularity due to nonsmoothness. First, we discuss replacing the nonsmooth objective functions via a SoftMax Q-learning approach, which directly addresses the trade-off between bias and variance of the maximum operation in the local asymptotic framework. Proofs are given in the appendix.
Nonregularity of the estimators for the parameters associated with the optimal treatment regimes is mainly due to the existence of non-responders to treatments. Therefore, it would be useful and important if we could identify these non-responders. In the second part, we review our existing work on non-responder identification via penalization. We also discuss how this penalization can alleviate, although not solve, some regularity issues.
For the third and final aspect we wish to discuss, we note that in some public health settings, the parameters in the dynamic treatment regime are not as important as the value function which reflects the overall population impact of the estimated regime and is perhaps the most important quantity to focus on for public health policy. We propose a truncated value function which only focuses on those subjects who are expected to have large treatment effects. We claim that this alternative value function is clinically meaningful and does not suffer from nonregularity.
2. SoftMax Q-Learning
In this section we study the effect of replacing the max operator with a smoother version of it in the two-stage Q-learning algorithm discussed by Laber et al. We show that this smoothing can reduce the bias and can be controlled under local alternatives. The proposed SoftMax approach also sheds light on the bias/variance tradeoff which can be obtained by using over/under smoothing. In what follows, we briefly describe the SoftMax Q-learning algorithm, and then present some theoretical and simulation results.
2.1. Proposed Algorithm
Consider the Q-learning algorithm discussed by Laber et al. in Section 2. In step 2 of the algorithm, the stage outcome is predicted by
We propose replacing Ỹ with a SoftMax version of it. Define the SoftMax function by (see Fig 1)
Fig 1.

The function log{exp(x)+exp(0)} in blue and the function max(x, 0) in red. Note that the functions roughly agree for x ∉ [−3, 3].
Let
The estimator β̂1 of β1 is given by . We note that the algorithm discussed by Laber et al. is obtained as the limit, as α goes to infinity, of the SoftMax Q-learning algorithm discussed here.
2.2. Theory
In the following we briefly discuss the asymptotic properties of β̂1. We first discuss the limiting distribution of . We then discuss this limiting distribution under local alternatives. Finally, we discuss the asymptotic bias. The proofs appear in the Appendix.
Theorem 1
Assume (A1)–(A2) from Laber et al., and let αn → ∞ such that for a∞ ∈ [0, ∞). then
- If a∞ = 0,
-
If 0 < a∞ < ∞, thenwhere
For local alternatives the limiting distribution is given below.
Theorem 2
Assume (A1)–(A3) from Laber et al., and let αn → ∞ such that for a∞ ∈ (0; ∞). Then
where
The bound of the bias, scaled by root-n, under both standard and local alternatives asymptotics, is given below.
Corollary 1
Assume (A1)–(A2) from Laber et al., and let αn → ∞ such that for a∞ ∈ (0, ∞). Fix c ∈ ℝp21. Then
When (A3) from Laber et al. also holds, then
The above results show that by choosing the scale of α, the bias can be controlled. Theorem 2 shows that this control of the bias directly influences the variance, at least under local alternatives.
For inference, we need to discuss two different settings. When holding α fixed, as n goes to infinity, standard inference for the parameters is valid, as the problem becomes regular. However, this comes with the price that the bias does not vanish even asymptotically (see also the discussion in Section 4 ). As proved in Theorem 2, when taking α to infinity, as n goes to infinity, the problem is nonregular. Thus, adaptive confidence intervals, such as the one suggested by Laber et al., are needed in order to perform valid inference.
2.3. Simulations for SoftMax
We compare the small-sample behaviour of SoftMax to that of soft-threshholding using the example setting discussed in Laber et al., Section 3. Let . The max estimator is defined by
A soft-thresholdling estimator is defined by
Finally, the SoftMax estimator is defined by
Let Y |A ~ N(μa, 1), a = 0, 1, and assume that the treatment assignment is perfectly balanced. We use 1000 Monte Carlo replicates to estimate the bias for each parameter setting. Figure 2 below shows the bias as a function of the treatment effect and with tuning parameters σ ∈ [0, 5] and α ∈ [1, 6] for the soft-thresholding and SoftMax, respectively. It appears that the SoftMax does not suffer from large bias on points away from . Also, as expected from Theorem 1, the bias decreases as α increases.
Fig 2.

Left: Bias for soft-thresholding. Right: Bias for SoftMax. In both panels the bias is measured in units of for n = 10, as a function of effect size and of the tuning parameters, σ and α, for the soft-thresholding and SoftMax, respectively.
3. Penalized and Adaptive Q-Learning
In Penalized Q-learning (Song et al., 2011 ) and adaptive Q-learning (Goldberg et al., 2013 ), penalties were imposed on the term for each individual. This use of penalized estimation allows us to simultaneously estimate the second stage parameters and select individuals whose value functions are not affected by treatments, i.e., those individuals whose true values of are zero. Although the penalized method does not solve the non-regularity issue in estimating β’s, our numerical studies have demonstrated that penalized Q-learning is not only able to reduce bias, but also provides better coverage of confidence intervals in a number of scenarios, as compared to the hard thresholding method of Moodie and Richardson (2010) and some soft thresholding methods including resampling approaches. Furthermore, the inference approach for penalized methods described in Zhang and (Zhang 2014) appears to be able to handle diverging model perturbations. Finally, a nice feature of our penalized learning is that it enables us to identify non-responders, who may also have small treatment benefits even under a local alternative. Since it is clinically and practically most useful to target groups whose treatment benefit is large, identifying those subjects with small treatment benefits is useful for better allocation of resources and for reducing costs.
4. Truncated Value Function
The non-regularity issue arises primarily in settings where there are some subjects who do not respond to treatments at the second stage and where inference focuses on effect size. In the context of public health policy, we think that (1) the overall benefit (value) may be of greater interest compared to individual effect sizes and (2) those subjects who are not sensitive to treatments (approximate non-responders) should not have a large impact on the overall decision making process. Thus, we propose an appropriate alternative criterion, namely the ε-truncated value, for evaluating the optimal policy as follows:
where δ(X1) and δ(X2) denote the expected treatment effects at the first and second stages respectively. Here, ε is a small constant indicating a clinically meaningful effect size.
Under a SMART trial with randomization probabilities πk at stage k (k = 1, 2), this truncated value is equal to
Compared to the usual value function, we can see that Vε(d1, d2) differs by at most O(ε). Using the Q-learning model, the above value function for the estimated rule is
One advantage of considering this value function is that non-regularity will no longer be an issue since we have excluded the non-responders from the above statistic. One can easily show converges to the same normal distribution under local alternatives whether P(β2X2 = 0) > 0 or not.
5. Concluding Remarks
We again thank the authors for their very interesting work which likely stimulate additional future research on this crucial topic. It is clear that there are many fundamental and unresolved computational, methodological and theoretical challenges remaining which will benefit from many diverse problem solving approaches. We look forward to seeing this intriguing research area continue to develop.
Appendix A: Proofs
Sketch of proof of Theorem 1
Using the same arguments that lead to Eq. 2 in Laber et al., we have
where
is smooth and asymptotically normal and
Note that
where the last equality follows by taking derivatives, and where
| (1) |
For the remainder term, note that Lemma B.6 shows the consistency of β̂2,1, and that the expectation of the Hessian of is bounded by Assumption (A1). Hence, by applying Lemma B.5 to the matrix Σ̂1 that appears in the remainder term, we conclude that
| (2) |
where
Recall that by assumption, αn and thus fact that
we obtain that for a given h2,1,
Define the function w: Dp1 × l
∞(
) × ℝp21 × [0, 1] ↦ ℝp1 by w(Σ, μ, ν, a) = Σ−1μ(g(ν, B1, H2,1, a)), where
and where
= {g(ν, b1, h2,1, a), ||ν|| ≤ K}. Using the same arguments as those used in Lemma B.11, one can show that w is continuous at (Σ1,∞, P, ℝp21, 0). Thus, using the continuous mapping theorem, it can be shown that
where
We now discuss
, the third term in (3). By Lemma 1 (i) below, when
,
which proves (i).
For (ii), let δn → 0 such that αnδn → ∞. Write
Note that by Lemma 1 (vi),
Let , and note that p(0) = 0 and thus p(δ) → 0 as δ → 0. Hence,
Summarizing, we obtain that
which proves (ii).
Sketch of proof of Theorem 2
Using the same arguments that lead to Eq. 2 in Laber et al., we have
where
is smooth and asymptotically normal, and
Similarly to the proof of Theorem 1, we have
| (3) |
where
We obtain that for a given h2,1, and , and ,
Using the same arguments given in the proof of Theorem 4.1 (see also proof of Theorem 1 above) it can be shown that
+
converges in distribution to
where
Note that
The first three terms can be bounded as follows:
Thus the limiting distribution of
depends on that of Dn. For Dn we have
which concludes the proof.
Proof of Corollary 1
We prove only the second assertion, the first can be proved similarly. Noting that both
and
have mean zero, it is enough to bound the bias of
Using Lemma 1 (i) with and , we obtain that the bias is bounded by
The following lemma is needed for the proofs of Theorems 1–2:
Lemma 1
Let . Then
0 < f(α, x) ≤ log(2)/α.
argmaxx f (α, x) = 0 and f (α, 0) = log(2)/α.
f(α, x) → 0 as α → ∞.
Fix δ > 0. Then .
The proof is technical and therefore omitted.
Footnotes
The first author was funded in part by ISF grant 1308/12. The other authors were funded in part by grant P01 CA142538 from the National Cancer Institute. The second author was also funded in part by NSF-DMS 1309465.
Contributor Information
Yair Goldberg, Email: ygoldberg@stat.haifa.ac.il, Department of Statistics, University of Haifa, Mount Carmel, Haifa 31905, Israel.
Rui Song, Email: rsong@ncsu.edu, Department of Statistics, North Carolina State University, Raleigh, NC 27695, U.S.A.
Donglin Zeng, Email: dzeng@email.unc.edu, Department of Biostatistics, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, U.S.A.
Michael R. Kosorok, Email: kosorok@unc.edu, Department of Biostatistics, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, U.S.A
References
- Goldberg Y, Song R, Kosorok MR. From Probability to Statistics and Back: High-Dimensional Models and Processes – A Festschrift in Honor of Jon A. Wellner. Institute of Mathematical Statistics, vol. 9 of IMS Collections, chap. Adaptive Q-Learning. 2013:150–162. [Google Scholar]
- Moodie EEM, Richardson TS. Estimating optimal dynamic regimes: correcting bias under the null. Scandinavian Journal of Statistics. 2010;37:126–146. doi: 10.1111/j.1467-9469.2009.00661.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Song R, Wang W, Zeng D, Kosorok MR. Penalized Q-Learning for Dynamic Treatment Regimes. 2011 doi: 10.5705/ss.2012.364. Submitted. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang CH, Zhang SS. Confidence intervals for low dimensional parameters in high dimensional linear models. Journal of the Royal Statistical Society: Series B. 2014;76:217–242. [Google Scholar]
