Abstract
Contextual bandit algorithms are increasingly replacing non-adaptive A/B tests in e-commerce, healthcare, and policymaking because they can both improve outcomes for study participants and increase the chance of identifying good or even best policies. To support credible inference on novel interventions at the end of the study, nonetheless, we still want to construct valid confidence intervals on average treatment effects, subgroup effects, or value of new policies. The adaptive nature of the data collected by contextual bandit algorithms, however, makes this difficult: standard estimators are no longer asymptotically normally distributed and classic confidence intervals fail to provide correct coverage. While this has been addressed in non-contextual settings by using stabilized estimators, the contextual setting poses unique challenges that we tackle for the first time in this paper. We propose the Contextual Adaptive Doubly Robust (CADR) estimator, the first estimator for policy value that is asymptotically normal under contextual adaptive data collection. The main technical challenge in constructing CADR is designing adaptive and consistent conditional standard deviation estimators for stabilization. Extensive numerical experiments using 57 OpenML datasets demonstrate that confidence intervals based on CADR uniquely provide correct coverage.
1. Introduction
Contextual bandits, where personalized decisions are made sequentially and simultaneously with data collection, are increasingly used to address important decision-making problems where data is limited and/or expensive to collect, with applications in product recommendation [Li et al., 2010], revenue management [Kallus and Udell, 2020, Qiang and Bayati, 2016], and personalized medicine [Tewari and Murphy, 2017]. Adaptive experiments, whether based on bandit algorithms or Bayesian optimization, are increasingly being considered in place of classic randomized trials in order to improve both the outcomes for study participants and the chance of identifying the best treatment allocations [Athey et al., 2018, Quinn et al., 2019, Kasy and Sautmann, 2021, Bakshy et al., 2018].
But, at the end of the study, we still want to construct valid confidence intervals on average treatment effects, subgroup effects, or the value of new personalized interventions. Such confidence intervals are, for example, crucial for enabling credible inference on the presence or absence of improvement of novel policies. However, due to the adaptive nature of the data collection, unlike classic randomized trials, standard estimates and their confidence intervals actually fail to provide correct coverage, that is, contain the true parameter with the desired confidence probability (e.g., 95%). A variety of recent work has recognized this and offered remedies [Hadad et al., 2019, Luedtke and van der Laan, 2016], but only for the case of non-contextual adaptive data collection. Like classic confidence intervals, when data comes from a contextual bandit – or any other context-dependent adaptive data collection – these intervals also fail to provide correct coverage. In this paper, we propose the first asymptotically normal estimator for the value of a (possibly contextual) policy from context-dependent adaptively collected data. This asymptotic normality leads directly to the construction of valid confidence intervals.
Our estimator takes the form of a stabilized doubly robust estimator, that is, a weighted time average of an estimate of the so-called canonical gradient using plug in estimators for the outcome model, where each time point is inversely weighted by its estimated conditional standard deviation given the past. We term this the Contextual Adaptive Doubly Robust (CADR) estimator. We show that, given consistent conditional variance estimates which at each time point only depend on previous data, the CADR estimator is asymptotically normal, and as a result we can easily construct asymptotically valid confidence intervals. This normality is in fact robust to misspecifying the outcome model. A significant technical challenge is actually constructing such variance estimators. We resolve this using an adaptive variance estimator based on the importance-sampling ratio of current to past (adaptive) policies at each time point. We also show that we can reliably estimate outcome models from the adaptively-collected data so that we can plug them in. Extensive experiments using 57 OpenML datasets demonstrate the failure of previous approaches and the success of ours at constructing confidence intervals with correct coverage.
1.1. Problem Statement and Notation
The data.
Our data consists of a sequence of observations indexed t = 1, …, T comprising of context , action , and outcome generated by an adaptive experiment, such as a contextual bandit algorithm. Roughly, at each round t = 1, 2, …, T, an agent formed a contextual policy gt(a | x) based on all past observations, then observed an independently drawn context vector X(t) ~ Q0,X, carried out an action A(t) drawn from its current policy gt(· | X(t)), and observed an outcome Y (t) ~ Q0,Y (· | A(t), X(t)) depending only on the present context and action. The action and context measurable spaces , are arbitrary, e.g., finite or continuous.
More formally, we let O(t) ≔ (X(t), A(t), Y (t)) and make the following assumptions about the sequence O(1), …, O(T) comprising our dataset. First, we assume X(t) is independent of all else given A(t) and has a time-independent marginal distribution that we denote by Q0,X. Second, we assume A(t) is independent of all else given O(1), …, O(t − 1), X(t) and we set gt(· | X(t)) to its (random) conditional distribution given O(1), …, O(t − 1), X(t). Third, we assume Y (t) is independent of all else given X(t), A(t) and has a time-independent conditional distribution given X(t) = x, A(t) = a that is denoted by Q0,Y (· | A, X). The distributions Q0,X and Q0,Y are unknown, while the policies gt(a | x) are known, as would be the case when running an adaptive experiment. To simplify presentation we endow with a base measure (e.g., counting for finite actions or Lebesgue for continuous actions) and identify policies gt with conditional densities with respect to (w.r.t.) . In the case of K < ∞ actions, policies are maps from to the K-simplex.
Note that, as the agent updates its policy based on already collected observations, gt is a random O(1), …, O(t − 1)-measurable object. This is the major departure from the setting considered in other literature on off-policy evaluation, which only consider a fixed logging policy, gt = g, that is independent of the data. See Section 1.2.
The target parameter.
We are interested in inference on a generalized average causal effect expressed as a functional of the unknown distributions above, Ψ0 = Ψ(Q0,X, Q0,Y), where for any distributions QX, QY, we define
where is a given fixed, bounded function. Two examples are: (a) when g* is a policy (conditional density), then Ψ0 is its value; (b) when g* is the difference between two policies then Ψ0 is the difference between their values. A prominent example of the latter is when and g*(a | x) = a, which is known as the average treatment effect. If we include an indicator for x being in some set, then we get the subgroup effect.
Defining the conditional mean outcome,
we note that the target parameter only depends on Q0,Y via , so we also overload notation and write for any function . Note that when and is the counting measure, the integral over a is a simple sum.
Canonical gradient.
We will make repeated use of the following function: for any conditional density (a, x) ↦ g(a | x), any probability distribution QX over the context space , and any function , we define the function by
Further, define , which coincides with the so-called canonical gradient of the target parameter Ψ w.r.t. the usual nonparametric statistical model comprising all joint distributions over [van der Vaart, 2000, van der Laan and Robins, 2003].
Integration operator notation.
For any policy g and distributions QX, QY, denote by PQ,g the induced distribution on . For any function , we use the integration operator notation
that is, the expectation w.r.t. PQ,g alone. Then, for example, for any O(1), …, O(s − 1)-measurable random function , we have that .
1.2. Related Literature and Challenges for Post-Contextual-Bandit Inference
Off-policy evaluation.
In non-adaptive settings, where gt = g is fixed and does not depend on previous observations, common off-the shelf estimators for the mean outcome under g* include the Inverse Propensity Scoring (IPS) estimator [Beygelzimer and Langford, 2009, Li et al., 2011] and and the Doubly Robust (DR) estimator [Dudík et al., 2011, Robins et al., 1994]:
where is an estimator of the outcome model . If we use cross-fitting to estimate [Chernozhukov et al., 2018], then both the IPS and DR estimators are unbiased and asymptotically normal, permitting straightforward inference using Wald confidence intervals (i.e., ±1.96 of the estimated standard error). There also exist many variants of the IPS and DR estimators that, rather than plugging in the importance sampling (IS) ratios (g*/gt)(A(t) | X(t)) and/or outcome-model estimators, instead choose them directly with the aim to minimize error [e.g. Kallus, 2018, Farajtabar et al., 2018, Thomas and Brunskill, 2016, Wang et al., 2017, Kallus and Uehara, 2019b].
Inference challenges in adaptive settings.
In the adaptive setting, it is easy to see that, if in the tth term for DR we use an outcome model fit using only the observations O(1), …, O(t − 1), then both the IPS and DR estimators both remain unbiased. However, neither generally converges to a normal distribution. One key difference between the non-adaptive and adaptive settings is that the IS ratios (g*/gt)(A(t) | X(t)) can both diverge to infinity or converge to zero. As a result of this, the above two estimators may either be dominated by their first terms or their last terms. At a more theoretical level, this violates the classical condition of martingale central limit theorems that the conditional variance of the terms given previous observations stabilizes asymptotically.
Stabilized DR estimators in non-contextual settings.
The issue for inference due to instability of the DR estimator terms was recognized by Luedtke and van der Laan [2016] in another setting. They work in the non-adaptive setting but consider the problem of inferring the maximum mean outcome over all policies when the optimal policy is non-unique. Their proposal is a so-called stabilized estimator, in which each term is inversely weighted by an estimate of its conditional standard deviation given the previous terms. This stabilization trick has been also been reused for off-policy inference from non-contextual bandit data by Hadad et al. [2019], as the stabilized estimator remains asymptotically normal, permitting inference. In their non-contextual setting, an estimate of the conditional standard deviation of the terms can easily be obtained by the inverse square root propensities. In contrast, in our contextual setting, obtaining valid stabilization weights is more challenging and requires a construction involving adaptive training on past data.
1.3. Contributions
In this paper, we construct and analyze a stabilized estimator for policy evaluation from context-dependent adaptively collected data, such as the result of running a contextual bandit algorithm. This then immediately enables inference. After constructing a generic extension of the stabilization trick, the main technical challenge is to construct a sequence of estimators of the conditional standard deviations that are both consistent and such that for each t, only uses the previous data points O(1), …, O(t − 1). We show in extensive experiments across a large set of contextual bandit environments that our confidence intervals uniquely achieve close to nominal coverage.
2. Construction and Analysis of the Generic Contextual Stabilized Estimator
In this section, we give a generic construction of a stabilized estimator in our contextual and adaptive setting. That is, given generic plug-ins for outcome model and conditional standard deviation. We then provide conditions under which the estimator is asymptotically normal, as desired. To develop CADR, we will then proceed to construct appropriate plug in estimators in the proceeding sections.
2.1. Construction of the Estimator
Outcome and variance estimators.
Our estimator uses a sequence of estimators of the outcome model , such that, for every t, is O(1), …, O(t)-measurable, that is, is trained using only the data up to time t. A key part of our estimator are the conditional variance estimators.
Additionally, we require estimates of the conditional standard deviation of the canonical gradient. Define
where .
Let be a given sequence of estimates of σ0,t such that is O(1), …, O(t − 1)-measurable, that is, is estimated using only the data up to time t.
The generic form of the estimator.
The generic contextual stabilized estimator is then defined as:
| (1) |
2.2. Asymptotic normality guarantees
We next characterize the asymptotic distribution of under some assumptions.
Assumption 1 (Non degenerate efficiency bound). .
Assumption 1 states that there is no fixed logging policy g such that the efficiency bound for estimation of in the nonparametric model, from i.i.d. draws of , is zero. If assumption 1 does not hold, there exists a logging policy g such that, if , then (g*(A | X)/g(A | X))Y equals with probability 1. In other words, if assumption 1 does not hold, there exists a logging policy g such that can be estimated with no error with probability 1 from a single draw of . Thus, it is very lax. An easy sufficient condition for Assumption 1 is that the outcome model has nontrivial variance in that .

Assumption 2 (Consistent standard deviation estimators.). almost surely.
In the next section we will proceed to construct specific estimators that satisfy Assumption 2, leading to our proposed CADR estimator and confidence intervals.
Assumption 3 (Exploration rate). For any t ≥ 1, we have that almost surely.
Here, at ≳ bt means that for some constant c > 0, we have at ≥ cbt for all t ≥ 1. Assumption 3 requires that the exploration rate of the adaptive experiment does not decay too quickly.
Based on these assumptions, we have the following asymptotic normality result:
Theorem 1. Denote . Under Assumptions 1 to 3, it holds that .
Remark 1. Theorem 1 does not require the outcome model estimator to converge at all. As we will see in Section 3, our conditional variance estimator does require that the outcome model converges to a fixed limit , but this limit does not have to be the true outcome model . In other words, consistency of the outcome model is not required at any point of our analysis.
3. Construction of the Conditional Variance Estimator and CADR
We now tackle the construction of satisfying our assumptions; namely, they must be adaptively trained only on past data at each t and they must be consistent. Observe that , where we define
Designing an O(1), …, O(t − 1)-measurable estimator of presents several challenges. First, while we can only use observations O(1), …, O(t − 1) to estimate it, is defined as a function of integrals w.r.t. , from which we have only one observation, namely O(t). Second, our estimation target is random as it depends on gt and . Third, gt, depend on the same observations O(1), …, O(t − 1) that we have at our disposal to estimate .
Representation via importance sampling.
We can overcome the first difficulty via importance sampling, which allows us to write , i = 1, 2 as integrals w.r.t. , s = 1, …, t − 1, i.e., the conditional distributions of observations O(s), s = 1, …, t − 1 given their respective past. Namely, for any s ≥ 1, i = 1, 2, we have that
| (2) |
Dealing with the randomness of the estimation target.
We now turn to second challenge. Since can be written in terms of for i = 1, 2, Eq. (2) suggests perhaps an approach based on sample averages of over s. However, whenever s < t, the latter is an O(1), …, O(t − 1)-measurable function due to the dependence on gt and . Namely, does not coincide in general with the conditional expectation , as would arise from a sample average. We now look at solutions to overcome this difficulty, considering first and then gt.
Dealing with the randomness of .
We propose an estimator of for any fixed g. While requiring that converges to the true outcome regression function is a strong requirement, most reasonable estimators will at least converge to some fixed limit . As a result, under an appropriate stochastic convergence condition on , can be reasonably approximated by the corresponding Cesaro averages, defined for i = 1, 2 as
These are easy to estimate from the corresponding sample averages, defined for i = 1, 2 as
since for each i = 1, 2, the difference is the average of a martingale difference sequence (MDS). We then define our estimator of as
| (3) |
From fixed g to random gt.
So far, we have proposed and justified the construction of as an estimator of for a fixed g. We now discuss conditions under which is valid estimator of . When g is fixed, for each i = 1, 2, the error decomposes as the sum of the MDS average and of the Cesaro approximation error . Both differences are straightforward to bound. For a random gt, the term is no longer an MDS average. Fortunately, under a complexity condition on the logging policy class , we can bound the supremum of the martingale empirical processes , which in turn gives us a bound on .
Consistency guarantee for .
Our formal consistency result relies on the following assumptions.
Assumption 4 (Outcome regression estimator convergence). There exists β > 0, and a fixed function such that almost surely.
The next assumption is a bound on the bracketing entropy (see, e.g., [van der Vaart and Wellner, 1996] for definition) of the logging policy class.
Assumption 5 (Complexity of the logging policy class). There exists a class of conditional densities such that almost surely, there exists G > 0 such that , and for some p > 0
where .
Next, we require a condition on the exploration rate that is stronger than Assumption 3.
Assumption 6 (Exploration rate (stronger)). For ant t ≥ 1, we have that almost surely, where α(β, p) ≔ min(1/((3 + p)), 1/(1 + 2p), β).
Theorem 2. Suppose that Assumptions 4 to 6 hold. Then, almost surely.
Remark 2. While we theoretically require the existence of a logging policy class with controlled complexity, we do not actually need to know to construct our estimator. Moreover, while we require a bound on the bracketing entropy of the logging policy class , we impose no restriction on the outcome regression model complexity, permitting us to use flexible black-box regression methods.
Remark 3. Assumption 4 requires to be a sequence of regression estimator, such that for every t ≥ 1, is fitted on O(1), …, O(t) and for which we can guarantee a rate of convergence to some fixed limit . Note that this can at first glance pose a challenge since observations O(1), …, O(t) are adaptively collected. In the appendix, we give guarantees for outcome regression estimation over a nonparametric model using an importance sampling weighted empirical risk minimization.
CADR asymptotics.
Our proposed CADR estimator is now given by plugging our estimates from Eq. (3) into Eq. (1), as summarized in Algorithm 1 As an immediate corollary of Theorems 1 and 2 we have our main guarantee for this final estimator, showing CADR is asymptotically normal, whence we immediately obtain asymptotically valid confidence intervals.
Corollary 1 (CADR Asymptotics and Inference). Suppose that Assumptions 1 and 4 to 6 hold. Let be given as in Eq. (3). Denote . Then,
Moreover, letting ζα denote the α-quantile of the standard normal distribution,
4. Empirical Evaluation
We next present computational results on public datasets that demonstrate the robustness of CADR confidence intervals using contextual bandit data with comparison to several baselines. Our experiments focus on the case of finitely-many actions, .
4.1. Baseline Estimators
We compare CADR to several benchmarks. All take the following form for a choice of wt, ωt, :
where .
The Direct Method (DM) sets wt = 1, ωt = 0 and fits by running some regression method for each a on the data {(X(s), Y (s)) : 1 ≤ s ≤ t − 1, A(s) = a}. We will use either linear regression or decision-tree regression, both using default sklearn parameters. Note that even in non-contextual settings, where is a simple per-arm sample average, may be biased due to adaptive data collection [Xu et al., 2013, Luedtke and van der Laan, 2016, Bowden and Trippa, 2017, Nie et al., 2018, Hadad et al., 2019, Shin et al., 2019]. Inverse Propensity Score Weighting (IPW) sets wt = 1, ωt = (g*/gt)(A(t) | X(t)), . Doubly Robust (DR) sets wt = 1, ωt = (g*/gt)(A(t) | X(t)) and fits as in DM. More Robust Doubly Robust (MRDR) [Farajtabar et al., 2018] is the same as DR but when fitting we reweight each data point by . None of the above are generally asymptotically normal under adaptive data collection [Hadad et al., 2019]. Adaptive Doubly Robust (ADR; a.k.a. stabilized onestep estimator for multi-armed bandit data) [Luedtke and van der Laan, 2016, Hadad et al., 2019] is the same as DR but sets . ADR is unbiased and asymptotically normal for multi-armed bandit logging policies but is biased for context-measurable adaptive logging policies, which is the focus of this paper. Finally, note that our proposal CADR takes the same form as DR but with using our adaptive conditional standard deviation estimators in Eq. (3).
4.2. Contextual Bandit Data from Multiclass Classification Data
To construct our data, we turn K-class classification tasks into a K-armed contextual bandit problems [Dudík et al., 2014, Dimakopoulou et al., 2017, Su et al., 2019], which has the benefits of reproducibility using public datasets and being able to make uncontroversial comparisons using actual ground truth data with counterfactuals. We use the public OpenML Curated Classification benchmarking suite 2018 (OpenML-CC18; BSD 3-Clause license) [Bischl et al., 2017], which has datasets that vary in domain, number of observations, number of classes and number of features. Among these, we select the classification datasets which have less than 100 features. This results in 57 classification datasets from OpenML-CC18 used for evaluation and Table 1 summarizes the characteristics of these datasets.
Table 1:
Characteristics of the 57 OpenML-CC18 datasets used for evaluation.
| Samples | Count | Classes | Count | Features | Count | ||
|---|---|---|---|---|---|---|---|
| < 1000 | 17 | = 2 | 31 | ≥ 2 and < 10 | 14 | ||
| ≥ 1000 and < 10000 | 30 | > 2 and < 10 | 17 | ≥ 10 and < 50 | 34 | ||
| ≥ 10000 | 10 | ≥ 10 | 9 | ≥ 50 and ≤ 100 | 9 |
Each dataset is a collection of pairs of covariates X and labels L ∈ {1, …, K}. We transform each dataset to the contextual bandit problem as follows. At each round, we draw X(t), L(t) uniformly at random with replacement from the dataset. We reveal the context X(t) to the agent, and given an arm pull A(t), we draw and return the reward . To generate our data, we set T = 10000 and use the following ϵ-greedy procedure. We pull arms uniformly at random until each arm has been pulled at least once. Then at each subsequent round t, we fit using the data up to that time in the same fashion as used for the DM estimator above using decision-tree regressions. We set and ϵt = 0.01 · t−1/3. We then let gt(a | x) = ϵt/K for and . That is, with probability ϵt we pull a random arm, and otherwise we pull .
We then consider four candidate policies to evaluate: (1) “arm 1 non-contextual”: g*(1 | x) = 1 and otherwise g*(a | x) = 0 (note that the meaning of label “1” changes by dataset), (2) “arm 2 non-contextual”: g*(2 | x) = 1 and otherwise g*(a | x) = 0, (3) “linear contextual”: we sample a new dataset of size T using a uniform exploration policy, then fit as above using linear regression, fix , and set g*(a* | x) = 1 and otherwise g*(a | x) = 0, (4) “tree contextual”: same as “linear contextual” but fit using decision-tree regression.
4.3. Results
Figure 1 shows the comparison of CADR estimator against DM, IPW, DR, ADR, and MRDR w.r.t. coverage, that is, the frequency over 64 replications of the 95% confidence interval covering the true Ψ0, for each of the 57 OpenML-CC18 datasets and 4 target policies. In each subfigure, each dot represents a dataset, the y-axis corresponds to the coverage of the CADR estimator and the x-axis corresponds to the coverage of one of the baseline estimators. The lines represent one standard error over the 64 replications. The dot is depicted in blue if for that dataset CADR has significantly better coverage than the baseline estimator, in red if it has significantly worse coverage, and in black if the difference in coverage of both estimators is within one standard error. In Fig. 1, outcome models for CADR, DM, DR, ADR, and MRDR are fit using linear regression (with default sklearn parameters). In the appendix, we provide additional empirical results where we use decision-tree regressions, or where we use the MRDR outcome model for CADR, or where we use cross-fold estimation across time.
Figure 1:

Comparison of CADR estimator against DM, IPW, DR, ADR and MRDR w.r.t. 95% confidence interval coverage on 57 OpenML-CC18 datasets and 4 target policies.
Across all of our experiments, we observe that the confidence interval of CADR has better coverage of the ground truth than any other baseline, which can be attributed to its asymptotic normality. The second best estimator in terms of coverage is DR. The advantages of CADR over DR are most pronounced when either (a) there is a mismatch between the logging policy and the target policy (e.g., compare the 1st and 2nd rows in Fig. 1; the tree target policy is most similar to the logging policy, which also uses trees) or (b) when the outcome model is bad (either due to model misspecification such as with a linear model on real data or due to small sample size).
5. Conclusions
Adaptive experiments hold great promise for better, more efficient, and even more ethical experiments. However, they complicate post-experiment inference, which is a cornerstone of drawing credible conclusions from controlled experiments. We provided here the first asymptotically normal estimator for policy value and causal effects when data were generated from a contextual adaptive experiment, such as a contextual bandit algorithm. This led to simple and effective confidence intervals given by adding and subtracting multiples of the standard error, making contextual adaptive experiments a more viable option for experimentation in practice.
6. Societal Impact and Limitations
Adaptive experiments hold particular promise in settings where experimentation is costly and/or dangerous, such as in medicine and policymaking. By adapting treatment allocation, harmful interventions can be avoided, outcomes for study participants improved, and smaller studies enabled. Being able to draw credible conclusions from such experiments make them viable replacements for classic randomized trials. Our confidence intervals offer one way to do so. At the same time, and especially subject to our assumption of vanishing but nonzero exploration, these experiments must be subject to the same ethical guidelines as classic randomized experiments. Additionally, the usual caveats of frequentist confidence intervals hold here, such as its interpretation only as a guarantee over data collection, this guarantee only being approximate in finite samples when we rely on asymptotic normality, and the risks of multiple comparisons and of p-hacking. Finally, we note that our inference focused on an average quantity, as such it focuses on social welfare and need not capture the risk to individuals or groups. Subgroup analyses may therefore be helpful in complementing the analysis; these can be conducted by setting g*(a | x) to zero for some x’s. Future work may be necessary to further extend our results to conducting inference on risk metrics such as quantiles of outcomes.
Supplementary Material
Contributor Information
Aurélien Bibaut, Netflix.
Antoine Chambaz, Université Paris Descartes.
Maria Dimakopoulou, Netflix.
Nathan Kallus, Cornell University and Netflix.
Mark van der Laan, University of California, Berkeley.
References
- Athey Susan, Baird Sarah, Jamison Julian, McIntosh Craig, Özler Berk, and Sama Dohbit. A sequential and adaptive experiment to increase the uptake of long-acting reversible contraceptives in cameroon, 2018. URL http://pubdocs.worldbank.org/en/606341582906195532/Study-Protocol-Adaptive-experiment-on-FP-counseling-and-uptake-of-MCs.pdf. Study protocol.
- Bakshy Eytan, Dworkin Lili, Karrer Brian, Kashin Konstantin, Letham Benjamin, Murthy Ashwin, and Singh Shaun. Ae: A domain-agnostic platform for adaptive experimentation. In Workshop on System for ML, 2018. [Google Scholar]
- Beygelzimer Alina and Langford John. The offset tree for learning with partial labels. In Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 129–138, 2009. [Google Scholar]
- Bibaut Aurelien, Dimakopoulou Maria, Chambaz Antoine, Kallus Nathan, and van der Laan Mark. Risk minimization from adaptively collected data: Guarantees for supervised and policy learning. 2021. [PMC free article] [PubMed]
- Bischl Bernd, Casalicchio Giuseppe, Feurer Matthias, Hutter Frank, Lang Michel, Mantovani Rafael G, van Rijn Jan N, and Vanschoren Joaquin. Openml benchmarking suites. arXiv preprint arXiv:1708.03731, 2017. [Google Scholar]
- Bowden Jack and Trippa Lorenzo. Unbiased estimation for response adaptive clinical trials. Statistical methods in medical research, 26(5):2376–2388, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chernozhukov Victor, Chetverikov Denis, Demirer Mert, Duflo Esther, Hansen Christian, Newey Whitney, and Robins James. Double/debiased machine learning for treatment and structural parameters. The Econometrics Journal, 21(1):C1–C68, 2018. [Google Scholar]
- Dimakopoulou Maria, Zhou Zhengyuan, Athey Susan, and Imbens Guido. Estimation considerations in contextual bandits. arXiv preprint arXiv:1711.07077, 2017. [Google Scholar]
- Dudík Miroslav, Langford John, and Li Lihong. Doubly robust policy evaluation and learning. In Proceedings of the 28th International Conference on International Conference on Machine Learning, pages 1097–1104, 2011. [Google Scholar]
- Dudík Miroslav, Erhan Dumitru, Langford John, Li Lihong, et al. Doubly robust policy evaluation and optimization. Statistical Science, 29(4):485–511, 2014. [Google Scholar]
- Farajtabar Mehrdad, Chow Yinlam, and Ghavamzadeh Mohammad. More robust doubly robust off-policy evaluation. In International Conference on Machine Learning, pages 1447–1456. PMLR, 2018. [Google Scholar]
- Hadad Vitor, Hirshberg David A, Zhan Ruohan, Wager Stefan, and Athey Susan. Confidence intervals for policy evaluation in adaptive experiments. arXiv preprint arXiv:1911.02768, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kallus Nathan. Balanced policy evaluation and learning. In Advances in Neural Information Processing Systems, pages 8895–8906, 2018. [Google Scholar]
- Kallus Nathan and Udell Madeleine. Dynamic assortment personalization in high dimensions. Operations Research, 68(4):1020–1037, 2020. [Google Scholar]
- Kallus Nathan and Uehara Masatoshi. Efficiently breaking the curse of horizon in off-policy evaluation with double reinforcement learning. arXiv preprint arXiv:1909.05850, 2019a. [Google Scholar]
- Kallus Nathan and Uehara Masatoshi. Intrinsically efficient, stable, and bounded off-policy evaluation for reinforcement learning. Advances in neural information processing systems, 32, 2019b. [Google Scholar]
- Kasy Maximilian and Sautmann Anja. Adaptive treatment assignment in experiments for policy choice. Econometrica, 89(1):113–132, 2021. [Google Scholar]
- Li Lihong, Chu Wei, Langford John, and Schapire Robert E. A contextual-bandit approach to personalized news article recommendation. In Proceedings of the 19th international conference on World wide web, pages 661–670, 2010. [Google Scholar]
- Li Lihong, Chu Wei, Langford John, and Wang Xuanhui. Unbiased offline evaluation of contextual-bandit-based news article recommendation algorithms. In Proceedings of the fourth ACM international conference on Web search and data mining, pages 297–306, 2011. [Google Scholar]
- Luedtke Alexander R. and van der Laan Mark J.. Statistical inference for the mean outcome under a possibly non-unique optimal treatment strategy. The Annals of Statistics, 44(2):713–742, 2016. Doi: 10.1214/15-AOS1384. URL 10.1214/15-AOS1384. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nie Xinkun, Tian Xiaoying, Taylor Jonathan, and Zou James. Why adaptively collected data have negative bias and how to correct for it. In International Conference on Artificial Intelligence and Statistics, pages 1261–1269. PMLR, 2018. [Google Scholar]
- Qiang Sheng and Bayati Mohsen. Dynamic pricing with demand covariates. arXiv preprint arXiv:1604.07463, 2016. [Google Scholar]
- Quinn Simon, Teytelboym Alex, Kasy Maximilian, Gordon Grant, and Caria Stefano. A sequential and adaptive experiment to increase the uptake of long-acting reversible contraceptives in cameroon, 2019. URL https://www.socialscienceregistry.org/trials/3870. Study registration.
- Robins James M, Rotnitzky Andrea, and Zhao Lue Ping. Estimation of regression coefficients when some regressors are not always observed. Journal of the American statistical Association, 89 (427):846–866, 1994. [Google Scholar]
- Shin Jaehyeok, Ramdas Aaditya, and Rinaldo Alessandro. On the bias, risk and consistency of sample means in multi-armed bandits. arXiv preprint arXiv:1902.00746, 2019. [Google Scholar]
- Su Yi, Wang Lequn, Santacatterina Michele, and Joachims Thorsten. Cab: Continuous adaptive blending for policy evaluation and learning. In International Conference on Machine Learning, pages 6005–6014. PMLR, 2019. [Google Scholar]
- Tewari Ambuj and Murphy Susan A. From ads to interventions: Contextual bandits in mobile health. In Mobile Health, pages 495–517. Springer, 2017. [Google Scholar]
- Thomas Philip and Brunskill Emma. Data-efficient off-policy policy evaluation for reinforcement learning. In International Conference on Machine Learning, pages 2139–2148. PMLR, 2016. [Google Scholar]
- van der Laan Mark J and Robins James M. Unified methods for censored longitudinal data and causality. Springer Science & Business Media, 2003. [Google Scholar]
- van der Vaart A and Wellner J. Weak Convergence and Empirical Processes. Springer-Verlag; New York, 03 1996. ISBN 9781475725452. [Google Scholar]
- van der Vaart Aad W. Asymptotic statistics. Cambridge university press, 2000. [Google Scholar]
- van Handel R. On the minimal penalty for Markov order estimation. Probability Theory and Related Fields, 150:709–738, 2011. [Google Scholar]
- Wang Yu-Xiang, Agarwal Alekh, and Dudık Miroslav. Optimal and adaptive off-policy evaluation in contextual bandits. In International Conference on Machine Learning, pages 3589–3597. PMLR, 2017. [Google Scholar]
- Xu Min, Qin Tao, and Liu Tie-Yan. Estimation bias in multi-armed bandit algorithms for search advertising. Advances in Neural Information Processing Systems, 26:2400–2408, 2013. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
