Abstract
Estimation of causal effects in non-randomized studies comprises two distinct phases: design, without outcome data, and analysis of the outcome data according to a specified protocol. Recently, Gutman and Rubin (2013) proposed a new analysis-phase method for estimating treatment effects when the outcome is binary and there is only one covariate, which viewed causal effect estimation explicitly as a missing data problem. Here, we extend this method to situations with continuous outcomes and multiple covariates and compare it with other commonly used methods (such as matching, subclassification, weighting, and covariance adjustment). We show, using an extensive simulation, that of all methods considered, and in many of the experimental conditions examined, our new ‘multiple-imputation using two subclassification splines’ method appears to be the most efficient and has coverage levels that are closest to nominal. In addition, it can estimate finite population average causal effects as well as non-linear causal estimands. This type of analysis also allows the identification of subgroups of units for which the effect appears to be especially beneficial or harmful.
Keywords: causal inference, multiple imputation, propensity score, Rubin causal model, spline, subgroup analysis
1. Introduction
1.1. Overview
The goal of many studies in biomedical and social science is to estimate the causal effects of a binary treatment W on an outcome Y. Although randomized trials are the gold standard for inferring causal effects, in many settings, they may be impractical because of financial, logistical, or ethical considerations, and researchers are forced to use non-randomized observational studies, where it is difficult to attribute differences in outcomes to differences in treatments received because baseline characteristics that influence treatment decisions also influence the outcomes.
To overcome this issue, observational studies first should be designed to approximate randomized experiments as closely as possible and second should be analyzed using a predefined estimation procedure [1]. These two phases are commonly referred to as the design phase and the analysis phase. The design phase includes contemplating, collecting, organizing, and analyzing data without seeing any outcome data [1,2]. In observational studies, methods such as subclassification [3], matching on covariates or functions of them [4–7], and calculation of weights [8, 9] have been proposed to help approximate hypothetical randomized experiments and thus are part of the design phase. These methods do not rely on any outcome data, and Rubin [1] advocated that they should be conducted without any access to outcome data to assure the objectivity of the design. A critical additional activity in the design phase is the specification of a protocol for the estimation procedure to be used once outcome data are revealed.
The statistical literature includes many procedures for estimating treatment effects in observational studies. Gutman and Rubin [10] showed that when Y and X are scalar and continuous, and treatment assignment only depends on X, only a few of the common procedures are generally statistically valid, with the most promising ones being matching with replacement for effect estimation, with or without covariance adjustment, combined with within-treatment group matching for sampling variance estimation, suggested by Abadie and Imbens [11], and called in Gutman and Rubin [10] M–N–m and M–C–m (M, cross-group matching for point estimate; C/N, with/without covariance adjustment; m, matching within treatment group for sampling variance estimate).
Recently, Gutman and Rubin [12] described a new procedure called multiple imputation (MI) with two subclassification splines (MITSS) that explicitly views causal effect estimation as a missing data problem [13, 14] but only considered binary outcomes and a scalar covariate. Here, we extend this procedure to settings with multiple covariates and multiple continuous outcomes. The proposed extended procedure shares conceptual similarities with a procedure proposed by Little and An [15] for missing data imputation and relies on an initial propensity score estimation.
We compare the validity and performance of MITSS to the most promising methods identified by Gutman and Rubin [10] using Neyman’s framework of frequentist operating characteristics, in which an α-level interval estimate is ‘valid’ if, under repeated sampling from the population (finite or super), it covers the estimand in at least α percent of the samples. In addition to validity, we compare the accuracy, biases, and mean square errors (MSEs) of point estimates. Of all the methods considered, and across most of the conditions examined, when the distributions of the covariates in the control and treatment groups have adequate overlap, as defined in Sections 3 and 4, the new procedure, MITSS, is superior. When there is limited overlap, no method can be trusted because they all rely on unassailable extrapolation assumptions.
1.2. Framework
For a group of N units in the population indexed by i = 1, …, n ≤ N, n0 units receive the control treatment (Wi = 0) and n1 = n − n0 units receive the active treatment (Wi = 1). Let Yi(0) and Yi(1) be the ‘potential’ outcomes corresponding to the appropriate treatment level at a specific point time; the use of potential outcomes to define causal effects is due to Neyman [16] in the context of randomization-based inference in randomized experiments and generalized to other settings in Rubin [17]. Assuming the stable unit treatment value assumption [18, 19], we can write the observable outcome for unit i as
(1) |
For each unit i, there is also a vector of covariates with dimension L that are unaffected by W, Xi = (Xi1, …, XiL) (Xi could also include higher order interactions and non-linear terms in basic covariates). For causal inference, we need the probability that each unit received the active rather than the control treatment, that is, the assignment mechanism. Assuming unconfoundedness [19] and that units can be modeled independently given the parameters, then the assignment mechanism for unit i can be written as
(2) |
where e(Xi) is the propensity score for unit i [6], and the dependence on ϕ is notationally suppressed in (2). Throughout, we assume, as is typical in practice in observational studies, that e(Xi) is unknown and only an estimate of it, ê(Xi), is available. Estimating e(Xi) is part of the design phase of observational studies [1], and many procedures have been proposed to perform it (see [20] for a review). In this manuscript, we assume that the propensity scores have been estimated to the satisfaction of the researcher without observing any outcome variables. Thus, the observed data for person i comprise the vector { , Xi, Wi, ê(Xi)}.
1.3. Estimands
Causal effects are summarized by estimands, which are functions of the unit-level potential outcomes on a common set of units [13, 17]. There are two main types of estimands in statistics: finite population (N < ∞) and super-population (N = ∞). A commonly used estimand is the super-population average treatment effect (ATE), γ1 ≡ E(Y(1) − Y(0)), where the expectation is taken over the entire population. The finite population ATE is . A generalization of the ATE can be obtained by summarizing the effects for a subpopulation of units based on X. Letting E(Y(w)|X) = hw(X) for w = 0, 1, one generalization of the ATE is the average difference between the two response surfaces at X = x, h1(x) − h0(x), and in the finite population, , where N(x) is the number of units in the population for which X = x.
Another generalization of the ATE uses non-linear functions of the potential outcomes. For example, when Yi(w), w ∈ {0, 1} are continuous, an estimand of interest is the proportion of units for whom the treatment is not useful or even harmful. More specifically, γ2 = E(δ(Y(1) − Y(0))), where δ(a) is one if a ≤ 0 and zero otherwise, and its finite population counterpart is .
Most of the commonly used causal inference procedures are implicitly intended for super-population linear estimands. We propose new procedures that rely on MI [21, 22], which allow the estimation of both finite and super-population estimands as well as non-linear estimands, but focus here on the more common super-population estimands.
1.4. Current methods
When Xi and Yi are scalar and continuous, and the distributions of X in the control and active treatment groups differ, Gutman and Rubin [10] identified the two most promising procedures for inference for the ATE: matching for effect estimation combined with within group matching for sampling variance estimation (M–N–m) [11], and matching for effect estimation with covariance adjustments combined with within group matching for sampling variance estimation (M–C–m) [11, 23]. This sampling variance estimation allows for heteroskedasticity across treatment groups and levels of the propensity score as described in Section 4 of Abadie and Imbens [11]. Because procedures that perform marginally worse with scalar X may perform better than the most promising scalar procedures when dealing with multiple X, three additional mostly valid procedures were included here. The three procedures include full matching (FM) [24, 25], weighting (IPW1) [9], and a so-called doubly robust method (DR) [8, 9]. For concise description of these procedures, see Gutman and Rubin [10]. These five procedures are compared here with the newly proposed procedure for both scalar X and multivariate X. Other procedures such as regression adjusted subclassification were not included, because they were generally invalid even with scalar X [10].
2. MI with two splines in subclasses (MITSS)
We propose a new procedure that addresses causal effect inference from a missing data perspective [14] by multiply imputing the missing potential outcomes.
The response surfaces, h0(X) and h1(X), are considered unknown and are approximated by h̃0(X, θ0) and h̃1(X, θ1), respectively, where h̃w has a known parametric form, and θw (w ∈ {0, 1}) are unknown vector parameters.
MITSS shares some similarities with the penalized spline propensity prediction model [15, 26] proposed for missing–data imputation. Relying on the missing at random assumption [14], penalized spline propensity prediction imputes missing values using an additive model that combines a penalized spline on the probability of being missing and linear model adjustments on all other orthogonalized covariates. Formally, MITSS assumes that
(3) |
where g is commonly referred to as the link function [27], fw is a spline with K knots, u(x) = log(x/(1 − x)), is a vector parameter defining the splines and linear adjustments orthogonal to them, and , with being Xil orthogonal to u(ê(Xi)) [28]. Operationally, regress Xil on u(ê(Xi)), and set to the residuals of this regression. Throughout, we assume that g is the identity link function.
The intuition behind functional form (3) is first to subclassify on the estimated propensity score, so that within each subclass, Xort is well balanced across treatment groups; second, to fit a spline across propensity subclasses, because Bayesian models for missing data that include e(Xi) are likely to be well calibrated [29]. Causal effects estimation can be viewed as a missing data problem [13], and Bayesian modeling has been used to make inferences when missing data occur [30]. Here, we use Bayesian modeling to make causal inferences; and third, to use simple linear model adjustments for Xort to gain efficiency and control for minor residual imbalances. A more complex function of Xort, such as a spline function for each dimension of Xort or an interaction term between each subclass indicator and each dimension of Xort, can be used to replace the linear adjustments; however, because the differences in covariates’ distributions are mostly captured by the spline along the propensity score, and Xort is well balanced within each subclass, the gain in efficiency might not justify the added model complexity and were not examined here.
Formally, let γ= τ(Y(0), Y(1))), where τ is a known function, and Y(w) = {Y1(w), …, Yn(w)}T. A Bayesian approach to estimating γ will condition on the observed values, Yobs, X, W, and approximate the distribution p(γ | Yobs, X, W). Assuming (2), this distribution is
(4) |
where and . Equation (4) shows that in order to estimate γ, we only need to estimate the probability of p(Ymis | Yobs, X), because p(γ | Yobs, Ymis, X, W) is a degenerate distribution for finite population estimands and a known function for super-population estimands.
Let ψw(θw | Yobs, X) be the posterior density of θw (i.e., θw’s conditional distribution given Yobs and X). Assuming that Yi(1) and Yi(0) are conditionally independent given X and θw, and that the prior distributions of θ1 and θ0 are independent, then ψ0 and ψ1 are independent given Yobs and X. This assumption is also known as no contamination of imputation across treatments [31], and it is sometimes implicitly made (e.g., with DR methods), and sometimes explicitly not made [31]. Based on this assumption, we have the following:
(5) |
where Y(0)mis comprise the potential outcomes under control for all the units with Wi = 1 and Y(1)mis comprise the potential outcomes under the active treatment for all the units with Wi = 0.
Using Equation (5), an estimate for is obtained by , which reduces to
where
(6) |
Analogously, letting Δ(X) be the density of Xi in the super-population, estimate γ1 by
Estimating Δ(X) is part of the design phase of observational studies, because it does not involve any outcomes. The standard error of and γ̂1 is obtained using MI. This point estimate may have some vague resemblance to the DR estimate, but it differs by at least three important aspects. First, it does not rely on weights, which may result in large biases and sampling variance in certain situations [32, 33]. Second, it introduces flexible spline models and adjustments for orthogonalized covariates. Third, it integrates over the posterior distribution of the parameters and explicitly imputes the missing potential outcomes. Thus, the proposed procedure is not restricted to specific parameter estimation and can provide point and interval estimates for estimands that do not have a closed form (e.g., and γ2).
We now explicate and summarize this procedure:
Partition the n units into at most six subclasses, such that there are at least three units from each treatment group in each of the subclasses: the partitioning is performed using Xi when it is scalar, or ê(Xi) when Xi comprises multiple covariates, for example, using the method described in Imbens and Rubin [34]. The number 3 was chosen so that all of the parameters of the spline model can be estimated even when using an improper prior distribution.
From observed data { , Xi, i ∈ 1, …, N}, independently estimate E(Yi(0) | Xi, θ0) = h̃0(Xi, θ0) and E(Yi(1) | Xi, θ1) = h̃1(Xi, θ1), using regression splines with knots located at the borders of the subclasses and a linear combination of Xort, as in (3); the location of the knots is guided by step 1, so that the distribution of in each subclass is similar in the control and treatment groups.
Draw the vector parameters θ1 and θ0 from their independent posterior distributions.
Using the drawn values of θ1 and θ0, independently impute the missing control potential outcomes in the active treatment group, the Yi(0) with Wi = 1, and the missing active treatment potential outcomes in the control group, the Yi(1) with Wi = 0.
Repeat steps 4 and 5 M times (in our simulations, M = 25).
Estimate the treatment effect γ and its sampling variance in each of the imputed datasets; from imputed dataset m, let γ̂(m) be the estimated treatment effect, and let V̂(m) be its estimated sampling variance, m = 1, …, M. When γ is a finite-sample estimand, V̂(m) = 0.
- Estimate γ by and its standard error by
Calculate the interval estimate for γ, using the t approximation for the posterior variance of γ given by Barnard and Rubin [35].
When M is large (e.g., > 100), the last two steps may be replaced by using percentiles of γ̂(m), to obtain posterior interval for γ.
3. Scalar Xi
To investigate the performance of MITSS in comparison with current statistically approximately valid procedures, we used simulation analysis with scalar Xi, because methods that perform poorly with a single covariate will generally perform poorly with multiple covariates. The simulation configurations are identical to these in Gutman and Rubin [10] and comprise two types of situational factors. The first type is either known to the investigator, such as sample sizes, or can be easily estimated without examining any outcome data, such as the covariate (Xi) distributions. The second type of factor is not known to the investigator and cannot be estimated for the data at the design phase; these define the response surfaces generating the outcome data from the covariates. See Table II from Gutman and Rubin [10] for more details. The changes in the current scalar Xi simulations are that MITSS is now included, and of the procedures listed in Table I in Gutman and Rubin [10], only M–N–m, M–C–m, FM, IPW1, and DR are included, because they were the most promising.
Table II.
Method | Type | Percent | Absolute bias | Interval width | RMSE |
---|---|---|---|---|---|
IPW1 | A | 59 | 1.06 | 0.22 | 0.28 |
B | 10 | 0.32 | 0.78 | 0.65 | |
C | 20 | 1.70 | −0.68 | 0.11 | |
D | 11 | 0.85 | −0.15 | 0.56 | |
DR | A | 51 | 0.50 | −0.03 | −0.02 |
B | 4 | 0.90 | 0.84 | 0.77 | |
C | 28 | 1.07 | −0.60 | −0.06 | |
D | 17 | 0.42 | −0.27 | 0.29 | |
M–N–m | A | 73 | 0.43 | 0.17 | 0.18 |
B | 11 | −0.02 | 0.31 | 0.22 | |
C | 6 | 0.76 | −0.57 | −0.14 | |
D | 9 | 0.40 | 0.15 | 0.31 | |
M–C–m | A | 73 | 0.26 | 0.17 | 0.18 |
B | 12 | 0.19 | 0.32 | 0.24 | |
C | 7 | 0.38 | −0.66 | −0.40 | |
D | 9 | 0.25 | 0.12 | 0.19 | |
Full match | A | 55 | 0.47 | 0.03 | 0.05 |
B | 9 | −0.32 | 0.02 | −0.05 | |
C | 24 | 0.61 | −0.41 | −0.24 | |
D | 12 | 0.42 | −0.35 | 0.22 |
Differences are displayed as alternative methods to MITSS.
The types are configurations for which 95% interval estimates has the following coverages characteristics: A, MITSS ≥ 0.9 Method ≥ 0.9; B, MITSS < 0.9 Method ≥ 0.9; C, MITSS ≥ 0.9 Method < 0.9; D, MITSS < 0.9 Method < 0.9.
Overall, there are 35 × 2 different configurations of known factors, and 36 (35) monotone (non-monotone) response surface configurations, resulting in a 311 × 2 (310 × 2) factorial design for monotone (non-monotone) response surfaces. For each configuration, Nrep = 100 replications were produced. As in Gutman and Rubin [10], throughout this simulation, we restrict the analysis of results to configurations for which the Jensen–Shannon divergence (JSD) [36] was less than 0.3, so there is reasonable overlap in the X distributions between the two treatment groups. JSD is a measure that quantifies differences between distributions and is defined by
(7) |
where G0 and G1 are the distributions of X in the control and active treatment, respectively, QKL is the Kullback–Leibler divergence [37], and λ is the proportion of treated units in the population (see [10] for additional details).
3.1. Results for interval coverages
We compare the coverages of the five promising methods identified by Gutman and Rubin [10] to MITSS. Estimation of the propensity score for the weighting methods was accomplished using the algorithm described by Imai and Ratkovic [38]. For MITSS, we assumed that Yi(W)|X, θW, σW ~ 𝒩(h̃W(θW, X), σW), where 𝒩 is the is the Normal distribution, and, for analysis, we also assumed the commonly used independent improper prior distribution for θW and σW:
(8) |
For simplicity, the estimand of interest was the super-population ATE. For each of the five procedures and MITSS, at each configuration, and at each of the 100 replications, we calculated the estimated treatment effect, the estimated sampling variance, and the corresponding interval width and determined whether the 95% interval covered the true treatment effect. Using these values, we calculated for each configuration the estimated coverage rate, the bias, the mean sampling variance, and the mean interval width. Throughout, coverage rates are summarized by the proportions of configurations with over 90% coverage for a 95% interval estimate that are expressed as percentages, as well as by the medians, the inter-quartile ranges, and the 25% percentiles across configurations that are expressed as decimals.
Table I displays the proportion of configurations in which the different methods lead to over 90% coverage, as well as the median and inter-quartile range of the coverage rate of the intervals, for the six methods when response surfaces were monotone. MITSS has 79% of the configurations with at least 90% coverage, which is second only to M–N–m and M–C–m. The rest of the methods had less than 75% of the configurations with over 90% coverage.
Table I.
Method | Overall
|
Null
|
Non-Null
|
||||
---|---|---|---|---|---|---|---|
% of Configurations with over 90% coverage | Median | 25% | 75% | Median | 25% | 75% | |
IPW1 | 70 | 0.98 | 0.86 | 1 | 0.97 | 0.82 | 1 |
DR | 55 | 0.92 | 0.70 | 0.96 | 0.91 | 0.65 | 0.96 |
FM | 64 | 0.96 | 0.93 | 0.99 | 0.93 | 0.83 | 0.98 |
M–N–m | 84 | 0.94 | 0.90 | 0.97 | 0.98 | 0.94 | 1 |
M–C–m | 84 | 0.93 | 0.89 | 0.96 | 0.98 | 0.94 | 1 |
MITSS | 79 | 0.94 | 0.91 | 0.96 | 0.94 | 0.91 | 0.96 |
Null configurations are simulation conditions in which h1(x) = h0(x) ∀x ∈ ℝ, and non-null configurations are all other configurations. For both null and non-null configurations, MITSS has median interval coverages that are close to and also most concentrated around 95%. Similar results with slightly different numbers occur for non-monotone response surfaces (Supporting Information).
Figure 1 summarizes the median and the 25th percentile of the coverages as functions of the JSD for monotone response surfaces. The median coverages of M–N–m and M–C–m are above 95% for every value of the JSD, which suggests over-coverage. MITSS and FM generally result in median coverages that are close to nominal. However, the 25th percentile plot shows that all methods have decreased coverage as the JSD increases, with MITSS, M–N–m, and M–C–m having 25th percentiles still close to 95%, even for large JSD values. IPW1 and DR have median coverage that is higher than 95% for lower JSD values, but they have substantial under-coverage for larger JSD values. Similar trends are observed for non-monotone response surfaces (Supporting Information).
3.2. Results for biases, RMSEs, and intervals widths
Table II compares the median log differences in absolute bias, interval widths, and root MSEs (RMSEs) across configurations with monotone response surfaces for the six procedures. We compared MITSS with each of the five procedures separately, by classifying the configurations into four types (A, B, C, D). Type A comprises configurations where both MITSS and the compared procedure have 95% interval estimates with coverages that are higher than 0.9. For all of the methods, the majority of the configurations are classified as type A (IPW1 = 59%, DR = 51%, M–N–m = 73%, M–C–m = 73%, Full Match = 55%). In these configurations, MITSS had smaller median bias, and except for DR, it also had smaller median RMSE and median interval width. In configurations where MITSS had lower than 0.9 coverage and the other methods had higher than 0.9 coverage (labeled type B), MITSS had larger bias than M–N–m and FM, but smaller bias than M–C–m and the weighting methods; MITSS’s interval widths in these configurations were generally too short. In configurations where MITSS had intervals with coverages that were higher than 0.9, and the other methods had coverages that were lower than 0.9 (labeled type C), MITSS had smaller median bias and wider median interval width; similar trends are observed for non-monotone response surfaces (Supporting Information). Other than M–N–m and M–C–m, MITSS had significantly more type C configurations than type B (Table II).
This analysis shows that, in comparison with the other methods, MITSS is generally valid and it has point estimates with smaller biases and generates shorter 95% intervals. It is also interesting to note that IPW1 and DR had the largest biases, and IPW1 also had the largest interval widths. Similar results appear for non-monotone response surfaces. These results support the analyses performed by Kang and Schafer [32] and Waernbaum [33], that IPW1 and DR methods are often statistically inferior methods.
When the distributions of X in the treatment and control groups differ substantially, none of the methods are statistically valid (labeled type D). To avoid these situations, it is important, at the design stage, to examine the overlap of the distributions of X in the control and treated groups. Units in one group that have no close match in the other group should be discarded, and causal inferences restricted to the region of X where there is overlap. The new estimand differs from the original estimand, which is generally impossible to estimate without making empirically unassailable assumptions. Our summary of the simulations with scalar X is that, of the methods considered, MITSS has the best statistical properties.
4. Multivariate X
When investigating multiple covariates, we consider the methods described in Section 3 modified so that they are matched/subclassified based on the estimated propensity score, instead of scalar X. Covariance adjustment for M–C–m was performed using linear regression that included Xort. We also considered three versions of MITSS with cubic splines along the propensity score: (i) the original MITSS; (ii) MITSS with 15 instead of six subclasses (MITSS-15); and (iii) MITSS with six subclasses but without linear adjustment for covariates orthogonal to the propensity score (MITSS-PS). All three versions of MITSS assumed that Yi(W) ~ 𝒩(h̃W(θW, X), σW) with corresponding h̃W and prior distributions (8).
Based on the results of the scalar covariate simulation, we also supplemented all of the methods with a preliminary step that discarded all units without ‘close’ Xi matches in the other treatment group. This step was introduced because the chance that units in the treatment and control groups will have close Xi matches in the other group decreases with the number of covariates.
The general advice for discarding units is that it should be performed with scientific knowledge and is usually study specific. However, in a simulation, a context-free discarding rule needs to be specified explicitly. One possible rule is to keep all units with estimated propensity scores between [êmin(X), êmax(X)], where êmin(X) = max{min{ê(Xi) | Wi = 0}, min{ê(Xi | Wi = 1)} and êmax(X) = min{max{ê(Xi) | Wi = 0}, max{ê(Xi | Wi = 1)}. Crump et al. [39] proposed a different rule that enjoys an optimal precision criterion and keeps all units with propensity scores between (0.1, 0.9). The results using either of these rules are quite similar, with a slight bias advantage to the rule that discards more observations. Thus, we only display the results for the [êmin(X), êmax(X)] discarding rule. Overall, we compare eight procedures each of them with and without discarding.
4.1. Covariate distributions
Let , J is the dimension of ; X′ is a matrix with n rows and J columns; includes all the second-order interactions of . Formally, ; X″ is a n × (J(J − 1)/2) matrix, and X is an n × L matrix, where L = J + J(J −1)/2. The values of were generated from two different multivariate skew-t distributions [40].
where I is the identity matrix, μ̃ = (μ̃, 0, …, 0), Ω is a diagonal matrix with entries (ν1, ν2, …, ν2), η0 = (η0, 0, …, 0), η1 = (η1, 0, …, 0), and df are the degrees of freedom. We indexed the distance between the treated and the control groups’ means in terms of the standardized bias, . It should be noted that the initial bias is only in the first component, but it is a function of the dimension of . This idea also appeared in Rubin [41] and Rubin and Thomas [42] and, in the case of ellipsoidally symmetric distributions, corresponds to adding a constant bias to each component.
For each configuration defined by the factors SB, , η0 and η1, Nrep = 100 replications were generated. Table III describes the three levels of each of these factors. The factors J, df, r and n1 also vary in each of the settings, yielding a factorial design with 35 × 23 × 3 levels.
Table III.
Factor | Levels multivariate X | Description | ||
---|---|---|---|---|
r | {1, 2} | Ratio of sample sizes | ||
n | {300, 600, 1200} | Treatment group population size | ||
η0 | {−3.5, 0, 3.5} | Skewness of the first covariate in the treatment group | ||
η1 | {−3.5, 0, 3.5} | Skewness of the first covariate in the control group | ||
B |
|
Standardized bias for first covariate | ||
|
|
Ratio of variances for first covariate | ||
βX | {0.5, 1, 2} | Correlation between X1iand Yi(W) | ||
|
|
Ratio of variances of all covariates except the first covariate | ||
df | {7, ∞} | Degrees of freedom of skewed t-distribution | ||
J | {2, 5, 10} | Number of covariates | ||
CX | {2, 3, 5} | Complexity of multivariate function | ||
G0(x) and G1(x) | Described in the Supporting Information and response surfaces |
For all of the methods and all of the configurations, ê(Xi) was estimated using logistic regression model that included X, , and Xp1 × Xp2, ∀p1, p2 ∈{1, …, J}, p1 ≠ p2. In cases where the logistic regression did not converge within 25 iterations, linear discriminant analysis was used instead to estimate ê(Xi).
When the dimension of the covariate increases, estimating the JSD for two distributions can become computationally intensive owing to the need for estimating multivariate integrals. To save computation time, we estimated the JSD for the distributions of the estimated propensity scores in the treatment and control group using kernel-density estimation [43]. We considered only configurations that had JSD < 0.3 on the estimated propensity scores without any discarding, and in a separate analysis, we considered these configurations after discarding non-overlapping units.
4.2. Response surfaces for potential outcomes given covariates
In each simulation replication, we randomly generate the continuous outcome data from
(9) |
where F1 and F0 are normal distributions with conditional means f0(X, B0) and f1(X, B1), respectively, and conditional variances σY; σY, B1 and B0 are parameters unknown to the investigator. The functions f0 and f1 covered a wide range of possibilities. They were generated by an algorithm that was controlled by four factors responsible for different characteristics of the multivariate functions: βX ∈ BW governed the correlation between and Yi(W); CX ∈ BW governed the ‘complexity’, and , W ∈ {0, 1} controlled whether Yi(W) was influenced by second-order interactions. Detailed implementation of the procedure is described in the Supporting Information.
For each configuration of the factors that are known to the investigator, as well as βX, CX, and , the generating procedure was replicated 100 times, thereby investigating monotone and non-monotone response surfaces and the effect of different functional factors on the success of the estimation procedures.
4.3. Results
Using the same methodology as in Section 3.1, we examined the coverage of the methods described in Section 3 for 95% nominal intervals for configurations with JSD < 0.3 on the estimated propensity score (Table IV). MITSS, MITSS-15, FM, and M–N–m are the only methods with coverage that is higher than 90% in at least 60% of all configurations. The rest of the methods have less than 56% of the configurations with coverage over 90%. Figure 2 displays the median and 25% coverage across configurations for different JSD values. M–N–m, FM, MITSS-15, and MITSS all have median coverage that is close to nominal even for large JSD values. Each of IPW1 and DR have median coverage that decreases sharply with increasing JSD, and M–C–m and MITSS-PS have a more moderate median coverage decline. The 25% coverage figure shows similar trends, with more pronounced decreases in coverage for MITSS and FM as JSD increases.
Table IV.
Method | Overall
|
Null
|
Non-null
|
||||
---|---|---|---|---|---|---|---|
% of configurations with over 90% coverage | Median | 25% | 75% | Median | 25% | 75% | |
MITSS-PS | 49 | 0.9 | 0.76 | 0.96 | 0.89 | 0.75 | 0.95 |
MITSS | 63 | 0.92 | 0.83 | 0.96 | 0.92 | 0.85 | 0.95 |
MITSS-15 | 73 | 0.94 | 0.88 | 0.97 | 0.94 | 0.89 | 0.96 |
M–N–m | 72 | 0.96 | 0.90 | 0.98 | 0.95 | 0.87 | 0.98 |
M–C–m | 48 | 0.89 | 0.74 | 0.95 | 0.89 | 0.74 | 0.95 |
DR | 56 | 0.92 | 0.69 | 0.97 | 0.92 | 0.71 | 0.96 |
IPW1 | 52 | 0.93 | 0.58 | 1 | 0.90 | 0.54 | 0.99 |
FM | 64 | 0.94 | 0.86 | 0.97 | 0.92 | 0.84 | 0.96 |
Of the configurations with JSD < 0.3, 94.8% had at least 15 units in each treatment group that satisfied the [êmin(X), êmax(X)] rule in all of the 100 replications. Less than 2% of the configurations had fewer than 15 units in at least one of the treatment groups that satisfied this rule for all 100 replications and were removed from the analysis. Table V displays the 95% coverage for configurations with JSD < 0.3 after discarding observations that did not overlap. All of the methods, except for M–N–m, improved in terms of coverage. MITSS, MITSS-15, DR, and IPW1 are the only methods with over 70% of the configuration having at least 90% coverage. However, the IPW1 25th percentile is higher than nominal, indicating that its interval estimates may be inefficient. Except for IPW1, all of the methods have median coverages that are close to nominal after discarding non-overlapping units (Figure 3(a)). Figure 3(b) shows that the coverages at the 25th percentile drop sharply for IPW1 when JSD increases, and only moderate decreases are observed for the other methods.
Table V.
Method | Overall
|
Null
|
Non-Null
|
||||
---|---|---|---|---|---|---|---|
% of configurations with over 90% coverage | Median | 25% | 75% | Median | 25% | 75% | |
MITSS-PS | 65 | 0.96 | 0.84 | 0.99 | 0.94 | 0.82 | 0.97 |
MITSS | 74 | 0.94 | 0.89 | 0.97 | 0.94 | 0.9 | 0.96 |
MITSS-15 | 75 | 0.94 | 0.89 | 0.97 | 0.94 | 0.9 | 0.96 |
M–N–m | 68 | 0.96 | 0.85 | 0.98 | 0.94 | 0.85 | 0.97 |
M–C–m | 68 | 0.96 | 0.85 | 0.98 | 0.94 | 0.85 | 0.97 |
DR | 78 | 0.96 | 0.91 | 0.98 | 0.95 | 0.91 | 0.98 |
IPW1 | 80 | 1 | 0.98 | 1 | 1 | 0.96 | 1 |
FM | 68 | 0.96 | 0.87 | 0.98 | 0.93 | 0.86 | 0.96 |
Table VI compares the median log differences in absolute bias, interval widths, and RMSEs. For all of the methods, most of the configurations are classified as type A (both MITSS and the compared method have over 90% coverage). In configurations of type A, MITSS generally has smaller bias and RMSE, and except for M–N–m and FM, MITSS also has shorter interval width. In configurations of type B (MITSS has less than 90% coverage and the other method more than 90%), MITSS has lower median bias and median RMSE in comparison with all three methods but wider interval widths in comparison with FM and M–N–m. It is important to note that in these configurations, MITSS has coverage that is smaller than nominal in comparison with the other methods and is thus generally invalid. With configurations of type C (MITTS has coverage over 90% and the other method less than 90%), MITSS has smaller bias, interval width, and RMSE compared with MITSS-15, MITSS-PS, IPW1, and M-n-m, but larger bias and RMSE compared with FM, DR, and M–C–m. However, in these configurations, MITSS is generally statistically valid, and the other methods are not. In summary, without any discarding, M–N–m, FM, MITSS-15, and MITSS are generally valid for smaller JSD values, and MITSS generally has smaller bias and RMSE in comparison with these methods.
Table VI.
Method | Type | Percent | Absolute bias | Interval width | RMSE |
---|---|---|---|---|---|
MITSS-PS | A | 41 | 0.27 | 0.15 | 0.27 |
B | 8 | 0.42 | 0.09 | 0.42 | |
C | 22 | 0.19 | 0.38 | 0.19 | |
D | 29 | 0.34 | 0.19 | 0.34 | |
MITSS-15 | A | 61 | 0.41 | 0.52 | 0.41 |
B | 12 | 1.43 | 1.68 | 1.43 | |
C | 2 | 0.21 | 0.35 | 0.21 | |
D | 25 | 0.36 | 0.31 | 0.35 | |
M–N–m | A | 51 | 0.13 | −0.13 | 0.13 |
B | 22 | 0.21 | −0.37 | 0.21 | |
C | 12 | 0.08 | 0.19 | 0.08 | |
D | 15 | 0.18 | −0.27 | 0.18 | |
M–C–m | A | 37 | 0.22 | 0.1 | 0.22 |
B | 10 | 0.36 | 0.02 | 0.36 | |
C | 26 | −0.12 | 0.09 | −0.12 | |
D | 27 | 0.1 | −0.06 | 0.1 | |
DR | A | 46 | 0.28 | 1.05 | 0.28 |
B | 9 | 1.95 | 2.38 | 1.95 | |
C | 17 | −0.97 | −0.39 | −0.97 | |
D | 28 | −0.51 | −0.09 | −0.51 | |
IPW1 | A | 39 | 2.85 | 3.68 | 2.85 |
B | 13 | 1.94 | 2.28 | 1.94 | |
C | 24 | 1.73 | 3.13 | 1.77 | |
D | 24 | 0.86 | 1.61 | 0.88 | |
FM | A | 48 | 0.21 | 0.03 | 0.21 |
B | 16 | 0.33 | −0.15 | 0.33 | |
C | 15 | −0.08 | 0.00 | −0.08 | |
D | 21 | −0.10 | −0.44 | −0.11 |
Differences are displayed as alternative methods to MITSS.
The types are configurations for which 95% interval estimates has the following coverages characteristics: A, MITSS ≥ 0.9 and Method ≥ 0.9; B, MITSS < 0.9 and Method ≥ 0.9; C, MITSS ≥ 0.9 and Method < 0.9; D, MITSS < 0.9 and Method < 0.9.
Table VII compares the median log differences in absolute bias, interval widths, and RMSEs between the methods examined in Section 3 and MITSS. After discarding, the majority of the configurations are classified as type A. In configurations of type A, MITSS has smaller bias, shortest interval width, and smaller RMSE in comparison with all of the methods, except for MITSS-15, for which it has approximately equal bias, interval width, and RMSE. In fact, except for DR and MITSS-15, MITSS has smaller bias, shortest interval width, and smaller RMSE even in configurations of type C. Furthermore, the adjustments for the components of the covariates orthogonal to the estimated propensity score are beneficial as revealed by the comparison of MITSS with MITSS-PS. In summary, after discarding, almost all of the methods improved their coverage, and MITSS generally has smallest bias and RMSE, and shortest interval width.
Table VII.
Method | Type | Percent | Absolute bias | Interval width | RMSE |
---|---|---|---|---|---|
MITSS-PS | A | 58 | 0.27 | 0.11 | 0.27 |
B | 7 | 0.38 | −0.01 | 0.37 | |
C | 17 | 0.16 | 0.39 | 0.16 | |
D | 19 | 0.22 | 0.02 | 0.22 | |
MITSS-15 | A | 72 | 0.0 | 0.0 | 0.0 |
B | 3 | 0.01 | 0.11 | 0.01 | |
C | 2 | −0.01 | 0.02 | −0.01 | |
D | 23 | 0.0 | 0.03 | 0.0 | |
M–N–m | A | 60 | 0.4 | 0.27 | 0.4 |
B | 8 | 0.54 | 0.14 | 0.54 | |
C | 15 | 0.3 | 0.5 | 0.3 | |
D | 17 | 0.36 | 0.06 | 0.36 | |
M–C–m | A | 59 | 0.4 | 0.27 | 0.4 |
B | 8 | 0.54 | 0.14 | 0.54 | |
C | 15 | 0.3 | 0.5 | 0.3 | |
D | 17 | 0.36 | 0.06 | 0.36 | |
DR | A | 68 | 0.54 | 1.36 | 0.54 |
B | 10 | 1.54 | 2.11 | 1.55 | |
C | 6 | −0.26 | 0.31 | −0.26 | |
D | 16 | 0.01 | 0.23 | −0.01 | |
IPW1 | A | 68 | 4.83 | 3.92 | 4.83 |
B | 12 | 3.86 | 2.89 | 3.86 | |
C | 7 | 1.97 | 4.70 | 2.21 | |
D | 13 | 1.63 | 3.09 | 1.71 | |
FM | A | 60 | 0.49 | 0.43 | 0.49 |
B | 8 | 0.76 | 0.32 | 0.76 | |
C | 15 | 0.41 | 0.55 | 0.41 | |
D | 17 | 0.30 | 0.05 | 0.30 |
Differences are displayed as alternative methods to MITSS.
The types are configurations for which 95% interval estimates has the following coverages characteristics: A, MITSS ≥ 0.9 and Method ≥ 0.9; B, MITSS < 0.9 and Method ≥ 0.9; C, MITSS ≥ 0.9 and Method < 0.9; D, MITSS < 0.9 and Method < 0.9.
5. Simulations Using Real Data
In Sections 3 and 4, the covariate distributions and the response surfaces were simulated from analytic distribution. We now apply MITSS and M–N–m to real X and Yobs data, where we simulate only the unobserved potential outcomes. The dataset for this analysis is from Shadish et al. [44], which was designed to estimate the effects of a specific training program on mathematics and vocabulary scores of undergraduates; thus, the potential outcomes were two-dimensional vectors of the two scores. The study also includes a rich set of covariates. In our analysis, we treat all of the students as part of one observational study with an unconfounded assignment mechanism.
5.1. Partially simulated data
We conducted two sets of simulations. The first set assumed that the intervention had a constant (additive) effect, and the unconfounded assignment mechanism involved ten randomly selected covariates, five randomly selected squared covariates, and five randomly selected second-order interactions. In the second set of simulations, we imputed unobserved potential outcomes by searching for the ‘most similar’ student, in terms of the 28 basic covariates, among those who received the other training. The most similar student to student i was chosen by first finding a donor pool of Ki students (1 ≤ Ki ≤ n) with the smallest Hamming distance [45] from student i, based on sex, marital status, preference for literature, ethnicity, and whether the student was a math major. Among the Ki students, the student with the smallest Mahalanobis distance [46] to student i on the continuous covariates (such as pre–training scores, ACT and SAT scores, and GPA) was declared as the most similar student to student i. The observed values of the two potential outcomes for the most similar student were imputed for the missing potential outcomes for student i. Once all of the missing potential outcomes were imputed, treatment assignments were obtained in a way similar to the constant effects simulation setting. Each student’s potential outcome vector included the treatment (mathematics) and the control (vocabulary) scores. In our simulations, we generated 50 different treatment assignment probability models, and for each of these, we generated 40 realizations of potential outcomes.
MITSS used the following bivariate regression model for multiply imputing the missing outcomes:
where are the potential outcomes for the mathematics and vocabulary scores, respectively; for unit i receiving treatment W, and were defined by (3), and Φ2 is a bivariate Normal random variable, with mean vector zero and covariance matrix ΣW. The prior distributions were and .
5.2. Results
The estimated propensity score was obtained using the algorithm described in Imbens and Rubin [34, Chapter 13]. Following Shadish et al. [44], we defined the estimand of interest to be γ1. Because γ1 is now a two-component vector, there are several ways to construct interval estimates for it. For example, one can use either an ellipsoidal 95% region or a rectangular region obtained from a combination of single coordinate 95% interval estimates. The latter region analytically has less than 95% family-wise coverage. Table VIII displays the median bias, coverage, interval width, and RMSE across assignment probability models for both interval estimate constructions, as well as for configurations with constant and heterogeneous treatment effects. Both methods covered the true treatment effect at a rate that is higher than or equal to nominal for both configurations when using rectangular confidence intervals. In addition, MITSS produced valid ellipsoidal confidence regions. In principal, one could implement a multivariate normal analysis using M–N–m and obtain an ellipsoidal confidence region. However, here we illustrate the scalar analysis that is generally performed in practice. In both sets of simulations, MITSS has generally smaller bias than M–N–m, and it always enjoys shortest interval widths and smaller RMSEs.
Table VIII.
Constant effect
|
Heterogeneous effect
|
|||
---|---|---|---|---|
M–N–m | MITSS | M–N–m | MITSS | |
Bias math | 0.04 | 0.04 | 0.04 | 0.03 |
Bias vocabulary | 0.07 | 0.04 | 0.05 | 0.04 |
% Reduction in bias math | 0.81 | 0.76 | 0.72 | 0.79 |
% Reduction in bias vocabulary | 0.67 | 0.69 | 0.59 | 0.65 |
Coverage super-population math | 0.95 | 0.98 | 0.98 | 1.00 |
Coverage super-population vocabulary | 0.95 | 0.98 | 0.98 | 0.98 |
Coverage super-population joint | — | 0.98 | — | 1.00 |
Super-population interval width math | 0.77 | 0.68 | 0.85 | 0.76 |
Super-population interval width vocabulary | 0.82 | 0.73 | 0.83 | 0.74 |
RMSE math | 0.40 | 0.37 | 0.44 | 0.38 |
RMSE vocabulary | 0.42 | 0.38 | 0.43 | 0.38 |
Finally, we are also interested in an estimand that describes the proportion of people for whom the mathematics training is not useful or even harmful. This estimand cannot be obtained using simple weighting or subclassification methods, and its sampling variance is not readily available from M–N–m or M–C–m. However, both the estimand and its sampling variance are readily available from MITSS. MITSS’s coverage of this estimand was higher than nominal (97.7%) with bias that was less than 0.02. This type of analysis allows the identification of subgroups for which the active treatment could be harmful relative to the control.
6. Discussion
The simulation demonstrates that, in the configurations examined, for both scalar and multivariate X with overlapping distributions in the treatment and control groups, MITSS is a generally valid procedure that is more accurate and precise than the other valid procedures. However, when there are non-overlapping units, the splines in MITSS are implicitly used to extrapolate to units with covariate values that are not observed for that treatment, which is known to be ‘dangerous’ [47], leading to increased bias. Attempting to address this issue using more complex splines (e.g., MITSS-15) results in better coverage, but generally larger biases, interval widths, and RMSEs. This conclusion agrees with the previous literature, which suggests that there are only small gains in terms of bias reduction when more than six subclasses/knots are used [3,48]. Of course, the results here are confined to the sample sizes, and simulation configurations examined.
Although the general statistical validity of interval estimates obtained using MITSS arises from smaller biases and smaller RMSEs, it is not the most efficient procedure in all configurations. For example, in some cases, scientific knowledge may support the claim that Y(0) and Y(1) are nearly parallel linear combinations of X, which can be estimated accurately and efficiently using multiple linear regression. Other scientific knowledge may be that Y(0) and Y(1) are monotone in X, and more efficient estimators can be obtained by better approximating the response surfaces in step 2 of MITSS (e.g., [49]). This type of knowledge is likely to be more recondite when the dimension of X is larger. Another option is to assume that the values of the missing potential outcomes are all represented in the empirical distribution of the observed potential outcomes and employ predictive mean matching [50,51] for approximating the response surfaces. Nevertheless, any scientific knowledge used in situations where the distributions of X in the control and treatment groups do not overlap should be treated with extreme scrutiny, because model-based extrapolation is inevitably involved.
When it is impossible to obtain full overlap of the covariate distributions between the treatment group and the control group, no method can provide generally valid statistical inferences. In such cases, the investigator should consider discarding some of the units that do not overlap, so that a valid inference on a restricted population can be obtained [1]. The general advice is that such restrictions should be made with scientific knowledge, which is usually study specific. Here, we examined a rule that discards observations for which the propensity scores are in the intersection of the range of the two treatment groups and showed that MITSS was generally valid.
The results presented here for scalar X and continuous outcome parallel the results obtained for scalar X and dichotomous outcomes [12]. Because the same trends were also observed here for multivariate X, we conjecture that MITSS can be generally applied to both dichotomous and continuous outcomes with scalar or multivariate X.
Another advantage of MITSS over the previously proposed procedures is the freedom to define relevant estimands, such as finite population ones, non-linear ones, and estimands that are based on multivariate potential outcomes. Finite population estimands are sometimes of interest when a possible intervention is to be applied to all of the units in the population (e.g., the 50 states in the United States and all of the nursing homes in the United States). Because MITTS assumes that Yi(0) and Yi(1) are independent given X, the finite sample interval estimates may be too short when this assumption is not approximately correct. One possible solution is to include more covariates, including the second-order interaction terms. The selection of these terms should be performed before any examination of the observed outcomes, so that the design and analysis phases are performed separately. A different approach in situations where there is a priori knowledge that the treatment effect is constant is to change step 2 of MITSS and use a single spline with constant treatment effect instead of the two independent models. However, constant treatment effects are not always realistic, because it is hard to believe that units with different initial characteristics will respond to the active treatment versus control treatment in the same way. This situation is more prevalent in studies with heterogeneous units. A more reasonable solution is to use the sampling variance of the super-population estimate. The super-population sampling variance is equal to the largest finite sample variance in completely randomized experiments [52]. This approach was used throughout this manuscript and had good frequentist properties. Further discussion of this point with Shadish’s data simulation is discussed in the Supporting Information.
In conclusion, multiply imputing the potential outcomes and discarding observations that do not overlap in terms of covariates result in efficient inferences for the treatment effect.
Supplementary Material
Footnotes
Additional supporting information may be found in the online version of this article at the publisher’s web-site. The on-line supplemental file is composed of four main parts. The first part includes additional discussion on finite sample estimands. The second part expands on the specific parameters used in the implementation of MITSS in this paper. The third part includes description of the algorithm that was used to generate multivariate response surfaces. The fourth part includes additional tables and figures.
References
- 1.Rubin DB. The design versus the analysis of observational studies for causal effects: parallels with the design of randomized trials. Statistics in Medicine. 2007;26:20–36. doi: 10.1002/sim.2739. [DOI] [PubMed] [Google Scholar]
- 2.Rubin DB. For objective causal inference, design trumps analysis. The Annals of Applied Statistics. 2008;2(3):808–840. [Google Scholar]
- 3.Cochran WG. The effectiveness of adjustment by subclassification in removing bias in observational studies. Biometrics. 1968;24:295–313. [PubMed] [Google Scholar]
- 4.Rubin DB. Matching to remove bias in observational studies. Biometrics. 1973;29:159–183. [Google Scholar]
- 5.Rubin DB. Using multivariate matched sampling and regression adjustment to control bias. Journal of the American Statistical Association. 1979;74:318–328. [Google Scholar]
- 6.Rosenbaum PR, Rubin DB. The central role of the propensity score in observational studies for causal effects. Biometrika. 1983;70:41–55. [Google Scholar]
- 7.Rubin DB, Thomas N. Combining propensity score matching with additional adjustments for prognostic covariates. Journal of the American Statistical Association. 2000;95(450):573–585. [Google Scholar]
- 8.Bang H, Robins JM. Doubly robust estimation in missing data and causal inference models. Biometrics. 2005;61:962–973. doi: 10.1111/j.1541-0420.2005.00377.x. [DOI] [PubMed] [Google Scholar]
- 9.Lunceford JK, Davidian M. Stratification and weighting via the propensity score in estimation of causal treatment effects: a comparative study. Statistics in medicine. 2004;23(19):2937–2960. doi: 10.1002/sim.1903. [DOI] [PubMed] [Google Scholar]
- 10.Gutman R, Rubin DB. Comparing procedures for causal effect estimation with binary treatments in unconfounded studies with one continuous covariate. Statistical Methods in Medical Research. 2015 doi: 10.1177/0962280215570722. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Abadie A, Imbens GW. Large sample properties of matching estimators for average treatment effects. Econometrica. 2006;74(1):235–267. [Google Scholar]
- 12.Gutman R, Rubin DB. Robust estimation of causal effects of binary treatments in unconfounded studies with dichotomous outcomes. Statistics in Medicine. 2013;32(11):1795–1814. doi: 10.1002/sim.5627. [DOI] [PubMed] [Google Scholar]
- 13.Rubin DB. Bayesian inference for causal effects: the role of randomization. Annals of Statistics. 1978;6:34–58. [Google Scholar]
- 14.Rubin DB. Inference and missing data. Biometrika. 1976;63(3):581–592. [Google Scholar]
- 15.Little RJA, An H. Robust likelihood-based analysis of multivariate data with missing values. Statistica Sinica. 2004;14:933–952. [Google Scholar]
- 16.Neyman J. Sur les applications de la thar des probabilities aux experiences agaricales: Essay de principle. English translation of excerpts by Dabrowska, D. and Speed, T (1990) Statistical Science. 1923;5:465–472. [Google Scholar]
- 17.Rubin DB. Estimating causal effects of treatments in randomized and nonrandomized studies. Journal of Educational Psychology. 1974;66(5):688–701. [Google Scholar]
- 18.Rubin DB. Comment on randomization analysis of experimental data: The fisher randomization test. Journal of the American Statistical Association. 1980;75(371):591–593. [Google Scholar]
- 19.Rubin DB. Formal modes of statistical inference for causal effects. Journal of Statistical Planning and Inference. 1990;25:279–292. [Google Scholar]
- 20.Stuart EA. Matching methods for causal inference. Statistical Science. 2010;25:1–21. doi: 10.1214/09-STS313. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Rubin DB. Multiple Imputation for Nonresponse in Surveys. John Wiley & Sons Inc; New York, NY: 1987. [Google Scholar]
- 22.Rubin DB. Direct and indirect causal effects via potential outcomes. Scandinavian Journal of Statistics. 2004;31(2):161–170. [Google Scholar]
- 23.Abadie A, Imbens GW. Bias-corrected matching estimators for average treatment effects. Journal of Business & Economic Statistics. 2011;29(1):1–11. [Google Scholar]
- 24.Hansen BB. Full matching in an observational study of coaching for the SAT. Journal of the American Statistical Association. 2004;99(467):609–618. [Google Scholar]
- 25.Stuart EA, Green KM. Using full matching to estimate causal effects in nonexperimental studies: examining the relationship between adolescent marijuana use and adult outcomes. Developmental Psychology. 2008;44(2):395–406. doi: 10.1037/0012-1649.44.2.395. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Zhang G, Little RJA. Extensions of the penalized spline of propensity prediction method of imputation. Biometrics. 2009;65(3):911–918. doi: 10.1111/j.1541-0420.2008.01155.x. [DOI] [PubMed] [Google Scholar]
- 27.McCullagh P, Nelder JA. Generalized Linear Models. 2. Chapman and Hall; London: 1989. [Google Scholar]
- 28.Rubin DB. Using propensity scores to help design observational studies: application to the tobacco litigation. Health Services and Outcomes Research Methodology. 2001;2(3):169–188. [Google Scholar]
- 29.Rubin DB. The use of propensity score in applied Bayesian inference. In: Bernardo JM, De Groot MH, Lindley DV, Smith AFM, editors. Bayesian Statistics. Vol. 2. Elsevier Science Publisher B. V; North-Holland: 1985. pp. 463–472. [Google Scholar]
- 30.Little RJA, Rubin DB. Statistical Analysis with Missing Data, Second Edition. Wiley-Interscience; Hoboken: 2002. [Google Scholar]
- 31.Rubin DB. Statistical inference for causal effects, with emphasis on applications in epidemiology and medical statistics. Handbook of Statistics: Epidemiology and Medical Statistics. 2008;27:28–62. [Google Scholar]
- 32.Kang JDY, Schafer JL. Demystifying double robustness: a comparison of alternative strategies for estimating a population mean from incomplete data. Statistical Science. 2007;4:1–18. doi: 10.1214/07-STS227. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Waernbaum I. Model misspecification and robustness in causal inference: comparing matching with doubly robust estimation. Statistics in Medicine. 2012;31(15):1572–1581. doi: 10.1002/sim.4496. [DOI] [PubMed] [Google Scholar]
- 34.Imbens G, Rubin DB. Causal Inference in Statistics, and in the Social and Biomedical Sciences. Cambridge University Press; New York, NY: 2015. [Google Scholar]
- 35.Barnard J, Rubin DB. Small-sample degrees of freedom with multiple imputation. Biometrika. 1999;86(4):948–955. [Google Scholar]
- 36.Lin J. Divergence measures based on Shannon entropy. IEEE Transactions on Information Theory. 1991;37:145–151. [Google Scholar]
- 37.Kullback S, Leibler RA. On information and sufficiency. The Annals of Mathematical Statistics. 1951;22:79–86. [Google Scholar]
- 38.Imai K, Ratkovic M. Covariate balancing propensity score. Journal of the Royal Statistical Society, Series B (Statistical Methodology) 2014;76(1):243–263. [Google Scholar]
- 39.Crump RK, Hotz VJ, Imbens GW, Mitnik OA. Dealing with limited overlap in estimation of average treatment effects. Biometrika. 2009;96(1):187–199. [Google Scholar]
- 40.Azzalini A, Capitanio A. Distributions generated by perturbation of symmetry with emphasis on a multivariate skew t-distribution. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 2003;65(2):367–389. [Google Scholar]
- 41.Rubin DB. Multivariate matching methods that are equal percent bias reducing, II: maximums on bias reduction for fixed sample sizes. Biometrics. 1976:121–132. [Google Scholar]
- 42.Rubin DB, Thomas N. Affinely invariant matching methods with ellipsoidal distributions. The Annals of Statistics. 1992:1079–1093. [Google Scholar]
- 43.Silverman BW. Density Estimation. Chapman & Hall/CRC; London: 1986. [Google Scholar]
- 44.Shadish WR, Clark MH, Steiner PM. Can nonrandomized experiments yield accurate answers? A randomized experiment comparing random and nonrandom assignments. Journal of the American Statistical Association. 2008;103(484):1334–1344. [Google Scholar]
- 45.Hamming RW. Error detecting and error correcting codes. Bell System technical journal. 1950;29(2):147–160. [Google Scholar]
- 46.Mahalanobis PC. On the generalized distance in statistics. Proceedings of the National Institute of Sciences of India; New Delhi. 1936. pp. 49–55. [Google Scholar]
- 47.Hastie T, Tibshirani R, Friedman JJ. The Elements of Statistical Learning. Vol. 1. Springer; New York: 2001. [Google Scholar]
- 48.Rosenbaum PR, Rubin DB. Reducing bias in observational studies using subclassification on the propensity score. Journal of the American Statistical Association. 1984;79:516–524. [Google Scholar]
- 49.Shively TS, Sager TW, Walker SG. A Bayesian approach to non-parametric monotone function estimation. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 2009;71(1):159–175. [Google Scholar]
- 50.Rubin DB. Statistical matching using file concatenation with adjusted weights and multiple imputations. Journal of Business & Economic Statistics. 1986;4(1):87–94. [Google Scholar]
- 51.Little RJA. Missing-data adjustments in large surveys. Journal of Business & Economic Statistics. 1988;6(3):287–296. [Google Scholar]
- 52.Rubin DB. Comment: Neyman (1923) and causal inference in experiments and observational studies. Statistical Science. 1990;5(4):472–480. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.