Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2019 Feb 1.
Published in final edited form as: Prev Sci. 2018 Feb;19(Suppl 1):95–108. doi: 10.1007/s11121-017-0760-x

Limitations in using multiple imputation to harmonize individual participant data for meta-analysis

Juned Siddique 1, Peter J de Chavez 2, George Howe 3, Gracelyn Cruden 4, C Hendricks Brown 5
PMCID: PMC5572105  NIHMSID: NIHMS855914  PMID: 28243827

Abstract

Individual participant data (IPD) meta-analysis is meta-analysis in which the individual level data for each study are obtained and used for synthesis. A common challenge in IPD meta-analysis is when variables of interest are measured differently in different studies. The term harmonization has been coined to describe the procedure of placing variables on the same scale in order to permit pooling of data from a large number of studies. Using data from an IPD meta-analysis of 19 adolescent depression trials, we describe a multiple imputation approach for harmonizing 10 depression measures across the 19 trials by treating those depression measures that were not used in a study as missing data. We then apply diagnostics to address the fit of our imputation model. Even after reducing the scale of our application, we were still unable to produce accurate imputations of the missing values. We describe those features of the data that made it difficult to harmonize the depression measures and provide some guidelines for using multiple imputation for harmonization in IPD meta-analysis.

Keywords: multiple imputation, data synthesis, individual participant data meta-analysis, posterior predictive checking


In response to limitations imposed by traditional meta-analysis, an increasingly popular approach for data synthesis is individual participant data (IPD) meta-analysis in which the raw individual level data for each study are obtained and used for synthesis (Riley, Lambert, & Abo-Zaid 2010). With the raw data in hand, an analyst can adjust for patient-level covariates and take into account repeated measures, missing values, and differential follow-up times. In general, pooling data from multiple studies results in larger sample sizes, increased statistical power, increased variability on important measures, and the capacity to test more sophisticated models (Brown et al. 2016). When the combined samples are more heterogenous than any single trial, IPD meta-analysis may also provide increased confidence in generalization (Perrino et al. 2013).

Another advantage of IPD meta-analysis is increased frequencies of low base-rate behaviors such as suicide or drug use. The frequency of these behaviors may be too low to be modeled in any single study, but may be high enough when aggregated across multiple studies. When multiple longitudinal studies are combined, a much broader developmental period can be considered, given overlapping age ranges across the set of contributing studies (Curran & Hussong 2009). IPD meta-analysis can also substantially increase power to detect moderation. Dagne, Brown, Howe, Kellam, and Liu (2016) found that the power to detect moderator effects for individual-level moderators could be as much as 16 times greater for IPD meta-analysis as compared to standard meta-regression.

IPD meta-analysis has its challenges. In particular, a common situation is when variables of interest are measured differently in different studies. The term harmonization has been coined to describe the procedure of placing variables on the same scale in order to permit pooling of data from a large number of studies (Griffith et al. 2015, Hussong, Curran, & Bauer 2013).

There are a number of existing methods for data harmonization which make use of the fact that even if different studies use different outcomes, they are attempting to measure the same construct or constructs of interest. One approach is to treat the unobserved measures as missing data and then replace them with plausible values using multiple imputation (Gelman, King, & Liu 1998, Kline, Andridge, & Kaizar 2015, Resche-Rigon, White, Bartlett, Peters, & Thompson 2013, Rubin 1987, Siddique et al. 2015).

With multiple imputation, missing values are replaced with two or more plausible values to create two or more completed data sets. Analyses are then conducted separately on each data set and final estimates are obtained by combining the results from each of the imputed data sets using rules that account for within-imputation and between-imputation variability. See Harel and Zhou (2007) for a review.

In the context of harmonization for IPD meta-analysis, multiple imputation has a number of advantages. Once unmeasured variables have been imputed, analyses and their subsequent inferences are based on existing scales of interest. In addition, after the data set has been filled in, it can be shared with other investigators and can be used for numerous analyses using complete data methods. In fact, once a variable has been multiply imputed, it may be used as an outcome in one analysis, and as a covariate in another analysis.

Siddique et al. (2015) describe an imputation-based approach to harmonize outcome measures across 5 longitudinal depression trials where there is no overlap in outcome measures within trials. They extend previous methods for harmonization by addressing harmonization in a longitudinal setting where different studies have different follow-up times and the relationships between outcomes may change over time. They also discuss practical issues in developing an imputation model including taking into account treatment group and study and develop diagnostics for checking the fit of the imputation model.

In this article we describe a multiple imputation approach for harmonizing depression measures across 19 longitudinal intervention trials where there is no single outcome measure used by all 19 trials. We use the methods of Siddique et al. (2015) for harmonization in a longitudinal setting in order to account for differential follow-up times between studies and to account for the fact that the relationships between outcome variables may change over time. This paper extends the work of Siddique et al. (2015) by implementing the methods in a more challenging setting with 10 measures sparsely distributed across 19 heterogeneous trials. None of the trials used in Siddique et al. (2015) are among the 19 trials in this paper. We implement our methods using free and easily-available software and highlight those conditions where it is not possible to produce accurate imputations, either due to an inability to estimate the parameters in the imputation model or due to an inability to estimate study-specific effects.

This article is organized as follows. In Section 2, we describe the example that motivated this work, a study of 19 randomized trials for the prevention of depression among adolescents. In Section 3, we describe our imputation model and diagnostics for checking the quality of imputations when variables are missing for all participants within a study. Section 4 presents the results of applying our methods to the adolescent data and Section 5 offers discussion and areas for future work.

Motivating Example

Our motivating example is an ongoing IPD meta-analysis investigating moderators of treatment effectiveness for the prevention of depression among adolescents. The project consists of individual participant data from 19 adolescent depression prevention trials. In 9 of the 19 trials, the intervention was intended to specifically target youth depression. In the remaining 10 trials, the focus of the interventions were family-based interventions for behavioral health promotion, and for substance abuse and HIV/AIDS sexual risk behavior prevention. Each trial was an RCT with both an intervention and a control group. More details regarding this project are described in the accompanying article by Brown et al. (2016) in this same issue.

Table 1 lists the 19 trials and various study characteristics. The trials ranged in size from 41 to 697 participants, and were roughly half male and half female. Participants were mostly teenagers, with an age range of 7 to 21 years of age. Ten of the 19 studies were longer than two years, but for this IPD meta-analysis we only use data from the first two years of each trial. Based on these data, the average number of assessments (including baseline) was 4, and trial duration ranged from 6 to 24 months with an average duration of 17 months.

Table 1.

Trial names and trial-level descriptive statistics for the 19 adolescent depression trials consisting of 5547 participants.

Trial ID Trial Name N Percent Female Age (range) No. Assess Duration (months) No. Measures

1 New Beginnings 240 49 10 (9–12) 4 9 4
2 Family Bereavement 245 47 11 (7–16) 3 14 4
3 Familias Unidas I 258 53 13 (12–16) 4 24 1
6 Familias Unidas DJJ 242 36 15 (12–17) 3 12 3
8 Bridges (Puentes) 542 51 12 (11–14) 3 14 5
10 ADEPT 94 60 15 (13–18) 4 24 2
12 Project Alliance 2 592 48 12 (10–14) 3 24 2
14 Familias Unidas CDC 160 49 15 (14–18) 4 12 3
17 Family Talk 135 43 12 (9–15) 3 18 1
18 CATCH IT 83 57 17 (14–21) 6 12 2
28 Penn Resiliency Program (PRP) I 697 46 12 (9–15) 6 24 1
49 Prevention of Depression Study (PODS) 316 59 15 (13–17) 3 6 3
50 K-IPT AST 57 60 15 (13–17) 6 18 3
61 Narsad IPT 41 85 13 (11–16) 4 6 3
78 Familias Unidas II 213 36 14 (12–16) 3 15 3
84 Project Alliance 1 179 55 12 (11–14) 3 24 6
98 Preparing for the Drug Free Years (PDFY) and Iowa Strengthening Families Program (ISP) 667 52 11(10–13) 3 15 2
247 IPT-AST vs CBT 379 55 14 (13–18) 5 18 1
698 Penn Resiliency Program (PRP) II 407 48 12 (10–15) 6 24 1

The last column in Table 1 lists the number of depression measures used in each trial. While some trials only used one depression measure, most used more than one and the Project Alliance 1 trial used 6 depression measures.

Table 2 lists the 10 different depression measures used in each of the 19 studies and their average values at baseline. Several important points are worth noting. First, there is no measure that is used by all the trials. Second, some measures are self-reported (denoted by a (S) after the measure), some measures are parent-reported (denoted by a (P)), and one measure is clinician-rated (denoted by a (C)). The third point is that while 10 measures are listed, several of the measures are subscales of a larger measure. The Child Behavior Checklist (CBCL) Anxious/Depressed subscale (CBCL-A), the CBCL Withdrawn/Depressed subscale (CBCL-W) (Achenbach 1991), and the CBCL Depression scale (CBCL-D) (Clarke, Lewinsohn, Hops, & Seeley 1992) are all derived from the CBCL. Hence, all three of these parent-reported subscales are often measured in the same trials. Similarly, the Youth Self Report (YSR) Anxious/Depressed subscale (YSR-A) and the YSR Withdrawn/Depressed subscale (YSR-W) (Ebesutani, Bernstein, Martinez, Chorpita, & Weisz 2011) are both derived from the YSR and tend to be used in the same trials. Trials that used the YSR also tended to use the Revised Brief Problem Checklist Anxiety/Withdrawl subscale (RBPC) (Quay & Peterson 1996).

Table 2.

Baseline means and missing data patterns of the 19 adolescent depression trials. Measures followed by an (S) are self-reported measures. Those followed by a (P) are parent-reported measures. Those followed by a (C) are clinician-rated.

Trial ID CDI(S) CBCL-A(P) CBCL-W(P) CBCL-D(P) YSR-A(S) YSR-W(S) RBPC(P) CESD(S) CESD10(S) CDRS(C)
1 5.81 5.30 2.69 4.33
2 9.74 4.74 2.95 4.68
3 3.29
6 4.70 3.82 5.68
8 3.82 2.95 3.33 5.73 3.93
10 7.76 24.45
12 7.49
14 5.00 3.81 5.35
17 5.68
18* 22.38 12.19
28 8.78
49 15.66 8.89 28.81
50 26.37 14.02 28.14
61 25.22 13.15 27.88
78 4.96 4.24 6.55
84 9.38 2.99 2.25 3.58 5.75 3.92
98 4.49
247 9.93 17.20 9.62
698 10.88

CDI: Children’s Depression Inventory (self-reported)

CBCL-A: Child Behavior Checklist, Anxious/Depressed Subscale (parent-reported)

CBCL-W: Children’s Depression Inventory, Withdrawn/Depressed Subscale (parent-reported)

CBCL-D: Child Behavior Checklist Depression Scale (parent-reported)

YSR-A: Youth Self-Report, Anxious/Depressed Subscale (self-reported)

YSR-W: Youth Self-Report, Withdrawn/Depressed Subscale (self-reported)

RBPC: Revised Brief Problem Checklist, Anxiety/Withdrawal Subscale (parent-reported)

CESD: Center for Epidemiological Studies Depression Scale (self-reported)

CESD10: Center for Epidemiological Studies Depression Scale, 10 items only (self-reported)

CDRS: Children’s Depression Rating Scale (clinician-rated)

*

The CATCH IT Trial only used the full CESD at baseline

Table 2 also includes two measures, The Center for Epidemiologic Studies Depression Scale (CESD) (Eaton, Smith, Ybarra, Muntaner, & Tien 2004, Radloff 1977, 1991) and also what is referred to as the CESD10. For some follow-up time points, the CATCH IT trial (trial 18) only used 10 items from the CESD, and we refer to this measure as the CESD10 (Radloff 1977). Since the CESD10 is a subset of the CESD, we are able to calculate the CESD10 for all studies that used the CESD except for the ADEPT trial (trial 10) for which item level CESD data were not available. For those follow-up occasions where the CATCH-IT trial did not use the full CESD, we treat the CESD as missing data while recognizing that the CESD and the CESD10 are highly correlated and that we have 4 studies that contain both these measures.

Roughly speaking, the 19 trials can be placed into three categories based on which depression measures they used: 1) those trials that use the Children’s Depression Inventory (CDI) (Helsel & Matson 1984, Kovacs 1984) and the CBCL; 2) those that use the YSR; 3) those that use the CESD and the Children’s Depression Rating Scale (CDRS) (Mayes, Bernstein, Haley, Kennard, & Emslie 2010, Poznanski, Freeman, & Mokros 1985). Our imputation procedure relies on our ability to estimate the relationship among all these variables. In this regard, Trials 8, 10, 84, and 247 are particularly important because they provide connections between the three groups of measures. Trial 8 uses the CBCL and the YSR and Trial 84 uses the CDI, CBCL, and the YSR. Together, these two trials link categories 1 and 2. Trial 247 links categories 1 and 3 through the CDI and CESD and Trial 10 links categories 2 and 3 through the CBCL and CESD. Still, as highlighted by the shaded cells in Table 2, there is a great deal of sparseness in our data set. If we think of Table 2 as an 19 × 10 matrix, then only 50 of the 190 cells are filled-in. As will be shown, this sparseness will prevent us from filling in all missing cells accurately.

Methods

Our approach for harmonizing the depression data across the multiple trials follows that of Siddique et al. (2015) where the uncollected depression measures are considered missing data and missing observations are multiply imputed. To check the quality of our imputations, we perform diagnostics using the re-imputation strategy of He and Zaslavsky (2012) in which observed data are deleted and then imputed and quantities based on imputed values are compared to the same quantities using observed values.

Set Up

We begin by assembling the data in a vertical (long) format, so that each row represents a single participant at a single point in time. Columns are time, demographics, and the 10 different depression measures used across all trials. To account for skewness in outcomes and non-linear trends over time, all depression measures were transformed using a square root transformation. Imputations were also performed on the original scale. Time, measured as the number of months since baseline, was log transformed. Once imputation is complete, all depression measures are back-transformed to their original distributions.

Imputation Model

Our imputation model is a multivariate linear mixed-effects regression model as described by Schafer and Yucel (2002) and implemented in the R (R Core Team 2012) package PAN (Zhao & Schafer 2013). This model was used by Siddique et al. (2015) to harmonize multiple depression measures in a IPD setting. Using notation from Schafer and Yucel (2002), let yi denote an ni × r matrix of multivariate data for participant i, i = 1, …, m, where each row of yi is a joint realization of depression measures Y1, …, Yr, which are measured ni times. We assume that yi follows a multivariate linear mixed-effects model of the form

yi=Xiβ+Zibi+εi (1)

where Xi (ni × p) and Zi (ni × q) are known covariate matrices, β (p × r) is a matrix of regression coefficients common to all units (the “fixed effects”) and bi (q x r) is a matrix of coefficients specific to unit i (the “random effects”). We assume the ni rows of the error terms εi are independently normally distributed as N (0, Σ) and the random effects are distributed as vec(bi) ~ N (0, Ψ) (where the “vec” operator vectorizes a matrix by stacking its columns). In our model, fixed effects include an intercept term, months since baseline (log-transformed), the square of log transformed months since baseline, gender, and age. Random effects initially included an intercept term and a random months since baseline term.

Imputations of the missing components of yi are generated by drawing from the posterior predictive distribution of the missing data P(Ymis|Yobs). PAN does this using Markov Chain Monte Carlo (MCMC) (Schafer & Yucel 2002), which requires the specification of prior distributions for the parameters in the imputation model in Equation 1. Here, we use noninformative priors for both the fixed effects and random effects. Specifically, we assume an improper uniform density for the regression coefficents β and non-informative inverse-Wishart priors for the covariance matrix of the random effects and the error variance with with r × 2 and r degrees of freedom, respectively, and scale parameters equal to the identity matrix. We assessed convergence of our Markov chains by visual inspection of trace plots and autocorrelation plots as well as by using formal MCMC diagnostics (Cowles & Carlin 1996).

Imputations were performed separately by treatment group so that all of the parameters in Equation 1 can vary by treatment group. Since both sets of imputations (those based on untransformed values and those based on square root transformed values) assume the data are continuous, once missing values were imputed, we considered two strategies to put imputed values back on an ordinal scale: 1) rounding, where values were rounded to the nearest possible value; and 2) leaving imputed values as continuous which means that negative values remain negative, even though all of the scales in Table 1 are non-negative. Strategy 2 is motivated by research showing that when imputing limited-range variables it may be best to allow imputed values to remain out of range (Rodwell, Lee, Romaniuk, & Carlin 2014).

Associations among the depression measures

Fitting the parameters of the model in Equation 1 requires estimation of the association among all the measures listed in Table 2. That is, for every possible pair of measures, there must be at least one trial in which both measures are given to the same participants. For example, in order to estimate the pairwise association of the CDI with the other variables in Table 2 we can use Trials 1, 2, and 84 to estimate most of these associations. More problematic is the CDRS which is used in only 3 trials, and overlaps only with the CESD. So while we are able to estimate the relationship between the CESD and the CDRS, we cannot estimate the relationship of the CDRS with any of the other depression measures. This will ultimately prevent us from accurately harmonizing the CDRS across all 19 trials. Since the CDRS does not provide any information on the relationship between itself and any other variables besides the CESD, we have dropped the CDRS from our list of variables to be harmonized.

Similarly, the RBPC is either measured by itself (Trial 3) or with the YSR. In fact, in Trial 3 (Familias Unidas 1), the RBPC is the only measure given to participants. Thus, the only relationship we can measure using the RBPC is the relationship between the RBPC and the YSR. However, most trials that use the RBCP also use the YSR, and vice-versa. Furthermore, the YSR itself lacks overlap with most of the other measures. This is evident in Table 3 which displays the correlation matrix of all the depression measures. The numbers in parentheses below the correlations on the diagonal report the number of trials which use each measure. The numbers in parentheses below the correlations on the off-diagonal report the number of trials which use both of the measures listed in the row and column. For example, eight trials use the CDI and one trial uses both the CDI and the CESD whose correlation is 0.81. The shaded cells in Table 3 identify those pairs of measures in which there is no overlap. The YSR does not overlap with the CESD in any trial, so that it cannot be harmonized in those four trials that use the CESD and nothing else (Trials 18, 49, 50, 61). For this reason, and for the additional reason the the RBPC has very low correlation with the YSR, we also drop the RBPC from our list of measures to be harmonized. This has the undesirable consequence of requiring us to also remove Trial 3 from the 19 trials we wish to synthesize. A similar decision to drop Trial 3 was made by Brown et al. (2016) in the companion paper in this issue.

Table 3.

Correlation matrix of all depression measures at baseline across the 19 trials. The number in parentheses under the correlation is the number of trials which used the depression measure (diagonal) or the number of trials using both measures (off-diagonal). Measures followed by an (S) are self-reported measures. Those followed by a (P) are parent-reported measures. Those followed by a (C) are clinician-rated.

CDI(S) CBCL-A(P) CBCL-W(P) CBCL-D(P) YSR-A(S) YSR-W(S) RBPC(P) CESD(S) CESD10(S) CDRS(C)
CDI(S) 1.00 (8) 0.21 (3) 0.27 (3) 0.28 (3) 0.56 (1) 0.43 (1) NA (0) 0.81 (1) 0.77 (1) NA (0)
CBCL-A(P) 1.00 (4) 0.53 (4) 0.77 (4) 0.18 (2) 0.13 (2) NA (0) NA (0) NA (0) NA (0)
CBCL-W(P) 1.00 (4) 0.65 (4) 0.07 (2) 0.21 (2) NA (0) NA (0) NA (0) NA (0)
CBCL-D(P) 1.00 (5) 0.16 (2) 0.17 (2) NA (0) −0.03 (1) NA (0) NA (0)
YSR-A(S) 1.00 (6) 0.67 (5) 0.10 (3) NA (0) NA (0) NA (0)
YSR-W(S) 1.00 (5) 0.16 (3) NA (0) NA (0) NA (0)
RBPC(P) 1.00
4
NA (0) NA (0) NA (0)
CESD(S) 1.00
6
0.95 (5) 0.28 (3)
CESD10(S) 1.00 (5) 0.27 (3)
CDRS(C) 1.00 (3)

Looking at Table 3, only the CDI emerges as a potential variable for harmonization. The CDI overlaps with all of our remaining measures. The correlations between the CDI and all three CBCL subscales are low, but that is a nature of the two measures, one being self-report, the other parent report. Even if we did have item-level data from the ADEPT Trial (Trial 10) which would allow us to calculate the CESD10 for this trial (and thus estimate the association with the CBCL-D and the CESD10), the CBCL-D is not a good target for harmonization due to its low correlation with the CESD. Thus, while we will impute all of the depression measures using the model in Equation 1, we focus our attention on the CDI since it is the only outcome that has the potential to be imputed accurately.

Besides the large number of NA’s which indicate that the correlation was not estimable because there were no trials which used both measures on the same participants, what is notable about Table 3 is how low the correlations are. While the correlations among subscales of the same measures are moderately high, correlations across different measures are relatively low, especially considering that many of these scales are presumably measuring the same construct. This is likely due to the fact that the measures in Table 3 are a collection of self-reported, parent-reported, and clinican-rated measures. For example, the correlation between the CBCL-D (parent-reported) and the CDI (self-reported) is 0.28. And the correlation between the CBCL-A (parent-reported) and the YSR-A (self-reported) is 0.18. Most notable, the correlation between the CBCL-D (parent-reported) and the CESD (self-reported) is negative and equal to −0.03. The correlations of the RBPC (parent-reported) with the YSR-A (self-reported) and YSR-W (self-reported) are also low, 0.10 and 0.16 respectively. The only large correlation between two different scales is that of the CDI and CESD which are both self-reported and whose baseline correlation is 0.81.

Each depression measure is imputed based on a regression which conditions on the remaining depression measures. Thus, when imputing the CDI, not only do we need to be able to measure the pairwise association of the CDI with the other depression measures, we must also be able to measure the association of the other depression measures with themselves. This second condition is slightly problematic because, as mentioned before, not all the measures overlap with each other. The CBCL-A and the CBCL-W do not overlap with the CESD. For this reason, we also remove the CBCL-A and the CBCL-W from our measures to be imputed. The YSR-A and the YSR-W also do not overlap with the CESD, but we cannot remove both these measures because they are the only measures used by Trials 6, 14, 78, and 98. However, the partial correlation of the CDI and the YRS-W, controlling for the YSR-A is only 0.07 at baseline. Therefore, without much loss of information, we can also remove the YSR-W from our imputation model.

Table 4, is a revised version of Table 3, now only including those 5 measures and 18 studies in our reduced imputation model. The shaded cells are those measures with no overlap and thus inestimable covariances. In our Bayesian set-up, when parameters cannot be identified, their posterior distribution is equal to their prior distribution. Our non-informative inverse-Wishart prior for the covariance matrix sets these covariances to be centered around 0. If the unobserved correlations are small, the non-informative prior will have little effect on the resulting imputations. In the discussion, we describe alternative approaches for handling these inestimable parameters.

Table 4.

Correlation matrix of depression measures at baseline. Only the 5 measures and 18 studies in the reduced imputation model are included. The number in parentheses under the correlation is the number of trials which used the depression measure (diagonal) or the number of trials using both measures (off-diagonal). Measures followed by an (S) are self-reported measures. Those followed by a (P) are parent-reported measures.

CDI(S) CBCL-D(P) YSR-A(S) CESD(S) CESD10(S)
CDI(S) 1.00 (8) 0.28 (3) 0.56 (1) 0.81 (1) 0.77 (1)
CBCL-D(P) 1.00 (5) 0.16 (2) −0.03 (1) NA (0)
YSR-A(S) 1.00 (6) NA (0) NA (0)
CESD(S) 1.00 (6) 0.95 (5)
CESD10(S) 1.00 (5)

Imputation diagnostics

In our setting, where the amount of missing data is considerable, and where we are imputing values for every participant within a trial for some depression measures, it is particularly important to check the imputation model and the quality of its imputations. Here, we use posterior predictive checks using numerical summaries based on test statistics (Gelman, Meng, & Stern 1996). We focus on diagnostics that capture important features of the data that are relevant to our target analyses.

Our approach follows the posterior predictive checking and re-imputation strategy of He and Zaslavsky (2012). We do this by duplicating Trials 1, 2, 84, and 247 which contain the CDI and at least one of the other measures in Table 4. Next, we deleted all values of the CDI in these duplicated trials. Finally, we concatenated these deleted data sets with the original 18 trials treating the duplicated trials as if they were 4 additional trials. Table 5 describes the design of our re-imputation strategy. Note that a limitation of the strategy is that it does not allow us to investigate how well our imputation model imputes the CDI in those trials that do not use the CDI.

Table 5.

Design of the duplication and re-imputation strategy. Measures followed by an (S) are self-reported measures. Those followed by a (P) are parent-reported measures.

ID of Duplicated Trial Measures deleted in duplicate data set Measures remaining in duplicate data set

1 CDI(S) CBCL-D(P)
2 CDI(S) CBCL-D(P)
84 CDI(S) CBCL-D(P), YSR-A(S)
247 CDI(S) CESD(S), CESD10(S)

Next we generated imputations using the imputation model described above. Let Y be the observed values of the CDI from the duplicated data prior to deletion and Yimp the imputed version of the CDI in the duplicated data set. To compare observed data to imputed data, we use a test statistic, T(Y, θ), some scalar function of the data. Posterior predictive checking consists of comparing T(Y, θ) to the distribution of T(Yimp, θ) where T(Yimp, θ) is the test statistic based on imputed values of Y. Lack of fit of the imputed data to the observed data can be measured by the posterior predictive p-value (ppp), the probability that the imputed data are more extreme than the observed data, as measured by the test quantity (Gelman, Carlin, Stern, & Rubin 2004, Gelman et al. 1996).

A small ppp suggests that the proposed imputation model is not adequate to support the targeted post-imputation analysis (He & Zaslavsky 2012). We investigated 3 sets of test statistics that capture important relationships linked to our substantive analyses. These test statistics are: 1) the correlation between the CDI and an observed measure at each time point; 2) the means of the CDI at each time point; and 3) the slope of the control group, the treatment group, and the treatment effect from a regression model regressing the CDI values on (log) months since baseline.

Results based on application to adolescent data

We begin this section by presenting the results of the diagnostics to ensure that our imputations are reasonable and are replicating important relationships relevant to our target analyses. We then analyze the adolescent data using the CDI as our depression outcome of interest. First we only analyze those eight trials which used the CDI. We then analyze all 18 trials using both observed CDI and imputed CDI data.

MCMC diagnostics

The first step in evaluating our imputation model is checking the convergence of the Markov chain used to generate the imputations. We assessed convergence of our Markov chains by visual inspection of trace plots and autocorrelation plots. These diagnostics made it very apparent that we were not able to estimate all of the parameters in our imputation model. For many of the measures in Table 4 there was not enough overlap to measure correlations at both the within- and between-participant level. That is, we could not assume both the random effects and the error terms were correlated across measures. This caused us to consider a vastly reduced model in which the covariance matrix Ψ of the random effects vec(bi) ~ N(0, Ψ) is block diagonal such that the random effects across outcomes are independent (random effects within an outcome are still correlated). Thus, all the association between measures is via the error covariance. This simplified structure has a number of consequences. First, it assumes that the correlation between two outcomes over time is constant. This is a reasonable assumption as all the measures are presumably measuring the same construct of interest. But due to floor and ceiling effects (see for example (Siddique et al. 2015)) correlations between measures can change as a function of time. The second result is that different outcomes measured on the same participant at different times are assumed to be independent. For example, a participant’s value on the CDI at baseline is independent of their value on the CBCL-D (but not the CDI) at some follow-up time point.

MCMC diagnostics also suggested that our model did not have adequate data to estimate both random intercept and random slope terms for all measures, even with our block diagonal covariance structure. Thus, only random intercept terms were included in our imputation model. This has the result of assuming measure variances are constant over time.

We generated 500,000 parameter draws from our reduced imputation model from a single Markov chain. Trace plots and autocorrelation plots of those parameters associated with the CDI reflected convergence of the chain. Formal diagnostics based on the Geweke Diagnostic (Geweke 1992) and the Gelman-Rubin diagnostic (Gelman & Rubin 1992) based on 3 parallel chains also suggested convergence. After diagnosing convergence, we ran one of our chains for an additional 500,000 iterations and drew 100 imputations by drawing values from every 5,000 iterations.

Results from posterior predictive checking

Figures 1 and 2 (available online) display histograms of imputed and observed CDI values from the duplicated data sets. Figure 1 shows imputations based on the imputation model where depression measures were square root transformed prior to entering the imputation model. Imputed values were then squared. The panel on the left of Figure 1 is a histogram in which imputed values were squared but not rounded to the nearest observed value. As a result, there are a few imputed values greater than 54 which the maximum possible value on the CDI. The middle panel is a histogram in which imputed values were squared and rounded to the nearest observed value such that all values are within the range of the CDI (0 to 54). The panel on the right is a histogram of observed CDI values from the duplicated data.

Figure 2 are imputations based on the imputation model where depression measures were not transformed prior to entering the imputation model. The panel on the left of Figure 2 is a histogram in which imputed values were not rounded to the nearest observed value. As a result, there are negative imputed values which are not possible on the CDI. The middle panel is a histogram in which imputed values were rounded to the nearest observed value such that there is a spike at 0 from rounding negative imputed values to 0. The panel on the right is a histogram of observed CDI values from the duplicated data.

Although the imputed values in the middle panel of Figure 1 appear to best preserve the distribution of the observed data, this is not necessarily the goal of our imputation model. Instead we wish to preserve important features of the data that are relevant to our target analyses. To this end, we also performed posterior predictive checks of correlations, marginal means, and changes of CDI scores over time. Based on the results of these posterior predictive checks, we selected the imputation model where measures were imputed on their original scale and not rounded. Results from this model are presented below.

Table 6 displays the results of posterior predictive checks (based on 100 imputed data sets) for both control and treatment group participants of the correlation between the CDI and an observed measure in the duplicated trials described in Table 5. In both intervention groups, correlations at baseline and the first two follow-up time points were checked. As mentioned above, our imputation model assumes that partial correlations between any two measures is time-invariant. As a result, imputed analyses do not capture changes in correlations over time. Instead, correlations based on imputed values are averaged over time. For Trial 2, the correlation in the duplicated trial (with a deleted and imputed measure) is similar to the observed correlation and the two-sided posterior predictive p-values are larger than 0.05. For the remaining trials, the observed correlation and the correlation calculated using the duplicated data set are not similar and the posterior predictive p-values are small, sometimes even equal to 0. These results suggest that our imputation model is not preserving all the relationships among the data.

Table 6.

Posterior predictive checks of simple correlations at the first three time points in Trials 1, 2, 84, and 247. Results are based on imputed CDI values and their correlation with an observed measure. The CDI and CESD are self-reported measures. The CBCL-D is parent-reported.

Condition Trial ID Observed Measure Baseline Time 1 Time 2
Obs. Imp. ppp Obs. Imp. ppp Obs. Imp. ppp

Control 1 CBCL-D 0.23 0.06 0.16 0.29 0.09 0.10 0.15 0.05 0.42
2 CBCL-D 0.22 0.14 0.46 0.16 0.09 0.48 0.24 0.11 0.24
84 CBCL-D 0.50 0.16 0.00 0.19 0.12 0.64 −0.01 0.20 0.20
247 CESD 0.84 0.46 0.00 0.82 0.50 0.00 0.79 0.45 0.00
Treatment 1 CBCL-D 0.40 0.09 0.00 0.16 0.03 0.06 0.19 0.04 0.02
2 CBCL-D 0.19 0.11 0.46 0.14 0.05 0.26 0.05 0.08 0.68
84 CBCL-D 0.24 0.08 0.20 0.01 0.08 0.54 0.25 0.10 0.26
247 CESD 0.78 0.38 0.00 0.66 0.39 0.00 0.75 0.45 0.00

Table 7 displays the results for both control and treatment group participants of posterior predictive checks of the mean of the CDI in the duplicated trials at baseline and the first two follow-up time points. In both intervention groups, the results suggest that imputed means are inaccurate. There are two reasons for this innacuracy. The first is that failing to preserve relationships as demonstrated in Table 6 leads to inaccurate imputations. The second reason is that our model does not incorporate trial-level effects. For example, trials 1, 2, and 84 have similar baseline CBCL-D scores of 4.33, 4.68, and 3.58, respectively. However, their baseline CDI scores are not similar, 5.81, 9.74, and 9.38, respectively. Not accounting for trial-level effects in our imputation model results in pooling observations across trials. The result is that when imputing the CDI in Trial 1, the imputed values are skewed towards the CDI values in Trials 2 and 84 which are much larger than those in Trial 1.

Table 7.

Results from posterior predictive checks of CDI means at the first three time points for control and treatment group participants in Trials 1, 2, 84, and 247.

Condition Trial ID Baseline Time 1 Time 2
Obs. Imp. ppp Obs. Imp. ppp Obs. Imp. ppp

Control 1 5.33 7.64 0.00 4.51 7.14 0.00 3.68 6.56 0.00
2 9.57 8.77 0.32 8.32 7.52 0.36 8.02 7.01 0.24
84 9.46 9.35 0.88 9.53 8.39 0.20 11.02 7.90 0.00
247 10.86 10.65 0.74 12.24 11.07 0.02 10.49 9.95 0.42
Treatment 1 6.04 7.89 0.00 4.37 6.38 0.00 3.87 5.89 0.00
2 9.87 8.49 0.02 7.98 6.73 0.04 7.30 6.29 0.06
84 9.31 8.83 0.58 8.71 7.18 0.06 7.62 6.73 0.36
247 9.19 9.53 0.60 9.12 9.14 0.98 9.43 9.09 0.52

Table 8 displays the results of posterior predictive checks of the fixed coefficients of a random intercept and slope regression model of the imputed depression score as a function of log(number of months since baseline + 1) for each trial in Table 5. Results are for the control group slope, the treatment group slope, and the difference between the two slopes (i.e. the treatment effect). As with the marginal means and correlations, results are attenuated towards the average for all trials. However, because there is less treatment effect heterogenity in our data (at least for the 4 duplicated trials), treatment effects based on duplicated data are close to observed treatment effects and all posterior predictive p-values are large, the smallest being equal to 0.38. However, this result is not due to an imputation model that fits the data well. Instead, it is a fortunate result of Trials 1, 2, 84, and 247 having similar treatment effects.

Table 8.

Results from posterior predictive checks of intercept, slope, and difference between slopes in Trials 1, 2, 84, and 247.

Trial ID Control slope Tx slope Tx effect
Obs. Imp. ppp Obs. Imp. ppp Obs. Imp. ppp

1 −0.95 −0.56 0.16 −1.05 −0.93 0.54 −0.10 −0.37 0.38
2 −0.61 −0.60 0.94 −0.91 −0.85 0.66 −0.30 −0.24 0.90
84 0.30 −0.39 0.00 −0.37 −0.73 0.14 −0.68 −0.34 0.38
247 0.15 −0.23 0.00 −0.19 −0.47 0.06 −0.34 −0.24 0.68

Post-imputation analysis of adolescent trial data

Despite the findings from the imputation diagnostics, which suggested that our imputation model is not preserving important features of the data, we proceeded to discard the duplicated data and analyze the data from the 18 adolescent trials. We analyzed the CDI scores (both observed and imputed) as a function of treatment and time using the following random intercept and slope regression model. Let CDIijk be the CDI score for participant i at occasion j, j = 1, …, ni in trial k, k = 1, … 18. And let timeijk be the time since baseline and Ti a variable indicating whether participant i was randomized to the intervention or control group. Then our model is

CDIijk=β0+β1timeij+β2(timeijTi)+b0k+b0i+b1itimeijk+εijk. (2)

As in our imputation model, time has been transformed as log(months since baseline + 1). We did this so that we could model time linearly in order to simplify the presentation of our analyses and avoid having to include a quadratic effect for time in our model. The term b0k is a random trial effect with mean 0 and follows a normal distribution. The terms b0i and b1i are random intercept and slope terms, respectively, and follow a bivariate normal distribution, again with mean 0. The error term εijk also follows a normal distribution and is independent of the random effects.

In this model, inference focuses on the regression coefficient β2, the time by treatment interaction. This term is the difference in slopes between intervention and control groups. Table 9 presents the results of our analysis using only the observed CDI scores as well as using both observed and imputed CDI scores. Focusing on the treatment by time interaction in Table 9, the treatment effect is significant in both CDRS analyses. That is, overall, those who were assigned to one of the 8 preventive interventions had more improvement in symptoms than those assigned to the control condition. In terms of effect sizes, at 24-months, the effect size from the analysis which uses the observed data is −0.11. The effect size from the analyses which uses both observed and imputed data is −0.13. For the most part, there is very little difference between the two analyses. This result is not surprising. As was demonstrated, our imputation model was not able to incorporate information from other trials that did not use the CDI. Thus we see little difference between the two analyses. However, the variance components in the imputed analyses are smaller than those in the observed analyses. This likely reflects the fact that our imputation model did not include these between-person effects. Thus, random effect variances are smaller and residual variance is larger.

Table 9.

Observed-only and post-imputation analyses of CDI scores. Observed-only analyses are based on the 8 trials that used the CDI (n=2874 participants). Imputed analyses are based on 100 imputations for all missing CDI scores and include the 8 trials that used the CDI and the 10 trials that did not use the CDI (n=5289 participants).

Observed Observed and Imputed

Parameter Est SE t-val p-val Est SE t-val p-val
Intercept 8.53 0.70 12.18 <.001 9.38 0.44 21.49 <.001
Time −0.43 0.08 −5.36 <.001 −0.44 0.08 −5.59 <.001
Tx*Time −0.26 0.09 −2.76 .006 −0.29 0.09 −3.3 .001
SD(b0k) 1.92 1.64
SD(b0i) 6.43 5.95
SD(b1i) 1.65 1.13
Corr(b0i, b1i) −0.47 −0.37
SD(εijk) 4.16 4.48

SD(b0k): Standard deviation of random trial-level intercepts

SD(b0i): Standard deviation of random subject-level intercepts

SD(b1i): Standard deviation of random subject-level slopes

Corr(b0i, b1i): Correlation of random intercepts and slopes

SD(εijk): Standard deviation of residual error

Discussion

We have described a multiple imputation approach for harmonizing outcomes across multiple longitudinal trials. In our motivating example, we initially sought to harmonize 10 measures across 19 trials. This proved to not be possible using our methodology, because there was not enough overlap across measures to enable us to estimate their joint distribution. We then pursued a more modest goal, dropping one of the trials and attempting to harmonize only the CDI which was already used in 8 of the 18 remaining trials and overlapped with most of the other measures in at least one trial. Based on our imputation diagnostics, this reduced model did not appear to preserve relationships among variables or produce accurate imputations. Performing imputations on the original scale of the outcomes or after square root transformation did not improve our results, nor did rounding or not rounding the imputations. Still, this exercise was informative, because it highlighted those conditions that are necessary for harmonizing measures across multiple trials using multiple imputation. We now summarize each of these conditions.

Trial-level variability should be incorporated into the imputation model

Our imputation model was a two-level hierarchical model where repeated observations were nested within individual. As a result, clustering at the trial level was ignored, and observations on different participants within the same trial were assumed to be independent. See Siddique et al. (2015) for a formal presentation of the assumptions that are made when three-level IPD data is imputed under a two-level imputation model. Ignoring between-trial variability in our imputation model resulted in imputed values which underestimated between-study variability. As a result, imputations of marginal means were attenuated, as our imputation model assumed the conditional means across all trials were the same. Not including random time effects at the trial level assumes that treatment effects are the same across trial. Post-imputation treatment effects were then attenuated towards the overall treatment effect.

At first blush, in a data set which contains 19 trials, incorporating between-trial variability into our imputation model would appear to be feasible. However, in a setting where a depression measure can be missing for every participant within a trial, estimating random-effects at the trial level for each depression measure requires sufficient information to measure the correlation between measures at the trial level which requires that both measures must be used together in 3 or more trials. As can be seen in Table 2, most pairs of measures overlap in fewer than 3 trials.

Two sources of between-trial variability in our data set are the various interventions used in the different trials and the various patient populations. When it is not possible incorporate trial-level variability into the imputation model, one option is to restrict the number of trials to a more homogenous sample with respect to patient population and intervention type. This could potentially remove the need to model trial-level variability at the expense of addressing a different research question.

Relationships among variables over time must be allowed to change

Again due to the sparsity in our data, we were unable to estimate random slope effects at neither the trial nor the participant level in our imputation model. Thus, our models assume that variances of measures are constant over time and that the correlations between measures are time-invariant. However, looking at the observed columns in Table 6, correlations with the CDI appear to change over time. An imputation model that assumes this correlation is the same at all time points will generate inaccurate (or diffuse) imputations. Since our analyses are concerned with measuring change over time, it is essential that our imputed values preserve these relationships over time

Measures to be harmonized should be related to one another

This last condition seems obvious but it was an issue in our data. Although all studies sought to measure depression, some studies used self ratings, others used parent ratings, and some used clinician ratings. Some subscales sought to measure different components of depression. The result, as seen in Table 3 are ten measures that are for the most part, not highly correlated with one another. Contrast this with our prior work (Siddique et al. 2015) where we harmonized the CDRS and the Hamilton Depression Rating Scale (HDRS) (Hamilton 1960). In that study, the correlation between the CDRS and the HDRS was as high as 0.85. It is not enough that targets of harmonization putatively be measuring the same construct. The variables themselves need to be highly correlated.

A useful diagnostic in our setting is the fraction of missing information (Rubin 1987) which measures the additional inferential uncertainty in a parameter due to missing data. In some settings, a high rate of missing values for a variable does not automatically translate into high rates of missing information for its marginal parameters because the variable may be highly correlated with other variables that are more fully observed (Schafer, 1997). In those situations, multiple imputation can provide precise and valid inferences. The percentage of missing CDI values in our final set of 18 trials was 65%. In our analyses of the CDI with observed and imputed data reported in Table 9, the fraction of missing information for the time by treatment interaction term was 61%. The similarity of these two values suggest that the other depression measures in our data set did not help improve the accuracy of our imputation model and highlights the importance of having variables in the imputation model that are highly correlated with the variables that have missing values.

Conclusion

We sought to harmonize 10 measures across 19 trials and were unable to do so primarily due to the large amount of missing information, the lack of overlap across measures, and the low correlations among many of the measures that did overlap. We pursued harmonization via multiple imputation because, when done correctly, it has the following advantages: variables remain on their original scale, special analytical methods are not required after the data have been imputed, relationships among variables are preserved, and between-trial variability is accounted for. If the analyst is willing to forgo some of these advantages, other approaches may be feasible. The simplest approach is to standardize all measures and treat them as if they were identical on the transformed scale. Standardization can be easily applied in most situations with continuous measures and does not require specialized software. However, standardization does not take into account differences in the measurement properties of different scales and tends to mask heterogeneity between studies. Interpretation can be difficult because the analysis is no longer on the original scale (Griffith et al. 2015).

Latent variable methods which assume a single common factor may be more feasible in our setting but use sophisticated models and require assumptions regarding measurement invariance over time that can be hard to check. This is the approach taken by Brown et al. (2016) in this issue who imposed a two-level latent growth model on the depression measures. Their approach borrows strength across studies by relying on a common single depression factor whose measurement properties are assumed to be constant over the two years of data.

A promising approach in our opinion is to bring in additional sources of information. One way to obtain additional information is by drilling down to the item level and linking items across measures. When the same items occur in different measures, an item response theory (IRT) approach can be used (Bauer & Hussong 2009, Curran 2009, Curran & Hussong 2009, Curran et al. 2008). Even if there is no overlap in items across instruments, a bifactor IRT approach can be used to determine a single factor that is shared across all instruments and separate factors that account for differences in instruments. This approach has been investigated on these same data by Howe, Dagne, Brown, Brincks, and Beardslee (2016) who were able to identify invariant items (i.e., showed no differential item functioning) despite extreme sparseness in the overlap of instruments and items.

Another source of (external) information are “bridging studies” that provide overlap on measures when there is no overlap in the data set of interest. These bridging studies can be appended to the data set of interest in order to facilitate harmonization (Siddique et al. 2015). Two other potential approaches are to create synthetic data (Schifeling & Reiter 2015) or use informative priors (Rässler 2003). A careful simulation study investigating properties of the above methods—in addition to imputation—for large scale harmonization would be a useful contribution to the literature. In particular, how these methods perform when faced with the challenges encountered in this paper: little overlap among measures to be harmonized, substantial between-trial variability, low (and changing over time) correlation among variables.

Increasingly, researchers are collecting data from multiple studies in order to synthesize findings and perform more sophisticated analyes. These projects will continue to grow as federal funding agencies encourage data sharing (National Institutes of Health 2003, National Science Foundation 2011) and more journals require the release of data to accompany manuscripts. Methods that harmonize variables across data sets and facilitate analyses by many researchers are increasingly important in order to make full and efficient use of synthesized data and take advantage of the potential of IPD meta-analysis to address new questions not answerable by a single study.

Supplementary Material

11121_2017_760_MOESM1_ESM

Acknowledgments

Funding We gratefully acknowledge the National Institute of Mental Health Collaborative Synthesis for Adolescent Depression Trials Study Team, comprised of our many colleagues who generously provided their data to be used in this study, obtained access to key datasets, reviewed coding decisions, or provided substantive or methodologic recommendations. We also thank NIH for their support through Grant Number R01MH040859 (Collaborative Synthesis for Adolescent Depression Trials, Brown PI), and the following grants: Siddique-NCI CA154862-01, Garber, Brent, Beardslee, Clarke et al. NIMH MH64735, MH6454, MH64717, Gillham et al.- NIMH MH52270, Garber et al.-William T. Grant Foundation 961730, Dishion et al.- NIDA DA07031 and DA13773, Gillham et al.- NIMH MH52270, Szapocznik et al.- NIMH MH61143, Pantin et al.- NIDA DA017462, Prado et al.- NIDA DA025894, Prado et al.- CDC U01PS000671, Stormshak et al.- NIDA DA018374, Sandler et al.- NIMH MH49155, Wolchik et al.- NIMH MH068685, Young et al.- NARSAD, Spoth et al. - NIDA DA 007029, Clarke et al.- NIMH MH 48118, Young et al.- NIMH MH071320, Beardslee et al.- NIMH MH48696, VanVoorhees et al.-NIMH MH072918, and Gonzales et al. NIMH MH64707. The content of this paper is solely the responsibility of the authors and does not necessarily represent the official views of the funding agencies nor that of our collaborators who provided access to their data.

Footnotes

Compliance with Ethical Standards

Disclosure of potential conflicts of interest The authors declare that they have no conflicts of interest.

Ethical approval All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki declaration and its later amendments or comparable ethical standards.

Informed consent Informed consent was obtained from all individual participants included in the study.

Contributor Information

Juned Siddique, Department of Preventive Medicine, Northwestern University.

Peter J. de Chavez, Department of Preventive Medicine, Northwestern University

George Howe, Department of Psychology, George Washington University.

Gracelyn Cruden, Department of Psychiatry and Behavioral Sciences, Northwestern University.

C. Hendricks Brown, Department of Psychiatry and Behavioral Sciences, Northwestern University.

References

  1. Achenbach TM. Manual for the Child Behavior Checklist/4-18 and 1991 profile [Computer software manual] Burlington, VT: University of Vermont, Department of Psychiatry; 1991. [Google Scholar]
  2. Bauer DJ, Hussong AM. Psychometric approaches for developing commensurate measures across independent studies: Traditional and new models. Psychological Methods. 2009;14:101–125. doi: 10.1037/a0015583. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Brown CH, Brincks A, Huang S, Perrino T, Cruden G, Pantin H, … Sandler I. Two-year in impact of prevention programs on adolescent depression: An integrative data analysis approach. Prevention Science. 2016 doi: 10.1007/s11121-016-0737-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Clarke GN, Lewinsohn PM, Hops H, Seeley JR. A self-and parent-report measure of adolescent depression: The Child Behavior Checklist Depression scale (CBCL-D) Behavioral Assessment. 1992;14 [Google Scholar]
  5. Cowles MK, Carlin BP. Markov chain monte carlo convergence diagnostics: a comparative review. Journal of the American Statistical Association. 1996;91(434):883–904. [Google Scholar]
  6. Curran PJ. The seemingly quixotic pursuit of a cumulative psychological science: Introduction to the special issue. Psychological Methods. 2009;14:77–80. doi: 10.1037/a0015972. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Curran PJ, Hussong AM. Integrative data analysis: The simultaneous analysis of multiple data sets. Psychological Methods. 2009;14:81–100. doi: 10.1037/a0015914. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Curran PJ, Hussong AM, Cai L, Huang W, Chassin L, Sher KJ, Zucker RA. Pooling data from multiple longitudinal studies: The role of item response theory in integrative data analysis. Developmental Psychology. 2008;44:365–380. doi: 10.1037/0012-1649.44.2.365. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Dagne GA, Brown CH, Howe G, Kellam SG, Liu L. Testing moderation in network meta-analysis with individual participant data. Statistics in Medicine. 2016;34:2485–2502. doi: 10.1002/sim.6883. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Eaton WW, Smith C, Ybarra M, Muntaner C, Tien A. Center for epidemiologic studies depression scale: review and revision (CESD and CESD-R) In: Maruish ME, editor. The use of psychological testing for treatment planning and outcomes assessment: Volume 3: Instruments for adults. 3. Mahwah, NJ: Lawrence Erlbaum Associates Publishers; 2004. pp. 363–377. [Google Scholar]
  11. Ebesutani C, Bernstein A, Martinez JI, Chorpita BF, Weisz JR. The Youth Self Report: applicability and validity across younger and older youths. Journal of Clinical Child & Adolescent Psychology. 2011;40:338–346. doi: 10.1080/15374416.2011.546041. [DOI] [PubMed] [Google Scholar]
  12. Gelman A, Carlin JB, Stern HS, Rubin DB. Bayesian data analysis, 2nd edition. Boca Raton, FL: Chapman & Hall/CRC press; 2004. [Google Scholar]
  13. Gelman A, King G, Liu C. Not asked and not answered: Multiple imputation for multiple surveys. Journal of the American Statistical Association. 1998;93:846–857. [Google Scholar]
  14. Gelman A, Meng XL, Stern H. Posterior predictive assessment of model fitness via realized discrepancies. Statistica Sinica. 1996;6:733–760. [Google Scholar]
  15. Gelman A, Rubin DB. Inference from iterative simulation using multiple sequences. Statistical Science. 1992:457–472. [Google Scholar]
  16. Geweke J. Evaluating the accuracy of sampling-based approaches to calculating posterior moments. In: Bernado JM, Berger JO, Dawid AP, Smith AF, editors. Bayesian statistics. Vol. 4. Oxford: Clarendon Press; 1992. pp. 169–193. [Google Scholar]
  17. Griffith LE, Van Den Heuvel E, Fortier I, Sohel N, Hofer SM, … Payette H, et al. Statistical approaches to harmonize data on cognitive measures in systematic reviews are rarely reported. Journal of Clinical Epidemiology. 2015;68:154–162. doi: 10.1016/j.jclinepi.2014.09.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Hamilton M. A rating scale for depression. Journal of Neurology, Neurosurgery, and Psychiatry. 1960;23:56–62. doi: 10.1136/jnnp.23.1.56. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Harel O, Zhou XH. Multiple imputation: review of theory, implementation and software. Statistics in Medicine. 2007;26:3057–3077. doi: 10.1002/sim.2787. [DOI] [PubMed] [Google Scholar]
  20. He Y, Zaslavsky AM. Diagnosing imputation models by applying target analyses to posterior replicates of completed data. Statistics in Medicine. 2012;31:1–18. doi: 10.1002/sim.4413. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Helsel WJ, Matson JL. The assessment of depression in children: The internal structure of the Child Depression Inventory CDI. Behaviour Research and Therapy. 1984;22:289–298. doi: 10.1016/0005-7967(84)90009-3. [DOI] [PubMed] [Google Scholar]
  22. Howe GW, Dagne G, Brown CH, Brincks A, Beardslee W. Evaluating construct equivalence and harmonizing measurement of adolescent depression when synthesizing results across multiple studies. 2016 In preparation. [Google Scholar]
  23. Hussong AM, Curran PJ, Bauer DJ. Integrative data analysis in clinical psychology research. Annual Review of Clinical Psychology. 2013;9:61–89. doi: 10.1146/annurev-clinpsy-050212-185522. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Kline D, Andridge R, Kaizar E. Comparing multiple imputation methods for systematically missing subject-level data. Research Synthesis Methods. 2015:1–13. doi: 10.1002/jrsm.1192. [DOI] [PubMed] [Google Scholar]
  25. Kovacs M. The Children’s Depression Inventory CDI. Psychopharmacology Bulletin. 1984;21:995–998. [PubMed] [Google Scholar]
  26. Mayes TL, Bernstein IH, Haley CL, Kennard BD, Emslie GJ. Psychometric properties of the Children’s Depression Rating Scale–revised in adolescents. Journal of Child and Adolescent Psychopharmacology. 2010;20:513–516. doi: 10.1089/cap.2010.0063. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. National Institutes of Health. [Accessed 3-March-2014];Final NIH statement on sharing research data. 2003 http://grants.nih.gov/grants/guide/notice-files/NOT-OD-03-032.html.
  28. National Science Foundation. [Accessed 3-March-2014];Dissemination and sharing of research results. 2011 http://www.nsf.gov/pubs/policydocs/pappguide/nsf11001/aag_6.jsp#VID4.
  29. Perrino T, Howe G, Sperling A, Beardslee W, Sandler I, Shern D, … Brown CH. Advancing science through collaborative data sharing and synthesis. Perspectives on Psychological Science. 2013;8:433–444. doi: 10.1177/1745691613491579. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Poznanski EO, Freeman LN, Mokros HB. Children’s depression rating scale–revised (September 1984) Psychopharmacology Bulletin. 1985;21:979–989. [Google Scholar]
  31. Quay HC, Peterson DR. Revised Behavior Problem Checklist. Odessa, FL: Psychological Assessment Resources; 1996. [Google Scholar]
  32. R Core Team. R: A language and environment for statistical computing [Computer software manual] Vienna, Austria: 2012. Retrieved from http://www.R-project.org/ [Google Scholar]
  33. Radloff LS. The CES-D scale a self-report depression scale for research in the general population. Applied Psychological Measurement. 1977;1:385–401. [Google Scholar]
  34. Radloff LS. The use of the Center for Epidemiologic Studies Depression Scale in adolescents and young adults. Journal of Youth and Adolescence. 1991;20:149–166. doi: 10.1007/BF01537606. [DOI] [PubMed] [Google Scholar]
  35. Rässler S. A non-iterative Bayesian approach to statistical matching. Statistica Neerlandica. 2003;57:58–74. [Google Scholar]
  36. Resche-Rigon M, White IR, Bartlett JW, Peters SA, Thompson SG. Multiple imputation for handling systematically missing confounders in meta-analysis of individual participant data. Statistics in Medicine. 2013;32:4890–4905. doi: 10.1002/sim.5894. [DOI] [PubMed] [Google Scholar]
  37. Riley RD, Lambert PC, Abo-Zaid G. Meta-analysis of individual participant data: rationale, conduct, and reporting. BMJ: British Medical Journal. 2010:521–525. doi: 10.1136/bmj.c221. [DOI] [PubMed] [Google Scholar]
  38. Rodwell L, Lee KJ, Romaniuk H, Carlin JB. Comparison of methods for imputing limited-range variables: a simulation study. BMC Medical Research Methodology. 2014;14:57. doi: 10.1186/1471-2288-14-57. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Rubin DB. Multiple imputation for nonresponse in surveys. New York: John Wiley and Sons; 1987. [Google Scholar]
  40. Schafer JL, Yucel RM. Computational strategies for multivariate linear mixed-effects models with missing values. Journal of Computational and Graphical Statistics. 2002;11:437–457. [Google Scholar]
  41. Schifeling TA, Reiter JP. Incorporating marginal prior information in latent class models. Bayesian Analysis 2015 [Google Scholar]
  42. Siddique J, Reiter JP, Brincks A, Gibbons RD, Crespi CM, Brown CH. Multiple imputation for harmonizing longitudinal non-commensurate measures in individual participant data meta-analysis. Statistics in Medicine. 2015;34:3399–3414. doi: 10.1002/sim.6562. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Zhao JH, Schafer JL. R package version 0.9. 2013. pan: Multiple imputation for multivariate panel or clustered data [Computer software manual] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

11121_2017_760_MOESM1_ESM

RESOURCES