Abstract
In psychiatric and social epidemiology studies, it is common to measure multiple different outcomes using a comprehensive battery of tests thought to be related to an underlying construct of interest. In the research that motivates our work, researchers wanted to assess the impact of in utero alcohol exposure on child cognition and neuropsychological development, which are evaluated using a range of different psychometric tests. Statistical analysis of the resulting multiple outcomes data can be challenging, because the outcomes measured on the same individual are not independent. Moreover, it is unclear, a priori, which outcomes are impacted by the exposure under study. While researchers will typically have some hypotheses about which outcomes are important, a framework is needed to help identify outcomes that are sensitive to the exposure and to quantify the associated treatment or exposure effects of interest. We propose such a framework using a modification of stochastic search variable selection, a popular Bayesian variable selection model and use it to quantify an overall effect of the exposure on the affected outcomes. The performance of the method is investigated empirically and an illustration is given through application using data from our motivating study.
Keywords: Bayesian methods, biostatistics, variable selection
1 |. INTRODUCTION
In psychological and social epidemiology studies, participants are typically assessed using a comprehensive battery of tests or tasks designed to measure psychological, neurological, or cognitive outcomes that are difficult to measure directly. Analysts then face the challenge of how to best handle the resulting multiple outcomes. Often, a large number of outcomes are collected, and it can be challenging to decide which outcomes to include in the analysis. Scientists typically rely on previous studies, in combination with expert knowledge, to select the outcomes on which to focus. No statistical framework has been available for identifying outcomes that are sensitive to an exposure, nor has such a framework been developed to quantify the magnitude of effects.
There is a rich literature on statistical methods for the analysis of multiple outcomes data. The simplest approach is to analyze each outcome separately, but such an analysis requires adjustment for multiple comparisons (Lefkopoulou & Ryan, 1993). Structural equation models (SEMs) can also be used to model correlated outcomes by treating the outcomes as manifestations of the latent variables (Budtz-Jørgensen et al., 2002; Dunson, 2000; Sánchez et al., 2005). However, the regression coefficients characterizing the relationship between the exposure and the latent factor can be problematic to interpret, and inference is sensitive to model misspecification (Sammel & Ryan, 2002). Meta-analysis is another popular approach to synthesis of multiple outcomes data, but relatively little work has been carried out for dealing with highly correlated outcomes in observational settings (Akkaya Hocagil et al., 2022; Berkey et al., 1998; Ryan, 2008; van den Noortgate et al., 2015). Generalized estimating equations (Lefkopoulou et al., 1989; Liang & Zeger, 1986) have also been used to analyze multiple outcome data, with working covariance matrices specified to accommodate correlations across outcomes, because the repeated observations on each individual can be viewed as a special type of clustered data. Generalized linear mixed models offer another framework to model the effect of exposure on multiple outcomes (Sammel et al., 1999; Thurston et al., 2009). In this paper, we extend the generalized linear modeling approach for the analysis of multiple outcomes.
A limitation of the available statistical methods for analyzing multiple outcomes data is that researchers must specify the outcomes to be included in the analysis. As mentioned above, this is usually done using expert knowledge or following some gatekeeping procedure to select the subset of affected outcomes (see, for example, Turk et al., 2008). However, this can be challenging when outcomes are high dimensional or when expert knowledge does not provide strong guidance. Moreover, using exploratory data analysis to guide the decision-making increases the risk of distorting statistical inference due to multiple comparisons. We develop and evaluate a principled statistical approach for identification of relevant outcomes on which to model the exposure effects, while accounting for the correlation among the outcomes.
We refer to the challenge of identifying which of many observed outcomes are sensitive to an exposure as the outcome selection problem and show that it can be reframed as a classical variable selection problem. Variable selection is typically carried out to choose a subset of candidate predictors that together explain most of the variation in a single response variable. The variable selection literature has a long history, from earlier frequentist approaches such as “best subset” regression, model selection based on Akaike/Bayesian information criterion (Akaike, 1998; Schwarz, 1978), backward and forward stepwise regression, to the more recent Bayesian methods that involve a wide range of “slab-and-spike” or shrinkage priors; see Hastie et al. (2020), O’Hara and Sillanpää (2009), and van Erp et al. (2019) for some recent reviews. To the best of our knowledge, these ideas and approaches have not been adapted to deal with the setting where one aims to select which outcomes in a large set of candidate outcomes are sensitive to an exposure.
In this paper, we first show how the problem of interest can be reframed as one of variable selection. We adopt a Bayesian approach to analyze outcomes and identify those that are strongly affected by the exposure. The model is motivated by the popular stochastic search variable selection (SSVS) method, but we extend the SSVS prior to allow estimation of a mean effect among the sensitive outcomes. A random effects model is used to account for the correlation among outcomes measured on the same individuals.
The paper is organized as follows: In Section 2, we present the basic model and show how the outcome selection problem can be reframed as one of variable selection. We also discuss the associated computing approach. In Section 3, we assess the performance of our method in comparison to other variable selection models based on a simulation study. In Section 4, we use the model and method to analyze data from our motivating application regarding the effect of in utero alcohol exposure on different measures of child cognition. In Section 5, we present some conclusions and discussion.
2 |. METHODOLOGY
2.1 |. Addressing the outcome selection problem
Suppose we observe continuous outcomes for each of independent individuals. The outcomes will typically be correlated because they are measures from the same individual, though they may be of different scales and nature. For example, the outcomes may be measuring different domains of a person’s cognitive function (verbal versus mathematical). Therefore, exposure effects are not expected to be exactly the same across affected outcomes but vary around a mean level , which we identify as the parameter of interest. For each individual, we observe an exposure value and some other observed predictor variables, which we denote as . In the application discussed in Section 4, is a propensity score computed for each individual to adjust for confounders.
We now show how to express our multiple outcomes data as panel data in long format. Consider a sample of independent individuals labeled , where each individual has measurements on outcomes labeled . For now, assume there are no missing data and that all outcomes have been measured on all individuals. Now suppose we stack the observations from all individuals together, giving us a data set with observations in total. Row of this data set records the observed outcome corresponding to individual and outcome .
To represent the multiple outcome problem as a multiple predictors problem, we first define a new set of covariates :
for . The ith value of is the interaction between the exposure level of individual and a dummy variable indicating whether the value of the outcome is . Including this exposure by outcome interaction term is critical because it allows for a potentially different exposure effect, depending on outcome. An example of the dataframe format and how to map from the multiple outcome format to the stacked format for individuals and outcome variables is presented in Figure 1. This “trick” of expressing the multiple outcomes problem in terms of repeated measures has been widely used in the literature, making it straightforward then to analyze multiple outcomes using standard mixed modeling or GEE software (Lefkopoulou et al., 1989).
We will base our analysis on a linear mixed model, as follows:
(2.1) |
for , individual and outcome . The error terms are independent and normally distributed, and the random effect accounts for the within-individual correlation and for .
The parameters and are outcome-specific intercepts and coefficients for , and the coefficients represents the exposure effect on outcome . In a classical multiple outcomes setting, it is typical to assume that all outcomes are associated in a similar way with the exposure or treatment of interest, with effects varying around a mean level . It is natural to assume
for , and then assign appropriate priors for and to estimate the model using a Bayesian estimation procedure. We now want to generalize this framework to allow for the possibility that not all outcomes are affected by the exposure. Asking the question of which outcomes should be included becomes a problem of variable selection, based on the covariates . From a modeling perspective, allowing for some of the outcomes to be unaffected by the exposure simply corresponds to setting for those variables.
Variable selection methodologies have been extended to deal with random effects; see for example Bondell et al. (2010), Fan and Li (2012), and Yang et al. (2020). For our model, because we do not need to perform variable selection for the random effects, it is straightforward to use existing techniques for the independent predictors . This means that we can potentially use any of a variety of sparsity priors, such as SSVS (George & McCulloch, 1993), Bayesian LASSO (Figueiredo, 2003; Park & Casella, 2008), and the horseshoe prior (Carvalho et al., 2009) for outcome selection. However, because we are interested in selecting the subset of affected outcomes variables and quantifying the mean exposure effect on these variables, in this study we focus on the use of the SSVS method. We discuss this further in the next section.
2.2 |. Stochastic search variable selection
There is a large literature on Bayesian variable selection methods; however, in this paper, we only discuss the “slab and spike” type of priors, as they are suitable for our problem of identifying sensitive outcomes from a large number of outcomes. Methods that compare models by Bayes Factor (Kass & Raftery, 1995) or criteria such as DIC (Spiegelhalter et al., 2002) or WAIC (Watanabe & Opper, 2010) require fitting all candidate models and hence only applicable when comparing a small number of models. Therefore, they are not suitable for the outcome selection problem.
Methods involving a “slab and spike” prior can be divided, broadly, into two categories: Methods that specify a prior that approximate the “slab and spike” shape for the coefficients ; and methods that use latent indicator variables that indicate whether a covariate is included in the model. Shrinkage priors such as the Bayesian LASSO (Figueiredo, 2003; Park & Casella, 2008) and the horseshoe prior (Carvalho et al., 2009) belong to the first category. The implementation of these methods is straightforward and they have had extensive use in recent years. However, it is not clear how to modify these priors to incorporate a common mean of the nonzero coefficients. Yang et al. (2020) proposed using SSVS for selection of fixed effects in linear mixed models; however, they also did not consider estimating the mean effect.
The second category of approach defines a latent variable that indicates whether a coefficient is nonzero. In the approaches proposed by Kuo and Mallick (1998) and Dellaportas et al. (2002), a coefficient is set to 0 if . Both methods specify and hence require an appropriate prior for . These approaches can be challenging to tune to ensure that the iterates of do not get stuck at 0 or 1. For example, mixing may be poor for the Kuo and Mallick (1998) approach if the prior for is too vague (O’Hara & Sillanpää, 2009). We can assume with unknown and , but the model will be hard to fit and we cannot interpret as the mean of all of which .
Our method is motivated by the SSVS method (George & McCulloch, 1993), which defines a mixture prior for instead: Let be a latent indicator variable, with means covariate is included in the model, means it is not. The indicator affects the prior of , so we can define a joint prior for as
Conditioning on , the prior of is
For our outcome selection framework, we propose to modify the SSVS prior to incorporate the mean exposure effect on the sensitive outcomes. Conditioning on , we now have a mixture prior for
(2.2) |
To improve the performance of the model, we follow Meuwissen and Goddard (2004) and modify the prior in (2.2) to
(2.3) |
The tuning parameter should be chosen to ensure good separation between the “in” and “out” variables. The standard deviations , and are assigned log-normal priors in our simulation study and application.
Note that the posterior mean of will be the posterior probability that outcome is included in the model; hence, it will be important in terms of interpreting the results of our model fit. The prior, , can simply be a categorical distribution with a fixed probability parameter. This prior probability may be different across outcomes, based on the experts’ knowledge, or fixed at 0.5 so that the prior is non-informative. The prior probability can also be treated as a parameter to be estimated (O’Hara & Sillanpää, 2009). For the examples in this paper, we simply set a prior probability for each .
For the examples in this paper, an outcome is classified as “relevant” if the posterior mean of the corresponding is greater than 0.5. We note that the threshold may affect the conclusion on the relevance of each outcome variable but does not change the estimate of the mean effect . Of course, other thresholds could be used. We suggest that it is best to report the posterior probabilities of for all . An alternative approach is to look at the whole vector to identify the most frequently sampled subsets of outcomes.
3 |. SIMULATION STUDY
3.1 |. Setup
In this section, we demonstrate the performance of the method using simulated data. The aim of this simulation study is to assess the performance of our prior for outcome selection for different effect sizes. In the exercise, we set the number of outcomes to and a moderate sample size of . We examine the performance of the proposed model with different numbers of relevant outcomes .
We simulated 10 data sets from the model (2.1) described in Section 2.1. We set the parameters value to generate the data sets as follows: The intercepts values are randomly picked from and standard deviations are generated from for . The coefficients corresponding to the relevant outcomes were generated from to ensure that the are scattered closely enough around . We used two different values for the mean common effect and .
Given these “true” parameters values, in each simulation, we create a data set by first generate the exogenous variable from and exposure and then generate the outcome according to model (2.1).
We then fit model (2.1) using the prior in (2.3) with to each of the 10 data sets. We examine both versions of SSVS: Our proposed model in which is a parameter and the standard SSVS prior where . The prior probability of is for all outcome . For comparison, we also fit the model where are assumed a hierarchical prior with unknown and . We also fit the model that only uses the correct relevant outcomes, assuming . We call this the “subset model.” The result of the subset model is treated as the “standard” because it is the model that uses the correct set of outcomes.
The rest of the parameters are assigned fairly flat priors. For example, we use a normal prior for and . The parameters and are assigned log-normal(0,10) priors. In all models, is assigned a log-normal(0,1) prior.
For each simulation, we record the number of outcomes identified as relevant, the number of correctly identified outcomes, the number of false positives, and the estimated . An outcome is classified as “relevant” if the posterior mean of the corresponding is greater than 0.5. The results presented here represent the average over the 10 simulations for each setting. The SSVS models are fitted using the software JAGS (Plummer, 2003) and the other models are implemented with STAN (Carpenter et al., 2017).
Note that in this simulation study, the first setting with represents a situation when the effect is weak with small data, so that the posterior standard deviation is large. In this case, it would be difficult for the model to decide whether a is 0 or not. The second setting mimics the situation in which the effect is stronger, and the selection method is expected to work better.
3.2 |. Results
The performance of the SSVS algorithm in detecting the affected outcomes for different and is presented in Table 1. The results suggest that the original and our modified SSVS algorithms have very similar performance, though neither do well in detecting the affected outcomes when is small. This is expected as the overall effect is small, so some of the relevant would be close to 0. Because here we used uninformative priors for , the algorithm will keep switching between stage and for these outcomes. This is similar to the phenomenon observed by O’Hara and Sillanpää (2009), where the posterior probabilities of are close to 0.5 and some outcomes are classified incorrectly by chance.
TABLE 1.
# identified as relevant | # correctly identified | # incorrectly identified | |||||
---|---|---|---|---|---|---|---|
True μ | K 1 | μ unknown | μ = 0 | μ unknown | μ = 0 | μ unknown | μ = 0 |
−0.1 | 5 | 4.9 | 4.4 | 1.6 | 1.5 | 3.3 | 2.9 |
10 | 5.2 | 5.3 | 1.9 | 2.1 | 3.3 | 3.2 | |
15 | 5.1 | 4.4 | 4.5 | 3.5 | 0.6 | 0.9 | |
−3 | 5 | 5 | 5 | 5 | 5 | 0 | 0 |
10 | 10 | 10.1 | 10 | 10 | 0 | 0.1 | |
15 | 15 | 15 | 15 | 15 | 0 | 0 |
Table 2 shows the mean squared errors of estimating the individual coefficient :
(3.1) |
where we take the estimate to be the posterior mean of if the posterior mean of is greater than 0.5; otherwise we set . For both large and small values of , the SSVS priors provide more accurate estimates of in terms of MSE, compared to the model without variable selection.
TABLE 2.
SSVS – μ unknown | SSVS - μ = 0 | No variable selection | ||||
---|---|---|---|---|---|---|
K 1 | Small μ | Large μ | Small μ | Large μ | Small μ | Large μ |
5 | 0.012 | 0.004 | 0.007 | 0.007 | 0.032 | 0.066 |
10 | 0.022 | 0.017 | 0.014 | 0.032 | 0.036 | 0.075 |
15 | 0.014 | 0.032 | 0.011 | 0.053 | 0.036 | 0.073 |
Lastly, Table 3 shows the estimated of , averaged over 10 simulations, by different priors. Table 3 shows that our modified SSVS can provide estimates of that are closer to the result from the subset model, especially for large . However, when the effect is weak, the model is not able to estimate accurately because it fails to identify the correct set of sensitive outcomes.
TABLE 3.
True μ | K 1 | SSVS | No selection | Subset model |
---|---|---|---|---|
−0.1 | 5 | −0.025 (0.066) |
−0.120 (0.145) |
−0.209 (0.143) |
10 | 0.046 (0.069) |
0.034 (0.157) |
−0.015 (0.169) |
|
15 | −0.026 (0.091) |
−0.059 (0.188) |
−0.077 (0.213) |
|
−3 | 5 | −3.281 (0.040) |
−0.908 (0.150) |
−3.365 (0.144) |
10 | −2.949 (0.098) |
−1.393 (0.152) |
−2.875 (0.175) |
|
15 | −2.678 (0.115) |
−2.002 (0.210) |
−2.688 (0.221) |
The simulation example shows that SSVS priors can provide accurate estimates of the coefficients and accurately identify the affected outcomes and estimate the mean effect when is far from 0. However, it may require more informative priors for and better tuning to capture small effects accurately. The R code for the study is provided on Github–see https://github.com/khuedung91/BayesianOutcomeSelection/.
4 |. EFFECT OF PRENATAL ALCOHOL EXPOSURE ON CHILDREN IN DETROIT, MICHIGAN
In this section, we apply our proposed framework to data collected as part of an investigation of the long-term effect of prenatal alcohol exposure (PAE) on a child’s cognitive and behavioral function. Numerous studies have shown that high levels of PAE can result in a distinct pattern of craniofacial anomalies, growth restriction, and cognitive and behavioral deficits, a condition known as fetal alcohol syndrome (FAS) (Hoyme et al., 2005; Hoyme et al., 2016), the most severe of a continuum of fetal alcohol syndrome disorders (FASD) (Carter et al., 2016; Jacobson et al., 2004; Jacobson et al., 2008; Mattson et al., 2019). Alternatively, some individuals with PAE exhibit cognitive and/or behavioral impairment without the characteristic craniofacial dysmorphology and/or growth restriction, a disorder known as alcohol-related neurodevelopmental disorder (ARND).
Our data come from a longitudinal study, funded by the US National Institutes of Health and conducted in Detroit, Michigan. In this study, the mothers were interviewed prenatally about their alcohol consumption during pregnancy, and the children were followed throughout childhood, many of them up until they were 20 years of age. The study collected a large number of variables reflecting responses on various neuro-cognitive tests and behavioral outcomes assessed on the children throughout childhood. Each of the administered tests could be classified as relevant to one of several different domains including cognition, executive function, and behavior, among others. Previous neurocognitive studies have suggested that the impact of PAE on all of these domains will not be the same, given that alcohol may have a stronger effect on certain parts of the brain, while other areas may be relatively unaffected or spared, depending on the timing, genetic vulnerability, and ethnic or racial group of the exposure (Jacobson et al., 2004; Jacobson & Jacobson, 1999; 2002). Recent analyses by our group made use of expert knowledge to select outcomes for analysis and simply assumed that each had been affected by PAE to some extent (Jacobson et al., 2021).
To illustrate our methodology, in this paper we focus on a set of 14 outcomes collected when the children were approximately 7 years of age. The first eight outcomes come from the Achenbach Child Behavior Checklist (CBCL) and Teacher’s Report Form (TRF) at age 7 (Achenbach, 1991). The CBCL is a checklist completed by the parent and designed to detect emotional and behavioral problems in children and adolescents, whereas the TRF represents the child’s principal teacher’s report of the similar. These assessments include the child’s internalizing and externalizing behaviors, and social and attention problems. The remaining six outcomes correspond to the results of various cognitive and neuro-developmental tests related to IQ assessed on the Wechsler Intelligence Scales for Children–III (Wechsler, 1991), academic achievement in reading and arithmetic, learning and memory abilities, and executive function. Recent analyses have reported that, of these 14 outcomes, the first eight are relatively less affected by PAE, whereas the last six are more sensitive to alcohol exposure (Jacobson et al., 2021). After preprocessing, the data include outcomes from 336 children. PAE is computed based on the mother’s average daily dose of absolute alcohol consumed (in ounces) during pregnancy (AA/day). Because the distribution of alcohol exposure is positively skewed with a minimum level 0, we compute log(AA/day + 1) and use this as the measure of PAE in the analysis.
4.1 |. Model and setup
We fit the model (2.1) with the prior in (2.3) to the data set. To adjust for confounders associated with both alcohol exposure and cognitive function, we add a propensity score , which was computed beforehand. For details on the covariates included in the propensity score and how it was constructed, we refer readers to Akkaya Hocagil et al. (2021). Before running the analysis, we rescaled all outcomes to have mean 0 and variance 1.
We fit our proposed model with a few different settings. We start with an uninformative prior for the indicator and set for all . We use the prior in (2.3) with for where is assigned a log-normal (0,1) prior. As a comparison, we also try the prior in (2.2) where we fix and a shrinkage prior. For the shrinkage prior, we simply follow Figueiredo (2003) and assign a Laplace(0,1) prior for the . We also attempt the horseshoe prior (Carvalho et al., 2009) for but the MCMC has convergence issue and hence the result is not presented here.
To assess how sensitive the result is to the prior probability , we also fit the model with a more informative set of ,
These prior probabilities were chosen by utilizing expertise knowledge and set the probability of the outcomes that are known to be relevant to be closer to 1. In practice, more informative priors may help the MCMC to have better mixing.
We also fit the model (2.1) to only those outcomes chosen by our SSVS model with a hierarchical prior . We call this model the “subset” model. Similar to in the simulation study, we will compare the estimates of and from this reduced model with our approach.
We assigned a normal prior for . We found the appropriate prior’s parameters by fitting model (2.1) to the data using the R package Ime4. The estimates of from Ime4 suggested that the average effect on the affected outcomes may be around −1, and therefore we used the prior that covers this value. For the rest of the parameters, we chose diffuse priors. The prior for and is log-normal(0,10). The prior for and are normal and normal , respectively. The SSVS models were fitted using the software JAGS (Plummer, 2003), running three chains each with 200,000 burn-in and 200,000 samples with thinning = 10. The other models were implemented in STAN (Carpenter et al., 2017).
4.2 |. Results
The results are presented in Tables 4 and 5. Table 4 shows the mean posterior probability of . For the Laplace shrinkage prior, we report whether 0 is outside of the 95% credible intervals of the parameters. The table shows that all SSVS models choose the same set of relevant outcomes in different settings. The informative prior on results in different posterior mean of ; however, it does not affect the inference on the outcomes’ relevance for most outcome variables. The only exception is CBCL Externalizing at age 7, of which the posterior probability of being affected is slightly less than 0.5 (0.436 and 0.496) when using a noninformative prior and slightly higher than 0.5 (0.530) when using an informative prior. These results are also similar to that of the Laplace shrinkage prior; however, this prior shrinks more toward 0 than the SSVS priors.
TABLE 4.
Informative |
||||
---|---|---|---|---|
Laplace prior | ||||
CBCL Social Problem | 0 | 0.286 | 0.384 | 0.374 |
CBCL Attention Problem | 0 | 0.505 | 0.533 | 0.576 |
CBCL Internalizing | 0 | 0.195 | 0.247 | 0.057 |
CBCL Externalizing | 0 | 0.436 | 0.496 | 0.530 |
TRF Social Problem | 1 | 0.856 | 0.741 | 0.947 |
TRF Attention Problem | 1 | 0.873 | 0.742 | 0.947 |
TRF Internalizing | 0 | 0.208 | 0.267 | 0.065 |
TRF Externalizing | 1 | 0.864 | 0.746 | 0.947 |
Verbal IQ | 0 | 0.291 | 0.393 | 0.388 |
Performance IQ | 0 | 0.345 | 0.432 | 0.442 |
Freedom from distractibility | 1 | 0.966 | 0.809 | 0.969 |
Verbal fluency | 0 | 0.670 | 0.629 | 0.900 |
Digit span backwards | 1 | 0.871 | 0.737 | 0.945 |
Story memory | 0 | 0.267 | 0.360 | 0.334 |
TABLE 5.
Informative |
|||||
---|---|---|---|---|---|
Laplace prior | Subset | ||||
CBCL Social Problem | −0.090 | 0.000 | 0.000 | 0.000 | |
CBCL Attention Problem | −0.230 | −0.274 | −0.268 | −0.339 | −0.475 |
CBCL Internalizing | 0.083 | 0.000 | 0.000 | 0.000 | |
CBCL Externalizing | −0.195 | 0.000 | 0.000 | −0.324 | |
TRF Social Problem | −0.486 | −0.404 | −0.412 | −0.460 | −0.612 |
TRF Attention Problem | −0.464 | −0.396 | −0.404 | −0.451 | −0.602 |
TRF Internalizing | 0.064 | 0.000 | 0.000 | 0.000 | |
TRF Externalizing | −0.499 | −0.407 | −0.417 | −0.464 | −0.617 |
Verbal IQ | −0.112 | 0.000 | 0.000 | 0.000 | |
Performance IQ | −0.141 | 0.000 | 0.000 | 0.000 | |
Freedom from distractibility | −0.561 | −0.447 | −0.468 | −0.505 | −0.622 |
Verbal fluency | −0.327 | −0.332 | −0.330 | −0.388 | −0.525 |
Digit span backwards | −0.453 | −0.389 | −0.400 | −0.446 | −0.574 |
Story memory | −0.061 | 0.000 | 0.000 | 0.000 | |
μ | −0.324 | −0.303 | −0.398 | −0.572 | |
τ | 0.219 | 0.233 | 0.187 | 0.172 |
Table 5 presents the estimates of and overall effect . The SSVS with informative and the subset model suggest a strong negative effect of PAE on the cognitive outcomes. These findings are consistent with those in Jacobson et al. (2021). The estimate of is similar for both models (0.187 vs. 0.172). On the other hand, the SSVS models with the noninformative prior suggest a weaker effect (−0.324 and −0.303 versus −0.398). The noninformative prior also results in larger estimates of (0.219 and 0.233 vs. 0.187).
Table 5 shows that all SSVS models produce smaller estimates for the coefficients of the affected outcomes and hence , compared with the subset model. The result here is consistent with our observation in the simulation study in Section 3. However, as shown in Tables 4 and 5, our proposed model produces very similar estimates of compared with the Laplace prior in all settings. Table 5 also indicates that the informative prior for the indicator produces estimates of and that are closer to the subset model that only includes the affected outcomes.
5 |. CONCLUSION
In this paper, we propose a statistical method for identifying outcomes from a large number of observed variables that are directly affected by an exposure variable. Our method is an extension of standard Bayesian variable selection models to multiple outcomes data, which also provides an estimate of the overall effect of the exposure variable in the subset of affected outcomes. We demonstrate the performance and limitations of our method in a simulation exercise and a real data application.
Our application in modeling the effect of PAE on cognition identified a set of neurodevelopmental tests that are significantly affected by fetal alcohol exposure. In addition, the model indicates a negative overall effect of PAE on the sensitive outcomes. A limitation of the current model is that we only use an individual random intercept to capture the correlations among the outcomes. This approach may not be ideal, and we may consider a more sophisticated correlation structure in future work.
Finally, the proposed framework is shown to be effective in identifying sensitive outcomes in various scenarios. However, it may underestimate the outcome-specific effect size and mean effect when the effects are mild. This is a common issue with variable selection priors; we expect that the result can be improved by using more informative priors for the indicators .
ACKNOWLEDGEMENTS
This research was funded by grants from the National Institutes of Health/National Institute on Alcohol Abuse and Alcoholism (NIH/NIAAA; R01-AA025905) and the Lycaki-Young Fund from the State of Michigan to Sandra W. Jacobson and Joseph L. Jacobson. Much of the work was accomplished while Khue-Dung Dang was a postdoctoral research fellow at UTS, supported by the Australian Research Council Centre of Excellence for Mathematical and Statistical Frontiers (ACEMS) grant CE140100049. This work was also supported by a grant from the Canadian Institutes of Health Research to Richard J. Cook and Louise Ryan (FRN PJT-180551). Open access publishing facilitated by The University of Melbourne, as part of the Wiley - The University of Melbourne agreement via the Council of Australian University Librarians.
Funding information
Australian Research Council Centre of Excellence for Mathematical and Statistical Frontiers, Grant/Award Number: CE140100049; Canadian Institutes of Health Research, Grant/Award Number: FRN PJT-180551; Lycaki-Young Fund; National Institutes of Health/ National Institute on Alcohol Abuse and Alcoholism, Grant/Award Number: R01-AA025905
REFERENCES
- Achenbach TM (1991). Manual for the Child Behavior Checklist/4–18 and 1991 Profile. University of Vermont, Department of Psychiatry. [Google Scholar]
- Akaike H (1998). Information theory and an extension of the maximum likelihood principle, Selected Papers of Hirotugu Akaike (pp. 199–213). Springer. [Google Scholar]
- Akkaya Hocagil T, Cook RJ, Jacobson SW, Jacobson JL, & Ryan LM (2021). Propensity score analysis for a semi-continuous exposure variable: A study of gestational alcohol exposure and childhood cognition. Journal of the Royal Statistical Society: Series A (Statistics in Society), 184(4), 1390–1413. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Akkaya Hocagil T, Ryan LM, Cook RJ, Jacobson SW, Richardson GA, Day NL, Coles CD, Carmichael Olson H, & Jacobson JL (2022). A hierarchical meta-analysis for settings involving multiple outcomes across multiple cohorts. Stat, 11(1), e462. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Berkey CS, Hoaglin DC, Antczak-Bouckoms A, Mosteller F, & Colditz GA (1998). Meta-analysis of multiple outcomes by regression with random effects. Statistics in Medicine, 17(22), 2537–2550. [DOI] [PubMed] [Google Scholar]
- Bondell HD, Krishna A, & Ghosh SK (2010). Joint variable selection for fixed and random effects in linear mixed-effects models. Biometrics, 66(4), 1069–1077. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Budtz-Jørgensen E, Keiding N, Grandjean P, & Weihe P (2002). Estimation of health effects of prenatal methylmercury exposure using structural equation models. Environmental Health, 1(1), 1–22. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Carpenter B, Gelman A, Hoffman MD, Lee D, Goodrich B, Betancourt M, Brubaker M, Guo J, Li P, & Riddell A (2017). STAN: A probabilistic programming language. Journal of Statistical Software, 76(1). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Carter RC, Jacobson JL, Molteno CD, Dodge NC, Meintjes EM, & Jacobson SW (2016). Fetal alcohol growth restriction and cognitive impairment. Pediatrics, 138(2), e20160775. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Carvalho CM, Polson NG, & Scott JG (2009). Handling sparsity via the horseshoe. In Proceedings of the 12th International Conference on Artificial Intelligence and Statistics (pp. 73–80). PMLR. [Google Scholar]
- Dellaportas P, Forster JJ, & Ntzoufras I (2002). On Bayesian model and variable selection using MCMC. Statistics and Computing, 12(1), 27–36. [Google Scholar]
- Dunson DB (2000). Bayesian latent variable models for clustered mixed outcomes. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 62(2), 355–366. [Google Scholar]
- Fan Y, & Li R (2012). Variable selection in linear mixed effects models. Annals of Statistics, 40(4), 2043. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Figueiredo MAT (2003). Adaptive sparseness for supervised learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 25(9), 1150–1159. [Google Scholar]
- George EI, & McCulloch RE (1993). Variable selection via Gibbs sampling. Journal of the American Statistical Association, 88(423), 881–889. [Google Scholar]
- Hastie T, Tibshirani R, & Tibshirani R (2020). Best subset, forward stepwise or Lasso? Analysis and recommendations based on extensive comparisons. Statistical Science, 35(4), 579–592. [Google Scholar]
- Hoyme HE, Kalberg WO, Elliott AJ, Blankenship J, Buckley D, Marais A-S, Manning MA, Robinson LK, Adam MP, Abdul-Rahman O, Jewett T, Coles CD, Chambers C, Jones KL, Adnams CM, Shah PE, Riley EP, Charness ME, Warren KR, & May PA (2016). Updated clinical guidelines for diagnosing fetal alcohol spectrum disorders. Pediatrics, 138(2), e20154256. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hoyme HE, May PA, Kalberg WO, Kodituwakku P, Gossage JP, Trujillo PM, Buckley DG, Miller JH, Aragon AS, Khaole N, Viljoen DL, Jones KL, & Robinson LK (2005). A practical clinical approach to diagnosis of fetal alcohol spectrum disorders: Clarification of the 1996 Institute of Medicine criteria. Pediatrics, 115(1), 39–47. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jacobson JL, Akkaya-Hocagil T, Ryan LM, Dodge NC, Richardson GA, Olson HC, Coles CD, Day NL, Cook RJ, & Jacobson SW (2021). Effects of prenatal alcohol exposure on cognitive and behavioral development: Findings from a hierarchical meta-analysis of data from six prospective longitudinal US cohorts. Alcoholism: Clinical and Experimental Research, 45(10), 2040–2058. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jacobson JL, & Jacobson SW (1999). Drinking moderately and pregnancy: Effects on child development. Alcohol Research & Health, 23(1), 25–30. [PMC free article] [PubMed] [Google Scholar]
- Jacobson JL, & Jacobson SW (2002). Effects of prenatal alcohol exposure on child development. Alcohol Research & Health, 26(4), 282–286. [PMC free article] [PubMed] [Google Scholar]
- Jacobson SW, Jacobson JL, Sokol RJ, Chiodo LM, & Corobana R (2004). Maternal age, alcohol abuse history, and quality of parenting as moderators of the effects of prenatal alcohol exposure on 7.5-year intellectual function. Alcoholism: Clinical and Experimental Research, 28(11), 1732–1745. [DOI] [PubMed] [Google Scholar]
- Jacobson SW, Stanton ME, Molteno CD, Burden MJ, Fuller DS, Hoyme HE, Robinson LK, Khaole N, & Jacobson JL (2008). Impaired eyeblink conditioning in children with fetal alcohol syndrome. Alcoholism: Clinical and Experimental Research, 32(2), 365–372. [DOI] [PubMed] [Google Scholar]
- Kass RE, & Raftery AE (1995). Bayes factors. Journal of the American Statistical Association, 90(430), 773–795. [Google Scholar]
- Kuo L, & Mallick B (1998). Variable selection for regression models. Sankhyā: The Indian Journal of Statistics, Series B, 60(1), 65–81. [Google Scholar]
- Lefkopoulou M, Moore D, & Ryan L (1989). The analysis of multiple correlated binary outcomes: Application to rodent teratology experiments. Journal of the American Statistical Association, 84(407), 810–815. [Google Scholar]
- Lefkopoulou M, & Ryan L (1993). Global tests for multiple binary outcomes. Biometrics, 975–988. [PubMed] [Google Scholar]
- Liang K-Y, & Zeger SL (1986). Longitudinal data analysis using generalized linear models. Biometrika, 73(1), 13–22. [Google Scholar]
- Mattson SN, Bernes GA, & Doyle LR (2019). Fetal alcohol spectrum disorders: A review of the neurobehavioral deficits associated with prenatal alcohol exposure. Alcoholism: Clinical and Experimental Research, 43(6), 1046–1062. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Meuwissen T. heoH. E., & Goddard ME (2004). Mapping multiple QTL using linkage disequilibrium and linkage analysis information and multitrait data. Genetics Selection Evolution, 36(3), 261–279. [DOI] [PMC free article] [PubMed] [Google Scholar]
- O’Hara RB, & Sillanpää MJ (2009). A review of Bayesian variable selection methods: What, how and which. Bayesian Analysis, 4(1), 85–117. [Google Scholar]
- Park T, & Casella G (2008). The Bayesian lasso. Journal of the American Statistical Association, 103(482), 681–686. [Google Scholar]
- Plummer M (2003). JAGS: A program for analysis of Bayesian graphical models using Gibbs sampling. In Proceedings of the 3rd International Workshop on Distributed Statistical Computing (pp. 1–10). [Google Scholar]
- Ryan L (2008). Combining data from multiple sources, with applications to environmental risk assessment. Statistics in Medicine, 27(5), 698–710. [DOI] [PubMed] [Google Scholar]
- Sammel M, Lin X, & Ryan L (1999). Multivariate linear mixed models for multiple outcomes. Statistics in Medicine, 18(17–18), 2479–2492. [DOI] [PubMed] [Google Scholar]
- Sammel MD, & Ryan LM (2002). Effects of covariance misspecification in a latent variable model for multiple outcomes. Statistica Sinica, 1207–1222. [Google Scholar]
- Sánchez BN, Budtz-Jørgensen E, Ryan LM, & Hu H (2005). Structural equation models: a review with applications to environmental epidemiology. Journal of the American Statistical Association, 100(472), 1443–1455. [Google Scholar]
- Schwarz G (1978). Estimating the dimension of a model. The Annals of Statistics, 461–464. [Google Scholar]
- Spiegelhalter DJ, Best NG, Carlin BP, & Linde AVD (2002). Bayesian measures of model complexity and fit. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 64(4), 583–639. [Google Scholar]
- Thurston SW, Ruppert D, & Davidson PW (2009). Bayesian models for multiple outcomes nested in domains. Biometrics, 65(4), 1078–1086. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Turk DC, Dworkin RH, McDermott MP, Bellamy N, Burke LB, Chandler JM, Cleeland CS, Cowan P, Dimitrova R, Farrar JT, Hertz S, Heyse JF, lyengar S, Jadad AR, Jay GW, Jermano JA, Katz NP, Manning DC, Martin S, … Witter J (2008). Analyzing multiple endpoints in clinical trials of pain treatments: IMMPACT recommendations. Pain, 139(3), 485–493. [DOI] [PubMed] [Google Scholar]
- van den Noortgate W, López-López JA, Marín-Martínez F, & Sánchez-Meca J (2015). Meta-analysis of multiple outcomes: A multilevel approach. Behavior Research Methods, 47(4), 1274–1294. [DOI] [PubMed] [Google Scholar]
- van Erp S, Oberski DL, & Mulder J (2019). Shrinkage priors for Bayesian penalized regression. Journal of Mathematical Psychology, 89, 31–50. [Google Scholar]
- Watanabe S, & Opper M (2010). Asymptotic equivalence of Bayes cross validation and widely applicable information criterion in singular learning theory. Journal of Machine Learning Research, 11(12). [Google Scholar]
- Wechsler D (1991). Manual for the Wechsler intelligence scale for children (3rd ed.). Psychological Corporation. [Google Scholar]
- Yang M, Wang M, & Dong G (2020). Bayesian variable selection for mixed effects model with shrinkage prior. Computational Statistics, 35(1), 227–243. [Google Scholar]